Batch Gradient Descent

I happened to stumble on Prof. Andrew Ng’s Machine Learning classes which are available online as part of Stanford Center for Professional Development. The first lecture in the series discuss the topic of fitting parameters for a given data set using linear regression.  For understanding this concept, I chose to take data from the top…

Read More

GATE-2012 ECE Q13 (circuits)

Question 13 on analog electronics from GATE (Graduate Aptitude Test in Engineering) 2012 Electronics and Communication Engineering paper. Q13. The diodes and the capacitors in the circuit shown are ideal. The voltage  across the diode  is (A)  (B)   (C)  (D) Solution The first half of the circuit is a negative clamper circuit and the second half…

Read More

Stochastic Gradient Descent

For curve fitting using linear regression, there exists a minor variant of Batch Gradient Descent algorithm, called Stochastic Gradient Descent. In the Batch Gradient Descent, the parameter vector  is updated as, . (loop over all elements of training set in one iteration) For Stochastic Gradient Descent, the vector gets updated as, at each iteration the…

Read More

GATE-2012 ECE Q7 (digital)

Question 7 on digital from GATE (Graduate Aptitude Test in Engineering) 2012 Electronics and Communication Engineering paper. Q7. The output Y of a 2-bit comparator is logic 1 whenever the 2 bit input A is greater than 2 bit input B. The number of combinations for which output is logic 1 is (A) 4 (B)…

Read More

GATE-2012 ECE Q24 (math)

Question 24 on math from GATE (Graduate Aptitude Test in Engineering) 2012 Electronics and Communication Engineering paper. Q24. Two independent random variables X and Y are uniformly distributed in the interval [-1, 1]. The probability that max[X,Y] is less than 1/2 is (A) 3/4 (B) 9/16 (C) 1/4 (D) 2/3

Read More