Closed form solution for linear regression

In the previous post on Batch Gradient Descent and Stochastic Gradient Descent, we looked at two iterative methods for finding the parameter vector  which minimizes the square of the error between the predicted value  and the actual output  for all  values in the training set. A closed form solution for finding the parameter vector  is possible, and in this post…

Read More

Batch Gradient Descent

I happened to stumble on Prof. Andrew Ng’s Machine Learning classes which are available online as part of Stanford Center for Professional Development. The first lecture in the series discuss the topic of fitting parameters for a given data set using linear regression.  For understanding this concept, I chose to take data from the top…

Read More

Blog on DSP

I happened to visit ‘The Digital Signal Processing Blog’, maintained by Mr. Andres Kwasinski, Ph. D. In the blog one can find details about the upcoming IEEE conferences pertaining to communication and multimedia processing. Further, in some of the posts, author shares his thoughts on topics like fixed point arithmetic (here) and wavelets (here) etc….

Read More

2nd order sigma delta modulator

In a previous post, the variance of the in-band quantization noise for a first order sigma delta modulator was derived. Taking it one step furhter, let us find the variance of the quantization noise filtered by a second order filter. With a first order filter, the quantization noise passes through a system with transfer function…

Read More