Solved!

SOLVED the Rubik’s cube !!!   After 6 months, 2 cube’s and countless twists and turns, extremely glad to reach here. Will enjoy the beauty of the solved cube for couple of days before breaking it and going over the whole journey again…. (Thanks dear Kunju for introducing me to the cube) Disclosure : After solving…

Read More

Convolutional code

Coding is a technique where redundancy is added to original bit sequence to increase the reliability of the communication. In this article, lets discuss a simple binary convolutional coding scheme at the transmitter and the associated Viterbi (maximum likelihood) decoding scheme at the receiver. Update: For some reason, the blog is unable to display the…

Read More

Support Vibha’s Dream Mile event

My friend Mr. Balaji volunteers for Vibha, a non-profit  organization whose mission is to ensure that every underprivileged child attains his or her right to education, health and opportunity. Vibha, which was founded in 1991 has a volunteer network of 825 members spread across Atlanta, Austin, Bay Area, Boston, Chicago, Dallas, Houston, Jacksonville, Los Angeles,…

Read More

Scaling factor in QAM

When QAM (Quadrature Amplitude Modulation) is used, typically one may find a scaling factor associated with the constellation mapping operation. It may be reasonably obvious that this scaling factor is for normalizing the average energy to one. This post attempts to compute the average energy of the 16-QAM, 64-QAM and M-QAM constellation (where is a…

Read More

Selection Diversity

This is the first post in the series discussing receiver diversity in a wireless link. Receiver diversity is a form of space diversity, where there are multiple antennas at the receiver. The presence of receiver diversity poses an interesting problem – how do we use ‘effectively‘ the information from all the antennas to demodulate the…

Read More

Stochastic Gradient Descent

For curve fitting using linear regression, there exists a minor variant of Batch Gradient Descent algorithm, called Stochastic Gradient Descent. In the Batch Gradient Descent, the parameter vector  is updated as, . (loop over all elements of training set in one iteration) For Stochastic Gradient Descent, the vector gets updated as, at each iteration the…

Read More

Blog on DSP

I happened to visit ‘The Digital Signal Processing Blog’, maintained by Mr. Andres Kwasinski, Ph. D. In the blog one can find details about the upcoming IEEE conferences pertaining to communication and multimedia processing. Further, in some of the posts, author shares his thoughts on topics like fixed point arithmetic (here) and wavelets (here) etc….

Read More