Support Vibha’s Dream Mile event

My friend Mr. Balaji volunteers for Vibha, a non-profit  organization whose mission is to ensure that every underprivileged child attains his or her right to education, health and opportunity. Vibha, which was founded in 1991 has a volunteer network of 825 members spread across Atlanta, Austin, Bay Area, Boston, Chicago, Dallas, Houston, Jacksonville, Los Angeles,…

Read More

Understanding an OFDM transmission

Let us try to understand simulation of a typical Orthogonal Frequency Division Multiplexing (OFDM) transmission defined per IEEE 802.11a specification. Orthogonal pulses In a previous post (here ), we have understood that the minimum frequency separation for two sinusoidals with arbitrary phases to be orthogonal is , where is the symbol period. In Orthogonal Frequency…

Read More

Migration to new template (skin)

Hi, Those visiting the blog might have noticed a fresh look to the dspLog. This new feel is thanks to the Thesis Magazine Skin provided by FourBlogger Skins. Click here to view more details. There some more tinkering required at some places. But, in general most of the settings are taken care. Hope you like…

Read More

BPSK BER with OFDM modulation

Oflate, I am getting frequent requests for bit error rate simulations using OFDM (Orthogonal Frequency Division Multiplexing) modulation. In this post, we will discuss a simple OFDM transmitter and receiver, find the relation between Eb/No (Bit to Noise ratio) and Es/No (Signal to Noise ratio) and compute the bit error rate with BPSK.

Read More

Viterbi decoder

Coding is a technique where redundancy is added to original bit sequence to increase the reliability of the communication. Lets discuss a simple binary convolutional coding scheme at the transmitter and the associated Viterbi (maximum likelihood) decoding scheme at the receiver. Update: For some reason, the blog is unable to display the article which discuss…

Read More

Stochastic Gradient Descent

For curve fitting using linear regression, there exists a minor variant of Batch Gradient Descent algorithm, called Stochastic Gradient Descent. In the Batch Gradient Descent, the parameter vector  is updated as, . (loop over all elements of training set in one iteration) For Stochastic Gradient Descent, the vector gets updated as, at each iteration the…

Read More