Books

Happy holidays! 🙂 Wishing every one merry Christmas and a great year 2009 and beyond. I will list down some of the books which I have on my desk. They help me with the math and simulations Digital Communication: Third Edition, by John R. Barry, Edward A. Lee, David G. Messerschmitt

Read More

Bit Error Rate (BER) for frequency shift keying with coherent demodulation

Following the request by Siti Naimah, this post discuss the bit error probability for coherent demodulation of binary Frequency Shift Keying (BFSK) along with a small Matlab code snippet. Using the definition provided in Sec 4.4.4 of [DIG-COMM-SKLAR]), in binary Frequency shift keying (BFSK), the bits 0’s and 1’s are represented by signals and having…

Read More

Scaling factor in QAM

When QAM (Quadrature Amplitude Modulation) is used, typically one may find a scaling factor associated with the constellation mapping operation. It may be reasonably obvious that this scaling factor is for normalizing the average energy to one. This post attempts to compute the average energy of the 16-QAM, 64-QAM and M-QAM constellation (where is a…

Read More

Deriving PDF of Rayleigh random variable

In the post on Rayleigh channel model, we stated that a circularly symmetric random variable is of the form , where real and imaginary parts are zero mean independent and identically distributed (iid) Gaussian random variables. The magnitude which has the probability density, is called a Rayleigh random variable. Further, the phase is uniformly distributed from…

Read More

Chi Square Random Variable

While trying to derive the theoretical bit error rate (BER) for BPSK modulation in a Rayleigh fading channel, I realized that I need to discuss chi square random variable prior. What is chi-square random variable? Let there be independent and identically distributed Gaussian random variables with mean and variance and we form a new random…

Read More

Stochastic Gradient Descent

For curve fitting using linear regression, there exists a minor variant of Batch Gradient Descent algorithm, called Stochastic Gradient Descent. In the Batch Gradient Descent, the parameter vector  is updated as, . (loop over all elements of training set in one iteration) For Stochastic Gradient Descent, the vector gets updated as, at each iteration the…

Read More