Books

Happy holidays! 🙂 Wishing every one merry Christmas and a great year 2009 and beyond. I will list down some of the books which I have on my desk. They help me with the math and simulations Digital Communication: Third Edition, by John R. Barry, Edward A. Lee, David G. Messerschmitt

Read More

GATE-2012 ECE Q24 (math)

Question 24 on math from GATE (Graduate Aptitude Test in Engineering) 2012 Electronics and Communication Engineering paper. Q24. Two independent random variables X and Y are uniformly distributed in the interval [-1, 1]. The probability that max[X,Y] is less than 1/2 is (A) 3/4 (B) 9/16 (C) 1/4 (D) 2/3

Read More

GATE-2012 ECE Q2 (communication)

Question 52 on communication from GATE (Graduate Aptitude Test in Engineering) 2012 Electronics and Communication Engineering paper. Q2. The power spectral density of a real process for positive frequencies is shown below. The values of  and , respectively are (A)  (B)  (C)  (D)  Solution For a wide sense stationary function, the auto-correlation with delay  is defined as,…

Read More

Receive diversity in AWGN

Some among you will be aware that in a wireless link having multiple antenna’s at the receiver (aka receive diversity) improves the bit error rate (BER) performance. In this post, let us try to understand the BER improvement with receive diversity. And, since we are just getting started, let us limit ourselves to additive white…

Read More

Linear to log conversion

In signal processing blocks like power estimation used in digital communication, it may be required to represent the estimate in log scale. This post explains a simple linear to log conversion scheme proposed in the DSP Guru column on DSP Trick: Quick-and-Dirty Logarithms. The scheme makes implementation of a linear to log conversion simple and…

Read More

Stochastic Gradient Descent

For curve fitting using linear regression, there exists a minor variant of Batch Gradient Descent algorithm, called Stochastic Gradient Descent. In the Batch Gradient Descent, the parameter vector  is updated as, . (loop over all elements of training set in one iteration) For Stochastic Gradient Descent, the vector gets updated as, at each iteration the…

Read More