GATE-2012 ECE Q11 (signals)

Question 11 on signals from GATE (Graduate Aptitude Test in Engineering) 2012 Electronics and Communication Engineering paper. Q11. The unilateral Laplace transform of is . The unilateral Laplace transform ofis (A)  (B)  (C)  (D)  Solution From the definition of Laplace transform for a function defined for all real numbers  is,  , where  with real numbers  and . To find the Laplace…

Read More

Batch Gradient Descent

I happened to stumble on Prof. Andrew Ng’s Machine Learning classes which are available online as part of Stanford Center for Professional Development. The first lecture in the series discuss the topic of fitting parameters for a given data set using linear regression.  For understanding this concept, I chose to take data from the top…

Read More

Symbol Error Rate (SER) for QPSK (4-QAM) modulation

Given that we have discussed symbol error rate probability for a 4-PAM modulation, let us know focus on finding the symbol error probability for a QPSK (4-QAM) modulation scheme. Background Consider that the alphabets used for a QPSK (4-QAM) is (Refer example 5-35 in [DIG-COMM-BARRY-LEE-MESSERSCHMITT]). Download free e-Book discussing theoretical and simulated error rates for…

Read More

Scaling factor in QAM

When QAM (Quadrature Amplitude Modulation) is used, typically one may find a scaling factor associated with the constellation mapping operation. It may be reasonably obvious that this scaling factor is for normalizing the average energy to one. This post attempts to compute the average energy of the 16-QAM, 64-QAM and M-QAM constellation (where is a…

Read More

2nd order sigma delta modulator

In a previous post, the variance of the in-band quantization noise for a first order sigma delta modulator was derived. Taking it one step furhter, let us find the variance of the quantization noise filtered by a second order filter. With a first order filter, the quantization noise passes through a system with transfer function…

Read More

Closed form solution for linear regression

In the previous post on Batch Gradient Descent and Stochastic Gradient Descent, we looked at two iterative methods for finding the parameter vector  which minimizes the square of the error between the predicted value  and the actual output  for all  values in the training set. A closed form solution for finding the parameter vector  is possible, and in this post…

Read More