Get inspired by the success stories of our students in IIT JAM MS, ISI  MStat, CMI MSc DS.  Learn More

# ISI MStat PSB 2013 Problem 4 | Linear Regression This is a sample problem from ISI MStat PSB 2013 Problem 4. It is based on the simple linear regression model, finding the estimates, and MSEs. But think over the "Food for Thought" any kind of discussion will be appreciated. Give it a try!

## Problem- ISI MStat PSB 2013 Problem 4

Consider n independent observation from the model ,

where is normal with mean 0 and variance . Let and be the maximum likelihood estimators of and , respectively. Let and be the estimated values of and , respectively.

(a) What is the estimated mean of Y, when when ? Estimate the mean squared error of this estimator .

(b) What is the predicted value of Y, when when ? Estimate the mean squared error of this predictor .

### Prerequisites

Linear Regression

Method of Least Squares

Maximum likelihood Estimators.

Mean Squared Error.

## Solution :

Here for the given model,

we have , the random errors, , and the maximum likelihood estimators (MLE), of the model parameters are given by and . The interesting thing about this model is, since the random errors are Gaussian Random variables, the Ordinary Least Square Estimates of the model parameters and , are identical to their Maximum Likelihood Estimators, ( which are already given!). How ?? Verify it yourself and once and remember it henceforth.

So, here and there also the OLS estimates of the model parameters respectively.

And By Gauss-Markov Theorem, the OLS estimates of the model parameters are the BLUE (Best Linear Unbiased Estimator), for the model parameters. So, here and are also the unbiased estimators of and respectively.

(a) Now we need to find the estimated mean Y given , is the estimated mean of Y given .

Now since, the given MLEs ( OLSEs) are also unbiased for their respective parameters, = = = So, .

(b) Similarly, when , the predicted value of Y would be, is the predicted value of Y when is given.

Using similar arguments, as in (a) Properties of independence between the model parameters , verify that, . Hence we are done !

## Food For Thought

Now, can you explain Why, the Maximum Likelihood Estimators and Ordinary Least Square Estimates are identical, when the model assumes Gaussian errors ??

Wait!! Not done yet. The main course is served below !!

In a game of dart, a thrower throws a dart randomly and uniformly in a unit circle. Let be the angle between the line segment joining the dart and the center and the horizontal axis, now consider Z be a random variable. When the thrower is lefty , Z=-1 and when the thrower is right-handed , Z=1 . Assume that getting a Left-handed and Right-handed thrower is equally likely ( is it really equally likely, in real scenario ?? ). Can you construct a regression model, for regressing on Z.

Think over it, if you want to discuss, we can do that too !!

## Subscribe to Cheenta at Youtube

This is a sample problem from ISI MStat PSB 2013 Problem 4. It is based on the simple linear regression model, finding the estimates, and MSEs. But think over the "Food for Thought" any kind of discussion will be appreciated. Give it a try!

## Problem- ISI MStat PSB 2013 Problem 4

Consider n independent observation from the model ,

where is normal with mean 0 and variance . Let and be the maximum likelihood estimators of and , respectively. Let and be the estimated values of and , respectively.

(a) What is the estimated mean of Y, when when ? Estimate the mean squared error of this estimator .

(b) What is the predicted value of Y, when when ? Estimate the mean squared error of this predictor .

### Prerequisites

Linear Regression

Method of Least Squares

Maximum likelihood Estimators.

Mean Squared Error.

## Solution :

Here for the given model,

we have , the random errors, , and the maximum likelihood estimators (MLE), of the model parameters are given by and . The interesting thing about this model is, since the random errors are Gaussian Random variables, the Ordinary Least Square Estimates of the model parameters and , are identical to their Maximum Likelihood Estimators, ( which are already given!). How ?? Verify it yourself and once and remember it henceforth.

So, here and there also the OLS estimates of the model parameters respectively.

And By Gauss-Markov Theorem, the OLS estimates of the model parameters are the BLUE (Best Linear Unbiased Estimator), for the model parameters. So, here and are also the unbiased estimators of and respectively.

(a) Now we need to find the estimated mean Y given , is the estimated mean of Y given .

Now since, the given MLEs ( OLSEs) are also unbiased for their respective parameters, = = = So, .

(b) Similarly, when , the predicted value of Y would be, is the predicted value of Y when is given.

Using similar arguments, as in (a) Properties of independence between the model parameters , verify that, . Hence we are done !

## Food For Thought

Now, can you explain Why, the Maximum Likelihood Estimators and Ordinary Least Square Estimates are identical, when the model assumes Gaussian errors ??

Wait!! Not done yet. The main course is served below !!

In a game of dart, a thrower throws a dart randomly and uniformly in a unit circle. Let be the angle between the line segment joining the dart and the center and the horizontal axis, now consider Z be a random variable. When the thrower is lefty , Z=-1 and when the thrower is right-handed , Z=1 . Assume that getting a Left-handed and Right-handed thrower is equally likely ( is it really equally likely, in real scenario ?? ). Can you construct a regression model, for regressing on Z.

Think over it, if you want to discuss, we can do that too !!

## Subscribe to Cheenta at Youtube

This site uses Akismet to reduce spam. Learn how your comment data is processed.

### Knowledge Partner  