Last time | Next time |
One problem is using the wrong equations, and sometimes the units will tell us that we have a problem. Paying attention to units pays!
Just get that to me by the end of the day.
If your group has any questions, feel free to stop by and talk to me!
I'll be assigning groups and towns next Friday, after we discuss the fit to the Keeling data in class.
Hopefully you now understand vector and matrix notation, and a few important operations (e.g. transpose and inverse).
Matrices will be important in other models we consider.
There are three things I always consider about a model:
The standard errors of the parameters pop out of the inverse matrix we compute, multiplied by the mean SSE. Once we have those, we have everything we need for confidence intervals.
We're often interested to know whether we can exclude a certain value from a confidence interval -- e.g., can we conclude that the slope parameter $b$ in a linear regression $y(x)=a+bx$ is not 0? If so, we conclude that there is a non-zero slope, and the model suggests that $x$ drives values up or down, depending on the sign of $b$.
Patrick proposed an exponential model for the Keeling data, but it is non-linear in the parameters: \[ y(t)=ae^{bt} \]
Yet we can still use linear regression to fit it: how so?
Alexander also illustrates that one cannot simply invert the regression equation $y=a+bx$ to get the regression equation $x=\frac{y-a}{b}$. So it really matters in linear regression which variable is considered "independent" and which is considered "dependent".