|
| 1 | +--- |
| 2 | +layout: definition |
| 3 | +mathjax: true |
| 4 | + |
| 5 | +author: "Joram Soch" |
| 6 | +affiliation: "BCCN Berlin" |
| 7 | +e_mail: "joram.soch@bccn-berlin.de" |
| 8 | +date: 2021-10-27 07:07:00 |
| 9 | + |
| 10 | +title: "Simple linear regression" |
| 11 | +chapter: "Statistical Models" |
| 12 | +section: "Univariate normal data" |
| 13 | +topic: "Simple linear regression" |
| 14 | +definition: "Definition" |
| 15 | + |
| 16 | +sources: |
| 17 | + - authors: "Wikipedia" |
| 18 | + year: 2021 |
| 19 | + title: "Simple linear regression" |
| 20 | + in: "Wikipedia, the free encyclopedia" |
| 21 | + pages: "retrieved on 2021-10-27" |
| 22 | + url: "https://en.wikipedia.org/wiki/Simple_linear_regression#Fitting_the_regression_line" |
| 23 | + |
| 24 | +def_id: "D163" |
| 25 | +shortcut: "slr" |
| 26 | +username: "JoramSoch" |
| 27 | +--- |
| 28 | + |
| 29 | + |
| 30 | +**Definition:** Let $y$ and $x$ be two $n \times 1$ vectors. |
| 31 | + |
| 32 | +Then, a statement asserting a linear relationship between $x$ and $y$ |
| 33 | + |
| 34 | +$$ \label{eq:slr-model} |
| 35 | +y = \beta_0 + \beta_1 x + \varepsilon \; , |
| 36 | +$$ |
| 37 | + |
| 38 | +together with a statement asserting a [normal distribution](/D/mvn) for $\varepsilon$ |
| 39 | + |
| 40 | +$$ \label{eq:slr-noise} |
| 41 | +\varepsilon \sim \mathcal{N}(0, \sigma^2 V) |
| 42 | +$$ |
| 43 | + |
| 44 | +is called a univariate simple regression model or simply, "simple linear regression". |
| 45 | + |
| 46 | +* $y$ is called "dependent variable", "measured data" or "signal"; |
| 47 | + |
| 48 | +* $x$ is called "independent variable", "predictor" or "covariate"; |
| 49 | + |
| 50 | +* $V$ is called "covariance matrix" or "covariance structure"; |
| 51 | + |
| 52 | +* $\beta_1$ is called "slope of the [regression line](/D/regline)"; |
| 53 | + |
| 54 | +* $\beta_0$ is called "intercept of the [regression line](/D/regline)"; |
| 55 | + |
| 56 | +* $\varepsilon$ is called "noise", "errors" or "error terms"; |
| 57 | + |
| 58 | +* $\sigma^2$ is called "noise variance" or "error variance"; |
| 59 | + |
| 60 | +* $n$ is the number of observations. |
| 61 | + |
| 62 | +When the covariance structure $V$ is equal to the $n \times n$ identity matrix, this is called simple linear regression with independent and identically distributed (i.i.d.) observations: |
| 63 | + |
| 64 | +$$ \label{eq:mlr-noise-iid} |
| 65 | +V = I_n \quad \Rightarrow \quad \varepsilon \sim \mathcal{N}(0, \sigma^2 I_n) \quad \Rightarrow \quad \varepsilon_i \overset{\text{i.i.d.}}{\sim} \mathcal{N}(0, \sigma^2) \; . |
| 66 | +$$ |
| 67 | + |
| 68 | +In this case, the linear regression model can also be written as |
| 69 | + |
| 70 | +$$ \label{eq:slr-model-sum} |
| 71 | +y_i = \beta_0 + \beta_1 x_i + \varepsilon_i, \; \varepsilon_i \sim \mathcal{N}(0, \sigma^2) \; . |
| 72 | +$$ |
| 73 | + |
| 74 | +Otherwise, it is called simple linear regression with correlated observations. |
0 commit comments