How to compute R-squared for a simple linear model
Description
Let’s say we have fit a linear model to two columns of data, one for a single independent variable $x$ and the other for a single dependent variable $y$. How can we compute $R^2$ for that model, to measure its goodness of fit?
Related tasks:
Solution, in Julia
We assume you have already fit a linear model to the data, as in the code below, which is explained fully in a separate task, how to fit a linear model to two columns of data.
1
2
3
4
5
using GLM, DataFrames
xs = [ 393, 453, 553, 679, 729, 748, 817 ]
ys = [ 24, 25, 27, 36, 55, 68, 84 ]
data = DataFrame( xs=xs, ys=ys )
model = lm( @formula( ys ~ xs ), data )
1
2
3
4
5
6
7
8
9
10
11
StatsModels.TableRegressionModel{LinearModel{GLM.LmResp{Vector{Float64}}, GLM.DensePredChol{Float64, LinearAlgebra.CholeskyPivoted{Float64, Matrix{Float64}, Vector{Int64}}}}, Matrix{Float64}}
ys ~ 1 + xs
Coefficients:
───────────────────────────────────────────────────────────────────────────
Coef. Std. Error t Pr(>|t|) Lower 95% Upper 95%
───────────────────────────────────────────────────────────────────────────
(Intercept) -37.3214 18.9954 -1.96 0.1066 -86.1508 11.5079
xs 0.13272 0.029589 4.49 0.0065 0.0566587 0.20878
───────────────────────────────────────────────────────────────────────────
You can get the $R^2$ value from your model using the r2
function in the GLM package.
1
r2( model )
1
0.8009488239830588
Content last modified on 24 July 2023.
See a problem? Tell us or edit the source.
Using SciPy, in Python
We assume you have already fit a linear model to the data, as in the code below, which is explained fully in a separate task, how to fit a linear model to two columns of data.
1
2
3
4
import scipy.stats as stats
xs = [ 393, 453, 553, 679, 729, 748, 817 ]
ys = [ 24, 25, 27, 36, 55, 68, 84 ]
model = stats.linregress( xs, ys )
The $R$ value is part of the model object that stats.linregress
returns.
1
model.rvalue
1
0.8949574425541466
You can compute $R^2$ just by squaring it.
1
model.rvalue ** 2
1
0.8009488239830586
Content last modified on 24 July 2023.
See a problem? Tell us or edit the source.
Solution, in R
We assume you have already fit a linear model to the data, as in the code below, which is explained fully in a separate task, how to fit a linear model to two columns of data.
1
2
3
xs <- c( 393, 453, 553, 679, 729, 748, 817 )
ys <- c( 24, 25, 27, 36, 55, 68, 84 )
model <- lm( ys ~ xs )
You can get a lot of information about your model from its summary.
1
summary( model )
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
Call:
lm(formula = ys ~ xs)
Residuals:
1 2 3 4 5 6 7
9.163 2.199 -9.072 -16.795 -4.431 6.047 12.890
Coefficients:
Estimate Std. Error t value Pr(>|t|)
(Intercept) -37.32142 18.99544 -1.965 0.10664
xs 0.13272 0.02959 4.485 0.00649 **
---
Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1
Residual standard error: 11.62 on 5 degrees of freedom
Multiple R-squared: 0.8009, Adjusted R-squared: 0.7611
F-statistic: 20.12 on 1 and 5 DF, p-value: 0.006486
In particular, it contains the $R^2$ value.
1
summary( model )$r.squared
1
[1] 0.8009488
Content last modified on 24 July 2023.
See a problem? Tell us or edit the source.
Topics that include this task
Opportunities
This website does not yet contain a solution for this task in any of the following software packages.
- Excel
If you can contribute a solution using any of these pieces of software, see our Contributing page for how to help extend this website.