Does your regression model have a low R-squared? That seems like a problem—but it might not be. Learn what a low R-squared does and does not mean for your model.

If your regression model contains independent variables that are statistically significant, a reasonably high R-squared value makes sense. The statistical significance indicates that changes in the independent variables correlate with shifts in the dependent variable. Correspondingly, the good R-squared value signifies that your model explains a good proportion of the variability in the dependent variable.

That seems like a logical combination, right?

However, what if your model has independent variables that are statistically significant but a *low* R-squared value? This combination indicates that the independent variables are correlated with the dependent variable, but they do not explain much of the variability in the dependent variable. Huh?

Over the years, I’ve had many questions about how to interpret this combination. Some people have wondered whether the significant variables are meaningful. Do these results even make sense? Yes, they do!

In this post, I show how to interpret regression models that have significant independent variables but a low R-squared. To do this, I’ll compare regression models with low and high R-squared values so you can really grasp the similarities and differences and what it all means.

**Related post**: When Should I Use Regression Analysis?

## Comparing Regression Models with Low and High R-squared Values

Like many concepts in statistics, it’s so much easier to understand this one using graphs. In fact, research finds that charts are crucial to convey certain information about regression models accurately.

Consequently, I’ll use fitted line plots to illustrate the concepts for models with one independent variable. However, these interpretations remain valid for multiple regression.

Let’s consider two regression models that assess the relationship between Input and Output. In both models, Input is statistically significant. The equations for these models are below:

- Output1 = 44.53 + 2.024*Input
- Output2 = 44.86 + 2.134*Input

These two regression equations are almost exactly equal. If you saw only the equations, you’d think the models are very similar. Now consider that the R-squared for the Output1 model is 14.7% and for Output2 it is 86.5%. The models aren’t as similar as they first appear.

Graphs can really bring the differences to life. Let’s see what these models and data actually look like. In the two graphs below, the scales are the same to make the comparison easier. You can download the CSV data file: HighLowRsquaredData.

Whoa! Did you expect that much of a difference?

To understand how to interpret a regression model with significant independent variables but a low R-squared, we’ll compare the similarities and the differences between these two models.

## Regression Model Similarities

The models are similar in the following ways:

- The equations are nearly equal: Output = 44 + 2 * Input
- Input is significant with a p-value < 0.001

Additionally, the regression lines in both plots provide an unbiased fit to the upward trend in both datasets. They have the same upward slope of 2.

Interpreting a regression coefficient that is statistically significant does not change based on the R-squared value. Both graphs show that if you move to the right on the x-axis by one unit of Input, Output increases on the y-axis by an average of two units. This mean change in output is the same for both models even though the R-squared values are different.

Furthermore, if you enter the same Input value in the two equations, you’ll obtain approximately equal predicted values for Output. For example, an Input of 10 produces predicted values of 66.2 and 64.8. These values represent the predicted *mean* value of the dependent variable.

## Regression Model Differences

The similarities all focus on the mean—the mean change and the mean predicted value. However, the biggest difference between the two models is the *variability* around those means. In fact, I’d guess that the difference in variability is the first thing about the plots that grabbed your attention. Understanding this topic boils down to grasping the separate concepts of central tendency and variability, and how they relate to the distribution of data points around the fitted line.

While the regression coefficients and predicted values focus on the mean, R-squared measures the scatter of the data around the regression lines. That’s why the two R-squared values are so different. For a given dataset, higher variability around the regression line produces a lower R-squared value.

Take a look at the chart with the low R-squared. Even these relatively noisy data have a significant trend. You can see that as the Input value increases, the Output value also increases. This statistically significant relationship between the variables tells us that knowing the value of Input provides information about the value of Output. The difference between the models is the spread of the data points around the predicted mean at any given location along the regression line.

Be sure to keep the low R-squared graph in mind if you need to comprehend a model that has significant independent variables but a low R-squared!

While the two models produce mean predictions that are almost the same, the variability (i.e., the precision) around the predictions is different. I’ll show you how to assess precision using prediction intervals. This method is particularly useful when you have more than one independent variable and can’t graph the models to see the spread of data around the regression line.

## Using Prediction Intervals to See the Variability

A prediction interval is a range where a single new observation is likely to fall given values of the independent variables that you specify. Narrower prediction intervals represent more precise predictions. Most statistical software can calculate prediction intervals.

**Related posts**: Making Predictions with Regression Analysis and Understand Precision in Applied Regression to Avoid Costly Mistakes

The statistical output below displays the fitted values and prediction intervals that are associated with an Input value of 10 for both models. The first output is for the model with the low R-squared.

As I mentioned earlier, the mean predicted values (i.e., the fit) are nearly equal. However, the prediction intervals are very different because they incorporate the variability. The high variability/low R-squared model has a prediction interval of approximately -500 to 630. That’s over 1100 units!

On the other hand, the low variability/high R-squared model has a much narrower prediction interval of roughly -30 to 160, about 190 units.

After seeing the variability in the data, the differing levels of precision should make sense.

## Key Points about Low R-squared Values

Let’s go over the key points.

- Regression coefficients and fitted values represent means.
- R-squared and prediction intervals represent variability.
- You interpret the coefficients for significant variables the same way regardless of the R-squared value.
- Low R-squared values can warn of imprecise predictions.

What can be done about that low R-squared value? That’s the next question I usually hear in this context. Often, the first thought is to add more variables to the model to increase R-squared.

**Related post**: How High Does R-squared Need to Be?

If you can find legitimate predictors, that can work in some cases. However, for every study area there is an inherent amount of unexplainable variability. For instance, studies that attempt to predict human behavior generally have R-squared values less than 50%. People are hard to predict. You can force a regression model to go past this point but it comes at the cost of misleading regression coefficients, p-values, and R-squared.

Adjusted R-squared and predicted R-squared are tools that help you avoid this problem.

If you are mainly interested in understanding the relationships between the variables, the good news is that a low R-squared does not negate the importance of any significant variables. Even with a low R-squared, statistically significant P-values continue to identify relationships and coefficients have the same interpretation. Generally, you have no additional cause to discount these findings.

For more information about choosing the correct regression model, see my post about model specification.

Ethan says

Dear Jim,

I found a strange phenomenon while checking the R square change in my study.

Could you please help identify some potential reasons why I see this phenomenon?

Also, do you think is this normal that could happen while applying the fixed-effect panel model?

Basically, I apply the fixed-effect panel model on my sample (~9 yearly observations per company).

My sample included 3000 companies and 12250 observations.

When I apply the fixed-effect panel and include the year fixed-effect, the R square do not change much between models when I add new regressors. The adjusted R square increase from 0.1938 to 0.1951 when I add 5 more variables into the base model, which can be said almost no increment.

Nevertheless, if I did not apply the year fixed-effect, and only add industry dummies in using fixed-effect model. The adjusted R square change become normal from 0.1741 to 0.1951 from the base model to main model (add 5 more independent variables).

I do not know how to deal with this situation. Do you mind offering some suggestions and explanations on this issue?

Thank you very much in advance.

Best regards,

Ethan

Vicky says

What if you add other independent variables that correlates with the independent variable you use in the model? Will a model with extra such independent variables decrease effect size and increase p-value of this independent variable?

Janet Acke says

Great article – thank you.

How would you answer the question ‘how much of the variation is explained by these 2 different factors’? If one of the explanatory variables had lots of variation between observations and the other one didn’t, in trying to explain for example wage levels.

If you had 4 individuals who all had very similar levels of schooling but very different levels of IQ, and the dependent variable is wage levels. Would it be best to run a regression on each individually and see which has the higher coefficient, or the higher R^2. Or run them together and look at coefficients?

Say schooling is very significant but doesn’t vary much, would it be fair to say that “schooling explains much of the difference in wages”. Or vice versa and something is hardly significant (lower coefficient, or indeed insignificant) but it varies a lot between observations.

Or is there another value aside from r^2 or the coefficient which captures this idea of variance in both X and Y, as well as variation between them?

Is there a single value that quantifies the notion of ‘how much variation in Y is explained by X1 vs X2… etc’

Thanks so much!

Jim Frost says

Hi Janet,

I’ve written an article that I think will help you out: Identifying the Most Important Variables in Regression Models. It talks about some of the issues you mention and ways to answer that question. If you have more questions after reading that one, please post them in the comments for that post. Thanks!

And, I’ll just add that basing conclusions on four people is an extremely small sample size. I would think it would be hard to draw any firm conclusions from such a small sample.

Akanksha says

Thank you for your easy and to the point concept clarity articles. You are awesome. Every small and big term and concept is explained so intricately.

sajad ahmad mir says

which is true: if r squared increases, variable is significant. if r squared decreases variable is not significant

Jim Frost says

Hi Sajad,

Neither statement is necessarily true. The first can be true or false. If you add a variable to a model, R-squared always increases by some amount whether or not the variable is significant. It does not mean the variable is significant. The second statement is false because the R-squared never decreases when you add a variable even when it is not significant.

If you’re talking about adjusted R-squared, the answers are slightly different.

Matt G says

Hi Jim,

What does it mean when the ANOVA table has reported the model to be significant, but the individual coefficient variables are insignificant? Does this suggest that the independent variables cannot significantly predict the outcome of the dependent variable, but they can better predict the dependent variable once combined into a model?

Many thanks.

Jim Frost says

Hi Matt,

I write about that in my post about the overall F-test. I think you’ll find your answer there. If you more questions after reading that, don’t hesitate to post them!

Zakia Siddiqui says

Dear Sir,

I hope you are doing well.

It seems my regression model isn’t significant (R square is 0.18, Adjusted R square is 0.12, F is 3.28 and p is 0.13 (Intercept), right? However, beta is 1.36 and p-value of the IV is 0.09, Can you please suggest how to do the interpretation?

Thanks in advance.

Jim Frost says

Hi Zakia,

That f-value and it’s corresponding p-value is for the overall significance of your regression model. Clink that link to learn how to interpret it.

From the from what you write, it looks like neither your model nor the IV is significant. Click to learn how to interpret regression coefficients and their p-values.

Sheran says

Hello Sir, What does a null model means in simpler terms? and what does a statistically significant model means?

Jim Frost says

Hi Sheran, I discuss null models and what a statistically significant model means in my post about the overall F-test for regression.

Simon S says

Dear Jim!

We are writhing a master thesis, and we are having the same problems of the adjusted R-square being too low for most our independent variables. What are the minimum requirements for the of beta, R-square and significance levels? Do we also need to calculate the statistical power of our test?

Thank you in advance!

mandakini das says

sir,

My R squared value for moderator variables is 22% only which is low. But my main effects and interaction effects (moderator) statistically significant. Is the model is good one?

Deepa Chandwani says

Hi sir..

Thanks a lot for the explaination.

Please help me with my query.

There are two independent variables in my study where my hypothesis is one has positive impact i.e. direct relationship and the other has negative impact i.e. inverse relationship with the dependent variable.

i carried out multiple regression in Microsoft excel. R2 was 0.19.. F value = 3.79 and Signifance F =0.03

But since the relationship is different , how should i interpret p values…

If multiple regression is not to be done then which test should be carried out?

Thanks SIR.

Jim Frost says

Hi Deepa,

It certainly sounds like the right type of problem for multiple regression. You don’t state what the p-values for the independent variables are? It looks like overall you have a significant model. It sounds like you have two IVs. As one of the IV increases, the mean DV also tends to increase. For other IV, as it increases the mean of the DV tends to decrease. Although, without the p-values for the IVs, I don’t know if both of those relationships are statistically significant.

For more information, read my post about how to interpret regression coefficients and their p-values.

I’m not sure exactly what you’re asking?

Grace Ison says

Thank you so much, Mr. Frost! This is really a great help for my study. Thanks again! 🙂

Jim Frost says

You’re very welcome, Grace. Best of luck with your studies! 🙂

Grace Ison says

Hi Mr. Frost! Thank you for your prompt reply. This helps me a lot in my thesis. I just have one more question. How about if there is a moderating variable like for example, ‘age’ in the multiple regression. Should I also state it this way: “Is there a significant relationship between the independent and dependent variable when age is used as test factor?” or is there another way of stating the research question? Thank you so much!

Jim Frost says

Hi again Grace, a moderating variable exists when there is a significant interaction effect. Suppose you have DV Y, IV X, and a moderating variable M. If there is a statistically significant interaction between X*M, then it indicates that the relationship between X and Y changes based on the value of M. Or, posed as a question, does the relationship between X and Y change based on the value of M?

For more information, read my post about interaction effects.

Grace Ison says

Hi, I just have a question. What is the right research question for a multiple regression analysis? Is it, “is there a significant relationship between dependent variable and independent variable” or “is there a significant difference between dependent variable and independent variable”. Thank you so much!

Jim Frost says

Hi Grace,

Regression determines whether there is a significant relationship between an independent variable and the dependent variable. You mention “multiple” regression, which means your model contains at least two IVs. The analysis tests the relationship between each IV and the DV.

leeterhao says

hi sir. I have a question.

If my anova result get the r=0.2195 and my r^2 = 0.04. So how can i interpret the value

Jim Frost says

Hi, you’re r-squared value is extremely low. In fact, it might not be significantly different from zero. Check your F-test of overall significance to check that.

Your model accounts for 4% of the variance in your dependent variable.

Narendra says

Hello Sir,

In my case, Predicted model is adequate as F value is greater than tabulated F value. But I am getting pure error 29.71 and R-squared value is 77.00%. How can I interpret the results? Please help!

pradip says

sir, how we know over all F test is significant when independent variable is not signficant

Jim Frost says

Hi Pradip,

I’m not 100% sure that I understand what you’re asking. However, you can determine whether the overall F-test is significant by comparing its p-value to your significance level. If its p-value is less than your significance level, the overall F-test is significant. You can read more in my post about the overall F-test.

I hope this helps!

Mostafa Haj Hamidi says

I thank you for your efforts and I wish you a healthy health

Mackenzie says

Hi Jim,

Thank you for the quick response and helpful discussion. I really appreciate it!

Mackenzie says

Hi Jim,

How would you interpret a model with a significant (p<.001), but low R-squared (16%) and only one of the seven predictors are significant (β=.23, p <.01).

Jim Frost says

Hi Mackenzie,

Because only one predictor is significant, you have enough evidence to conclude that there is relationship between only that predictor and the response (DV). The low R-squared indicates that there is a lot of variability around the fitted line.

Look at the images of the fitted line plots for the two models in this blog post. Your model more closely resembles the plot for the low R-squared model. You can see that there is a trend, but the distance between the data points and the lines are greater.

A low R-squared like that is a problem only when you’re trying to produce accurate predictions. But, if you’re just want to understand the relationships between variables, you still have that relationship which is statistically significant.

I hope this helps!

Waseem Majeed says

How can we calculate or say that our regression model is significant?

Jim Frost says

Hi Waseem,

You can assess the coefficient p-values and the F-test of overall significance. I’ve written blog posts on both, which you should read.

Also, consider buying my ebook about regression analysis for a more thorough look!

I hope that helps!

lesley l says

HI when reporting results from a linear regression with a t test, ( rsquared has a low value (the x axis is temperature, the y axis is change in species richness) ,and high variabiltiy abut the line of best fit, how do I report these resutls, do I include

I am not sure which numbers to report basically and how to make sense of them once the test has been run .do i include the slope of line, ? the t stat ? how do I incorporate the standard error into reporting statistical evidence

I have coefficient for intercept

and x variable ,SE and t stat values but I don’t know how to use them in my report

Hari Poudel says

Dear Jim,

I need to ask a follow-up question to the above explanation. I run a logistic regression in which one of the IVs is significant across various models. Other IVs are not siginificant. I also ran a postestimation model -test and it also gave me similar kinds of relationships. What would be better way to explain such type of relationships? Thnka you for your time and help.

Jim Frost says

Hi Hari,

It sounds like your sample only provides sufficient evidence to suggest that the one IV has a relationship with the DV in the population. For your other IVs, the sample does not provide sufficient evidence to conclude that those relationships exist in the population. Whatever sample effects you’re seeing for the non-significant IVs might well be random error rather than representing a relationship that exists in the population.

Pranesh Debnath says

Hello Jim nice to see your explanation!!!!

I am giving following data. Kindly Calrify my question…

Multiple R=0.339958

R-Square=0.115571

Adj. R-Square=0.109458

Standard Error=0.221261

Observation (sample size)=875

No. of Independent variables=6

My question: Is there any mathematical relationship with “Standard Error” and other information given above as like R-Square and Adj. R-Square?

Jim Frost says

Hi Pranesh,

I’d say that the standard error of the regression is probably the most closely related to adjusted R-squared. Both goodness-of-fit measures use the adjusted sums of squares, which incorporates the error degrees of freedom. However, it would also be related to R-squared even though it does not make that adjustment. In general, as R-squared and, particularly, adjusted R-squared increase for a particular dataset, the standard error tends to decrease.

Jerry says

This is indeed an excellent explanation of R-squared with good links to definitions and other articles. Thank you! I’d like to know more about prediction intervals though. Is a prediction interval the range of all output values that occur in the data from the specified input value? And can SPSS produce prediction intervals? thanks

Jim Frost says

Hi Jerry,

Prediction intervals aren’t based on the range of output given the input but rather the mean square error, which is a measure of the variability of the residuals. The mean square error is small when the observed values are close to the fitted values–which in turn produces tighter prediction intervals.

As for SPSS, I don’t regularly use that software. However, I checked around and you can create prediction intervals for simple regression at least. I’m not sure about multiple regression. SPSS refers to them as individual confidence intervals in the regression context. You can create them in SPSS scatterplots. Go to the Chart Editor dialog for scatterplots and click the Fit Line tab. Under Confidence Intervals, choose Individual.

I hope that helps!

Chris Spadi says

Hi Jim,

I made scatter plots where x is the year (spanning over 70 years) and Y is the count (fires) for each month in a specific region. My R2 values range between .22 to .62. Showing a positive trend.

Things like increased drought, rising temperature, decreased precipitation, and wind factor into each year.

Is a R2 value range of .22 to .62 necessarily bad in this case?

Thanks,

Chris

Jim Frost says

Hi Chris, it sounds like in addition to year, you have other independent (X) variables? And, the R-squared varies based on the specific combination of those variables you include in the variable?

I’ve noticed that statisticians sometimes want to categorize the strength of the results based on things like the value of R-squared. A well-known one even created ranges of R-squared values and attributed labels such as strong, moderate, and weak.

I don’t think that’s the way to go. Some study areas are inherently more unpredictable than others. When an area is inherently more unpredictable, you’d expect lower R-squared values as a matter of course. You have to adjust your idea of what constitutes a higher R-squared based on the subject area.

So, that all said, those R-squared values aren’t necessarily bad. However, you’ll need to use subject area knowledge and compare your results to similar studies to make that determination. As along as you have significant predictors, your model is supplying important information about the factors that increase fires. One potential downside occurs if you want to use your model to make predictions rather than just understanding the role of underlying factors. Models with R-squared values in the range you mention often produce imprecise predictions (wide prediction intervals).

Best of luck with your analysis!

Ivan says

Hello Jim,

I am trying to predict house prices based on the number of basic house characteristics. R-sq is 0.52, but most of the independent variables have a p-value within 0.2 and 0.99. Does it mean they are all insignificant? How do I know which ones to include and which ones to eliminate? Thank you very much

Jim Frost says

Hi Ivan,

Yes, you those variables are not statistically significant. You’d often remove them from the model. For more information, read my post about specifying the correct regression model.

Keryn says

Hi Jim, if my regression model is not significant (R ² = .011, adjusted R ² = .005, F (3, 574) = 2.031, p = .108). But my beta value for one of the IV is significant (just) beta =.082, p=.048, how do I interpret this?

Jim Frost says

Hi Keryn,

It can happen that the overall significance doesn’t necessarily match the fact of whether there are any significant independent variables, such as in your model. These tests usually agree. If you have a significant IV, you usually obtain a significant overall test of significance. But, when you have borderline results, like in your model, they can disagree. For more information, read my post about the overall F-test of significance.

As for having a significant variable but a very low R-squared, interpret it exactly as I describe in this post. The relationship between the IV and the DV is statistically significant. In other words, knowing the value of the IV provides some information about the DV. However, there is a lot variability around the fitted values that your model doesn’t explain.

A couple of suggestions for you. Given the extremely low R-squared value, I’d double check the residual plots to be sure that they look good. If you see a non-random pattern in the residual plots, you might be able to improve your model.

Also, given the borderline nature of the p-value for your IV, combined with the non-significant overall test and low R-squared, you should consider these results as suggestive or preliminary.

Best of luck with your analysis!

Alvinn-emmanuel Yao says

Hello Sir,

I have a question. if we have a low R squared value but the P value of F is significant, is the model a good one?

Jim Frost says

Hi,

The first thing you should check is to determine whether you have an independent variable (IV) that has a significant p-value. If you do, then I’d say that the model is probably a good one (always assuming that your residuals look good). You can link an effect to a specific IV or IVs.

If you don’t have an IV that is significant but the overall F-test is significant, it gets a little tricky. In this case, you don’t have sufficient evidence to conclude that any particular IV has a statistically significant relationship with the dependent variable (DV). However, all the IVs in the model taken together have a significant relationship with the DV. Unfortunately, that relationship, as measured by the low R-squared, is fairly weak. You’re pretty much at the minimum limits of useful knowledge in this scenario. You can’t pinpoint the effect to specific IVs and it’s a weak effect to boot. I’d say that a study like this potentially provides evidence that some effect is present but you’d need additional, larger studies to really learn something useful.

So, first step is to look at the p-values for the IVs to see which general scenario I describe above your model fits in. Then, take it from there!

I hope this helps!

Thomas Mkhabela says

Very precise, i really appreciate the answers.

Jim Frost says

Hi Thomas, Thanks! I’m glad it was helpful!

Mark Davis says

Very clear and concise article. You did a great job helping me figure out the right questions to ask!

Abhisek Guha says

Excellent concept clearing article.

Jim Frost says

Thank you, Abhisek!