Multiple regression results help The 2019 Stack Overflow Developer Survey Results Are In Announcing the arrival of Valued Associate #679: Cesar Manara Planned maintenance scheduled April 17/18, 2019 at 00:00UTC (8:00pm US/Eastern)Not-significant F but a significant coefficient in multiple linear regressionWhen to transform predictor variables when doing multiple regression?Combining multiple imputation results for hierarchical regression in SPSSHow to deal with different outcomes between pairwise correlations and multiple regressionProbing effects in a multivariate multiple regressionPredictor flipping sign in regression with no multicollinearityMeta analysis of Multiple regressionWhat is the difference between each predictor's standardized betas (from multiple regression) and it's Pearson's correlation coefficient?Multiple linear regression coefficients meaningWhat is the correct way to follow up a multivariate multiple regression?

How did the audience guess the pentatonic scale in Bobby McFerrin's presentation?

Why can't devices on different VLANs, but on the same subnet, communicate?

What force causes entropy to increase?

How long does the line of fire that you can create as an action using the Investiture of Flame spell last?

Is this wall load bearing? Blueprints and photos attached

Reference for the teaching of not-self

What is this lever in Argentinian toilets?

Working through the single responsibility principle (SRP) in Python when calls are expensive

How to colour the US map with Yellow, Green, Red and Blue to minimize the number of states with the colour of Green

Is it ethical to upload a automatically generated paper to a non peer-reviewed site as part of a larger research?

Is there a writing software that you can sort scenes like slides in PowerPoint?

How is simplicity better than precision and clarity in prose?

Difference between "generating set" and free product?

Would an alien lifeform be able to achieve space travel if lacking in vision?

Can the DM override racial traits?

Why did all the guest students take carriages to the Yule Ball?

ELI5: Why do they say that Israel would have been the fourth country to land a spacecraft on the Moon and why do they call it low cost?

In horse breeding, what is the female equivalent of putting a horse out "to stud"?

University's motivation for having tenure-track positions

I could not break this equation. Please help me

Finding the path in a graph from A to B then back to A with a minimum of shared edges

Does Parliament need to approve the new Brexit delay to 31 October 2019?

The variadic template constructor of my class cannot modify my class members, why is that so?

Do warforged have souls?



Multiple regression results help



The 2019 Stack Overflow Developer Survey Results Are In
Announcing the arrival of Valued Associate #679: Cesar Manara
Planned maintenance scheduled April 17/18, 2019 at 00:00UTC (8:00pm US/Eastern)Not-significant F but a significant coefficient in multiple linear regressionWhen to transform predictor variables when doing multiple regression?Combining multiple imputation results for hierarchical regression in SPSSHow to deal with different outcomes between pairwise correlations and multiple regressionProbing effects in a multivariate multiple regressionPredictor flipping sign in regression with no multicollinearityMeta analysis of Multiple regressionWhat is the difference between each predictor's standardized betas (from multiple regression) and it's Pearson's correlation coefficient?Multiple linear regression coefficients meaningWhat is the correct way to follow up a multivariate multiple regression?



.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty,.everyoneloves__bot-mid-leaderboard:empty margin-bottom:0;








2












$begingroup$


For my first ever research paper I've run a hierarchal multiple linear regression with two predictors and one outcome variable, however I don't understand my results. I've found predictor A to be a significant predictor for my outcome variable alone. However, when both my predictors are in the model, predictor A is not a significant predictor, only predictor B is. How can this be if predictor A was significant in the first model? How does predictor B change how significant predictor A is?



Thank you!










share|cite|improve this question







New contributor




ummmm is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
Check out our Code of Conduct.







$endgroup$


















    2












    $begingroup$


    For my first ever research paper I've run a hierarchal multiple linear regression with two predictors and one outcome variable, however I don't understand my results. I've found predictor A to be a significant predictor for my outcome variable alone. However, when both my predictors are in the model, predictor A is not a significant predictor, only predictor B is. How can this be if predictor A was significant in the first model? How does predictor B change how significant predictor A is?



    Thank you!










    share|cite|improve this question







    New contributor




    ummmm is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
    Check out our Code of Conduct.







    $endgroup$














      2












      2








      2





      $begingroup$


      For my first ever research paper I've run a hierarchal multiple linear regression with two predictors and one outcome variable, however I don't understand my results. I've found predictor A to be a significant predictor for my outcome variable alone. However, when both my predictors are in the model, predictor A is not a significant predictor, only predictor B is. How can this be if predictor A was significant in the first model? How does predictor B change how significant predictor A is?



      Thank you!










      share|cite|improve this question







      New contributor




      ummmm is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.







      $endgroup$




      For my first ever research paper I've run a hierarchal multiple linear regression with two predictors and one outcome variable, however I don't understand my results. I've found predictor A to be a significant predictor for my outcome variable alone. However, when both my predictors are in the model, predictor A is not a significant predictor, only predictor B is. How can this be if predictor A was significant in the first model? How does predictor B change how significant predictor A is?



      Thank you!







      multiple-regression mlr






      share|cite|improve this question







      New contributor




      ummmm is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.











      share|cite|improve this question







      New contributor




      ummmm is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.









      share|cite|improve this question




      share|cite|improve this question






      New contributor




      ummmm is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.









      asked 9 hours ago









      ummmmummmm

      111




      111




      New contributor




      ummmm is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.





      New contributor





      ummmm is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.






      ummmm is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.




















          3 Answers
          3






          active

          oldest

          votes


















          1












          $begingroup$

          regression coefficients reflect the simultaneous effects of multiple predictors. If the two predictors are inter-dependent (i.e. correlated) the results can differ from single input models.






          share|cite|improve this answer









          $endgroup$




















            1












            $begingroup$

            The tests in multiple regression are "added last" tests. That means they test whether the model significantly improves after including the extra variable in a regression that contains all other predictors.



            In your model with no predictors, adding A improves the model, so the test of A is significant in the model with only A.



            In a model with A already in the model, adding B improves the model, so the test of B is significant in the model with A and B. But in a model with B already in the model, adding A doesn't improve the model, so the test of A is not significant in the model with A and B. B is doing all the work that A would do, so adding A doesn't improve the model beyond B.



            As @IrishStat mentioned, this can occur when A and B are correlated (positively or negatively) with each other. It's a fairly common occurrence in regression modeling. The conclusion you might draw is that A predicts the outcome when B is not in the model (i.e., unavailable), but after including B, A doesn't do much more to predict the outcome. Unfortunately, without more information about the causal structure of your variables, there is little more interpretation available.






            share|cite|improve this answer









            $endgroup$




















              0












              $begingroup$

              To expand a little on @Noah and @IrishStat's answers, in a multiple regression, coefficients for each independent variable/predictor are estimated to obtain the direct effect of each variable, using variation unique to that variable and the variable's correlation with the outcome variable, not using variation shared by predictors. (In technical terms, we are talking about variance and covariance of these variables.) The less unique variation there is, the less significant the estimate will become.



              So why, in your example, did you end up with an insignificant predictor A when B was added, and not with a significant predictor A and insignificant predictor B? It is likely because the proportion of variance of predictor A that it has in common with predictor B is larger than the proportion of variance of predictor B that it has in common with predictor A.






              share|cite|improve this answer









              $endgroup$













                Your Answer








                StackExchange.ready(function()
                var channelOptions =
                tags: "".split(" "),
                id: "65"
                ;
                initTagRenderer("".split(" "), "".split(" "), channelOptions);

                StackExchange.using("externalEditor", function()
                // Have to fire editor after snippets, if snippets enabled
                if (StackExchange.settings.snippets.snippetsEnabled)
                StackExchange.using("snippets", function()
                createEditor();
                );

                else
                createEditor();

                );

                function createEditor()
                StackExchange.prepareEditor(
                heartbeatType: 'answer',
                autoActivateHeartbeat: false,
                convertImagesToLinks: false,
                noModals: true,
                showLowRepImageUploadWarning: true,
                reputationToPostImages: null,
                bindNavPrevention: true,
                postfix: "",
                imageUploader:
                brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
                contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
                allowUrls: true
                ,
                onDemand: true,
                discardSelector: ".discard-answer"
                ,immediatelyShowMarkdownHelp:true
                );



                );






                ummmm is a new contributor. Be nice, and check out our Code of Conduct.









                draft saved

                draft discarded


















                StackExchange.ready(
                function ()
                StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f402894%2fmultiple-regression-results-help%23new-answer', 'question_page');

                );

                Post as a guest















                Required, but never shown

























                3 Answers
                3






                active

                oldest

                votes








                3 Answers
                3






                active

                oldest

                votes









                active

                oldest

                votes






                active

                oldest

                votes









                1












                $begingroup$

                regression coefficients reflect the simultaneous effects of multiple predictors. If the two predictors are inter-dependent (i.e. correlated) the results can differ from single input models.






                share|cite|improve this answer









                $endgroup$

















                  1












                  $begingroup$

                  regression coefficients reflect the simultaneous effects of multiple predictors. If the two predictors are inter-dependent (i.e. correlated) the results can differ from single input models.






                  share|cite|improve this answer









                  $endgroup$















                    1












                    1








                    1





                    $begingroup$

                    regression coefficients reflect the simultaneous effects of multiple predictors. If the two predictors are inter-dependent (i.e. correlated) the results can differ from single input models.






                    share|cite|improve this answer









                    $endgroup$



                    regression coefficients reflect the simultaneous effects of multiple predictors. If the two predictors are inter-dependent (i.e. correlated) the results can differ from single input models.







                    share|cite|improve this answer












                    share|cite|improve this answer



                    share|cite|improve this answer










                    answered 8 hours ago









                    IrishStatIrishStat

                    21.4k42342




                    21.4k42342























                        1












                        $begingroup$

                        The tests in multiple regression are "added last" tests. That means they test whether the model significantly improves after including the extra variable in a regression that contains all other predictors.



                        In your model with no predictors, adding A improves the model, so the test of A is significant in the model with only A.



                        In a model with A already in the model, adding B improves the model, so the test of B is significant in the model with A and B. But in a model with B already in the model, adding A doesn't improve the model, so the test of A is not significant in the model with A and B. B is doing all the work that A would do, so adding A doesn't improve the model beyond B.



                        As @IrishStat mentioned, this can occur when A and B are correlated (positively or negatively) with each other. It's a fairly common occurrence in regression modeling. The conclusion you might draw is that A predicts the outcome when B is not in the model (i.e., unavailable), but after including B, A doesn't do much more to predict the outcome. Unfortunately, without more information about the causal structure of your variables, there is little more interpretation available.






                        share|cite|improve this answer









                        $endgroup$

















                          1












                          $begingroup$

                          The tests in multiple regression are "added last" tests. That means they test whether the model significantly improves after including the extra variable in a regression that contains all other predictors.



                          In your model with no predictors, adding A improves the model, so the test of A is significant in the model with only A.



                          In a model with A already in the model, adding B improves the model, so the test of B is significant in the model with A and B. But in a model with B already in the model, adding A doesn't improve the model, so the test of A is not significant in the model with A and B. B is doing all the work that A would do, so adding A doesn't improve the model beyond B.



                          As @IrishStat mentioned, this can occur when A and B are correlated (positively or negatively) with each other. It's a fairly common occurrence in regression modeling. The conclusion you might draw is that A predicts the outcome when B is not in the model (i.e., unavailable), but after including B, A doesn't do much more to predict the outcome. Unfortunately, without more information about the causal structure of your variables, there is little more interpretation available.






                          share|cite|improve this answer









                          $endgroup$















                            1












                            1








                            1





                            $begingroup$

                            The tests in multiple regression are "added last" tests. That means they test whether the model significantly improves after including the extra variable in a regression that contains all other predictors.



                            In your model with no predictors, adding A improves the model, so the test of A is significant in the model with only A.



                            In a model with A already in the model, adding B improves the model, so the test of B is significant in the model with A and B. But in a model with B already in the model, adding A doesn't improve the model, so the test of A is not significant in the model with A and B. B is doing all the work that A would do, so adding A doesn't improve the model beyond B.



                            As @IrishStat mentioned, this can occur when A and B are correlated (positively or negatively) with each other. It's a fairly common occurrence in regression modeling. The conclusion you might draw is that A predicts the outcome when B is not in the model (i.e., unavailable), but after including B, A doesn't do much more to predict the outcome. Unfortunately, without more information about the causal structure of your variables, there is little more interpretation available.






                            share|cite|improve this answer









                            $endgroup$



                            The tests in multiple regression are "added last" tests. That means they test whether the model significantly improves after including the extra variable in a regression that contains all other predictors.



                            In your model with no predictors, adding A improves the model, so the test of A is significant in the model with only A.



                            In a model with A already in the model, adding B improves the model, so the test of B is significant in the model with A and B. But in a model with B already in the model, adding A doesn't improve the model, so the test of A is not significant in the model with A and B. B is doing all the work that A would do, so adding A doesn't improve the model beyond B.



                            As @IrishStat mentioned, this can occur when A and B are correlated (positively or negatively) with each other. It's a fairly common occurrence in regression modeling. The conclusion you might draw is that A predicts the outcome when B is not in the model (i.e., unavailable), but after including B, A doesn't do much more to predict the outcome. Unfortunately, without more information about the causal structure of your variables, there is little more interpretation available.







                            share|cite|improve this answer












                            share|cite|improve this answer



                            share|cite|improve this answer










                            answered 7 hours ago









                            NoahNoah

                            3,6811417




                            3,6811417





















                                0












                                $begingroup$

                                To expand a little on @Noah and @IrishStat's answers, in a multiple regression, coefficients for each independent variable/predictor are estimated to obtain the direct effect of each variable, using variation unique to that variable and the variable's correlation with the outcome variable, not using variation shared by predictors. (In technical terms, we are talking about variance and covariance of these variables.) The less unique variation there is, the less significant the estimate will become.



                                So why, in your example, did you end up with an insignificant predictor A when B was added, and not with a significant predictor A and insignificant predictor B? It is likely because the proportion of variance of predictor A that it has in common with predictor B is larger than the proportion of variance of predictor B that it has in common with predictor A.






                                share|cite|improve this answer









                                $endgroup$

















                                  0












                                  $begingroup$

                                  To expand a little on @Noah and @IrishStat's answers, in a multiple regression, coefficients for each independent variable/predictor are estimated to obtain the direct effect of each variable, using variation unique to that variable and the variable's correlation with the outcome variable, not using variation shared by predictors. (In technical terms, we are talking about variance and covariance of these variables.) The less unique variation there is, the less significant the estimate will become.



                                  So why, in your example, did you end up with an insignificant predictor A when B was added, and not with a significant predictor A and insignificant predictor B? It is likely because the proportion of variance of predictor A that it has in common with predictor B is larger than the proportion of variance of predictor B that it has in common with predictor A.






                                  share|cite|improve this answer









                                  $endgroup$















                                    0












                                    0








                                    0





                                    $begingroup$

                                    To expand a little on @Noah and @IrishStat's answers, in a multiple regression, coefficients for each independent variable/predictor are estimated to obtain the direct effect of each variable, using variation unique to that variable and the variable's correlation with the outcome variable, not using variation shared by predictors. (In technical terms, we are talking about variance and covariance of these variables.) The less unique variation there is, the less significant the estimate will become.



                                    So why, in your example, did you end up with an insignificant predictor A when B was added, and not with a significant predictor A and insignificant predictor B? It is likely because the proportion of variance of predictor A that it has in common with predictor B is larger than the proportion of variance of predictor B that it has in common with predictor A.






                                    share|cite|improve this answer









                                    $endgroup$



                                    To expand a little on @Noah and @IrishStat's answers, in a multiple regression, coefficients for each independent variable/predictor are estimated to obtain the direct effect of each variable, using variation unique to that variable and the variable's correlation with the outcome variable, not using variation shared by predictors. (In technical terms, we are talking about variance and covariance of these variables.) The less unique variation there is, the less significant the estimate will become.



                                    So why, in your example, did you end up with an insignificant predictor A when B was added, and not with a significant predictor A and insignificant predictor B? It is likely because the proportion of variance of predictor A that it has in common with predictor B is larger than the proportion of variance of predictor B that it has in common with predictor A.







                                    share|cite|improve this answer












                                    share|cite|improve this answer



                                    share|cite|improve this answer










                                    answered 28 mins ago









                                    AlexKAlexK

                                    1908




                                    1908




















                                        ummmm is a new contributor. Be nice, and check out our Code of Conduct.









                                        draft saved

                                        draft discarded


















                                        ummmm is a new contributor. Be nice, and check out our Code of Conduct.












                                        ummmm is a new contributor. Be nice, and check out our Code of Conduct.











                                        ummmm is a new contributor. Be nice, and check out our Code of Conduct.














                                        Thanks for contributing an answer to Cross Validated!


                                        • Please be sure to answer the question. Provide details and share your research!

                                        But avoid


                                        • Asking for help, clarification, or responding to other answers.

                                        • Making statements based on opinion; back them up with references or personal experience.

                                        Use MathJax to format equations. MathJax reference.


                                        To learn more, see our tips on writing great answers.




                                        draft saved


                                        draft discarded














                                        StackExchange.ready(
                                        function ()
                                        StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f402894%2fmultiple-regression-results-help%23new-answer', 'question_page');

                                        );

                                        Post as a guest















                                        Required, but never shown





















































                                        Required, but never shown














                                        Required, but never shown












                                        Required, but never shown







                                        Required, but never shown

































                                        Required, but never shown














                                        Required, but never shown












                                        Required, but never shown







                                        Required, but never shown







                                        Popular posts from this blog

                                        Log på Navigationsmenu

                                        Creating second map without labels using QGIS?How to lock map labels for inset map in Print Composer?How to Force the Showing of Labels of a Vector File in QGISQGIS Valmiera, Labels only show for part of polygonsRemoving duplicate point labels in QGISLabeling every feature using QGIS?Show labels for point features outside map canvasAbbreviate Road Labels in QGIS only when requiredExporting map from composer in QGIS - text labels have moved in output?How to make sure labels in qgis turn up in layout map?Writing label expression with ArcMap and If then Statement?

                                        Detroit Tigers Spis treści Historia | Skład zespołu | Sukcesy | Członkowie Baseball Hall of Fame | Zastrzeżone numery | Przypisy | Menu nawigacyjneEncyclopedia of Detroit - Detroit TigersTigers Stadium, Detroit, MITigers Timeline 1900sDetroit Tigers Team History & EncyclopediaTigers Timeline 1910s1935 World Series1945 World Series1945 World Series1984 World SeriesComerica Park, Detroit, MI2006 World Series2012 World SeriesDetroit Tigers 40-Man RosterDetroit Tigers Coaching StaffTigers Hall of FamersTigers Retired Numberse