On George Box, Galit Shmueli and the scientific method?Paradox in model selection (AIC, BIC, to explain or to predict?)Practical thoughts on explanatory vs. predictive modelingPractical thoughts on explanatory vs predictive modelingHow to estimate the prediction power for a combination of MARS and CART model?Paradox in model selection (AIC, BIC, to explain or to predict?)Match model selection strategies with modelling objectivesWhat is the point of dividing data into training and test parts to assess prediction properties when we have AIC?How exactly does ridge regression helps in the case of multicollinearity?Minimizing bias in explanatory modeling, why? (Galit Shmueli's “To Explain or to Predict”)Time series analysis VAR model: AIC and BIC test criteriaBias and Variance in underspecified models

Is adding a new player (or players) a DM decision, or a group decision?

How dangerous are set-size assumptions?

Where can I find a database of galactic spectra?

How risky is real estate?

Why cruise at 7000' in an A319?

Why do all the teams that I have worked with always finish a sprint without completion of all the stories?

Why do some professors with PhDs leave their professorships to teach high school?

Apply brace expansion in "reverse order"

Impossible darts scores

Is my Rep in Stack-Exchange Form?

Why do textbooks often include the solutions to odd or even numbered problems but not both?

Does squid ink pasta bleed?

How long would it take to cross the Channel in 1890's?

Should developer taking test phones home or put in office?

Employer wants to use my work email account after I quit

Inverse-quotes-quine

Is there a maximum distance from a planet that a moon can orbit?

Can Ogre clerics use Purify Food and Drink on humanoid characters?

Iterate over colors on next pie chart with pgf-pie

How was Hillel permitted to go to the skylight to hear the shiur

Accidentals and ties

Using sed to replace "A" with a "B" or "C"

Is it damaging to turn off a small fridge for two days every week?

Fedora boot screen shows both Fedora logo and Lenovo logo. Why and How?



On George Box, Galit Shmueli and the scientific method?


Paradox in model selection (AIC, BIC, to explain or to predict?)Practical thoughts on explanatory vs. predictive modelingPractical thoughts on explanatory vs predictive modelingHow to estimate the prediction power for a combination of MARS and CART model?Paradox in model selection (AIC, BIC, to explain or to predict?)Match model selection strategies with modelling objectivesWhat is the point of dividing data into training and test parts to assess prediction properties when we have AIC?How exactly does ridge regression helps in the case of multicollinearity?Minimizing bias in explanatory modeling, why? (Galit Shmueli's “To Explain or to Predict”)Time series analysis VAR model: AIC and BIC test criteriaBias and Variance in underspecified models






.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty,.everyoneloves__bot-mid-leaderboard:empty margin-bottom:0;








14












$begingroup$


(This question might seem like it is better suited for the Philosophy SE. I am hoping that statisticians can clarify my misconceptions about Box's and Shmueli's statements, hence I am posting it here).



George Box (of ARIMA fame said):



"All models are wrong, but some are useful."



Galit Shmueli in her famous paper "To Explain or to Predict", argues (and cites others who agree with her) that:



Explaining and predicting are not the same, and that some models do a good job of explaining, even though they do a poor job at predicting.



I feel that these to principles are somehow contradictory.



If a model doesn't predict well, is it useful?



More importantly, if a model explains well (but doesn't necessarily predict well), then it has to be true (i.e not wrong) in some way or another. So how does that mesh with Box's "all models are wrong"?



Finally, if a model explains well, but doesn't predict well, how is it even scientific? Most scientific demarcation criteria (verificationism, falsificstionism, etc...) imply that a scientific statement has to have predictive power, or colloquially: A theory or model is correct only if it can be empirically tested (or falsified), which means that it has to predict future outcomes.



My questions:



  • Are Box's statement and Shmueli's ideas indeed contradictory, or am I missing something, e.g can a model not have predictive power yet still be useful?

  • If the statements of Box and Shmueli are not contradictory, then what does it mean for a model to be wrong and not predict well, yet still have explanatory power? Put it differently: If one takes away both correctness and predictive ability, what is left of a model?

What empirical validations are possible when a model has explanatory power, but not predictive power? Shmueli mentions things like: use the AIC for explanation and the BIC for prediction, etc,...but I don't see how that solves the problem. With predictive models, you can use the AIC, or the BIC, or the $R^2$, or $L1$ regularization, etc...but ultimately out of sample testing and performance in production is what determines the quality of the model. But for models that explain well, I don't see how any loss function can ever truly evaluate a model. In philosophy of science, there is the concept of underdetermination which seems pertinent here: For any given data set, one can always judiciously choose some distribution (or mixture of distributions) and loss function $L$ in such a way that they fit the data (and therefore can be claimed to explain it). Moreover, the threshold that $L$ should be under for someone to claim that the model adequately explains the data is arbitrary (kind of like p-values, why is it $p < 0.05$ and not $p < 0.1$ or $p < 0.01$?).



  • Based on the above, how can one objectively validate a model that explains well, but doesn't predict well, since out of sample testing is not possible?









share|cite|improve this question











$endgroup$











  • $begingroup$
    See also a related question: Paradox in model selection (AIC, BIC, to explain or to predict?).
    $endgroup$
    – Richard Hardy
    Jun 14 at 8:27







  • 3




    $begingroup$
    As far as I can remember Shmueli, for her to explain well means to get the functional form right (but possibly have huge estimation imprecission when it comes to parameters of that functional form), while to predict well means to get the bias-variance tradeoff right (compromise on the functional form so as to increase estimation precision). Also, the tag philosophical could be useful here.
    $endgroup$
    – Richard Hardy
    Jun 14 at 8:28







  • 1




    $begingroup$
    I do not think the "or" needs to be exclusive. Simplicity and unification are non-controversial selection criteria for theories of equal predictive value, and if so, sacrificing precision for them may well be justified in many contexts. Box's motto reflects a different view of science, advanced e.g. in van Fraassen's Scientific Image (and going back to Kant): it is to construct adequate/useful descriptions of the observed, not to tell literally true stories about unobservable "reality". Rough adequacy may suffice for many tasks, and "the one correct model" may well be a naive figment.
    $endgroup$
    – Conifold
    Jun 14 at 8:46






  • 2




    $begingroup$
    @Conifold Indeed, Richard Levins' loop analysis of complex causal systems (not a statistical method, though there are applications of it that relate directly to statistical predictions) sacrifices almost all precision in favor of model realism (variables & relationships between them) and generality (applicability of analytic results on a model to all variables sharing the same causal structure). See Levins, R. (1966). The Strategy of Model Building In Population Biology. American Scientist, 54(4), 421–431.
    $endgroup$
    – Alexis
    Jun 14 at 18:37











  • $begingroup$
    (+1. I'm still hoping to find time to post an answer here...)
    $endgroup$
    – amoeba
    yesterday


















14












$begingroup$


(This question might seem like it is better suited for the Philosophy SE. I am hoping that statisticians can clarify my misconceptions about Box's and Shmueli's statements, hence I am posting it here).



George Box (of ARIMA fame said):



"All models are wrong, but some are useful."



Galit Shmueli in her famous paper "To Explain or to Predict", argues (and cites others who agree with her) that:



Explaining and predicting are not the same, and that some models do a good job of explaining, even though they do a poor job at predicting.



I feel that these to principles are somehow contradictory.



If a model doesn't predict well, is it useful?



More importantly, if a model explains well (but doesn't necessarily predict well), then it has to be true (i.e not wrong) in some way or another. So how does that mesh with Box's "all models are wrong"?



Finally, if a model explains well, but doesn't predict well, how is it even scientific? Most scientific demarcation criteria (verificationism, falsificstionism, etc...) imply that a scientific statement has to have predictive power, or colloquially: A theory or model is correct only if it can be empirically tested (or falsified), which means that it has to predict future outcomes.



My questions:



  • Are Box's statement and Shmueli's ideas indeed contradictory, or am I missing something, e.g can a model not have predictive power yet still be useful?

  • If the statements of Box and Shmueli are not contradictory, then what does it mean for a model to be wrong and not predict well, yet still have explanatory power? Put it differently: If one takes away both correctness and predictive ability, what is left of a model?

What empirical validations are possible when a model has explanatory power, but not predictive power? Shmueli mentions things like: use the AIC for explanation and the BIC for prediction, etc,...but I don't see how that solves the problem. With predictive models, you can use the AIC, or the BIC, or the $R^2$, or $L1$ regularization, etc...but ultimately out of sample testing and performance in production is what determines the quality of the model. But for models that explain well, I don't see how any loss function can ever truly evaluate a model. In philosophy of science, there is the concept of underdetermination which seems pertinent here: For any given data set, one can always judiciously choose some distribution (or mixture of distributions) and loss function $L$ in such a way that they fit the data (and therefore can be claimed to explain it). Moreover, the threshold that $L$ should be under for someone to claim that the model adequately explains the data is arbitrary (kind of like p-values, why is it $p < 0.05$ and not $p < 0.1$ or $p < 0.01$?).



  • Based on the above, how can one objectively validate a model that explains well, but doesn't predict well, since out of sample testing is not possible?









share|cite|improve this question











$endgroup$











  • $begingroup$
    See also a related question: Paradox in model selection (AIC, BIC, to explain or to predict?).
    $endgroup$
    – Richard Hardy
    Jun 14 at 8:27







  • 3




    $begingroup$
    As far as I can remember Shmueli, for her to explain well means to get the functional form right (but possibly have huge estimation imprecission when it comes to parameters of that functional form), while to predict well means to get the bias-variance tradeoff right (compromise on the functional form so as to increase estimation precision). Also, the tag philosophical could be useful here.
    $endgroup$
    – Richard Hardy
    Jun 14 at 8:28







  • 1




    $begingroup$
    I do not think the "or" needs to be exclusive. Simplicity and unification are non-controversial selection criteria for theories of equal predictive value, and if so, sacrificing precision for them may well be justified in many contexts. Box's motto reflects a different view of science, advanced e.g. in van Fraassen's Scientific Image (and going back to Kant): it is to construct adequate/useful descriptions of the observed, not to tell literally true stories about unobservable "reality". Rough adequacy may suffice for many tasks, and "the one correct model" may well be a naive figment.
    $endgroup$
    – Conifold
    Jun 14 at 8:46






  • 2




    $begingroup$
    @Conifold Indeed, Richard Levins' loop analysis of complex causal systems (not a statistical method, though there are applications of it that relate directly to statistical predictions) sacrifices almost all precision in favor of model realism (variables & relationships between them) and generality (applicability of analytic results on a model to all variables sharing the same causal structure). See Levins, R. (1966). The Strategy of Model Building In Population Biology. American Scientist, 54(4), 421–431.
    $endgroup$
    – Alexis
    Jun 14 at 18:37











  • $begingroup$
    (+1. I'm still hoping to find time to post an answer here...)
    $endgroup$
    – amoeba
    yesterday














14












14








14


6



$begingroup$


(This question might seem like it is better suited for the Philosophy SE. I am hoping that statisticians can clarify my misconceptions about Box's and Shmueli's statements, hence I am posting it here).



George Box (of ARIMA fame said):



"All models are wrong, but some are useful."



Galit Shmueli in her famous paper "To Explain or to Predict", argues (and cites others who agree with her) that:



Explaining and predicting are not the same, and that some models do a good job of explaining, even though they do a poor job at predicting.



I feel that these to principles are somehow contradictory.



If a model doesn't predict well, is it useful?



More importantly, if a model explains well (but doesn't necessarily predict well), then it has to be true (i.e not wrong) in some way or another. So how does that mesh with Box's "all models are wrong"?



Finally, if a model explains well, but doesn't predict well, how is it even scientific? Most scientific demarcation criteria (verificationism, falsificstionism, etc...) imply that a scientific statement has to have predictive power, or colloquially: A theory or model is correct only if it can be empirically tested (or falsified), which means that it has to predict future outcomes.



My questions:



  • Are Box's statement and Shmueli's ideas indeed contradictory, or am I missing something, e.g can a model not have predictive power yet still be useful?

  • If the statements of Box and Shmueli are not contradictory, then what does it mean for a model to be wrong and not predict well, yet still have explanatory power? Put it differently: If one takes away both correctness and predictive ability, what is left of a model?

What empirical validations are possible when a model has explanatory power, but not predictive power? Shmueli mentions things like: use the AIC for explanation and the BIC for prediction, etc,...but I don't see how that solves the problem. With predictive models, you can use the AIC, or the BIC, or the $R^2$, or $L1$ regularization, etc...but ultimately out of sample testing and performance in production is what determines the quality of the model. But for models that explain well, I don't see how any loss function can ever truly evaluate a model. In philosophy of science, there is the concept of underdetermination which seems pertinent here: For any given data set, one can always judiciously choose some distribution (or mixture of distributions) and loss function $L$ in such a way that they fit the data (and therefore can be claimed to explain it). Moreover, the threshold that $L$ should be under for someone to claim that the model adequately explains the data is arbitrary (kind of like p-values, why is it $p < 0.05$ and not $p < 0.1$ or $p < 0.01$?).



  • Based on the above, how can one objectively validate a model that explains well, but doesn't predict well, since out of sample testing is not possible?









share|cite|improve this question











$endgroup$




(This question might seem like it is better suited for the Philosophy SE. I am hoping that statisticians can clarify my misconceptions about Box's and Shmueli's statements, hence I am posting it here).



George Box (of ARIMA fame said):



"All models are wrong, but some are useful."



Galit Shmueli in her famous paper "To Explain or to Predict", argues (and cites others who agree with her) that:



Explaining and predicting are not the same, and that some models do a good job of explaining, even though they do a poor job at predicting.



I feel that these to principles are somehow contradictory.



If a model doesn't predict well, is it useful?



More importantly, if a model explains well (but doesn't necessarily predict well), then it has to be true (i.e not wrong) in some way or another. So how does that mesh with Box's "all models are wrong"?



Finally, if a model explains well, but doesn't predict well, how is it even scientific? Most scientific demarcation criteria (verificationism, falsificstionism, etc...) imply that a scientific statement has to have predictive power, or colloquially: A theory or model is correct only if it can be empirically tested (or falsified), which means that it has to predict future outcomes.



My questions:



  • Are Box's statement and Shmueli's ideas indeed contradictory, or am I missing something, e.g can a model not have predictive power yet still be useful?

  • If the statements of Box and Shmueli are not contradictory, then what does it mean for a model to be wrong and not predict well, yet still have explanatory power? Put it differently: If one takes away both correctness and predictive ability, what is left of a model?

What empirical validations are possible when a model has explanatory power, but not predictive power? Shmueli mentions things like: use the AIC for explanation and the BIC for prediction, etc,...but I don't see how that solves the problem. With predictive models, you can use the AIC, or the BIC, or the $R^2$, or $L1$ regularization, etc...but ultimately out of sample testing and performance in production is what determines the quality of the model. But for models that explain well, I don't see how any loss function can ever truly evaluate a model. In philosophy of science, there is the concept of underdetermination which seems pertinent here: For any given data set, one can always judiciously choose some distribution (or mixture of distributions) and loss function $L$ in such a way that they fit the data (and therefore can be claimed to explain it). Moreover, the threshold that $L$ should be under for someone to claim that the model adequately explains the data is arbitrary (kind of like p-values, why is it $p < 0.05$ and not $p < 0.1$ or $p < 0.01$?).



  • Based on the above, how can one objectively validate a model that explains well, but doesn't predict well, since out of sample testing is not possible?






predictive-models aic bic philosophical explanatory-models






share|cite|improve this question















share|cite|improve this question













share|cite|improve this question




share|cite|improve this question








edited Jun 14 at 17:03







Skander H.

















asked Jun 14 at 7:15









Skander H.Skander H.

4,52813 silver badges41 bronze badges




4,52813 silver badges41 bronze badges











  • $begingroup$
    See also a related question: Paradox in model selection (AIC, BIC, to explain or to predict?).
    $endgroup$
    – Richard Hardy
    Jun 14 at 8:27







  • 3




    $begingroup$
    As far as I can remember Shmueli, for her to explain well means to get the functional form right (but possibly have huge estimation imprecission when it comes to parameters of that functional form), while to predict well means to get the bias-variance tradeoff right (compromise on the functional form so as to increase estimation precision). Also, the tag philosophical could be useful here.
    $endgroup$
    – Richard Hardy
    Jun 14 at 8:28







  • 1




    $begingroup$
    I do not think the "or" needs to be exclusive. Simplicity and unification are non-controversial selection criteria for theories of equal predictive value, and if so, sacrificing precision for them may well be justified in many contexts. Box's motto reflects a different view of science, advanced e.g. in van Fraassen's Scientific Image (and going back to Kant): it is to construct adequate/useful descriptions of the observed, not to tell literally true stories about unobservable "reality". Rough adequacy may suffice for many tasks, and "the one correct model" may well be a naive figment.
    $endgroup$
    – Conifold
    Jun 14 at 8:46






  • 2




    $begingroup$
    @Conifold Indeed, Richard Levins' loop analysis of complex causal systems (not a statistical method, though there are applications of it that relate directly to statistical predictions) sacrifices almost all precision in favor of model realism (variables & relationships between them) and generality (applicability of analytic results on a model to all variables sharing the same causal structure). See Levins, R. (1966). The Strategy of Model Building In Population Biology. American Scientist, 54(4), 421–431.
    $endgroup$
    – Alexis
    Jun 14 at 18:37











  • $begingroup$
    (+1. I'm still hoping to find time to post an answer here...)
    $endgroup$
    – amoeba
    yesterday

















  • $begingroup$
    See also a related question: Paradox in model selection (AIC, BIC, to explain or to predict?).
    $endgroup$
    – Richard Hardy
    Jun 14 at 8:27







  • 3




    $begingroup$
    As far as I can remember Shmueli, for her to explain well means to get the functional form right (but possibly have huge estimation imprecission when it comes to parameters of that functional form), while to predict well means to get the bias-variance tradeoff right (compromise on the functional form so as to increase estimation precision). Also, the tag philosophical could be useful here.
    $endgroup$
    – Richard Hardy
    Jun 14 at 8:28







  • 1




    $begingroup$
    I do not think the "or" needs to be exclusive. Simplicity and unification are non-controversial selection criteria for theories of equal predictive value, and if so, sacrificing precision for them may well be justified in many contexts. Box's motto reflects a different view of science, advanced e.g. in van Fraassen's Scientific Image (and going back to Kant): it is to construct adequate/useful descriptions of the observed, not to tell literally true stories about unobservable "reality". Rough adequacy may suffice for many tasks, and "the one correct model" may well be a naive figment.
    $endgroup$
    – Conifold
    Jun 14 at 8:46






  • 2




    $begingroup$
    @Conifold Indeed, Richard Levins' loop analysis of complex causal systems (not a statistical method, though there are applications of it that relate directly to statistical predictions) sacrifices almost all precision in favor of model realism (variables & relationships between them) and generality (applicability of analytic results on a model to all variables sharing the same causal structure). See Levins, R. (1966). The Strategy of Model Building In Population Biology. American Scientist, 54(4), 421–431.
    $endgroup$
    – Alexis
    Jun 14 at 18:37











  • $begingroup$
    (+1. I'm still hoping to find time to post an answer here...)
    $endgroup$
    – amoeba
    yesterday
















$begingroup$
See also a related question: Paradox in model selection (AIC, BIC, to explain or to predict?).
$endgroup$
– Richard Hardy
Jun 14 at 8:27





$begingroup$
See also a related question: Paradox in model selection (AIC, BIC, to explain or to predict?).
$endgroup$
– Richard Hardy
Jun 14 at 8:27





3




3




$begingroup$
As far as I can remember Shmueli, for her to explain well means to get the functional form right (but possibly have huge estimation imprecission when it comes to parameters of that functional form), while to predict well means to get the bias-variance tradeoff right (compromise on the functional form so as to increase estimation precision). Also, the tag philosophical could be useful here.
$endgroup$
– Richard Hardy
Jun 14 at 8:28





$begingroup$
As far as I can remember Shmueli, for her to explain well means to get the functional form right (but possibly have huge estimation imprecission when it comes to parameters of that functional form), while to predict well means to get the bias-variance tradeoff right (compromise on the functional form so as to increase estimation precision). Also, the tag philosophical could be useful here.
$endgroup$
– Richard Hardy
Jun 14 at 8:28





1




1




$begingroup$
I do not think the "or" needs to be exclusive. Simplicity and unification are non-controversial selection criteria for theories of equal predictive value, and if so, sacrificing precision for them may well be justified in many contexts. Box's motto reflects a different view of science, advanced e.g. in van Fraassen's Scientific Image (and going back to Kant): it is to construct adequate/useful descriptions of the observed, not to tell literally true stories about unobservable "reality". Rough adequacy may suffice for many tasks, and "the one correct model" may well be a naive figment.
$endgroup$
– Conifold
Jun 14 at 8:46




$begingroup$
I do not think the "or" needs to be exclusive. Simplicity and unification are non-controversial selection criteria for theories of equal predictive value, and if so, sacrificing precision for them may well be justified in many contexts. Box's motto reflects a different view of science, advanced e.g. in van Fraassen's Scientific Image (and going back to Kant): it is to construct adequate/useful descriptions of the observed, not to tell literally true stories about unobservable "reality". Rough adequacy may suffice for many tasks, and "the one correct model" may well be a naive figment.
$endgroup$
– Conifold
Jun 14 at 8:46




2




2




$begingroup$
@Conifold Indeed, Richard Levins' loop analysis of complex causal systems (not a statistical method, though there are applications of it that relate directly to statistical predictions) sacrifices almost all precision in favor of model realism (variables & relationships between them) and generality (applicability of analytic results on a model to all variables sharing the same causal structure). See Levins, R. (1966). The Strategy of Model Building In Population Biology. American Scientist, 54(4), 421–431.
$endgroup$
– Alexis
Jun 14 at 18:37





$begingroup$
@Conifold Indeed, Richard Levins' loop analysis of complex causal systems (not a statistical method, though there are applications of it that relate directly to statistical predictions) sacrifices almost all precision in favor of model realism (variables & relationships between them) and generality (applicability of analytic results on a model to all variables sharing the same causal structure). See Levins, R. (1966). The Strategy of Model Building In Population Biology. American Scientist, 54(4), 421–431.
$endgroup$
– Alexis
Jun 14 at 18:37













$begingroup$
(+1. I'm still hoping to find time to post an answer here...)
$endgroup$
– amoeba
yesterday





$begingroup$
(+1. I'm still hoping to find time to post an answer here...)
$endgroup$
– amoeba
yesterday











4 Answers
4






active

oldest

votes


















8












$begingroup$

Let me start with the pithy quote by George Box, that "all models are wrong, but some are useful". This statement is an encapsulation of the methodological approach of "positivism", which is a philosophical approach that is highly influential in the sciences. This approach is described in detail (in the context of economic theory) in the classic methodological essay of Friedman (1966). In that essay, Friedman argues that any useful scientific theory necessarily constitutes a simplification of reality, and thus its assumptions must always depart from reality to some degree, and may even depart substantially from reality. He argues that the value of a scientific theory should not be judged by the closeness of its assumptions to reality --- instead it should be judged by its simplicity in reducing the complexity of the world to a manageable set of principles, and its accuracy in making predictions about reality, and generating new testable hypotheses about reality. Thus, Friedman argues that "all models are wrong" insofar as they all contain assumptions that simplify (and therefore depart from) reality, but that "some are useful" insofar as they give a simple framework to make useful predictions about reality.



Now, if you read Box (1976) (the paper where he first states that "all models are wrong"), you will see that he does not cite Friedman, nor does he mention methodological positivism. Nevertheless, his explanation of the scientific method and its characteristics is extremely close to that developed by Friedman. In particular, both authors stress that a scientific theory will make predictions about reality that can be tested against observed facts, and the error in the prediction can then be used as a basis for revision of the theory.



Now, on to the dichotomy discussed by Galit Shmueli in Shmueli (2001). In this paper, Shmueli compares causal explanation and prediction of observed outcomes and argues that these are distinct activities. Specifically, she argues that causal relations are based on underlying constructs that do not manifest directly in measureable outcomes, and so "measurable data are not accurate representations of their underlying constructs" (p. 293). She therefore argues that there is an aspect of statistical analysis that involves making inferences about unobservable underlying causal relations that are not manifested in measureable counterfactual differences in outcomes.



Unless I am misunderstanding something, I think it is fair to say that this idea is in tension with the positivist views of Box and Friedman, as represented in the quote by Box. The positivist viewpoint essentially says that there are no admissible metaphysical "constructs" beyond those that manifest in measureable outcomes. Positivism confines itself to consideration of observable data, and concepts built on this data; it excludes consideration of a priori metaphysical concepts. Thus, a positivist would argue that the concept of causality can only be valid to the extent that it is defined in terms of measureable outcomes in reality --- to the extent it is defined as something distinct from this (as Shmueli treats it), this would be regarded as metaphysical speculation, and would be treated as inadmissible in scientific discourse.



So I think you're right --- these two approaches are essentially in conflict. The positivist approach used by Box insists that valid scientific concepts be grounded entirely in their manifestations in reality, whereas the alternative approach used by Shmueli says that there are some "constructs" that are important scientific concepts (that we want to explain) but which cannot be perfectly represented when they are "operationalised" by relating them to measureable outcomes in reality.






share|cite|improve this answer











$endgroup$












  • $begingroup$
    Exactly!!! Shmueli seems to be contradicting most positivist (and falsificationist) definitions of what a scientific statement is, and I am wondering whether her intention is indeed to make such a bold philosophical statement? or whether she, as a statistician, isn't aware of how bold her statements really are?
    $endgroup$
    – Skander H.
    Jun 14 at 17:07










  • $begingroup$
    I posted a related question on the philosophy SE, care to weigh in on it?
    $endgroup$
    – Skander H.
    Jun 14 at 17:10






  • 1




    $begingroup$
    Although positivism and pragmatism share anti-realist traits, Box's approach is the latter, and positivism is not influential in sciences at least since late 1960-s. This is why Box does not mention Friedman, or other positivists. Pragmatism does not confine itself to observable data, and has no problem with a priori concepts or metaphysical constructions. It just does not see them as approaching "reality", and hence aiming at uniquely "right" features, they can be plural, task-dependent, and revisable. Causal relations are such constructs, so there is no conflict between Box and Shmueli.
    $endgroup$
    – Conifold
    Jun 14 at 21:37











  • $begingroup$
    Whether or not Box was a pragmatist, the particular quote he gives here is, in my view, more conducive to positivism than pragmatism. The latter philosophy takes a pluralistic view of metaphysics, which holds that there are multiple coherent ways of conceptualising reality, and that all these are "true" in the sense of being useful conceptualisations of reality. Thus, positivism would say, "all models are wrong, but some are useful", whereas the maxim of pragmatism would be closer to "many models are right, because they are useful".
    $endgroup$
    – Ben
    yesterday










  • $begingroup$
    The identification of true with useful is just a folk misconception about pragmatism. All models are wrong because "right" suggests that they have something to correspond to, which pragmatists deny. And the principle of tolerance, "models are right as long as they serve a purpose", is due to Carnap, the father of logical positivism.
    $endgroup$
    – Conifold
    5 hours ago


















4












$begingroup$

A model, when used to explain things, is a simplification of reality. Simplification is just another word for "wrong in some useful way". For example, if we round the number 3.1415926535898 to 3.14 we are making an error, but this error allows us humans to focus on the most important part of that number. This is how models are used in explaining, it provides insights on some problem, but by necessity has to abstract away from a lot of other things: We humans are just not very good at looking at a thousands things simultaneously. If we primarily care about predicting we want to include those thousands things whenever possible, but with explaining the trade-off is different.






share|cite|improve this answer









$endgroup$








  • 1




    $begingroup$
    "but this error allows us humans to focus on the most important part of that number." this makes sense and helps me understand what "explanation" means, but then it also confirms my point that explanation is more of an artistic/aesthetic concept than a scientific one. Based on you $pi$ example, consider the following as well: A novel dimensionality reduction technique allows for plotting very elegant and intuitive graphs of high dimensional data, which makes for good explanation, but then there is no way to objectively asses the accuracy of this technique, its value is purely subjective.
    $endgroup$
    – Skander H.
    Jun 14 at 8:41






  • 1




    $begingroup$
    @SkanderH. To the extent that "elegant and intuitive graphs" facilitate engineering applications, or development of new theories, their value is not purely subjective, or non-scientific, it is pragmatic. Unification, explanatory power, simplicity and coherence are broadly acknowledged as epistemic, not aesthetic, values. The choice between Lorentz's theory of ether and special relativity was made based on just such considerations, they are predictively equivalent.
    $endgroup$
    – Conifold
    Jun 14 at 9:33


















2












$begingroup$

An example of a model that is excellent at prediction but does not explain anything is given in the Wikipedia article “All models are wrong”. The example is Newton’s model of gravitation. Newton’s model almost always gives predictions that are indistinguishable from empirical observations. Yet the model is extremely implausible: because it postulates a force that can act instantaneously over arbitrarily large distances.



Newton’s model has been supplanted by the model given in Einstein’s general theory of relativity. With general relativity, gravitational forces travel through space at finite speed (the speed of light).



Newton’s model is not a simplification of the general-relativistic model. To illustrate that, consider an apple falling down from a tree. According to general relativity, the apple falls without Earth exerting any force on the apple. (The primary reason the apple falls is that Earth warps time, so that clocks near the base of the tree run more slowly than clocks high up in the tree.) Thus, as the Wikipedia article notes, Newton’s model is completely wrong from an explanatory perspective.



The paper by Shmueli [2010] presumes that there are two purposes for a model: prediction and explanation. In fact, several authors have stated that there are three purposes (see e.g. Konishi & Kitagawa [Information Criteria and Statistical Modeling, 2008: §1.1] and Friendly & Meyer [Discrete Data Analysis, 2016: §11.6]). The three purposes correspond to the three kinds of logical reasoning:



  • prediction (corresponding to deduction);

  • parameter estimation (corresponding to induction);

  • description of structure (corresponding to abduction).





share|cite|improve this answer










New contributor



SolidPhase is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
Check out our Code of Conduct.





$endgroup$












  • $begingroup$
    To say that Newton's model of gravitation "does not explain anything" is, frankly, ludicrous. -1.
    $endgroup$
    – amoeba
    Jun 16 at 20:45










  • $begingroup$
    amoeba, Newton's model does not explain anything about how gravitation works, under the assumption that general relativity is accurate. If an apple falls, Newton's model postulates that Earth exerts a force on the apple and that postulate is completely false. I ask you to consider my answer further. If you still do not understand, then kindly tell me what is unclear.
    $endgroup$
    – SolidPhase
    Jun 16 at 20:53











  • $begingroup$
    What you say is quite clear but I strongly disagree with it.
    $endgroup$
    – amoeba
    Jun 16 at 22:07










  • $begingroup$
    amoeba, I ask you to explain why you disagree: do you have a reason? (Note that I have added an extra sentence to the answer.)
    $endgroup$
    – SolidPhase
    Jun 16 at 22:16











  • $begingroup$
    Thanks. I will look up the references you mention. I understand how a model can predict even if it doesn't explain. What I don't get is the opposite direction: How can a model explain without predicting. You Newton vs. Einstein examples just muddies everything even more: The whole reason Einstein's theory supplanted Newton's was because it predicted better. Look at it another way: If we have competing explanatory models, how can we evaluated them unless we test which one has the most predictive power?
    $endgroup$
    – Skander H.
    Jun 17 at 5:35


















1












$begingroup$

I'm an undergraduate in Statistics, so I won't call myself an expert, but here are my two cents.



Models don't explain themselves; humans interpret them. Linear models are easier to understand than neural networks and random forests because they are closer to how we make decisions. Indeed, ANNs imitate the human brain, but you don't decide which restaurant to go tomorrow by doing a series of matrix multiplications. Instead, you weight some factors in your mind by their importance, which is essentially a linear combination.



"Explanatory power" measures how well a model gets along with humans' intuition, whereas "predictive power" measures how well it aligns with the underlying mechanism of the process in interest. The contradiction between them is essentially the gap between what the world is and how we can perceive/understand it. I hope this explains why "some models do a good job of explaining, even though they do a poor job at predicting".



Ian Stewart once said, "If our brains were simple enough for us to understand them, we'd be so simple that we couldn't." Unfortunately, our little human brains are actually very simple compared to the universe, or even a stock market (which involves a lot of brains :). Up to now, all models are products of human brains, so it must be more-or-less inaccurate, which leads to Box's "All models are wrong". On the other hand, a model doesn't have to be technically correct to be useful. For example, Newton's laws of motion has been disproved by Einstein, but it remains useful when an object is not ridiculously big or fast.



To address your question, I honestly can't see the incompatibility between Box and Shmueli's points. It seems that you consider "explanatory power" and "predictive power" to be binomial properties, but I think they sit at the two ends of a spectrum.






share|cite|improve this answer











$endgroup$












  • $begingroup$
    @RichardHardy Please see my edited answer.
    $endgroup$
    – nalzok
    Jun 14 at 13:11













Your Answer








StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "65"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);

else
createEditor();

);

function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: false,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: null,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);



);













draft saved

draft discarded


















StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f412980%2fon-george-box-galit-shmueli-and-the-scientific-method%23new-answer', 'question_page');

);

Post as a guest















Required, but never shown

























4 Answers
4






active

oldest

votes








4 Answers
4






active

oldest

votes









active

oldest

votes






active

oldest

votes









8












$begingroup$

Let me start with the pithy quote by George Box, that "all models are wrong, but some are useful". This statement is an encapsulation of the methodological approach of "positivism", which is a philosophical approach that is highly influential in the sciences. This approach is described in detail (in the context of economic theory) in the classic methodological essay of Friedman (1966). In that essay, Friedman argues that any useful scientific theory necessarily constitutes a simplification of reality, and thus its assumptions must always depart from reality to some degree, and may even depart substantially from reality. He argues that the value of a scientific theory should not be judged by the closeness of its assumptions to reality --- instead it should be judged by its simplicity in reducing the complexity of the world to a manageable set of principles, and its accuracy in making predictions about reality, and generating new testable hypotheses about reality. Thus, Friedman argues that "all models are wrong" insofar as they all contain assumptions that simplify (and therefore depart from) reality, but that "some are useful" insofar as they give a simple framework to make useful predictions about reality.



Now, if you read Box (1976) (the paper where he first states that "all models are wrong"), you will see that he does not cite Friedman, nor does he mention methodological positivism. Nevertheless, his explanation of the scientific method and its characteristics is extremely close to that developed by Friedman. In particular, both authors stress that a scientific theory will make predictions about reality that can be tested against observed facts, and the error in the prediction can then be used as a basis for revision of the theory.



Now, on to the dichotomy discussed by Galit Shmueli in Shmueli (2001). In this paper, Shmueli compares causal explanation and prediction of observed outcomes and argues that these are distinct activities. Specifically, she argues that causal relations are based on underlying constructs that do not manifest directly in measureable outcomes, and so "measurable data are not accurate representations of their underlying constructs" (p. 293). She therefore argues that there is an aspect of statistical analysis that involves making inferences about unobservable underlying causal relations that are not manifested in measureable counterfactual differences in outcomes.



Unless I am misunderstanding something, I think it is fair to say that this idea is in tension with the positivist views of Box and Friedman, as represented in the quote by Box. The positivist viewpoint essentially says that there are no admissible metaphysical "constructs" beyond those that manifest in measureable outcomes. Positivism confines itself to consideration of observable data, and concepts built on this data; it excludes consideration of a priori metaphysical concepts. Thus, a positivist would argue that the concept of causality can only be valid to the extent that it is defined in terms of measureable outcomes in reality --- to the extent it is defined as something distinct from this (as Shmueli treats it), this would be regarded as metaphysical speculation, and would be treated as inadmissible in scientific discourse.



So I think you're right --- these two approaches are essentially in conflict. The positivist approach used by Box insists that valid scientific concepts be grounded entirely in their manifestations in reality, whereas the alternative approach used by Shmueli says that there are some "constructs" that are important scientific concepts (that we want to explain) but which cannot be perfectly represented when they are "operationalised" by relating them to measureable outcomes in reality.






share|cite|improve this answer











$endgroup$












  • $begingroup$
    Exactly!!! Shmueli seems to be contradicting most positivist (and falsificationist) definitions of what a scientific statement is, and I am wondering whether her intention is indeed to make such a bold philosophical statement? or whether she, as a statistician, isn't aware of how bold her statements really are?
    $endgroup$
    – Skander H.
    Jun 14 at 17:07










  • $begingroup$
    I posted a related question on the philosophy SE, care to weigh in on it?
    $endgroup$
    – Skander H.
    Jun 14 at 17:10






  • 1




    $begingroup$
    Although positivism and pragmatism share anti-realist traits, Box's approach is the latter, and positivism is not influential in sciences at least since late 1960-s. This is why Box does not mention Friedman, or other positivists. Pragmatism does not confine itself to observable data, and has no problem with a priori concepts or metaphysical constructions. It just does not see them as approaching "reality", and hence aiming at uniquely "right" features, they can be plural, task-dependent, and revisable. Causal relations are such constructs, so there is no conflict between Box and Shmueli.
    $endgroup$
    – Conifold
    Jun 14 at 21:37











  • $begingroup$
    Whether or not Box was a pragmatist, the particular quote he gives here is, in my view, more conducive to positivism than pragmatism. The latter philosophy takes a pluralistic view of metaphysics, which holds that there are multiple coherent ways of conceptualising reality, and that all these are "true" in the sense of being useful conceptualisations of reality. Thus, positivism would say, "all models are wrong, but some are useful", whereas the maxim of pragmatism would be closer to "many models are right, because they are useful".
    $endgroup$
    – Ben
    yesterday










  • $begingroup$
    The identification of true with useful is just a folk misconception about pragmatism. All models are wrong because "right" suggests that they have something to correspond to, which pragmatists deny. And the principle of tolerance, "models are right as long as they serve a purpose", is due to Carnap, the father of logical positivism.
    $endgroup$
    – Conifold
    5 hours ago















8












$begingroup$

Let me start with the pithy quote by George Box, that "all models are wrong, but some are useful". This statement is an encapsulation of the methodological approach of "positivism", which is a philosophical approach that is highly influential in the sciences. This approach is described in detail (in the context of economic theory) in the classic methodological essay of Friedman (1966). In that essay, Friedman argues that any useful scientific theory necessarily constitutes a simplification of reality, and thus its assumptions must always depart from reality to some degree, and may even depart substantially from reality. He argues that the value of a scientific theory should not be judged by the closeness of its assumptions to reality --- instead it should be judged by its simplicity in reducing the complexity of the world to a manageable set of principles, and its accuracy in making predictions about reality, and generating new testable hypotheses about reality. Thus, Friedman argues that "all models are wrong" insofar as they all contain assumptions that simplify (and therefore depart from) reality, but that "some are useful" insofar as they give a simple framework to make useful predictions about reality.



Now, if you read Box (1976) (the paper where he first states that "all models are wrong"), you will see that he does not cite Friedman, nor does he mention methodological positivism. Nevertheless, his explanation of the scientific method and its characteristics is extremely close to that developed by Friedman. In particular, both authors stress that a scientific theory will make predictions about reality that can be tested against observed facts, and the error in the prediction can then be used as a basis for revision of the theory.



Now, on to the dichotomy discussed by Galit Shmueli in Shmueli (2001). In this paper, Shmueli compares causal explanation and prediction of observed outcomes and argues that these are distinct activities. Specifically, she argues that causal relations are based on underlying constructs that do not manifest directly in measureable outcomes, and so "measurable data are not accurate representations of their underlying constructs" (p. 293). She therefore argues that there is an aspect of statistical analysis that involves making inferences about unobservable underlying causal relations that are not manifested in measureable counterfactual differences in outcomes.



Unless I am misunderstanding something, I think it is fair to say that this idea is in tension with the positivist views of Box and Friedman, as represented in the quote by Box. The positivist viewpoint essentially says that there are no admissible metaphysical "constructs" beyond those that manifest in measureable outcomes. Positivism confines itself to consideration of observable data, and concepts built on this data; it excludes consideration of a priori metaphysical concepts. Thus, a positivist would argue that the concept of causality can only be valid to the extent that it is defined in terms of measureable outcomes in reality --- to the extent it is defined as something distinct from this (as Shmueli treats it), this would be regarded as metaphysical speculation, and would be treated as inadmissible in scientific discourse.



So I think you're right --- these two approaches are essentially in conflict. The positivist approach used by Box insists that valid scientific concepts be grounded entirely in their manifestations in reality, whereas the alternative approach used by Shmueli says that there are some "constructs" that are important scientific concepts (that we want to explain) but which cannot be perfectly represented when they are "operationalised" by relating them to measureable outcomes in reality.






share|cite|improve this answer











$endgroup$












  • $begingroup$
    Exactly!!! Shmueli seems to be contradicting most positivist (and falsificationist) definitions of what a scientific statement is, and I am wondering whether her intention is indeed to make such a bold philosophical statement? or whether she, as a statistician, isn't aware of how bold her statements really are?
    $endgroup$
    – Skander H.
    Jun 14 at 17:07










  • $begingroup$
    I posted a related question on the philosophy SE, care to weigh in on it?
    $endgroup$
    – Skander H.
    Jun 14 at 17:10






  • 1




    $begingroup$
    Although positivism and pragmatism share anti-realist traits, Box's approach is the latter, and positivism is not influential in sciences at least since late 1960-s. This is why Box does not mention Friedman, or other positivists. Pragmatism does not confine itself to observable data, and has no problem with a priori concepts or metaphysical constructions. It just does not see them as approaching "reality", and hence aiming at uniquely "right" features, they can be plural, task-dependent, and revisable. Causal relations are such constructs, so there is no conflict between Box and Shmueli.
    $endgroup$
    – Conifold
    Jun 14 at 21:37











  • $begingroup$
    Whether or not Box was a pragmatist, the particular quote he gives here is, in my view, more conducive to positivism than pragmatism. The latter philosophy takes a pluralistic view of metaphysics, which holds that there are multiple coherent ways of conceptualising reality, and that all these are "true" in the sense of being useful conceptualisations of reality. Thus, positivism would say, "all models are wrong, but some are useful", whereas the maxim of pragmatism would be closer to "many models are right, because they are useful".
    $endgroup$
    – Ben
    yesterday










  • $begingroup$
    The identification of true with useful is just a folk misconception about pragmatism. All models are wrong because "right" suggests that they have something to correspond to, which pragmatists deny. And the principle of tolerance, "models are right as long as they serve a purpose", is due to Carnap, the father of logical positivism.
    $endgroup$
    – Conifold
    5 hours ago













8












8








8





$begingroup$

Let me start with the pithy quote by George Box, that "all models are wrong, but some are useful". This statement is an encapsulation of the methodological approach of "positivism", which is a philosophical approach that is highly influential in the sciences. This approach is described in detail (in the context of economic theory) in the classic methodological essay of Friedman (1966). In that essay, Friedman argues that any useful scientific theory necessarily constitutes a simplification of reality, and thus its assumptions must always depart from reality to some degree, and may even depart substantially from reality. He argues that the value of a scientific theory should not be judged by the closeness of its assumptions to reality --- instead it should be judged by its simplicity in reducing the complexity of the world to a manageable set of principles, and its accuracy in making predictions about reality, and generating new testable hypotheses about reality. Thus, Friedman argues that "all models are wrong" insofar as they all contain assumptions that simplify (and therefore depart from) reality, but that "some are useful" insofar as they give a simple framework to make useful predictions about reality.



Now, if you read Box (1976) (the paper where he first states that "all models are wrong"), you will see that he does not cite Friedman, nor does he mention methodological positivism. Nevertheless, his explanation of the scientific method and its characteristics is extremely close to that developed by Friedman. In particular, both authors stress that a scientific theory will make predictions about reality that can be tested against observed facts, and the error in the prediction can then be used as a basis for revision of the theory.



Now, on to the dichotomy discussed by Galit Shmueli in Shmueli (2001). In this paper, Shmueli compares causal explanation and prediction of observed outcomes and argues that these are distinct activities. Specifically, she argues that causal relations are based on underlying constructs that do not manifest directly in measureable outcomes, and so "measurable data are not accurate representations of their underlying constructs" (p. 293). She therefore argues that there is an aspect of statistical analysis that involves making inferences about unobservable underlying causal relations that are not manifested in measureable counterfactual differences in outcomes.



Unless I am misunderstanding something, I think it is fair to say that this idea is in tension with the positivist views of Box and Friedman, as represented in the quote by Box. The positivist viewpoint essentially says that there are no admissible metaphysical "constructs" beyond those that manifest in measureable outcomes. Positivism confines itself to consideration of observable data, and concepts built on this data; it excludes consideration of a priori metaphysical concepts. Thus, a positivist would argue that the concept of causality can only be valid to the extent that it is defined in terms of measureable outcomes in reality --- to the extent it is defined as something distinct from this (as Shmueli treats it), this would be regarded as metaphysical speculation, and would be treated as inadmissible in scientific discourse.



So I think you're right --- these two approaches are essentially in conflict. The positivist approach used by Box insists that valid scientific concepts be grounded entirely in their manifestations in reality, whereas the alternative approach used by Shmueli says that there are some "constructs" that are important scientific concepts (that we want to explain) but which cannot be perfectly represented when they are "operationalised" by relating them to measureable outcomes in reality.






share|cite|improve this answer











$endgroup$



Let me start with the pithy quote by George Box, that "all models are wrong, but some are useful". This statement is an encapsulation of the methodological approach of "positivism", which is a philosophical approach that is highly influential in the sciences. This approach is described in detail (in the context of economic theory) in the classic methodological essay of Friedman (1966). In that essay, Friedman argues that any useful scientific theory necessarily constitutes a simplification of reality, and thus its assumptions must always depart from reality to some degree, and may even depart substantially from reality. He argues that the value of a scientific theory should not be judged by the closeness of its assumptions to reality --- instead it should be judged by its simplicity in reducing the complexity of the world to a manageable set of principles, and its accuracy in making predictions about reality, and generating new testable hypotheses about reality. Thus, Friedman argues that "all models are wrong" insofar as they all contain assumptions that simplify (and therefore depart from) reality, but that "some are useful" insofar as they give a simple framework to make useful predictions about reality.



Now, if you read Box (1976) (the paper where he first states that "all models are wrong"), you will see that he does not cite Friedman, nor does he mention methodological positivism. Nevertheless, his explanation of the scientific method and its characteristics is extremely close to that developed by Friedman. In particular, both authors stress that a scientific theory will make predictions about reality that can be tested against observed facts, and the error in the prediction can then be used as a basis for revision of the theory.



Now, on to the dichotomy discussed by Galit Shmueli in Shmueli (2001). In this paper, Shmueli compares causal explanation and prediction of observed outcomes and argues that these are distinct activities. Specifically, she argues that causal relations are based on underlying constructs that do not manifest directly in measureable outcomes, and so "measurable data are not accurate representations of their underlying constructs" (p. 293). She therefore argues that there is an aspect of statistical analysis that involves making inferences about unobservable underlying causal relations that are not manifested in measureable counterfactual differences in outcomes.



Unless I am misunderstanding something, I think it is fair to say that this idea is in tension with the positivist views of Box and Friedman, as represented in the quote by Box. The positivist viewpoint essentially says that there are no admissible metaphysical "constructs" beyond those that manifest in measureable outcomes. Positivism confines itself to consideration of observable data, and concepts built on this data; it excludes consideration of a priori metaphysical concepts. Thus, a positivist would argue that the concept of causality can only be valid to the extent that it is defined in terms of measureable outcomes in reality --- to the extent it is defined as something distinct from this (as Shmueli treats it), this would be regarded as metaphysical speculation, and would be treated as inadmissible in scientific discourse.



So I think you're right --- these two approaches are essentially in conflict. The positivist approach used by Box insists that valid scientific concepts be grounded entirely in their manifestations in reality, whereas the alternative approach used by Shmueli says that there are some "constructs" that are important scientific concepts (that we want to explain) but which cannot be perfectly represented when they are "operationalised" by relating them to measureable outcomes in reality.







share|cite|improve this answer














share|cite|improve this answer



share|cite|improve this answer








edited Jun 14 at 11:02









kjetil b halvorsen

34.7k9 gold badges90 silver badges265 bronze badges




34.7k9 gold badges90 silver badges265 bronze badges










answered Jun 14 at 10:50









BenBen

32.8k2 gold badges38 silver badges144 bronze badges




32.8k2 gold badges38 silver badges144 bronze badges











  • $begingroup$
    Exactly!!! Shmueli seems to be contradicting most positivist (and falsificationist) definitions of what a scientific statement is, and I am wondering whether her intention is indeed to make such a bold philosophical statement? or whether she, as a statistician, isn't aware of how bold her statements really are?
    $endgroup$
    – Skander H.
    Jun 14 at 17:07










  • $begingroup$
    I posted a related question on the philosophy SE, care to weigh in on it?
    $endgroup$
    – Skander H.
    Jun 14 at 17:10






  • 1




    $begingroup$
    Although positivism and pragmatism share anti-realist traits, Box's approach is the latter, and positivism is not influential in sciences at least since late 1960-s. This is why Box does not mention Friedman, or other positivists. Pragmatism does not confine itself to observable data, and has no problem with a priori concepts or metaphysical constructions. It just does not see them as approaching "reality", and hence aiming at uniquely "right" features, they can be plural, task-dependent, and revisable. Causal relations are such constructs, so there is no conflict between Box and Shmueli.
    $endgroup$
    – Conifold
    Jun 14 at 21:37











  • $begingroup$
    Whether or not Box was a pragmatist, the particular quote he gives here is, in my view, more conducive to positivism than pragmatism. The latter philosophy takes a pluralistic view of metaphysics, which holds that there are multiple coherent ways of conceptualising reality, and that all these are "true" in the sense of being useful conceptualisations of reality. Thus, positivism would say, "all models are wrong, but some are useful", whereas the maxim of pragmatism would be closer to "many models are right, because they are useful".
    $endgroup$
    – Ben
    yesterday










  • $begingroup$
    The identification of true with useful is just a folk misconception about pragmatism. All models are wrong because "right" suggests that they have something to correspond to, which pragmatists deny. And the principle of tolerance, "models are right as long as they serve a purpose", is due to Carnap, the father of logical positivism.
    $endgroup$
    – Conifold
    5 hours ago
















  • $begingroup$
    Exactly!!! Shmueli seems to be contradicting most positivist (and falsificationist) definitions of what a scientific statement is, and I am wondering whether her intention is indeed to make such a bold philosophical statement? or whether she, as a statistician, isn't aware of how bold her statements really are?
    $endgroup$
    – Skander H.
    Jun 14 at 17:07










  • $begingroup$
    I posted a related question on the philosophy SE, care to weigh in on it?
    $endgroup$
    – Skander H.
    Jun 14 at 17:10






  • 1




    $begingroup$
    Although positivism and pragmatism share anti-realist traits, Box's approach is the latter, and positivism is not influential in sciences at least since late 1960-s. This is why Box does not mention Friedman, or other positivists. Pragmatism does not confine itself to observable data, and has no problem with a priori concepts or metaphysical constructions. It just does not see them as approaching "reality", and hence aiming at uniquely "right" features, they can be plural, task-dependent, and revisable. Causal relations are such constructs, so there is no conflict between Box and Shmueli.
    $endgroup$
    – Conifold
    Jun 14 at 21:37











  • $begingroup$
    Whether or not Box was a pragmatist, the particular quote he gives here is, in my view, more conducive to positivism than pragmatism. The latter philosophy takes a pluralistic view of metaphysics, which holds that there are multiple coherent ways of conceptualising reality, and that all these are "true" in the sense of being useful conceptualisations of reality. Thus, positivism would say, "all models are wrong, but some are useful", whereas the maxim of pragmatism would be closer to "many models are right, because they are useful".
    $endgroup$
    – Ben
    yesterday










  • $begingroup$
    The identification of true with useful is just a folk misconception about pragmatism. All models are wrong because "right" suggests that they have something to correspond to, which pragmatists deny. And the principle of tolerance, "models are right as long as they serve a purpose", is due to Carnap, the father of logical positivism.
    $endgroup$
    – Conifold
    5 hours ago















$begingroup$
Exactly!!! Shmueli seems to be contradicting most positivist (and falsificationist) definitions of what a scientific statement is, and I am wondering whether her intention is indeed to make such a bold philosophical statement? or whether she, as a statistician, isn't aware of how bold her statements really are?
$endgroup$
– Skander H.
Jun 14 at 17:07




$begingroup$
Exactly!!! Shmueli seems to be contradicting most positivist (and falsificationist) definitions of what a scientific statement is, and I am wondering whether her intention is indeed to make such a bold philosophical statement? or whether she, as a statistician, isn't aware of how bold her statements really are?
$endgroup$
– Skander H.
Jun 14 at 17:07












$begingroup$
I posted a related question on the philosophy SE, care to weigh in on it?
$endgroup$
– Skander H.
Jun 14 at 17:10




$begingroup$
I posted a related question on the philosophy SE, care to weigh in on it?
$endgroup$
– Skander H.
Jun 14 at 17:10




1




1




$begingroup$
Although positivism and pragmatism share anti-realist traits, Box's approach is the latter, and positivism is not influential in sciences at least since late 1960-s. This is why Box does not mention Friedman, or other positivists. Pragmatism does not confine itself to observable data, and has no problem with a priori concepts or metaphysical constructions. It just does not see them as approaching "reality", and hence aiming at uniquely "right" features, they can be plural, task-dependent, and revisable. Causal relations are such constructs, so there is no conflict between Box and Shmueli.
$endgroup$
– Conifold
Jun 14 at 21:37





$begingroup$
Although positivism and pragmatism share anti-realist traits, Box's approach is the latter, and positivism is not influential in sciences at least since late 1960-s. This is why Box does not mention Friedman, or other positivists. Pragmatism does not confine itself to observable data, and has no problem with a priori concepts or metaphysical constructions. It just does not see them as approaching "reality", and hence aiming at uniquely "right" features, they can be plural, task-dependent, and revisable. Causal relations are such constructs, so there is no conflict between Box and Shmueli.
$endgroup$
– Conifold
Jun 14 at 21:37













$begingroup$
Whether or not Box was a pragmatist, the particular quote he gives here is, in my view, more conducive to positivism than pragmatism. The latter philosophy takes a pluralistic view of metaphysics, which holds that there are multiple coherent ways of conceptualising reality, and that all these are "true" in the sense of being useful conceptualisations of reality. Thus, positivism would say, "all models are wrong, but some are useful", whereas the maxim of pragmatism would be closer to "many models are right, because they are useful".
$endgroup$
– Ben
yesterday




$begingroup$
Whether or not Box was a pragmatist, the particular quote he gives here is, in my view, more conducive to positivism than pragmatism. The latter philosophy takes a pluralistic view of metaphysics, which holds that there are multiple coherent ways of conceptualising reality, and that all these are "true" in the sense of being useful conceptualisations of reality. Thus, positivism would say, "all models are wrong, but some are useful", whereas the maxim of pragmatism would be closer to "many models are right, because they are useful".
$endgroup$
– Ben
yesterday












$begingroup$
The identification of true with useful is just a folk misconception about pragmatism. All models are wrong because "right" suggests that they have something to correspond to, which pragmatists deny. And the principle of tolerance, "models are right as long as they serve a purpose", is due to Carnap, the father of logical positivism.
$endgroup$
– Conifold
5 hours ago




$begingroup$
The identification of true with useful is just a folk misconception about pragmatism. All models are wrong because "right" suggests that they have something to correspond to, which pragmatists deny. And the principle of tolerance, "models are right as long as they serve a purpose", is due to Carnap, the father of logical positivism.
$endgroup$
– Conifold
5 hours ago













4












$begingroup$

A model, when used to explain things, is a simplification of reality. Simplification is just another word for "wrong in some useful way". For example, if we round the number 3.1415926535898 to 3.14 we are making an error, but this error allows us humans to focus on the most important part of that number. This is how models are used in explaining, it provides insights on some problem, but by necessity has to abstract away from a lot of other things: We humans are just not very good at looking at a thousands things simultaneously. If we primarily care about predicting we want to include those thousands things whenever possible, but with explaining the trade-off is different.






share|cite|improve this answer









$endgroup$








  • 1




    $begingroup$
    "but this error allows us humans to focus on the most important part of that number." this makes sense and helps me understand what "explanation" means, but then it also confirms my point that explanation is more of an artistic/aesthetic concept than a scientific one. Based on you $pi$ example, consider the following as well: A novel dimensionality reduction technique allows for plotting very elegant and intuitive graphs of high dimensional data, which makes for good explanation, but then there is no way to objectively asses the accuracy of this technique, its value is purely subjective.
    $endgroup$
    – Skander H.
    Jun 14 at 8:41






  • 1




    $begingroup$
    @SkanderH. To the extent that "elegant and intuitive graphs" facilitate engineering applications, or development of new theories, their value is not purely subjective, or non-scientific, it is pragmatic. Unification, explanatory power, simplicity and coherence are broadly acknowledged as epistemic, not aesthetic, values. The choice between Lorentz's theory of ether and special relativity was made based on just such considerations, they are predictively equivalent.
    $endgroup$
    – Conifold
    Jun 14 at 9:33















4












$begingroup$

A model, when used to explain things, is a simplification of reality. Simplification is just another word for "wrong in some useful way". For example, if we round the number 3.1415926535898 to 3.14 we are making an error, but this error allows us humans to focus on the most important part of that number. This is how models are used in explaining, it provides insights on some problem, but by necessity has to abstract away from a lot of other things: We humans are just not very good at looking at a thousands things simultaneously. If we primarily care about predicting we want to include those thousands things whenever possible, but with explaining the trade-off is different.






share|cite|improve this answer









$endgroup$








  • 1




    $begingroup$
    "but this error allows us humans to focus on the most important part of that number." this makes sense and helps me understand what "explanation" means, but then it also confirms my point that explanation is more of an artistic/aesthetic concept than a scientific one. Based on you $pi$ example, consider the following as well: A novel dimensionality reduction technique allows for plotting very elegant and intuitive graphs of high dimensional data, which makes for good explanation, but then there is no way to objectively asses the accuracy of this technique, its value is purely subjective.
    $endgroup$
    – Skander H.
    Jun 14 at 8:41






  • 1




    $begingroup$
    @SkanderH. To the extent that "elegant and intuitive graphs" facilitate engineering applications, or development of new theories, their value is not purely subjective, or non-scientific, it is pragmatic. Unification, explanatory power, simplicity and coherence are broadly acknowledged as epistemic, not aesthetic, values. The choice between Lorentz's theory of ether and special relativity was made based on just such considerations, they are predictively equivalent.
    $endgroup$
    – Conifold
    Jun 14 at 9:33













4












4








4





$begingroup$

A model, when used to explain things, is a simplification of reality. Simplification is just another word for "wrong in some useful way". For example, if we round the number 3.1415926535898 to 3.14 we are making an error, but this error allows us humans to focus on the most important part of that number. This is how models are used in explaining, it provides insights on some problem, but by necessity has to abstract away from a lot of other things: We humans are just not very good at looking at a thousands things simultaneously. If we primarily care about predicting we want to include those thousands things whenever possible, but with explaining the trade-off is different.






share|cite|improve this answer









$endgroup$



A model, when used to explain things, is a simplification of reality. Simplification is just another word for "wrong in some useful way". For example, if we round the number 3.1415926535898 to 3.14 we are making an error, but this error allows us humans to focus on the most important part of that number. This is how models are used in explaining, it provides insights on some problem, but by necessity has to abstract away from a lot of other things: We humans are just not very good at looking at a thousands things simultaneously. If we primarily care about predicting we want to include those thousands things whenever possible, but with explaining the trade-off is different.







share|cite|improve this answer












share|cite|improve this answer



share|cite|improve this answer










answered Jun 14 at 8:34









Maarten BuisMaarten Buis

16.4k19 silver badges43 bronze badges




16.4k19 silver badges43 bronze badges







  • 1




    $begingroup$
    "but this error allows us humans to focus on the most important part of that number." this makes sense and helps me understand what "explanation" means, but then it also confirms my point that explanation is more of an artistic/aesthetic concept than a scientific one. Based on you $pi$ example, consider the following as well: A novel dimensionality reduction technique allows for plotting very elegant and intuitive graphs of high dimensional data, which makes for good explanation, but then there is no way to objectively asses the accuracy of this technique, its value is purely subjective.
    $endgroup$
    – Skander H.
    Jun 14 at 8:41






  • 1




    $begingroup$
    @SkanderH. To the extent that "elegant and intuitive graphs" facilitate engineering applications, or development of new theories, their value is not purely subjective, or non-scientific, it is pragmatic. Unification, explanatory power, simplicity and coherence are broadly acknowledged as epistemic, not aesthetic, values. The choice between Lorentz's theory of ether and special relativity was made based on just such considerations, they are predictively equivalent.
    $endgroup$
    – Conifold
    Jun 14 at 9:33












  • 1




    $begingroup$
    "but this error allows us humans to focus on the most important part of that number." this makes sense and helps me understand what "explanation" means, but then it also confirms my point that explanation is more of an artistic/aesthetic concept than a scientific one. Based on you $pi$ example, consider the following as well: A novel dimensionality reduction technique allows for plotting very elegant and intuitive graphs of high dimensional data, which makes for good explanation, but then there is no way to objectively asses the accuracy of this technique, its value is purely subjective.
    $endgroup$
    – Skander H.
    Jun 14 at 8:41






  • 1




    $begingroup$
    @SkanderH. To the extent that "elegant and intuitive graphs" facilitate engineering applications, or development of new theories, their value is not purely subjective, or non-scientific, it is pragmatic. Unification, explanatory power, simplicity and coherence are broadly acknowledged as epistemic, not aesthetic, values. The choice between Lorentz's theory of ether and special relativity was made based on just such considerations, they are predictively equivalent.
    $endgroup$
    – Conifold
    Jun 14 at 9:33







1




1




$begingroup$
"but this error allows us humans to focus on the most important part of that number." this makes sense and helps me understand what "explanation" means, but then it also confirms my point that explanation is more of an artistic/aesthetic concept than a scientific one. Based on you $pi$ example, consider the following as well: A novel dimensionality reduction technique allows for plotting very elegant and intuitive graphs of high dimensional data, which makes for good explanation, but then there is no way to objectively asses the accuracy of this technique, its value is purely subjective.
$endgroup$
– Skander H.
Jun 14 at 8:41




$begingroup$
"but this error allows us humans to focus on the most important part of that number." this makes sense and helps me understand what "explanation" means, but then it also confirms my point that explanation is more of an artistic/aesthetic concept than a scientific one. Based on you $pi$ example, consider the following as well: A novel dimensionality reduction technique allows for plotting very elegant and intuitive graphs of high dimensional data, which makes for good explanation, but then there is no way to objectively asses the accuracy of this technique, its value is purely subjective.
$endgroup$
– Skander H.
Jun 14 at 8:41




1




1




$begingroup$
@SkanderH. To the extent that "elegant and intuitive graphs" facilitate engineering applications, or development of new theories, their value is not purely subjective, or non-scientific, it is pragmatic. Unification, explanatory power, simplicity and coherence are broadly acknowledged as epistemic, not aesthetic, values. The choice between Lorentz's theory of ether and special relativity was made based on just such considerations, they are predictively equivalent.
$endgroup$
– Conifold
Jun 14 at 9:33




$begingroup$
@SkanderH. To the extent that "elegant and intuitive graphs" facilitate engineering applications, or development of new theories, their value is not purely subjective, or non-scientific, it is pragmatic. Unification, explanatory power, simplicity and coherence are broadly acknowledged as epistemic, not aesthetic, values. The choice between Lorentz's theory of ether and special relativity was made based on just such considerations, they are predictively equivalent.
$endgroup$
– Conifold
Jun 14 at 9:33











2












$begingroup$

An example of a model that is excellent at prediction but does not explain anything is given in the Wikipedia article “All models are wrong”. The example is Newton’s model of gravitation. Newton’s model almost always gives predictions that are indistinguishable from empirical observations. Yet the model is extremely implausible: because it postulates a force that can act instantaneously over arbitrarily large distances.



Newton’s model has been supplanted by the model given in Einstein’s general theory of relativity. With general relativity, gravitational forces travel through space at finite speed (the speed of light).



Newton’s model is not a simplification of the general-relativistic model. To illustrate that, consider an apple falling down from a tree. According to general relativity, the apple falls without Earth exerting any force on the apple. (The primary reason the apple falls is that Earth warps time, so that clocks near the base of the tree run more slowly than clocks high up in the tree.) Thus, as the Wikipedia article notes, Newton’s model is completely wrong from an explanatory perspective.



The paper by Shmueli [2010] presumes that there are two purposes for a model: prediction and explanation. In fact, several authors have stated that there are three purposes (see e.g. Konishi & Kitagawa [Information Criteria and Statistical Modeling, 2008: §1.1] and Friendly & Meyer [Discrete Data Analysis, 2016: §11.6]). The three purposes correspond to the three kinds of logical reasoning:



  • prediction (corresponding to deduction);

  • parameter estimation (corresponding to induction);

  • description of structure (corresponding to abduction).





share|cite|improve this answer










New contributor



SolidPhase is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
Check out our Code of Conduct.





$endgroup$












  • $begingroup$
    To say that Newton's model of gravitation "does not explain anything" is, frankly, ludicrous. -1.
    $endgroup$
    – amoeba
    Jun 16 at 20:45










  • $begingroup$
    amoeba, Newton's model does not explain anything about how gravitation works, under the assumption that general relativity is accurate. If an apple falls, Newton's model postulates that Earth exerts a force on the apple and that postulate is completely false. I ask you to consider my answer further. If you still do not understand, then kindly tell me what is unclear.
    $endgroup$
    – SolidPhase
    Jun 16 at 20:53











  • $begingroup$
    What you say is quite clear but I strongly disagree with it.
    $endgroup$
    – amoeba
    Jun 16 at 22:07










  • $begingroup$
    amoeba, I ask you to explain why you disagree: do you have a reason? (Note that I have added an extra sentence to the answer.)
    $endgroup$
    – SolidPhase
    Jun 16 at 22:16











  • $begingroup$
    Thanks. I will look up the references you mention. I understand how a model can predict even if it doesn't explain. What I don't get is the opposite direction: How can a model explain without predicting. You Newton vs. Einstein examples just muddies everything even more: The whole reason Einstein's theory supplanted Newton's was because it predicted better. Look at it another way: If we have competing explanatory models, how can we evaluated them unless we test which one has the most predictive power?
    $endgroup$
    – Skander H.
    Jun 17 at 5:35















2












$begingroup$

An example of a model that is excellent at prediction but does not explain anything is given in the Wikipedia article “All models are wrong”. The example is Newton’s model of gravitation. Newton’s model almost always gives predictions that are indistinguishable from empirical observations. Yet the model is extremely implausible: because it postulates a force that can act instantaneously over arbitrarily large distances.



Newton’s model has been supplanted by the model given in Einstein’s general theory of relativity. With general relativity, gravitational forces travel through space at finite speed (the speed of light).



Newton’s model is not a simplification of the general-relativistic model. To illustrate that, consider an apple falling down from a tree. According to general relativity, the apple falls without Earth exerting any force on the apple. (The primary reason the apple falls is that Earth warps time, so that clocks near the base of the tree run more slowly than clocks high up in the tree.) Thus, as the Wikipedia article notes, Newton’s model is completely wrong from an explanatory perspective.



The paper by Shmueli [2010] presumes that there are two purposes for a model: prediction and explanation. In fact, several authors have stated that there are three purposes (see e.g. Konishi & Kitagawa [Information Criteria and Statistical Modeling, 2008: §1.1] and Friendly & Meyer [Discrete Data Analysis, 2016: §11.6]). The three purposes correspond to the three kinds of logical reasoning:



  • prediction (corresponding to deduction);

  • parameter estimation (corresponding to induction);

  • description of structure (corresponding to abduction).





share|cite|improve this answer










New contributor



SolidPhase is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
Check out our Code of Conduct.





$endgroup$












  • $begingroup$
    To say that Newton's model of gravitation "does not explain anything" is, frankly, ludicrous. -1.
    $endgroup$
    – amoeba
    Jun 16 at 20:45










  • $begingroup$
    amoeba, Newton's model does not explain anything about how gravitation works, under the assumption that general relativity is accurate. If an apple falls, Newton's model postulates that Earth exerts a force on the apple and that postulate is completely false. I ask you to consider my answer further. If you still do not understand, then kindly tell me what is unclear.
    $endgroup$
    – SolidPhase
    Jun 16 at 20:53











  • $begingroup$
    What you say is quite clear but I strongly disagree with it.
    $endgroup$
    – amoeba
    Jun 16 at 22:07










  • $begingroup$
    amoeba, I ask you to explain why you disagree: do you have a reason? (Note that I have added an extra sentence to the answer.)
    $endgroup$
    – SolidPhase
    Jun 16 at 22:16











  • $begingroup$
    Thanks. I will look up the references you mention. I understand how a model can predict even if it doesn't explain. What I don't get is the opposite direction: How can a model explain without predicting. You Newton vs. Einstein examples just muddies everything even more: The whole reason Einstein's theory supplanted Newton's was because it predicted better. Look at it another way: If we have competing explanatory models, how can we evaluated them unless we test which one has the most predictive power?
    $endgroup$
    – Skander H.
    Jun 17 at 5:35













2












2








2





$begingroup$

An example of a model that is excellent at prediction but does not explain anything is given in the Wikipedia article “All models are wrong”. The example is Newton’s model of gravitation. Newton’s model almost always gives predictions that are indistinguishable from empirical observations. Yet the model is extremely implausible: because it postulates a force that can act instantaneously over arbitrarily large distances.



Newton’s model has been supplanted by the model given in Einstein’s general theory of relativity. With general relativity, gravitational forces travel through space at finite speed (the speed of light).



Newton’s model is not a simplification of the general-relativistic model. To illustrate that, consider an apple falling down from a tree. According to general relativity, the apple falls without Earth exerting any force on the apple. (The primary reason the apple falls is that Earth warps time, so that clocks near the base of the tree run more slowly than clocks high up in the tree.) Thus, as the Wikipedia article notes, Newton’s model is completely wrong from an explanatory perspective.



The paper by Shmueli [2010] presumes that there are two purposes for a model: prediction and explanation. In fact, several authors have stated that there are three purposes (see e.g. Konishi & Kitagawa [Information Criteria and Statistical Modeling, 2008: §1.1] and Friendly & Meyer [Discrete Data Analysis, 2016: §11.6]). The three purposes correspond to the three kinds of logical reasoning:



  • prediction (corresponding to deduction);

  • parameter estimation (corresponding to induction);

  • description of structure (corresponding to abduction).





share|cite|improve this answer










New contributor



SolidPhase is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
Check out our Code of Conduct.





$endgroup$



An example of a model that is excellent at prediction but does not explain anything is given in the Wikipedia article “All models are wrong”. The example is Newton’s model of gravitation. Newton’s model almost always gives predictions that are indistinguishable from empirical observations. Yet the model is extremely implausible: because it postulates a force that can act instantaneously over arbitrarily large distances.



Newton’s model has been supplanted by the model given in Einstein’s general theory of relativity. With general relativity, gravitational forces travel through space at finite speed (the speed of light).



Newton’s model is not a simplification of the general-relativistic model. To illustrate that, consider an apple falling down from a tree. According to general relativity, the apple falls without Earth exerting any force on the apple. (The primary reason the apple falls is that Earth warps time, so that clocks near the base of the tree run more slowly than clocks high up in the tree.) Thus, as the Wikipedia article notes, Newton’s model is completely wrong from an explanatory perspective.



The paper by Shmueli [2010] presumes that there are two purposes for a model: prediction and explanation. In fact, several authors have stated that there are three purposes (see e.g. Konishi & Kitagawa [Information Criteria and Statistical Modeling, 2008: §1.1] and Friendly & Meyer [Discrete Data Analysis, 2016: §11.6]). The three purposes correspond to the three kinds of logical reasoning:



  • prediction (corresponding to deduction);

  • parameter estimation (corresponding to induction);

  • description of structure (corresponding to abduction).






share|cite|improve this answer










New contributor



SolidPhase is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
Check out our Code of Conduct.








share|cite|improve this answer



share|cite|improve this answer








edited Jun 16 at 22:14





















New contributor



SolidPhase is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
Check out our Code of Conduct.








answered Jun 16 at 18:49









SolidPhaseSolidPhase

294 bronze badges




294 bronze badges




New contributor



SolidPhase is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
Check out our Code of Conduct.




New contributor




SolidPhase is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
Check out our Code of Conduct.













  • $begingroup$
    To say that Newton's model of gravitation "does not explain anything" is, frankly, ludicrous. -1.
    $endgroup$
    – amoeba
    Jun 16 at 20:45










  • $begingroup$
    amoeba, Newton's model does not explain anything about how gravitation works, under the assumption that general relativity is accurate. If an apple falls, Newton's model postulates that Earth exerts a force on the apple and that postulate is completely false. I ask you to consider my answer further. If you still do not understand, then kindly tell me what is unclear.
    $endgroup$
    – SolidPhase
    Jun 16 at 20:53











  • $begingroup$
    What you say is quite clear but I strongly disagree with it.
    $endgroup$
    – amoeba
    Jun 16 at 22:07










  • $begingroup$
    amoeba, I ask you to explain why you disagree: do you have a reason? (Note that I have added an extra sentence to the answer.)
    $endgroup$
    – SolidPhase
    Jun 16 at 22:16











  • $begingroup$
    Thanks. I will look up the references you mention. I understand how a model can predict even if it doesn't explain. What I don't get is the opposite direction: How can a model explain without predicting. You Newton vs. Einstein examples just muddies everything even more: The whole reason Einstein's theory supplanted Newton's was because it predicted better. Look at it another way: If we have competing explanatory models, how can we evaluated them unless we test which one has the most predictive power?
    $endgroup$
    – Skander H.
    Jun 17 at 5:35
















  • $begingroup$
    To say that Newton's model of gravitation "does not explain anything" is, frankly, ludicrous. -1.
    $endgroup$
    – amoeba
    Jun 16 at 20:45










  • $begingroup$
    amoeba, Newton's model does not explain anything about how gravitation works, under the assumption that general relativity is accurate. If an apple falls, Newton's model postulates that Earth exerts a force on the apple and that postulate is completely false. I ask you to consider my answer further. If you still do not understand, then kindly tell me what is unclear.
    $endgroup$
    – SolidPhase
    Jun 16 at 20:53











  • $begingroup$
    What you say is quite clear but I strongly disagree with it.
    $endgroup$
    – amoeba
    Jun 16 at 22:07










  • $begingroup$
    amoeba, I ask you to explain why you disagree: do you have a reason? (Note that I have added an extra sentence to the answer.)
    $endgroup$
    – SolidPhase
    Jun 16 at 22:16











  • $begingroup$
    Thanks. I will look up the references you mention. I understand how a model can predict even if it doesn't explain. What I don't get is the opposite direction: How can a model explain without predicting. You Newton vs. Einstein examples just muddies everything even more: The whole reason Einstein's theory supplanted Newton's was because it predicted better. Look at it another way: If we have competing explanatory models, how can we evaluated them unless we test which one has the most predictive power?
    $endgroup$
    – Skander H.
    Jun 17 at 5:35















$begingroup$
To say that Newton's model of gravitation "does not explain anything" is, frankly, ludicrous. -1.
$endgroup$
– amoeba
Jun 16 at 20:45




$begingroup$
To say that Newton's model of gravitation "does not explain anything" is, frankly, ludicrous. -1.
$endgroup$
– amoeba
Jun 16 at 20:45












$begingroup$
amoeba, Newton's model does not explain anything about how gravitation works, under the assumption that general relativity is accurate. If an apple falls, Newton's model postulates that Earth exerts a force on the apple and that postulate is completely false. I ask you to consider my answer further. If you still do not understand, then kindly tell me what is unclear.
$endgroup$
– SolidPhase
Jun 16 at 20:53





$begingroup$
amoeba, Newton's model does not explain anything about how gravitation works, under the assumption that general relativity is accurate. If an apple falls, Newton's model postulates that Earth exerts a force on the apple and that postulate is completely false. I ask you to consider my answer further. If you still do not understand, then kindly tell me what is unclear.
$endgroup$
– SolidPhase
Jun 16 at 20:53













$begingroup$
What you say is quite clear but I strongly disagree with it.
$endgroup$
– amoeba
Jun 16 at 22:07




$begingroup$
What you say is quite clear but I strongly disagree with it.
$endgroup$
– amoeba
Jun 16 at 22:07












$begingroup$
amoeba, I ask you to explain why you disagree: do you have a reason? (Note that I have added an extra sentence to the answer.)
$endgroup$
– SolidPhase
Jun 16 at 22:16





$begingroup$
amoeba, I ask you to explain why you disagree: do you have a reason? (Note that I have added an extra sentence to the answer.)
$endgroup$
– SolidPhase
Jun 16 at 22:16













$begingroup$
Thanks. I will look up the references you mention. I understand how a model can predict even if it doesn't explain. What I don't get is the opposite direction: How can a model explain without predicting. You Newton vs. Einstein examples just muddies everything even more: The whole reason Einstein's theory supplanted Newton's was because it predicted better. Look at it another way: If we have competing explanatory models, how can we evaluated them unless we test which one has the most predictive power?
$endgroup$
– Skander H.
Jun 17 at 5:35




$begingroup$
Thanks. I will look up the references you mention. I understand how a model can predict even if it doesn't explain. What I don't get is the opposite direction: How can a model explain without predicting. You Newton vs. Einstein examples just muddies everything even more: The whole reason Einstein's theory supplanted Newton's was because it predicted better. Look at it another way: If we have competing explanatory models, how can we evaluated them unless we test which one has the most predictive power?
$endgroup$
– Skander H.
Jun 17 at 5:35











1












$begingroup$

I'm an undergraduate in Statistics, so I won't call myself an expert, but here are my two cents.



Models don't explain themselves; humans interpret them. Linear models are easier to understand than neural networks and random forests because they are closer to how we make decisions. Indeed, ANNs imitate the human brain, but you don't decide which restaurant to go tomorrow by doing a series of matrix multiplications. Instead, you weight some factors in your mind by their importance, which is essentially a linear combination.



"Explanatory power" measures how well a model gets along with humans' intuition, whereas "predictive power" measures how well it aligns with the underlying mechanism of the process in interest. The contradiction between them is essentially the gap between what the world is and how we can perceive/understand it. I hope this explains why "some models do a good job of explaining, even though they do a poor job at predicting".



Ian Stewart once said, "If our brains were simple enough for us to understand them, we'd be so simple that we couldn't." Unfortunately, our little human brains are actually very simple compared to the universe, or even a stock market (which involves a lot of brains :). Up to now, all models are products of human brains, so it must be more-or-less inaccurate, which leads to Box's "All models are wrong". On the other hand, a model doesn't have to be technically correct to be useful. For example, Newton's laws of motion has been disproved by Einstein, but it remains useful when an object is not ridiculously big or fast.



To address your question, I honestly can't see the incompatibility between Box and Shmueli's points. It seems that you consider "explanatory power" and "predictive power" to be binomial properties, but I think they sit at the two ends of a spectrum.






share|cite|improve this answer











$endgroup$












  • $begingroup$
    @RichardHardy Please see my edited answer.
    $endgroup$
    – nalzok
    Jun 14 at 13:11















1












$begingroup$

I'm an undergraduate in Statistics, so I won't call myself an expert, but here are my two cents.



Models don't explain themselves; humans interpret them. Linear models are easier to understand than neural networks and random forests because they are closer to how we make decisions. Indeed, ANNs imitate the human brain, but you don't decide which restaurant to go tomorrow by doing a series of matrix multiplications. Instead, you weight some factors in your mind by their importance, which is essentially a linear combination.



"Explanatory power" measures how well a model gets along with humans' intuition, whereas "predictive power" measures how well it aligns with the underlying mechanism of the process in interest. The contradiction between them is essentially the gap between what the world is and how we can perceive/understand it. I hope this explains why "some models do a good job of explaining, even though they do a poor job at predicting".



Ian Stewart once said, "If our brains were simple enough for us to understand them, we'd be so simple that we couldn't." Unfortunately, our little human brains are actually very simple compared to the universe, or even a stock market (which involves a lot of brains :). Up to now, all models are products of human brains, so it must be more-or-less inaccurate, which leads to Box's "All models are wrong". On the other hand, a model doesn't have to be technically correct to be useful. For example, Newton's laws of motion has been disproved by Einstein, but it remains useful when an object is not ridiculously big or fast.



To address your question, I honestly can't see the incompatibility between Box and Shmueli's points. It seems that you consider "explanatory power" and "predictive power" to be binomial properties, but I think they sit at the two ends of a spectrum.






share|cite|improve this answer











$endgroup$












  • $begingroup$
    @RichardHardy Please see my edited answer.
    $endgroup$
    – nalzok
    Jun 14 at 13:11













1












1








1





$begingroup$

I'm an undergraduate in Statistics, so I won't call myself an expert, but here are my two cents.



Models don't explain themselves; humans interpret them. Linear models are easier to understand than neural networks and random forests because they are closer to how we make decisions. Indeed, ANNs imitate the human brain, but you don't decide which restaurant to go tomorrow by doing a series of matrix multiplications. Instead, you weight some factors in your mind by their importance, which is essentially a linear combination.



"Explanatory power" measures how well a model gets along with humans' intuition, whereas "predictive power" measures how well it aligns with the underlying mechanism of the process in interest. The contradiction between them is essentially the gap between what the world is and how we can perceive/understand it. I hope this explains why "some models do a good job of explaining, even though they do a poor job at predicting".



Ian Stewart once said, "If our brains were simple enough for us to understand them, we'd be so simple that we couldn't." Unfortunately, our little human brains are actually very simple compared to the universe, or even a stock market (which involves a lot of brains :). Up to now, all models are products of human brains, so it must be more-or-less inaccurate, which leads to Box's "All models are wrong". On the other hand, a model doesn't have to be technically correct to be useful. For example, Newton's laws of motion has been disproved by Einstein, but it remains useful when an object is not ridiculously big or fast.



To address your question, I honestly can't see the incompatibility between Box and Shmueli's points. It seems that you consider "explanatory power" and "predictive power" to be binomial properties, but I think they sit at the two ends of a spectrum.






share|cite|improve this answer











$endgroup$



I'm an undergraduate in Statistics, so I won't call myself an expert, but here are my two cents.



Models don't explain themselves; humans interpret them. Linear models are easier to understand than neural networks and random forests because they are closer to how we make decisions. Indeed, ANNs imitate the human brain, but you don't decide which restaurant to go tomorrow by doing a series of matrix multiplications. Instead, you weight some factors in your mind by their importance, which is essentially a linear combination.



"Explanatory power" measures how well a model gets along with humans' intuition, whereas "predictive power" measures how well it aligns with the underlying mechanism of the process in interest. The contradiction between them is essentially the gap between what the world is and how we can perceive/understand it. I hope this explains why "some models do a good job of explaining, even though they do a poor job at predicting".



Ian Stewart once said, "If our brains were simple enough for us to understand them, we'd be so simple that we couldn't." Unfortunately, our little human brains are actually very simple compared to the universe, or even a stock market (which involves a lot of brains :). Up to now, all models are products of human brains, so it must be more-or-less inaccurate, which leads to Box's "All models are wrong". On the other hand, a model doesn't have to be technically correct to be useful. For example, Newton's laws of motion has been disproved by Einstein, but it remains useful when an object is not ridiculously big or fast.



To address your question, I honestly can't see the incompatibility between Box and Shmueli's points. It seems that you consider "explanatory power" and "predictive power" to be binomial properties, but I think they sit at the two ends of a spectrum.







share|cite|improve this answer














share|cite|improve this answer



share|cite|improve this answer








edited Jun 14 at 13:11

























answered Jun 14 at 10:57









nalzoknalzok

2642 silver badges13 bronze badges




2642 silver badges13 bronze badges











  • $begingroup$
    @RichardHardy Please see my edited answer.
    $endgroup$
    – nalzok
    Jun 14 at 13:11
















  • $begingroup$
    @RichardHardy Please see my edited answer.
    $endgroup$
    – nalzok
    Jun 14 at 13:11















$begingroup$
@RichardHardy Please see my edited answer.
$endgroup$
– nalzok
Jun 14 at 13:11




$begingroup$
@RichardHardy Please see my edited answer.
$endgroup$
– nalzok
Jun 14 at 13:11

















draft saved

draft discarded
















































Thanks for contributing an answer to Cross Validated!


  • Please be sure to answer the question. Provide details and share your research!

But avoid


  • Asking for help, clarification, or responding to other answers.

  • Making statements based on opinion; back them up with references or personal experience.

Use MathJax to format equations. MathJax reference.


To learn more, see our tips on writing great answers.




draft saved


draft discarded














StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f412980%2fon-george-box-galit-shmueli-and-the-scientific-method%23new-answer', 'question_page');

);

Post as a guest















Required, but never shown





















































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown

































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown







Popular posts from this blog

Get product attribute by attribute group code in magento 2get product attribute by product attribute group in magento 2Magento 2 Log Bundle Product Data in List Page?How to get all product attribute of a attribute group of Default attribute set?Magento 2.1 Create a filter in the product grid by new attributeMagento 2 : Get Product Attribute values By GroupMagento 2 How to get all existing values for one attributeMagento 2 get custom attribute of a single product inside a pluginMagento 2.3 How to get all the Multi Source Inventory (MSI) locations collection in custom module?Magento2: how to develop rest API to get new productsGet product attribute by attribute group code ( [attribute_group_code] ) in magento 2

Category:9 (number) SubcategoriesMedia in category "9 (number)"Navigation menuUpload mediaGND ID: 4485639-8Library of Congress authority ID: sh85091979ReasonatorScholiaStatistics

Magento 2.3: How do i solve this, Not registered handle, on custom form?How can i rewrite TierPrice Block in Magento2magento 2 captcha not rendering if I override layout xmlmain.CRITICAL: Plugin class doesn't existMagento 2 : Problem while adding custom button order view page?Magento 2.2.5: Overriding Admin Controller sales/orderMagento 2.2.5: Add, Update and Delete existing products Custom OptionsMagento 2.3 : File Upload issue in UI Component FormMagento2 Not registered handleHow to configured Form Builder Js in my custom magento 2.3.0 module?Magento 2.3. How to create image upload field in an admin form