On George Box, Galit Shmueli and the scientific method?Paradox in model selection (AIC, BIC, to explain or to predict?)Practical thoughts on explanatory vs. predictive modelingWhy do AIC and BIC show inversed outputs?How to estimate the prediction power for a combination of MARS and CART model?Is it inherently invalid to use BIC for model averaging?What is the point of dividing data into training and test parts to assess prediction properties when we have AIC?Using the BIC to select a modelIf the AIC and the BIC are asymptotically equivalent to cross validation, is it possible to dispense with a test set when using them?Minimizing bias in explanatory modeling, why? (Galit Shmueli's “To Explain or to Predict”)

Building a phone charger 500 years ago

Can a stolen Android phone with USB debugging enabled have screen lock bypassed?

What is it called when at university there are two subjects being held at the same time?

Why derailleur guard is present only on more affordable bicycles

Calculate the internal angles of a regular polygon with N sides

SQL Server trusted connection additional security

Practically, how does an 'observer' collapse a wave function?

Is it okay to have an email address called "SS"?

"When you Frankenstein a team together..." - Is "Frankenstein" a new verb?

Trading stock more quickly vs. holding it

Where are the NPC mages' spellbooks in the Lost Mine of Phandelver adventure?

The falling broom handle

Is there a name for this equivalence relation?

Can only rich people become president?

Why couldn't the Romulans simply circumvent Starfleet's blockade?

I can be found near gentle green hills and stony mountains

180W Laptop charged with 45W charger, is it dead?

ASCII Expansion

Authentication versus Authorisation

What websites can be protected by an SSL certificate?

What does it mean to support the brexit deal but need more time to scrutinise it?

Can the target of Feign Death speak, move, and dodge a fireball

What are the applications of the Mean Value Theorem?

How can I use Charisma instead of Strength or Dexterity for weapon attacks without being a Hexblade warlock?



On George Box, Galit Shmueli and the scientific method?


Paradox in model selection (AIC, BIC, to explain or to predict?)Practical thoughts on explanatory vs. predictive modelingWhy do AIC and BIC show inversed outputs?How to estimate the prediction power for a combination of MARS and CART model?Is it inherently invalid to use BIC for model averaging?What is the point of dividing data into training and test parts to assess prediction properties when we have AIC?Using the BIC to select a modelIf the AIC and the BIC are asymptotically equivalent to cross validation, is it possible to dispense with a test set when using them?Minimizing bias in explanatory modeling, why? (Galit Shmueli's “To Explain or to Predict”)






.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty,.everyoneloves__bot-mid-leaderboard:empty
margin-bottom:0;









16














$begingroup$


(This question might seem like it is better suited for the Philosophy SE. I am hoping that statisticians can clarify my misconceptions about Box's and Shmueli's statements, hence I am posting it here).



George Box (of ARIMA fame) said:



"All models are wrong, but some are useful."



Galit Shmueli in her famous paper "To Explain or to Predict", argues (and cites others who agree with her) that:



Explaining and predicting are not the same, and that some models do a good job of explaining, even though they do a poor job at predicting.



I feel that these to principles are somehow contradictory.



If a model doesn't predict well, is it useful?



More importantly, if a model explains well (but doesn't necessarily predict well), then it has to be true (i.e not wrong) in some way or another. So how does that mesh with Box's "all models are wrong"?



Finally, if a model explains well, but doesn't predict well, how is it even scientific? Most scientific demarcation criteria (verificationism, falsificstionism, etc...) imply that a scientific statement has to have predictive power, or colloquially: A theory or model is correct only if it can be empirically tested (or falsified), which means that it has to predict future outcomes.



My questions:



  • Are Box's statement and Shmueli's ideas indeed contradictory, or am I missing something, e.g can a model not have predictive power yet still be useful?

  • If the statements of Box and Shmueli are not contradictory, then what does it mean for a model to be wrong and not predict well, yet still have explanatory power? Put it differently: If one takes away both correctness and predictive ability, what is left of a model?

What empirical validations are possible when a model has explanatory power, but not predictive power? Shmueli mentions things like: use the AIC for explanation and the BIC for prediction, etc,...but I don't see how that solves the problem. With predictive models, you can use the AIC, or the BIC, or the $R^2$, or $L1$ regularization, etc...but ultimately out of sample testing and performance in production is what determines the quality of the model. But for models that explain well, I don't see how any loss function can ever truly evaluate a model. In philosophy of science, there is the concept of underdetermination which seems pertinent here: For any given data set, one can always judiciously choose some distribution (or mixture of distributions) and loss function $L$ in such a way that they fit the data (and therefore can be claimed to explain it). Moreover, the threshold that $L$ should be under for someone to claim that the model adequately explains the data is arbitrary (kind of like p-values, why is it $p < 0.05$ and not $p < 0.1$ or $p < 0.01$?).



  • Based on the above, how can one objectively validate a model that explains well, but doesn't predict well, since out of sample testing is not possible?









share|cite|improve this question












$endgroup$














  • $begingroup$
    See also a related question: Paradox in model selection (AIC, BIC, to explain or to predict?).
    $endgroup$
    – Richard Hardy
    Jun 14 at 8:27







  • 3




    $begingroup$
    As far as I can remember Shmueli, for her to explain well means to get the functional form right (but possibly have huge estimation imprecission when it comes to parameters of that functional form), while to predict well means to get the bias-variance tradeoff right (compromise on the functional form so as to increase estimation precision). Also, the tag philosophical could be useful here.
    $endgroup$
    – Richard Hardy
    Jun 14 at 8:28







  • 1




    $begingroup$
    I do not think the "or" needs to be exclusive. Simplicity and unification are non-controversial selection criteria for theories of equal predictive value, and if so, sacrificing precision for them may well be justified in many contexts. Box's motto reflects a different view of science, advanced e.g. in van Fraassen's Scientific Image (and going back to Kant): it is to construct adequate/useful descriptions of the observed, not to tell literally true stories about unobservable "reality". Rough adequacy may suffice for many tasks, and "the one correct model" may well be a naive figment.
    $endgroup$
    – Conifold
    Jun 14 at 8:46






  • 2




    $begingroup$
    @Conifold Indeed, Richard Levins' loop analysis of complex causal systems (not a statistical method, though there are applications of it that relate directly to statistical predictions) sacrifices almost all precision in favor of model realism (variables & relationships between them) and generality (applicability of analytic results on a model to all variables sharing the same causal structure). See Levins, R. (1966). The Strategy of Model Building In Population Biology. American Scientist, 54(4), 421–431.
    $endgroup$
    – Alexis
    Jun 14 at 18:37











  • $begingroup$
    (+1. I'm still hoping to find time to post an answer here...)
    $endgroup$
    – amoeba
    Jun 19 at 21:59


















16














$begingroup$


(This question might seem like it is better suited for the Philosophy SE. I am hoping that statisticians can clarify my misconceptions about Box's and Shmueli's statements, hence I am posting it here).



George Box (of ARIMA fame) said:



"All models are wrong, but some are useful."



Galit Shmueli in her famous paper "To Explain or to Predict", argues (and cites others who agree with her) that:



Explaining and predicting are not the same, and that some models do a good job of explaining, even though they do a poor job at predicting.



I feel that these to principles are somehow contradictory.



If a model doesn't predict well, is it useful?



More importantly, if a model explains well (but doesn't necessarily predict well), then it has to be true (i.e not wrong) in some way or another. So how does that mesh with Box's "all models are wrong"?



Finally, if a model explains well, but doesn't predict well, how is it even scientific? Most scientific demarcation criteria (verificationism, falsificstionism, etc...) imply that a scientific statement has to have predictive power, or colloquially: A theory or model is correct only if it can be empirically tested (or falsified), which means that it has to predict future outcomes.



My questions:



  • Are Box's statement and Shmueli's ideas indeed contradictory, or am I missing something, e.g can a model not have predictive power yet still be useful?

  • If the statements of Box and Shmueli are not contradictory, then what does it mean for a model to be wrong and not predict well, yet still have explanatory power? Put it differently: If one takes away both correctness and predictive ability, what is left of a model?

What empirical validations are possible when a model has explanatory power, but not predictive power? Shmueli mentions things like: use the AIC for explanation and the BIC for prediction, etc,...but I don't see how that solves the problem. With predictive models, you can use the AIC, or the BIC, or the $R^2$, or $L1$ regularization, etc...but ultimately out of sample testing and performance in production is what determines the quality of the model. But for models that explain well, I don't see how any loss function can ever truly evaluate a model. In philosophy of science, there is the concept of underdetermination which seems pertinent here: For any given data set, one can always judiciously choose some distribution (or mixture of distributions) and loss function $L$ in such a way that they fit the data (and therefore can be claimed to explain it). Moreover, the threshold that $L$ should be under for someone to claim that the model adequately explains the data is arbitrary (kind of like p-values, why is it $p < 0.05$ and not $p < 0.1$ or $p < 0.01$?).



  • Based on the above, how can one objectively validate a model that explains well, but doesn't predict well, since out of sample testing is not possible?









share|cite|improve this question












$endgroup$














  • $begingroup$
    See also a related question: Paradox in model selection (AIC, BIC, to explain or to predict?).
    $endgroup$
    – Richard Hardy
    Jun 14 at 8:27







  • 3




    $begingroup$
    As far as I can remember Shmueli, for her to explain well means to get the functional form right (but possibly have huge estimation imprecission when it comes to parameters of that functional form), while to predict well means to get the bias-variance tradeoff right (compromise on the functional form so as to increase estimation precision). Also, the tag philosophical could be useful here.
    $endgroup$
    – Richard Hardy
    Jun 14 at 8:28







  • 1




    $begingroup$
    I do not think the "or" needs to be exclusive. Simplicity and unification are non-controversial selection criteria for theories of equal predictive value, and if so, sacrificing precision for them may well be justified in many contexts. Box's motto reflects a different view of science, advanced e.g. in van Fraassen's Scientific Image (and going back to Kant): it is to construct adequate/useful descriptions of the observed, not to tell literally true stories about unobservable "reality". Rough adequacy may suffice for many tasks, and "the one correct model" may well be a naive figment.
    $endgroup$
    – Conifold
    Jun 14 at 8:46






  • 2




    $begingroup$
    @Conifold Indeed, Richard Levins' loop analysis of complex causal systems (not a statistical method, though there are applications of it that relate directly to statistical predictions) sacrifices almost all precision in favor of model realism (variables & relationships between them) and generality (applicability of analytic results on a model to all variables sharing the same causal structure). See Levins, R. (1966). The Strategy of Model Building In Population Biology. American Scientist, 54(4), 421–431.
    $endgroup$
    – Alexis
    Jun 14 at 18:37











  • $begingroup$
    (+1. I'm still hoping to find time to post an answer here...)
    $endgroup$
    – amoeba
    Jun 19 at 21:59














16












16








16


7



$begingroup$


(This question might seem like it is better suited for the Philosophy SE. I am hoping that statisticians can clarify my misconceptions about Box's and Shmueli's statements, hence I am posting it here).



George Box (of ARIMA fame) said:



"All models are wrong, but some are useful."



Galit Shmueli in her famous paper "To Explain or to Predict", argues (and cites others who agree with her) that:



Explaining and predicting are not the same, and that some models do a good job of explaining, even though they do a poor job at predicting.



I feel that these to principles are somehow contradictory.



If a model doesn't predict well, is it useful?



More importantly, if a model explains well (but doesn't necessarily predict well), then it has to be true (i.e not wrong) in some way or another. So how does that mesh with Box's "all models are wrong"?



Finally, if a model explains well, but doesn't predict well, how is it even scientific? Most scientific demarcation criteria (verificationism, falsificstionism, etc...) imply that a scientific statement has to have predictive power, or colloquially: A theory or model is correct only if it can be empirically tested (or falsified), which means that it has to predict future outcomes.



My questions:



  • Are Box's statement and Shmueli's ideas indeed contradictory, or am I missing something, e.g can a model not have predictive power yet still be useful?

  • If the statements of Box and Shmueli are not contradictory, then what does it mean for a model to be wrong and not predict well, yet still have explanatory power? Put it differently: If one takes away both correctness and predictive ability, what is left of a model?

What empirical validations are possible when a model has explanatory power, but not predictive power? Shmueli mentions things like: use the AIC for explanation and the BIC for prediction, etc,...but I don't see how that solves the problem. With predictive models, you can use the AIC, or the BIC, or the $R^2$, or $L1$ regularization, etc...but ultimately out of sample testing and performance in production is what determines the quality of the model. But for models that explain well, I don't see how any loss function can ever truly evaluate a model. In philosophy of science, there is the concept of underdetermination which seems pertinent here: For any given data set, one can always judiciously choose some distribution (or mixture of distributions) and loss function $L$ in such a way that they fit the data (and therefore can be claimed to explain it). Moreover, the threshold that $L$ should be under for someone to claim that the model adequately explains the data is arbitrary (kind of like p-values, why is it $p < 0.05$ and not $p < 0.1$ or $p < 0.01$?).



  • Based on the above, how can one objectively validate a model that explains well, but doesn't predict well, since out of sample testing is not possible?









share|cite|improve this question












$endgroup$




(This question might seem like it is better suited for the Philosophy SE. I am hoping that statisticians can clarify my misconceptions about Box's and Shmueli's statements, hence I am posting it here).



George Box (of ARIMA fame) said:



"All models are wrong, but some are useful."



Galit Shmueli in her famous paper "To Explain or to Predict", argues (and cites others who agree with her) that:



Explaining and predicting are not the same, and that some models do a good job of explaining, even though they do a poor job at predicting.



I feel that these to principles are somehow contradictory.



If a model doesn't predict well, is it useful?



More importantly, if a model explains well (but doesn't necessarily predict well), then it has to be true (i.e not wrong) in some way or another. So how does that mesh with Box's "all models are wrong"?



Finally, if a model explains well, but doesn't predict well, how is it even scientific? Most scientific demarcation criteria (verificationism, falsificstionism, etc...) imply that a scientific statement has to have predictive power, or colloquially: A theory or model is correct only if it can be empirically tested (or falsified), which means that it has to predict future outcomes.



My questions:



  • Are Box's statement and Shmueli's ideas indeed contradictory, or am I missing something, e.g can a model not have predictive power yet still be useful?

  • If the statements of Box and Shmueli are not contradictory, then what does it mean for a model to be wrong and not predict well, yet still have explanatory power? Put it differently: If one takes away both correctness and predictive ability, what is left of a model?

What empirical validations are possible when a model has explanatory power, but not predictive power? Shmueli mentions things like: use the AIC for explanation and the BIC for prediction, etc,...but I don't see how that solves the problem. With predictive models, you can use the AIC, or the BIC, or the $R^2$, or $L1$ regularization, etc...but ultimately out of sample testing and performance in production is what determines the quality of the model. But for models that explain well, I don't see how any loss function can ever truly evaluate a model. In philosophy of science, there is the concept of underdetermination which seems pertinent here: For any given data set, one can always judiciously choose some distribution (or mixture of distributions) and loss function $L$ in such a way that they fit the data (and therefore can be claimed to explain it). Moreover, the threshold that $L$ should be under for someone to claim that the model adequately explains the data is arbitrary (kind of like p-values, why is it $p < 0.05$ and not $p < 0.1$ or $p < 0.01$?).



  • Based on the above, how can one objectively validate a model that explains well, but doesn't predict well, since out of sample testing is not possible?






predictive-models aic bic philosophical explanatory-models






share|cite|improve this question
















share|cite|improve this question













share|cite|improve this question




share|cite|improve this question








edited 2 days ago







Skander H.

















asked Jun 14 at 7:15









Skander H.Skander H.

5,44618 silver badges47 bronze badges




5,44618 silver badges47 bronze badges














  • $begingroup$
    See also a related question: Paradox in model selection (AIC, BIC, to explain or to predict?).
    $endgroup$
    – Richard Hardy
    Jun 14 at 8:27







  • 3




    $begingroup$
    As far as I can remember Shmueli, for her to explain well means to get the functional form right (but possibly have huge estimation imprecission when it comes to parameters of that functional form), while to predict well means to get the bias-variance tradeoff right (compromise on the functional form so as to increase estimation precision). Also, the tag philosophical could be useful here.
    $endgroup$
    – Richard Hardy
    Jun 14 at 8:28







  • 1




    $begingroup$
    I do not think the "or" needs to be exclusive. Simplicity and unification are non-controversial selection criteria for theories of equal predictive value, and if so, sacrificing precision for them may well be justified in many contexts. Box's motto reflects a different view of science, advanced e.g. in van Fraassen's Scientific Image (and going back to Kant): it is to construct adequate/useful descriptions of the observed, not to tell literally true stories about unobservable "reality". Rough adequacy may suffice for many tasks, and "the one correct model" may well be a naive figment.
    $endgroup$
    – Conifold
    Jun 14 at 8:46






  • 2




    $begingroup$
    @Conifold Indeed, Richard Levins' loop analysis of complex causal systems (not a statistical method, though there are applications of it that relate directly to statistical predictions) sacrifices almost all precision in favor of model realism (variables & relationships between them) and generality (applicability of analytic results on a model to all variables sharing the same causal structure). See Levins, R. (1966). The Strategy of Model Building In Population Biology. American Scientist, 54(4), 421–431.
    $endgroup$
    – Alexis
    Jun 14 at 18:37











  • $begingroup$
    (+1. I'm still hoping to find time to post an answer here...)
    $endgroup$
    – amoeba
    Jun 19 at 21:59

















  • $begingroup$
    See also a related question: Paradox in model selection (AIC, BIC, to explain or to predict?).
    $endgroup$
    – Richard Hardy
    Jun 14 at 8:27







  • 3




    $begingroup$
    As far as I can remember Shmueli, for her to explain well means to get the functional form right (but possibly have huge estimation imprecission when it comes to parameters of that functional form), while to predict well means to get the bias-variance tradeoff right (compromise on the functional form so as to increase estimation precision). Also, the tag philosophical could be useful here.
    $endgroup$
    – Richard Hardy
    Jun 14 at 8:28







  • 1




    $begingroup$
    I do not think the "or" needs to be exclusive. Simplicity and unification are non-controversial selection criteria for theories of equal predictive value, and if so, sacrificing precision for them may well be justified in many contexts. Box's motto reflects a different view of science, advanced e.g. in van Fraassen's Scientific Image (and going back to Kant): it is to construct adequate/useful descriptions of the observed, not to tell literally true stories about unobservable "reality". Rough adequacy may suffice for many tasks, and "the one correct model" may well be a naive figment.
    $endgroup$
    – Conifold
    Jun 14 at 8:46






  • 2




    $begingroup$
    @Conifold Indeed, Richard Levins' loop analysis of complex causal systems (not a statistical method, though there are applications of it that relate directly to statistical predictions) sacrifices almost all precision in favor of model realism (variables & relationships between them) and generality (applicability of analytic results on a model to all variables sharing the same causal structure). See Levins, R. (1966). The Strategy of Model Building In Population Biology. American Scientist, 54(4), 421–431.
    $endgroup$
    – Alexis
    Jun 14 at 18:37











  • $begingroup$
    (+1. I'm still hoping to find time to post an answer here...)
    $endgroup$
    – amoeba
    Jun 19 at 21:59
















$begingroup$
See also a related question: Paradox in model selection (AIC, BIC, to explain or to predict?).
$endgroup$
– Richard Hardy
Jun 14 at 8:27





$begingroup$
See also a related question: Paradox in model selection (AIC, BIC, to explain or to predict?).
$endgroup$
– Richard Hardy
Jun 14 at 8:27





3




3




$begingroup$
As far as I can remember Shmueli, for her to explain well means to get the functional form right (but possibly have huge estimation imprecission when it comes to parameters of that functional form), while to predict well means to get the bias-variance tradeoff right (compromise on the functional form so as to increase estimation precision). Also, the tag philosophical could be useful here.
$endgroup$
– Richard Hardy
Jun 14 at 8:28





$begingroup$
As far as I can remember Shmueli, for her to explain well means to get the functional form right (but possibly have huge estimation imprecission when it comes to parameters of that functional form), while to predict well means to get the bias-variance tradeoff right (compromise on the functional form so as to increase estimation precision). Also, the tag philosophical could be useful here.
$endgroup$
– Richard Hardy
Jun 14 at 8:28





1




1




$begingroup$
I do not think the "or" needs to be exclusive. Simplicity and unification are non-controversial selection criteria for theories of equal predictive value, and if so, sacrificing precision for them may well be justified in many contexts. Box's motto reflects a different view of science, advanced e.g. in van Fraassen's Scientific Image (and going back to Kant): it is to construct adequate/useful descriptions of the observed, not to tell literally true stories about unobservable "reality". Rough adequacy may suffice for many tasks, and "the one correct model" may well be a naive figment.
$endgroup$
– Conifold
Jun 14 at 8:46




$begingroup$
I do not think the "or" needs to be exclusive. Simplicity and unification are non-controversial selection criteria for theories of equal predictive value, and if so, sacrificing precision for them may well be justified in many contexts. Box's motto reflects a different view of science, advanced e.g. in van Fraassen's Scientific Image (and going back to Kant): it is to construct adequate/useful descriptions of the observed, not to tell literally true stories about unobservable "reality". Rough adequacy may suffice for many tasks, and "the one correct model" may well be a naive figment.
$endgroup$
– Conifold
Jun 14 at 8:46




2




2




$begingroup$
@Conifold Indeed, Richard Levins' loop analysis of complex causal systems (not a statistical method, though there are applications of it that relate directly to statistical predictions) sacrifices almost all precision in favor of model realism (variables & relationships between them) and generality (applicability of analytic results on a model to all variables sharing the same causal structure). See Levins, R. (1966). The Strategy of Model Building In Population Biology. American Scientist, 54(4), 421–431.
$endgroup$
– Alexis
Jun 14 at 18:37





$begingroup$
@Conifold Indeed, Richard Levins' loop analysis of complex causal systems (not a statistical method, though there are applications of it that relate directly to statistical predictions) sacrifices almost all precision in favor of model realism (variables & relationships between them) and generality (applicability of analytic results on a model to all variables sharing the same causal structure). See Levins, R. (1966). The Strategy of Model Building In Population Biology. American Scientist, 54(4), 421–431.
$endgroup$
– Alexis
Jun 14 at 18:37













$begingroup$
(+1. I'm still hoping to find time to post an answer here...)
$endgroup$
– amoeba
Jun 19 at 21:59





$begingroup$
(+1. I'm still hoping to find time to post an answer here...)
$endgroup$
– amoeba
Jun 19 at 21:59











4 Answers
4






active

oldest

votes


















9
















$begingroup$

Let me start with the pithy quote by George Box, that "all models are wrong, but some are useful". This statement is an encapsulation of the methodological approach of "positivism", which is a philosophical approach that is highly influential in the sciences. This approach is described in detail (in the context of economic theory) in the classic methodological essay of Friedman (1966). In that essay, Friedman argues that any useful scientific theory necessarily constitutes a simplification of reality, and thus its assumptions must always depart from reality to some degree, and may even depart substantially from reality. He argues that the value of a scientific theory should not be judged by the closeness of its assumptions to reality --- instead it should be judged by its simplicity in reducing the complexity of the world to a manageable set of principles, and its accuracy in making predictions about reality, and generating new testable hypotheses about reality. Thus, Friedman argues that "all models are wrong" insofar as they all contain assumptions that simplify (and therefore depart from) reality, but that "some are useful" insofar as they give a simple framework to make useful predictions about reality.



Now, if you read Box (1976) (the paper where he first states that "all models are wrong"), you will see that he does not cite Friedman, nor does he mention methodological positivism. Nevertheless, his explanation of the scientific method and its characteristics is extremely close to that developed by Friedman. In particular, both authors stress that a scientific theory will make predictions about reality that can be tested against observed facts, and the error in the prediction can then be used as a basis for revision of the theory.



Now, on to the dichotomy discussed by Galit Shmueli in Shmueli (2001). In this paper, Shmueli compares causal explanation and prediction of observed outcomes and argues that these are distinct activities. Specifically, she argues that causal relations are based on underlying constructs that do not manifest directly in measureable outcomes, and so "measurable data are not accurate representations of their underlying constructs" (p. 293). She therefore argues that there is an aspect of statistical analysis that involves making inferences about unobservable underlying causal relations that are not manifested in measureable counterfactual differences in outcomes.



Unless I am misunderstanding something, I think it is fair to say that this idea is in tension with the positivist views of Box and Friedman, as represented in the quote by Box. The positivist viewpoint essentially says that there are no admissible metaphysical "constructs" beyond those that manifest in measureable outcomes. Positivism confines itself to consideration of observable data, and concepts built on this data; it excludes consideration of a priori metaphysical concepts. Thus, a positivist would argue that the concept of causality can only be valid to the extent that it is defined in terms of measureable outcomes in reality --- to the extent it is defined as something distinct from this (as Shmueli treats it), this would be regarded as metaphysical speculation, and would be treated as inadmissible in scientific discourse.



So I think you're right --- these two approaches are essentially in conflict. The positivist approach used by Box insists that valid scientific concepts be grounded entirely in their manifestations in reality, whereas the alternative approach used by Shmueli says that there are some "constructs" that are important scientific concepts (that we want to explain) but which cannot be perfectly represented when they are "operationalised" by relating them to measureable outcomes in reality.






share|cite|improve this answer












$endgroup$














  • $begingroup$
    Exactly!!! Shmueli seems to be contradicting most positivist (and falsificationist) definitions of what a scientific statement is, and I am wondering whether her intention is indeed to make such a bold philosophical statement? or whether she, as a statistician, isn't aware of how bold her statements really are?
    $endgroup$
    – Skander H.
    Jun 14 at 17:07










  • $begingroup$
    I posted a related question on the philosophy SE, care to weigh in on it?
    $endgroup$
    – Skander H.
    Jun 14 at 17:10






  • 1




    $begingroup$
    Although positivism and pragmatism share anti-realist traits, Box's approach is the latter, and positivism is not influential in sciences at least since late 1960-s. This is why Box does not mention Friedman, or other positivists. Pragmatism does not confine itself to observable data, and has no problem with a priori concepts or metaphysical constructions. It just does not see them as approaching "reality", and hence aiming at uniquely "right" features, they can be plural, task-dependent, and revisable. Causal relations are such constructs, so there is no conflict between Box and Shmueli.
    $endgroup$
    – Conifold
    Jun 14 at 21:37











  • $begingroup$
    Whether or not Box was a pragmatist, the particular quote he gives here is, in my view, more conducive to positivism than pragmatism. The latter philosophy takes a pluralistic view of metaphysics, which holds that there are multiple coherent ways of conceptualising reality, and that all these are "true" in the sense of being useful conceptualisations of reality. Thus, positivism would say, "all models are wrong, but some are useful", whereas the maxim of pragmatism would be closer to "many models are right, because they are useful".
    $endgroup$
    – Ben
    Jun 20 at 1:15










  • $begingroup$
    The identification of true with useful is just a folk misconception about pragmatism. All models are wrong because "right" suggests that they have something to correspond to, which pragmatists deny. And the principle of tolerance, "models are right as long as they serve a purpose", is due to Carnap, the father of logical positivism.
    $endgroup$
    – Conifold
    Jun 21 at 0:08


















4
















$begingroup$

A model, when used to explain things, is a simplification of reality. Simplification is just another word for "wrong in some useful way". For example, if we round the number 3.1415926535898 to 3.14 we are making an error, but this error allows us humans to focus on the most important part of that number. This is how models are used in explaining, it provides insights on some problem, but by necessity has to abstract away from a lot of other things: We humans are just not very good at looking at a thousands things simultaneously. If we primarily care about predicting we want to include those thousands things whenever possible, but with explaining the trade-off is different.






share|cite|improve this answer










$endgroup$










  • 1




    $begingroup$
    "but this error allows us humans to focus on the most important part of that number." this makes sense and helps me understand what "explanation" means, but then it also confirms my point that explanation is more of an artistic/aesthetic concept than a scientific one. Based on you $pi$ example, consider the following as well: A novel dimensionality reduction technique allows for plotting very elegant and intuitive graphs of high dimensional data, which makes for good explanation, but then there is no way to objectively asses the accuracy of this technique, its value is purely subjective.
    $endgroup$
    – Skander H.
    Jun 14 at 8:41






  • 1




    $begingroup$
    @SkanderH. To the extent that "elegant and intuitive graphs" facilitate engineering applications, or development of new theories, their value is not purely subjective, or non-scientific, it is pragmatic. Unification, explanatory power, simplicity and coherence are broadly acknowledged as epistemic, not aesthetic, values. The choice between Lorentz's theory of ether and special relativity was made based on just such considerations, they are predictively equivalent.
    $endgroup$
    – Conifold
    Jun 14 at 9:33


















3
















$begingroup$

An example of a model that is excellent at prediction but does not explain anything is given in the Wikipedia article “All models are wrong”. The example is Newton’s model of gravitation. Newton’s model almost always gives predictions that are indistinguishable from empirical observations. Yet the model is extremely implausible: because it postulates a force that can act instantaneously over arbitrarily large distances.



Newton’s model has been supplanted by the model given in Einstein’s general theory of relativity. With general relativity, gravitational forces travel through space at finite speed (the speed of light).



Newton’s model is not a simplification of the general-relativistic model. To illustrate that, consider an apple falling down from a tree. According to general relativity, the apple falls without Earth exerting any force on the apple. (The primary reason the apple falls is that Earth warps time, so that clocks near the base of the tree run more slowly than clocks high up in the tree.) Thus, as the Wikipedia article notes, Newton’s model is completely wrong from an explanatory perspective.



The paper by Shmueli [2010] presumes that there are two purposes for a model: prediction and explanation. In fact, several authors have stated that there are three purposes (see e.g. Konishi & Kitagawa [Information Criteria and Statistical Modeling, 2008: §1.1] and Friendly & Meyer [Discrete Data Analysis, 2016: §11.6]). The three purposes correspond to the three kinds of logical reasoning:



  • prediction (corresponding to deduction);

  • parameter estimation (corresponding to induction);

  • description of structure (corresponding to abduction).





share|cite|improve this answer












$endgroup$














  • $begingroup$
    To say that Newton's model of gravitation "does not explain anything" is, frankly, ludicrous. -1.
    $endgroup$
    – amoeba
    Jun 16 at 20:45










  • $begingroup$
    amoeba, Newton's model does not explain anything about how gravitation works, under the assumption that general relativity is accurate. If an apple falls, Newton's model postulates that Earth exerts a force on the apple and that postulate is completely false. I ask you to consider my answer further. If you still do not understand, then kindly tell me what is unclear.
    $endgroup$
    – SolidPhase
    Jun 16 at 20:53











  • $begingroup$
    What you say is quite clear but I strongly disagree with it.
    $endgroup$
    – amoeba
    Jun 16 at 22:07










  • $begingroup$
    amoeba, I ask you to explain why you disagree: do you have a reason? (Note that I have added an extra sentence to the answer.)
    $endgroup$
    – SolidPhase
    Jun 16 at 22:16











  • $begingroup$
    Thanks. I will look up the references you mention. I understand how a model can predict even if it doesn't explain. What I don't get is the opposite direction: How can a model explain without predicting. You Newton vs. Einstein examples just muddies everything even more: The whole reason Einstein's theory supplanted Newton's was because it predicted better. Look at it another way: If we have competing explanatory models, how can we evaluated them unless we test which one has the most predictive power?
    $endgroup$
    – Skander H.
    Jun 17 at 5:35


















1
















$begingroup$

I'm an undergraduate in Statistics, so I won't call myself an expert, but here are my two cents.



Models don't explain themselves; humans interpret them. Linear models are easier to understand than neural networks and random forests because they are closer to how we make decisions. Indeed, ANNs imitate the human brain, but you don't decide which restaurant to go tomorrow by doing a series of matrix multiplications. Instead, you weight some factors in your mind by their importance, which is essentially a linear combination.



"Explanatory power" measures how well a model gets along with humans' intuition, whereas "predictive power" measures how well it aligns with the underlying mechanism of the process in interest. The contradiction between them is essentially the gap between what the world is and how we can perceive/understand it. I hope this explains why "some models do a good job of explaining, even though they do a poor job at predicting".



Ian Stewart once said, "If our brains were simple enough for us to understand them, we'd be so simple that we couldn't." Unfortunately, our little human brains are actually very simple compared to the universe, or even a stock market (which involves a lot of brains :). Up to now, all models are products of human brains, so it must be more-or-less inaccurate, which leads to Box's "All models are wrong". On the other hand, a model doesn't have to be technically correct to be useful. For example, Newton's laws of motion has been disproved by Einstein, but it remains useful when an object is not ridiculously big or fast.



To address your question, I honestly can't see the incompatibility between Box and Shmueli's points. It seems that you consider "explanatory power" and "predictive power" to be binomial properties, but I think they sit at the two ends of a spectrum.






share|cite|improve this answer












$endgroup$
















    Your Answer








    StackExchange.ready(function()
    var channelOptions =
    tags: "".split(" "),
    id: "65"
    ;
    initTagRenderer("".split(" "), "".split(" "), channelOptions);

    StackExchange.using("externalEditor", function()
    // Have to fire editor after snippets, if snippets enabled
    if (StackExchange.settings.snippets.snippetsEnabled)
    StackExchange.using("snippets", function()
    createEditor();
    );

    else
    createEditor();

    );

    function createEditor()
    StackExchange.prepareEditor(
    heartbeatType: 'answer',
    autoActivateHeartbeat: false,
    convertImagesToLinks: false,
    noModals: true,
    showLowRepImageUploadWarning: true,
    reputationToPostImages: null,
    bindNavPrevention: true,
    postfix: "",
    imageUploader:
    brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
    contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/4.0/"u003ecc by-sa 4.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
    allowUrls: true
    ,
    onDemand: true,
    discardSelector: ".discard-answer"
    ,immediatelyShowMarkdownHelp:true
    );



    );














    draft saved

    draft discarded
















    StackExchange.ready(
    function ()
    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f412980%2fon-george-box-galit-shmueli-and-the-scientific-method%23new-answer', 'question_page');

    );

    Post as a guest















    Required, but never shown


























    4 Answers
    4






    active

    oldest

    votes








    4 Answers
    4






    active

    oldest

    votes









    active

    oldest

    votes






    active

    oldest

    votes









    9
















    $begingroup$

    Let me start with the pithy quote by George Box, that "all models are wrong, but some are useful". This statement is an encapsulation of the methodological approach of "positivism", which is a philosophical approach that is highly influential in the sciences. This approach is described in detail (in the context of economic theory) in the classic methodological essay of Friedman (1966). In that essay, Friedman argues that any useful scientific theory necessarily constitutes a simplification of reality, and thus its assumptions must always depart from reality to some degree, and may even depart substantially from reality. He argues that the value of a scientific theory should not be judged by the closeness of its assumptions to reality --- instead it should be judged by its simplicity in reducing the complexity of the world to a manageable set of principles, and its accuracy in making predictions about reality, and generating new testable hypotheses about reality. Thus, Friedman argues that "all models are wrong" insofar as they all contain assumptions that simplify (and therefore depart from) reality, but that "some are useful" insofar as they give a simple framework to make useful predictions about reality.



    Now, if you read Box (1976) (the paper where he first states that "all models are wrong"), you will see that he does not cite Friedman, nor does he mention methodological positivism. Nevertheless, his explanation of the scientific method and its characteristics is extremely close to that developed by Friedman. In particular, both authors stress that a scientific theory will make predictions about reality that can be tested against observed facts, and the error in the prediction can then be used as a basis for revision of the theory.



    Now, on to the dichotomy discussed by Galit Shmueli in Shmueli (2001). In this paper, Shmueli compares causal explanation and prediction of observed outcomes and argues that these are distinct activities. Specifically, she argues that causal relations are based on underlying constructs that do not manifest directly in measureable outcomes, and so "measurable data are not accurate representations of their underlying constructs" (p. 293). She therefore argues that there is an aspect of statistical analysis that involves making inferences about unobservable underlying causal relations that are not manifested in measureable counterfactual differences in outcomes.



    Unless I am misunderstanding something, I think it is fair to say that this idea is in tension with the positivist views of Box and Friedman, as represented in the quote by Box. The positivist viewpoint essentially says that there are no admissible metaphysical "constructs" beyond those that manifest in measureable outcomes. Positivism confines itself to consideration of observable data, and concepts built on this data; it excludes consideration of a priori metaphysical concepts. Thus, a positivist would argue that the concept of causality can only be valid to the extent that it is defined in terms of measureable outcomes in reality --- to the extent it is defined as something distinct from this (as Shmueli treats it), this would be regarded as metaphysical speculation, and would be treated as inadmissible in scientific discourse.



    So I think you're right --- these two approaches are essentially in conflict. The positivist approach used by Box insists that valid scientific concepts be grounded entirely in their manifestations in reality, whereas the alternative approach used by Shmueli says that there are some "constructs" that are important scientific concepts (that we want to explain) but which cannot be perfectly represented when they are "operationalised" by relating them to measureable outcomes in reality.






    share|cite|improve this answer












    $endgroup$














    • $begingroup$
      Exactly!!! Shmueli seems to be contradicting most positivist (and falsificationist) definitions of what a scientific statement is, and I am wondering whether her intention is indeed to make such a bold philosophical statement? or whether she, as a statistician, isn't aware of how bold her statements really are?
      $endgroup$
      – Skander H.
      Jun 14 at 17:07










    • $begingroup$
      I posted a related question on the philosophy SE, care to weigh in on it?
      $endgroup$
      – Skander H.
      Jun 14 at 17:10






    • 1




      $begingroup$
      Although positivism and pragmatism share anti-realist traits, Box's approach is the latter, and positivism is not influential in sciences at least since late 1960-s. This is why Box does not mention Friedman, or other positivists. Pragmatism does not confine itself to observable data, and has no problem with a priori concepts or metaphysical constructions. It just does not see them as approaching "reality", and hence aiming at uniquely "right" features, they can be plural, task-dependent, and revisable. Causal relations are such constructs, so there is no conflict between Box and Shmueli.
      $endgroup$
      – Conifold
      Jun 14 at 21:37











    • $begingroup$
      Whether or not Box was a pragmatist, the particular quote he gives here is, in my view, more conducive to positivism than pragmatism. The latter philosophy takes a pluralistic view of metaphysics, which holds that there are multiple coherent ways of conceptualising reality, and that all these are "true" in the sense of being useful conceptualisations of reality. Thus, positivism would say, "all models are wrong, but some are useful", whereas the maxim of pragmatism would be closer to "many models are right, because they are useful".
      $endgroup$
      – Ben
      Jun 20 at 1:15










    • $begingroup$
      The identification of true with useful is just a folk misconception about pragmatism. All models are wrong because "right" suggests that they have something to correspond to, which pragmatists deny. And the principle of tolerance, "models are right as long as they serve a purpose", is due to Carnap, the father of logical positivism.
      $endgroup$
      – Conifold
      Jun 21 at 0:08















    9
















    $begingroup$

    Let me start with the pithy quote by George Box, that "all models are wrong, but some are useful". This statement is an encapsulation of the methodological approach of "positivism", which is a philosophical approach that is highly influential in the sciences. This approach is described in detail (in the context of economic theory) in the classic methodological essay of Friedman (1966). In that essay, Friedman argues that any useful scientific theory necessarily constitutes a simplification of reality, and thus its assumptions must always depart from reality to some degree, and may even depart substantially from reality. He argues that the value of a scientific theory should not be judged by the closeness of its assumptions to reality --- instead it should be judged by its simplicity in reducing the complexity of the world to a manageable set of principles, and its accuracy in making predictions about reality, and generating new testable hypotheses about reality. Thus, Friedman argues that "all models are wrong" insofar as they all contain assumptions that simplify (and therefore depart from) reality, but that "some are useful" insofar as they give a simple framework to make useful predictions about reality.



    Now, if you read Box (1976) (the paper where he first states that "all models are wrong"), you will see that he does not cite Friedman, nor does he mention methodological positivism. Nevertheless, his explanation of the scientific method and its characteristics is extremely close to that developed by Friedman. In particular, both authors stress that a scientific theory will make predictions about reality that can be tested against observed facts, and the error in the prediction can then be used as a basis for revision of the theory.



    Now, on to the dichotomy discussed by Galit Shmueli in Shmueli (2001). In this paper, Shmueli compares causal explanation and prediction of observed outcomes and argues that these are distinct activities. Specifically, she argues that causal relations are based on underlying constructs that do not manifest directly in measureable outcomes, and so "measurable data are not accurate representations of their underlying constructs" (p. 293). She therefore argues that there is an aspect of statistical analysis that involves making inferences about unobservable underlying causal relations that are not manifested in measureable counterfactual differences in outcomes.



    Unless I am misunderstanding something, I think it is fair to say that this idea is in tension with the positivist views of Box and Friedman, as represented in the quote by Box. The positivist viewpoint essentially says that there are no admissible metaphysical "constructs" beyond those that manifest in measureable outcomes. Positivism confines itself to consideration of observable data, and concepts built on this data; it excludes consideration of a priori metaphysical concepts. Thus, a positivist would argue that the concept of causality can only be valid to the extent that it is defined in terms of measureable outcomes in reality --- to the extent it is defined as something distinct from this (as Shmueli treats it), this would be regarded as metaphysical speculation, and would be treated as inadmissible in scientific discourse.



    So I think you're right --- these two approaches are essentially in conflict. The positivist approach used by Box insists that valid scientific concepts be grounded entirely in their manifestations in reality, whereas the alternative approach used by Shmueli says that there are some "constructs" that are important scientific concepts (that we want to explain) but which cannot be perfectly represented when they are "operationalised" by relating them to measureable outcomes in reality.






    share|cite|improve this answer












    $endgroup$














    • $begingroup$
      Exactly!!! Shmueli seems to be contradicting most positivist (and falsificationist) definitions of what a scientific statement is, and I am wondering whether her intention is indeed to make such a bold philosophical statement? or whether she, as a statistician, isn't aware of how bold her statements really are?
      $endgroup$
      – Skander H.
      Jun 14 at 17:07










    • $begingroup$
      I posted a related question on the philosophy SE, care to weigh in on it?
      $endgroup$
      – Skander H.
      Jun 14 at 17:10






    • 1




      $begingroup$
      Although positivism and pragmatism share anti-realist traits, Box's approach is the latter, and positivism is not influential in sciences at least since late 1960-s. This is why Box does not mention Friedman, or other positivists. Pragmatism does not confine itself to observable data, and has no problem with a priori concepts or metaphysical constructions. It just does not see them as approaching "reality", and hence aiming at uniquely "right" features, they can be plural, task-dependent, and revisable. Causal relations are such constructs, so there is no conflict between Box and Shmueli.
      $endgroup$
      – Conifold
      Jun 14 at 21:37











    • $begingroup$
      Whether or not Box was a pragmatist, the particular quote he gives here is, in my view, more conducive to positivism than pragmatism. The latter philosophy takes a pluralistic view of metaphysics, which holds that there are multiple coherent ways of conceptualising reality, and that all these are "true" in the sense of being useful conceptualisations of reality. Thus, positivism would say, "all models are wrong, but some are useful", whereas the maxim of pragmatism would be closer to "many models are right, because they are useful".
      $endgroup$
      – Ben
      Jun 20 at 1:15










    • $begingroup$
      The identification of true with useful is just a folk misconception about pragmatism. All models are wrong because "right" suggests that they have something to correspond to, which pragmatists deny. And the principle of tolerance, "models are right as long as they serve a purpose", is due to Carnap, the father of logical positivism.
      $endgroup$
      – Conifold
      Jun 21 at 0:08













    9














    9










    9







    $begingroup$

    Let me start with the pithy quote by George Box, that "all models are wrong, but some are useful". This statement is an encapsulation of the methodological approach of "positivism", which is a philosophical approach that is highly influential in the sciences. This approach is described in detail (in the context of economic theory) in the classic methodological essay of Friedman (1966). In that essay, Friedman argues that any useful scientific theory necessarily constitutes a simplification of reality, and thus its assumptions must always depart from reality to some degree, and may even depart substantially from reality. He argues that the value of a scientific theory should not be judged by the closeness of its assumptions to reality --- instead it should be judged by its simplicity in reducing the complexity of the world to a manageable set of principles, and its accuracy in making predictions about reality, and generating new testable hypotheses about reality. Thus, Friedman argues that "all models are wrong" insofar as they all contain assumptions that simplify (and therefore depart from) reality, but that "some are useful" insofar as they give a simple framework to make useful predictions about reality.



    Now, if you read Box (1976) (the paper where he first states that "all models are wrong"), you will see that he does not cite Friedman, nor does he mention methodological positivism. Nevertheless, his explanation of the scientific method and its characteristics is extremely close to that developed by Friedman. In particular, both authors stress that a scientific theory will make predictions about reality that can be tested against observed facts, and the error in the prediction can then be used as a basis for revision of the theory.



    Now, on to the dichotomy discussed by Galit Shmueli in Shmueli (2001). In this paper, Shmueli compares causal explanation and prediction of observed outcomes and argues that these are distinct activities. Specifically, she argues that causal relations are based on underlying constructs that do not manifest directly in measureable outcomes, and so "measurable data are not accurate representations of their underlying constructs" (p. 293). She therefore argues that there is an aspect of statistical analysis that involves making inferences about unobservable underlying causal relations that are not manifested in measureable counterfactual differences in outcomes.



    Unless I am misunderstanding something, I think it is fair to say that this idea is in tension with the positivist views of Box and Friedman, as represented in the quote by Box. The positivist viewpoint essentially says that there are no admissible metaphysical "constructs" beyond those that manifest in measureable outcomes. Positivism confines itself to consideration of observable data, and concepts built on this data; it excludes consideration of a priori metaphysical concepts. Thus, a positivist would argue that the concept of causality can only be valid to the extent that it is defined in terms of measureable outcomes in reality --- to the extent it is defined as something distinct from this (as Shmueli treats it), this would be regarded as metaphysical speculation, and would be treated as inadmissible in scientific discourse.



    So I think you're right --- these two approaches are essentially in conflict. The positivist approach used by Box insists that valid scientific concepts be grounded entirely in their manifestations in reality, whereas the alternative approach used by Shmueli says that there are some "constructs" that are important scientific concepts (that we want to explain) but which cannot be perfectly represented when they are "operationalised" by relating them to measureable outcomes in reality.






    share|cite|improve this answer












    $endgroup$



    Let me start with the pithy quote by George Box, that "all models are wrong, but some are useful". This statement is an encapsulation of the methodological approach of "positivism", which is a philosophical approach that is highly influential in the sciences. This approach is described in detail (in the context of economic theory) in the classic methodological essay of Friedman (1966). In that essay, Friedman argues that any useful scientific theory necessarily constitutes a simplification of reality, and thus its assumptions must always depart from reality to some degree, and may even depart substantially from reality. He argues that the value of a scientific theory should not be judged by the closeness of its assumptions to reality --- instead it should be judged by its simplicity in reducing the complexity of the world to a manageable set of principles, and its accuracy in making predictions about reality, and generating new testable hypotheses about reality. Thus, Friedman argues that "all models are wrong" insofar as they all contain assumptions that simplify (and therefore depart from) reality, but that "some are useful" insofar as they give a simple framework to make useful predictions about reality.



    Now, if you read Box (1976) (the paper where he first states that "all models are wrong"), you will see that he does not cite Friedman, nor does he mention methodological positivism. Nevertheless, his explanation of the scientific method and its characteristics is extremely close to that developed by Friedman. In particular, both authors stress that a scientific theory will make predictions about reality that can be tested against observed facts, and the error in the prediction can then be used as a basis for revision of the theory.



    Now, on to the dichotomy discussed by Galit Shmueli in Shmueli (2001). In this paper, Shmueli compares causal explanation and prediction of observed outcomes and argues that these are distinct activities. Specifically, she argues that causal relations are based on underlying constructs that do not manifest directly in measureable outcomes, and so "measurable data are not accurate representations of their underlying constructs" (p. 293). She therefore argues that there is an aspect of statistical analysis that involves making inferences about unobservable underlying causal relations that are not manifested in measureable counterfactual differences in outcomes.



    Unless I am misunderstanding something, I think it is fair to say that this idea is in tension with the positivist views of Box and Friedman, as represented in the quote by Box. The positivist viewpoint essentially says that there are no admissible metaphysical "constructs" beyond those that manifest in measureable outcomes. Positivism confines itself to consideration of observable data, and concepts built on this data; it excludes consideration of a priori metaphysical concepts. Thus, a positivist would argue that the concept of causality can only be valid to the extent that it is defined in terms of measureable outcomes in reality --- to the extent it is defined as something distinct from this (as Shmueli treats it), this would be regarded as metaphysical speculation, and would be treated as inadmissible in scientific discourse.



    So I think you're right --- these two approaches are essentially in conflict. The positivist approach used by Box insists that valid scientific concepts be grounded entirely in their manifestations in reality, whereas the alternative approach used by Shmueli says that there are some "constructs" that are important scientific concepts (that we want to explain) but which cannot be perfectly represented when they are "operationalised" by relating them to measureable outcomes in reality.







    share|cite|improve this answer















    share|cite|improve this answer




    share|cite|improve this answer








    edited Jun 14 at 11:02









    kjetil b halvorsen

    38k9 gold badges92 silver badges296 bronze badges




    38k9 gold badges92 silver badges296 bronze badges










    answered Jun 14 at 10:50









    BenBen

    39.4k2 gold badges51 silver badges170 bronze badges




    39.4k2 gold badges51 silver badges170 bronze badges














    • $begingroup$
      Exactly!!! Shmueli seems to be contradicting most positivist (and falsificationist) definitions of what a scientific statement is, and I am wondering whether her intention is indeed to make such a bold philosophical statement? or whether she, as a statistician, isn't aware of how bold her statements really are?
      $endgroup$
      – Skander H.
      Jun 14 at 17:07










    • $begingroup$
      I posted a related question on the philosophy SE, care to weigh in on it?
      $endgroup$
      – Skander H.
      Jun 14 at 17:10






    • 1




      $begingroup$
      Although positivism and pragmatism share anti-realist traits, Box's approach is the latter, and positivism is not influential in sciences at least since late 1960-s. This is why Box does not mention Friedman, or other positivists. Pragmatism does not confine itself to observable data, and has no problem with a priori concepts or metaphysical constructions. It just does not see them as approaching "reality", and hence aiming at uniquely "right" features, they can be plural, task-dependent, and revisable. Causal relations are such constructs, so there is no conflict between Box and Shmueli.
      $endgroup$
      – Conifold
      Jun 14 at 21:37











    • $begingroup$
      Whether or not Box was a pragmatist, the particular quote he gives here is, in my view, more conducive to positivism than pragmatism. The latter philosophy takes a pluralistic view of metaphysics, which holds that there are multiple coherent ways of conceptualising reality, and that all these are "true" in the sense of being useful conceptualisations of reality. Thus, positivism would say, "all models are wrong, but some are useful", whereas the maxim of pragmatism would be closer to "many models are right, because they are useful".
      $endgroup$
      – Ben
      Jun 20 at 1:15










    • $begingroup$
      The identification of true with useful is just a folk misconception about pragmatism. All models are wrong because "right" suggests that they have something to correspond to, which pragmatists deny. And the principle of tolerance, "models are right as long as they serve a purpose", is due to Carnap, the father of logical positivism.
      $endgroup$
      – Conifold
      Jun 21 at 0:08
















    • $begingroup$
      Exactly!!! Shmueli seems to be contradicting most positivist (and falsificationist) definitions of what a scientific statement is, and I am wondering whether her intention is indeed to make such a bold philosophical statement? or whether she, as a statistician, isn't aware of how bold her statements really are?
      $endgroup$
      – Skander H.
      Jun 14 at 17:07










    • $begingroup$
      I posted a related question on the philosophy SE, care to weigh in on it?
      $endgroup$
      – Skander H.
      Jun 14 at 17:10






    • 1




      $begingroup$
      Although positivism and pragmatism share anti-realist traits, Box's approach is the latter, and positivism is not influential in sciences at least since late 1960-s. This is why Box does not mention Friedman, or other positivists. Pragmatism does not confine itself to observable data, and has no problem with a priori concepts or metaphysical constructions. It just does not see them as approaching "reality", and hence aiming at uniquely "right" features, they can be plural, task-dependent, and revisable. Causal relations are such constructs, so there is no conflict between Box and Shmueli.
      $endgroup$
      – Conifold
      Jun 14 at 21:37











    • $begingroup$
      Whether or not Box was a pragmatist, the particular quote he gives here is, in my view, more conducive to positivism than pragmatism. The latter philosophy takes a pluralistic view of metaphysics, which holds that there are multiple coherent ways of conceptualising reality, and that all these are "true" in the sense of being useful conceptualisations of reality. Thus, positivism would say, "all models are wrong, but some are useful", whereas the maxim of pragmatism would be closer to "many models are right, because they are useful".
      $endgroup$
      – Ben
      Jun 20 at 1:15










    • $begingroup$
      The identification of true with useful is just a folk misconception about pragmatism. All models are wrong because "right" suggests that they have something to correspond to, which pragmatists deny. And the principle of tolerance, "models are right as long as they serve a purpose", is due to Carnap, the father of logical positivism.
      $endgroup$
      – Conifold
      Jun 21 at 0:08















    $begingroup$
    Exactly!!! Shmueli seems to be contradicting most positivist (and falsificationist) definitions of what a scientific statement is, and I am wondering whether her intention is indeed to make such a bold philosophical statement? or whether she, as a statistician, isn't aware of how bold her statements really are?
    $endgroup$
    – Skander H.
    Jun 14 at 17:07




    $begingroup$
    Exactly!!! Shmueli seems to be contradicting most positivist (and falsificationist) definitions of what a scientific statement is, and I am wondering whether her intention is indeed to make such a bold philosophical statement? or whether she, as a statistician, isn't aware of how bold her statements really are?
    $endgroup$
    – Skander H.
    Jun 14 at 17:07












    $begingroup$
    I posted a related question on the philosophy SE, care to weigh in on it?
    $endgroup$
    – Skander H.
    Jun 14 at 17:10




    $begingroup$
    I posted a related question on the philosophy SE, care to weigh in on it?
    $endgroup$
    – Skander H.
    Jun 14 at 17:10




    1




    1




    $begingroup$
    Although positivism and pragmatism share anti-realist traits, Box's approach is the latter, and positivism is not influential in sciences at least since late 1960-s. This is why Box does not mention Friedman, or other positivists. Pragmatism does not confine itself to observable data, and has no problem with a priori concepts or metaphysical constructions. It just does not see them as approaching "reality", and hence aiming at uniquely "right" features, they can be plural, task-dependent, and revisable. Causal relations are such constructs, so there is no conflict between Box and Shmueli.
    $endgroup$
    – Conifold
    Jun 14 at 21:37





    $begingroup$
    Although positivism and pragmatism share anti-realist traits, Box's approach is the latter, and positivism is not influential in sciences at least since late 1960-s. This is why Box does not mention Friedman, or other positivists. Pragmatism does not confine itself to observable data, and has no problem with a priori concepts or metaphysical constructions. It just does not see them as approaching "reality", and hence aiming at uniquely "right" features, they can be plural, task-dependent, and revisable. Causal relations are such constructs, so there is no conflict between Box and Shmueli.
    $endgroup$
    – Conifold
    Jun 14 at 21:37













    $begingroup$
    Whether or not Box was a pragmatist, the particular quote he gives here is, in my view, more conducive to positivism than pragmatism. The latter philosophy takes a pluralistic view of metaphysics, which holds that there are multiple coherent ways of conceptualising reality, and that all these are "true" in the sense of being useful conceptualisations of reality. Thus, positivism would say, "all models are wrong, but some are useful", whereas the maxim of pragmatism would be closer to "many models are right, because they are useful".
    $endgroup$
    – Ben
    Jun 20 at 1:15




    $begingroup$
    Whether or not Box was a pragmatist, the particular quote he gives here is, in my view, more conducive to positivism than pragmatism. The latter philosophy takes a pluralistic view of metaphysics, which holds that there are multiple coherent ways of conceptualising reality, and that all these are "true" in the sense of being useful conceptualisations of reality. Thus, positivism would say, "all models are wrong, but some are useful", whereas the maxim of pragmatism would be closer to "many models are right, because they are useful".
    $endgroup$
    – Ben
    Jun 20 at 1:15












    $begingroup$
    The identification of true with useful is just a folk misconception about pragmatism. All models are wrong because "right" suggests that they have something to correspond to, which pragmatists deny. And the principle of tolerance, "models are right as long as they serve a purpose", is due to Carnap, the father of logical positivism.
    $endgroup$
    – Conifold
    Jun 21 at 0:08




    $begingroup$
    The identification of true with useful is just a folk misconception about pragmatism. All models are wrong because "right" suggests that they have something to correspond to, which pragmatists deny. And the principle of tolerance, "models are right as long as they serve a purpose", is due to Carnap, the father of logical positivism.
    $endgroup$
    – Conifold
    Jun 21 at 0:08













    4
















    $begingroup$

    A model, when used to explain things, is a simplification of reality. Simplification is just another word for "wrong in some useful way". For example, if we round the number 3.1415926535898 to 3.14 we are making an error, but this error allows us humans to focus on the most important part of that number. This is how models are used in explaining, it provides insights on some problem, but by necessity has to abstract away from a lot of other things: We humans are just not very good at looking at a thousands things simultaneously. If we primarily care about predicting we want to include those thousands things whenever possible, but with explaining the trade-off is different.






    share|cite|improve this answer










    $endgroup$










    • 1




      $begingroup$
      "but this error allows us humans to focus on the most important part of that number." this makes sense and helps me understand what "explanation" means, but then it also confirms my point that explanation is more of an artistic/aesthetic concept than a scientific one. Based on you $pi$ example, consider the following as well: A novel dimensionality reduction technique allows for plotting very elegant and intuitive graphs of high dimensional data, which makes for good explanation, but then there is no way to objectively asses the accuracy of this technique, its value is purely subjective.
      $endgroup$
      – Skander H.
      Jun 14 at 8:41






    • 1




      $begingroup$
      @SkanderH. To the extent that "elegant and intuitive graphs" facilitate engineering applications, or development of new theories, their value is not purely subjective, or non-scientific, it is pragmatic. Unification, explanatory power, simplicity and coherence are broadly acknowledged as epistemic, not aesthetic, values. The choice between Lorentz's theory of ether and special relativity was made based on just such considerations, they are predictively equivalent.
      $endgroup$
      – Conifold
      Jun 14 at 9:33















    4
















    $begingroup$

    A model, when used to explain things, is a simplification of reality. Simplification is just another word for "wrong in some useful way". For example, if we round the number 3.1415926535898 to 3.14 we are making an error, but this error allows us humans to focus on the most important part of that number. This is how models are used in explaining, it provides insights on some problem, but by necessity has to abstract away from a lot of other things: We humans are just not very good at looking at a thousands things simultaneously. If we primarily care about predicting we want to include those thousands things whenever possible, but with explaining the trade-off is different.






    share|cite|improve this answer










    $endgroup$










    • 1




      $begingroup$
      "but this error allows us humans to focus on the most important part of that number." this makes sense and helps me understand what "explanation" means, but then it also confirms my point that explanation is more of an artistic/aesthetic concept than a scientific one. Based on you $pi$ example, consider the following as well: A novel dimensionality reduction technique allows for plotting very elegant and intuitive graphs of high dimensional data, which makes for good explanation, but then there is no way to objectively asses the accuracy of this technique, its value is purely subjective.
      $endgroup$
      – Skander H.
      Jun 14 at 8:41






    • 1




      $begingroup$
      @SkanderH. To the extent that "elegant and intuitive graphs" facilitate engineering applications, or development of new theories, their value is not purely subjective, or non-scientific, it is pragmatic. Unification, explanatory power, simplicity and coherence are broadly acknowledged as epistemic, not aesthetic, values. The choice between Lorentz's theory of ether and special relativity was made based on just such considerations, they are predictively equivalent.
      $endgroup$
      – Conifold
      Jun 14 at 9:33













    4














    4










    4







    $begingroup$

    A model, when used to explain things, is a simplification of reality. Simplification is just another word for "wrong in some useful way". For example, if we round the number 3.1415926535898 to 3.14 we are making an error, but this error allows us humans to focus on the most important part of that number. This is how models are used in explaining, it provides insights on some problem, but by necessity has to abstract away from a lot of other things: We humans are just not very good at looking at a thousands things simultaneously. If we primarily care about predicting we want to include those thousands things whenever possible, but with explaining the trade-off is different.






    share|cite|improve this answer










    $endgroup$



    A model, when used to explain things, is a simplification of reality. Simplification is just another word for "wrong in some useful way". For example, if we round the number 3.1415926535898 to 3.14 we are making an error, but this error allows us humans to focus on the most important part of that number. This is how models are used in explaining, it provides insights on some problem, but by necessity has to abstract away from a lot of other things: We humans are just not very good at looking at a thousands things simultaneously. If we primarily care about predicting we want to include those thousands things whenever possible, but with explaining the trade-off is different.







    share|cite|improve this answer













    share|cite|improve this answer




    share|cite|improve this answer










    answered Jun 14 at 8:34









    Maarten BuisMaarten Buis

    16.8k21 silver badges46 bronze badges




    16.8k21 silver badges46 bronze badges










    • 1




      $begingroup$
      "but this error allows us humans to focus on the most important part of that number." this makes sense and helps me understand what "explanation" means, but then it also confirms my point that explanation is more of an artistic/aesthetic concept than a scientific one. Based on you $pi$ example, consider the following as well: A novel dimensionality reduction technique allows for plotting very elegant and intuitive graphs of high dimensional data, which makes for good explanation, but then there is no way to objectively asses the accuracy of this technique, its value is purely subjective.
      $endgroup$
      – Skander H.
      Jun 14 at 8:41






    • 1




      $begingroup$
      @SkanderH. To the extent that "elegant and intuitive graphs" facilitate engineering applications, or development of new theories, their value is not purely subjective, or non-scientific, it is pragmatic. Unification, explanatory power, simplicity and coherence are broadly acknowledged as epistemic, not aesthetic, values. The choice between Lorentz's theory of ether and special relativity was made based on just such considerations, they are predictively equivalent.
      $endgroup$
      – Conifold
      Jun 14 at 9:33












    • 1




      $begingroup$
      "but this error allows us humans to focus on the most important part of that number." this makes sense and helps me understand what "explanation" means, but then it also confirms my point that explanation is more of an artistic/aesthetic concept than a scientific one. Based on you $pi$ example, consider the following as well: A novel dimensionality reduction technique allows for plotting very elegant and intuitive graphs of high dimensional data, which makes for good explanation, but then there is no way to objectively asses the accuracy of this technique, its value is purely subjective.
      $endgroup$
      – Skander H.
      Jun 14 at 8:41






    • 1




      $begingroup$
      @SkanderH. To the extent that "elegant and intuitive graphs" facilitate engineering applications, or development of new theories, their value is not purely subjective, or non-scientific, it is pragmatic. Unification, explanatory power, simplicity and coherence are broadly acknowledged as epistemic, not aesthetic, values. The choice between Lorentz's theory of ether and special relativity was made based on just such considerations, they are predictively equivalent.
      $endgroup$
      – Conifold
      Jun 14 at 9:33







    1




    1




    $begingroup$
    "but this error allows us humans to focus on the most important part of that number." this makes sense and helps me understand what "explanation" means, but then it also confirms my point that explanation is more of an artistic/aesthetic concept than a scientific one. Based on you $pi$ example, consider the following as well: A novel dimensionality reduction technique allows for plotting very elegant and intuitive graphs of high dimensional data, which makes for good explanation, but then there is no way to objectively asses the accuracy of this technique, its value is purely subjective.
    $endgroup$
    – Skander H.
    Jun 14 at 8:41




    $begingroup$
    "but this error allows us humans to focus on the most important part of that number." this makes sense and helps me understand what "explanation" means, but then it also confirms my point that explanation is more of an artistic/aesthetic concept than a scientific one. Based on you $pi$ example, consider the following as well: A novel dimensionality reduction technique allows for plotting very elegant and intuitive graphs of high dimensional data, which makes for good explanation, but then there is no way to objectively asses the accuracy of this technique, its value is purely subjective.
    $endgroup$
    – Skander H.
    Jun 14 at 8:41




    1




    1




    $begingroup$
    @SkanderH. To the extent that "elegant and intuitive graphs" facilitate engineering applications, or development of new theories, their value is not purely subjective, or non-scientific, it is pragmatic. Unification, explanatory power, simplicity and coherence are broadly acknowledged as epistemic, not aesthetic, values. The choice between Lorentz's theory of ether and special relativity was made based on just such considerations, they are predictively equivalent.
    $endgroup$
    – Conifold
    Jun 14 at 9:33




    $begingroup$
    @SkanderH. To the extent that "elegant and intuitive graphs" facilitate engineering applications, or development of new theories, their value is not purely subjective, or non-scientific, it is pragmatic. Unification, explanatory power, simplicity and coherence are broadly acknowledged as epistemic, not aesthetic, values. The choice between Lorentz's theory of ether and special relativity was made based on just such considerations, they are predictively equivalent.
    $endgroup$
    – Conifold
    Jun 14 at 9:33











    3
















    $begingroup$

    An example of a model that is excellent at prediction but does not explain anything is given in the Wikipedia article “All models are wrong”. The example is Newton’s model of gravitation. Newton’s model almost always gives predictions that are indistinguishable from empirical observations. Yet the model is extremely implausible: because it postulates a force that can act instantaneously over arbitrarily large distances.



    Newton’s model has been supplanted by the model given in Einstein’s general theory of relativity. With general relativity, gravitational forces travel through space at finite speed (the speed of light).



    Newton’s model is not a simplification of the general-relativistic model. To illustrate that, consider an apple falling down from a tree. According to general relativity, the apple falls without Earth exerting any force on the apple. (The primary reason the apple falls is that Earth warps time, so that clocks near the base of the tree run more slowly than clocks high up in the tree.) Thus, as the Wikipedia article notes, Newton’s model is completely wrong from an explanatory perspective.



    The paper by Shmueli [2010] presumes that there are two purposes for a model: prediction and explanation. In fact, several authors have stated that there are three purposes (see e.g. Konishi & Kitagawa [Information Criteria and Statistical Modeling, 2008: §1.1] and Friendly & Meyer [Discrete Data Analysis, 2016: §11.6]). The three purposes correspond to the three kinds of logical reasoning:



    • prediction (corresponding to deduction);

    • parameter estimation (corresponding to induction);

    • description of structure (corresponding to abduction).





    share|cite|improve this answer












    $endgroup$














    • $begingroup$
      To say that Newton's model of gravitation "does not explain anything" is, frankly, ludicrous. -1.
      $endgroup$
      – amoeba
      Jun 16 at 20:45










    • $begingroup$
      amoeba, Newton's model does not explain anything about how gravitation works, under the assumption that general relativity is accurate. If an apple falls, Newton's model postulates that Earth exerts a force on the apple and that postulate is completely false. I ask you to consider my answer further. If you still do not understand, then kindly tell me what is unclear.
      $endgroup$
      – SolidPhase
      Jun 16 at 20:53











    • $begingroup$
      What you say is quite clear but I strongly disagree with it.
      $endgroup$
      – amoeba
      Jun 16 at 22:07










    • $begingroup$
      amoeba, I ask you to explain why you disagree: do you have a reason? (Note that I have added an extra sentence to the answer.)
      $endgroup$
      – SolidPhase
      Jun 16 at 22:16











    • $begingroup$
      Thanks. I will look up the references you mention. I understand how a model can predict even if it doesn't explain. What I don't get is the opposite direction: How can a model explain without predicting. You Newton vs. Einstein examples just muddies everything even more: The whole reason Einstein's theory supplanted Newton's was because it predicted better. Look at it another way: If we have competing explanatory models, how can we evaluated them unless we test which one has the most predictive power?
      $endgroup$
      – Skander H.
      Jun 17 at 5:35















    3
















    $begingroup$

    An example of a model that is excellent at prediction but does not explain anything is given in the Wikipedia article “All models are wrong”. The example is Newton’s model of gravitation. Newton’s model almost always gives predictions that are indistinguishable from empirical observations. Yet the model is extremely implausible: because it postulates a force that can act instantaneously over arbitrarily large distances.



    Newton’s model has been supplanted by the model given in Einstein’s general theory of relativity. With general relativity, gravitational forces travel through space at finite speed (the speed of light).



    Newton’s model is not a simplification of the general-relativistic model. To illustrate that, consider an apple falling down from a tree. According to general relativity, the apple falls without Earth exerting any force on the apple. (The primary reason the apple falls is that Earth warps time, so that clocks near the base of the tree run more slowly than clocks high up in the tree.) Thus, as the Wikipedia article notes, Newton’s model is completely wrong from an explanatory perspective.



    The paper by Shmueli [2010] presumes that there are two purposes for a model: prediction and explanation. In fact, several authors have stated that there are three purposes (see e.g. Konishi & Kitagawa [Information Criteria and Statistical Modeling, 2008: §1.1] and Friendly & Meyer [Discrete Data Analysis, 2016: §11.6]). The three purposes correspond to the three kinds of logical reasoning:



    • prediction (corresponding to deduction);

    • parameter estimation (corresponding to induction);

    • description of structure (corresponding to abduction).





    share|cite|improve this answer












    $endgroup$














    • $begingroup$
      To say that Newton's model of gravitation "does not explain anything" is, frankly, ludicrous. -1.
      $endgroup$
      – amoeba
      Jun 16 at 20:45










    • $begingroup$
      amoeba, Newton's model does not explain anything about how gravitation works, under the assumption that general relativity is accurate. If an apple falls, Newton's model postulates that Earth exerts a force on the apple and that postulate is completely false. I ask you to consider my answer further. If you still do not understand, then kindly tell me what is unclear.
      $endgroup$
      – SolidPhase
      Jun 16 at 20:53











    • $begingroup$
      What you say is quite clear but I strongly disagree with it.
      $endgroup$
      – amoeba
      Jun 16 at 22:07










    • $begingroup$
      amoeba, I ask you to explain why you disagree: do you have a reason? (Note that I have added an extra sentence to the answer.)
      $endgroup$
      – SolidPhase
      Jun 16 at 22:16











    • $begingroup$
      Thanks. I will look up the references you mention. I understand how a model can predict even if it doesn't explain. What I don't get is the opposite direction: How can a model explain without predicting. You Newton vs. Einstein examples just muddies everything even more: The whole reason Einstein's theory supplanted Newton's was because it predicted better. Look at it another way: If we have competing explanatory models, how can we evaluated them unless we test which one has the most predictive power?
      $endgroup$
      – Skander H.
      Jun 17 at 5:35













    3














    3










    3







    $begingroup$

    An example of a model that is excellent at prediction but does not explain anything is given in the Wikipedia article “All models are wrong”. The example is Newton’s model of gravitation. Newton’s model almost always gives predictions that are indistinguishable from empirical observations. Yet the model is extremely implausible: because it postulates a force that can act instantaneously over arbitrarily large distances.



    Newton’s model has been supplanted by the model given in Einstein’s general theory of relativity. With general relativity, gravitational forces travel through space at finite speed (the speed of light).



    Newton’s model is not a simplification of the general-relativistic model. To illustrate that, consider an apple falling down from a tree. According to general relativity, the apple falls without Earth exerting any force on the apple. (The primary reason the apple falls is that Earth warps time, so that clocks near the base of the tree run more slowly than clocks high up in the tree.) Thus, as the Wikipedia article notes, Newton’s model is completely wrong from an explanatory perspective.



    The paper by Shmueli [2010] presumes that there are two purposes for a model: prediction and explanation. In fact, several authors have stated that there are three purposes (see e.g. Konishi & Kitagawa [Information Criteria and Statistical Modeling, 2008: §1.1] and Friendly & Meyer [Discrete Data Analysis, 2016: §11.6]). The three purposes correspond to the three kinds of logical reasoning:



    • prediction (corresponding to deduction);

    • parameter estimation (corresponding to induction);

    • description of structure (corresponding to abduction).





    share|cite|improve this answer












    $endgroup$



    An example of a model that is excellent at prediction but does not explain anything is given in the Wikipedia article “All models are wrong”. The example is Newton’s model of gravitation. Newton’s model almost always gives predictions that are indistinguishable from empirical observations. Yet the model is extremely implausible: because it postulates a force that can act instantaneously over arbitrarily large distances.



    Newton’s model has been supplanted by the model given in Einstein’s general theory of relativity. With general relativity, gravitational forces travel through space at finite speed (the speed of light).



    Newton’s model is not a simplification of the general-relativistic model. To illustrate that, consider an apple falling down from a tree. According to general relativity, the apple falls without Earth exerting any force on the apple. (The primary reason the apple falls is that Earth warps time, so that clocks near the base of the tree run more slowly than clocks high up in the tree.) Thus, as the Wikipedia article notes, Newton’s model is completely wrong from an explanatory perspective.



    The paper by Shmueli [2010] presumes that there are two purposes for a model: prediction and explanation. In fact, several authors have stated that there are three purposes (see e.g. Konishi & Kitagawa [Information Criteria and Statistical Modeling, 2008: §1.1] and Friendly & Meyer [Discrete Data Analysis, 2016: §11.6]). The three purposes correspond to the three kinds of logical reasoning:



    • prediction (corresponding to deduction);

    • parameter estimation (corresponding to induction);

    • description of structure (corresponding to abduction).






    share|cite|improve this answer















    share|cite|improve this answer




    share|cite|improve this answer








    edited Jun 16 at 22:14

























    answered Jun 16 at 18:49









    SolidPhaseSolidPhase

    394 bronze badges




    394 bronze badges














    • $begingroup$
      To say that Newton's model of gravitation "does not explain anything" is, frankly, ludicrous. -1.
      $endgroup$
      – amoeba
      Jun 16 at 20:45










    • $begingroup$
      amoeba, Newton's model does not explain anything about how gravitation works, under the assumption that general relativity is accurate. If an apple falls, Newton's model postulates that Earth exerts a force on the apple and that postulate is completely false. I ask you to consider my answer further. If you still do not understand, then kindly tell me what is unclear.
      $endgroup$
      – SolidPhase
      Jun 16 at 20:53











    • $begingroup$
      What you say is quite clear but I strongly disagree with it.
      $endgroup$
      – amoeba
      Jun 16 at 22:07










    • $begingroup$
      amoeba, I ask you to explain why you disagree: do you have a reason? (Note that I have added an extra sentence to the answer.)
      $endgroup$
      – SolidPhase
      Jun 16 at 22:16











    • $begingroup$
      Thanks. I will look up the references you mention. I understand how a model can predict even if it doesn't explain. What I don't get is the opposite direction: How can a model explain without predicting. You Newton vs. Einstein examples just muddies everything even more: The whole reason Einstein's theory supplanted Newton's was because it predicted better. Look at it another way: If we have competing explanatory models, how can we evaluated them unless we test which one has the most predictive power?
      $endgroup$
      – Skander H.
      Jun 17 at 5:35
















    • $begingroup$
      To say that Newton's model of gravitation "does not explain anything" is, frankly, ludicrous. -1.
      $endgroup$
      – amoeba
      Jun 16 at 20:45










    • $begingroup$
      amoeba, Newton's model does not explain anything about how gravitation works, under the assumption that general relativity is accurate. If an apple falls, Newton's model postulates that Earth exerts a force on the apple and that postulate is completely false. I ask you to consider my answer further. If you still do not understand, then kindly tell me what is unclear.
      $endgroup$
      – SolidPhase
      Jun 16 at 20:53











    • $begingroup$
      What you say is quite clear but I strongly disagree with it.
      $endgroup$
      – amoeba
      Jun 16 at 22:07










    • $begingroup$
      amoeba, I ask you to explain why you disagree: do you have a reason? (Note that I have added an extra sentence to the answer.)
      $endgroup$
      – SolidPhase
      Jun 16 at 22:16











    • $begingroup$
      Thanks. I will look up the references you mention. I understand how a model can predict even if it doesn't explain. What I don't get is the opposite direction: How can a model explain without predicting. You Newton vs. Einstein examples just muddies everything even more: The whole reason Einstein's theory supplanted Newton's was because it predicted better. Look at it another way: If we have competing explanatory models, how can we evaluated them unless we test which one has the most predictive power?
      $endgroup$
      – Skander H.
      Jun 17 at 5:35















    $begingroup$
    To say that Newton's model of gravitation "does not explain anything" is, frankly, ludicrous. -1.
    $endgroup$
    – amoeba
    Jun 16 at 20:45




    $begingroup$
    To say that Newton's model of gravitation "does not explain anything" is, frankly, ludicrous. -1.
    $endgroup$
    – amoeba
    Jun 16 at 20:45












    $begingroup$
    amoeba, Newton's model does not explain anything about how gravitation works, under the assumption that general relativity is accurate. If an apple falls, Newton's model postulates that Earth exerts a force on the apple and that postulate is completely false. I ask you to consider my answer further. If you still do not understand, then kindly tell me what is unclear.
    $endgroup$
    – SolidPhase
    Jun 16 at 20:53





    $begingroup$
    amoeba, Newton's model does not explain anything about how gravitation works, under the assumption that general relativity is accurate. If an apple falls, Newton's model postulates that Earth exerts a force on the apple and that postulate is completely false. I ask you to consider my answer further. If you still do not understand, then kindly tell me what is unclear.
    $endgroup$
    – SolidPhase
    Jun 16 at 20:53













    $begingroup$
    What you say is quite clear but I strongly disagree with it.
    $endgroup$
    – amoeba
    Jun 16 at 22:07




    $begingroup$
    What you say is quite clear but I strongly disagree with it.
    $endgroup$
    – amoeba
    Jun 16 at 22:07












    $begingroup$
    amoeba, I ask you to explain why you disagree: do you have a reason? (Note that I have added an extra sentence to the answer.)
    $endgroup$
    – SolidPhase
    Jun 16 at 22:16





    $begingroup$
    amoeba, I ask you to explain why you disagree: do you have a reason? (Note that I have added an extra sentence to the answer.)
    $endgroup$
    – SolidPhase
    Jun 16 at 22:16













    $begingroup$
    Thanks. I will look up the references you mention. I understand how a model can predict even if it doesn't explain. What I don't get is the opposite direction: How can a model explain without predicting. You Newton vs. Einstein examples just muddies everything even more: The whole reason Einstein's theory supplanted Newton's was because it predicted better. Look at it another way: If we have competing explanatory models, how can we evaluated them unless we test which one has the most predictive power?
    $endgroup$
    – Skander H.
    Jun 17 at 5:35




    $begingroup$
    Thanks. I will look up the references you mention. I understand how a model can predict even if it doesn't explain. What I don't get is the opposite direction: How can a model explain without predicting. You Newton vs. Einstein examples just muddies everything even more: The whole reason Einstein's theory supplanted Newton's was because it predicted better. Look at it another way: If we have competing explanatory models, how can we evaluated them unless we test which one has the most predictive power?
    $endgroup$
    – Skander H.
    Jun 17 at 5:35











    1
















    $begingroup$

    I'm an undergraduate in Statistics, so I won't call myself an expert, but here are my two cents.



    Models don't explain themselves; humans interpret them. Linear models are easier to understand than neural networks and random forests because they are closer to how we make decisions. Indeed, ANNs imitate the human brain, but you don't decide which restaurant to go tomorrow by doing a series of matrix multiplications. Instead, you weight some factors in your mind by their importance, which is essentially a linear combination.



    "Explanatory power" measures how well a model gets along with humans' intuition, whereas "predictive power" measures how well it aligns with the underlying mechanism of the process in interest. The contradiction between them is essentially the gap between what the world is and how we can perceive/understand it. I hope this explains why "some models do a good job of explaining, even though they do a poor job at predicting".



    Ian Stewart once said, "If our brains were simple enough for us to understand them, we'd be so simple that we couldn't." Unfortunately, our little human brains are actually very simple compared to the universe, or even a stock market (which involves a lot of brains :). Up to now, all models are products of human brains, so it must be more-or-less inaccurate, which leads to Box's "All models are wrong". On the other hand, a model doesn't have to be technically correct to be useful. For example, Newton's laws of motion has been disproved by Einstein, but it remains useful when an object is not ridiculously big or fast.



    To address your question, I honestly can't see the incompatibility between Box and Shmueli's points. It seems that you consider "explanatory power" and "predictive power" to be binomial properties, but I think they sit at the two ends of a spectrum.






    share|cite|improve this answer












    $endgroup$



















      1
















      $begingroup$

      I'm an undergraduate in Statistics, so I won't call myself an expert, but here are my two cents.



      Models don't explain themselves; humans interpret them. Linear models are easier to understand than neural networks and random forests because they are closer to how we make decisions. Indeed, ANNs imitate the human brain, but you don't decide which restaurant to go tomorrow by doing a series of matrix multiplications. Instead, you weight some factors in your mind by their importance, which is essentially a linear combination.



      "Explanatory power" measures how well a model gets along with humans' intuition, whereas "predictive power" measures how well it aligns with the underlying mechanism of the process in interest. The contradiction between them is essentially the gap between what the world is and how we can perceive/understand it. I hope this explains why "some models do a good job of explaining, even though they do a poor job at predicting".



      Ian Stewart once said, "If our brains were simple enough for us to understand them, we'd be so simple that we couldn't." Unfortunately, our little human brains are actually very simple compared to the universe, or even a stock market (which involves a lot of brains :). Up to now, all models are products of human brains, so it must be more-or-less inaccurate, which leads to Box's "All models are wrong". On the other hand, a model doesn't have to be technically correct to be useful. For example, Newton's laws of motion has been disproved by Einstein, but it remains useful when an object is not ridiculously big or fast.



      To address your question, I honestly can't see the incompatibility between Box and Shmueli's points. It seems that you consider "explanatory power" and "predictive power" to be binomial properties, but I think they sit at the two ends of a spectrum.






      share|cite|improve this answer












      $endgroup$

















        1














        1










        1







        $begingroup$

        I'm an undergraduate in Statistics, so I won't call myself an expert, but here are my two cents.



        Models don't explain themselves; humans interpret them. Linear models are easier to understand than neural networks and random forests because they are closer to how we make decisions. Indeed, ANNs imitate the human brain, but you don't decide which restaurant to go tomorrow by doing a series of matrix multiplications. Instead, you weight some factors in your mind by their importance, which is essentially a linear combination.



        "Explanatory power" measures how well a model gets along with humans' intuition, whereas "predictive power" measures how well it aligns with the underlying mechanism of the process in interest. The contradiction between them is essentially the gap between what the world is and how we can perceive/understand it. I hope this explains why "some models do a good job of explaining, even though they do a poor job at predicting".



        Ian Stewart once said, "If our brains were simple enough for us to understand them, we'd be so simple that we couldn't." Unfortunately, our little human brains are actually very simple compared to the universe, or even a stock market (which involves a lot of brains :). Up to now, all models are products of human brains, so it must be more-or-less inaccurate, which leads to Box's "All models are wrong". On the other hand, a model doesn't have to be technically correct to be useful. For example, Newton's laws of motion has been disproved by Einstein, but it remains useful when an object is not ridiculously big or fast.



        To address your question, I honestly can't see the incompatibility between Box and Shmueli's points. It seems that you consider "explanatory power" and "predictive power" to be binomial properties, but I think they sit at the two ends of a spectrum.






        share|cite|improve this answer












        $endgroup$



        I'm an undergraduate in Statistics, so I won't call myself an expert, but here are my two cents.



        Models don't explain themselves; humans interpret them. Linear models are easier to understand than neural networks and random forests because they are closer to how we make decisions. Indeed, ANNs imitate the human brain, but you don't decide which restaurant to go tomorrow by doing a series of matrix multiplications. Instead, you weight some factors in your mind by their importance, which is essentially a linear combination.



        "Explanatory power" measures how well a model gets along with humans' intuition, whereas "predictive power" measures how well it aligns with the underlying mechanism of the process in interest. The contradiction between them is essentially the gap between what the world is and how we can perceive/understand it. I hope this explains why "some models do a good job of explaining, even though they do a poor job at predicting".



        Ian Stewart once said, "If our brains were simple enough for us to understand them, we'd be so simple that we couldn't." Unfortunately, our little human brains are actually very simple compared to the universe, or even a stock market (which involves a lot of brains :). Up to now, all models are products of human brains, so it must be more-or-less inaccurate, which leads to Box's "All models are wrong". On the other hand, a model doesn't have to be technically correct to be useful. For example, Newton's laws of motion has been disproved by Einstein, but it remains useful when an object is not ridiculously big or fast.



        To address your question, I honestly can't see the incompatibility between Box and Shmueli's points. It seems that you consider "explanatory power" and "predictive power" to be binomial properties, but I think they sit at the two ends of a spectrum.







        share|cite|improve this answer















        share|cite|improve this answer




        share|cite|improve this answer








        edited Jun 14 at 13:11

























        answered Jun 14 at 10:57









        nalzoknalzok

        6827 silver badges17 bronze badges




        6827 silver badges17 bronze badges































            draft saved

            draft discarded















































            Thanks for contributing an answer to Cross Validated!


            • Please be sure to answer the question. Provide details and share your research!

            But avoid


            • Asking for help, clarification, or responding to other answers.

            • Making statements based on opinion; back them up with references or personal experience.

            Use MathJax to format equations. MathJax reference.


            To learn more, see our tips on writing great answers.




            draft saved


            draft discarded














            StackExchange.ready(
            function ()
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f412980%2fon-george-box-galit-shmueli-and-the-scientific-method%23new-answer', 'question_page');

            );

            Post as a guest















            Required, but never shown





















































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown

































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown









            Popular posts from this blog

            Tamil (spriik) Luke uk diar | Nawigatjuun

            Align equal signs while including text over equalitiesAMS align: left aligned text/math plus multicolumn alignmentMultiple alignmentsAligning equations in multiple placesNumbering and aligning an equation with multiple columnsHow to align one equation with another multline equationUsing \ in environments inside the begintabularxNumber equations and preserving alignment of equal signsHow can I align equations to the left and to the right?Double equation alignment problem within align enviromentAligned within align: Why are they right-aligned?

            Training a classifier when some of the features are unknownWhy does Gradient Boosting regression predict negative values when there are no negative y-values in my training set?How to improve an existing (trained) classifier?What is effect when I set up some self defined predisctor variables?Why Matlab neural network classification returns decimal values on prediction dataset?Fitting and transforming text data in training, testing, and validation setsHow to quantify the performance of the classifier (multi-class SVM) using the test data?How do I control for some patients providing multiple samples in my training data?Training and Test setTraining a convolutional neural network for image denoising in MatlabShouldn't an autoencoder with #(neurons in hidden layer) = #(neurons in input layer) be “perfect”?