Statistically-Indistinguishable Ensembles and the Evaluation of - - PowerPoint PPT Presentation

statistically indistinguishable ensembles and the
SMART_READER_LITE
LIVE PREVIEW

Statistically-Indistinguishable Ensembles and the Evaluation of - - PowerPoint PPT Presentation

Intro Opportunity Understanding Evaluation Outro References Statistically-Indistinguishable Ensembles and the Evaluation of Climate Models Corey Dethier University of Notre Dame Philosophy Department corey.dethier@gmail.com Feb 28, 2020


slide-1
SLIDE 1

Intro Opportunity Understanding Evaluation Outro References

Statistically-Indistinguishable Ensembles and the Evaluation of Climate Models

Corey Dethier

University of Notre Dame Philosophy Department corey.dethier@gmail.com

Feb 28, 2020

slide-2
SLIDE 2

Intro Opportunity Understanding Evaluation Outro References

Intro

slide-3
SLIDE 3

Intro Opportunity Understanding Evaluation Outro References

A problem

There are many different global climate models, and sometimes they don’t agree.

slide-4
SLIDE 4

Intro Opportunity Understanding Evaluation Outro References

A problem

There are many different global climate models, and sometimes they don’t agree. Example: global climate models deliver a range for “CO2 sensitivity” of 2.1˝ C to 4.7˝ C (IPCC Working Group 1 2013, 817). Seems to provide evidence that the true value is in this range.

slide-5
SLIDE 5

Intro Opportunity Understanding Evaluation Outro References

The standing view

Both climate scientists and philosophers have registered skepticism.

E.g.: Baumberger, Knutti, and Hadorn (2017), Justus (2012), Knutti, Allen, et al. (2008), Knutti, Furrer, et al. (2010), Parker (2011, 2018), Pirtle, Meyer, and Hamilton (2010), and Winsberg (2018)

slide-6
SLIDE 6

Intro Opportunity Understanding Evaluation Outro References

The standing view

Both climate scientists and philosophers have registered skepticism.

E.g.: Baumberger, Knutti, and Hadorn (2017), Justus (2012), Knutti, Allen, et al. (2008), Knutti, Furrer, et al. (2010), Parker (2011, 2018), Pirtle, Meyer, and Hamilton (2010), and Winsberg (2018)

The standard diagnosis: the group of models is a “ensemble of

  • pportunity.” Read: not like a random sample.
slide-7
SLIDE 7

Intro Opportunity Understanding Evaluation Outro References

My thesis

I think there’s a deeper problem. My diagnosis: uncertainty about (constraints on) the space of possible models. Recognizing this deeper problem helps us better understand and evaluate contemporary work within climate science.

slide-8
SLIDE 8

Intro Opportunity Understanding Evaluation Outro References

Plan for the talk

  • 1. (What’s wrong with) The ensemble of opportunity diagnosis.
  • 2. Understanding the statistically-indistinguishable paradigm.
  • 3. Evaluating the statistically-indistinguishable paradigm.
  • 4. Conclusion: “Are the models so out of touch? No, it’s the

meta-model that is wrong.”

slide-9
SLIDE 9

Intro Opportunity Understanding Evaluation Outro References

Ensembles of opportunity

slide-10
SLIDE 10

Intro Opportunity Understanding Evaluation Outro References

How to draw conclusions of groups of models

Treat a group of models like a sample from a population—that is, use statistics.

slide-11
SLIDE 11

Intro Opportunity Understanding Evaluation Outro References

How to draw conclusions of groups of models

Treat a group of models like a sample from a population—that is, use statistics. The standard diagnosis: the method of construction of actual ensembles isn’t like random sampling. My diagnosis: there’s uncertainty about the space of possible models.

slide-12
SLIDE 12

Intro Opportunity Understanding Evaluation Outro References

A thorough method

Method 1: just build a model for every possibility. Problems: Impractical. Only works if the possibilities are equally likely.

slide-13
SLIDE 13

Intro Opportunity Understanding Evaluation Outro References

Independent sampling

Method 2: build models that are representative of each component taken independently. Maybe what’s intended by “principled.”

slide-14
SLIDE 14

Intro Opportunity Understanding Evaluation Outro References

Independent sampling

Method 2: build models that are representative of each component taken independently. Maybe what’s intended by “principled.” But only works if each component is independent.

slide-15
SLIDE 15

Intro Opportunity Understanding Evaluation Outro References

Independent sampling

Method 2: build models that are representative of each component taken independently. Maybe what’s intended by “principled.” But only works if each component is independent.

slide-16
SLIDE 16

Intro Opportunity Understanding Evaluation Outro References

The problem, then

Takeaway: in order to even say what a “principled” construction method is, we need background knowledge about the constraints

  • n the set of models.

And that knowledge isn’t being invoked in theoretical discussions

  • f evaluation.
slide-17
SLIDE 17

Intro Opportunity Understanding Evaluation Outro References

Understanding “statistically-indistinguishable” ensembles

slide-18
SLIDE 18

Intro Opportunity Understanding Evaluation Outro References

Forgetting about construction

An alternative means of justifying inferences from a given ensemble: use proxies to check whether the ensemble is representative.

slide-19
SLIDE 19

Intro Opportunity Understanding Evaluation Outro References

Forgetting about construction

An alternative means of justifying inferences from a given ensemble: use proxies to check whether the ensemble is representative. A different problem: proxies indicate that extant ensembles aren’t representative.

slide-20
SLIDE 20

Intro Opportunity Understanding Evaluation Outro References

First, the problem

The problem, very roughly pictured:

(a) Ensemble is representative (b) Ensemble is too wide (c) Ensemble is too narrow

slide-21
SLIDE 21

Intro Opportunity Understanding Evaluation Outro References

First, the problem

The problem, very roughly pictured:

(a) Ensemble is representative (b) Ensemble is too wide (c) Ensemble is too narrow

slide-22
SLIDE 22

Intro Opportunity Understanding Evaluation Outro References

The solution

A number of climate scientists—most prominently Annan and Hargreaves (2010, 2011, 2017)—have argued that this result is misleading, because it relies on a particular statistical “paradigm.”

slide-23
SLIDE 23

Intro Opportunity Understanding Evaluation Outro References

The solution

A number of climate scientists—most prominently Annan and Hargreaves (2010, 2011, 2017)—have argued that this result is misleading, because it relies on a particular statistical “paradigm.” “Truth-centered” paradigm: ensemble-proxy relationship is like that between a sample and a population mean. “Statistically indistinguishable” paradigm: ensemble-proxy relationship is like that between a sample and a population member.

slide-24
SLIDE 24

Intro Opportunity Understanding Evaluation Outro References

The statistically-indistinguishable advantage

Given the SI paradigm:

(a) Ensemble is representative (b) Ensemble is too wide (c) Ensemble is too narrow

slide-25
SLIDE 25

Intro Opportunity Understanding Evaluation Outro References

The statistically-indistinguishable advantage

Given the SI paradigm:

(a) Ensemble is representative (b) Ensemble is too wide (c) Ensemble is too narrow

slide-26
SLIDE 26

Intro Opportunity Understanding Evaluation Outro References

Understanding the framework

The upshot: if SI is the right paradigm, we can draw some conclusions from groups of models. Not because we have a new construction method. But because model evaluation provides us with sufficient background knowledge about the relationship between ensemble and world to justify said conclusions.

slide-27
SLIDE 27

Intro Opportunity Understanding Evaluation Outro References

Evaluating “statistically-indistinguishable” ensembles

slide-28
SLIDE 28

Intro Opportunity Understanding Evaluation Outro References

Are they right?

slide-29
SLIDE 29

Intro Opportunity Understanding Evaluation Outro References

Are they right?

Yes and no. More specifically: I don’t think this buys all the inferences we want—particularly when it comes to the future.

slide-30
SLIDE 30

Intro Opportunity Understanding Evaluation Outro References

Paradigms and predictions

Evaluation provides justification iff the proxy and the target can be assumed to be similar.

slide-31
SLIDE 31

Intro Opportunity Understanding Evaluation Outro References

Paradigms and predictions

Evaluation provides justification iff the proxy and the target can be assumed to be similar. In the context of future predictions about the climate, however, the assumption that the proxy (contemporary climate) is like the future in any sense is substantive.

slide-32
SLIDE 32

Intro Opportunity Understanding Evaluation Outro References

Whence the extra power?

Recall: the truth-centered worry was the existence of models more extreme than extant ensembles.

slide-33
SLIDE 33

Intro Opportunity Understanding Evaluation Outro References

Whence the extra power?

Recall: the truth-centered worry was the existence of models more extreme than extant ensembles. If we take the shift in paradigm to provide us with (extra) justification for future predictions, we essentially rule this worry out by fiat. That is: by way of an assumption about the nature of the space of possible models.

slide-34
SLIDE 34

Intro Opportunity Understanding Evaluation Outro References

The main point

Note that this assumption may well be justified.

slide-35
SLIDE 35

Intro Opportunity Understanding Evaluation Outro References

The main point

Note that this assumption may well be justified. My point is that the evaluation of the SI paradigm turns on our knowledge about the space of possible models. And doesn’t have anything much to do with construction methods.

slide-36
SLIDE 36

Intro Opportunity Understanding Evaluation Outro References

Outro

slide-37
SLIDE 37

Intro Opportunity Understanding Evaluation Outro References

The takeaway

I’ve argued that the problem that we face is uncertainty about the space of possible models. I could be wrong—particularly about the evaluative point.

slide-38
SLIDE 38

Intro Opportunity Understanding Evaluation Outro References

The takeaway

I’ve argued that the problem that we face is uncertainty about the space of possible models. I could be wrong—particularly about the evaluative point. Maybe we still haven’t identified the right the meta-“paradigm”; after all, both SI and the traditional alternative assume that the ensemble is like a random sample of something.

slide-39
SLIDE 39

Intro Opportunity Understanding Evaluation Outro References

Thank you

Thank you!

slide-40
SLIDE 40

Intro Opportunity Understanding Evaluation Outro References

Annan, James D. and Julia C. Hargreaves (2010). Reliability of the CMIP3 Ensemble. Geophysical Research Letters 37: 1–5. – (2011). Understanding the CMIP3 Model Ensemble. Journal of Climate 24: 4529–38. – (2017). On the Meaning of Independence in Climate Science. Earth Systems Dynamics 8: 211–24. Baumberger, Christoph, Reto Knutti, and Gertrude Hirsch Hadorn (2017). Building Confidence in Climate Model Projections: An Analysis of Inferences From Fit. Wiley Interdisciplinary Reviews: Climate Change 8.3: e454. IPCC Working Group 1 (2013). Climate Change 2013: The Physical Science Basis. Ed. by Thomas F. Stocker et al. Fifth Assessment Report of the Intergovernmental Panel on Climate

  • Change. Cambridge: Cambridge University Press.

Justus, James (2012). The Elusive Basis of Inferential Robustness. Philosophy of Science 79.5: 795–807.

slide-41
SLIDE 41

Intro Opportunity Understanding Evaluation Outro References

Knutti, Reto, Myles R. Allen, et al. (2008). A Review of Uncertainties in Global Temperature Projections over the Twenty-First Century. Journal of Climate 21.11: 2651–63. Knutti, Reto, Reinhard Furrer, et al. (2010). Challenges in Combining Projections from Multiple Climate Models. Journal of Climate 25.10: 2739–58. Parker, Wendy S. (2011). When Climate Models Agree: The Significance of Robust Model Predictions. Philosophy of Science 78.4: 579–600. – (2018). The Significance of Robust Climate Projections. In: Climate Modeling: Philosophical and Conceptual Issues. Ed. by Elisabeth A. Lloyd and Eric Winsberg. Cham: Palgrave Macmillan: 273–96. Pirtle, Zach, Ryan Meyer, and Andrew Hamilton (2010). What Does it Mean when Climate Models Agree? A Case for Assessing Independence Among General Circulation Models. Environmental Science & Policy 13.5: 351–61.

slide-42
SLIDE 42

Intro Opportunity Understanding Evaluation Outro References

Winsberg, Eric (2018). What does Robustness Teach us in Climate Science: A Re-Appraisal. Synthese (online first).