<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.0 20120330//EN" "JATS-journalpublishing1.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">NEJSDS</journal-id>
<journal-title-group><journal-title>The New England Journal of Statistics in Data Science</journal-title></journal-title-group>
<issn pub-type="ppub">2693-7166</issn><issn-l>2693-7166</issn-l>
<publisher>
<publisher-name>New England Statistical Society</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">NEJSDS55</article-id>
<article-id pub-id-type="doi">10.51387/23-NEJSDS55</article-id>
<article-id pub-id-type="arxiv">2308.14959</article-id>
<article-categories>
<subj-group subj-group-type="heading"><subject>Methodology Article</subject></subj-group>
<subj-group subj-group-type="area"><subject>Statistical Methodology</subject></subj-group>
</article-categories>
<title-group>
<article-title>Improving Data Analysis by Testing by Betting: Optional Continuation and Descriptive Statistics</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Shafer</surname><given-names>Glenn</given-names></name><email xlink:href="mailto:gshafer@business.rutgers.edu">gshafer@business.rutgers.edu</email><xref ref-type="aff" rid="j_nejsds55_aff_001"/>
</contrib>
<aff id="j_nejsds55_aff_001"><institution>Rutgers Business School and Department of Statistics</institution>, 1 Washington Park, Newark, New Jersey 07102, <country>USA</country>. E-mail address: <email xlink:href="mailto:gshafer@business.rutgers.edu">gshafer@business.rutgers.edu</email></aff>
</contrib-group>
<pub-date pub-type="ppub"><year>2024</year></pub-date><pub-date pub-type="epub"><day>13</day><month>12</month><year>2023</year></pub-date><volume>2</volume><issue>2</issue><fpage>215</fpage><lpage>228</lpage><history><date date-type="accepted"><day>2</day><month>12</month><year>2023</year></date></history>
<permissions><copyright-statement>© 2024 New England Statistical Society</copyright-statement><copyright-year>2024</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>Open access article under the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/">CC BY</ext-link> license.</license-p></license></permissions>
<abstract>
<p>When testing a statistical hypothesis, is it legitimate to deliberate on the basis of initial data about whether and how to collect further data? Game-theoretic probability’s <italic>fundamental principle for testing by betting</italic> says yes, provided that you are testing the hypothesis’s predictions by betting and do not risk more capital than initially committed. Standard statistical theory uses <italic>Cournot’s principle</italic>, which does not allow such <italic>optional continuation</italic>. Cournot’s principle can be extended to allow optional continuation when testing is carried out by multiplying likelihood ratios, but the extension lacks the simplicity and generality of testing by betting.</p>
<p>Testing by betting can also help us with descriptive data analysis. To obtain a purely and honestly descriptive analysis using competing probability distributions, we have them bet against each other using the principle. The place of confidence intervals is then taken by sets of distributions that do relatively well in the competition. In the simplest implementation, these sets coincide with R. A. Fisher’s likelihood ranges.</p>
</abstract>
<kwd-group>
<label>Keywords and phrases</label>
<kwd>Game-theoretic probability</kwd>
<kwd>Game-theoretic statistics</kwd>
<kwd>Optional continuation</kwd>
<kwd>Optional stopping</kwd>
<kwd>Cournot’s principle</kwd>
<kwd>Fundamental principle of testing by betting</kwd>
<kwd>Ville’s inequality</kwd>
<kwd>Descriptive statistics</kwd>
<kwd>Kelly betting</kwd>
<kwd>Likelihood</kwd>
<kwd>Probability forecasting</kwd>
<kwd>Convenience sample</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="j_nejsds55_s_001">
<label>1</label>
<title>Introduction</title>
<p>Game-theoretic probability studies games in which announcements by Player I define rates at which Player II can bet on outcomes. The games can have one or many rounds. Player I’s announcements can take various forms. One possibility is that Player I announces a probability distribution and Player II is authorized to buy any payoff for its expected value.</p>
<p>A probability distribution for a discrete-time stochastic process can be used as a strategy for Player I; it tells Player I what probability distribution for the <italic>n</italic>th outcome to announce after seeing the first <inline-formula id="j_nejsds55_ineq_001"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$n-1$]]></tex-math></alternatives></inline-formula> outcomes. When Player I is required to follow a particular strategy of this form, strategies for Player II define martingales. Theorems about discrete-time stochastic processes become theorems in game theory [<xref ref-type="bibr" rid="j_nejsds55_ref_052">52</xref>, <xref ref-type="bibr" rid="j_nejsds55_ref_054">54</xref>].</p>
<p>The betting intuition brought to the fore by game-theoretic probability has proven productive for statistical theory, sometimes helping mathematical statisticians find more efficient or more powerful methods [<xref ref-type="bibr" rid="j_nejsds55_ref_043">43</xref>]. More importantly, game-theoretic probability can help us go beyond today’s standard mental framework for mathematical statistics, where we begin with a collection of probability distributions and test whether one of these distributions could have “generated” our data or try to decide which one did so. This article discusses two ways in which game-theoretic probability goes beyond the standard framework to help make data analysis more flexible and more honest.</p>
<p>Considered purely as mathematics, game-theoretic probability generalizes standard probability theory, because it does not assume that any of the players are required to follow a strategy specified in advance.<xref ref-type="fn" rid="j_nejsds55_fn_001">1</xref><fn id="j_nejsds55_fn_001"><label><sup>1</sup></label>
<p>Game-theoretic probability also allows Player I to make more limited betting offers, such as those considered in the theory of imprecise probabilities [<xref ref-type="bibr" rid="j_nejsds55_ref_002">2</xref>, <xref ref-type="bibr" rid="j_nejsds55_ref_056">56</xref>] or more comprehensive betting offers, such as those considered by [<xref ref-type="bibr" rid="j_nejsds55_ref_057">57</xref>]. I do not study these aspects of game-theoretic probability’s flexibility in this paper.</p></fn> This allows us to authorize, in a simple and explicit way, deliberation on the basis of initial data about (1) whether and how to collect further data and (2) how to use it in statistical testing. We already see such deliberation in practice, but squaring it with standard probability theory is not so simple and limits the tests authorized to those used in the game-theoretic picture. I discuss this point in detail in Section <xref rid="j_nejsds55_s_002">2</xref>.</p>
<p>In Section <xref rid="j_nejsds55_s_012">3</xref>, I consider a second way game-theoretic probability can help us improve data analysis. Ever since statistical testing began to be widely used in the early 19th century, its greatest abuse has been unjustified and often clearly erroneous assumptions of randomness. Sometimes the assumption is that successive observations are chosen randomly from a population. Sometimes the assumption is that the deviations of successive observations from a model are random – i.e., independent of each other and of the explanatory variables in the model. In some cases, practitioners acknowledge that the assumptions are unjustified but claim that their analyses are nevertheless interesting as descriptions of data. Because the statistical language being used is at least implicitly causal, this usually sounds like double talk. The betting picture can help here by providing a language that is clearly descriptive rather than causal. We rank the probability distributions in a model by having them bet against each other’s predictions. This produces a set of distributions that did best in the competition. This set is not a confidence set. There is no suggestion that we are confident about anything, and we are certainly not confident that one of the distributions in the set “generated” the data. The set is simply a set of distributions that predicted the data better than the other distributions in the model. The betting language provides an intuitive scale for this predictive success.</p>
<p>These two improvements in data analysis are not so much improvements in what we do as improvements in how we understand and explain what we do. In practice, researchers already deliberate about how to continue experimentation. Some already use the descriptive sets we obtain in Section <xref rid="j_nejsds55_s_012">3</xref>; they are called “likelihood ranges”. Aside from setting aside some types of testing that cannot be legitimately authorized when we deliberate about continuation, the proposed improvements concern how we can communicate data analyses more clearly and more honestly to ourselves and to others. But it is also notable and important that the betting language allows us to generalize seamlessly to the case where predictions being tested or compared are not necessarily produced by probability models. They may instead be produced by algorithms of a very different kind — perhaps neural nets, perhaps physical models like those used in weather forecasting.</p>
</sec>
<sec id="j_nejsds55_s_002">
<label>2</label>
<title>Statistical Testing with Optional Continuation</title>
<p>Statistical testing requires a mathematical theory of probability together with a principle that specifies how probabilities can be discredited by observations. 
<list>
<list-item id="j_nejsds55_li_001">
<label>•</label>
<p>The principle used to make traditional probability theory into a theory of statistical testing is sometimes called <italic>Cournot’s principle</italic>.<xref ref-type="fn" rid="j_nejsds55_fn_002">2</xref><fn id="j_nejsds55_fn_002"><label><sup>2</sup></label>
<p>See [<xref ref-type="bibr" rid="j_nejsds55_ref_053">53</xref>] and [<xref ref-type="bibr" rid="j_nejsds55_ref_046">46</xref>, <xref ref-type="bibr" rid="j_nejsds55_ref_051">51</xref>] for the history of Cournot’s principle. The principle is sometimes ridiculed by philosophers; see for example [<xref ref-type="bibr" rid="j_nejsds55_ref_035">35</xref>, p. 49] and [<xref ref-type="bibr" rid="j_nejsds55_ref_014">14</xref>, pp. 66–67]. But it has been articulated in one way or another by a panoply of mathematicians and statisticians, including Jacob Bernoulli, Antoine-Augustin Cournot himself, Émile Borel, Richard von Mises, Andrei Kolmgorov, Abraham Wald, Joseph L. Doob, William Feller, Harold Jeffreys, Charles Stein, and Philip Dawid.</p></fn> This principle authorizes a statistician to select an event to which a probability distribution assigns small probability and to regard its happening as evidence against the distribution.</p>
</list-item>
<list-item id="j_nejsds55_li_002">
<label>•</label>
<p>To make game-theoretic probability into a theory of statistical testing, we can use a principle that I have called <italic>the fundamental principle for testing by betting</italic>.<xref ref-type="fn" rid="j_nejsds55_fn_003">3</xref><fn id="j_nejsds55_fn_003"><label><sup>3</sup></label>
<p>Vovk and I have used various other names for this principle. In 2001, we called it <italic>the fundamental interpretative hypothesis of probability</italic> [<xref ref-type="bibr" rid="j_nejsds55_ref_052">52</xref>, pp. 5, 14, 62]. In 2019, we called it <italic>the game-theoretic version of Cournot’s principle</italic> [<xref ref-type="bibr" rid="j_nejsds55_ref_054">54</xref>, pp. 226–227].</p></fn> This principle, which is related to but distinct from Cournot’s principle, authorizes a statistician to interpret success in betting against a probability distribution as evidence against the distribution.</p>
</list-item>
</list> 
Do these principles authorize optional continuation?</p>
<p>As the term is used here, <italic>optional continuation</italic> refers to the practice of deliberating, after seeing some initial data, about whether and how to continue collecting and analyzing data. Such continuation may involve observations or experiments not contemplated at the outset. It is distinguished from optional stopping, which refers in established usage only to the possibility of adopting at the outset a plan specifying circumstances under which we will curtail a fully planned sequential experiment or observational study.</p>
<p>The fundamental principle for testing by betting asserts the validity of optional continuation for the type of testing it considers. Cournot’s principle, in its classical formulation, does not. It can be extended to assert the validity of optional continuation when testing is carried out by multiplying likelihood ratios, but as I will explain, the extension lacks the simplicity and generality of the fundamental principle for testing by betting.</p>
<sec id="j_nejsds55_s_003">
<label>2.1</label>
<title>Optional Continuation in Practice and Theory</title>
<p>Optional continuation has long been part of statistical practice. It is implicit, for example, in the idea of meta-analysis. But it has proven difficult to bring it under the purview of statistical theory.</p>
<p>The term “optional continuation” with the meaning used here first appeared in print in Allard Hendriksen’s master’s thesis at the University of Leiden, written under the supervision of Peter Grünwald [<xref ref-type="bibr" rid="j_nejsds55_ref_033">33</xref>]. Hendriksen wrote on page 3 of the thesis, <disp-quote>
<p>“Optional continuation” is the practice of combining evidence of studies that were done because of promising results of previous research on the same subject.</p></disp-quote> The term has subsequently been used in other work by Günwald’s machine-learning research group at CWI in Amsterdam [<xref ref-type="bibr" rid="j_nejsds55_ref_032">32</xref>, <xref ref-type="bibr" rid="j_nejsds55_ref_031">31</xref>]. But as of June 13, 2023, it had not yet appeared in any of the 34 statistics journals in JSTOR.</p>
<p>The older term “optional stopping” was introduced by the Duke mathematician Joseph Albert Greenwood [<xref ref-type="bibr" rid="j_nejsds55_ref_029">29</xref>]. Greenwood sought empirical adjustments to account for the way Joseph Rhine’s laboratory was conducting and analyzing its experiments on extra-sensory perception. Rhine stopped experimenting with each subject when a success rate thought to be statistically significant was achieved, then combined the <italic>z</italic>-scores achieved by successive subjects.</p>
<p>Greenwood’s problem was brought to wider attention in mathematical statistics by William Feller’s critiques of the ESP work [<xref ref-type="bibr" rid="j_nejsds55_ref_020">20</xref>, pp. 272, 286–292], [<xref ref-type="bibr" rid="j_nejsds55_ref_019">19</xref>, pp. 140, 190, 197]. In subsequent work in probability, “optional stopping” has referred to stopping rules that can be adopted in advance without annulling a desired property of a stochastic process, usually the property of being a martingale [<xref ref-type="bibr" rid="j_nejsds55_ref_016">16</xref>].</p>
<p>In his book on sequential analysis [<xref ref-type="bibr" rid="j_nejsds55_ref_059">59</xref>], Abraham Wald considered only “sequential sampling plans” chosen in advance. While allowing early stopping when there was enough evidence to make a decision, these plans specified whether or not to stop and how to continue if stopping was not mandated, all as a function of outcomes so far. In a review of the book, George Barnard wrote that sequential analysis marked “the entry of statistical considerations into the very process of experimentation itself” [<xref ref-type="bibr" rid="j_nejsds55_ref_003">3</xref>]. We know that the process of experimentation often involves not only plans adopted in advance but also opportunistic changes in plans, based on new insights and unexpected information.</p>
<p>Barnard seems not to have followed up on his insight concerning the role of statistics in the process of experimentation; he did not discuss it, for example in his major article on statistical inference [<xref ref-type="bibr" rid="j_nejsds55_ref_004">4</xref>]. But in a subsequent article entitled “Sequential experimentation”, R. A. Fisher wrote about the need for sequential deliberation in these terms [<xref ref-type="bibr" rid="j_nejsds55_ref_021">21</xref>, p. 183]: <disp-quote>
<p>The present use of the term sequential is intended to be of a broader import than the formal use of the word as associated with the systematic procedure known as sequential analysis. The experimenter does not regard his material as wholly passive but instead looks to what may be learnt from it with a view to the improvement and extension of the enquiry. This willingness to learn from it how to proceed is the essential quality of sequential procedures. Wald introduced the sequential test, but the sequential idea is much older. For example, what is the policy of a research unit? It is that in time we may learn to do better and follow up our more promising results. The essence of sequential experimentation is a series of experiments each of which depends on what has gone before. For example, in a sample survey scheme, as explained by Yates, a pilot survey is intended to supply a basis for efficiently planning the subsequent stages of a survey. …</p></disp-quote> Until the recent work on optional continuation, this insight about statistical practice has remained outside the ambit of statistical theory.</p>
</sec>
<sec id="j_nejsds55_s_004">
<label>2.2</label>
<title>A Betting Game with Optional Continuation</title>
<p>The simplest game used in game-theoretic probability has three players: Forecaster makes probability predictions, Skeptic bets against them, and Reality announces the outcomes. The game is a perfect-information game, meaning that the players move in turn and see each other’s moves. We can vary the rules of the game, but we need not impose any further condition on what information any player might have or acquire in the course of the game, or on how the players might collaborate. Forecaster and Skeptic might be the same person. Forecaster and Reality might be the same person.</p>
<p>If Forecaster keeps forecasting, Skeptic can keep betting. Forecaster need not follow a plan or strategy about what to forecast next or how to forecast it.<xref ref-type="fn" rid="j_nejsds55_fn_004">4</xref><fn id="j_nejsds55_fn_004"><label><sup>4</sup></label>
<p>To see how probability’s limit theorems can be generalized to accommodate Forecaster’s freedom, see [<xref ref-type="bibr" rid="j_nejsds55_ref_054">54</xref>, ğ7.5].</p></fn> Even if Forecaster follows a strategy known to Skeptic, Skeptic need not have a plan or strategy for when or how to bet against the forecasts. Thus optional continuation is built into the game, for both Forecaster and Skeptic. Skeptic can decide whether and how to continue selecting from Forecaster’s betting offers, but Forecaster can decide what experiments or observations to make and what forecasts (perhaps probabilities) to give for them.</p>
<p>Vovk and I have used the example of quantum mechanics to illustrate game-theoretic probability’s capacity for optional continuation; see [<xref ref-type="bibr" rid="j_nejsds55_ref_052">52</xref>, pp. 189–191] and [<xref ref-type="bibr" rid="j_nejsds55_ref_054">54</xref>, pp. 215–217]. In this example, we split Forecaster into two players, Observer and Quantum Mechanics. Observer selects the experiment, and Quantum Mechanics makes the probability forecast. Formally, the game continues indefinitely, and both Observer and Skeptic can effectively stop it by making null moves.</p>
<p>Although optional continuation is built into the game, we need this principle to use the game in statistical testing:</p><statement id="j_nejsds55_stat_001"><label>Principle 1</label>
<title>(Fundamental principle for testing by betting).</title>
<p><italic>Successive bets against a forecaster that begin with unit capital and never risk more discredit the forecaster to the extent that the final capital is large.</italic><xref ref-type="fn" rid="j_nejsds55_fn_005">5</xref><fn id="j_nejsds55_fn_005"><label><sup>5</sup></label>
<p>I first formulated the principle in these words in my SIPTA lectures [<xref ref-type="bibr" rid="j_nejsds55_ref_049">49</xref>].</p></fn></p></statement>
<p>In one sense, this says it all. But some elaboration may be useful:</p>
<list>
<list-item id="j_nejsds55_li_003">
<label>1.</label>
<p>The principle is <italic>fundamental</italic>, not the consequence of some more extensive philosophy or methodology. We do not begin by saying that the forecaster’s probabilities are or should be objective, subjective, personal, “frequentist”, or whatever. We are testing the forecaster qua forecaster, and so we are testing his forecasts qua forecasts; the question is only whether they are good forecasts, relative to the knowledge and skill of whoever is doing the testing.</p>
</list-item>
<list-item id="j_nejsds55_li_004">
<label>2.</label>
<p>The forecaster may give a probability for a single event <italic>A</italic>, a probability distribution for an outcome <italic>X</italic>, or something less than a probability or a probability distribution:</p>
<list>
<list-item id="j_nejsds55_li_005">
<label>•</label>
<p>If the forecaster gives a probability, you may bet on either side at the corresponding odds.</p>
</list-item>
<list-item id="j_nejsds55_li_006">
<label>•</label>
<p>If the forecaster gives a probability distribution for <italic>X</italic>, you may buy or sell any payoff <inline-formula id="j_nejsds55_ineq_002"><alternatives><mml:math>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$S(X)$]]></tex-math></alternatives></inline-formula> for its expected value.</p>
</list-item>
<list-item id="j_nejsds55_li_007">
<label>•</label>
<p>If the forecaster gives only an estimate <italic>E</italic> of <italic>X</italic>, you may buy or sell <italic>X</italic> for <italic>E</italic>.</p>
</list-item>
<list-item id="j_nejsds55_li_008">
<label>•</label>
<p>If the forecaster repeatedly gives a new probability for <italic>A</italic> or new estimate for <italic>X</italic>, say daily, you may buy or sell tomorrow’s price for today’s price.</p>
</list-item>
<list-item id="j_nejsds55_li_009">
<label>•</label>
<p>If the forecaster gives upper and lower previsions, you may buy at the upper or sell at the lower.</p>
</list-item>
</list>
</list-item>
<list-item id="j_nejsds55_li_010">
<label>3.</label>
<p>You <italic>begin with unit capital</italic> only for mathematical convenience. The discredit is measured by the ratio (final capital)/(initial capital), which I call the <italic>betting score</italic>.<xref ref-type="fn" rid="j_nejsds55_fn_006">6</xref><fn id="j_nejsds55_fn_006"><label><sup>6</sup></label>
<p>The alternative name <italic>e-value</italic>, designed to resemble <italic>p-value</italic> and avoid reference to betting, has recently become popular [<xref ref-type="bibr" rid="j_nejsds55_ref_058">58</xref>, <xref ref-type="bibr" rid="j_nejsds55_ref_043">43</xref>].</p></fn></p>
</list-item>
<list-item id="j_nejsds55_li_011">
<label>4.</label>
<p>If you make several bets against the same forecaster (or the same theory or closely related theories), each starting with its own capital, then you are not allowed to report only the cases where you discredited the forecaster. Instead, you must report the overall result, the sum of your final capital over all the bets divided by the sum of your initial capital over all the bets.</p>
</list-item>
<list-item id="j_nejsds55_li_012">
<label>5.</label>
<p>When betting against successive forecasts, each bet uses only the cumulative capital remaining after the previous bet. You may not borrow or otherwise raise more capital in order to continue betting. This is what <italic>never risk more</italic> than the initial capital means.</p>
</list-item>
<list-item id="j_nejsds55_li_013">
<label>6.</label>
<p>When you stop, you must compare your initial capital with your <italic>final</italic> capital. You cannot claim to have discredited the forecaster because you had reached a higher level of capital in the interim. You do not have the money if you kept betting and lost it.<xref ref-type="fn" rid="j_nejsds55_fn_007">7</xref><fn id="j_nejsds55_fn_007"><label><sup>7</sup></label>
<p>The anonymous 13th-century author who left us with the earliest surviving calculation of the chances for a throw of three dice warned us [<xref ref-type="bibr" rid="j_nejsds55_ref_034">34</xref>, p. 172]: “Addeque, quod lusor se continuare lucrando nescit, perdendo nescit dimittere ludum.” Not knowing how to maintain his luck when winning, the gambler does not know how to quit when losing.</p></fn></p>
</list-item>
</list>
<p>I have stated the fundamental principle for testing by betting in 26 words, then taken a page to explain it. Is the principle simple? In any case, it is coherent and teachable. In contexts where the forecasts are only single probabilities or estimates, the principle can be taught even to those who have never studied mathematical probability. Moreover, the principle builds on ideas about betting that most people acquire before ever studying mathematical probability. Too many predictions contradicted by experience discredit the person making them. If you lose too much money betting on something, you are not much of an expert about it. Etc.</p>
</sec>
<sec id="j_nejsds55_s_005">
<label>2.3</label>
<title>Cournot’s Principle in Classical Form</title>
<p>What principles must we add to traditional probability theory to allow optional continuation?</p>
<p>Before addressing this question, I discuss a more basic question: How are we authorized to discredit a probability distribution <italic>P</italic> using observations? As I have already mentioned, the classical answer is Cournot’s principle: we select an event <italic>E</italic> that has small probability <inline-formula id="j_nejsds55_ineq_003"><alternatives><mml:math>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$P(E)$]]></tex-math></alternatives></inline-formula> (call <italic>E</italic> our <italic>test event</italic>). The probability distribution <italic>P</italic> is discredited if <italic>E</italic> happens; we prefer to believe that the probabilities are incorrect rather than think that this improbable event happened.</p><statement id="j_nejsds55_stat_002"><label>Principle 2</label>
<title>(Cournot’s principle).</title>
<p><italic>If we specify an event E in advance, and E happens, then we may take α, the probability of E, as a measure of evidence against P. The magnitude of discredit is measured by how small α and thus how large</italic> <inline-formula id="j_nejsds55_ineq_004"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi></mml:math><tex-math><![CDATA[$1/\alpha $]]></tex-math></alternatives></inline-formula> <italic>is.</italic></p></statement>
<p>We may call <inline-formula id="j_nejsds55_ineq_005"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi></mml:math><tex-math><![CDATA[$1/\alpha $]]></tex-math></alternatives></inline-formula> our <italic>test score</italic>: 
<disp-formula id="j_nejsds55_eq_001">
<label>(2.1)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mtext>test score</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mfenced separators="" open="{" close="">
<mml:mrow>
<mml:mtable columnspacing="10.0pt" equalrows="false" columnlines="none" equalcolumns="false" columnalign="left left">
<mml:mtr>
<mml:mtd class="array">
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd class="array">
<mml:mtext>if</mml:mtext>
<mml:mspace width="2.5pt"/>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mspace width="2.5pt"/>
<mml:mtext>happens</mml:mtext>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="array">
<mml:mn>0</mml:mn>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd class="array">
<mml:mtext>if</mml:mtext>
<mml:mspace width="2.5pt"/>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mspace width="2.5pt"/>
<mml:mtext>does not happen</mml:mtext>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \text{test score}=\left\{\begin{array}{l@{\hskip10.0pt}l}1/\alpha \hspace{1em}& \text{if}\hspace{2.5pt}E\hspace{2.5pt}\text{happens}\\ {} 0\hspace{1em}& \text{if}\hspace{2.5pt}E\hspace{2.5pt}\text{does not happen}.\end{array}\right.\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>Cournot’s principle can be considered a special case of the fundamental principle for testing, because <inline-formula id="j_nejsds55_ineq_006"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi></mml:math><tex-math><![CDATA[$1/\alpha $]]></tex-math></alternatives></inline-formula> is the capital that would result from <italic>E</italic>’s happening if you bet unit capital on <italic>E</italic>. The test score (<xref rid="j_nejsds55_eq_001">2.1</xref>) is also a betting score.</p>
<p>Although Cournot’s principle has long been fundamental to statistical theory, current philosophical fashion has made it difficult to teach. A frequent objection is that some event of small probability always happens. When we hear this objection, we emphasize “specified in advance”, which requires less emphasis when testing by betting, because a bet must be made in advance.</p>
<p>In some cases, we say “simple to describe” instead of or in addition to “specified in advance”. Simplicity is also implicit to some extent when testing by betting, because a bet cannot be made and implemented unless the event is relatively simple.</p>
</sec>
<sec id="j_nejsds55_s_006">
<label>2.4</label>
<title>Extending Cournot’s Principle to Test Variables</title>
<p>This extension of Cournot’s principle does not require us to specify in an advance a goal <inline-formula id="j_nejsds55_ineq_007"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi></mml:math><tex-math><![CDATA[$1/\alpha $]]></tex-math></alternatives></inline-formula> for the strength of the evidence.</p>
<p>Suppose <italic>S</italic> is a nonnegative random variable, chosen in advance and so not too hard to describe, with expected value <inline-formula id="j_nejsds55_ineq_008"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[${E_{P}}(S)=1$]]></tex-math></alternatives></inline-formula> (call <italic>S</italic> our <italic>test variable</italic>). Our next principle says that a realized value <italic>s</italic> of <italic>S</italic> discredits <italic>P</italic> to the extent that <italic>s</italic> is much larger than 1.</p><statement id="j_nejsds55_stat_003"><label>Principle 3</label>
<title>(Authorization to test with a test variable).</title>
<p><italic>If we specify a test variable S in advance, then we may take s, the observed value of S, as a measure of evidence against P. We interpret s (our test score) on the same scale as we use in Cournot’s principle. In other words, when</italic> <inline-formula id="j_nejsds55_ineq_009"><alternatives><mml:math>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi></mml:math><tex-math><![CDATA[$s=1/\alpha $]]></tex-math></alternatives></inline-formula><italic>, it has the same force against P as the happening of a pre-specified event E when</italic> <inline-formula id="j_nejsds55_ineq_010"><alternatives><mml:math>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi></mml:math><tex-math><![CDATA[$P(E)=\alpha $]]></tex-math></alternatives></inline-formula><italic>.</italic></p></statement>
<p>Cournot’s principle is the special case of Principle <xref rid="j_nejsds55_stat_003">3</xref> where <italic>S</italic> is given by (<xref rid="j_nejsds55_eq_001">2.1</xref>). Principle <xref rid="j_nejsds55_stat_003">3</xref> adds to Cournot’s principle the possibility of a more graduated report on the strength of the evidence against <italic>P</italic>. The test scores it provides can be compared with p-values, which also provide a graduated report. A p-value <italic>p</italic> discredits <italic>P</italic> to the extent that it is small or <inline-formula id="j_nejsds55_ineq_011"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi></mml:math><tex-math><![CDATA[$1/p$]]></tex-math></alternatives></inline-formula> is large. But a given value of <inline-formula id="j_nejsds55_ineq_012"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi></mml:math><tex-math><![CDATA[$1/p$]]></tex-math></alternatives></inline-formula> carries less force than the same value <italic>s</italic> of a test score from a test variable. See ğ<xref rid="j_nejsds55_s_010">2.8</xref>.</p>
<p>It might seem that the greater flexibility offered by a test variable <italic>S</italic> comes at a price. When <italic>s</italic> is the realized value, the events <inline-formula id="j_nejsds55_ineq_013"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$\{S=s\}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds55_ineq_014"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$\{S\le s\}$]]></tex-math></alternatives></inline-formula> happen, and Markov’s inequality tells us that our score <inline-formula id="j_nejsds55_ineq_015"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$1/P(E)$]]></tex-math></alternatives></inline-formula> would have been at least as great, often greater, had we chosen one of these events as our test event <italic>E</italic>. But of course we could not have made these choices, because we did not know <italic>s</italic> in advance.</p>
<p>Like the classical form of Cournot’s principle, Principle 3 can be considered a special case of the fundamental principle for testing by betting. The observed value <italic>s</italic> of the test variable <italic>S</italic> is the capital that would result from buying <italic>S</italic> for its expected value. The test variable <italic>S</italic> is our bet, and <italic>s</italic> is our betting score.</p>
</sec>
<sec id="j_nejsds55_s_007">
<label>2.5</label>
<title>Extending Cournot’s Principle to Test Martingales</title>
<p>Now suppose we want to test a probability distribution <italic>P</italic> for a stochastic process <inline-formula id="j_nejsds55_ineq_016"><alternatives><mml:math>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>:</mml:mo>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mspace width="0.1667em"/></mml:math><tex-math><![CDATA[$X:={X_{1}},{X_{2}},\dots \hspace{0.1667em}$]]></tex-math></alternatives></inline-formula>, and we observe the <inline-formula id="j_nejsds55_ineq_017"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{t}}$]]></tex-math></alternatives></inline-formula> successively. We use a <italic>test martingale</italic>, a nonnegative martingale <inline-formula id="j_nejsds55_ineq_018"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mspace width="0.1667em"/></mml:math><tex-math><![CDATA[${S_{1}},{S_{2}},\dots \hspace{0.1667em}$]]></tex-math></alternatives></inline-formula> with <inline-formula id="j_nejsds55_ineq_019"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[${E_{P}}({S_{1}})=1$]]></tex-math></alternatives></inline-formula>, again chosen in advance and hence relatively simple. The value <inline-formula id="j_nejsds55_ineq_020"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${s_{t}}$]]></tex-math></alternatives></inline-formula> of <inline-formula id="j_nejsds55_ineq_021"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${S_{t}}$]]></tex-math></alternatives></inline-formula> may become known to us only when we have observed <inline-formula id="j_nejsds55_ineq_022"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{1}},\dots ,{X_{t}}$]]></tex-math></alternatives></inline-formula>. To interpret <inline-formula id="j_nejsds55_ineq_023"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${s_{t}}$]]></tex-math></alternatives></inline-formula>, we adopt this principle:</p><statement id="j_nejsds55_stat_004"><label>Principle 4</label>
<title>(Authorization to test with a test martingale).</title>
<p><italic>If we specify a test martingale</italic> <inline-formula id="j_nejsds55_ineq_024"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mspace width="0.1667em"/></mml:math><tex-math><![CDATA[${S_{1}},{S_{2}},\dots \hspace{0.1667em}$]]></tex-math></alternatives></inline-formula> <italic>in advance, then at all times t we may take</italic> <inline-formula id="j_nejsds55_ineq_025"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${s_{t}}$]]></tex-math></alternatives></inline-formula><italic>, the observed value of</italic> <inline-formula id="j_nejsds55_ineq_026"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${S_{t}}$]]></tex-math></alternatives></inline-formula><italic>, as the current measure of evidence against P. We interpret each</italic> <inline-formula id="j_nejsds55_ineq_027"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${s_{t}}$]]></tex-math></alternatives></inline-formula> <italic>(each test score) on the same scale as we use in Principles</italic> <xref rid="j_nejsds55_stat_002"><italic>2</italic></xref> <italic>and</italic> <xref rid="j_nejsds55_stat_003"><italic>3</italic></xref><italic>. In other words, when</italic> <inline-formula id="j_nejsds55_ineq_028"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi></mml:math><tex-math><![CDATA[${s_{t}}=1/\alpha $]]></tex-math></alternatives></inline-formula><italic>, it has the same force against P as the happening of a pre-specified event E with</italic> <inline-formula id="j_nejsds55_ineq_029"><alternatives><mml:math>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi></mml:math><tex-math><![CDATA[$P(E)=\alpha $]]></tex-math></alternatives></inline-formula><italic>.</italic></p></statement>
<p>Principle <xref rid="j_nejsds55_stat_003">3</xref> is the special case of Principle <xref rid="j_nejsds55_stat_004">4</xref> where all the <inline-formula id="j_nejsds55_ineq_030"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${S_{t}}$]]></tex-math></alternatives></inline-formula> for <inline-formula id="j_nejsds55_ineq_031"><alternatives><mml:math>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo stretchy="false">≥</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$t\ge 2$]]></tex-math></alternatives></inline-formula> are equal to the constant 1.</p>
<p>Like Cournot’s principle and our previous extensions of it, Principle <xref rid="j_nejsds55_stat_004">4</xref> can be considered a special case of the fundamental principle of testing by betting. For each <italic>t</italic>, <inline-formula id="j_nejsds55_ineq_032"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${s_{t}}$]]></tex-math></alternatives></inline-formula> is our capital after observing <inline-formula id="j_nejsds55_ineq_033"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{t}}$]]></tex-math></alternatives></inline-formula> if we first buy <inline-formula id="j_nejsds55_ineq_034"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${S_{1}}$]]></tex-math></alternatives></inline-formula> and then, for each time <inline-formula id="j_nejsds55_ineq_035"><alternatives><mml:math>
<mml:mi mathvariant="italic">u</mml:mi>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$u\gt 1$]]></tex-math></alternatives></inline-formula>, invest all our capital after observing <inline-formula id="j_nejsds55_ineq_036"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">u</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{u}}$]]></tex-math></alternatives></inline-formula> in <inline-formula id="j_nejsds55_ineq_037"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">u</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${S_{u+1}}$]]></tex-math></alternatives></inline-formula>. But we are still testing a mathematical object, a probability distribution <italic>P</italic>. Forecaster is following a fixed strategy, which tells him to use <italic>P</italic>’s successive conditional probabilities as forecasts, and Skeptic’s strategy (the test martingale) is also specified in advance. Neither Forecaster nor Skeptic have any discretion after the process of begins. So Principle <xref rid="j_nejsds55_stat_004">4</xref> is not a principle of optional continuation in the sense of this article. It allows optional stopping only in Doob’s sense; before the process begins, Skeptic can modify a proposed test martingale by adopting a plan for stopping.</p>
<p>Although the statement of Principle <xref rid="j_nejsds55_stat_004">4</xref> does not mention betting, I do not recall seeing the principle explained without a betting story. I do not know any other way to motivate it.</p>
</sec>
<sec id="j_nejsds55_s_008">
<label>2.6</label>
<title>Improvised Testing (Optional Continuation for Skeptic)</title>
<p>Principle <xref rid="j_nejsds55_stat_004">4</xref> authorizes the statistician to use a test martingale specified in advance. Improvisation is not yet authorized. For this, we need some further principle. As with Principle <xref rid="j_nejsds55_stat_004">4</xref>, we are testing a probability distribution <italic>P</italic> for a stochastic process <inline-formula id="j_nejsds55_ineq_038"><alternatives><mml:math>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>:</mml:mo>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mspace width="0.1667em"/></mml:math><tex-math><![CDATA[$X:={X_{1}},{X_{2}},\dots \hspace{0.1667em}$]]></tex-math></alternatives></inline-formula>, and we observe the <inline-formula id="j_nejsds55_ineq_039"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{t}}$]]></tex-math></alternatives></inline-formula> successively. When <inline-formula id="j_nejsds55_ineq_040"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${x_{1}},\dots ,{x_{t-1}}$]]></tex-math></alternatives></inline-formula> are possible values of <inline-formula id="j_nejsds55_ineq_041"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{1}},\dots ,{X_{t-1}}$]]></tex-math></alternatives></inline-formula>, we call a nonnegative variable <inline-formula id="j_nejsds55_ineq_042"><alternatives><mml:math>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$S({X_{t}})$]]></tex-math></alternatives></inline-formula> a <italic>round-t test variable given</italic> <inline-formula id="j_nejsds55_ineq_043"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${x_{1}},\dots ,{x_{t-1}}$]]></tex-math></alternatives></inline-formula> if <inline-formula id="j_nejsds55_ineq_044"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[${E_{P}}(S({X_{t}})|{x_{1}},\dots ,{x_{t-1}})=1$]]></tex-math></alternatives></inline-formula>; when <inline-formula id="j_nejsds55_ineq_045"><alternatives><mml:math>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$t=1$]]></tex-math></alternatives></inline-formula>, this reduces to <inline-formula id="j_nejsds55_ineq_046"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[${E_{P}}(S({X_{1}}))=1$]]></tex-math></alternatives></inline-formula>. We can formulate a principle for improvisation in testing as follows:</p><statement id="j_nejsds55_stat_005"><label>Principle 5</label>
<title>(Authorization to improvise when testing).</title>
<p><italic>Suppose we set</italic> <inline-formula id="j_nejsds55_ineq_047"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[${s_{0}}=1$]]></tex-math></alternatives></inline-formula><italic>, specify a round-1 test variable, say</italic> <inline-formula id="j_nejsds55_ineq_048"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${S_{1}}({X_{1}})$]]></tex-math></alternatives></inline-formula><italic>. Then, beginning with</italic> <inline-formula id="j_nejsds55_ineq_049"><alternatives><mml:math>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$t=1$]]></tex-math></alternatives></inline-formula><italic>,</italic> 
<list>
<list-item id="j_nejsds55_li_014">
<label>1.</label>
<p><italic>we observe</italic> <inline-formula id="j_nejsds55_ineq_050"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{t}}$]]></tex-math></alternatives></inline-formula><italic>’s value</italic> <inline-formula id="j_nejsds55_ineq_051"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${x_{t}}$]]></tex-math></alternatives></inline-formula><italic>,</italic></p>
</list-item>
<list-item id="j_nejsds55_li_015">
<label>2.</label>
<p><italic>we set</italic> <inline-formula id="j_nejsds55_ineq_052"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>:</mml:mo>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${s_{t}}:={s_{t-1}}{S_{t}}({x_{t}})$]]></tex-math></alternatives></inline-formula><italic>, and</italic></p>
</list-item>
<list-item id="j_nejsds55_li_016">
<label>3.</label>
<p><italic>we specify a round-</italic><inline-formula id="j_nejsds55_ineq_053"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(t+1)$]]></tex-math></alternatives></inline-formula> <italic>test variable given</italic> <inline-formula id="j_nejsds55_ineq_054"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${x_{1}},\dots ,{x_{t}}$]]></tex-math></alternatives></inline-formula><italic>, say</italic> <inline-formula id="j_nejsds55_ineq_055"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${S_{t+1}}({X_{t+1}})$]]></tex-math></alternatives></inline-formula><italic>.</italic></p>
</list-item>
</list> 
<italic>Suppose we continue so long as we still want to continue and stop whenever we please (after step 2 for some t). Then at all times t until after we stop, we may take</italic> <inline-formula id="j_nejsds55_ineq_056"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${s_{t}}$]]></tex-math></alternatives></inline-formula> <italic>as the current measure of evidence against P. We interpret</italic> <inline-formula id="j_nejsds55_ineq_057"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${s_{t}}$]]></tex-math></alternatives></inline-formula> <italic>on the same scale as we use in Principles</italic> <xref rid="j_nejsds55_stat_002"><italic>2</italic></xref><italic>,</italic> <xref rid="j_nejsds55_stat_003"><italic>3</italic></xref><italic>, and</italic> <xref rid="j_nejsds55_stat_004"><italic>4</italic></xref><italic>.</italic></p></statement>
<p>Principle <xref rid="j_nejsds55_stat_005">5</xref> generalizes Principle <xref rid="j_nejsds55_stat_004">4</xref>, and like Principle <xref rid="j_nejsds55_stat_004">4</xref>, it can be considered a special case of the fundamental principle for testing by betting. Skeptic is now a free player, not constrained to follow a strategy specified in advance.</p>
</sec>
<sec id="j_nejsds55_s_009">
<label>2.7</label>
<title>Improvised Probabilities (Optional Continuation for Forecaster)</title>
<p>Principle <xref rid="j_nejsds55_stat_005">5</xref> authorizes a statistician testing a probability distribution to improvise. But this still does not bring us to R. A. Fisher’s vision, where a statistician helps construct over time not only a test but also the probabilities being tested. In this vision, the statistician brainstorms with other scientists to design an experiment with outcome <inline-formula id="j_nejsds55_ineq_058"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{1}}$]]></tex-math></alternatives></inline-formula>, to which they assign probabilities based on some theory they want to test, and after observing <inline-formula id="j_nejsds55_ineq_059"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{1}}={x_{1}}$]]></tex-math></alternatives></inline-formula>, they brainstorm again about what they have learned and design a possibly unanticipated experiment with outcome <inline-formula id="j_nejsds55_ineq_060"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{2}}$]]></tex-math></alternatives></inline-formula>, and so on.</p>
<p>It is tempting to try to square traditional probability with Fisher’s vision by imagining that this collaboration defines a probability distribution <italic>P</italic> progressively. on the first step we define a probability distribution <inline-formula id="j_nejsds55_ineq_061"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{1}}$]]></tex-math></alternatives></inline-formula> for <inline-formula id="j_nejsds55_ineq_062"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{1}}$]]></tex-math></alternatives></inline-formula>. On the second, we define a probability distribution <inline-formula id="j_nejsds55_ineq_063"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{2}}$]]></tex-math></alternatives></inline-formula> for <inline-formula id="j_nejsds55_ineq_064"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{2}}$]]></tex-math></alternatives></inline-formula>, and so on. We are tempted to say that we are testing the product <inline-formula id="j_nejsds55_ineq_065"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>×</mml:mo>
<mml:mo stretchy="false">⋯</mml:mo>
<mml:mo>×</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{1}}\times \cdots \times {P_{k}}$]]></tex-math></alternatives></inline-formula>, where <italic>k</italic> is where the research team stops. But the statistician did not set out to test <inline-formula id="j_nejsds55_ineq_066"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>×</mml:mo>
<mml:mo stretchy="false">⋯</mml:mo>
<mml:mo>×</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{1}}\times \cdots \times {P_{k}}$]]></tex-math></alternatives></inline-formula>. She and her colleagues waited to design the second experiment and its <inline-formula id="j_nejsds55_ineq_067"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{2}}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds55_ineq_068"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{2}}$]]></tex-math></alternatives></inline-formula> until they had seen <inline-formula id="j_nejsds55_ineq_069"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${x_{1}}$]]></tex-math></alternatives></inline-formula>. Had <inline-formula id="j_nejsds55_ineq_070"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${x_{1}}$]]></tex-math></alternatives></inline-formula> come out differently, their subsequent brainstorming might have produced a different <inline-formula id="j_nejsds55_ineq_071"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{2}}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds55_ineq_072"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{2}}$]]></tex-math></alternatives></inline-formula>, and so on. If there is a single comprehensive probability distribution being tested, it is not <inline-formula id="j_nejsds55_ineq_073"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>×</mml:mo>
<mml:mo stretchy="false">⋯</mml:mo>
<mml:mo>×</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{1}}\times \cdots \times {P_{k}}$]]></tex-math></alternatives></inline-formula>; instead it must give conditional probabilities for each <inline-formula id="j_nejsds55_ineq_074"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{t}}$]]></tex-math></alternatives></inline-formula> given all the different ways the previous <italic>x</italic>s might have come out and all the different ways the research team’s information and thinking might evolve while the previous experiments were being performed and analyzed.</p>
<p>Several decades ago Philip Dawid [<xref ref-type="bibr" rid="j_nejsds55_ref_012">12</xref>, <xref ref-type="bibr" rid="j_nejsds55_ref_013">13</xref>] bravely argued that these dependencies should not matter—that we can design significance tests, confidence intervals, and Bayesian procedures that are unaffected by probabilities, somehow true or somehow invented, involving the might-have-beens. As these might-have-beens do not matter, we can just pretend that we have the desired independence. This is Dawid’s <italic>prequential</italic> model. Although some statisticians (including myself) found it appealing, others found it confusing. What are we really testing? Are we testing a huge and not fully specified probability distribution <italic>P</italic> whose unspecified probabilities include probabilities for actions of the research team doing the testing? In my experience, some mathematicians say yes, but the formulation has a complicated and paradoxical air that hardly lends itself to communication with scientists and their public.</p>
<p>Leaving aside the problem of communication, can we formulate a principle that authorizes us to use Dawid’s insight to construct test scores? Here’s a try.</p><statement id="j_nejsds55_stat_006"><label>Principle 6</label>
<title>(A prequential testing principle).</title>
<p><italic>Suppose we set</italic> <inline-formula id="j_nejsds55_ineq_075"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[${s_{0}}=1$]]></tex-math></alternatives></inline-formula><italic>, construct an experiment that will produce a variable</italic> <inline-formula id="j_nejsds55_ineq_076"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{1}}$]]></tex-math></alternatives></inline-formula><italic>, select a probability distribution</italic> <inline-formula id="j_nejsds55_ineq_077"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{1}}$]]></tex-math></alternatives></inline-formula> <italic>for</italic> <inline-formula id="j_nejsds55_ineq_078"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{1}}$]]></tex-math></alternatives></inline-formula><italic>, and select a test variable</italic> <inline-formula id="j_nejsds55_ineq_079"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${S_{1}}$]]></tex-math></alternatives></inline-formula> <italic>for</italic> <inline-formula id="j_nejsds55_ineq_080"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{1}}$]]></tex-math></alternatives></inline-formula><italic>. Then, beginning with</italic> <inline-formula id="j_nejsds55_ineq_081"><alternatives><mml:math>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$t=1$]]></tex-math></alternatives></inline-formula><italic>,</italic> 
<list>
<list-item id="j_nejsds55_li_017">
<label>1.</label>
<p><italic>we observe</italic> <inline-formula id="j_nejsds55_ineq_082"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{t}}$]]></tex-math></alternatives></inline-formula><italic>’s value</italic> <inline-formula id="j_nejsds55_ineq_083"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${x_{t}}$]]></tex-math></alternatives></inline-formula><italic>,</italic></p>
</list-item>
<list-item id="j_nejsds55_li_018">
<label>2.</label>
<p><italic>we set</italic> <inline-formula id="j_nejsds55_ineq_084"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>:</mml:mo>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${s_{t}}:={s_{t-1}}{S_{t}}({x_{t}})$]]></tex-math></alternatives></inline-formula><italic>, and</italic></p>
</list-item>
<list-item id="j_nejsds55_li_019">
<label>3.</label>
<p><italic>we construct an experiment (perhaps newly conceived) that will yield a variable</italic> <inline-formula id="j_nejsds55_ineq_085"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{t+1}}$]]></tex-math></alternatives></inline-formula><italic>, a probability distribution</italic> <inline-formula id="j_nejsds55_ineq_086"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{t+1}}$]]></tex-math></alternatives></inline-formula> <italic>for</italic> <inline-formula id="j_nejsds55_ineq_087"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{t+1}}$]]></tex-math></alternatives></inline-formula><italic>, and a test variable</italic> <inline-formula id="j_nejsds55_ineq_088"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${S_{t+1}}$]]></tex-math></alternatives></inline-formula> <italic>for</italic> <inline-formula id="j_nejsds55_ineq_089"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{t+1}}$]]></tex-math></alternatives></inline-formula><italic>.</italic></p>
</list-item>
</list> 
<italic>Suppose we continue so long as we still want to continue and stop whenever we please (after step 2 for some t). Then at all times t until after we stop, we may take</italic> <inline-formula id="j_nejsds55_ineq_090"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${s_{t}}$]]></tex-math></alternatives></inline-formula> <italic>as the current measure of evidence against the</italic> <inline-formula id="j_nejsds55_ineq_091"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{t}}$]]></tex-math></alternatives></inline-formula> <italic>we have constructed so far all being valid. We may interpret</italic> <inline-formula id="j_nejsds55_ineq_092"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${s_{t}}$]]></tex-math></alternatives></inline-formula> <italic>on the same scale as we use in Principles</italic> <xref rid="j_nejsds55_stat_002"><italic>2</italic></xref><italic>,</italic> <xref rid="j_nejsds55_stat_003"><italic>3</italic></xref><italic>,</italic> <xref rid="j_nejsds55_stat_004"><italic>4</italic></xref><italic>, and</italic> <xref rid="j_nejsds55_stat_005"><italic>5</italic></xref><italic>.</italic></p></statement>
<p>Principle <xref rid="j_nejsds55_stat_005">5</xref> is a special case of Principle <xref rid="j_nejsds55_stat_006">6</xref>. And Principle <xref rid="j_nejsds55_stat_006">6</xref>, like our preceding extensions of Cournot’s principle, can be considered a special case of the fundamental principle for testing by betting. Now both Forecaster and Skeptic are free agents, not constrained to follow any strategy specified in advance.</p>
<p>The principle’s consistency with testing in the game-theoretic framework is not surprising, as that framework was partly inspired by Dawid’s prequential model.</p>
</sec>
<sec id="j_nejsds55_s_010">
<label>2.8</label>
<title>The Role of Ville’s Inequality</title>
<p>Ville’s inequality says that if <inline-formula id="j_nejsds55_ineq_093"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mspace width="0.1667em"/></mml:math><tex-math><![CDATA[${S_{1}},{S_{2}},\dots \hspace{0.1667em}$]]></tex-math></alternatives></inline-formula> is a test martingale, then 
<disp-formula id="j_nejsds55_eq_002">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">sup</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo stretchy="false">≥</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:munder>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">≥</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ P\left(\underset{t\ge 1}{\sup }{S_{t}}\ge \frac{1}{\alpha }\right)\le \alpha \]]]></tex-math></alternatives>
</disp-formula> 
for every <inline-formula id="j_nejsds55_ineq_094"><alternatives><mml:math>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[$\alpha \gt 0$]]></tex-math></alternatives></inline-formula>. Some people (including myself) have sometimes said that Ville’s inequality authorizes optional continuation. This is a careless formulation. First because a theorem is never more than mathematics; it cannot authorize anything. Second because the principle Ville’s inequality suggests is not an optional continuation principle in the sense developed in this article. It begins with the choice of a test martingale and so does not help us understand optional continuation for Skeptic or optional continuation for Forecaster.</p>
<p>Recall that a random variable <italic>W</italic> satisfying <inline-formula id="j_nejsds55_ineq_095"><alternatives><mml:math>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">W</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi></mml:math><tex-math><![CDATA[$P(W\le \alpha )\le \alpha $]]></tex-math></alternatives></inline-formula> for all <inline-formula id="j_nejsds55_ineq_096"><alternatives><mml:math>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[$\alpha \gt 0$]]></tex-math></alternatives></inline-formula> is called a <italic>p-variable</italic>, and that a realized value of a p-variable is called a <italic>p-value</italic>.<xref ref-type="fn" rid="j_nejsds55_fn_008">8</xref><fn id="j_nejsds55_fn_008"><label><sup>8</sup></label>
<p>See [<xref ref-type="bibr" rid="j_nejsds55_ref_055">55</xref>, p. 88] for an explanation of how this definition of <italic>p-value</italic> is equivalent to the traditional definition in terms of test statistics. In many branches of theoretical and applied statistics, the distinction between <italic>p-variable</italic> and <italic>p-value</italic> is ignored; both are called p-values.</p></fn> Ville’s inequality tells us that <inline-formula id="j_nejsds55_ineq_097"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo movablelimits="false">sup</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo stretchy="false">≥</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$1/{\sup _{t\ge 1}}{S_{t}}$]]></tex-math></alternatives></inline-formula> is a p-variable, and so <inline-formula id="j_nejsds55_ineq_098"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo movablelimits="false">sup</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo stretchy="false">≥</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$1/{\sup _{t\ge 1}}{s_{t}}$]]></tex-math></alternatives></inline-formula> is a p-value. Well, almost. It is at least implicit in the notion of a p-value, as statisticians understand and use the term, that we have observed it and know we have observed it. We do not expect this for <inline-formula id="j_nejsds55_ineq_099"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo movablelimits="false">sup</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo stretchy="false">≥</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$1/{\sup _{t\ge 1}}{s_{t}}$]]></tex-math></alternatives></inline-formula>. But we do observe upper bounds. At time <italic>t</italic>, we have observed the upper bound <inline-formula id="j_nejsds55_ineq_100"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo movablelimits="false">sup</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$1/{\sup _{1\le i\le t}}{s_{i}}$]]></tex-math></alternatives></inline-formula>, and an upper bound on a p-value is a p-value. So most statisticians who use p-values would probably accept this principle:</p><statement id="j_nejsds55_stat_007"><label>Principle 7</label>
<title>(The dynamic p-value principle).</title>
<p><italic>At any time t as we observe a sequence</italic> <inline-formula id="j_nejsds55_ineq_101"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mspace width="0.1667em"/></mml:math><tex-math><![CDATA[${S_{1}},{S_{2}},\dots \hspace{0.1667em}$]]></tex-math></alternatives></inline-formula> <italic>that is a martingale under P, we may interpret</italic> <inline-formula id="j_nejsds55_ineq_102"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo movablelimits="false">sup</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$1/{\sup _{1\le i\le t}}{s_{i}}$]]></tex-math></alternatives></inline-formula> <italic>as evidence against P just as statisticians usually interpret a p-value.</italic></p></statement>
<p>Principle <xref rid="j_nejsds55_stat_007">7</xref> can be called an optional stopping principle. It is not an optional continuation principle in the sense of this article, because it does not authorize us to change the test martingale or later experiments and the probabilities for their <italic>X</italic>s.</p>
<p>Because the <inline-formula id="j_nejsds55_ineq_103"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo movablelimits="false">sup</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">u</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$1/{\sup _{1\le i\le u}}{s_{i}}$]]></tex-math></alternatives></inline-formula> for <inline-formula id="j_nejsds55_ineq_104"><alternatives><mml:math>
<mml:mi mathvariant="italic">u</mml:mi>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mi mathvariant="italic">t</mml:mi></mml:math><tex-math><![CDATA[$u\lt t$]]></tex-math></alternatives></inline-formula> are no smaller than <inline-formula id="j_nejsds55_ineq_105"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo movablelimits="false">sup</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$1/{\sup _{1\le i\le t}}{s_{i}}$]]></tex-math></alternatives></inline-formula>, these earlier <inline-formula id="j_nejsds55_ineq_106"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo movablelimits="false">sup</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">u</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$1/{\sup _{1\le i\le u}}{s_{i}}$]]></tex-math></alternatives></inline-formula> still have the force of a p-value against <italic>P</italic> when we give <inline-formula id="j_nejsds55_ineq_107"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo movablelimits="false">sup</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$1/{\sup _{1\le i\le t}}{s_{i}}$]]></tex-math></alternatives></inline-formula> the force of a p-value against <italic>P</italic>. This observation is related to the concept of a confidence sequence, which goes back to [<xref ref-type="bibr" rid="j_nejsds55_ref_011">11</xref>].</p>
<p>As an optional stopping principle for martingales, Principle <xref rid="j_nejsds55_stat_007">7</xref> can be compared with Principle <xref rid="j_nejsds55_stat_004">4</xref>. Neither is stronger than the other. Principle <xref rid="j_nejsds55_stat_004">4</xref> authorizes us to use <inline-formula id="j_nejsds55_ineq_108"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${s_{t}}$]]></tex-math></alternatives></inline-formula> as a measure of our evidence against <italic>P</italic> and to continue doing so if we stop. But it does not allow us to continue using <inline-formula id="j_nejsds55_ineq_109"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${s_{t}}$]]></tex-math></alternatives></inline-formula> if we do not stop and hence does not authorize us to use the sometimes larger <inline-formula id="j_nejsds55_ineq_110"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mo movablelimits="false">sup</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\sup _{1\le i\le t}}{s_{i}}$]]></tex-math></alternatives></inline-formula> [<xref ref-type="bibr" rid="j_nejsds55_ref_055">55</xref>]. But it gives <inline-formula id="j_nejsds55_ineq_111"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$1/{s_{t}}$]]></tex-math></alternatives></inline-formula> the force of a fixed significance level, which is greater than the force of a p-value. When we observe a p-value <italic>p</italic>, we are observing the happening of the event <inline-formula id="j_nejsds55_ineq_112"><alternatives><mml:math>
<mml:mi mathvariant="italic">W</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi></mml:math><tex-math><![CDATA[$W\le p$]]></tex-math></alternatives></inline-formula>, which has probability <italic>p</italic> or less. But we did not specify this event in advance; we only specified the p-variable <italic>W</italic>.</p>
<p>Ville’s inequality and Principle <xref rid="j_nejsds55_stat_007">7</xref> have generalizations in game-theoretic statistics, where they use game-theoretic definitions of upper and lower probability and expected value [<xref ref-type="bibr" rid="j_nejsds55_ref_054">54</xref>, Exercise 2.10].</p>
</sec>
<sec id="j_nejsds55_s_011">
<label>2.9</label>
<title>Conclusion</title>
<p>We have discussed how Cournot’s principle, the traditional principle for testing a probability distribution, can be extended so that it fully accommodates optional continuation and yet does not explicitly use game-theoretic probability or ideas about betting. Is this extension worth the trouble?</p>
<p>The clear message of the exercise is that the fundamental principle of testing by betting, coupled with game-theoretic probability, provides a theoretical basis for optional continuation that is simpler, clearer, and more general. Readers will judge for themselves, but I submit that Principle <xref rid="j_nejsds55_stat_006">6</xref> is overly complex, ill-motivated, and impossible to teach without reference to betting. It remains, moreover, less general than the fundamental principle of testing by betting, because it requires Forecaster’s moves on each round of a forecasting game to be a probability distribution.</p>
<p>Our exercise has also illustrated the new clarity brought to statistical theory by game-theoretic probability’s distinction between Forecaster and Skeptic. This distinction has helped us see the complexity of the notion of optional continuation. Optional continuation for Forecaster is a step further than optional continuation for Skeptic.</p>
</sec>
</sec>
<sec id="j_nejsds55_s_012">
<label>3</label>
<title>Game-Theoretic Descriptive Data Analysis</title>
<p>Researchers often construct statistical models that cannot be taken seriously as anything more than descriptions of their <italic>study populations</italic> — the populations for which they have data. Unfortunately, our methodology and terminology for constructing such models (estimation, significance tests, confidence intervals, credible regions, etc.) can only be understood in terms of inferences about larger populations or observations not yet made.</p>
<p>A study population is often merely a <italic>convenience sample</italic> (examples we managed to find). Sometimes it is an <italic>entire population</italic> (perhaps the eight highly industrialized nations, or the five hundred corporations in the S&amp;P 500 index). Describing such populations means summarizing — identifying general features rather than details.</p>
<p>When a study population is merely a collection of numbers, we may be able to summarize it by giving a few descriptive statistics, such as the average <inline-formula id="j_nejsds55_ineq_113"><alternatives><mml:math><mml:mover accent="false">
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mo accent="true">‾</mml:mo></mml:mover></mml:math><tex-math><![CDATA[$\overline{y}$]]></tex-math></alternatives></inline-formula> and the standard deviation <italic>s</italic>. But confusion arises as soon as we ask about the precision of these statistics. If <inline-formula id="j_nejsds55_ineq_114"><alternatives><mml:math><mml:mover accent="false">
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mo accent="true">‾</mml:mo></mml:mover>
<mml:mo>=</mml:mo>
<mml:mn>5.346</mml:mn></mml:math><tex-math><![CDATA[$\overline{y}=5.346$]]></tex-math></alternatives></inline-formula>, then is 5 just as good a description? What about 6 or 4? Maybe. If the numbers are very spread out, then saying that 0 is in the middle might be just as good as saying that 5.346 is in the middle. Our usual response to this difficulty is to replace <inline-formula id="j_nejsds55_ineq_115"><alternatives><mml:math><mml:mover accent="false">
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mo accent="true">‾</mml:mo></mml:mover></mml:math><tex-math><![CDATA[$\overline{y}$]]></tex-math></alternatives></inline-formula> with a confidence interval, such as <inline-formula id="j_nejsds55_ineq_116"><alternatives><mml:math><mml:mover accent="false">
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mo accent="true">‾</mml:mo></mml:mover>
<mml:mo>±</mml:mo>
<mml:mn>1.96</mml:mn>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msqrt>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msqrt></mml:math><tex-math><![CDATA[$\overline{y}\pm 1.96s/\sqrt{n}$]]></tex-math></alternatives></inline-formula>, but then we are pretending to make an inference to a theoretical mean or a larger population.</p>
<p>The first thesis of this section is that when our goal is merely to describe a study population, we should talk about prediction, not about inference. The average of a collection of numbers can be used to predict each number, and asking whether alternatives to the average predict the numbers in the collection as well or nearly as well does not sidetrack us into thinking about inference to some larger population. The predictions considered here are probability distributions. Following [<xref ref-type="bibr" rid="j_nejsds55_ref_054">54</xref>], I will call them <italic>probability forecasts</italic> or simply <italic>forecasts</italic>.</p>
<p>The second thesis is that betting can give us a scale for comparing probability forecasts that has intuitive meaning without recourse to inferential ideas. The key is to have the different probability distributions in a model (or different algorithms that produce probability forecasts) bet against each other. The factor by which one distribution or algorithm multiplies the money it risks betting against another is a measure of how much better it did as a forecaster, a measure that has an intuitive meaning even for those not trained in probability theory.</p>
<sec id="j_nejsds55_s_013">
<label>3.1</label>
<title>Theory</title>
<p>As explained in Section <xref rid="j_nejsds55_s_002">2</xref>, game-theoretic probability uses probability distributions as forecasting strategies. It recasts the notion of independent observations as a property a forecasting strategy might have: the strategy always makes the same forecast or uses the same forecasting rule. The question whether given observations are random with respect to a probability distribution is replaced by the question whether such a forecasting strategy withstands bets against its forecasts.</p>
<p>In addition to serving as a forecasting strategy, a probability distribution can also serve as a strategy for betting against a probability forecaster. This was explained by John L. Kelly Jr. [<xref ref-type="bibr" rid="j_nejsds55_ref_037">37</xref>], and the strategy has been called <italic>Kelly betting</italic>; see also [<xref ref-type="bibr" rid="j_nejsds55_ref_054">54</xref>, Ch. 10] and [<xref ref-type="bibr" rid="j_nejsds55_ref_050">50</xref>]. In the simplest case, Kelly betting produces a likelihood ratio as the payoff of a gamble.</p>
<p>The duality in the way probability distributions can be used — as forecasting strategies and as strategies for betting against forecasts, is key to the simple idea I am proposing here: Choose a statistical model (i.e., a collection of probability distributions), have them bet against each other’s forecasts for the study population, and take the distribution or distributions that do best in the competition as our description of the study population. The result is purely descriptive; it does not suppose that the winners will forecast well in other data, and it does not even make any claims about the model we have chosen being best for the study population. The choice may be purely conventional.</p>
<sec id="j_nejsds55_s_014">
<label>3.1.1</label>
<title>From Probability Forecasts to Description</title>
<p>Any method of description requires relatively arbitrary choices and conventions. Here we begin by choosing variables that have values for each individual in our study population: a variable <italic>Y</italic> (the <italic>target</italic> variable) and variables <inline-formula id="j_nejsds55_ineq_117"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">K</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{1}},\dots ,{X_{K}}$]]></tex-math></alternatives></inline-formula> (the <italic>forecasting</italic> variables). Inferential statistical theory has accustomed us to think of these choices as causal modeling, but when description is our goal, <italic>Y</italic> and the <inline-formula id="j_nejsds55_ineq_118"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{k}}$]]></tex-math></alternatives></inline-formula> are simply what we want to describe. For whatever reason, we want to know how they are related in the study population.</p>
<p>Next we choose a family of algorithms <inline-formula id="j_nejsds55_ineq_119"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="normal">Θ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${({P_{\theta }})_{\theta \in \Theta }}$]]></tex-math></alternatives></inline-formula>, where each algorithm <inline-formula id="j_nejsds55_ineq_120"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{\theta }}$]]></tex-math></alternatives></inline-formula> uses the <inline-formula id="j_nejsds55_ineq_121"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{k}}$]]></tex-math></alternatives></inline-formula> to give probability distributions for <italic>Y</italic>. The <inline-formula id="j_nejsds55_ineq_122"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{\theta }}$]]></tex-math></alternatives></inline-formula> are our forecasters, and the probability distributions they give are our forecasts.</p>
<p>We will want to limit the complexity of the <inline-formula id="j_nejsds55_ineq_123"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{\theta }}$]]></tex-math></alternatives></inline-formula>. In inferential statistics, simplicity is said to be a virtue because complex forecasters are likely to overfit — i.e., fail to generalize beyond the study populations. When are goal is description rather than inference, either because we are uninterested in other populations or because we cannot make assumptions that would justify inference to other populations, a more immediate virtue of simplicity is salient. Description is description only when it is simple enough to be understood.</p>
<p>Being familiar and conventional is another virtue of a descriptive forecasting family. Description requires convention, and it can only be communicated to those who know the convention.</p>
<p>Once the variables and the forecasting family are chosen, we can evaluate the <italic>θ</italic> according to their relative forecasting success within the study population. Let <inline-formula id="j_nejsds55_ineq_124"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="normal">Θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>D</mml:mtext>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\Theta _{\text{D}}}$]]></tex-math></alternatives></inline-formula> be the subset of Θ consisting of <italic>θ</italic> whose forecasts perform reasonably well in our judgment; we may call it the <italic>description range</italic>. Its elements are the <italic>descriptive forecasters of Y</italic>. Often we will be most interested in some particular aspect of the descriptive forecasters. When <inline-formula id="j_nejsds55_ineq_125"><alternatives><mml:math>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">μ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\theta =(\mu ,{\sigma ^{2}})$]]></tex-math></alternatives></inline-formula>, for example, we might be interested in <italic>μ</italic>. In other cases, we might be interested in the difference <inline-formula id="j_nejsds55_ineq_126"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>A</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>B</mml:mtext>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${Y_{\text{A}}}-{Y_{\text{B}}}$]]></tex-math></alternatives></inline-formula> when values of the <inline-formula id="j_nejsds55_ineq_127"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{k}}$]]></tex-math></alternatives></inline-formula> for two individuals A and B are given. In general, the range of <inline-formula id="j_nejsds55_ineq_128"><alternatives><mml:math>
<mml:mi mathvariant="italic">h</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$h(\theta )$]]></tex-math></alternatives></inline-formula> when <italic>θ</italic> ranges over <inline-formula id="j_nejsds55_ineq_129"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="normal">Θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>D</mml:mtext>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\Theta _{\text{D}}}$]]></tex-math></alternatives></inline-formula> is our <italic>description range</italic> for <inline-formula id="j_nejsds55_ineq_130"><alternatives><mml:math>
<mml:mi mathvariant="italic">h</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$h(\theta )$]]></tex-math></alternatives></inline-formula>.</p>
</sec>
<sec id="j_nejsds55_s_015">
<label>3.1.2</label>
<title>The Betting Competition</title>
<p>For each ordered pair <inline-formula id="j_nejsds55_ineq_131"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({\theta _{1}},{\theta _{2}})$]]></tex-math></alternatives></inline-formula> of elements of Θ, we pit <inline-formula id="j_nejsds55_ineq_132"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{2}}$]]></tex-math></alternatives></inline-formula> bet against <inline-formula id="j_nejsds55_ineq_133"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{1}}$]]></tex-math></alternatives></inline-formula>. Forecaster predicts <italic>Y</italic> with the distribution <inline-formula id="j_nejsds55_ineq_134"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{{\theta _{1}}}}$]]></tex-math></alternatives></inline-formula>; Skeptic predicts <italic>Y</italic> with the distribution <inline-formula id="j_nejsds55_ineq_135"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{{\theta _{2}}}}$]]></tex-math></alternatives></inline-formula>. Skeptic has unit capital. Forecaster offers to sell Skeptic any non-negative payoff <inline-formula id="j_nejsds55_ineq_136"><alternatives><mml:math>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$S(Y)$]]></tex-math></alternatives></inline-formula> for <inline-formula id="j_nejsds55_ineq_137"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${E_{{\theta _{1}}}}(S)$]]></tex-math></alternatives></inline-formula>, his expected value for <italic>S</italic>. In deciding what nonnegative payoff <italic>S</italic> to buy, Skeptic can use his distribution <inline-formula id="j_nejsds55_ineq_138"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${P_{{\theta _{2}}}}$]]></tex-math></alternatives></inline-formula> in different ways, but when he uses Kelly betting he buys the payoff <italic>S</italic> given by 
<disp-formula id="j_nejsds55_eq_003">
<label>(3.1)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>:</mml:mo>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ S(Y):={P_{{\theta _{2}}}}(Y)/{P_{{\theta _{1}}}}(Y)\]]]></tex-math></alternatives>
</disp-formula> 
at the price Forecaster requires: 
<disp-formula id="j_nejsds55_eq_004">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="2.5pt"/>
<mml:mtext>which is equal to</mml:mtext>
<mml:mspace width="2.5pt"/>
<mml:mn>1</mml:mn>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {E_{{\theta _{1}}}}\left(\frac{{P_{{\theta _{2}}}}(Y)}{{P_{{\theta _{1}}}}(Y)}\right),\hspace{2.5pt}\text{which is equal to}\hspace{2.5pt}1.\]]]></tex-math></alternatives>
</disp-formula> 
This bet turns Skeptic’s initial unit capital into <inline-formula id="j_nejsds55_ineq_139"><alternatives><mml:math>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$S(y)={P_{{\theta _{2}}}}(y)/{P_{{\theta _{1}}}}(y)$]]></tex-math></alternatives></inline-formula>. This ratio, the betting score, provides a measure of <inline-formula id="j_nejsds55_ineq_140"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{2}}$]]></tex-math></alternatives></inline-formula>’s forecasting success relative to <inline-formula id="j_nejsds55_ineq_141"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{1}}$]]></tex-math></alternatives></inline-formula>.</p>
<p>For our descriptive purposes, Kelly betting can be considered one more convention. It does, however, have well known optimization properties for Skeptic when Skeptic is confident in <italic>Q</italic> as a forecast.<xref ref-type="fn" rid="j_nejsds55_fn_009">9</xref><fn id="j_nejsds55_fn_009"><label><sup>9</sup></label>
<p>See [<xref ref-type="bibr" rid="j_nejsds55_ref_008">8</xref>]. For more on Kelly betting, see [<xref ref-type="bibr" rid="j_nejsds55_ref_018">18</xref>, Chapter 10] and [<xref ref-type="bibr" rid="j_nejsds55_ref_060">60</xref>]. For other roles Kelly betting can play in statistical theory, see [<xref ref-type="bibr" rid="j_nejsds55_ref_050">50</xref>] and [<xref ref-type="bibr" rid="j_nejsds55_ref_058">58</xref>].</p></fn> An important alternative to Kelly betting is fractional Kelly betting, which risks only a fraction of one’s capital. Being more cautious, this penalizes Skeptic less when <inline-formula id="j_nejsds55_ineq_142"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${P_{{\theta _{2}}}}(y)/{P_{{\theta _{1}}}}(y)$]]></tex-math></alternatives></inline-formula> is low. This makes the competition between forecasters less sensitive to individuals that are unusual with respect to the whole study population.</p>
<p>The ratio <inline-formula id="j_nejsds55_ineq_143"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${P_{{\theta _{2}}}}(y)/{P_{{\theta _{1}}}}(y)$]]></tex-math></alternatives></inline-formula> is familiar to statisticians under the name <italic>likelihood ratio</italic>. We will usually chose our forecasting family so that a unique maximum of <inline-formula id="j_nejsds55_ineq_144"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${P_{\theta }}(y)$]]></tex-math></alternatives></inline-formula> always exists. The value of <italic>θ</italic> that achieves the maximum, say <inline-formula id="j_nejsds55_ineq_145"><alternatives><mml:math><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover></mml:math><tex-math><![CDATA[$\hat{\theta }$]]></tex-math></alternatives></inline-formula>, is the <italic>maximum-likelihood estimate</italic> when <italic>y</italic> is observed, and the likelihood ratio 
<disp-formula id="j_nejsds55_eq_005">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ L(\theta )=\frac{{P_{\theta }}(y)}{{P_{\hat{\theta }}}(y)}\]]]></tex-math></alternatives>
</disp-formula> 
is a number between zero and one that measures <italic>θ</italic>’s forecasting performance.</p>
<p>Using cutoffs suggested by R. A. Fisher in 1956,<xref ref-type="fn" rid="j_nejsds55_fn_010">10</xref><fn id="j_nejsds55_fn_010"><label><sup>10</sup></label>
<p>See [<xref ref-type="bibr" rid="j_nejsds55_ref_022">22</xref>, p. 71] or page 75 of the posthumous third edition (1973).</p></fn> we may classify the performance of the <italic>θ</italic> according to their value of <inline-formula id="j_nejsds55_ineq_146"><alternatives><mml:math>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$L(\theta )$]]></tex-math></alternatives></inline-formula> as follows: 
<disp-formula id="j_nejsds55_eq_006">
<label>(3.2)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mtable columnspacing="10.0pt" equalrows="false" columnlines="none" equalcolumns="false" columnalign="left center">
<mml:mtr>
<mml:mtd class="array">
<mml:mtext>Very good</mml:mtext>
</mml:mtd>
<mml:mtd class="array">
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">≥</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>2</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="array">
<mml:mtext>Good</mml:mtext>
</mml:mtd>
<mml:mtd class="array">
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">≥</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>5</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="array">
<mml:mtext>Satisfactory</mml:mtext>
</mml:mtd>
<mml:mtd class="array">
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">≥</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>15</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="array">
<mml:mtext>Unsatisfactory</mml:mtext>
</mml:mtd>
<mml:mtd class="array">
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>15</mml:mn>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \begin{array}{l@{\hskip10.0pt}c}\text{Very good}& L(\theta )\ge 1/2\\ {} \text{Good}& 1/2\gt L(\theta )\ge 1/5\\ {} \text{Satisfactory}& 1/5\gt L(\theta )\ge 1/15\\ {} \text{Unsatisfactory}& 1/15\gt L(\theta )\end{array}\]]]></tex-math></alternatives>
</disp-formula> 
The names in (<xref rid="j_nejsds55_eq_006">3.2</xref>) are my suggestions; Fisher did not provide names for the intervals. Whatever names and cutoffs we choose will be arbitrary conventions, but no more arbitrary than the terminology and the cutoffs <inline-formula id="j_nejsds55_ineq_147"><alternatives><mml:math>
<mml:mn>5</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$5\% $]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds55_ineq_148"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$1\% $]]></tex-math></alternatives></inline-formula> used for statistical inference. If equally accepted as conventions, they can be equally serviceable. The meaning in terms of betting will be readily understood by the public, including those not trained in mathematical probability.</p>
<p>For Fisher, the intervals were inferences. A number of other authors, beginning with A. W. F. Edwards [<xref ref-type="bibr" rid="j_nejsds55_ref_017">17</xref>] and Richard Royall [<xref ref-type="bibr" rid="j_nejsds55_ref_045">45</xref>], have adopted Fisher’s proposal to use <inline-formula id="j_nejsds55_ineq_149"><alternatives><mml:math>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$L(\theta )$]]></tex-math></alternatives></inline-formula> for inference. These authors have developed methods for computing, tabulating, and displaying <inline-formula id="j_nejsds55_ineq_150"><alternatives><mml:math>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$L(\theta )$]]></tex-math></alternatives></inline-formula>, intervals of its values, and other summaries for a large variety of models. I will not attempt to review this very extensive work, but it is obviously an asset for the descriptive proposal I am making here.</p>
<p>Fisher called <inline-formula id="j_nejsds55_ineq_151"><alternatives><mml:math>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$L(\theta )$]]></tex-math></alternatives></inline-formula> the <italic>likelihood</italic> of <italic>θ</italic>. This name has endured for a century, and no mathematical statistician will be able to put it out of mind while reading the rest of this paper. Yet I will avoid it as much as possible, because its inferential connotation cannot be circumvented. On the other hand, I am using the familiar notation: <inline-formula id="j_nejsds55_ineq_152"><alternatives><mml:math><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover></mml:math><tex-math><![CDATA[$\hat{\theta }$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_nejsds55_ineq_153"><alternatives><mml:math>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$L(\theta )$]]></tex-math></alternatives></inline-formula>, and also <inline-formula id="j_nejsds55_ineq_154"><alternatives><mml:math>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$l(\theta )$]]></tex-math></alternatives></inline-formula> for <inline-formula id="j_nejsds55_ineq_155"><alternatives><mml:math>
<mml:mo movablelimits="false">ln</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\ln (L(\theta ))$]]></tex-math></alternatives></inline-formula>.</p>
</sec>
</sec>
<sec id="j_nejsds55_s_016">
<label>3.2</label>
<title>Examples</title>
<p>I offer three very simple examples. The first is purely formal and about as simple as possible: the forecast is a single probability, always the same probability. The second involves a classic convenience sample. The third, a fictional instantiation of a problem of current interest, involves an entire population.</p>
<sec id="j_nejsds55_s_017">
<label>3.2.1</label>
<title>Forecasting with a Single Probability</title>
<p>Suppose we observe successive trials of an event, and each algorithm in our forecasting family has a fixed probability that it uses each time as its forecast. Formally, <inline-formula id="j_nejsds55_ineq_156"><alternatives><mml:math>
<mml:mi mathvariant="normal">Θ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$\Theta =[0,1]$]]></tex-math></alternatives></inline-formula>, and Forecaster <italic>θ</italic> always gives <italic>θ</italic> as its forecast.</p>
<p>If we observe 100 trials, and the event happens 70 times, then <inline-formula id="j_nejsds55_ineq_157"><alternatives><mml:math><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
<mml:mo>=</mml:mo>
<mml:mn>0.7</mml:mn></mml:math><tex-math><![CDATA[$\hat{\theta }=0.7$]]></tex-math></alternatives></inline-formula>, and 
<disp-formula id="j_nejsds55_eq_007">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>:</mml:mo>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0.7</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>70</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0.3</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>30</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ L(\theta ):={\left(\frac{\theta }{0.7}\right)^{70}}{\left(\frac{1-\theta }{0.3}\right)^{30}}.\]]]></tex-math></alternatives>
</disp-formula> 
Our scheme for rating the forecasters yields these approximate description ranges: 
<table-wrap id="j_nejsds55_tab_001">
<table>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">Very good</td>
<td style="vertical-align: top; text-align: center"><inline-formula id="j_nejsds55_ineq_158"><alternatives><mml:math>
<mml:mn>0.64</mml:mn>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mn>0.76</mml:mn></mml:math><tex-math><![CDATA[$0.64\lt \theta \lt 0.76$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Good</td>
<td style="vertical-align: top; text-align: center"><inline-formula id="j_nejsds55_ineq_159"><alternatives><mml:math>
<mml:mn>0.61</mml:mn>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mn>0.78</mml:mn></mml:math><tex-math><![CDATA[$0.61\lt \theta \lt 0.78$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Satisfactory</td>
<td style="vertical-align: top; text-align: center"><inline-formula id="j_nejsds55_ineq_160"><alternatives><mml:math>
<mml:mn>0.59</mml:mn>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mn>0.80</mml:mn></mml:math><tex-math><![CDATA[$0.59\lt \theta \lt 0.80$]]></tex-math></alternatives></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap> 
The forecaster <inline-formula id="j_nejsds55_ineq_161"><alternatives><mml:math>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$\theta =1/2$]]></tex-math></alternatives></inline-formula> may have been of particular interest, and we may want to emphasize that its performance was unsatisfactory.</p>
<p>Not surprisingly, Fisher’s categories are roughly consistent with inferential practice. The standard error of the maximum-likelihood estimate 0.7 is 0.046, suggesting a <inline-formula id="j_nejsds55_ineq_162"><alternatives><mml:math>
<mml:mn>95</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$95\% $]]></tex-math></alternatives></inline-formula> confidence interval of <inline-formula id="j_nejsds55_ineq_163"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0.61</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.79</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(0.61,0.79)$]]></tex-math></alternatives></inline-formula>. Like this confidence interval, our description ranges do not contain <inline-formula id="j_nejsds55_ineq_164"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$1/2$]]></tex-math></alternatives></inline-formula>. But unlike the confidence interval, the description ranges merely describe; they merely tell us which constant forecasts perform relatively well in the data. This does not involve attribution of independence in any sense to the trials themselves.</p>
</sec>
<sec id="j_nejsds55_s_018">
<label>3.2.2</label>
<title>Fourier’s Masculine Generation</title>
<p>The calculation of error probabilities from statistical data was first made practical by Laplace’s central limit theorem, and the calculation was explained to statisticians by Joseph Fourier (1768–1830). Fourier had been an impassioned participant in the French revolution and an administrator under Napoleon. After the royalists regained power, a prominent royalist who had been Fourier’s student, Chabrol de Volvic, rescued him from impoverishment with an appointment to the Paris statistics bureau. This assignment left him time to perfect the theory of heat diffusion for which he is best known, but as part of his work at the statistics bureau, he published two marvelously clear essays on the use of probability in statistics, in 1826 and 1829. According to Bernard Bru, Marie-France Bru, and Olivier Bienaymé, these were the only works on mathematical probability read by statisticians in the early 19th century.<xref ref-type="fn" rid="j_nejsds55_fn_011">11</xref><fn id="j_nejsds55_fn_011"><label><sup>11</sup></label>
<p>See [<xref ref-type="bibr" rid="j_nejsds55_ref_009">9</xref>, p. 198]. The annual reports issued by the bureau during Fourier’s tenure list no editor on their cover pages. Fourier was no doubt primarily responsible for editing them, and I identify him as the editor in the references [<xref ref-type="bibr" rid="j_nejsds55_ref_023">23</xref>, <xref ref-type="bibr" rid="j_nejsds55_ref_024">24</xref>].</p></fn></p>
<p>To illustrate Laplace’s asymptotic theory, Fourier studied data on births and marriages gleaned from 18th-century parish and governmental records in Paris. He was particularly interested in the length of a masculine generation — the average time, for fathers of sons, from the father’s birth to the birth of his first son. On the basis of 505 cases, he estimated this average time to be 33.31 years. In his bureau’s report for 1829 [<xref ref-type="bibr" rid="j_nejsds55_ref_024">24</xref>, Table 64, p. 143ff], he gave the bounds on the estimate’s error shown in Table <xref rid="j_nejsds55_tab_002">1</xref>.</p>
<table-wrap id="j_nejsds55_tab_002">
<label>Table 1</label>
<caption>
<p>Fourier’s error probabilities.</p>
</caption>
<table>
<tbody>
<tr>
<td style="vertical-align: top; text-align: center; border-top: double">probability</td>
<td style="vertical-align: top; text-align: center; border-top: double">1/2</td>
<td style="vertical-align: top; text-align: center; border-top: double">1/20</td>
<td style="vertical-align: top; text-align: center; border-top: double">1/200</td>
<td style="vertical-align: top; text-align: center; border-top: double">1/2000</td>
<td style="vertical-align: top; text-align: center; border-top: double">1/20000</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center">error</td>
<td style="vertical-align: top; text-align: center">±2.7528</td>
<td style="vertical-align: top; text-align: center">±7.9932</td>
<td style="vertical-align: top; text-align: center">±11.4516</td>
<td style="vertical-align: top; text-align: center">±14.2044</td>
<td style="vertical-align: top; text-align: center">±16.5480</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center">level</td>
<td style="vertical-align: top; text-align: center">50%</td>
<td style="vertical-align: top; text-align: center">95%</td>
<td style="vertical-align: top; text-align: center">99.5%</td>
<td style="vertical-align: top; text-align: center">99.95%</td>
<td style="vertical-align: top; text-align: center">99.995%</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin">interval</td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin">(33.1, 33.5)</td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin">(32.6, 34.0)</td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin">(32.4, 34.3)</td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin">(32.1, 34.5)</td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin">(31.9, 34.7)</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>The first two lines give the probabilities for errors <italic>in months</italic> that Fourier calculated for his estimate of 33.31 <italic>years</italic>. He gave <inline-formula id="j_nejsds55_ineq_165"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>20</mml:mn></mml:math><tex-math><![CDATA[$1/20$]]></tex-math></alternatives></inline-formula>, for example, as the probability of the estimate erring by more than 11.4516 months. The second two lines translate this information into confidence intervals in years. The 95% interval, for example, is 32.6 to 34.0 years.</p>
</table-wrap-foot>
</table-wrap>
<p>Laplace’s theory is applicable, of course, only if the 505 cases are a random sample, and they are not. They are a convenience sample, consisting of 18th-century fathers of sons in Paris for which the needed parish records could be found. This convenience sample is of interest, but Fourier’s inferential analysis of it is unjustified. A descriptive analysis is needed.</p>
<p>For description, we do not need Fourier’s implicit assumption that the 505 cases were a random sample. We need a conventional forecasting family. Let us use the most conventional family, the normal family with mean <italic>μ</italic> and variance <inline-formula id="j_nejsds55_ineq_166"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${\sigma ^{2}}$]]></tex-math></alternatives></inline-formula>. Here <inline-formula id="j_nejsds55_ineq_167"><alternatives><mml:math>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">μ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\theta =(\mu ,{\sigma ^{2}})$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds55_ineq_168"><alternatives><mml:math><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
<mml:mo>=</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo><mml:mover accent="false">
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mo accent="true">‾</mml:mo></mml:mover>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>33.31</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>7.642</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\hat{\theta }=(\overline{y},s)=(33.31,7.642)$]]></tex-math></alternatives></inline-formula>, where <inline-formula id="j_nejsds55_ineq_169"><alternatives><mml:math><mml:mover accent="false">
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mo accent="true">‾</mml:mo></mml:mover></mml:math><tex-math><![CDATA[$\overline{y}$]]></tex-math></alternatives></inline-formula> is the average of the 505 ages, and <italic>s</italic> is their standard deviation. Fourier did not report the data, but he reported the average <inline-formula id="j_nejsds55_ineq_170"><alternatives><mml:math><mml:mover accent="false">
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mo accent="true">‾</mml:mo></mml:mover>
<mml:mo>=</mml:mo>
<mml:mn>33.31</mml:mn></mml:math><tex-math><![CDATA[$\overline{y}=33.31$]]></tex-math></alternatives></inline-formula> years, and we can calculate the standard deviation <inline-formula id="j_nejsds55_ineq_171"><alternatives><mml:math>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>7.642</mml:mn></mml:math><tex-math><![CDATA[$s=7.642$]]></tex-math></alternatives></inline-formula> years from the error probabilities he reported.</p>
<p>Writing <inline-formula id="j_nejsds55_ineq_172"><alternatives><mml:math>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$l(\theta )$]]></tex-math></alternatives></inline-formula> for <inline-formula id="j_nejsds55_ineq_173"><alternatives><mml:math>
<mml:mo movablelimits="false">ln</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\ln (L(\theta ))$]]></tex-math></alternatives></inline-formula>, we have 
<disp-formula id="j_nejsds55_eq_008">
<label>(3.3)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">μ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mo movablelimits="false">ln</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>−</mml:mo>
<mml:mo movablelimits="false">ln</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">σ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>−</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>+</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo><mml:mover accent="false">
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mo accent="true">‾</mml:mo></mml:mover>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">μ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>+</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ l(\mu ,{\sigma ^{2}})=n\left(\ln (s)-\ln (\sigma )-\frac{{s^{2}}+{(\overline{y}-\mu )^{2}}}{2{\sigma ^{2}}}+\frac{1}{2}\right),\]]]></tex-math></alternatives>
</disp-formula> 
where <inline-formula id="j_nejsds55_ineq_174"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>505</mml:mn></mml:math><tex-math><![CDATA[$n=505$]]></tex-math></alternatives></inline-formula>. The values of <inline-formula id="j_nejsds55_ineq_175"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">μ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\mu ,{\sigma ^{2}})$]]></tex-math></alternatives></inline-formula> for which (<xref rid="j_nejsds55_eq_008">3.3</xref>) exceeds <inline-formula id="j_nejsds55_ineq_176"><alternatives><mml:math>
<mml:mo movablelimits="false">ln</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\ln (1/2)$]]></tex-math></alternatives></inline-formula> constitute the very good description range for <inline-formula id="j_nejsds55_ineq_177"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">μ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\mu ,{\sigma ^{2}})$]]></tex-math></alternatives></inline-formula>, those for which it exceeds <inline-formula id="j_nejsds55_ineq_178"><alternatives><mml:math>
<mml:mo movablelimits="false">ln</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\ln (1/5)$]]></tex-math></alternatives></inline-formula> the good range, those for which it exceeds <inline-formula id="j_nejsds55_ineq_179"><alternatives><mml:math>
<mml:mo movablelimits="false">ln</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>15</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\ln (1/15)$]]></tex-math></alternatives></inline-formula> the satisfactory range.</p>
<p>Following Fourier, we are interested only in description ranges for <italic>μ</italic>. So our question is what values of <italic>μ</italic> are included in these three description ranges. So we maximize (<xref rid="j_nejsds55_eq_008">3.3</xref>) for each <italic>μ</italic>, by setting <inline-formula id="j_nejsds55_ineq_180"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${\sigma ^{2}}$]]></tex-math></alternatives></inline-formula> equal to <inline-formula id="j_nejsds55_ineq_181"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>+</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo><mml:mover accent="false">
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mo accent="true">‾</mml:mo></mml:mover>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">μ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${s^{2}}+{(\overline{y}-\mu )^{2}}$]]></tex-math></alternatives></inline-formula>. This gives 
<disp-formula id="j_nejsds55_eq_009">
<label>(3.4)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">μ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>+</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo><mml:mover accent="false">
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mo accent="true">‾</mml:mo></mml:mover>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">μ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mo movablelimits="false">ln</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>−</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo movablelimits="false">ln</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>+</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo><mml:mover accent="false">
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mo accent="true">‾</mml:mo></mml:mover>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">μ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ l(\mu ,{s^{2}}+{(\overline{y}-\mu )^{2}})=n\left(\ln (s)-\frac{1}{2}\ln ({s^{2}}+{(\overline{y}-\mu )^{2}})\right).\]]]></tex-math></alternatives>
</disp-formula> 
This is greater than <inline-formula id="j_nejsds55_ineq_182"><alternatives><mml:math>
<mml:mo movablelimits="false">ln</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">C</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\ln (1/C)$]]></tex-math></alternatives></inline-formula> when <italic>μ</italic> is in the interval 
<disp-formula id="j_nejsds55_eq_010">
<label>(3.5)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mover accent="false">
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mo accent="true">‾</mml:mo></mml:mover>
<mml:mo>±</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:msqrt>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msqrt>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \overline{y}\pm s\sqrt{{C^{2/n}}-1}.\]]]></tex-math></alternatives>
</disp-formula> 
Here we have obtained the log-likelihood (<xref rid="j_nejsds55_eq_009">3.4</xref>) for the parameter of interest <italic>μ</italic> by maximizing over the unwanted parameter <inline-formula id="j_nejsds55_ineq_183"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${\sigma ^{2}}$]]></tex-math></alternatives></inline-formula>. In inferential likelihood theory, the result of such a maximization is sometimes called a <italic>profile likelihood</italic>. The inferential use of profile likelihoods is hard to justify and sometimes misleading [<xref ref-type="bibr" rid="j_nejsds55_ref_045">45</xref>, p. 159], but their game-theoretic descriptive use is unobjectionable. We want to know how well each value of <italic>μ</italic> can perform in the betting competition. Its performance depends on the value of <inline-formula id="j_nejsds55_ineq_184"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${\sigma ^{2}}$]]></tex-math></alternatives></inline-formula> it is paired with. We pair it with the value of <inline-formula id="j_nejsds55_ineq_185"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${\sigma ^{2}}$]]></tex-math></alternatives></inline-formula> that enables it to do its best to see how well it can perform.</p>
<table-wrap id="j_nejsds55_tab_003">
<label>Table 2</label>
<caption>
<p>Description ranges for the masculine generation in Fourier’s study population, calculated from (<xref rid="j_nejsds55_eq_010">3.5</xref>).</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: center; border-top: double; border-bottom: solid thin"><italic>C</italic></td>
<td style="vertical-align: top; text-align: left; border-top: double; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: center; border-top: double; border-bottom: solid thin">description range</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: center">2</td>
<td style="vertical-align: top; text-align: left">Very good</td>
<td style="vertical-align: top; text-align: center"><inline-formula id="j_nejsds55_ineq_186"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>32.9</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>33.7</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(32.9,33.7)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center">5</td>
<td style="vertical-align: top; text-align: left">Good</td>
<td style="vertical-align: top; text-align: center"><inline-formula id="j_nejsds55_ineq_187"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>32.7</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>33.9</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(32.7,33.9)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin">15</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">Satisfactory</td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin"><inline-formula id="j_nejsds55_ineq_188"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>32.5</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>34.1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(32.5,34.1)$]]></tex-math></alternatives></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Table <xref rid="j_nejsds55_tab_003">2</xref> uses (<xref rid="j_nejsds55_eq_010">3.5</xref>) to calculate description ranges. Comparing it with Table <xref rid="j_nejsds55_tab_002">1</xref>, we see that the width of Fourier’s 95% confidence interval is between that of our good description range and our satisfactory description range. The relation as <italic>n</italic> increases between confidence intervals and the description ranges given by (<xref rid="j_nejsds55_eq_010">3.5</xref>) has been studied by [<xref ref-type="bibr" rid="j_nejsds55_ref_010">10</xref>] and [<xref ref-type="bibr" rid="j_nejsds55_ref_042">42</xref>].</p>
</sec>
<sec id="j_nejsds55_s_019">
<label>3.2.3</label>
<title>A Fictional Survey of Perceptions</title>
<p>Some organizations in the United States have recently surveyed their employees about perceptions of discrimination. To avoid the complexities involved in real examples, consider the following fictional example.</p>
<p>An organization wants to know whether its employees of different genders and racial identities differ systematically in their perception of discrimination. Most of the employees respond to a survey asking whether they have suffered discrimination because of their gender or race. The employees saying yes are distributed as shown in Table <xref rid="j_nejsds55_tab_004">3</xref>.</p>
<p>According to the usual test for the difference between two proportions, the difference between the rows (male vs female) and the difference between the columns (BIPOC vs White) are both statistically significant at the 5% level. But the 20 percentage-point difference between BIPOC males and BIPOC females is not, as its standard error is 
<disp-formula id="j_nejsds55_eq_011">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msqrt>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>20</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>+</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:msqrt>
<mml:mo stretchy="false">≈</mml:mo>
<mml:mn>0.18</mml:mn>
<mml:mo>=</mml:mo>
<mml:mn>18</mml:mn>
<mml:mspace width="2.5pt"/>
<mml:mtext>percentage points</mml:mtext>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \sqrt{\frac{1}{3}\frac{2}{3}\left(\frac{1}{20}+\frac{1}{10}\right)}\approx 0.18=18\hspace{2.5pt}\text{percentage points}.\]]]></tex-math></alternatives>
</disp-formula> 
These simple significance tests seem informative. The differences declared statistically significant seem general enough to be regarded as features of the organization, but we hesitate to say this about the difference declared not statistically significant.</p>
<p>Yet the theory of significance testing does not fit the occasion. Have the individuals in the study (or their responses to the survey) been chosen at random from some larger population? Certainly not. For anyone who has been inside an organization long enough to see its employees come and go, seeing or guessing the reasons, the idea that they constitute a random sample is phantasmagoria. Nor can we agree that their responses are independent with respect to some “data-generating mechanism”. Many of them see the same media and talk with each other.</p>
<p>If we took the theory of significance testing seriously for this example, we would also worry about multiple testing. The 5% error rate we claim for our tests is valid under the theory’s assumptions only when we make just a single comparison. We have made three comparisons and might make more.</p>
<p>The theory’s assumptions are not met, and we have abused the theory. But there is a larger issue. The theory is irrelevant from the outset, because its goals are irrelevant. The organization did not undertake the survey in order to make inferences about a larger or a different population or about some data-generating mechanism. The organization wanted only to know about itself. It wanted to know how its employees’ perceptions vary with gender and race. This calls for a descriptive analysis.</p>
<table-wrap id="j_nejsds55_tab_004">
<label>Table 3</label>
<caption>
<p>A fictional study.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: center; border-top: double; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: center; border-top: double; border-bottom: solid thin">Female</td>
<td style="vertical-align: top; text-align: center; border-top: double; border-bottom: solid thin">Male</td>
<td style="vertical-align: top; text-align: center; border-top: double; border-bottom: solid thin">Totals</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: center">BIPOC</td>
<td style="vertical-align: top; text-align: center"><inline-formula id="j_nejsds55_ineq_189"><alternatives><mml:math><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mn>8</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mpadded width="0pt">
<mml:mphantom>
<mml:mi mathvariant="italic">A</mml:mi></mml:mphantom></mml:mpadded>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>=</mml:mo>
<mml:mn>80</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$\frac{{8^{\phantom{A}}}}{10}=80\% $]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: center"><inline-formula id="j_nejsds55_ineq_190"><alternatives><mml:math><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>12</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>20</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>=</mml:mo>
<mml:mn>60</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$\frac{12}{20}=60\% $]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: center"><inline-formula id="j_nejsds55_ineq_191"><alternatives><mml:math><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>20</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>30</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo stretchy="false">≈</mml:mo>
<mml:mn>67</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$\frac{20}{30}\approx 67\% $]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center">White</td>
<td style="vertical-align: top; text-align: center"><inline-formula id="j_nejsds55_ineq_192"><alternatives><mml:math><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>20</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>50</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>=</mml:mo>
<mml:mn>40</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$\frac{20}{50}=40\% $]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: center"><inline-formula id="j_nejsds55_ineq_193"><alternatives><mml:math><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>20</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>120</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo stretchy="false">≈</mml:mo>
<mml:mn>17</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$\frac{20}{120}\approx 17\% $]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: center"><inline-formula id="j_nejsds55_ineq_194"><alternatives><mml:math><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>40</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>170</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo stretchy="false">≈</mml:mo>
<mml:mn>24</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$\frac{40}{170}\approx 24\% $]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin">Totals</td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin"><inline-formula id="j_nejsds55_ineq_195"><alternatives><mml:math><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>28</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>60</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo stretchy="false">≈</mml:mo>
<mml:mn>47</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$\frac{28}{60}\approx 47\% $]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin"><inline-formula id="j_nejsds55_ineq_196"><alternatives><mml:math><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>32</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>140</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo stretchy="false">≈</mml:mo>
<mml:mn>23</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$\frac{32}{140}\approx 23\% $]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin"><inline-formula id="j_nejsds55_ineq_197"><alternatives><mml:math><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>60</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>200</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>=</mml:mo>
<mml:mn>30</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$\frac{60}{200}=30\% $]]></tex-math></alternatives></inline-formula></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Numbers and proportions of positive responses, in a fictional study of the employees of a fictional organization, to the question whether one has experienced discrimination in the organization as the result of one’s identity. Here BIPOC means Black, indigenous, and people of color.</p>
</table-wrap-foot>
</table-wrap>
<p>For a descriptive analysis, we can use the obvious forecasting family, in which each cell in the <inline-formula id="j_nejsds55_ineq_198"><alternatives><mml:math>
<mml:mn>2</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$2\times 2$]]></tex-math></alternatives></inline-formula> table has its own forecast: 
<disp-formula id="j_nejsds55_eq_012">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>bf</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>bm</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>wf</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>wm</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \theta =({\theta _{\text{bf}}},{\theta _{\text{bm}}},{\theta _{\text{wf}}},{\theta _{\text{wm}}}),\]]]></tex-math></alternatives>
</disp-formula> 
where <inline-formula id="j_nejsds55_ineq_199"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>bf</mml:mtext>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{\text{bf}}}$]]></tex-math></alternatives></inline-formula> is the forecast that a BIPOC female will say yes to the survey, etc. According to the data in Table <xref rid="j_nejsds55_tab_004">3</xref>, 
<disp-formula id="j_nejsds55_eq_013">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
<mml:mo>=</mml:mo>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>8</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>12</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>20</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>20</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>50</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>20</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>120</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \hat{\theta }=\left(\frac{8}{10},\frac{12}{20},\frac{20}{50},\frac{20}{120}\right),\]]]></tex-math></alternatives>
</disp-formula> 
and 
<disp-formula id="j_nejsds55_eq_014">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
</mml:mtd>
<mml:mtd class="align-even">
<mml:msup>
<mml:mrow>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>bf</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mn>4</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>8</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>bf</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>bm</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>12</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>bm</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>8</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:msup>
<mml:mrow>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>wf</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>20</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>wf</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>30</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>wm</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>6</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>20</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>wm</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>5</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>6</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>100</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mspace width="-0.1667em"/>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}L(\theta )=& {\left(\frac{{\theta _{\text{bf}}}}{4/5}\right)^{8}}{\left(\frac{(1-{\theta _{\text{bf}}})}{1/5}\right)^{2}}{\left(\frac{{\theta _{\text{bm}}}}{3/5}\right)^{12}}{\left(\frac{(1-{\theta _{\text{bm}}})}{2/5}\right)^{8}}\\ {} & {\left(\frac{{\theta _{\text{wf}}}}{2/5}\right)^{20}}{\left(\frac{(1-{\theta _{\text{wf}}})}{3/5}\right)^{30}}{\left(\frac{{\theta _{\text{wm}}}}{1/6}\right)^{20}}{\left(\frac{(1-{\theta _{\text{wm}}})}{5/6}\right)^{100}}\hspace{-0.1667em}.\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>We found earlier that the 20 percentage-point difference between BIPOC males and BIPOC females is not statistically significant. In this descriptive analysis, the question can be reframed this way: what differences between BIPOC males and BIPOC females within the study population are forecast by very good forecasters? We can answer the question by looking at all the <inline-formula id="j_nejsds55_ineq_200"><alternatives><mml:math>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>bf</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>bm</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>wf</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>wm</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\theta =({\theta _{\text{bf}}},{\theta _{\text{bm}}},{\theta _{\text{wf}}},{\theta _{\text{wm}}})$]]></tex-math></alternatives></inline-formula> that rank as very good forecasters by having a value of <inline-formula id="j_nejsds55_ineq_201"><alternatives><mml:math>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$L(\theta )$]]></tex-math></alternatives></inline-formula> greater than <inline-formula id="j_nejsds55_ineq_202"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$1/2$]]></tex-math></alternatives></inline-formula> and finding the range of their values for <inline-formula id="j_nejsds55_ineq_203"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>bf</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>bm</mml:mtext>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{\text{bf}}}-{\theta _{\text{bm}}}$]]></tex-math></alternatives></inline-formula>. The range is from a little more than 0 to about 0.4. We can say that there are very good forecasters who give nearly the same forecasts for the two groups.</p>
<p>When the individuals responding to a yes-no survey are categorized in more than one way, or when other data is collected about them, we may prefer to use a more sophisticated forecasting family, such as logistic regression. The logic will remain the same. For particular interesting values of the forecasting variables, we can calculate the range of forecasts given by good forecasters. We can similarly calculate ranges for odds ratios. The computations involved are not trivial, but software environments adequate to the task would not be need to be more complex for the user than those that now use logistic regression for nominally inferential analyses.</p>
<p>The descriptive approach can be compared with the inferential approach used in 2016 by the University of Michigan’s Diversity, Equity &amp; Inclusion Initiative. Michigan sought inferential legitimacy by using random samples. As they explained in their report on the faculty survey [<xref ref-type="bibr" rid="j_nejsds55_ref_041">41</xref>, p. 6], <disp-quote>
<p>Given the large faculty population at the University of Michigan, this study used a sample survey approach rather than a census of all faculty. A carefully selected sample, with randomization, allows researchers to make scientifically based inferences to the population as a whole.</p></disp-quote> The second sentence of this quotation raises the question, unanswered by the authors, of how they would have made scientifically based inferences had they performed a whole census. How would they then have decided which differences were meaningful?</p>
<p>In any case, the authors chose <inline-formula id="j_nejsds55_ineq_204"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>500</mml:mn></mml:math><tex-math><![CDATA[$1,500$]]></tex-math></alternatives></inline-formula> out of <inline-formula id="j_nejsds55_ineq_205"><alternatives><mml:math>
<mml:mn>6</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>700</mml:mn></mml:math><tex-math><![CDATA[$6,700$]]></tex-math></alternatives></inline-formula> faculty members at random to complete the survey. The survey results were then analyzed using logistic regression, and a number of differences were observed to be statistically significant. It was found, for example, that female faculty were 130% more likely to feel discriminated against than male faculty (i.e., the odds ratio for a positive response to the question was equal to 2.3 and significantly different from 1). The results of the survey were clearly meaningful, but the inferential logic is problematic. As David A. Freedman [<xref ref-type="bibr" rid="j_nejsds55_ref_027">27</xref>] has shown, randomization probabilities do not justify logistic regression. Our descriptive theory is not affected by this problem and is just as applicable to a complete census as to a random or non-random sample.</p>
</sec>
</sec>
<sec id="j_nejsds55_s_020">
<label>3.3</label>
<title>Discussion</title>
<p>There is no need to document here the persistence and prevalence over the past two centuries of the use of inferential methods with non-random samples. These abuses have been repeatedly documented and deplored.<xref ref-type="fn" rid="j_nejsds55_fn_012">12</xref><fn id="j_nejsds55_fn_012"><label><sup>12</sup></label>
<p>See [<xref ref-type="bibr" rid="j_nejsds55_ref_028">28</xref>, pp. 212–217] for a particularly concise review. See also [<xref ref-type="bibr" rid="j_nejsds55_ref_001">1</xref>, <xref ref-type="bibr" rid="j_nejsds55_ref_006">6</xref>, <xref ref-type="bibr" rid="j_nejsds55_ref_009">9</xref>, <xref ref-type="bibr" rid="j_nejsds55_ref_026">26</xref>, <xref ref-type="bibr" rid="j_nejsds55_ref_039">39</xref>, <xref ref-type="bibr" rid="j_nejsds55_ref_040">40</xref>, <xref ref-type="bibr" rid="j_nejsds55_ref_048">48</xref>].</p></fn> There have also been numerous efforts to promote descriptive alternatives, but they have gained limited traction in the natural and social sciences. Why can we hope that the proposal to measure relative descriptive success by betting success might fare better? Because, as I have been arguing, our culture already knows something about interpreting betting success.</p>
<p>What are the descriptive alternatives?</p>
<sec id="j_nejsds55_s_021">
<label>3.3.1</label>
<title>Refuse to Calculate Precision</title>
<p>One obvious alternative is to fit models — i.e., calculate descriptive statistics such as means, standard deviations, and regression coefficients — while de-emphasizing the question of their precision and refraining from calculating significance tests or confidence intervals. This has often been the practice in fields, such as geodesy, that have been more influenced by the Gaussian tradition than by the Laplacean tradition.<xref ref-type="fn" rid="j_nejsds55_fn_013">13</xref><fn id="j_nejsds55_fn_013"><label><sup>13</sup></label>
<p>Marie-Françoise Jozeau has documented the competition of the two traditions in geodesy [<xref ref-type="bibr" rid="j_nejsds55_ref_036">36</xref>, <xref ref-type="bibr" rid="j_nejsds55_ref_047">47</xref>].</p></fn> A surveyor’s customers, for example, need a boundary line, not a confidence band. Efforts to establish the same practice in the social sciences have usually had limited and ephemeral impact. During much of the 20th century, some sociologists avoided using significance testing except for random samples, but this avoidance did not endure. In a study of the use of significance testing in two prominent sociology journals from 1935 to 2000 [<xref ref-type="bibr" rid="j_nejsds55_ref_038">38</xref>], Erin Leahey found a shift around 1975 among authors who had data on an entire population. Before that date, some of these authors declined to use significance tests, afterwards few did.</p>
<p>In 2004, in his <italic>Regression Analysis: A Constructive Critique</italic> [<xref ref-type="bibr" rid="j_nejsds55_ref_005">5</xref>], the sociologist Richard A. Berk gave three cheers for a purely descriptive use of multiple regression: estimate the regression coefficients, but do not calculate p-values or confidence intervals. The book has been widely cited, but so far as I know few researchers have followed this advice. What use is a regression coefficient if you have no sense of its precision?</p>
<p>John Tukey’s advocacy of data analysis was another effort to shift attention from inference to description [<xref ref-type="bibr" rid="j_nejsds55_ref_015">15</xref>]. But Tukey’s distinction between exploratory data analysis and confirmatory data analysis undermines the effort to promote description to a goal in itself. The terminology suggests that we explore in order to discover what we will then try to confirm.</p>
</sec>
<sec id="j_nejsds55_s_022">
<label>3.3.2</label>
<title>Reinterpreting Inferential Quantities</title>
<p>In 1995, Richard A. Berk and his colleagues Bruce Western and Robert E. Weiss proposed that the neo-Bayesian philosophy could justify statistical modeling for entire populations [<xref ref-type="bibr" rid="j_nejsds55_ref_007">7</xref>]. I have not been able to understand their argument.</p>
<p>Another notable effort to make inferential tools descriptive was mounted earlier by [<xref ref-type="bibr" rid="j_nejsds55_ref_025">25</xref>]. In the simplest cases, their proposal involves permuting residuals and interpreting a p-value as a measure of how unusual the actual data is in the population of alternative data thus generated. Along with many other mathematical statisticians, I was very intrigued by this proposal when it appeared, but I always found its intuition elusive.</p>
<p>The most common justification of using inferential tools with non-random samples is, of course, the argument that assumptions are never exactly satisfied. George Box’s slogan, “all models are wrong but some are useful” is evoked to justify calculating p-values and confidence intervals. This leaves unanswered, however, the question of what we are being asked to have confidence in. Any claim that the calculations are mere description is contradicted by the language being used.</p>
</sec>
<sec id="j_nejsds55_s_023">
<label>3.3.3</label>
<title>Other Justifications for the Likelihood Ratio?</title>
<p>I have proposed scoring relative success by the likelihood ratio <inline-formula id="j_nejsds55_ineq_206"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${P_{1}}(y)/{P_{2}}(y)$]]></tex-math></alternatives></inline-formula>. This is hardly a new idea; it goes back at least to Laplace. What this paper adds is a betting interpretation. The ratio is the factor by which algorithm 1 multiplies its capital betting against algorithm 2.</p>
<p>The ratio <inline-formula id="j_nejsds55_ineq_207"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${P_{1}}(y)/{P_{2}}(y)$]]></tex-math></alternatives></inline-formula> can also be understood in terms of information theory. John L. Kelly Jr. entitled his 1956 paper “A new interpretation of information rate”; for him Kelly betting was merely an interesting sidelight to information theory. The connection has been elaborated using the idea of minimum description length [<xref ref-type="bibr" rid="j_nejsds55_ref_044">44</xref>, <xref ref-type="bibr" rid="j_nejsds55_ref_030">30</xref>]. But these ideas too have failed to gain traction in the natural and social sciences, and we cannot expect that they will. As beautiful and simple as they are for mathematicians, the ideas are not part of our wider culture.</p>
</sec>
<sec id="j_nejsds55_s_024">
<label>3.3.4</label>
<title>Other Loss Functions?</title>
<p>If we set the betting interpretation aside and change the scale by taking logarithms, <inline-formula id="j_nejsds55_ineq_208"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${P_{1}}(y)/{P_{2}}(y)$]]></tex-math></alternatives></inline-formula> becomes <inline-formula id="j_nejsds55_ineq_209"><alternatives><mml:math>
<mml:mo movablelimits="false">ln</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>−</mml:mo>
<mml:mo movablelimits="false">ln</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\ln ({P_{1}}(y))-\ln ({P_{2}}(y))$]]></tex-math></alternatives></inline-formula>. As this makes clear, we are comparing the two forecasts using log loss. There are many other loss functions. Why shouldn’t we give others equal attention?</p>
<p>The answer, of course, is that we are not setting the betting interpretation aside. A vast amount of work has been done, in decision theory and more recently in machine learning, on evaluating predictions using loss functions. But in most fields of natural and social science, this work has gained little to no traction. I think this is because the scales of measurement have no meaning to the scientists and their larger public. Even though the cutoffs and other conventions we must use will be relatively arbitrary, the idea of measuring relative success in prediction by success in betting and resistance against betting does have meaning for scientists and the publics with whom they want to communicate.</p>
</sec>
</sec>
</sec>
</body>
<back>
<ack id="j_nejsds55_ack_001">
<title>Acknowledgements</title>
<p>I have had helpful discussions on these topics with countless colleagues, most recently with Marshall Abrams, John Aldrich, Gert de Cooman, Harry Crane, Nancy DiTomaso, Ruobin Gong, Peter Grünwald, Zev Hirsch, Wouter Koolen, Kitae Kum, Barry Loewer, Aaditya Ramdas, Judith ter Schure, Vladimir Vovk, Ruodu Wang, Sandy Zabell, and Xueyin Zhang. I have also benefited from comments by the editors and several anonymous referees. On the topic of optional continuation, the workshop <italic>Safe, Anytime-Valid Inference (SAVI) and Game-theoretic Statistics</italic> (May 30–June 3, 2022 in Eindhoven, Netherlands) was especially useful.</p></ack>
<ref-list id="j_nejsds55_reflist_001">
<title>References</title>
<ref id="j_nejsds55_ref_001">
<label>[1]</label><mixed-citation publication-type="journal"> <string-name><surname>Alexander</surname>, <given-names>N.</given-names></string-name> (<year>2015</year>). <article-title>What’s more general than a whole population?</article-title> <source>Emerging Themes in Epidemiology</source> <volume>12</volume>(<issue>1</issue>) <fpage>1</fpage>–<lpage>5</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_002">
<label>[2]</label><mixed-citation publication-type="book"> <string-name><surname>Augustin</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Coolen</surname>, <given-names>F. P. A.</given-names></string-name>, <string-name><surname>de Cooman</surname>, <given-names>G.</given-names></string-name> and <string-name><surname>Troffaes</surname>, <given-names>M. C. M.</given-names></string-name>, eds. (<year>2014</year>) <source>Introduction to Imprecise Probabilities</source>. <publisher-name>Wiley</publisher-name>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1002/9781118763117" xlink:type="simple">https://doi.org/10.1002/9781118763117</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3236913">MR3236913</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_003">
<label>[3]</label><mixed-citation publication-type="journal"> <string-name><surname>Barnard</surname>, <given-names>G.</given-names></string-name> (<year>1947</year>). <article-title>Review of <italic>Sequential Analysis</italic> by Abraham Wald</article-title>. <source>Journal of the American Statistical Association</source> <volume>42</volume>(<issue>240</issue>) <fpage>658</fpage>–<lpage>665</lpage>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0020764">MR0020764</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_004">
<label>[4]</label><mixed-citation publication-type="journal"> <string-name><surname>Barnard</surname>, <given-names>G. A.</given-names></string-name> (<year>1949</year>). <article-title>Statistical inference</article-title>. <source>Journal of the Royal Statistical Society B</source> <volume>11</volume>(<issue>2</issue>) <fpage>115</fpage>–<lpage>149</lpage>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0034983">MR0034983</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_005">
<label>[5]</label><mixed-citation publication-type="book"> <string-name><surname>Berk</surname>, <given-names>R. A.</given-names></string-name> (<year>2004</year>) <source>Regression Analysis: A Constructive Critique</source>. <publisher-name>SAGE</publisher-name>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_006">
<label>[6]</label><mixed-citation publication-type="chapter"> <string-name><surname>Berk</surname>, <given-names>R. A.</given-names></string-name> and <string-name><surname>Freedman</surname>, <given-names>D. A.</given-names></string-name> (<year>2003</year>). <chapter-title>Statistical assumptions as empirical commitments</chapter-title>. In <source>Law, Punishment, and Social Control: Essays in Honor of Sheldon Messinger</source>, <edition>2nd</edition> ed. <fpage>235</fpage>–<lpage>254</lpage>. <publisher-name>Aldine de Gruyter</publisher-name>, <publisher-loc>Berlin</publisher-loc>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_007">
<label>[7]</label><mixed-citation publication-type="journal"> <string-name><surname>Berk</surname>, <given-names>R. A.</given-names></string-name>, <string-name><surname>Western</surname>, <given-names>B.</given-names></string-name> and <string-name><surname>Weiss</surname>, <given-names>R. E.</given-names></string-name> (<year>1995</year>). <article-title>Statistical inference for apparent populations</article-title>. <source>Sociological Methodology</source> <volume>25</volume> <fpage>421</fpage>–<lpage>458</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_008">
<label>[8]</label><mixed-citation publication-type="chapter"> <string-name><surname>Breiman</surname>, <given-names>L.</given-names></string-name> (<year>1961</year>). <chapter-title>Optimal gambling systems for favorable games</chapter-title>. In <source>Fourth Berkeley Symposium on Probability and Mathematical Statistics</source> (<string-name><given-names>J.</given-names> <surname>Neyman</surname></string-name>, ed.) <volume>1</volume> <fpage>65</fpage>–<lpage>78</lpage>. <publisher-name>University of California Press</publisher-name>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0135630">MR0135630</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_009">
<label>[9]</label><mixed-citation publication-type="journal"> <string-name><surname>Bru</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Bru</surname>, <given-names>M.-F.</given-names></string-name> and <string-name><surname>Bienaymé</surname>, <given-names>O.</given-names></string-name> (<year>1997</year>). <article-title>La statistique critiquée par le calcul des probabilités: Deux manuscrits inédits d’Irenée Jules Bienaymé</article-title>. <source>Revue d’Histoire des Mathématiques</source> <volume>3</volume> <fpage>137</fpage>–<lpage>239</lpage>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1620388">MR1620388</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_010">
<label>[10]</label><mixed-citation publication-type="book"> <string-name><surname>Cahusac</surname>, <given-names>P.</given-names></string-name> (<year>2020</year>) <source>Evidence-Based Statistics: An Introduction to the Evidential Approach — from Likelihood Principle to Statistical Practice</source>. <publisher-name>Wiley</publisher-name>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_011">
<label>[11]</label><mixed-citation publication-type="journal"> <string-name><surname>Darling</surname>, <given-names>D. A.</given-names></string-name> and <string-name><surname>Robbins</surname>, <given-names>H.</given-names></string-name> (<year>1967</year>). <article-title>Confidence sequences for mean, variance, and median</article-title>. <source>Proceedings of the National Academy of Sciences of the United States of America</source> <volume>58</volume>(<issue>1</issue>) <fpage>66</fpage>–<lpage>68</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1073/pnas.58.1.66" xlink:type="simple">https://doi.org/10.1073/pnas.58.1.66</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0215406">MR0215406</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_012">
<label>[12]</label><mixed-citation publication-type="journal"> <string-name><surname>Dawid</surname>, <given-names>A. P.</given-names></string-name> (<year>1984</year>). <article-title>Present position and potential developments: Some personal views. Statistical theory, the prequential approach</article-title>. <source>Journal of the Royal Statistical Society: Series A</source> <volume>147</volume>(<issue>2</issue>) <fpage>278</fpage>–<lpage>290</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.2307/2981683" xlink:type="simple">https://doi.org/10.2307/2981683</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0763811">MR0763811</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_013">
<label>[13]</label><mixed-citation publication-type="journal"> <string-name><surname>Dawid</surname>, <given-names>A. P.</given-names></string-name> (<year>1991</year>). <article-title>Fisherian inference in likelihood and prequential frames of reference</article-title>. <source>Journal of the Royal Statistical Society: Series B</source> <volume>53</volume>(<issue>1</issue>) <fpage>79</fpage>–<lpage>109</lpage>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1094276">MR1094276</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_014">
<label>[14]</label><mixed-citation publication-type="book"> <string-name><surname>Diaconis</surname>, <given-names>P.</given-names></string-name> and <string-name><surname>Skyrms</surname>, <given-names>B.</given-names></string-name> (<year>2018</year>) <source>Ten Great Ideas about Chance</source>. <publisher-name>Princeton</publisher-name>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3702017">MR3702017</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_015">
<label>[15]</label><mixed-citation publication-type="journal"> <string-name><surname>Donoho</surname>, <given-names>D.</given-names></string-name> (<year>2017</year>). <article-title>50 years of data science</article-title>. <source>Journal of Computational and Graphical Statistics</source> <volume>26</volume>(<issue>4</issue>) <fpage>745</fpage>–<lpage>766</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/10618600.2017.1384734" xlink:type="simple">https://doi.org/10.1080/10618600.2017.1384734</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3765335">MR3765335</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_016">
<label>[16]</label><mixed-citation publication-type="book"> <string-name><surname>Doob</surname>, <given-names>J. L.</given-names></string-name> (<year>1953</year>) <source>Stochastic Processes</source>. <publisher-name>Wiley</publisher-name>, <publisher-loc>New York</publisher-loc>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0058896">MR0058896</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_017">
<label>[17]</label><mixed-citation publication-type="book"> <string-name><surname>Edwards</surname>, <given-names>A. W. F.</given-names></string-name> (<year>1972</year>) <source>Likelihood: An Account of the Statistical Concept of Likelihood and its Application to Scientific Inference</source>. <publisher-name>Cambridge</publisher-name>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0348869">MR0348869</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_018">
<label>[18]</label><mixed-citation publication-type="book"> <string-name><surname>Ethier</surname>, <given-names>S.</given-names></string-name> (<year>2010</year>) <source>The Doctrine of Chances: Probabilistic Aspects of Gambling</source>. <publisher-name>Springer</publisher-name>, <publisher-loc>Berlin</publisher-loc>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/978-3-540-78783-9" xlink:type="simple">https://doi.org/10.1007/978-3-540-78783-9</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2656351">MR2656351</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_019">
<label>[19]</label><mixed-citation publication-type="book"> <string-name><surname>Feller</surname>, <given-names>W.</given-names></string-name> (<year>1950</year>) <source>An Introduction to Probability Theory and Its Applications</source>, <edition>1st</edition> ed. <publisher-name>Wiley</publisher-name>, <publisher-loc>New York</publisher-loc>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0038583">MR0038583</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_020">
<label>[20]</label><mixed-citation publication-type="journal"> <string-name><surname>Feller</surname>, <given-names>W. K.</given-names></string-name> (<year>1940</year>). <article-title>Statistical aspects of ESP</article-title>. <source>The Journal of Parapsychology</source> <volume>4</volume>(<issue>2</issue>) <fpage>271</fpage>–<lpage>298</lpage>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0004461">MR0004461</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_021">
<label>[21]</label><mixed-citation publication-type="journal"> <string-name><surname>Fisher</surname>, <given-names>R. A.</given-names></string-name> (<year>1952</year>). <article-title>Sequential experimentation</article-title>. <source>Biometrics</source> <volume>8</volume> <fpage>183</fpage>–<lpage>187</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_022">
<label>[22]</label><mixed-citation publication-type="book"> <string-name><surname>Fisher</surname>, <given-names>R. A.</given-names></string-name> (<year>1956</year>) <source>Statistical Methods and Scientific Inference</source>. <publisher-name>Hafner</publisher-name>. <comment>Later editions in 1959 and 1973</comment>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0131909">MR0131909</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_023">
<label>[23]</label><mixed-citation publication-type="chapter"> <string-name><surname>Fourier</surname>, <given-names>J.</given-names></string-name> (<year>1826</year>). <chapter-title>Mémoire sur les résultats moyens déduits d’un grand nombre d’observations</chapter-title>. In <source>Recherches Statistiques sur la Ville de Paris et le dÉpartement de la Seine</source> (<string-name><given-names>J.</given-names> <surname>Fourier</surname></string-name>, ed.). <publisher-name>Imprimerie Royale</publisher-name>, <publisher-loc>Paris</publisher-loc>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_024">
<label>[24]</label><mixed-citation publication-type="chapter"> <string-name><surname>Fourier</surname>, <given-names>J.</given-names></string-name> (<year>1829</year>). <chapter-title>Second mémoire sur les résultats moyens et sur les erreurs des mesures</chapter-title>. In <source>Recherches statistiques sur la ville de Paris et le département de la Seine</source> (<string-name><given-names>J.</given-names> <surname>Fourier</surname></string-name>, ed.) <publisher-name>Imprimerie royale</publisher-name>, <publisher-loc>Paris</publisher-loc>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_025">
<label>[25]</label><mixed-citation publication-type="journal"> <string-name><surname>Freedman</surname>, <given-names>D.</given-names></string-name> and <string-name><surname>Lane</surname>, <given-names>D.</given-names></string-name> (<year>1983</year>). <article-title>A nonstochastic interpretation of reported significance levels</article-title>. <source>Journal of Business and Economic Statistics</source> <volume>1</volume> <fpage>292</fpage>–<lpage>298</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_026">
<label>[26]</label><mixed-citation publication-type="journal"> <string-name><surname>Freedman</surname>, <given-names>D. A.</given-names></string-name> (<year>1991</year>). <article-title>Statistical models and shoe leather</article-title>. <source>Sociological Methodology (with discussion)</source> <volume>21</volume> <fpage>291</fpage>–<lpage>358</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_027">
<label>[27]</label><mixed-citation publication-type="journal"> <string-name><surname>Freedman</surname>, <given-names>D. A.</given-names></string-name> (<year>2008</year>). <article-title>Randomization does not justify logistic regression</article-title>. <source>Statistical Science</source> <volume>23</volume>(<issue>2</issue>) <fpage>237</fpage>–<lpage>249</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/08-STS262" xlink:type="simple">https://doi.org/10.1214/08-STS262</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2516822">MR2516822</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_028">
<label>[28]</label><mixed-citation publication-type="book"> <string-name><surname>Freedman</surname>, <given-names>D. A.</given-names></string-name> (<year>2009</year>) <source>Statistical Models: Theory and Practice, Revised Edition</source>. <publisher-name>Cambridge</publisher-name>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1017/CBO9780511815867" xlink:type="simple">https://doi.org/10.1017/CBO9780511815867</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2489600">MR2489600</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_029">
<label>[29]</label><mixed-citation publication-type="journal"> <string-name><surname>Greenwood</surname>, <given-names>J. A.</given-names></string-name> (<year>1938</year>). <article-title>An empirical investigation of some sampling problems</article-title>. <source>The Journal of Parapsychology</source> <volume>3</volume>(<issue>2</issue>) <fpage>222</fpage>–<lpage>230</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_030">
<label>[30]</label><mixed-citation publication-type="book"> <string-name><surname>Grünwald</surname>, <given-names>P.</given-names></string-name> (<year>2007</year>) <source>The Minimum Description Length Principle</source>. <publisher-name>MIT</publisher-name>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_031">
<label>[31]</label><mixed-citation publication-type="chapter"> <string-name><surname>Grünwald</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>de Heide</surname>, <given-names>R.</given-names></string-name> and <string-name><surname>Koolen</surname>, <given-names>W. M.</given-names></string-name> (<year>2020</year>). <chapter-title>Safe testing</chapter-title>. In <source>2020 Information Theory and Applications Workshop (ITA)</source> <fpage>1</fpage>–<lpage>54</lpage>. <publisher-name>IEEE</publisher-name>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_032">
<label>[32]</label><mixed-citation publication-type="journal"> <string-name><surname>Grünwald</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Ly</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Pérez-Ortiz</surname>, <given-names>M.</given-names></string-name> and <string-name><surname>ter Schure</surname>, <given-names>J.</given-names></string-name> (<year>2021</year>). <article-title>The safe logrank test: Error control under optional stopping, continuation and prior misspecification</article-title>. <source>Proceedings of Machine Learning Research</source> <volume>146</volume> <fpage>107</fpage>–<lpage>117</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_033">
<label>[33]</label><mixed-citation publication-type="other"> <string-name><surname>Hendriksen</surname>, <given-names>A. A.</given-names></string-name> (2017). <ext-link ext-link-type="uri" xlink:href="https://www.universiteitleiden.nl/binaries/content/assets/science/mi/scripties/master/hendriksen-aa—master—26-juni-2017.pdf"><italic>Betting as an alternative to p-values</italic></ext-link>. Master’s thesis, University of Leiden, under the direction of Peter Grünwald.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_034">
<label>[34]</label><mixed-citation publication-type="book"> <string-name><surname>Hexter</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Pfuntner</surname>, <given-names>L.</given-names></string-name> and <string-name><surname>Haynes</surname>, <given-names>J.</given-names></string-name>, eds. (<year>2020</year>) <source>Appendix Ovidiana: Latin Poems Ascribed to Ovid in the Middle Ages</source>. <publisher-name>Harvard</publisher-name>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_035">
<label>[35]</label><mixed-citation publication-type="book"> <string-name><surname>Howson</surname>, <given-names>C.</given-names></string-name> and <string-name><surname>Urbach</surname>, <given-names>P.</given-names></string-name> (<year>2006</year>) <source>Scientific Reasoning: The Bayesian Approach</source>, <edition>Third</edition> ed. <publisher-name>Open Court</publisher-name>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_036">
<label>[36]</label><mixed-citation publication-type="other"> <string-name><surname>Jozeau</surname>, <given-names>M.-F.</given-names></string-name> Géodésie au XIXème Siècle: De l’hégémonie française à l’hégémonie allemande. Regards belges (1997). PhD thesis, Université Denis Diderot Paris VII, Paris.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_037">
<label>[37]</label><mixed-citation publication-type="journal"> <string-name><surname>Kelly Jr.</surname>, <given-names>J. L.</given-names></string-name> (<year>1956</year>). <article-title>A new interpretation of information rate</article-title>. <source>Bell System Technical Journal</source> <volume>35</volume>(<issue>4</issue>) <fpage>917</fpage>–<lpage>926</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1002/j.1538-7305.1956.tb03809.x" xlink:type="simple">https://doi.org/10.1002/j.1538-7305.1956.tb03809.x</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0090494">MR0090494</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_038">
<label>[38]</label><mixed-citation publication-type="journal"> <string-name><surname>Leahey</surname>, <given-names>E.</given-names></string-name> (<year>2005</year>). <article-title>Alphas and asterisks: The development of statistical significance testing standards in sociology</article-title>. <source>Social Forces</source> <volume>84</volume>(<issue>1</issue>) <fpage>1</fpage>–<lpage>24</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_039">
<label>[39]</label><mixed-citation publication-type="journal"> <string-name><surname>Mason</surname>, <given-names>W. M.</given-names></string-name> (<year>1991</year>). <article-title>Freedman is right as far as he goes, but there is more, and it’s worse. Statisticians could help</article-title>. <source>Sociological Methodology</source> <volume>21</volume> <fpage>337</fpage>–<lpage>351</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_040">
<label>[40]</label><mixed-citation publication-type="book"> <string-name><surname>Matthews</surname>, <given-names>J. R.</given-names></string-name> (<year>1995</year>) <source>Quantification and the Quest for Medical Certainty</source>. <publisher-name>Princeton</publisher-name>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_041">
<label>[41]</label><mixed-citation publication-type="other"> <string-name><surname>Office of Diversity</surname>, <given-names>U. o. M.</given-names> <suffix>Equity &amp; Inclusion</suffix></string-name> (2017). <italic>Results of the 2016 University of Michigan Faculty Campus Climate Survey on Diversity, Equity &amp; Inclusion</italic>. <uri>https://diversity.umich.edu/wp-content/uploads/2017/11/DEI-FACULTY-REPORT-FINAL.pdf</uri></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_042">
<label>[42]</label><mixed-citation publication-type="other"> <string-name><surname>Pawel</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Ly</surname>, <given-names>A.</given-names></string-name> and <string-name><surname>Wagenmakers</surname>, <given-names>E.-J.</given-names></string-name> (2023). Evidential calibration of confidence intervals. <italic>American Statistician</italic>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_043">
<label>[43]</label><mixed-citation publication-type="journal"> <string-name><surname>Ramdas</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Grünwald</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Vovk</surname>, <given-names>V.</given-names></string-name> and <string-name><surname>Shafer</surname>, <given-names>G.</given-names></string-name> (<year>2023</year>). <article-title>Game-theoretic statistics and safe anytime-valid inference</article-title>. <source>Statistical Science</source> <volume>38</volume>(<issue>4</issue>) <fpage>576</fpage>–<lpage>601</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/23-sts894" xlink:type="simple">https://doi.org/10.1214/23-sts894</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4665027">MR4665027</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_044">
<label>[44]</label><mixed-citation publication-type="book"> <string-name><surname>Rissanen</surname>, <given-names>J.</given-names></string-name> (<year>1989</year>) <source>Stochastic Complexity in Statistical Inquiry</source>. <publisher-name>World Scientific</publisher-name>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1082556">MR1082556</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_045">
<label>[45]</label><mixed-citation publication-type="book"> <string-name><surname>Royall</surname>, <given-names>R. M.</given-names></string-name> (<year>1997</year>) <source>Statistical Evidence: A Likelihood Paradigm</source>. <publisher-name>Chapman &amp; Hall</publisher-name>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1629481">MR1629481</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_046">
<label>[46]</label><mixed-citation publication-type="chapter"> <string-name><surname>Shafer</surname>, <given-names>G.</given-names></string-name> (<year>2007</year>). <chapter-title>From Cournot’s principle to market efficiency</chapter-title>. In <source>Augustin Cournot: Modelling Economics</source> (<string-name><given-names>J.-P.</given-names> <surname>Touffut</surname></string-name>, ed.) <fpage>55</fpage>–<lpage>95</lpage>. <publisher-name>Edward Elgar</publisher-name>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_047">
<label>[47]</label><mixed-citation publication-type="other"> <string-name><surname>Shafer</surname>, <given-names>G.</given-names></string-name> (2019). <italic>On the nineteenth-century origins of significance testing and p-hacking</italic>. Working paper 55, <ext-link ext-link-type="uri" xlink:href="http://probabilityandfinance.com">Game-Theoretic Probability Project</ext-link>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_048">
<label>[48]</label><mixed-citation publication-type="journal"> <string-name><surname>Shafer</surname>, <given-names>G.</given-names></string-name> (<year>2019</year>). <article-title>Pascal’s and Huygens’s game-theoretic foundations for probability</article-title>. <source>Sartoniana</source> <volume>32</volume> <fpage>117</fpage>–<lpage>145</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_049">
<label>[49]</label><mixed-citation publication-type="other"> <string-name><surname>Shafer</surname>, <given-names>G.</given-names></string-name> (2020). <italic>Game-theoretic foundations for statistical testing and imprecise probabilities</italic>. SIPTA School 2020/2021, <ext-link ext-link-type="uri" xlink:href="https://sites.google.com/view/sipta-school-2021/home">Slides and videos</ext-link>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_050">
<label>[50]</label><mixed-citation publication-type="journal"> <string-name><surname>Shafer</surname>, <given-names>G.</given-names></string-name> (<year>2021</year>). <article-title>Testing by betting: A strategy for statistical and scientific communication (with discussion)</article-title>. <source>Journal of the Royal Statistical Society: Series A</source> <volume>184</volume>(<issue>2</issue>) <fpage>407</fpage>–<lpage>478</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1111/rssa.12647" xlink:type="simple">https://doi.org/10.1111/rssa.12647</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4255905">MR4255905</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_051">
<label>[51]</label><mixed-citation publication-type="other"> <string-name><surname>Shafer</surname>, <given-names>G.</given-names></string-name> (2022). <italic>That’s what all the old guys said: The many faces of Cournot’s principle</italic>. Working Paper 60, <ext-link ext-link-type="uri" xlink:href="http://probabilityandfinance.com/">Game-Theoretic Probability Project</ext-link>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_052">
<label>[52]</label><mixed-citation publication-type="book"> <string-name><surname>Shafer</surname>, <given-names>G.</given-names></string-name> and <string-name><surname>Vovk</surname>, <given-names>V.</given-names></string-name> (<year>2001</year>) <source>Probability and Finance: It’s Only a Game</source>. <publisher-name>Wiley</publisher-name>, <publisher-loc>New York</publisher-loc>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1002/0471249696" xlink:type="simple">https://doi.org/10.1002/0471249696</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1852450">MR1852450</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_053">
<label>[53]</label><mixed-citation publication-type="journal"> <string-name><surname>Shafer</surname>, <given-names>G.</given-names></string-name> and <string-name><surname>Vovk</surname>, <given-names>V.</given-names></string-name> (<year>2006</year>). <article-title>The sources of Kolmogorov’s <italic>Grundbegriffe</italic></article-title>. <source>Statistical Science</source> <volume>21</volume>(<issue>1</issue>) <fpage>70</fpage>–<lpage>98</lpage>. <comment>See also Working Paper 4, <ext-link ext-link-type="uri" xlink:href="http://probabilityandfinance.com/">Game-Theoretic Probability Project</ext-link></comment>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_054">
<label>[54]</label><mixed-citation publication-type="book"> <string-name><surname>Shafer</surname>, <given-names>G.</given-names></string-name> and <string-name><surname>Vovk</surname>, <given-names>V.</given-names></string-name> (<year>2019</year>) <source>Game-Theoretic Foundations for Probability and Finance</source>. <publisher-name>Wiley</publisher-name>, <publisher-loc>New York</publisher-loc>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1002/0471249696" xlink:type="simple">https://doi.org/10.1002/0471249696</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1852450">MR1852450</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_055">
<label>[55]</label><mixed-citation publication-type="journal"> <string-name><surname>Shafer</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Shen</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Vereshchagin</surname>, <given-names>N.</given-names></string-name> and <string-name><surname>Vovk</surname>, <given-names>V.</given-names></string-name> (<year>2011</year>). <article-title>Test martingales, Bayes factors and p-values</article-title>. <source>Statistical Science</source> <volume>26</volume>(<issue>1</issue>) <fpage>84</fpage>–<lpage>101</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/10-STS347" xlink:type="simple">https://doi.org/10.1214/10-STS347</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2849911">MR2849911</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_056">
<label>[56]</label><mixed-citation publication-type="book"> <string-name><surname>Troffaes</surname>, <given-names>M. C. M.</given-names></string-name> and <string-name><surname>de Cooman</surname>, <given-names>G.</given-names></string-name> (<year>2014</year>) <source>Lower Previsions</source>. <publisher-name>Wiley</publisher-name>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1002/9781118762622" xlink:type="simple">https://doi.org/10.1002/9781118762622</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3222242">MR3222242</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_057">
<label>[57]</label><mixed-citation publication-type="other"> <string-name><surname>Vovk</surname>, <given-names>V.</given-names></string-name> (2023). <italic>The diachronic Baysian</italic>. Working paper 64, The Game-Theoretic Probability Project, <uri>http://probabilityandfinance.com</uri>.</mixed-citation>
</ref>
<ref id="j_nejsds55_ref_058">
<label>[58]</label><mixed-citation publication-type="journal"> <string-name><surname>Vovk</surname>, <given-names>V.</given-names></string-name> and <string-name><surname>Wang</surname>, <given-names>R.</given-names></string-name> (<year>2021</year>). <article-title>E-values: Calibration, combination, and applications</article-title>. <source>Annals of Statistics</source> <volume>49</volume>(<issue>3</issue>) <fpage>1736</fpage>–<lpage>1754</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/20-aos2020" xlink:type="simple">https://doi.org/10.1214/20-aos2020</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4298879">MR4298879</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_059">
<label>[59]</label><mixed-citation publication-type="book"> <string-name><surname>Wald</surname>, <given-names>A.</given-names></string-name> (<year>1947</year>) <source>Sequential Analysis</source>. <publisher-name>Wiley</publisher-name>, <publisher-loc>New York</publisher-loc>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0020764">MR0020764</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds55_ref_060">
<label>[60]</label><mixed-citation publication-type="journal"> <string-name><surname>Ziemba</surname>, <given-names>W. T.</given-names></string-name> (<year>2015</year>). <article-title>A response to Professor Paul A. Samuelson’s objections to Kelly capital growth investing</article-title>. <source>The Journal of Portfolio Management</source> <volume>42</volume>(<issue>1</issue>) <fpage>153</fpage>–<lpage>167</lpage>.</mixed-citation>
</ref>
</ref-list>
</back>
</article>
