<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.0 20120330//EN" "JATS-journalpublishing1.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">NEJSDS</journal-id>
<journal-title-group><journal-title>The New England Journal of Statistics in Data Science</journal-title></journal-title-group>
<issn pub-type="ppub">2693-7166</issn><issn-l>2693-7166</issn-l>
<publisher>
<publisher-name>New England Statistical Society</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">NEJSDS73</article-id>
<article-id pub-id-type="doi">10.51387/24-NEJSDS73</article-id>
<article-categories><subj-group subj-group-type="area">
<subject>Machine Learning and Data Mining</subject></subj-group><subj-group subj-group-type="heading">
<subject>Methodology Article</subject></subj-group></article-categories>
<title-group>
<article-title>Fairness and Randomness in Machine Learning: Statistical Independence and Relativization</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Derr</surname><given-names>Rabanus</given-names></name><email xlink:href="mailto:rabanus.derr@uni-tuebingen.de">rabanus.derr@uni-tuebingen.de</email><xref ref-type="aff" rid="j_nejsds73_aff_001"/><xref ref-type="corresp" rid="cor1">∗</xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Williamson</surname><given-names>Robert C.</given-names></name><email xlink:href="mailto:bob.williamson@uni-tuebingen.de">bob.williamson@uni-tuebingen.de</email><xref ref-type="aff" rid="j_nejsds73_aff_002"/>
</contrib>
<aff id="j_nejsds73_aff_001">Maria-von-Linden-Str. 6, 72076 Tübingen, <institution>University of Tübingen and Tübingen AI Center</institution>, <country>Germany</country>. E-mail address: <email xlink:href="mailto:rabanus.derr@uni-tuebingen.de">rabanus.derr@uni-tuebingen.de</email></aff>
<aff id="j_nejsds73_aff_002">Maria-von-Linden-Str. 6, 72076 Tübingen, <institution>University of Tübingen and Tübingen AI Center</institution>, <country>Germany</country>. E-mail address: <email xlink:href="mailto:bob.williamson@uni-tuebingen.de">bob.williamson@uni-tuebingen.de</email></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>∗</label>Corresponding author.</corresp>
</author-notes>
<pub-date pub-type="ppub"><year>2025</year></pub-date><pub-date pub-type="epub"><day>19</day><month>11</month><year>2024</year></pub-date><volume>3</volume><issue>1</issue><fpage>55</fpage><lpage>72</lpage>
<permissions><copyright-statement>© 2025 New England Statistical Society</copyright-statement><copyright-year>2025</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>Open access article under the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/">CC BY</ext-link> license.</license-p></license></permissions>
<abstract>
<p>Fair Machine Learning endeavors to prevent unfairness arising in the context of machine learning applications embedded in society. To this end, several mathematical fairness notions have been proposed. The most known and used notions turn out to be expressed in terms of statistical independence, which is taken to be a primitive and unambiguous notion. However, two choices remain (and are largely unexamined to date): what exactly is the meaning of statistical independence and what are the groups to which we ought to be fair? We answer both questions by leveraging Richard Von Mises’ theory of probability, which starts with data, and then builds the machinery of probability from the ground up. In particular, his theory places a relative definition of randomness as statistical independence at the center of statistical modelling. Much in contrast to the classically used, absolute i.i.d.-randomness, which turns out to be “orthogonal” to his conception. We show how Von Mises’ frequential modeling approach fits well to the problem of fair machine learning and show how his theory (suitably interpreted) demonstrates the equivalence between the contestability of the choice of groups in the fairness criterion and the contestability of the choice of relative randomness. We thus conclude that the problem of being fair in machine learning is precisely as hard as the problem of defining what is meant by being random. In both cases there is a consequential choice, yet there is no universal “right” choice possible.</p>
</abstract>
<kwd-group>
<label>Keywords and phrases</label>
<kwd>Fairness in machine learning</kwd>
<kwd>Richard von Mises</kwd>
<kwd>Randomness</kwd>
<kwd>Statistical independence</kwd>
</kwd-group>
<funding-group><funding-statement>This work was funded in part by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under Germany’s Excellence Strategy — EXC number 2064/1 — Project number 390727645; it was also supported by the German Federal Ministry of Education and Research (BMBF): Tübingen AI Center.</funding-statement></funding-group>
</article-meta>
</front>
<body>
<sec id="j_nejsds73_s_001">
<label>1</label>
<title>Introduction</title>
<p>Under the name “Fair Machine Learning” researchers have attempted to tackle problems of injustice, fairness, discrimination arising in the context of machine learning applications embedded in society [<xref ref-type="bibr" rid="j_nejsds73_ref_005">5</xref>]. Despite the variety of definitions of fairness and proposed “fair algorithms,” we still lack a conceptual understanding of fairness in machine learning [<xref ref-type="bibr" rid="j_nejsds73_ref_075">75</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_084">84</xref>]. What does it mean for predictions to be fair? How does the statistical frame influence fairness? Is there fair data and what would it look like? For instance more concretely, how does a population of individuals and their corresponding predictions look like if a provided definition of fairness is fulfilled?</p>
<p>We focus on a collection of widely used fairness notions which are based on statistical independence e.g., [<xref ref-type="bibr" rid="j_nejsds73_ref_016">16</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_046">46</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_022">22</xref>], but examine them from a new perspective. Surprisingly, debates concerning these notions have not questioned the role and meaning of statistical independence upon which they are based. As we shall argue, statistical independence is far from being a mathematical concept linked to one unique interpretation (see ğ<xref rid="j_nejsds73_s_008">4.2</xref>). This paper, in contrast to much of the literature on fairness in machine learning, e.g., [<xref ref-type="bibr" rid="j_nejsds73_ref_016">16</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_033">33</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_046">46</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_022">22</xref>], investigates what many definitions of fairness take for granted: a well-defined and meaningful notion of statistical independence.</p>
<p>Another, less popular, strand of research investigates the role of randomness in machine learning [<xref ref-type="bibr" rid="j_nejsds73_ref_094">94</xref>]. The standard randomness notion, independently and often identically distributed data points, suffers the longstanding critique of being inadequate (cf. [<xref ref-type="bibr" rid="j_nejsds73_ref_089">89</xref>]). Again, statistical independence lies at the foundation of this, hitherto unrelated to fairness, concept. (We will justify below our use of “randomness” as used here.)</p>
<p>At the core of both our observations is the unreflective use of a convenient mathematical theory of probability. Kolmogorov’s axiomatization of probability theory, developed 1933 in his book [<xref ref-type="bibr" rid="j_nejsds73_ref_055">55</xref>] (translated in [<xref ref-type="bibr" rid="j_nejsds73_ref_056">56</xref>]), dominates most research in machine learning. As Kolmogorov explicitly stated, his theory was designed as a purely axiomatic, mathematical theory detached from meaning and interpretation. In particular, Kolmogorov’s statistical independence lacks such reference. However, the modeling nature of machine learning and the arising ethical complications within machine learning applied in society ask for semantics of probabilistic notions.</p>
<p>In this work, we focus on statistical independence. We leverage a theory of probability axiomatized by Von Mises [<xref ref-type="bibr" rid="j_nejsds73_ref_107">107</xref>] in order to obtain <italic>meaningful</italic> access to probabilistic notions. (In leaning on Von Mises we are directly following the explicit advice of Kolmogorov [<xref ref-type="bibr" rid="j_nejsds73_ref_056">56</xref>, page 3, footnote 4].) This theory construes probability theory as “scientific”<xref ref-type="fn" rid="j_nejsds73_fn_001">1</xref><fn id="j_nejsds73_fn_001"><label><sup>1</sup></label>
<p>We further use the term “scientific” in order to describe a theory modeling a phenomenon in the world providing interpretations and verifiability in the sense of Von Mises (spelled out more concretely in Section <xref rid="j_nejsds73_s_009">4.3</xref>). Any more detailed discussion, e.g., along the lines of [<xref ref-type="bibr" rid="j_nejsds73_ref_076">76</xref>], is out of the scope of this paper.</p></fn> (as opposed to purely mathematical) with the aim to describe the world and provide interpretations and verifiability [<xref ref-type="bibr" rid="j_nejsds73_ref_107">107</xref>, pages 1 and 14]. Von Mises’ theory of probability provides a mathematical definition of statistical independence which describes statistical phenomena observable in the world. In particular, Von Mises’ statistical independence is mathematically, but not conceptually, related to Kolmogorov’s definition.</p>
<p>In this paper, we, to the best of our knowledge, are the first to apply Von Mises’ randomness to machine learning and to interpret randomness in machine learning in a Von Mises’ way. The paper is structured as follows:</p>
<p>In Section <xref rid="j_nejsds73_s_002">2</xref>, we outline our statistical perspective on machine learning. We present the “independent and identically distributed”-assumption (i.i.d.-assumption) as one commonly used choice for modeling randomness. The further occurrence of statistical independence as fundamental ingredient of fairness notions in machine learning (ğ<xref rid="j_nejsds73_s_003">3</xref>) pushes us to the question: “How to interpret statistical independence in (fair) machine learning?” Remarkably, “Independence” governs many discussions around fairness in machine learning without getting to a concrete meaning of this term. Its deeper semantics remain untouched even in the considerably exhaustive book by Barocas et al. [<xref ref-type="bibr" rid="j_nejsds73_ref_005">5</xref>, Chapter 3, p. 13].</p>
<p>We first dissect Kolmogorov’s widely used definition of statistical independence in Section <xref rid="j_nejsds73_s_006">4</xref> before we propose another mathematical notion following Von Mises. Von Mises uses his notion of independence in order to define randomness. We contrast his definition and the i.i.d.-assumption in Section <xref rid="j_nejsds73_s_012">5</xref>. This reveals a general typification of mathematical definitions of randomness which most importantly differ in the <italic>absoluteness</italic> respectively <italic>relativity</italic> to the problem under consideration (ğ<xref rid="j_nejsds73_s_014">5.2</xref>).</p>
<p>Finally, we leverage Von Mises’ definition of statistical independence to redefine three fairness notions from machine learning (ğ<xref rid="j_nejsds73_s_015">6</xref>). Against the background of Von Mises’ probability theory, we then link randomness and fairness both expressed as statistical independence (ğ<xref rid="j_nejsds73_s_016">7</xref>). Thereby, we reveal an unexpected hypothesis: randomness and fairness can be considered equivalent concepts in machine learning. Randomness becomes a <italic>relative</italic>, even an ethical choice. Fairness, however, turns out to be a modeling assumption about the data used in the machine learning system.</p>
<p>Due to the frequent use of the word “independence” with different meanings in this paper, we differentiate. By “independence” we mean an abstract concept of unrelatedness and non-influence [<xref ref-type="bibr" rid="j_nejsds73_ref_092">92</xref>]. We use it interchangeably with “statistical independence” which emphasizes the probabilistic and statistical context. When referring to the later introduced, formal definitions of statistical independence following Kolmogorov or Von Mises we explicitly state this. Finally, we assign “Independence” (capital “I”) to one of the fairness criteria in machine learning which demands for statistical independence of predictions and sensitive attribute [<xref ref-type="bibr" rid="j_nejsds73_ref_005">5</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_083">83</xref>]. Despite the abstract appearance of this work, we consider it as part of the project to make the very abstract notion of “independence” or “statistical independence” at least a bit more concrete — specifically, we describe independence in terms of samples, not abstractions such as countably additive probability measures. We ask (and try to assist) the reader to become aware of the implicit assumptions taken about the concept of “independence”.</p>
</sec>
<sec id="j_nejsds73_s_002">
<label>2</label>
<title>A Statistical Perspective on Machine Learning</title>
<p>Machine Learning ingests data and provides decisions or inferences. In this sense, at its core, it is statistics.<xref ref-type="fn" rid="j_nejsds73_fn_002">2</xref><fn id="j_nejsds73_fn_002"><label><sup>2</sup></label>
<p>Machine learning could be viewed as classical statistics with a stronger focus on algorithmic realizations (cf. [<xref ref-type="bibr" rid="j_nejsds73_ref_091">91</xref>, p. 6] or [<xref ref-type="bibr" rid="j_nejsds73_ref_058">58</xref>]). While there are ML approaches that do not seem statistical in the classical sense (e.g., worst case online sequence prediction), our general description still holds.</p></fn> Adopting this perspective, we understand machine learning as “modeling data generative processes”. Statistics, respectively machine learning, asks for properties of data generating processes given a collection of data [<xref ref-type="bibr" rid="j_nejsds73_ref_109">109</xref>, p. ix], [<xref ref-type="bibr" rid="j_nejsds73_ref_004">4</xref>, p. 1].</p>
<p>We assume that a data generative process occurs somehow in the world. We are confronted with a collection of numbers, the data, produced by the process and acquired by measurement. A “model” is a mathematical description of such a data generating process. This description should allow us to make predictions in an algorithmic fashion. Thus, we require, <italic>inter alia</italic>, a mathematical description of data — “a model for data collection” [<xref ref-type="bibr" rid="j_nejsds73_ref_018">18</xref>, p. 207]. What is arguably <italic>the</italic> standard model of data is stated in [<xref ref-type="bibr" rid="j_nejsds73_ref_030">30</xref>, p. 11]: <disp-quote>
<p>We shall assume in this book that <inline-formula id="j_nejsds73_ineq_001"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({X_{1}},{Y_{1}}),\dots ,({X_{n}},{Y_{n}})$]]></tex-math></alternatives></inline-formula>, the data, is a sequence of independent identically distributed (i.i.d.) random pairs with the same distribution[…].</p></disp-quote> Similar definitions can be found in many machine learning or statistics textbooks (e.g., [<xref ref-type="bibr" rid="j_nejsds73_ref_018">18</xref>, Def. 5.1.1]). Data (measurements from the world) is conceived of (mathematically) as a collection of random variables which share the same distribution and which are (statistically) independent to each other. Implicit in this definition is that the data indeed <italic>has</italic> a stable distribution. The assumed independence can be interpreted as presumption of randomness of the data. Each data point was “drawn independently”<xref ref-type="fn" rid="j_nejsds73_fn_003">3</xref><fn id="j_nejsds73_fn_003"><label><sup>3</sup></label>
<p>Observe that as a mathematical theory of data this already leaves a lot to be desired: you will not find in any text a precise and constructive explanation of this process of “drawn independently”. To be sure, the notion of “statistical independence” is well defined (see further below), as is a collection of random variables. But not the mysterious process of “drawing from”. The closest we can come to such a description of mechanism is the indirect abstract version: the data are created (by the world) in a manner that the (mathematical theorem) of the law of large numbers holds, and that their empirical distribution converges to the distribution which was given in the first place.</p></fn> from all others, with the obvious interpretation that each data point does not give a hint about the value of any other.</p>
<p>The i.i.d. assumption is two-fold: 1) the assumption of identical distributions of a sample, and 2) the mutual independence of points in the sample. The second assumption alone captures and pertains to randomness [<xref ref-type="bibr" rid="j_nejsds73_ref_050">50</xref>, Section 3]. However, since the use of i.i.d. is more common, we refer to this more specific assumption.</p>
<p>Even though many results in statistics and machine learning rely on the i.i.d. assumption (e.g., law of large numbers and central limit theorem in statistics [<xref ref-type="bibr" rid="j_nejsds73_ref_018">18</xref>], generalization bounds in statistical learning theory [<xref ref-type="bibr" rid="j_nejsds73_ref_091">91</xref>] and computationally feasible expressions in probabilistic machine learning [<xref ref-type="bibr" rid="j_nejsds73_ref_030">30</xref>]), it has always been subject of fundamental critique.<xref ref-type="fn" rid="j_nejsds73_fn_004">4</xref><fn id="j_nejsds73_fn_004"><label><sup>4</sup></label>
<p>Nicely summarized by Glenn Shafer in a comment on [<xref ref-type="bibr" rid="j_nejsds73_ref_040">40</xref>] “The i.i.d. case has also been central to statistics ever since Jacob Bernoulli proved the law of large numbers at the end of the 17th century, but its inadequacy was always obvious.” [<xref ref-type="bibr" rid="j_nejsds73_ref_089">89</xref>].</p></fn> Other randomness definitions are rarely applied, but exceptions exist [<xref ref-type="bibr" rid="j_nejsds73_ref_108">108</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_096">96</xref>].</p>
<p>In summary, statistical conclusions often rely on the i.i.d.-description of data. This description embraces a model of randomness making randomness a substantial assumption about the data in statistics and machine learning. Interestingly, statistical independence lies at the foundation of another, hitherto unrelated, concept: many fairness criteria in fair machine learning are expressed in terms of statistical independence.</p>
</sec>
<sec id="j_nejsds73_s_003">
<label>3</label>
<title>Fair Machine Learning Relies on Statistical Independence</title>
<p>With the broad use of machine learning algorithms in many socially relevant domains, e.g., recidivism risk prediction or algorithmic hiring, machine learning algorithms turned out to be part of discriminatory practices [<xref ref-type="bibr" rid="j_nejsds73_ref_022">22</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_080">80</xref>]. These revelations were accompanied by the rise of an entire research field, called “fair machine learning” (cf. [<xref ref-type="bibr" rid="j_nejsds73_ref_016">16</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_022">22</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_005">5</xref>]). We do not attempt to summarize this large literature here. Instead, we simply take a snapshot of the most widely known fairness criteria in machine learning [<xref ref-type="bibr" rid="j_nejsds73_ref_005">5</xref>, p. 45].</p>
<sec id="j_nejsds73_s_004">
<label>3.1</label>
<title>Three Fairness Criteria in Machine Learning</title>
<p>The three so called observational fairness criteria, which are expressed in terms of statistical independence, encompass a large part of fair machine learning literature:<xref ref-type="fn" rid="j_nejsds73_fn_005">5</xref><fn id="j_nejsds73_fn_005"><label><sup>5</sup></label>
<p>Obviously, there are other fairness notions in machine learning which we have not listed above and which are not expressed as statistical independence, e.g., [<xref ref-type="bibr" rid="j_nejsds73_ref_033">33</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_053">53</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_111">111</xref>].</p></fn></p>
<def-list><def-item id="x1-4001x-3.1"><term>Independence</term><def>
<p>demands that the predictions <inline-formula id="j_nejsds73_ineq_002"><alternatives><mml:math><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover></mml:math><tex-math><![CDATA[$\hat{Y}$]]></tex-math></alternatives></inline-formula> are statistically independent of group membership <italic>S</italic> in socially salient, morally relevant groups [<xref ref-type="bibr" rid="j_nejsds73_ref_016">16</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_052">52</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_033">33</xref>].</p></def></def-item><def-item id="x1-4002x-3.1"><term>Separation</term><def>
<p>is formalized as <inline-formula id="j_nejsds73_ineq_003"><alternatives><mml:math><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
<mml:mo stretchy="false">⊥</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi></mml:math><tex-math><![CDATA[$\hat{Y}\perp S|Y$]]></tex-math></alternatives></inline-formula>, i.e., the prediction <inline-formula id="j_nejsds73_ineq_004"><alternatives><mml:math><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover></mml:math><tex-math><![CDATA[$\hat{Y}$]]></tex-math></alternatives></inline-formula> is conditionally independent of the sensitive attribute <italic>S</italic> given the true label <italic>Y</italic> [<xref ref-type="bibr" rid="j_nejsds73_ref_046">46</xref>].</p></def></def-item><def-item id="x1-4003x-3.1"><term>Sufficiency</term><def>
<p>is fulfilled if and only if <inline-formula id="j_nejsds73_ineq_005"><alternatives><mml:math>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo stretchy="false">⊥</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo stretchy="false">|</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover></mml:math><tex-math><![CDATA[$Y\perp S|\hat{Y}$]]></tex-math></alternatives></inline-formula> [<xref ref-type="bibr" rid="j_nejsds73_ref_054">54</xref>].</p></def></def-item></def-list>
<p>For the sake of distinguishing between the fairness criteria “Independence” and statistical independence, we henceforth mark all fairness criteria by a leading capital letter. Each of the notions appear in a variety of ways and under different names [<xref ref-type="bibr" rid="j_nejsds73_ref_005">5</xref>, p. 45ff]. From the perspective of ethics, the fairness criteria have been substantiated via loss egalitarianism [<xref ref-type="bibr" rid="j_nejsds73_ref_011">11</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_111">111</xref>], absence of discrimination [<xref ref-type="bibr" rid="j_nejsds73_ref_011">11</xref>], affirmative action [<xref ref-type="bibr" rid="j_nejsds73_ref_009">9</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_083">83</xref>] or equality of opportunity [<xref ref-type="bibr" rid="j_nejsds73_ref_046">46</xref>].</p>
<p>Certainly, statistical independence is not equivalent to fairness in general (a constellation of concepts sharing a common name, and perhaps little else agreed by all) [<xref ref-type="bibr" rid="j_nejsds73_ref_033">33</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_046">46</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_083">83</xref>]. The nature of fairness has been discussed for decades, e.g., in political philosophy [<xref ref-type="bibr" rid="j_nejsds73_ref_082">82</xref>], moral philosophy [<xref ref-type="bibr" rid="j_nejsds73_ref_061">61</xref>] and actuarial science [<xref ref-type="bibr" rid="j_nejsds73_ref_001">1</xref>]. The “essentially contested” nature of fairness suggests that no universal, statistical criterion of fairness exists [<xref ref-type="bibr" rid="j_nejsds73_ref_039">39</xref>]. How fairness should be defined is a context-specific decision [<xref ref-type="bibr" rid="j_nejsds73_ref_084">84</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_047">47</xref>].</p>
<p>Nevertheless, in order to incorporate fairness notions into algorithmic tools we require mathematical formalisations of fairness definitions. The three named criteria dominate most of the practical fair machine learning tools [<xref ref-type="bibr" rid="j_nejsds73_ref_005">5</xref>, p. 45], presumably because their simple definitions make it easy to incorporate them in learning procedures in a pre-, in- or post-processing way [<xref ref-type="bibr" rid="j_nejsds73_ref_005">5</xref>, Chapter 3, p. 20]. Regarding both the reductionist definition of fairness and the pragmatic justification, we emphasize that <italic>our argumentation is solely with respect to the fairness criteria named above</italic>.</p>
<p>The three fairness criteria are described as <italic>group</italic> fairness notions since each of the definitions is intrinsically relativized with respect to sensitive groups. The definition of sensitive groups substantially influences the notion of fairness. For instance, via custom categorization one can provide fairness by group-design (see [<xref ref-type="bibr" rid="j_nejsds73_ref_067">67</xref>, Section H.3] for a detailed discussion of the question of choice of groups). In addition, the meaning of groups in a societal context influences the choice of groups as elaborated in [<xref ref-type="bibr" rid="j_nejsds73_ref_049">49</xref>], and as explored in a long line of work in social psychology [<xref ref-type="bibr" rid="j_nejsds73_ref_017">17</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_097">97</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_048">48</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_066">66</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_019">19</xref>]. We contribute to the debate by drawing a connection between the choice of groups and the choice of randomness in ğ<xref rid="j_nejsds73_s_018">7.1.1</xref>.</p>
</sec>
<sec id="j_nejsds73_s_005">
<label>3.2</label>
<title>Independence in Mathematics and the World</title>
<p>Behind the formalization of fairness as statistical independence, there is an apparently rigid, mathematical definition of statistical independence. The fairness criteria presume that we have the machinery of probability theory at our disposal and the relationship of mathematics and the world is clear and unambiguous. However, as we elaborate further in the following, there is no single notion of “the” mathematical theory of probability [<xref ref-type="bibr" rid="j_nejsds73_ref_036">36</xref>]. Furthermore, it is not clear what it means to be statistically independent when talking of measurements in the world. Respectively, it is not obvious that the standard formulation of statistical independence is the right one to use.</p>
<p>The current definitions of fairness in machine learning fail and even hurt in practice, because debates on fairness notions take for granted a commonly agreed meaning of statistical independence. This common ground fatally does not exist given just the standard probability theory. Hence, given a specified scenario, e.g., hiring algorithm for public service in Kenya, fair machine learning research should enable founded debates on the meaning and purposefulness of deploying a specific notion of fairness, in particular those which require independence statements because their are algorithmically attractive. If the debate in Kenya would rely on everybody’s – who are involved in the deployment process – intuitive understanding of statistical independence, it would result in a fatal misalignment of reasonings.</p>
<p>We perceive this work as part of the project to emphasize the substantive character formal notions of fairness should have. In other words, fairness is a societal and ethical concept that does not allow for the separation of machine learning as technical tool on one side and the idea of fairness in society on the other side. Debates on fairness have to consider machine learning <italic>in</italic> society.</p>
<p>Hence, if we desire statistical independence to capture a fairness notion applicable to the world, we ought to understand what the mathematical formulae signify in the world. Thus, in addition to the debate about the fairness criteria, a debate on the interpretation of statistical concepts in ethical context is required.</p>
<p>In this work, we contribute to the understanding by scrutinizing the standard definition of statistical independence. Motivated by the occurrence of statistical independence as fundamental ingredient in randomness as well as fairness in machine learning, we first detail the standard account due to Kolmogorov. What is statistical independence? How does statistical independence relate to an independence in the world?</p>
</sec>
</sec>
<sec id="j_nejsds73_s_006">
<label>4</label>
<title>Statistical Independence Revisited</title>
<p>To make any sense of phenomena in our world, we need to ignore large parts of it in order to avoid being overwhelmed. Hence, we usually assume or presume the phenomena of interest depends only on a few factors and to be <italic>independent</italic> of everything else [<xref ref-type="bibr" rid="j_nejsds73_ref_072">72</xref>]. Thus the concept of independence is inherent in a variety of subjects ranging from causal reasoning [<xref ref-type="bibr" rid="j_nejsds73_ref_093">93</xref>], to logic [<xref ref-type="bibr" rid="j_nejsds73_ref_042">42</xref>], accounting [<xref ref-type="bibr" rid="j_nejsds73_ref_024">24</xref>], public law [<xref ref-type="bibr" rid="j_nejsds73_ref_069">69</xref>] and many more. Independence, as we understand it, grasps the concept of incapability of an entity to be expressed, derived or deduced by something else [<xref ref-type="bibr" rid="j_nejsds73_ref_092">92</xref>].</p>
<sec id="j_nejsds73_s_007">
<label>4.1</label>
<title>From Independence to Statistical Independence</title>
<p>Of special interest to us is the concept of independence in probability theories and statistics [<xref ref-type="bibr" rid="j_nejsds73_ref_060">60</xref>][<xref ref-type="bibr" rid="j_nejsds73_ref_036">36</xref>, Section IIF, IIIG and VH]. Independence in a probabilistic context should somehow capture the unrelatedness between the occurrence of events, as has been understood for centuries: <disp-quote>
<p>Two Events are independent, when they have no connexion [sic] one with the other, and that the happening of one neither forwards nor obstructs the happening of the other. [<xref ref-type="bibr" rid="j_nejsds73_ref_028">28</xref>, Introduction, p. 6].</p></disp-quote> Modern probability theory loosely follows this intuition as we see in the following.</p>
</sec>
<sec id="j_nejsds73_s_008">
<label>4.2</label>
<title>Statistical Independence As We Know It Lacks Semantics</title>
<p>Since the axiomatization of probability theory developed by [<xref ref-type="bibr" rid="j_nejsds73_ref_055">55</xref>] (translated in [<xref ref-type="bibr" rid="j_nejsds73_ref_056">56</xref>]), it displaced many other approaches. Mathematically, Kolmogorov’s measure-theoretic axiomatization dominates all other mathematical formalizations to probability and related concepts.<xref ref-type="fn" rid="j_nejsds73_fn_006">6</xref><fn id="j_nejsds73_fn_006"><label><sup>6</sup></label>
<p>Exceptions exist, e.g., in quantum theory [<xref ref-type="bibr" rid="j_nejsds73_ref_044">44</xref>] or in statistics [<xref ref-type="bibr" rid="j_nejsds73_ref_108">108</xref>].</p></fn> In particular, his definition of statistical independence developed a well-accepted, ubiquitous notion. In a simple form it is given by:<xref ref-type="fn" rid="j_nejsds73_fn_007">7</xref><fn id="j_nejsds73_fn_007"><label><sup>7</sup></label>
<p>For a rigorous definition see Appendix <xref rid="j_nejsds73_stat_006">5</xref>.</p></fn></p><statement id="j_nejsds73_stat_001"><label>Definition 1</label>
<title>(Simplified Statistical Independence following Kolmogorov).</title>
<p>Two events <inline-formula id="j_nejsds73_ineq_006"><alternatives><mml:math>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi></mml:math><tex-math><![CDATA[$A,B$]]></tex-math></alternatives></inline-formula> are statistically independent iff 
<disp-formula id="j_nejsds73_eq_001">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo>∩</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ P(A\cap B)=P(A)P(B).\]]]></tex-math></alternatives>
</disp-formula>
</p></statement>
<p>Independence plays a central role in Kolmogorov’s probability theory: “measure theory ends and probability begins with the definition of independence” [<xref ref-type="bibr" rid="j_nejsds73_ref_032">32</xref>, p. 37] (quoted in [<xref ref-type="bibr" rid="j_nejsds73_ref_098">98</xref>]). However, Kolmogorov’s definition of independence is subtle and requires closer investigation.</p>
<p>We employ a small toy example in order to convey the semantic emptiness of Kolmogorov’s definition: consider the experiment of throwing a die. The events under observation are <inline-formula id="j_nejsds73_ineq_007"><alternatives><mml:math>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$A=\{1,2\}$]]></tex-math></alternatives></inline-formula>, seeing one or two pips, respectively <inline-formula id="j_nejsds73_ineq_008"><alternatives><mml:math>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$B=\{2,3\}$]]></tex-math></alternatives></inline-formula>, seeing two or three pips. If the die were fair, so each face has equal probability <inline-formula id="j_nejsds73_ineq_009"><alternatives><mml:math><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>6</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle></mml:math><tex-math><![CDATA[$\frac{1}{6}$]]></tex-math></alternatives></inline-formula> to show up, the events <italic>A</italic> and <italic>B</italic> would turn out to not be independent. In contrast, if the die were loaded in a very special way <inline-formula id="j_nejsds73_ineq_010"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle></mml:math><tex-math><![CDATA[${p_{2}}={p_{3}}=\frac{1}{2}$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_nejsds73_ineq_011"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>6</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[${p_{1}}={p_{4}}={p_{5}}={p_{6}}=0$]]></tex-math></alternatives></inline-formula>, where <inline-formula id="j_nejsds73_ineq_012"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{i}}$]]></tex-math></alternatives></inline-formula> refers to the probability of seeing <italic>i</italic> pips, the events <italic>A</italic> and <italic>B</italic> would be independent following Kolmogorov’s definition.</p>
<p>Thus, statistical independence, even though defined over events, manifests in the correspondence of how probabilities are mapped to events and why. The definition focuses on events. But, the crucial ingredient is the probability. Thus, there is no unique interpretation of statistical independence. A more detailed interpretation and meaning heavily depends on the interpretation of probability in the first place.</p>
<p>Given this observation, we may ask for a notion of independence in the world, which is somehow captured by Kolmogorov’s definition. Kolmogorov himself underlined the avowedly mathematical, axiomatic nature of probability. His theory is in principle detached from any meaning of probabilistic concepts such as statistical independence in the world [<xref ref-type="bibr" rid="j_nejsds73_ref_056">56</xref>, p. 1]. He even questioned the validity of his axioms as reasonable descriptions of the world [<xref ref-type="bibr" rid="j_nejsds73_ref_056">56</xref>, p. 17].</p>
<p>However, one can possibly construct a notion of independence in world which is captured by Kolmogorov’s definition. If one assumes one’s calculations about one’s beliefs on the happening of events are governed by the mathematical rules laid out by Kolmogorov, then Kolmogorov’s definition of statistical independence captures one’s (in-the-world) understanding of an independence of beliefs on the happening of events. This sketch of a purely subjectivist account to statistical independence neglects a justification for the choice of mathematical formulation and skips over any reference to an objective world. In conclusion, Kolmogorov’s independence might capture a worldly concept. But, this independence in the world is not uniquely attached to Kolmogorov’s definition.</p>
<p>There is a third major irritation arising from Kolmogorov’s definition. As observed already above, Kolmogorov treats events as the entities of independence. Though, against the background of De Moivre [<xref ref-type="bibr" rid="j_nejsds73_ref_028">28</xref>]’s intuition on statistical independence, we wonder about this focus. Statistical independence, as De Moivre [<xref ref-type="bibr" rid="j_nejsds73_ref_028">28</xref>] already emphasized, refers to altering “the happening of the event,” but not the event itself. It is not the independence between the shown numbers of the die (whatever this means), but the independence of the processes how the number showed up (loading the die, throwing the die, etc.) which are captured by statistical independence.</p>
<p>This critique is not new. Von Mises already criticized the measure theoretical definition by Kolmogorov in his book [<xref ref-type="bibr" rid="j_nejsds73_ref_107">107</xref>, pp. 36–39]. In summary, he argued that there is no interpretation of statistical independence of “single events.” The unrelatedness, which the probabilistic notion of statistical independence is trying to capture, locates in the process of reoccuring events, but not single events themselves. More recently, Von Collani [<xref ref-type="bibr" rid="j_nejsds73_ref_104">104</xref>] argued in a similar way. It is probability which brings the definition of statistical independence to life and it is the question what probability means and why we use it which links the mathematical definition to a concept in the real world.</p>
<p>In machine learning and statistics it is often presumed that the mathematical definition of independence captures a worldly concept. As we argued in this section, this link is far from being well-defined. However, if we consider machine learning as worldly data modeling, then the natural question arises: what do we model when we leverage Kolmogorov’s statistical independence? What do we mean by independence of events? In order to circumvent these questions, we propose to look into another mathematical theory of probability. This theory was led by the idea of modeling statistical phenomena in the world.</p>
</sec>
<sec id="j_nejsds73_s_009">
<label>4.3</label>
<title>Statistical Independence and a Probability Theory with Inherent Semantics</title>
<p>Around 15 years before Kolmogorov’s <italic>Grundbegriffe der Wahrscheinlichkeitsrechnung</italic> [<xref ref-type="bibr" rid="j_nejsds73_ref_055">55</xref>] (translated as [<xref ref-type="bibr" rid="j_nejsds73_ref_056">56</xref>]), Von Mises proposed an earlier axiomatization of probability theory [<xref ref-type="bibr" rid="j_nejsds73_ref_105">105</xref>]. His less known theory approached the problem of a mathematical theory of probability through the lens of physics. Von Mises aimed for a “mathematical theory of repetitive events.”</p>
<p>This aim included the emphasis on the link between real-world and mathematical idealization. In particular, he offers interpretability and verifiability of his theory [<xref ref-type="bibr" rid="j_nejsds73_ref_107">107</xref>, p. 1 and 14].<xref ref-type="fn" rid="j_nejsds73_fn_008">8</xref><fn id="j_nejsds73_fn_008"><label><sup>8</sup></label>
<p>Von Mises’ discussion pre-dates most of the modern work done in the philosophy of science. Popper [<xref ref-type="bibr" rid="j_nejsds73_ref_076">76</xref>], for instance, was inspired by Von Mises’ conception of randomness and probability. While Von Mises’ definition of verifiability and interpretability is arguably somewhat ill-conceived, nevertheless, he works, in contrast to Kolmogorov’s theory which is of purely syntactical nature, on a semantical project, where the connection of the real world and the mathematical formulation is of vital importance.</p></fn> For interpretation he defined probabilities in a frequency-based way (see Definition <xref rid="j_nejsds73_stat_002">2</xref>). This inherently reflects the repetitive nature of the phenomena under description. By verifiability he referred to the ability to <italic>approximately</italic> verify the probabilistic statements made about the world [<xref ref-type="bibr" rid="j_nejsds73_ref_107">107</xref>, p. 45].</p>
<p>In summary, Von Mises’ theory, in our conception of machine learning, starts the “modeling of data generating processes” on an even more fundamental level then it is currently done via the use of Kolmogorov’s axiomatization. His aim for a mathematical description of data-generating processes (the sequence of repetitive events) aligns to our perspective on machine learning as laid out earlier (cf. Section <xref rid="j_nejsds73_s_002">2</xref>). With Von Mises we obtain access to <italic>meaningful</italic> foundations for statistical concepts in machine learning. In particular, we redefine and reinterpret statistical independence in a Von Misesean way. This suggests new perspectives on the problem of fair machine learning and the concepts of fairness and randomness themselves.</p>
<p>For the further discussion, we summarize the major ingredients of Von Mises’ theory. Fortunately, it turns out that Von Mises’ notion of statistical independence, central to our discussion, is mathematically analogous to the well-known Kolmogorovian definition. Thus, one’s intuition on statistical independence is refined but its mathematical applicability remains.</p>
</sec>
<sec id="j_nejsds73_s_010">
<label>4.4</label>
<title>Von Mises’ Theory of Probability and Randomness in a Nutshell</title>
<p>Von Mises’ axiomatization of probability theory is based on random sequences of events and the interpretation of probability as the limiting frequency that an event occurs in such a sequence [<xref ref-type="bibr" rid="j_nejsds73_ref_105">105</xref>]. These random sequences, called collectives, are the main ingredients of his theory. For the sake of simplicity, we stick to binary collectives. Thus, collectives are 0-1-sequences with certain randomness properties which define probabilities for the labels 0 and 1. Nevertheless, it is possible to define collectives respectively probabilities on richer label sets. Collectives can be extended up to a continuum [<xref ref-type="bibr" rid="j_nejsds73_ref_107">107</xref>, II.B].</p>
<p>For notational economy, we note that a sequence taking values in <inline-formula id="j_nejsds73_ineq_013"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$\{0,1\}$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_nejsds73_ineq_014"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${({s_{i}})_{i\in \mathbb{N}}}$]]></tex-math></alternatives></inline-formula> can be identified with a <italic>function</italic> <inline-formula id="j_nejsds73_ineq_015"><alternatives><mml:math>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$s:\mathbb{N}\to \{0,1\}$]]></tex-math></alternatives></inline-formula>.</p><statement id="j_nejsds73_stat_002"><label>Definition 2</label>
<title>(Collective [<xref ref-type="bibr" rid="j_nejsds73_ref_107">107</xref>, p. 12]).</title>
<p>Let <inline-formula id="j_nejsds73_ineq_016"><alternatives><mml:math>
<mml:mi mathvariant="script">S</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{S}$]]></tex-math></alternatives></inline-formula> be a set of sequences <inline-formula id="j_nejsds73_ineq_017"><alternatives><mml:math>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$s:\mathbb{N}\to \{0,1\}$]]></tex-math></alternatives></inline-formula> with <inline-formula id="j_nejsds73_ineq_018"><alternatives><mml:math>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$s(j)=1$]]></tex-math></alternatives></inline-formula> for infinitely many <inline-formula id="j_nejsds73_ineq_019"><alternatives><mml:math>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi></mml:math><tex-math><![CDATA[$j\in \mathbb{N}$]]></tex-math></alternatives></inline-formula>. In mathematical terms, <italic>collectives</italic> with respect to <inline-formula id="j_nejsds73_ineq_020"><alternatives><mml:math>
<mml:mi mathvariant="script">S</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{S}$]]></tex-math></alternatives></inline-formula> are sequences <inline-formula id="j_nejsds73_ineq_021"><alternatives><mml:math>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$x:\mathbb{N}\to \{0,1\}$]]></tex-math></alternatives></inline-formula> for which the following two conditions hold. 
<list>
<list-item id="j_nejsds73_li_001">
<label>1.</label>
<p>The limit of relative frequencies of 1s, 
<disp-formula id="j_nejsds73_eq_002">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \underset{n\to \infty }{\lim }\frac{|\{i\in \mathbb{N}:x(i)=1,1\le i\le n\}|}{n}\]]]></tex-math></alternatives>
</disp-formula> 
exists.<xref ref-type="fn" rid="j_nejsds73_fn_009">9</xref><fn id="j_nejsds73_fn_009"><label><sup>9</sup></label>
<p>The mathematically inclined reader might notice the requirement for an order structure to obtain a definition of limit here. We use <italic>x</italic> as sequences with the standard order structure on the natural numbers. However, <italic>x</italic> can be generalized to be a net on more arbitrary base sets [<xref ref-type="bibr" rid="j_nejsds73_ref_051">51</xref>].</p></fn> If it exists, then the limit of relative frequencies of 0s exists, too. We define <inline-formula id="j_nejsds73_ineq_022"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{1}}$]]></tex-math></alternatives></inline-formula>, respectively <inline-formula id="j_nejsds73_ineq_023"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{0}}=1-{p_{1}}$]]></tex-math></alternatives></inline-formula>, to be its value.</p>
</list-item>
<list-item id="j_nejsds73_li_002">
<label>2.</label>
<p>For all <inline-formula id="j_nejsds73_ineq_024"><alternatives><mml:math>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="script">S</mml:mi></mml:math><tex-math><![CDATA[$s\in \mathcal{S}$]]></tex-math></alternatives></inline-formula>, 
<disp-formula id="j_nejsds73_eq_003">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mspace width="2.5pt"/>
<mml:mtext>and</mml:mtext>
<mml:mspace width="2.5pt"/>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mspace width="1em"/>
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}& \underset{n\to \infty }{\lim }\frac{|\{i\in \mathbb{N}:x(i)=1\hspace{2.5pt}\text{and}\hspace{2.5pt}s(i)=1,1\le i\le n\}|}{|\{j\in \mathbb{N}:s(j)=1,1\le j\le n\}|}\\ {} & \hspace{1em}=\underset{n\to \infty }{\lim }\frac{|\{i\in \mathbb{N}:x(i)=1,1\le i\le n\}|}{n}={p_{1}}.\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
</p>
</list-item>
</list>
</p></statement>
<p>We call <inline-formula id="j_nejsds73_ineq_025"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{0}}$]]></tex-math></alternatives></inline-formula> the <italic>probability</italic> of label 0. Conversely, <inline-formula id="j_nejsds73_ineq_026"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{1}}$]]></tex-math></alternatives></inline-formula> is the probability of label 1.<xref ref-type="fn" rid="j_nejsds73_fn_010">10</xref><fn id="j_nejsds73_fn_010"><label><sup>10</sup></label>
<p>Von Mises called <inline-formula id="j_nejsds73_ineq_027"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{0}}$]]></tex-math></alternatives></inline-formula> chance as long as <italic>x</italic> is a sequence. When <italic>x</italic> is a collective, he called it probability.</p></fn> The existence of the limit (Condition 1) is a non-vacuous condition. One can easily construct sequences whose frequencies do not converge [<xref ref-type="bibr" rid="j_nejsds73_ref_037">37</xref>].</p>
<p>The sequences <inline-formula id="j_nejsds73_ineq_028"><alternatives><mml:math>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="script">S</mml:mi></mml:math><tex-math><![CDATA[$s\in \mathcal{S}$]]></tex-math></alternatives></inline-formula> are called <italic>selection rules</italic>. A selection rule selects the <italic>j</italic>th element of <italic>x</italic> whenever <inline-formula id="j_nejsds73_ineq_029"><alternatives><mml:math>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$s(j)=1$]]></tex-math></alternatives></inline-formula>.<xref ref-type="fn" rid="j_nejsds73_fn_011">11</xref><fn id="j_nejsds73_fn_011"><label><sup>11</sup></label>
<p>We sweep under the carpet a substantial difference in Von Mises’ definition of selection rules and our definition. Von Mises allowed the selection rules to “see” the first <italic>n</italic> elements of the collective when deciding whether to choose the <inline-formula id="j_nejsds73_ineq_030"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$n+1$]]></tex-math></alternatives></inline-formula>th element [<xref ref-type="bibr" rid="j_nejsds73_ref_107">107</xref>, p. 9]. Our definition is more restrictive. We require the selected position to be determined before “seeing” the <italic>entire</italic> sequence. We focus on the ex ante nature of Von Mises randomness but neglect his recursive formalism.</p></fn> Informally, a collective (w.r.t <inline-formula id="j_nejsds73_ineq_031"><alternatives><mml:math>
<mml:mi mathvariant="script">S</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{S}$]]></tex-math></alternatives></inline-formula>) is a sequence which has invariant frequency limits with respect to all selection rules in <inline-formula id="j_nejsds73_ineq_032"><alternatives><mml:math>
<mml:mi mathvariant="script">S</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{S}$]]></tex-math></alternatives></inline-formula>. We call any selection rule which does not change the frequency limit of a collective <italic>admissible</italic>. This invariance property of collectives is often called “law of excluded gambling strategy” [<xref ref-type="bibr" rid="j_nejsds73_ref_107">107</xref>]. When thinking of a sequence of coin tosses, a gambler is not able to gain an advantage by just considering specific selected coin tosses. The probability of seeing “heads” or “tails” remains unchanged.</p>
<p>Von Mises introduced the “law of excluded gambling strategy” with the goal to define randomness of a collective [<xref ref-type="bibr" rid="j_nejsds73_ref_107">107</xref>, p. 8]. A collective is called <italic>random</italic> with respect to <inline-formula id="j_nejsds73_ineq_033"><alternatives><mml:math>
<mml:mi mathvariant="script">S</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{S}$]]></tex-math></alternatives></inline-formula>. Consequently, Von Mises integrated randomness and probability into one theory. In fact, admissibility of selection rules is equivalent to statistical independence in the sense of Von Mises. But it is defined with respect to collectives instead of selection rules.</p><statement id="j_nejsds73_stat_003"><label>Definition 3</label>
<title>(Von Mises’ Definition of Statistical Independence of Collectives [<xref ref-type="bibr" rid="j_nejsds73_ref_107">107</xref>, p. 30, Def. 2]).</title>
<p>A collective <italic>x</italic> with respect to <inline-formula id="j_nejsds73_ineq_034"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mathcal{S}_{x}}$]]></tex-math></alternatives></inline-formula> is called statistically independent to the collective <italic>y</italic> with respect to <inline-formula id="j_nejsds73_ineq_035"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mathcal{S}_{y}}$]]></tex-math></alternatives></inline-formula> iff the following limits exist and 
<disp-formula id="j_nejsds73_eq_004">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mspace width="2.5pt"/>
<mml:mtext>and</mml:mtext>
<mml:mspace width="2.5pt"/>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mspace width="1em"/>
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}& \underset{n\to \infty }{\lim }\frac{|\{i\in \mathbb{N}:x(i)=1\hspace{2.5pt}\text{and}\hspace{2.5pt}y(i)=1,1\le i\le n\}|}{|\{j\in \mathbb{N}:y(j)=1,1\le j\le n\}|}\\ {} & \hspace{1em}=\underset{n\to \infty }{\lim }\frac{|\{i\in \mathbb{N}:x(i)=1,1\le i\le n\}|}{n}={p_{1}},\end{aligned}\]]]></tex-math></alternatives>
</disp-formula> 
where <inline-formula id="j_nejsds73_ineq_036"><alternatives><mml:math><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>:</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[$\frac{0}{0}:=0$]]></tex-math></alternatives></inline-formula>. When two collectives are independent of each other we write 
<disp-formula id="j_nejsds73_eq_005">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo stretchy="false">⊥</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ x\perp y.\]]]></tex-math></alternatives>
</disp-formula>
</p></statement>
<p>In comparison to admissibility, the collective <italic>y</italic> adopts the role of a selection rule. It <italic>is</italic> in fact an admissible selection rule with the difference that a potentially finite number of elements in <italic>x</italic> are selected (cf. [<xref ref-type="bibr" rid="j_nejsds73_ref_041">41</xref>, p. 120f]). Conversely, Von Mises’ randomness is statistical independence with respect to sequences with infinitely many ones and potentially no frequency limit. (For a general comparison between Kolmogorov’s and Von Mises’ theory of probability see Appendix <xref rid="j_nejsds73_app_001">A</xref> and Table <xref rid="j_nejsds73_tab_002">2</xref>.)</p>
</sec>
<sec id="j_nejsds73_s_011">
<label>4.5</label>
<title>Kolmogorov’s Independence versus Von Mises’ Independence</title>
<p>What is the relationship between Kolmogorov’s and Von Mises’ definition of statistical independence? On a conceptual level, the critique posed earlier, which questioned the meaning of statistical independence between events following Kolmogorov, gets resolved.</p>
<p>Von Mises adopted a strong frequential perspective on probabilities which clarifies the mapping from real world to mathematical definition. He idealized repetitive observations by infinite sequences and defined probabilities as limiting frequencies.<xref ref-type="fn" rid="j_nejsds73_fn_012">12</xref><fn id="j_nejsds73_fn_012"><label><sup>12</sup></label>
<p>Without the idealization, again the mathematical description would miss a link to a worldly phenomenon. The idealization in terms of infinite sequences is substantial. In fact, the legitimacy of this idealization is the subject of another debate [<xref ref-type="bibr" rid="j_nejsds73_ref_045">45</xref>]. Nevertheless, the idealization taken in Von Mises’ framework is explicitly and transparently stated. Kolmogorov’s axioms do not possess such a statement.</p></fn> Von Mises’ independence states that there is no difference in counting the frequency of occurrences of an event in the entirety of the sequences or in a subselected sequence. His independence forbids any statistical interference between processes described as sequences. No statistical patterns can be derived from one sequence by leveraging the other. Von Mises’ definition formalizes the concept of statistical independence between processes of reoccuring events.</p>
<p>In contrast to Kolmogorov, Von Mises’ definition does not evoke the conceptual obscurity. His focus on idealized sequences of repetitive events restricts his definition of statistical independence to specific applications with the gain of clarity in the goal of the mathematical description. Von Mises’ definition of independence makes statistical independence more concrete then Kolmogorov’s definition does.</p>
<p>On a more formal level, Kolmogorov defined statistical independence via the factorization of measure (cf. Definition <xref rid="j_nejsds73_stat_006">5</xref>), whereas Von Mises defined statistical independence via conditionalization of measures. The invariance of the frequency limit of a collective with regard to the subselection via another collective can be interpreted as the invariance of a probability of an event with regard to the conditioning on another event, i.e., “selecting with respect to” is “conditioning on” (cf. Theorem <xref rid="j_nejsds73_stat_009">1</xref> and Theorem <xref rid="j_nejsds73_stat_011">2</xref>).</p>
<p>Mathematically, it turns out that Kolmogorov’s definition and Von Mises’ definition are both special cases (modulo the measure zero problem in conditioning) of a more general form of measure-theoretic statistical independence. A selection rule with converging frequency limit is admissible (respectively, statistically independent), to a collective if and only if the two are statistically independent of each other in the sense of Kolmogorov, when generalized to finitely additive probability spaces (see Appendix <xref rid="j_nejsds73_app_001">A</xref> for a formal statement of this claim). Thus, we can replace the known definition of statistical independence by Kolmogorov with the definition by Von Mises. Thereby, we give a specific meaning to statistical independence.</p>
<p>We have been motivated to dissect the notion of statistical independence for its central role in fair machine learning. Von Mises’ definition drew us closer to a more transparent mathematical formalization of statistical independence for fairness notions in machine learning. However, our discussion of Von Mises’ theory skipped over a substantial part of his work so far. Von Mises included a definition of randomness in his theory of probability. Much in contrast to Kolmogorov: There is no definition of “randomness” in Kolmogorov’s <italic>Grundbegriffe der Wahrscheinlichkeitsrechnung</italic> [<xref ref-type="bibr" rid="j_nejsds73_ref_055">55</xref>] (translated as [<xref ref-type="bibr" rid="j_nejsds73_ref_056">56</xref>]). Even more interestingly, Von Mises’ definition of randomness is stated in terms of statistical independence. The reader might notice that in Section <xref rid="j_nejsds73_s_002">2</xref> we already stumbled upon a heavily used notion of randomness in machine learning, which is expressed as statistical independence (i.i.d.). How do i.i.d. and Von Mises’ randomness relate to each other? How does the close connection between statistical independence and randomness complement our picture of the three fairness criteria from machine learning?</p>
</sec>
</sec>
<sec id="j_nejsds73_s_012">
<label>5</label>
<title>Randomness as Statistical Independence</title>
<p>The nature and definition of randomness seems as “random” as the term itself [<xref ref-type="bibr" rid="j_nejsds73_ref_035">35</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_073">73</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_068">68</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_103">103</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_008">8</xref>]. Usually, a very broad distinction between two approaches to randomness is made: process randomness versus outcome randomness [<xref ref-type="bibr" rid="j_nejsds73_ref_035">35</xref>]. In this work, we focus on outcome randomness and more specifically the role of randomness in statistics and machine learning.</p>
<p>Randomness is a modeling assumption in statistics (cf. Section <xref rid="j_nejsds73_s_002">2</xref>). Upon looking into statistics and machine learning textbooks one often finds the assumption of independent and identically distributed (i.i.d.) data points as the expression of randomness [<xref ref-type="bibr" rid="j_nejsds73_ref_018">18</xref>, p. 207], [<xref ref-type="bibr" rid="j_nejsds73_ref_030">30</xref>, p. 4].</p>
<p>We adopt Von Mises’ differing account of randomness. The expression of randomness <italic>relative</italic> to the problem at hand, particularly in settings with data models such as statistics, turns out to be substantial.</p>
<sec id="j_nejsds73_s_013">
<label>5.1</label>
<title>Orthogonal Perspectives on Randomness as Independence in Machine Learning and Statistics</title>
<p>Von Mises defined a random sequence as a sequence which is statistically independent to a (pre-specified) set of selection rules respectively other sequences. In contrast, an i.i.d.-sequence consists of elements each statistically independent to all others.</p>
<p>Both definitions are stated in terms of statistical independence. But, the relationship of independence and randomness in terms of i.i.d. and in Von Mises’ theory differ substantially. Von Mises’ randomness is stated relative with respect to a set of selection rules. Furthermore, it is stated between sequences, respectively collectives. Whereas, in an i.i.d. sequence randomness is expressed between random variables. The randomness definitions are in an abstract sense “orthogonal.” We consider a concrete example for better understanding.</p>
<def-list><def-item id="x1-13002x-1"><term>1.</term><def>
<p><bold>Horizontal Randomness.</bold> Let <inline-formula id="j_nejsds73_ineq_037"><alternatives><mml:math>
<mml:mi mathvariant="normal">Ω</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi></mml:math><tex-math><![CDATA[$\Omega =\mathbb{N}$]]></tex-math></alternatives></inline-formula> be a penguin colony. Let <inline-formula id="j_nejsds73_ineq_038"><alternatives><mml:math>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">f</mml:mi></mml:math><tex-math><![CDATA[$s,f$]]></tex-math></alternatives></inline-formula> be two attributes of a penguin, namely sex and whether a penguin has the penguin flu or not. Mathematically: <inline-formula id="j_nejsds73_ineq_039"><alternatives><mml:math>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="normal">Ω</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$s:\Omega \to \{0,1\}$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_nejsds73_ineq_040"><alternatives><mml:math>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="normal">Ω</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$f:\Omega \to \{0,1\}$]]></tex-math></alternatives></inline-formula>. So, penguins are individuals <inline-formula id="j_nejsds73_ineq_041"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="normal">Ω</mml:mi></mml:math><tex-math><![CDATA[$n\in \Omega $]]></tex-math></alternatives></inline-formula> which we do not know individually, but we know some attributes of them. Suppose we are given a sequence <inline-formula id="j_nejsds73_ineq_042"><alternatives><mml:math>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo></mml:math><tex-math><![CDATA[$f(1),f(2),f(3),\dots $]]></tex-math></alternatives></inline-formula> of flu values with existing frequency limit. This allows us to state randomness of <italic>f</italic> with respect to the corresponding sequence of sex values <italic>s</italic>, containing infinitely many ones and having a frequency limit, by: the sequence of sex values <italic>s</italic> is admissible on <italic>f</italic>. Respectively, <italic>s</italic> and <italic>f</italic> are statistically independent of each other. In the context of colony Ω a penguin having flu is random with respect to the sex of the penguin.</p></def></def-item><def-item id="x1-13004x-2"><term>2.</term><def>
<p><bold>Vertical Randomness.</bold> This is different to the i.i.d.-setting in which each penguin <inline-formula id="j_nejsds73_ineq_043"><alternatives><mml:math>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi></mml:math><tex-math><![CDATA[$i\in \mathbb{N}$]]></tex-math></alternatives></inline-formula> obtains its own random variable <inline-formula id="j_nejsds73_ineq_044"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="normal">Ω</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[${F_{i}}:\Omega \to \{0,1\}$]]></tex-math></alternatives></inline-formula> on some probability space <inline-formula id="j_nejsds73_ineq_045"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="normal">Ω</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="script">F</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\Omega ,\mathcal{F},P)$]]></tex-math></alternatives></inline-formula>. Here, <inline-formula id="j_nejsds73_ineq_046"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${F_{i}}$]]></tex-math></alternatives></inline-formula> encodes whether penguin <italic>i</italic> has the penguin flu or not. The sequence <inline-formula id="j_nejsds73_ineq_047"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo></mml:math><tex-math><![CDATA[${F_{1}},{F_{2}},{F_{3}},\dots $]]></tex-math></alternatives></inline-formula> somehow represents the colony. The included random variables share their distribution and are statistically independent to each other. The attribute flu is not random with respect to the attribute sex here, but the penguins are random with respect to each other. The random variables are (often implicitly) defined on a standard probability space on Ω. The set Ω here does <italic>not</italic> model the colony. It shrivels to an abstract source of randomness and probability.</p></def></def-item></def-list><p>The choice of perspective, horizontal or vertical, on randomness expressed as statistical independence is a question of the data model. The two types of randomness definitions are distinct in a number of ways. For a summary see Table <xref ref-type="sec" rid="j_nejsds73_tab_001">1</xref>. Most importantly, horizontal randomness is inherently expressed <italic>with respect to</italic>some mathematical object. Vertical randomness lacks this explicit relativization. This typification of horizontal and vertical, mathematical definitions of randomness is actually more broadly applicable.</p> 
<table-wrap id="j_nejsds73_tab_001">
<label>Table 1</label>
<caption>
<p>Typification of horizontal and vertical randomness (“RV” = “random variable”).</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Horizontal Randomness</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Vertical Randomness</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">Data points are modelled as:</td>
<td style="vertical-align: top; text-align: left">Evaluations of RVs</td>
<td style="vertical-align: top; text-align: left">RVs</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Mathematical definition of randomness of:</td>
<td style="vertical-align: top; text-align: left">Sequences</td>
<td style="vertical-align: top; text-align: left">Sequences of RVs</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">Explicit relativization:</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">Yes</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">No</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To the set of vertical randomness notions one can add: exchangeability [<xref ref-type="bibr" rid="j_nejsds73_ref_027">27</xref>], <italic>α</italic>-mixing, <italic>β</italic>-mixing [<xref ref-type="bibr" rid="j_nejsds73_ref_094">94</xref>] and possibly many more. The set of horizontal randomness notions is spanned up by an entire branch of computer science and mathematics: algorithmic randomness.</p>
<p>Algorithmic randomness poses the question whether a sequence is random or not. This question arose in [<xref ref-type="bibr" rid="j_nejsds73_ref_105">105</xref>] within the attempt to axiomatize probability theory [<xref ref-type="bibr" rid="j_nejsds73_ref_010">10</xref>, p. 3]. In algorithmic randomness further definitions of random sequences have been proposed. For the sake of simplicity the considered sequences consist only of zeros and ones.</p>
<p>Four intuitions for random sequences crystallized [<xref ref-type="bibr" rid="j_nejsds73_ref_077">77</xref>, p. 280ff]: typicality, incompressibility, unpredictability and independence (see Appendix <xref rid="j_nejsds73_app_003">C</xref>). For our purposes, the key point to note is that a random sequence is typical, incompressible, unpredictable or independent with respect to “something” (they are all relativised in some way). Each of these intuitions has been expressed in various mathematical terms. In particular, formalizations of the same intuitions are not necessarily equivalent, and formalizations of different intuitions sometimes coincide or are logically related (see Appendix <xref rid="j_nejsds73_app_004">D</xref>).<xref ref-type="fn" rid="j_nejsds73_fn_013">13</xref><fn id="j_nejsds73_fn_013"><label><sup>13</sup></label>
<p>For an overview of algorithmic randomness see [<xref ref-type="bibr" rid="j_nejsds73_ref_099">99</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_068">68</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_103">103</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_031">31</xref>].</p></fn> We mainly stick to the intuition of independence in this paper. A random sequence is independent of “some” other sequences [<xref ref-type="bibr" rid="j_nejsds73_ref_105">105</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_023">23</xref>].</p>
</sec>
<sec id="j_nejsds73_s_014">
<label>5.2</label>
<title>Relative Randomness Instead of Absolute, Universal Randomness</title>
<p>The definition of randomness for sequences is inherently relative. Even though, the notion is relative with respect to “something,” most of the effort has been spent on finding <italic>the</italic> set of statistically independent sequences defining randomness [<xref ref-type="bibr" rid="j_nejsds73_ref_023">23</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_068">68</xref>].<xref ref-type="fn" rid="j_nejsds73_fn_014">14</xref><fn id="j_nejsds73_fn_014"><label><sup>14</sup></label>
<p>The analogous observation holds for all four intuitions (see Appendix <xref rid="j_nejsds73_app_003">C</xref> and [<xref ref-type="bibr" rid="j_nejsds73_ref_068">68</xref>]).</p></fn></p>
<p>Naively, one could attempt to define a random sequence as: a sequence is random if and only if it is independent with respect to <italic>all</italic> sequences. However, this approach is doomed to fail. There is no sequence fulfilling this condition except for trivial ones such as endless repetitions of zeros or ones (see Kamke’s critique of von Mises’ notion of randomness [<xref ref-type="bibr" rid="j_nejsds73_ref_100">100</xref>]).</p>
<p>So instead, research focused on computability expressed in various ways (because it was felt by those investigating these matters that computability was somehow given, or more primitive, and thus a natural way to resolve the relativity of the notion of randomness). Intuitively, randomness is considered the antithesis of computability [<xref ref-type="bibr" rid="j_nejsds73_ref_077">77</xref>, p. 288]: something which is computable is not random. Something which is random is not computable. If we then informally update the definition above we obtain: a sequence is random if and only if it is independent with respect to all computable sequences [<xref ref-type="bibr" rid="j_nejsds73_ref_023">23</xref>].<xref ref-type="fn" rid="j_nejsds73_fn_015">15</xref><fn id="j_nejsds73_fn_015"><label><sup>15</sup></label>
<p>To be precise, a random sequence following Church [<xref ref-type="bibr" rid="j_nejsds73_ref_023">23</xref>] is independent to all <italic>partially computable</italic> selection rules following Von Mises (see Footnote <xref rid="j_nejsds73_fn_011">11</xref>).</p></fn> Analogous to <italic>the</italic> definition of computability [<xref ref-type="bibr" rid="j_nejsds73_ref_077">77</xref>, p. 165], this is taken as an argument for the existence of <italic>the</italic> definition of randomness [<xref ref-type="bibr" rid="j_nejsds73_ref_077">77</xref>, p. 287].</p>
<p>In our work, we argue towards a relativized conception of randomness in line with work by [<xref ref-type="bibr" rid="j_nejsds73_ref_077">77</xref>], [<xref ref-type="bibr" rid="j_nejsds73_ref_050">50</xref>] and [<xref ref-type="bibr" rid="j_nejsds73_ref_107">107</xref>].<xref ref-type="fn" rid="j_nejsds73_fn_016">16</xref><fn id="j_nejsds73_fn_016"><label><sup>16</sup></label>
<p>Humphreys [<xref ref-type="bibr" rid="j_nejsds73_ref_050">50</xref>] presented randomness as relativized to a probabilistic hypothesis or reference class. Porter [<xref ref-type="bibr" rid="j_nejsds73_ref_077">77</xref>, p. 169] even postulated the “No-Thesis”-Thesis: Any notion of randomness neither defines a well-defined collection of random sequences nor captures all mathematical conceptions of randomness; confer the logic of “essentially contested” concepts [<xref ref-type="bibr" rid="j_nejsds73_ref_039">39</xref>], which, presumably, are unavoidably contested for the same reason.</p></fn> A <italic>relative</italic> definition of randomness is a definition of randomness which is <italic>relative</italic> with respect to the problem under consideration.<xref ref-type="fn" rid="j_nejsds73_fn_017">17</xref><fn id="j_nejsds73_fn_017"><label><sup>17</sup></label>
<p>In machine learning, a problem under consideration is, for instance, animal classification via neural networks.</p></fn> In contrast, an <italic>absolute and universal</italic> definition of randomness would preserve its validity in all problems. It presupposes the existence of <italic>the</italic> randomness.</p>
<p><italic>Relative</italic> randomness with respect to the problem which we want to describe aligns to Von Mises’ theory of probability and randomness. Von Mises emphasized the <italic>ex ante</italic> choice of randomness [<xref ref-type="bibr" rid="j_nejsds73_ref_106">106</xref>, p. 89] <italic>relative</italic> to the problem at hand [<xref ref-type="bibr" rid="j_nejsds73_ref_107">107</xref>, p. 12]. First, one formalizes randomness with respect to the underlying problem, then one can consider a sequence to be random or not. Otherwise, if we are given a sequence, it is easy to construct a set of selection rules, such that the sequence is random with respect to this set.<xref ref-type="fn" rid="j_nejsds73_fn_018">18</xref><fn id="j_nejsds73_fn_018"><label><sup>18</sup></label>
<p>This idea is transferable to other intuitions of randomness, for instance [<xref ref-type="bibr" rid="j_nejsds73_ref_102">102</xref>].</p></fn> This, however, undermines the concept of randomness, which should capture the pre-existing typicality, incompressibility, unpredictability or independence of a sequence (cf. [<xref ref-type="bibr" rid="j_nejsds73_ref_108">108</xref>, p. 321]). Von Mises’ randomness intrinsically possesses a modeling character, similar to our needs in machine learning and statistics.</p>
<p>Given its role as modeling assumption in statistics, randomness lacks substantial justification to be expressed in any <italic>absolute and universal</italic> manner in this context. Neither are there reasons why computability<xref ref-type="fn" rid="j_nejsds73_fn_019">19</xref><fn id="j_nejsds73_fn_019"><label><sup>19</sup></label>
<p>Computability is often taken (at least by computer scientists) as a purely mathematical notion, detached from the world. An alternate view, close in spirit to Von Mises, is that computation is part of physics, and thus needs to be viewed in a <italic>scientific</italic>, and not merely <italic>mathematical</italic> manner [<xref ref-type="bibr" rid="j_nejsds73_ref_029">29</xref>].</p></fn> is the only mathematical, expressive way to encode one of the four intuitions of randomness. The i.i.d. assumption, an <italic>absolute and universal</italic> definition of randomness, does not fit this purpose. To appropriately model data we require adjustable notions of randomness. Otherwise, we restrict our modeling choice without reason or gain.<xref ref-type="fn" rid="j_nejsds73_fn_020">20</xref><fn id="j_nejsds73_fn_020"><label><sup>20</sup></label>
<p>This is not entirely true, as specific computability notions of randomness and the i.i.d.-assumption deliver convergence results for random sequences, which can be used to guarantee low estimation error in the long run.</p></fn></p>
<p>Equipped with the interpretation of statistical independence as randomness we now return to our motivation for investigating statistical independence. ML-friendly fairness criteria are built upon statistical independence. In contrast to Kolmogorov, Von Mises’ statistical independence transparently refers to a concept of independence in the real world. To clarify the meaning of fairness expressed as statistical independence, we directly apply Von Mises’ independence to the fairness criteria listed in Section <xref rid="j_nejsds73_s_003">3</xref> in the following.</p>
</sec>
</sec>
<sec id="j_nejsds73_s_015">
<label>6</label>
<title>Von Mises’ Fairness</title>
<p>With Von Mises’ definition of statistical independence we have a notion at our disposal which is conceptually focused on a more “scientific” perspective (i.e., making claims about the world) of statistical concepts. Since it is mathematically related to Kolmogorov’s standard account of statistical independence, Kolmogorov’s definition can, at many places, be easily replaced by Von Mises’ definition.</p>
<p>Let us denote the three presented fairness criteria in a Von Mises’ way (cf. Section <xref rid="j_nejsds73_s_011">4.5</xref>).</p><statement id="j_nejsds73_stat_004"><label>Definition 4</label>
<title>(Fairness as Statistical Independence).</title>
<p>A collective <inline-formula id="j_nejsds73_ineq_048"><alternatives><mml:math>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$x:\mathbb{N}\to \{0,1\}$]]></tex-math></alternatives></inline-formula> (with respect to a family of selection rules <inline-formula id="j_nejsds73_ineq_049"><alternatives><mml:math>
<mml:mi mathvariant="script">S</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{S}$]]></tex-math></alternatives></inline-formula>) is <italic>fair with respect to a set of sensitive groups</italic> <inline-formula id="j_nejsds73_ineq_050"><alternatives><mml:math>
<mml:mi mathvariant="script">G</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">J</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$\mathcal{G}=\{{s^{j}}:\mathbb{N}\to \{0,1\}|j\in J\}$]]></tex-math></alternatives></inline-formula> if 
<disp-formula id="j_nejsds73_eq_006">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo stretchy="false">⊥</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mspace width="2.5pt"/>
<mml:mspace width="2.5pt"/>
<mml:mo>∀</mml:mo>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ x\perp {s^{j}}\hspace{2.5pt}\hspace{2.5pt}\forall j\in J\]]]></tex-math></alternatives>
</disp-formula>
</p></statement>
<p>The 0-1-sequences <inline-formula id="j_nejsds73_ineq_051"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${s^{j}}$]]></tex-math></alternatives></inline-formula> determine for each individual <italic>i</italic> whether it is part of the group or not (according to whether <inline-formula id="j_nejsds73_ineq_052"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[${s^{j}}(i)=1$]]></tex-math></alternatives></inline-formula> or <inline-formula id="j_nejsds73_ineq_053"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[${s^{j}}(i)=0$]]></tex-math></alternatives></inline-formula>). We call these groups “sensitive,” as these are the groups which are of moral and ethical concern. In philosophical literature these groups are often called “socially salient groups” [<xref ref-type="bibr" rid="j_nejsds73_ref_002">2</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_064">64</xref>].<xref ref-type="fn" rid="j_nejsds73_fn_021">21</xref><fn id="j_nejsds73_fn_021"><label><sup>21</sup></label>
<p>Intersections of sensitive groups are not necessarily independent.</p></fn> We see that the connection between Von Misean independence and fairness arises from the observation that the set of sensitive groups <inline-formula id="j_nejsds73_ineq_054"><alternatives><mml:math>
<mml:mi mathvariant="script">G</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{G}$]]></tex-math></alternatives></inline-formula> <italic>is</italic> a family of selection rules, so that if <inline-formula id="j_nejsds73_ineq_055"><alternatives><mml:math>
<mml:mi mathvariant="script">G</mml:mi>
<mml:mo stretchy="false">⊆</mml:mo>
<mml:mi mathvariant="script">S</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{G}\subseteq \mathcal{S}$]]></tex-math></alternatives></inline-formula>, then indeed the collective <italic>x</italic> will be fair for <inline-formula id="j_nejsds73_ineq_056"><alternatives><mml:math>
<mml:mi mathvariant="script">G</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{G}$]]></tex-math></alternatives></inline-formula>.</p>
<p>Following Von Mises’ interpretation of independence, the given definition reads as follows: we assume we are in the idealized setting of infinitely many individuals with values <inline-formula id="j_nejsds73_ineq_057"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${x_{i}}$]]></tex-math></alternatives></inline-formula>, e.g., binary predictions. The predictions are fair if and only if there is no difference in counting the frequency of 1-predictions in the entirety or in the sensitive group. (For an illustration see Appendix <xref rid="j_nejsds73_fig_006">6</xref>.) A proper conceptualization of fairness requires such immediate semantics, but a purely mathematical theory of probability cannot offer these (see Section <xref rid="j_nejsds73_s_008">4.2</xref>).</p>
<p>Each of the three fairness criteria is captured in Definition <xref rid="j_nejsds73_stat_004">4</xref>; the choice of fairness criterion manifests in the collective under consideration: 
<def-list><def-item id="x1-15002x-6"><term>Independence</term><def>
<p>The collective <inline-formula id="j_nejsds73_ineq_058"><alternatives><mml:math>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$x:\mathbb{N}\to \{0,1\}$]]></tex-math></alternatives></inline-formula> consists of predictions; i.e., <inline-formula id="j_nejsds73_ineq_059"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$\{0,1\}$]]></tex-math></alternatives></inline-formula> is the set of predictions.</p></def></def-item><def-item id="x1-15003x-6"><term>Separation</term><def>
<p>The collective <inline-formula id="j_nejsds73_ineq_060"><alternatives><mml:math>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$x:\mathbb{N}\to \{0,1\}$]]></tex-math></alternatives></inline-formula> is obtained via the subselection of predictions based on the sequence of true labels corresponding to the predictions.<xref ref-type="fn" rid="j_nejsds73_fn_022">22</xref><fn id="j_nejsds73_fn_022"><label><sup>22</sup></label>
<p>“Selecting with respect to” is “conditioning on.”</p></fn></p></def></def-item><def-item id="x1-15004x-6"><term>Sufficiency</term><def>
<p>The true labels are subselected by predictions.</p></def></def-item></def-list> To enable intuitive access to the Von Mises’ notions of fairness we provide a toy example in Appendix <xref rid="j_nejsds73_app_002">B</xref>. The three fairness criteria Independence, Separation and Sufficiency encompass a large part of fair machine learning [<xref ref-type="bibr" rid="j_nejsds73_ref_005">5</xref>, p. 45]. Von Mises’ statistical independence gives a consistent interpretation to all of them. In fact, Von Mises’ independence opens the door to further investigations. To this end, we recapitulate the strong linkage between statistical independence and randomness in Von Mises’ theory.</p>
</sec>
<sec id="j_nejsds73_s_016">
<label>7</label>
<title>The Ethical Implications of Modeling Assumptions</title>
<p>Machine learning methods try to model data in complex ways. Derived statements, such as predictions, then potentially get applied in society. In these cases one is obliged to ask which ought-state the machine learning model should reflect [<xref ref-type="bibr" rid="j_nejsds73_ref_075">75</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_083">83</xref>]. To enable a justified choice, statistical concepts in machine learning require relations to the real world. Furthermore, modeling even requires understanding of the entanglement of societal and statistical concepts.</p>
<p>We proposed one specific <italic>meaningful</italic> definition of statistical independence which can be directly applied to the three observational fairness criteria from fair machine learning. In addition, this Von Mises’ independence is key to a relativized notion of randomness. Pulling these threads together, we are now able to establish the following link: <italic>Randomness is fairness. Fairness is randomness.</italic></p>
<sec id="j_nejsds73_s_017">
<label>7.1</label>
<title>Randomness is Fairness. Fairness is Randomness</title>
<p>The concepts fairness and randomness frequently appear jointly: [<xref ref-type="bibr" rid="j_nejsds73_ref_014">14</xref>] argues that a random allocation of goods is fair under certain conditions. Literature on sortition argues for just representation of society by random selection of people [<xref ref-type="bibr" rid="j_nejsds73_ref_074">74</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_095">95</xref>].<xref ref-type="fn" rid="j_nejsds73_fn_023">23</xref><fn id="j_nejsds73_fn_023"><label><sup>23</sup></label>
<p>Representativity here can be interpreted as typicality, thus one of the four intuitions for randomness.</p></fn> Bennett [<xref ref-type="bibr" rid="j_nejsds73_ref_007">7</xref>, p. 633] even states that randomness encompasses fairness.</p>
<p>With Von Mises’ axiom 2 and Definition <xref rid="j_nejsds73_stat_004">4</xref> we can now tighten the conceptual relationship of fairness and randomness. The proposition directly follows from the definition of randomness respectively fairness in the sense of Von Mises.</p><statement id="j_nejsds73_stat_005"><label>Proposition 1</label>
<title>(Randomness is fairness. Fairness is randomness.).</title>
<p><italic>Let x be a collective with respect to ∅ (the empty set). It is fair with respect to a set of sensitive groups (0-1-sequences)</italic> <inline-formula id="j_nejsds73_ineq_061"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo fence="true" stretchy="false">}</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\{{s^{j}}\}_{j\in J}}$]]></tex-math></alternatives></inline-formula><italic>, if and only if it is random with respect to</italic> <inline-formula id="j_nejsds73_ineq_062"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo fence="true" stretchy="false">}</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">J</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\{{s^{j}}\}_{j\in J}}$]]></tex-math></alternatives></inline-formula><italic>.</italic></p></statement>
<p>The given proposition establishes a helpful link. It gives insights into both of the concepts. In particular, it substantiates the relativized conception of randomness in machine learning as it presents randomness as an ethical choice.</p>
<sec id="j_nejsds73_s_018">
<label>7.1.1</label>
<title>Randomness as Ethical Choice</title>
<p>Randomness in machine learning is a modeling assumption (Section <xref rid="j_nejsds73_s_002">2</xref>). Fairness is an ethical choice.<xref ref-type="fn" rid="j_nejsds73_fn_024">24</xref><fn id="j_nejsds73_fn_024"><label><sup>24</sup></label>
<p>More specifically, the choice of operationalized fairness, one of the fairness criteria, and the choice of groups.</p></fn> In light of Proposition <xref rid="j_nejsds73_stat_005">1</xref> randomness gets an ethical choice and fairness a modeling assumption. We now further detail this perspective.</p>
<p>We assume that we are given a fixed set of selection rules, which defines “the” randomness. As far-fetched as this may sound, if we, for example, accept the so called Martin-Löf randomness as <italic>absolute and universal</italic> definition, then we exactly do this and fix the set of selection rules to the partial computable ones (see Appendix <xref rid="j_nejsds73_s_027">D.1</xref>). A sequence which is random with respect to this specified set of selection rules is fair with respect to the groups defined by the selection rules. Rephrased in terms of Martin-Löf randomness: a Martin-Löf random sequence is fair with respect to all partial computable groups. Only non-partial-computable groups (respectively sequences) can be discriminated against in this setting. If we interpret statistical independence as fairness (Section <xref rid="j_nejsds73_s_003">3</xref>), then fairness is as <italic>absolute and universal</italic> as randomness here. Where did the “essentially contested” nature of fairness [<xref ref-type="bibr" rid="j_nejsds73_ref_039">39</xref>] leave the picture?</p>
<p>The set of admissible selection rules specifies the choice of sensitive groups, which indeed is a fraught and contestable choice [<xref ref-type="bibr" rid="j_nejsds73_ref_067">67</xref>, Section H.3]. Thus each selection rule gets ethically loaded. Furthermore, the choice of collective, which we consider as random, fixes the fairness criterion. In summary, the determination of randomness is analogous to the determination of fairness.</p>
<p>However one defines randomness, it is an ethical choice. For symmetry reasons one can equivalently state in machine learning: fairness is a modeling assumption. The randomness assumption has an ethical, moral and potentially legal implication. We need non-mathematical, contextual arguments to each problem at hand which justify the adjustable and explicit randomness assumptions.</p>
<p>Given that randomness is an ethical choice, an absolute, universal conception of randomness counteracts any ethical debate in machine learning. Discussions about sexism, racism and other kinds of discrimination and injustice persist over time without ever arrogating the discovery of “the” fairness [<xref ref-type="bibr" rid="j_nejsds73_ref_039">39</xref>]. But if “the” randomness as statistical independence would exist, then “the” fairness as statistical independence would be an accessible notion. For illustration, we reconsider Martin-Löf randomness. A Martin-Löf random sequence is independent, respectively fair, to the set of all partial computable selection rules. But, it is completely unclear what the ethical meaning of partial computable groups is. And, it remains unsolved whether the groups given by gender are partial computable, when we desire to be fair with respect to them. We conceive Proposition <xref rid="j_nejsds73_stat_005">1</xref> as further counterargument to an absolute, universal definition of randomness. Randomness is, like fairness, better interpreted as a <italic>relative</italic> notion.</p>
<p>Further concluding, the equivalence of randomness and fairness highlights the deficiency of fairness notions in machine learning. The equivalence only holds due to the very reductionist perspective on fairness in fair machine learning. Despite their regular co-occurrence [<xref ref-type="bibr" rid="j_nejsds73_ref_014">14</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_074">74</xref>], [<xref ref-type="bibr" rid="j_nejsds73_ref_007">7</xref>, p. 633], fairness and randomness are more multi-facetted and non-overlapping concepts as illustrated in Proposition <xref rid="j_nejsds73_stat_005">1</xref>.</p>
</sec>
<sec id="j_nejsds73_s_019">
<label>7.1.2</label>
<title>Fairy Tales of Fairness: “Perfectly Fair” Data</title>
<p>With the relationship between fairness and randomness in mind, we now turn towards random data as primitive. Discussions in fair machine learning sometimes seemingly presume the existence of “perfectly fair” data (e.g., as highlighted in [<xref ref-type="bibr" rid="j_nejsds73_ref_083">83</xref>, p. 134]), as if fair machine learning merely tackles the cases where “perfectly fair” data is not available.</p>
<p>We interpret “perfectly fair” data as a collective with respect to all possible selection rules. The data does not depend on any (sensitive) group at all. In other words, “perfectly fair” data is “totally random” data. As we saw in Section <xref rid="j_nejsds73_s_014">5.2</xref> this is self-contradicting except of the trivial constant case. “Perfectly fair” data does not exist or is statistically useless.</p>
</sec>
</sec>
<sec id="j_nejsds73_s_020">
<label>7.2</label>
<title>Demanding Fairness Is Randomization: Fair Predictors Are Randomizers</title>
<p>In practice, it is often unreasonable to <italic>assume</italic> random or fair data as in Proposition <xref rid="j_nejsds73_stat_005">1</xref>. Instead one <italic>demands</italic> for fairness respectively randomness of predictions. In these settings, fair machine learning techniques are deployed to exhibit ex post fulfillment of fairness criteria.</p>
<p>We assume for the following discussion that the collective <italic>x</italic> consists of predictions, as in the fairness criteria Independence or Separation. Fair machine learning techniques enforce statistical independence of predictions and sensitive attributes. Rephrased, fair machine learning techniques actually introduce randomness post-hoc into the predictions. Thus, fair machine learning techniques can potentially be interpreted as randomization techniques.</p>
<sec id="j_nejsds73_s_021">
<label>7.2.1</label>
<title>Fairness-Accuracy Trade-Off — Another Perspective</title>
<p>We noticed that fair predictions are random predictions with respect to the sensitive attribute. In contrast, accurate predictions exploit all dependencies between given attributes and predictive goal, including the sensitive attributes. Thus, in fair machine learning morally wrongful discriminative potential of sensitive attributes is thrown away by purpose. On these grounds, it is not surprising that an increase in fairness respectively randomness (usually) goes hand in hand with a decrease in accuracy [<xref ref-type="bibr" rid="j_nejsds73_ref_110">110</xref>]. Randomization of predictions leads to the so called fairness-accuracy trade-off.</p>
<p>Concluding, via Von Mises’ axiomatization we established: <italic>Randomness is fairness. Fairness is randomness</italic>. Exploiting this new perspective, we unlock another perspective on fair predictors as randomizers, demonstrate the nonexistence of “perfectly fair” data and treat randomness as an ethical choice, which can be neither universal nor total. In particular, the “essentially contested” nature of fairness is tied to the “essentially <italic>relative</italic>” nature of randomness.</p>
</sec>
</sec>
</sec>
<sec id="j_nejsds73_s_022">
<label>8</label>
<title>Conclusion</title>
<p>Fair machine learning attained an increasing interest in the last years. However, its conceptual maturity lags behind. In particular, the interplay between data, its mathematical representation and their relation to fairness is encompassed by a veil of nescience. In this paper, we contribute towards a better understanding of randomness and fairness in machine learning.</p>
<p>We started from the most commonly used definition of statistical independence and questioned its representation due to a lack of semantics. Generally, we observe that in machine learning, as in statistics, probability and its related concepts should be interpreted as modeling assumptions about the world (of data). Von Mises aimed for exactly this “scientific” perspective on probability theory. We lean on his statistical independence, which clarifies the relation to the real world, and his definition of randomness, which is <italic>relative</italic> and orthogonal to the i.i.d. assumption, but similarly expressed as statistical independence. Then by the three fairness criteria in machine learning we obtain a further interpretation of independence, which we finally exploit to argue for a <italic>relative</italic> conception of randomness, randomness as an ethical choice in machine learning and fair predictors as randomizers.</p>
<sec id="j_nejsds73_s_023">
<label>8.1</label>
<title>Future Work: Approximate Randomness and Fairness, Randomness as Fairness via Calibration</title>
<p>Despite future conclusions in-between the topics fairness and randomness in other research subjects as machine learning, we claim that a significant dimension is missing in the present discussion. Practitioners usually deal with approximate versions of randomness, statistical independence or fairness. Yet, approximation spans another dimension of choice beset with pitfalls [<xref ref-type="bibr" rid="j_nejsds73_ref_079">79</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_063">63</xref>]. Several questions ranging from the choice of approximation to the interference of concepts arise. Future work should detail the implications of this choice.</p>
<p>Second, we conjecture that “<italic>Randomness is Fairness. Fairness is randomness.</italic>” can be substantiated via the intuition of unpredictability. Starting from [<xref ref-type="bibr" rid="j_nejsds73_ref_090">90</xref>] definition of unpredictability randomness, which is closely related to the calibration idea presented in [<xref ref-type="bibr" rid="j_nejsds73_ref_025">25</xref>], we can bridge to fairness as calibration as given in [<xref ref-type="bibr" rid="j_nejsds73_ref_022">22</xref>]. A recent work by Cynthia Dwork and collaborators in fact show a formal link between pseudo-randomness and fairness as calibration [<xref ref-type="bibr" rid="j_nejsds73_ref_034">34</xref>]. This work, however, still misses a more thorough discussion of the concepts of individual versus group fairness in machine learning [<xref ref-type="bibr" rid="j_nejsds73_ref_012">12</xref>]. As a subproblem, which is contained therein, the categorization into (sensitive) groups in fair machine learning deserves its own work.</p>
<p>Third, regarding a more thorough definition of statistical independence within the fairness criteria, we are convinced that a subjectivist interpretation of probability might reveal yet another perspective on the problem. We assume that the interplay between different interpretations of probability and ethical concepts such as fairness still leaves room for many important investigations.</p>
<p>Fourth, there are certainly more frameworks to give an interpretation and concretization to current notions in (fair) machine learning (cf. [<xref ref-type="bibr" rid="j_nejsds73_ref_041">41</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_036">36</xref>]).</p>
<p>Last but not least, we already referred to sortition literature and random allocation. The somewhat different relation between fairness and randomness in this literature leads us to speculate that further fruitful discussions between the two concepts may develop.</p>
<p>In the jungle of statistical concepts such as probability, uncertainty, randomness, independence etc. further relations to social and ethical concepts wait to be brought to light. And machine learning research should care:</p><disp-quote>
<p>The arguments that justify inference from a sample to a population should explicitly refer to the variety of non-mathematical considerations involved. [<xref ref-type="bibr" rid="j_nejsds73_ref_006">6</xref>, p. 11]</p></disp-quote>
</sec>
</sec>
</body>
<back>
<app-group>
<app id="j_nejsds73_app_001"><label>Appendix A</label>
<title>Generalized Von Misesean Probability Theory</title>
<p>In this appendix we outline a theory of probability subsuming that of Kolmogorov and Von Mises.</p>
<sec id="j_nejsds73_s_024">
<label>A.1</label>
<title>Kolmogorov’s Notion of Independence</title>
<p>Kolmogorov axiomatized probability theory in his book [<xref ref-type="bibr" rid="j_nejsds73_ref_056">56</xref>] in a measure theoretical way. He defined a probability space <inline-formula id="j_nejsds73_ineq_063"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="normal">Ω</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="script">F</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\Omega ,\mathcal{F},P)$]]></tex-math></alternatives></inline-formula> as a measure space with base set Ω, <italic>σ</italic>-algebra <inline-formula id="j_nejsds73_ineq_064"><alternatives><mml:math>
<mml:mi mathvariant="script">F</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{F}$]]></tex-math></alternatives></inline-formula> and normalized measure <italic>P</italic>. Events are elements of <inline-formula id="j_nejsds73_ineq_065"><alternatives><mml:math>
<mml:mi mathvariant="script">F</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{F}$]]></tex-math></alternatives></inline-formula>, i.e., subsets of Ω, which obtain a probability via <italic>P</italic>. Statistical independence is defined as a specific assignment of probability to an intersection event.</p><statement id="j_nejsds73_stat_006"><label>Definition 5</label>
<title>(Kolmogorov’s Definition of Statistical Independence of Events [<xref ref-type="bibr" rid="j_nejsds73_ref_056">56</xref>, p. 9, Def. 1]).</title>
<p>Let <inline-formula id="j_nejsds73_ineq_066"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="normal">Ω</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="script">F</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\Omega ,\mathcal{F},P)$]]></tex-math></alternatives></inline-formula> be a probability space. Two events <inline-formula id="j_nejsds73_ineq_067"><alternatives><mml:math>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="script">F</mml:mi></mml:math><tex-math><![CDATA[$A,B\in \mathcal{F}$]]></tex-math></alternatives></inline-formula> are called statistically independent iff 
<disp-formula id="j_nejsds73_eq_007">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo>∩</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ P(A\cap B)=P(A)P(B).\]]]></tex-math></alternatives>
</disp-formula>
</p></statement>
<p>As we highlighted above, Kolmogorov’s axiomatization is, despite its success, not the only mathematical theory of probability. Specifically, one can weaken the structure of the probability space and still work with concepts such as statistical independence, expectation, conditioning (e.g., [<xref ref-type="bibr" rid="j_nejsds73_ref_043">43</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_021">21</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_070">70</xref>]).</p>
</sec>
<sec id="j_nejsds73_s_025">
<label>A.2</label>
<title>Finitely Additive Probability Space</title>
<p>We introduce a weaker measure structure, which we call finitely additive probability space. Interestingly, this weaker structure includes the axiomatization of Kolmogorov and Von Mises as special cases. We define a finitely additive probability space modified from [<xref ref-type="bibr" rid="j_nejsds73_ref_081">81</xref>, Def. 2.1.1 (7)]) as</p><statement id="j_nejsds73_stat_007"><label>Definition 6</label>
<title>(Finitely Additive Probability Space).</title>
<p>The tuple <inline-formula id="j_nejsds73_ineq_068"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">N</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="script">A</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">ν</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(N,\mathcal{A},\nu )$]]></tex-math></alternatives></inline-formula> is called a <italic>finitely additive probability space</italic> for a base set <italic>N</italic>, a set of measurable sets <inline-formula id="j_nejsds73_ineq_069"><alternatives><mml:math>
<mml:mi mathvariant="script">A</mml:mi>
<mml:mo stretchy="false">⊂</mml:mo>
<mml:mi mathvariant="script">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">N</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\mathcal{A}\subset \mathcal{P}(N)$]]></tex-math></alternatives></inline-formula> containing the empty set <inline-formula id="j_nejsds73_ineq_070"><alternatives><mml:math>
<mml:mi>∅</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="script">A</mml:mi></mml:math><tex-math><![CDATA[$\varnothing \in \mathcal{A}$]]></tex-math></alternatives></inline-formula> and a finitely additive probability measure <inline-formula id="j_nejsds73_ineq_071"><alternatives><mml:math>
<mml:mi mathvariant="italic">ν</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="script">A</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$\nu :\mathcal{A}\to [0,1]$]]></tex-math></alternatives></inline-formula> satisfying the following conditions: 
<list>
<list-item id="j_nejsds73_li_003">
<label>(1)</label>
<p><inline-formula id="j_nejsds73_ineq_072"><alternatives><mml:math>
<mml:mi mathvariant="italic">ν</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi>∅</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[$\nu (\varnothing )=0$]]></tex-math></alternatives></inline-formula>,</p>
</list-item>
<list-item id="j_nejsds73_li_004">
<label>(2)</label>
<p>if <inline-formula id="j_nejsds73_ineq_073"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>∪</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="script">A</mml:mi></mml:math><tex-math><![CDATA[${A_{1}},{A_{2}},{A_{1}}\cup {A_{2}}\in \mathcal{A}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds73_ineq_074"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mi>∅</mml:mi></mml:math><tex-math><![CDATA[${A_{1}}\cap {A_{2}}=\varnothing $]]></tex-math></alternatives></inline-formula> then <inline-formula id="j_nejsds73_ineq_075"><alternatives><mml:math>
<mml:mi mathvariant="italic">ν</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>∪</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">ν</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">ν</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\nu ({A_{1}}\cup {A_{2}})=\nu ({A_{1}})+\nu ({A_{2}})$]]></tex-math></alternatives></inline-formula>.</p>
</list-item>
</list>
</p></statement>
<p>Observe that this definition does not impose any structural restrictions on the set of subsets <inline-formula id="j_nejsds73_ineq_076"><alternatives><mml:math>
<mml:mi mathvariant="script">A</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{A}$]]></tex-math></alternatives></inline-formula>.</p>
<p>Kolmogorov’s probability space is certainly a specific finitely additive probability space in our sense, as every set <italic>σ</italic>-algebra contains the empty set and every countably additive probability is finitely additive.</p>
<p>Analogously, Von Mises implicitly uses a finitely additive probability space. This space is given by <inline-formula id="j_nejsds73_ineq_077"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\mathbb{N},{\mathcal{A}_{\mathrm{vM}}},{\nu _{\mathrm{vM}}})$]]></tex-math></alternatives></inline-formula>, where <inline-formula id="j_nejsds73_ineq_078"><alternatives><mml:math>
<mml:mi mathvariant="double-struck">N</mml:mi></mml:math><tex-math><![CDATA[$\mathbb{N}$]]></tex-math></alternatives></inline-formula> are the natural numbers and <inline-formula id="j_nejsds73_ineq_079"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mathcal{A}_{\mathrm{vM}}}$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_nejsds73_ineq_080"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\nu _{\mathrm{vM}}}$]]></tex-math></alternatives></inline-formula> are defined in the following.</p>
<p>First, we consider the finitely additive base measure 
<disp-formula id="j_nejsds73_eq_008">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>:</mml:mo>
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\nu _{\mathrm{vM}}}(A):=\underset{n\to \infty }{\lim }\frac{|A\cap {\mathbb{N}_{n}}|}{n},\]]]></tex-math></alternatives>
</disp-formula> 
where <inline-formula id="j_nejsds73_ineq_081"><alternatives><mml:math>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo stretchy="false">⊂</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi></mml:math><tex-math><![CDATA[$A\subset \mathbb{N}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds73_ineq_082"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[${\mathbb{N}_{n}}=\{1,\dots ,n\}$]]></tex-math></alternatives></inline-formula>. <inline-formula id="j_nejsds73_ineq_083"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\nu _{\mathrm{vM}}}$]]></tex-math></alternatives></inline-formula> is called the “natural density” in the number theory literature [<xref ref-type="bibr" rid="j_nejsds73_ref_071">71</xref>, p. 256]. From this definition is not clear whether the given limit exists. Thus, we define the set of measurable sets by 
<disp-formula id="j_nejsds73_eq_009">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>:</mml:mo>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo>:</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mspace width="2.5pt"/>
<mml:mtext>exists</mml:mtext>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\mathcal{A}_{\mathrm{vM}}}:=\{A:{\nu _{\mathrm{vM}}}(A)\hspace{2.5pt}\text{exists}\},\]]]></tex-math></alternatives>
</disp-formula> 
which is called the “density logic” in [<xref ref-type="bibr" rid="j_nejsds73_ref_078">78</xref>]. It is a pre-Dynkin-system [<xref ref-type="bibr" rid="j_nejsds73_ref_088">88</xref>].</p>
<p>We generalize Kolmogorov’s definition of statistical independence of events to finitely additive probability spaces.</p><statement id="j_nejsds73_stat_008"><label>Definition 7</label>
<title>(Statistical Independence of Events on a Finitely Additive Probability Space).</title>
<p>Let <inline-formula id="j_nejsds73_ineq_084"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">N</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="script">A</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">ν</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(N,\mathcal{A},\nu )$]]></tex-math></alternatives></inline-formula> be a finitely additive probability space. Two measurable sets <inline-formula id="j_nejsds73_ineq_085"><alternatives><mml:math>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="script">A</mml:mi></mml:math><tex-math><![CDATA[$A,B\in \mathcal{A}$]]></tex-math></alternatives></inline-formula> are independent iff 
<list>
<list-item id="j_nejsds73_li_005">
<label>1.</label>
<p><inline-formula id="j_nejsds73_ineq_086"><alternatives><mml:math>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo>∩</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="script">A</mml:mi></mml:math><tex-math><![CDATA[$A\cap B\in \mathcal{A}$]]></tex-math></alternatives></inline-formula></p>
</list-item>
<list-item id="j_nejsds73_li_006">
<label>2.</label>
<p><inline-formula id="j_nejsds73_ineq_087"><alternatives><mml:math>
<mml:mi mathvariant="italic">ν</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo>∩</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">ν</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">ν</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\nu (A\cap B)=\nu (A)\nu (B)$]]></tex-math></alternatives></inline-formula></p>
</list-item>
</list>
</p></statement>
<p>Observe that the first condition is naturally fulfilled in Kolmogorov’s <italic>σ</italic>-algebra. In the case of Von Mises’ density logic, this constraint is strict. A pre-Dynkin-system is not closed under arbitrary intersections.</p>
</sec>
<sec id="j_nejsds73_s_026">
<label>A.3</label>
<title>Von Mises’ Admissibility and Kolmogorov’s Independence Are Analogues</title>
<p>We now reconsider the definition of collectives and selection rules. Both are <inline-formula id="j_nejsds73_ineq_088"><alternatives><mml:math>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$0,1$]]></tex-math></alternatives></inline-formula>-sequences on the natural numbers, with the restriction that selection rules contain infinitely many 1’s and their frequency limit does not have to exist. Both sequences can be interpreted as indicator functions on the natural numbers. So for <inline-formula id="j_nejsds73_ineq_089"><alternatives><mml:math>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$x:\mathbb{N}\to \{0,1\}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds73_ineq_090"><alternatives><mml:math>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$s:\mathbb{N}\to \{0,1\}$]]></tex-math></alternatives></inline-formula> we write <inline-formula id="j_nejsds73_ineq_091"><alternatives><mml:math>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$X={x^{-1}}(1)$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds73_ineq_092"><alternatives><mml:math>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$S={s^{-1}}(1)$]]></tex-math></alternatives></inline-formula> for the corresponding subsets of the natural numbers.</p>
<p>We want to show that Von Mises’ admissibility condition is equivalent to the given definition of statistical independence on finitely additive probability spaces. Actually, the equivalence only holds for the slightly restricted case in which the selection rule itself possesses converging frequencies.</p><statement id="j_nejsds73_stat_009"><label>Theorem 1</label>
<title>(Admissibility in Von Mises setting implies Statistical Independence on Finitely Additive Probability Spaces).</title>
<p><italic>Let x be a collective with respect to s. Suppose furthermore that s has a converging frequency limit. Then X and S, the indexed sets corresponding to the collective (respectively selection rule) on the finitely additive probability space</italic> <inline-formula id="j_nejsds73_ineq_093"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\mathbb{N},{\mathcal{A}_{\mathrm{vM}}},{\nu _{\mathrm{vM}}})$]]></tex-math></alternatives></inline-formula> <italic>are statistically independent.</italic></p></statement><statement id="j_nejsds73_stat_010"><label>Proof.</label>
<p>We know that 
<disp-formula id="j_nejsds73_eq_010">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mspace width="2.5pt"/>
<mml:mtext>and</mml:mtext>
<mml:mspace width="2.5pt"/>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mspace width="1em"/>
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}& \underset{n\to \infty }{\lim }\frac{|\{i:x(i)=1\hspace{2.5pt}\text{and}\hspace{2.5pt}s(i)=1,1\le i\le n\}|}{|\{j:s(j)=1,1\le j\le n\}|}\\ {} & \hspace{1em}=\underset{n\to \infty }{\lim }\frac{|\{i:x(i)=1,1\le i\le n\}|}{n}\end{aligned}\]]]></tex-math></alternatives>
</disp-formula> 
which we can rewrite as 
<disp-formula id="j_nejsds73_eq_011">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>∩</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \underset{n\to \infty }{\lim }\frac{|X\cap S\cap {\mathbb{N}_{n}}|}{|S\cap {\mathbb{N}_{n}}|}=\underset{n\to \infty }{\lim }\frac{|X\cap {\mathbb{N}_{n}}|}{n}.\]]]></tex-math></alternatives>
</disp-formula> 
This gives 
<disp-formula id="j_nejsds73_eq_012">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>∩</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>∩</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>∩</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>∩</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}{\nu _{\mathrm{vM}}}(X\cap S)& =\underset{n\to \infty }{\lim }\frac{|X\cap S\cap {\mathbb{N}_{n}}|}{n}\\ {} & =\underset{n\to \infty }{\lim }\frac{|X\cap S\cap {\mathbb{N}_{n}}|}{|S\cap {\mathbb{N}_{n}}|}\frac{|S\cap {\mathbb{N}_{n}}|}{n}\\ {} & =\underset{n\to \infty }{\lim }\frac{|X\cap S\cap {\mathbb{N}_{n}}|}{|S\cap {\mathbb{N}_{n}}|}\underset{n\to \infty }{\lim }\frac{|S\cap {\mathbb{N}_{n}}|}{n}\\ {} & =\underset{n\to \infty }{\lim }\frac{|X\cap {\mathbb{N}_{n}}|}{n}\underset{n\to \infty }{\lim }\frac{|S\cap {\mathbb{N}_{n}}|}{n}\\ {} & ={\nu _{\mathrm{vM}}}(X){\nu _{\mathrm{vM}}}(S),\end{aligned}\]]]></tex-math></alternatives>
</disp-formula> 
by help of a standard result for the multiplication of sequence limits [<xref ref-type="bibr" rid="j_nejsds73_ref_062">62</xref>, Theorem 3.1.7].  □</p></statement><statement id="j_nejsds73_stat_011"><label>Theorem 2.</label>
<p><italic>Let X and S be two statistically independent events on the finitely additive probability space</italic> <inline-formula id="j_nejsds73_ineq_094"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\mathbb{N},{\mathcal{A}_{\mathrm{vM}}},{\nu _{\mathrm{vM}}})$]]></tex-math></alternatives></inline-formula> <italic>with</italic> <inline-formula id="j_nejsds73_ineq_095"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[${\nu _{\mathrm{vM}}}(S)\gt 0$]]></tex-math></alternatives></inline-formula><italic>, then the corresponding collective x, indicator function of X, has the admissible selection rule s, indicator function of S.</italic></p></statement><statement id="j_nejsds73_stat_012"><label>Proof.</label>
<p>It is given that 
<disp-formula id="j_nejsds73_eq_013">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>∩</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>∩</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mspace width="1em"/>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}& \underset{n\to \infty }{\lim }\frac{|X\cap S\cap {\mathbb{N}_{n}}|}{n}={\nu _{\mathrm{vM}}}(X\cap S)\\ {} & \hspace{1em}={\nu _{\mathrm{vM}}}(X){\nu _{\mathrm{vM}}}(S)=\underset{n\to \infty }{\lim }\frac{|X\cap {\mathbb{N}_{n}}|}{n}\underset{n\to \infty }{\lim }\frac{|S\cap {\mathbb{N}_{n}}|}{n}.\end{aligned}\]]]></tex-math></alternatives>
</disp-formula> 
Furthermore <inline-formula id="j_nejsds73_ineq_096"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[${\nu _{\mathrm{vM}}}(S)\gt 0$]]></tex-math></alternatives></inline-formula> implies that the corresponding selection rules selects infinitely many elements and <inline-formula id="j_nejsds73_ineq_097"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:msub><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle></mml:math><tex-math><![CDATA[${\lim \nolimits_{n\to \infty }}\frac{n}{|S\cap {\mathbb{N}_{n}}|}=\frac{1}{{\nu _{\mathrm{vM}}}(S)}$]]></tex-math></alternatives></inline-formula>.</p>
<p>This implies 
<disp-formula id="j_nejsds73_eq_014">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>∩</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mspace width="1em"/>
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>∩</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mspace width="1em"/>
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>∩</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mspace width="1em"/>
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mspace width="1em"/>
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">lim</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>∩</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}& \underset{n\to \infty }{\lim }\frac{|X\cap S\cap {\mathbb{N}_{n}}|}{|S\cap {\mathbb{N}_{n}}|}\\ {} & \hspace{1em}=\underset{n\to \infty }{\lim }\frac{|X\cap S\cap {\mathbb{N}_{n}}|}{n}\frac{n}{|S\cap {\mathbb{N}_{n}}|}\\ {} & \hspace{1em}=\underset{n\to \infty }{\lim }\frac{|X\cap S\cap {\mathbb{N}_{n}}|}{n}\underset{n\to \infty }{\lim }\frac{n}{|S\cap {\mathbb{N}_{n}}|}\\ {} & \hspace{1em}=\underset{n\to \infty }{\lim }\frac{|X\cap {\mathbb{N}_{n}}|}{n}\underset{n\to \infty }{\lim }\frac{|S\cap {\mathbb{N}_{n}}|}{n}\underset{n\to \infty }{\lim }\frac{n}{|S\cap {\mathbb{N}_{n}}|}\\ {} & \hspace{1em}=\underset{n\to \infty }{\lim }\frac{|X\cap {\mathbb{N}_{n}}|}{n}.\end{aligned}\]]]></tex-math></alternatives>
</disp-formula> 
 □</p></statement>
<p>We note some caveats regarding the preceding discussion. 
<list>
<list-item id="j_nejsds73_li_007">
<label>1.</label>
<p>We require the frequency limit for <italic>s</italic> to exist in Theorem <xref rid="j_nejsds73_stat_009">1</xref>. So the sequence <italic>S</italic> is not an entirely general selection rule.</p>
</list-item>
<list-item id="j_nejsds73_li_008">
<label>2.</label>
<p>The condition <inline-formula id="j_nejsds73_ineq_098"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[${\nu _{\mathrm{vM}}}(S)\gt 0$]]></tex-math></alternatives></inline-formula> in Theorem <xref rid="j_nejsds73_stat_011">2</xref> ensures that we do not condition on measure zero events. Furthermore, it guarantees at this point, that the indicator function of <italic>S</italic> is a selection rule containing infinitely many ones. Besides that, its frequency limit exists.</p>
</list-item>
<list-item id="j_nejsds73_li_009">
<label>3.</label>
<p>Even though given here only for the binary case. The argumentation can be extended to continuum labeled collectives [<xref ref-type="bibr" rid="j_nejsds73_ref_107">107</xref>, II.B].</p>
</list-item>
<list-item id="j_nejsds73_li_010">
<label>4.</label>
<p>In our discussion, we focus on admissibility instead of statistical independence in the sense of Von Mises, since our argument finally focuses on the randomness interpretation of statistical independence. Admissibility and Von Mises’ statistical independence are, despite the objects on which they are defined, equivalent.</p>
</list-item>
<list-item id="j_nejsds73_li_011">
<label>5.</label>
<p>Statistical independence in Von Mises’ setting and Kolmogorov’s setting are not equivalent. They are special cases of a more general form of statistical independence. So they are mathematically analogous under mild conditions. But we emphasize the conceptual difference. Von Mises refused to define statistical independence between mere events [<xref ref-type="bibr" rid="j_nejsds73_ref_107">107</xref>, p. 35]. Instead he demanded statistical independence to be defined between collectives underlining that independence is a concept about aggregates, the collectives, not single occurring events.</p>
</list-item>
</list>
</p>
</sec>
</app>
<app id="j_nejsds73_app_002"><label>Appendix B</label>
<title>Von Mises’ Notions of Fairness in Practice</title>
<p>In practice we never observe infinite sequences, which are the building blocks of Von Mises’ theory. Nevertheless, infinite sequences can be seen as idealizations of finite, observable sequences. We make the following crucial and debatable assumption here: <italic>Given a sequence of sufficient length, the frequencies observed in this sequences are arbitrarily close to the limit of the frequency in the continuation of this sequence.</italic></p>
<p>This assumption is critical for at least three parts: 
<list>
<list-item id="j_nejsds73_li_012">
<label>1.</label>
<p>We assume convergence (cf. [<xref ref-type="bibr" rid="j_nejsds73_ref_037">37</xref>]).</p>
</list-item>
<list-item id="j_nejsds73_li_013">
<label>2.</label>
<p>In finite data we can only observe fractional frequencies, but not irrational ones (converging frequencies do fill the entire [0,1]).</p>
</list-item>
<list-item id="j_nejsds73_li_014">
<label>3.</label>
<p>The frequency limit of any infinite sequence is not governed by any frequency observed on a finite sequence.</p>
</list-item>
</list> 
For further simplification, we constrain ourselves to <inline-formula id="j_nejsds73_ineq_099"><alternatives><mml:math>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$0,1$]]></tex-math></alternatives></inline-formula>-outcomes and <inline-formula id="j_nejsds73_ineq_100"><alternatives><mml:math>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$0,1$]]></tex-math></alternatives></inline-formula>-decisions. In the following, we call 1 as outcome or decision “positive”, comparable to a setting where the 1-label corresponds to “getting a loan” and 0 to “not getting a loan”.</p>
<p>We consider the following simple toy example: let <inline-formula id="j_nejsds73_ineq_101"><alternatives><mml:math>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="double-struck">R</mml:mi></mml:math><tex-math><![CDATA[$X=\mathbb{R}$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_nejsds73_ineq_102"><alternatives><mml:math>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$Y=\{0,1\}$]]></tex-math></alternatives></inline-formula>. The data points of the 0 label are distributed according to a Gaussian distribution with mean <inline-formula id="j_nejsds73_ineq_103"><alternatives><mml:math>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$-1$]]></tex-math></alternatives></inline-formula> and standard deviation 1. The data points of the positive labels are distributed according to a Gaussian distribution with mean 1 and standard deviation 0.5. We consider the simple logistic probabilistic predictor <inline-formula id="j_nejsds73_ineq_104"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>+</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfrac>
</mml:mstyle></mml:math><tex-math><![CDATA[$p(x)=\frac{{e^{cx+a}}}{1+{e^{cx+a}}}$]]></tex-math></alternatives></inline-formula> with <inline-formula id="j_nejsds73_ineq_105"><alternatives><mml:math>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$c=2$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds73_ineq_106"><alternatives><mml:math>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo>−</mml:mo>
<mml:mn>0.2</mml:mn></mml:math><tex-math><![CDATA[$a=-0.2$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds73_ineq_107"><alternatives><mml:math>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi></mml:math><tex-math><![CDATA[$x\in X$]]></tex-math></alternatives></inline-formula>. We consider two groups: group A consists of all data points with values between <inline-formula id="j_nejsds73_ineq_108"><alternatives><mml:math>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$-1$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds73_ineq_109"><alternatives><mml:math>
<mml:mo>−</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$-2$]]></tex-math></alternatives></inline-formula>, group B consists of all data points with values between smaller than <inline-formula id="j_nejsds73_ineq_110"><alternatives><mml:math>
<mml:mo>−</mml:mo>
<mml:mn>1.5</mml:mn></mml:math><tex-math><![CDATA[$-1.5$]]></tex-math></alternatives></inline-formula> or greater than 1.5. In this case, the input and sensitive groups are highly correlated. Figure <xref rid="j_nejsds73_fig_001">1</xref> and Figure <xref rid="j_nejsds73_fig_002">2</xref> illustrate the example.</p>
<fig id="j_nejsds73_fig_001">
<label>Figure 1</label>
<caption>
<p>Generating Distribution and Predictor of the Toy Example.</p>
</caption>
<graphic xlink:href="nejsds73_g001.jpg"/>
</fig>
<fig id="j_nejsds73_fig_002">
<label>Figure 2</label>
<caption>
<p>Groups in the Toy Example.</p>
</caption>
<graphic xlink:href="nejsds73_g002.jpg"/>
</fig>
<p>Since the predictor is probabilistic we have to introduce a threshold <inline-formula id="j_nejsds73_ineq_111"><alternatives><mml:math>
<mml:mi mathvariant="italic">q</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$q\in [0,1]$]]></tex-math></alternatives></inline-formula> which derives a <inline-formula id="j_nejsds73_ineq_112"><alternatives><mml:math>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$0,1$]]></tex-math></alternatives></inline-formula>-decision from the prediction. The following plots show how the frequency of positive outcomes respectively positive decisions depending on the notion of fairness changes for different decision-thresholds.</p>
<def-list><def-item id="x1-29009x-B"><term>Independence</term><def>
<p>We compute the frequency of positive decisions for each group and the entirety. Independence requires these frequencies to be equal. See Figure <xref rid="j_nejsds73_fig_003">3</xref>. 
<fig id="j_nejsds73_fig_003">
<label>Figure 3</label>
<caption>
<p>Frequency of Positive Decisions per Group depending on the Decision-Threshold.</p>
</caption>
<graphic xlink:href="nejsds73_g003.jpg"/>
</fig>
</p></def></def-item><def-item id="x1-29011x-B"><term>Separation</term><def>
<p>We compute the frequency of positive decisions given that the outcomes were positive for each group and the entirety. Separation requires these frequencies to be equal. See Figure <xref rid="j_nejsds73_fig_004">4</xref>. 
<fig id="j_nejsds73_fig_004">
<label>Figure 4</label>
<caption>
<p>Frequency of Positive Decisions given Positive Outcomes per Group depending on the Decision-Threshold.</p>
</caption>
<graphic xlink:href="nejsds73_g004.jpg"/>
</fig>
</p></def></def-item><def-item id="x1-29013x-B"><term>Sufficiency</term><def>
<p>We compute the frequency of positive outcomes given that the decision were positive for each group and the entirety. Sufficiency requires these frequencies to be equal. See Figure <xref rid="j_nejsds73_fig_005">5</xref>. 
<fig id="j_nejsds73_fig_005">
<label>Figure 5</label>
<caption>
<p>Frequency of Positive Outcomes given Positive Decisions per Group depending on the Decision-Threshold.</p>
</caption>
<graphic xlink:href="nejsds73_g005.jpg"/>
</fig>
</p></def></def-item></def-list> 
<p>What do we learn from this toy example:</p>
<list>
<list-item id="j_nejsds73_li_015">
<label>1.</label>
<p>Our suggested notions have an empirical counterpart. Even though, the assumption taken above is critical. Much in contrast to most literature in fair machine learning, however, we (can) specify the idealization assumption explicitly. The regular setup hides many of the concerns behind a curtain of acceptance.</p>
</list-item>
<list-item id="j_nejsds73_li_016">
<label>2.</label>
<p>We observe that there is nothing special about our definitions of fairness. The frequential approach is intuitive and simplifies communication of technical results about machine learning fairness in societal debates.</p>
</list-item>
<list-item id="j_nejsds73_li_017">
<label>3.</label>
<p>We did not allude to specific algorithms which guarantee accurate predictions under fairness constraints. All such existing algorithms can be deployed in order to provide fair predictions following our definitions. Our notions are reinterpretations, not entirely new setups.</p>
</list-item>
<list-item id="j_nejsds73_li_018">
<label>4.</label>
<p>There is still a further, important, debate missing in this work. As pointed out in the conclusion of the paper, further conceptual clarification has to account for approximations of fairness notions to further enable principled argumentation in fairness debates.</p>
</list-item>
</list>
</app>
<app id="j_nejsds73_app_003"><label>Appendix C</label>
<title>Four Intuitions of Randomness</title>
<def-list><def-item id="x1-30001x-C"><term>Typicality</term><def>
<p>A sequence is called random if it shares “some” characteristics of any possible sequence. Martin-Löf and Schnorr formalized this idea via statistical tests [<xref ref-type="bibr" rid="j_nejsds73_ref_065">65</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_087">87</xref>].</p></def></def-item><def-item id="x1-30002x-C"><term>Incompressibility</term><def>
<p>The information contained in a random sequences is (approximately) as large as the sequence itself. The sequences cannot be compressed by “some” procedure [<xref ref-type="bibr" rid="j_nejsds73_ref_057">57</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_020">20</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_086">86</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_059">59</xref>].</p></def></def-item><def-item id="x1-30003x-C"><term>Unpredictability</term><def>
<p>In a random sequence one can know all foregoing elements without being able to predict by “some” procedure the next element [<xref ref-type="bibr" rid="j_nejsds73_ref_102">102</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_068">68</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_090">90</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_038">38</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_026">26</xref>].</p></def></def-item><def-item id="x1-30004x-C"><term>Independence</term><def>
<p>A random sequence is independent of “some” other sequences [<xref ref-type="bibr" rid="j_nejsds73_ref_105">105</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_023">23</xref>].</p></def></def-item></def-list> </app>
<app id="j_nejsds73_app_004"><label>Appendix D</label>
<title>Relations Between Randomness Definitions</title>
<p>We briefly outline some of the known relationships between various notions of randomness.</p>
<p>On one hand, there are mathematical expressions capturing the same randomness intuitions meanwhile being mathematically distinct to each other.</p>
<p>For instance, the definition of a typical sequence following Martin-Löf [<xref ref-type="bibr" rid="j_nejsds73_ref_065">65</xref>] implies Schnorr’s definition of typicality [<xref ref-type="bibr" rid="j_nejsds73_ref_085">85</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_087">87</xref>] but not vice versa [<xref ref-type="bibr" rid="j_nejsds73_ref_099">99</xref>, p. 143]. Cooman and De Bock’s imprecise unpredictability randomness [<xref ref-type="bibr" rid="j_nejsds73_ref_026">26</xref>] is strictly more expressive than Vovk and Shafer’s unpredictability randomness [<xref ref-type="bibr" rid="j_nejsds73_ref_090">90</xref>, Section 1.1], [<xref ref-type="bibr" rid="j_nejsds73_ref_026">26</xref>, Theorem 37].</p>
<p>On the other hand, there are mathematical expressions capturing differing randomness intuitions meanwhile being necessary, sufficient or even equivalent to each other. For instance, the Levin-Schnorr theorem, simultaneously proven by Levin [<xref ref-type="bibr" rid="j_nejsds73_ref_059">59</xref>] and Schnorr [<xref ref-type="bibr" rid="j_nejsds73_ref_086">86</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_087">87</xref>], established an equivalence of typicality following Martin-Löf [<xref ref-type="bibr" rid="j_nejsds73_ref_065">65</xref>] and incompressibility following Levin [<xref ref-type="bibr" rid="j_nejsds73_ref_059">59</xref>] and Schnorr [<xref ref-type="bibr" rid="j_nejsds73_ref_086">86</xref>] based on the idea of [<xref ref-type="bibr" rid="j_nejsds73_ref_057">57</xref>] (e.g., see in [<xref ref-type="bibr" rid="j_nejsds73_ref_103">103</xref>, Theorem 5.3] and references therein). Cooman and de Bock expressed Schnorr [<xref ref-type="bibr" rid="j_nejsds73_ref_085">85</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_087">87</xref>] and Martin-Löf randomness [<xref ref-type="bibr" rid="j_nejsds73_ref_065">65</xref>] in terms of an unpredictability approach [<xref ref-type="bibr" rid="j_nejsds73_ref_026">26</xref>]. Muchnik showed in [<xref ref-type="bibr" rid="j_nejsds73_ref_068">68</xref>] that all incompressible sequences, again following Levin and Schnorr’s prefix-complexity approach [<xref ref-type="bibr" rid="j_nejsds73_ref_086">86</xref>, <xref ref-type="bibr" rid="j_nejsds73_ref_059">59</xref>] are unpredictable in his sense.</p>
<p>In Ville’s thesis [<xref ref-type="bibr" rid="j_nejsds73_ref_102">102</xref>] he attempted to generalize the idea of “excluding a gambling strategy” (see Section <xref rid="j_nejsds73_s_010">4.4</xref>) via a game-theoretic approach. He showed [<xref ref-type="bibr" rid="j_nejsds73_ref_102">102</xref>, p. 76] that for any set of admissible selection rules <inline-formula id="j_nejsds73_ineq_113"><alternatives><mml:math>
<mml:mi mathvariant="script">S</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{S}$]]></tex-math></alternatives></inline-formula> he could construct a gambling strategy, more exactly a capital process associated to a gambling strategy, which captures the same randomness definition. The opposite direction is, however, impossible [<xref ref-type="bibr" rid="j_nejsds73_ref_102">102</xref>, p. 39]. But Ambos-Spies et al. [<xref ref-type="bibr" rid="j_nejsds73_ref_003">3</xref>] introduced a weaker form of [<xref ref-type="bibr" rid="j_nejsds73_ref_102">102</xref>]’s randomness as unpredictability which is equivalent to Church’s randomness [<xref ref-type="bibr" rid="j_nejsds73_ref_023">23</xref>] as independence [<xref ref-type="bibr" rid="j_nejsds73_ref_031">31</xref>, Section 12.3]. Finally, van Lambalgen observed an abstract interpretation of “random with respect to something” as “independent to” in [<xref ref-type="bibr" rid="j_nejsds73_ref_101">101</xref>].</p>
<sec id="j_nejsds73_s_027">
<label>D.1</label>
<title>A Prototypical, Absolute and Universal Notion of Randomness</title>
<p>One often referred <italic>absolute and universal</italic> definition of randomness is Martin-Löf’s typicality approach [<xref ref-type="bibr" rid="j_nejsds73_ref_015">15</xref>]. Martin-Löf’s randomness as typicality has been equivalently formalized in terms of incompressibility and unpredictability (see Section <xref rid="j_nejsds73_app_004">D</xref>). Furthermore, a sequence, which is Martin-Löf random [<xref ref-type="bibr" rid="j_nejsds73_ref_065">65</xref>], is statistically independent to all partial computable selection rules [<xref ref-type="bibr" rid="j_nejsds73_ref_023">23</xref>], [<xref ref-type="bibr" rid="j_nejsds73_ref_096">96</xref>, Theorem 11]. Contrarily, not every collective with respect to all partial computable selection rules is a Martin-Löf random sequence [<xref ref-type="bibr" rid="j_nejsds73_ref_013">13</xref>, p. 193]. So Martin-Löf’s definition is linked to all four intuitions.</p>
</sec>
</app>
<app id="j_nejsds73_app_005"><label>Appendix E</label>
<title>Kolmogorov’s Versus Von Mises’ Probability Theories in a Table</title>
<p>See Table <xref rid="j_nejsds73_tab_002">2</xref>.</p>
<table-wrap id="j_nejsds73_tab_002">
<label>Table 2</label>
<caption>
<p>Summary of main difference between Kolmogorov’s and Von Mises’ probability theories.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Kolmogorov</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Von Mises</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">fundamental structure</td>
<td style="vertical-align: top; text-align: left">probability space <inline-formula id="j_nejsds73_ineq_114"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="normal">Ω</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="script">F</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\Omega ,\mathcal{F},P)$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">collectives <inline-formula id="j_nejsds73_ineq_115"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${({x_{i}})_{i\in \mathbb{N}}}$]]></tex-math></alternatives></inline-formula>, implicitly <inline-formula id="j_nejsds73_ineq_116"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\mathbb{N},{\mathcal{A}_{\mathrm{vM}}},{\nu _{\mathrm{vM}}})$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">base set</td>
<td style="vertical-align: top; text-align: left">(almost arbitrary) set Ω</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_nejsds73_ineq_117"><alternatives><mml:math>
<mml:mi mathvariant="double-struck">N</mml:mi></mml:math><tex-math><![CDATA[$\mathbb{N}$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">set of events</td>
<td style="vertical-align: top; text-align: left"><italic>σ</italic>-algebra <inline-formula id="j_nejsds73_ineq_118"><alternatives><mml:math>
<mml:mi mathvariant="script">F</mml:mi></mml:math><tex-math><![CDATA[$\mathcal{F}$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">Dynkin-system <inline-formula id="j_nejsds73_ineq_119"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mathcal{A}_{\mathrm{vM}}}$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">probability</td>
<td style="vertical-align: top; text-align: left">finite positive measure</td>
<td style="vertical-align: top; text-align: left">limit of frequency sequence</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">probability measure</td>
<td style="vertical-align: top; text-align: left">countably additive</td>
<td style="vertical-align: top; text-align: left">finitely additive <inline-formula id="j_nejsds73_ineq_120"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ν</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">vM</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\nu _{\mathrm{vM}}}$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">randomness</td>
<td style="vertical-align: top; text-align: left">no explicit mathematical definition</td>
<td style="vertical-align: top; text-align: left">explicit mathematical definition</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">statistical independence</td>
<td style="vertical-align: top; text-align: left">factorization of joint distribution</td>
<td style="vertical-align: top; text-align: left">frequency limit doesn’t change under subselection</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">data model</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">each data point a random variable <inline-formula id="j_nejsds73_ineq_121"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{i}}$]]></tex-math></alternatives></inline-formula> on <inline-formula id="j_nejsds73_ineq_122"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="normal">Ω</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="script">F</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\Omega ,\mathcal{F},P)$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">each data point an element in a collective <inline-formula id="j_nejsds73_ineq_123"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${({x_{i}})_{i\in \mathbb{N}}}$]]></tex-math></alternatives></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap>
</app>
<app id="j_nejsds73_app_006"><label>Appendix F</label>
<title>Penguin Colony Example for Fair Collective</title>
<p>See Figure <xref rid="j_nejsds73_fig_006">6</xref>.</p>
<fig id="j_nejsds73_fig_006">
<label>Figure 6</label>
<caption>
<p>Example for a subselection and fair collective. <inline-formula id="j_nejsds73_ineq_124"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${n_{b}}$]]></tex-math></alternatives></inline-formula> denotes the number of black penguins among the first <italic>n</italic>-penguins. Blackness of penguins is distributed fairly with respect to sex if <inline-formula id="j_nejsds73_ineq_125"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{b}}={p_{b|F}}$]]></tex-math></alternatives></inline-formula>.</p>
</caption>
<graphic xlink:href="nejsds73_g006.jpg"/>
</fig>
</app></app-group>
<ack id="j_nejsds73_ack_001">
<title>Acknowledgements</title>
<p>Many thanks to Christian Fröhlich, Eric Raidl, Sebastian Zezulka, Thomas Grote and Benedikt Höltgen for helpful discussions and feedback. Furthermore, the authors thank all participants of the Philosophy of Science meets Machine Learning Conference 2022 in Tübingen, for all helpful comments and debates.</p>
<p>The authors appreciate and thank the International Max Planck Research School for Intelligent System (IMPRS-IS) for supporting Rabanus Derr.</p></ack>
<ref-list id="j_nejsds73_reflist_001">
<title>References</title>
<ref id="j_nejsds73_ref_001">
<label>[1]</label><mixed-citation publication-type="journal"><string-name><surname>Abraham</surname>, <given-names>K. S.</given-names></string-name> <article-title>Efficiency and fairness in insurance risk classification</article-title>. <source>Virginia Law Review</source> <volume>71</volume>(<issue>3</issue>) <fpage>403</fpage>–<lpage>451</lpage> (<year>1985</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_002">
<label>[2]</label><mixed-citation publication-type="chapter"><string-name><surname>Altman</surname>, <given-names>A.</given-names></string-name> <chapter-title>Discrimination</chapter-title>. In (<string-name><given-names>E. N.</given-names> <surname>Zalta</surname></string-name>, ed.) <source>The Stanford Encyclopedia of Philosophy (Summerition)</source>, <edition>2020</edition>th ed. <publisher-name>Stanford University</publisher-name> (<year>2020</year>). <comment><uri>https://plato.stanford.edu/entries/discrimination/</uri></comment></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_003">
<label>[3]</label><mixed-citation publication-type="chapter"><string-name><surname>Ambos-Spies</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Mayordomo</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>Y.</given-names></string-name> and <string-name><surname>Zheng</surname>, <given-names>X.</given-names></string-name> <chapter-title>Resource-bounded balanced genericity, stochasticity and weak randomness</chapter-title>. In <source>Annual Symposium on Theoretical Aspects of Computer Science</source> <fpage>61</fpage>–<lpage>74</lpage>. <publisher-name>Springer</publisher-name> (<year>1996</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1462086">MR1462086</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_004">
<label>[4]</label><mixed-citation publication-type="chapter"><string-name><surname>Bandyopadhyay</surname>, <given-names>P. S.</given-names></string-name> and <string-name><surname>Forster</surname>, <given-names>M. R.</given-names></string-name> <chapter-title>Philosophy of statistics: An introduction</chapter-title>. In (<string-name><given-names>P. S.</given-names> <surname>Bandyopadhyay</surname></string-name> and <string-name><given-names>M. R.</given-names> <surname>Forster</surname></string-name>, eds.) <source>Philosophy of Statistics</source> <volume>7</volume>. <publisher-name>Elsevier</publisher-name> (<year>2011</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1016/B978-0-444-51862-0.50001-0" xlink:type="simple">https://doi.org/10.1016/B978-0-444-51862-0.50001-0</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3295937">MR3295937</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_005">
<label>[5]</label><mixed-citation publication-type="book"><string-name><surname>Solon</surname> <given-names>Barocas</given-names></string-name>, <string-name><surname>Hardt</surname>, <given-names>M.</given-names></string-name> and <string-name><surname>Narayanan</surname>, <given-names>A.</given-names></string-name> <source>Fairness and Machine Learning: Limitations and Opportunities</source> (<year>2019</year>). <comment><uri>http://www.fairmlbook.org</uri></comment></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_006">
<label>[6]</label><mixed-citation publication-type="chapter"><string-name><surname>Battersby</surname>, <given-names>M.</given-names></string-name> <chapter-title>The rhetoric of numbers: Statistical inference as argumentation</chapter-title>. In <source>OSSA Conference Archive</source> (<year>2003</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_007">
<label>[7]</label><mixed-citation publication-type="chapter"><string-name><surname>Bennett</surname>, <given-names>D.</given-names></string-name> <chapter-title>Defining randomness</chapter-title>. In (<string-name><given-names>P. S.</given-names> <surname>Bandyopadhyay</surname></string-name> and <string-name><given-names>M. R.</given-names> <surname>Forster</surname></string-name>, eds.) <source>Philosophy of Statistics</source> <volume>7</volume>. <publisher-name>Elsevier</publisher-name> (<year>2011</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1016/B978-0-444-51862-0.50001-0" xlink:type="simple">https://doi.org/10.1016/B978-0-444-51862-0.50001-0</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3295937">MR3295937</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_008">
<label>[8]</label><mixed-citation publication-type="journal"><string-name><surname>Berkovitz</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Frigg</surname>, <given-names>R.</given-names></string-name> and <string-name><surname>Kronz</surname>, <given-names>F.</given-names></string-name> <article-title>The ergodic hierarchy, randomness and hamiltonian chaos</article-title>. <source>Studies in History and Philosophy of Science Part B: Studies in History and Philosophy of Modern Physics</source> <volume>37</volume>(<issue>4</issue>) <fpage>661</fpage>–<lpage>691</lpage> (<year>2006</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1016/j.shpsb.2006.02.003" xlink:type="simple">https://doi.org/10.1016/j.shpsb.2006.02.003</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2344115">MR2344115</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_009">
<label>[9]</label><mixed-citation publication-type="book"><string-name><surname>Biddle</surname>, <given-names>D.</given-names></string-name> <source>Adverse Impact and Test Validation: A Practitioner’s Guide to Valid and Defensible Employment Testing</source>. <publisher-name>Gower Publishing, Ltd.</publisher-name> (<year>2006</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_010">
<label>[10]</label><mixed-citation publication-type="journal"><string-name><surname>Bienvenu</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Shafer</surname>, <given-names>G.</given-names></string-name> and <string-name><surname>Shen</surname>, <given-names>A.</given-names></string-name> <article-title>On the history of martingales in the study of randomness</article-title>. <source>Electronic Journal for History of Probability and Statistics</source> <volume>5</volume>(<issue>1</issue>) <fpage>1</fpage>–<lpage>40</lpage> (<year>2009</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2520666">MR2520666</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_011">
<label>[11]</label><mixed-citation publication-type="chapter"><string-name><surname>Binns</surname>, <given-names>R.</given-names></string-name> <chapter-title>Fairness in machine learning: Lessons from political philosophy</chapter-title>. In <source>Conference on Fairness, Accountability and Transparency</source> <fpage>149</fpage>–<lpage>159</lpage>. <publisher-name>PMLR</publisher-name> (<year>2018</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_012">
<label>[12]</label><mixed-citation publication-type="chapter"><string-name><surname>Binns</surname>, <given-names>R.</given-names></string-name> <chapter-title>On the apparent conflict between individual and group fairness</chapter-title>. In <source>Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency</source> <fpage>514</fpage>–<lpage>524</lpage> (<year>2020</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_013">
<label>[13]</label><mixed-citation publication-type="book"><string-name><surname>Blum</surname>, <given-names>N.</given-names></string-name> <source>Einführung in Formale Sprachen, Berechenbarkeit, Informations-und Lerntheorie</source>. <publisher-name>Oldenbourg Wissenschaftsverlag</publisher-name> (<year>2009</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_014">
<label>[14]</label><mixed-citation publication-type="journal"><string-name><surname>Broome</surname>, <given-names>J.</given-names></string-name> <article-title>Selecting people randomly</article-title>. <source>Ethics</source> <volume>95</volume>(<issue>1</issue>) <fpage>38</fpage>–<lpage>55</lpage> (<year>1984</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_015">
<label>[15]</label><mixed-citation publication-type="journal"><string-name><surname>Buss</surname>, <given-names>S.</given-names></string-name> and <string-name><surname>Minnes</surname>, <given-names>M.</given-names></string-name> <article-title>Probabilistic algorithmic randomness</article-title>. <source>The Journal of Symbolic Logic</source> <volume>78</volume>(<issue>2</issue>) <fpage>579</fpage>–<lpage>601</lpage> (<year>2013</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3145197">MR3145197</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_016">
<label>[16]</label><mixed-citation publication-type="journal"><string-name><surname>Toon</surname>, <given-names>C.</given-names></string-name> and <string-name><surname>Verwer</surname>, <given-names>S.</given-names></string-name> <article-title>Three naive Bayes approaches for discrimination-free classification</article-title>. <source>Data Mining and Knowledge Discovery</source> <volume>21</volume>(<issue>2</issue>) <fpage>277</fpage>–<lpage>292</lpage> (<year>2010</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/s10618-010-0190-x" xlink:type="simple">https://doi.org/10.1007/s10618-010-0190-x</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2720507">MR2720507</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_017">
<label>[17]</label><mixed-citation publication-type="journal"><string-name><surname>Campbell</surname>, <given-names>D. T.</given-names></string-name> <article-title>Common fate, similarity, and other indices of the status of aggregates of persons as social entities</article-title>. <source>Behavioral Science</source> <volume>3</volume>(<issue>1</issue>) <fpage>14</fpage>–<lpage>25</lpage> (<year>1958</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_018">
<label>[18]</label><mixed-citation publication-type="book"><string-name><surname>Casella</surname>, <given-names>G.</given-names></string-name> and <string-name><surname>Berger</surname>, <given-names>R. L.</given-names></string-name> <source>Statistical Inference</source> <edition>2</edition>nd ed. <series>Duxbury Advanced Series</series> (<year>2002</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1051420">MR1051420</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_019">
<label>[19]</label><mixed-citation publication-type="journal"><string-name><surname>Castano</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Yzerbyt</surname>, <given-names>V.</given-names></string-name>, <string-name><surname>Paladino</surname>, <given-names>M.-P.</given-names></string-name> and <string-name><surname>Sacchi</surname>, <given-names>S.</given-names></string-name> <article-title>I belong, therefore, I exist: Ingroup identification, ingroup entitativity, and ingroup bias</article-title>. <source>Personality and Social Psychology Bulletin</source> <volume>28</volume>(<issue>2</issue>) <fpage>135</fpage>–<lpage>143</lpage> (<year>2002</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_020">
<label>[20]</label><mixed-citation publication-type="journal"><string-name><surname>Chaitin</surname>, <given-names>G. J.</given-names></string-name> <article-title>On the length of programs for computing finite binary sequences</article-title>. <source>Journal of the ACM</source> <volume>13</volume>(<issue>4</issue>) <fpage>547</fpage>–<lpage>569</lpage> (<year>1966</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1145/321356.321363" xlink:type="simple">https://doi.org/10.1145/321356.321363</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0210520">MR0210520</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_021">
<label>[21]</label><mixed-citation publication-type="journal"><string-name><surname>Chichilnisky</surname>, <given-names>G.</given-names></string-name> <article-title>The foundations of statistics with black swans</article-title>. <source>Mathematical Social Sciences</source> <volume>59</volume>(<issue>2</issue>) <fpage>184</fpage>–<lpage>192</lpage> (<year>2010</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1016/j.mathsocsci.2009.09.007" xlink:type="simple">https://doi.org/10.1016/j.mathsocsci.2009.09.007</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2650318">MR2650318</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_022">
<label>[22]</label><mixed-citation publication-type="journal"><string-name><surname>Chouldechova</surname>, <given-names>A.</given-names></string-name> <article-title>Fair prediction with disparate impact: A study of bias in recidivism prediction instruments</article-title>. <source>Big Data</source> <volume>5</volume>(<issue>2</issue>) <fpage>153</fpage>–<lpage>163</lpage> (<year>2017</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_023">
<label>[23]</label><mixed-citation publication-type="journal"><string-name><surname>Church</surname>, <given-names>A.</given-names></string-name> <article-title>On the concept of a random sequence</article-title>. <source>Bulletin of the American Mathematical Society</source> <volume>46</volume>(<issue>2</issue>) <fpage>130</fpage>–<lpage>135</lpage> (<year>1940</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1090/S0002-9904-1940-07154-X" xlink:type="simple">https://doi.org/10.1090/S0002-9904-1940-07154-X</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0000911">MR0000911</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_024">
<label>[24]</label><mixed-citation publication-type="chapter"><string-name><surname>Cook</surname>, <given-names>D. C.</given-names></string-name> <chapter-title>The concept of independence in accounting</chapter-title>. In <source>Federal Securities Law and Accounting 1933–1970: Selected Addresses</source> <fpage>198</fpage>–<lpage>222</lpage>. <publisher-name>Routledge</publisher-name> (<year>2020</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_025">
<label>[25]</label><mixed-citation publication-type="journal"><string-name><surname>Dawid</surname>, <given-names>P.</given-names></string-name> <article-title>On individual risk</article-title>. <source>Synthese</source> <volume>194</volume>(<issue>9</issue>) <fpage>3445</fpage>–<lpage>3474</lpage> (<year>2017</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/s11229-015-0953-4" xlink:type="simple">https://doi.org/10.1007/s11229-015-0953-4</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3704899">MR3704899</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_026">
<label>[26]</label><mixed-citation publication-type="journal"><string-name><surname>De Cooman</surname>, <given-names>G.</given-names></string-name> and <string-name><surname>De Bock</surname>, <given-names>J.</given-names></string-name> <article-title>Randomness is inherently imprecise</article-title>. <source>International Journal of Approximate Reasoning</source> <volume>141</volume>. <fpage>28</fpage>–<lpage>68</lpage> (<year>2022</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1016/j.ijar.2021.06.018" xlink:type="simple">https://doi.org/10.1016/j.ijar.2021.06.018</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4364895">MR4364895</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_027">
<label>[27]</label><mixed-citation publication-type="chapter"><string-name><surname>De Finetti</surname>, <given-names>B.</given-names></string-name> <chapter-title>Funzione caratteristica di un fenomeno aleatorio</chapter-title>. In <source>Atti del Congresso Internazionale dei Matematici: Bologna del 3 al 10 de settembre di 1928</source> <fpage>179</fpage>–<lpage>190</lpage> (<year>1929</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_028">
<label>[28]</label><mixed-citation publication-type="book"><string-name><surname>De Moivre</surname>, <given-names>A.</given-names></string-name> <source>The Doctrine of Chances: A Method of Calculating the Probabilities of Events in Play</source> <edition>2</edition>nd ed. <publisher-name>Frank Cass and Company Limited</publisher-name> (<year>1738/1967</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0231695">MR0231695</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_029">
<label>[29]</label><mixed-citation publication-type="book"><string-name><surname>Deutsch</surname>, <given-names>D.</given-names></string-name> <source>The Beginning of Infinity: Explanations that Transform the World</source>. <publisher-name>Penguin</publisher-name> (<year>2011</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2984795">MR2984795</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_030">
<label>[30]</label><mixed-citation publication-type="book"><string-name><surname>Devroye</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Györfi</surname>, <given-names>L.</given-names></string-name> and <string-name><surname>Lugosi</surname>, <given-names>G.</given-names></string-name> <source>A Probabilistic Theory of Pattern Recognition</source>. <publisher-name>Springer</publisher-name> (<year>1996</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/978-1-4612-0711-5" xlink:type="simple">https://doi.org/10.1007/978-1-4612-0711-5</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1383093">MR1383093</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_031">
<label>[31]</label><mixed-citation publication-type="book"><string-name><surname>Downey</surname>, <given-names>R. G.</given-names></string-name> and <string-name><surname>Hirschfeldt</surname>, <given-names>D. R.</given-names></string-name> <source>Algorithmic Randomness and Complexity</source>. <publisher-name>Springer</publisher-name> (<year>2010</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/978-0-387-68441-3" xlink:type="simple">https://doi.org/10.1007/978-0-387-68441-3</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2732288">MR2732288</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_032">
<label>[32]</label><mixed-citation publication-type="book"><string-name><surname>Durrett</surname>, <given-names>R.</given-names></string-name> <source>Probability: Theory and Examples</source> <edition>5</edition>th ed. <publisher-name>Cambridge University Press</publisher-name> (<year>2019</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1017/9781108591034" xlink:type="simple">https://doi.org/10.1017/9781108591034</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3930614">MR3930614</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_033">
<label>[33]</label><mixed-citation publication-type="chapter"><string-name><surname>Dwork</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Hardt</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Pitassi</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Reingold</surname>, <given-names>O.</given-names></string-name> and <string-name><surname>Zemel</surname>, <given-names>R.</given-names></string-name> <chapter-title>Fairness through awareness</chapter-title>. In <source>Proceedings of the 3rd Innovations in Theoretical Computer Science Conference – ITCS 2012</source> <fpage>214</fpage>–<lpage>226</lpage> (<year>2012</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3388391">MR3388391</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_034">
<label>[34]</label><mixed-citation publication-type="chapter"><string-name><surname>Dwork</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Lee</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Lin</surname>, <given-names>H.</given-names></string-name> and <string-name><surname>Tankala</surname>, <given-names>P.</given-names></string-name> <chapter-title>From pseudorandomness to multi-group fairness and back</chapter-title>. In <source>The Thirty Sixth Annual Conference on Learning Theory</source> <fpage>3566</fpage>–<lpage>3614</lpage>. <publisher-name>PMLR</publisher-name> (<year>2023</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_035">
<label>[35]</label><mixed-citation publication-type="chapter"><string-name><surname>Eagle</surname>, <given-names>A.</given-names></string-name> <chapter-title>Chance versus Randomness</chapter-title>. In (<string-name><given-names>E. N.</given-names> <surname>Zalta</surname></string-name>, ed.) <source>The Stanford Encyclopedia of Philosophy</source>, <publisher-name>Metaphysics Research Lab, Stanford University</publisher-name>, Spring 2021 edition, <year>2021</year>.</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_036">
<label>[36]</label><mixed-citation publication-type="book"><string-name><surname>Fine</surname>, <given-names>T. L.</given-names></string-name> <source>Theories of Probability: An Examination of Foundations</source>. <publisher-name>Academic Press</publisher-name> (<year>1973</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0433529">MR0433529</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_037">
<label>[37]</label><mixed-citation publication-type="chapter"><string-name><surname>Fröhlich</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Derr</surname>, <given-names>R.</given-names></string-name> and <string-name><surname>Williamson</surname>, <given-names>R. C.</given-names></string-name> <chapter-title>Towards a strictly frequentist theory of imprecise probability</chapter-title>. In (<string-name><given-names>E.</given-names> <surname>Miranda</surname></string-name>, <string-name><given-names>I.</given-names> <surname>Montes</surname></string-name>, <string-name><given-names>E.</given-names> <surname>Quaeghebeur</surname></string-name> and <string-name><given-names>B.</given-names> <surname>Vantaggi</surname></string-name>, eds.) <source>Proceedings of the Thirteenth International Symposium on Imprecise Probability: Theories and Applications</source>. <series>Proceedings of Machine Learning Research</series> <volume>215</volume> <fpage>230</fpage>–<lpage>240</lpage>. <publisher-name>PMLR</publisher-name> (<year>2023</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4663313">MR4663313</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_038">
<label>[38]</label><mixed-citation publication-type="journal"><string-name><surname>Frongillo</surname>, <given-names>R. M.</given-names></string-name> and <string-name><surname>Nobel</surname>, <given-names>A. B.</given-names></string-name> <article-title>Memoryless sequences for general losses</article-title>. <source>J. Mach. Learn. Res.</source> <volume>21</volume> <elocation-id>80</elocation-id> (<year>2020</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4119148">MR4119148</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_039">
<label>[39]</label><mixed-citation publication-type="chapter"><string-name><surname>Gallie</surname>, <given-names>W. B.</given-names></string-name> <chapter-title>Essentially contested concepts</chapter-title>. In <source>Proceedings of the Aristotelian society</source> <volume>56</volume> <fpage>167</fpage>–<lpage>198</lpage> (<year>1955</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_040">
<label>[40]</label><mixed-citation publication-type="journal"><string-name><surname>Gammerman</surname>, <given-names>A.</given-names></string-name> and <string-name><surname>Vovk</surname>, <given-names>V.</given-names></string-name> <article-title>Hedging predictions in machine learning</article-title>. <source>The Computer Journal</source> <volume>50</volume>(<issue>2</issue>) <fpage>151</fpage>–<lpage>163</lpage> (<year>2007</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_041">
<label>[41]</label><mixed-citation publication-type="book"><string-name><surname>Gillies</surname>, <given-names>D.</given-names></string-name> <source>An Objective Theory of Probability</source>. <publisher-name>Routledge</publisher-name> (<year>2010</year>) <comment>(First published 1973)</comment>.</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_042">
<label>[42]</label><mixed-citation publication-type="journal"><string-name><surname>Grädel</surname>, <given-names>E.</given-names></string-name> and <string-name><surname>Väänänen</surname>, <given-names>J.</given-names></string-name> <article-title>Dependence and independence</article-title>. <source>Studia Logica</source> <volume>101</volume>(<issue>2</issue>) <fpage>399</fpage>–<lpage>410</lpage> (<year>2013</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/s11225-013-9479-2" xlink:type="simple">https://doi.org/10.1007/s11225-013-9479-2</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3038039">MR3038039</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_043">
<label>[43]</label><mixed-citation publication-type="journal"><string-name><surname>Gudder</surname>, <given-names>S. P.</given-names></string-name> <article-title>Quantum probability spaces</article-title>. <source>Proceedings of the American Mathematical Society</source> <volume>21</volume>(<issue>2</issue>) <fpage>296</fpage>–<lpage>302</lpage> (<year>1969</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.2307/2036988" xlink:type="simple">https://doi.org/10.2307/2036988</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0243793">MR0243793</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_044">
<label>[44]</label><mixed-citation publication-type="book"><string-name><surname>Gudder</surname>, <given-names>S. P.</given-names></string-name> <source>Stochastic Methods in Quantum Mechanics</source>. <publisher-name>Dover Publications Inc.</publisher-name> (<year>1979</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0543489">MR0543489</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_045">
<label>[45]</label><mixed-citation publication-type="journal"><string-name><surname>Hájek</surname>, <given-names>A.</given-names></string-name> <article-title>Fifteen arguments against hypothetical frequentism</article-title>. <source>Erkenntnis</source> <volume>70</volume>(<issue>2</issue>) <fpage>211</fpage>–<lpage>235</lpage> (<year>2009</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/s10670-009-9154-1" xlink:type="simple">https://doi.org/10.1007/s10670-009-9154-1</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2481794">MR2481794</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_046">
<label>[46]</label><mixed-citation publication-type="chapter"><string-name><surname>Hardt</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Price</surname>, <given-names>E.</given-names></string-name> and <string-name><surname>Srebro</surname>, <given-names>N.</given-names></string-name> <chapter-title>Equality of opportunity in supervised learning</chapter-title>. In <source>Advances in Neural Information Processing Systems</source> <fpage>3323</fpage>–<lpage>3331</lpage> (<year>2016</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_047">
<label>[47]</label><mixed-citation publication-type="chapter"><string-name><surname>Hertweck</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Heitz</surname>, <given-names>C.</given-names></string-name> and <string-name><surname>Loi</surname>, <given-names>M.</given-names></string-name> <chapter-title>On the moral justification of statistical parity</chapter-title>. In <source>Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency</source> <fpage>747</fpage>–<lpage>757</lpage> (<year>2021</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_048">
<label>[48]</label><mixed-citation publication-type="book"><string-name><surname>Hogg</surname>, <given-names>M. A.</given-names></string-name> and <string-name><surname>Abrams</surname>, <given-names>D.</given-names></string-name> <source>Social Identifications: A Social Psychology of Intergroup Relations and Group Processes</source>. <publisher-name>Routledge</publisher-name> (<year>1998</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_049">
<label>[49]</label><mixed-citation publication-type="other"><string-name><surname>Hu</surname>, <given-names>L.</given-names></string-name> and <string-name><surname>Kohler-Hausmann</surname>, <given-names>I.</given-names></string-name> What’s sex got to do with fair machine learning? arXiv preprint arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2006.01770">2006.01770</ext-link> (<year>2020</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_050">
<label>[50]</label><mixed-citation publication-type="journal"><string-name><surname>Humphreys</surname>, <given-names>P. W.</given-names></string-name> <article-title>Randomness, independence, and hypotheses</article-title>. <source>Synthese</source> <volume>36</volume> <fpage>415</fpage>–<lpage>426</lpage> (<year>1977</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/BF00486105" xlink:type="simple">https://doi.org/10.1007/BF00486105</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0517129">MR0517129</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_051">
<label>[51]</label><mixed-citation publication-type="book"><string-name><surname>Ivanenko</surname>, <given-names>V. I.</given-names></string-name> <source>Decision Systems and Nonstochastic Randomness</source>. <publisher-name>Springer</publisher-name> (<year>2010</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/978-1-4419-5548-7" xlink:type="simple">https://doi.org/10.1007/978-1-4419-5548-7</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2606231">MR2606231</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_052">
<label>[52]</label><mixed-citation publication-type="chapter"><string-name><surname>Kamishima</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Akaho</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Asoh</surname>, <given-names>H.</given-names></string-name> and <string-name><surname>Sakuma</surname>, <given-names>J.</given-names></string-name> <chapter-title>Fairness-aware classifier with prejudice remover regularizer</chapter-title>. In <source>Proceedings of the 2012th European Conference on Machine Learning and Knowledge Discovery in Databases-Volume Part II</source> <volume>ECMLPKDD’12</volume>: <fpage>35</fpage>–<lpage>50</lpage> (<year>2012</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_053">
<label>[53]</label><mixed-citation publication-type="chapter"><string-name><surname>Kilbertus</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Rojas Carulla</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Parascandolo</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Hardt</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Janzing</surname>, <given-names>D.</given-names></string-name> and <string-name><surname>Schölkopf</surname>, <given-names>B.</given-names></string-name> <chapter-title>Avoiding discrimination through causal reasoning</chapter-title>. In <source>Advances in Neural Information Processing Systems</source> <volume>30</volume> (<year>2017</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_054">
<label>[54]</label><mixed-citation publication-type="chapter"><string-name><surname>Kleinberg</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Mullainathan</surname>, <given-names>S.</given-names></string-name> and <string-name><surname>Raghavan</surname>, <given-names>M.</given-names></string-name> <chapter-title>Inherent trade-offs in the fair determination of risk scores</chapter-title>. In <source>Leibniz International Proceedings in Informatics, LIPIcs</source> <volume>67</volume> <fpage>1</fpage>–<lpage>22</lpage> (<year>2017</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3754967">MR3754967</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_055">
<label>[55]</label><mixed-citation publication-type="book"><string-name><surname>Kolmogorov</surname>, <given-names>A. N.</given-names></string-name> <source>Grundbegriffe de Wahrscheinlichkeitsrechnung</source>. <publisher-name>Springer</publisher-name> (<year>1933</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0362415">MR0362415</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_056">
<label>[56]</label><mixed-citation publication-type="book"><string-name><surname>Kolmogorov</surname>, <given-names>A. N.</given-names></string-name> <source>Foundations of the Theory of Probability: Second English Edition</source>. <publisher-name>Chelsea Publishing Company</publisher-name> (<year>1956</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0079843">MR0079843</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_057">
<label>[57]</label><mixed-citation publication-type="journal"><string-name><surname>Kolmogorov</surname>, <given-names>A. N.</given-names></string-name> <article-title>Three approaches to the definition of the concept “quantity of information”</article-title>. <source>Problemy Peredachi Informatsii</source> <volume>1</volume>(<issue>1</issue>) <fpage>3</fpage>–<lpage>11</lpage> (<year>1965</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0184801">MR0184801</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_058">
<label>[58]</label><mixed-citation publication-type="journal"><string-name><surname>Lafferty</surname>, <given-names>J.</given-names></string-name> and <string-name><surname>Wasserman</surname>, <given-names>L.</given-names></string-name> <article-title>Challenges in statistical machine learning</article-title>. <source>Statistica Sinica</source> <volume>16</volume>(<issue>2</issue>) <fpage>307</fpage> (<year>2006</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2267237">MR2267237</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_059">
<label>[59]</label><mixed-citation publication-type="journal"><string-name><surname>Levin</surname>, <given-names>L. A.</given-names></string-name> <article-title>The concept of random sequence</article-title>. <source>Soviet Mathematics Doklady</source> <volume>14</volume>. <fpage>1413</fpage>–<lpage>1416</lpage> (<year>1973</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0366096">MR0366096</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_060">
<label>[60]</label><mixed-citation publication-type="other"><string-name><surname>Levin</surname>, <given-names>L. A.</given-names></string-name> A concept of independence with applications in various fields of mathematics. Technical Report MIT/LCS/TR-235, MIT, Laboratory for Computer Science, April <year>1980</year>.</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_061">
<label>[61]</label><mixed-citation publication-type="journal"><string-name><surname>Lippert-Rasmussen</surname>, <given-names>K.</given-names></string-name> <article-title>The badness of discrimination</article-title>. <source>Ethical Theory and Moral Practice</source> <volume>9</volume>(<issue>2</issue>) <fpage>167</fpage>–<lpage>185</lpage> (<year>2006</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_062">
<label>[62]</label><mixed-citation publication-type="book"><string-name><surname>Vitali</surname>, <given-names>L.</given-names></string-name> <source>Analysis 1 Lecture Notes 2013/2014</source>. <publisher-name>University of Bristol</publisher-name>, <year>2014</year>.</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_063">
<label>[63]</label><mixed-citation publication-type="chapter"><string-name><surname>Lohaus</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Perrot</surname>, <given-names>M.</given-names></string-name> and <string-name><surname>Von Luxburg</surname>, <given-names>U.</given-names></string-name> <chapter-title>Too relaxed to be fair</chapter-title>. In <source>International Conference on Machine Learning</source> <fpage>6360</fpage>–<lpage>6369</lpage>. <publisher-name>PMLR</publisher-name> (<year>2020</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_064">
<label>[64]</label><mixed-citation publication-type="other"><string-name><surname>Loi</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Herlitz</surname>, <given-names>A.</given-names></string-name> and <string-name><surname>Heidari</surname>, <given-names>H.</given-names></string-name> A philosophical theory of fairness for prediction-based decisions. Technical report, Politecnico di Milano, <year>2019</year>. <uri>https://ssrn.com/abstract=3450300</uri></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_065">
<label>[65]</label><mixed-citation publication-type="journal"><string-name><surname>Martin-Löf</surname>, <given-names>P.</given-names></string-name> <article-title>The definition of random sequences</article-title>. <source>Information and Control</source> <volume>9</volume>(<issue>6</issue>) <fpage>602</fpage>–<lpage>619</lpage> (<year>1966</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0223179">MR0223179</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_066">
<label>[66]</label><mixed-citation publication-type="book"><string-name><surname>McGarty</surname>, <given-names>C.</given-names></string-name> <source>Categorization in Social Psychology</source>. <publisher-name>Sage Publications</publisher-name> (<year>1999</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_067">
<label>[67]</label><mixed-citation publication-type="chapter"><string-name><surname>Menon</surname>, <given-names>A. K.</given-names></string-name> and <string-name><surname>Williamson</surname>, <given-names>R. C.</given-names></string-name> <chapter-title>The cost of fairness in binary classification</chapter-title>. In <source>Conference on Fairness, Accountability and Transparency</source> <fpage>107</fpage>–<lpage>118</lpage>. <publisher-name>PMLR</publisher-name> (<year>2018</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_068">
<label>[68]</label><mixed-citation publication-type="journal"><string-name><surname>Muchnik</surname>, <given-names>A. A.</given-names></string-name>, <string-name><surname>Semenov</surname>, <given-names>A. L.</given-names></string-name> and <string-name><surname>Uspensky</surname>, <given-names>V. A.</given-names></string-name> <article-title>Mathematical metaphysics of randomness</article-title>. <source>Theoretical Computer Science</source> <volume>207</volume>(<issue>2</issue>) <fpage>263</fpage>–<lpage>317</lpage> (<year>1998</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1016/S0304-3975(98)00069-3" xlink:type="simple">https://doi.org/10.1016/S0304-3975(98)00069-3</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1643438">MR1643438</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_069">
<label>[69]</label><mixed-citation publication-type="journal"><string-name><surname>Murchison</surname>, <given-names>B. C.</given-names></string-name> <article-title>The concept of independence in public law</article-title>. <source>Emory Law Journal</source> <volume>41</volume>. <fpage>961</fpage> (<year>1992</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_070">
<label>[70]</label><mixed-citation publication-type="journal"><string-name><surname>Narens</surname>, <given-names>L.</given-names></string-name> <article-title>An introduction to lattice based probability theories</article-title>. <source>Journal of Mathematical Psychology</source> <volume>74</volume>. <fpage>66</fpage>–<lpage>81</lpage> (<year>2016</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1016/j.jmp.2016.04.013" xlink:type="simple">https://doi.org/10.1016/j.jmp.2016.04.013</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3552130">MR3552130</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_071">
<label>[71]</label><mixed-citation publication-type="book"><string-name><surname>Nathanson</surname>, <given-names>M. B.</given-names></string-name> <source>Elementary Methods in Number Theory</source> <volume>195</volume>. <publisher-name>Springer</publisher-name> (<year>2008</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1732941">MR1732941</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_072">
<label>[72]</label><mixed-citation publication-type="journal"><string-name><surname>Naylor</surname>, <given-names>A. W.</given-names></string-name> <article-title>On decomposition theory: generalized dependence</article-title>. <source>IEEE Transactions on Systems, Man, and Cybernetics</source> <volume>11</volume>(<issue>10</issue>) <fpage>699</fpage>–<lpage>713</lpage> (<year>1981</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1109/TSMC.1981.4308590" xlink:type="simple">https://doi.org/10.1109/TSMC.1981.4308590</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0641948">MR0641948</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_073">
<label>[73]</label><mixed-citation publication-type="journal"><string-name><surname>Donald</surname></string-name> and <string-name><surname>Ornstein</surname>, <given-names>S.</given-names></string-name> <article-title>Ergodic theory, randomness, and “chaos”</article-title>. <source>Science</source> <volume>243</volume>(<issue>4888</issue>) <fpage>182</fpage>–<lpage>187</lpage> (<year>1989</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1126/science.243.4888.182" xlink:type="simple">https://doi.org/10.1126/science.243.4888.182</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0981173">MR0981173</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_074">
<label>[74]</label><mixed-citation publication-type="other"><string-name><surname>Parker</surname>, <given-names>J. M.</given-names></string-name> <italic>Randomness and legitimacy in selecting democratic representatives</italic>. The University of Texas at Austin, <year>2011</year>.</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_075">
<label>[75]</label><mixed-citation publication-type="chapter"><string-name><surname>Passi</surname>, <given-names>S.</given-names></string-name> and <string-name><surname>Barocas</surname>, <given-names>S.</given-names></string-name> <chapter-title>Problem formulation and fairness</chapter-title>. In <source>Proceedings of the Conference on Fairness, Accountability, and Transparency</source> <fpage>39</fpage>–<lpage>48</lpage> (<year>2019</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_076">
<label>[76]</label><mixed-citation publication-type="book"><string-name><surname>Popper</surname>, <given-names>K.</given-names></string-name> <source>The Logic of Scientific Discovery</source>. <publisher-name>Routledge</publisher-name> (<year>2010</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1195353">MR1195353</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_077">
<label>[77]</label><mixed-citation publication-type="other"><string-name><surname>Porter</surname>, <given-names>C. P.</given-names></string-name> <italic>Mathematical and philosophical perspectives on algorithmic randomness</italic>. PhD thesis, University of Notre Dame, <year>2012</year>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3217940">MR3217940</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_078">
<label>[78]</label><mixed-citation publication-type="journal"><string-name><surname>Pták</surname>, <given-names>P.</given-names></string-name> <article-title>Concrete quantum logics</article-title>. <source>International Journal of Theoretical Physics</source> <volume>39</volume>(<issue>3</issue>) <fpage>827</fpage>–<lpage>837</lpage> (<year>2000</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1023/A:1003626929648" xlink:type="simple">https://doi.org/10.1023/A:1003626929648</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1792201">MR1792201</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_079">
<label>[79]</label><mixed-citation publication-type="book"><string-name><surname>Putnam</surname>, <given-names>H.</given-names></string-name> <source>The Meaning of the Concept of Probability in Application to Finite Sequences (Routledge Revivals)</source>. <publisher-name>Garland Publishing</publisher-name> (<year>1990</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_080">
<label>[80]</label><mixed-citation publication-type="chapter"><string-name><surname>Raghavan</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Barocas</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Kleinberg</surname>, <given-names>J.</given-names></string-name> and <string-name><surname>Levy</surname>, <given-names>K.</given-names></string-name> <chapter-title>Mitigating bias in algorithmic hiring: Evaluating claims and practices</chapter-title>. In <source>Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency</source> <fpage>469</fpage>–<lpage>481</lpage> (<year>2020</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_081">
<label>[81]</label><mixed-citation publication-type="book"><string-name><surname>Bhaskara Rao</surname>, <given-names>K. P. S.</given-names></string-name> and <string-name><surname>Bhaskara Rao</surname>, <given-names>M.</given-names></string-name> <source>Theory of Charges: A Study of Finitely Additive Measures</source>. <publisher-name>Academic Press</publisher-name> (<year>1983</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0751777">MR0751777</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_082">
<label>[82]</label><mixed-citation publication-type="book"><string-name><surname>Rawls</surname>, <given-names>J.</given-names></string-name> <source>A Theory of Justice</source>. <publisher-name>The Belknap Press of Harvard University Press</publisher-name> (<year>1971</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_083">
<label>[83]</label><mixed-citation publication-type="chapter"><string-name><surname>Räz</surname>, <given-names>T.</given-names></string-name> <chapter-title>Group fairness: Independence revisited</chapter-title>. In <source>Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency</source> <fpage>129</fpage>–<lpage>137</lpage> (<year>2021</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4297512">MR4297512</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_084">
<label>[84]</label><mixed-citation publication-type="journal"><string-name><surname>Scantamburlo</surname>, <given-names>T.</given-names></string-name> <article-title>Non-empirical problems in fair machine learning</article-title>. <source>Ethics and Information Technology</source> <volume>23</volume>(<issue>4</issue>) <fpage>703</fpage>–<lpage>712</lpage> (<year>2021</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_085">
<label>[85]</label><mixed-citation publication-type="book"><string-name><surname>Schnorr</surname>, <given-names>C.-P.</given-names></string-name> <source>Zufälligkeit und Wahrscheinlichkeit: eine algorithmische Begründung der Wahrscheinlichkeitstheorie</source>. <series>Lecture Notes in Mathematics</series> <volume>218</volume>. <publisher-name>Springer</publisher-name> (<year>1971</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0414225">MR0414225</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_086">
<label>[86]</label><mixed-citation publication-type="chapter"><string-name><surname>Schnorr</surname>, <given-names>C.-P.</given-names></string-name> <chapter-title>The process complexity and effective random tests</chapter-title>. In <source>Proceedings of the Fourth Annual ACM Symposium on Theory of Computing</source> <fpage>168</fpage>–<lpage>176</lpage> (<year>1972</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1016/S0022-0000(73)80030-3" xlink:type="simple">https://doi.org/10.1016/S0022-0000(73)80030-3</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0325366">MR0325366</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_087">
<label>[87]</label><mixed-citation publication-type="chapter"><string-name><surname>Schnorr</surname>, <given-names>C.-P.</given-names></string-name> <chapter-title>A survey of the theory of random sequences</chapter-title>. In <source>Basic Problems in Methodology and Linguistics</source> <fpage>193</fpage>–<lpage>211</lpage>. <publisher-name>Springer</publisher-name> (<year>1977</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0517133">MR0517133</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_088">
<label>[88]</label><mixed-citation publication-type="journal"><string-name><surname>Schurz</surname>, <given-names>G.</given-names></string-name> and <string-name><surname>Leitgeb</surname>, <given-names>H.</given-names></string-name> <article-title>Finitistic and frequentistic approximation of probability measures with or without <italic>σ</italic>-additivity</article-title>. <source>Studia Logica</source> <volume>89</volume>(<issue>2</issue>) <fpage>257</fpage>–<lpage>283</lpage> (<year>2008</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/s11225-008-9128-3" xlink:type="simple">https://doi.org/10.1007/s11225-008-9128-3</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2429951">MR2429951</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_089">
<label>[89]</label><mixed-citation publication-type="journal"><string-name><surname>Shafer</surname>, <given-names>G.</given-names></string-name> <article-title>Discussion on hedging predictions in machine learning by A. Gammerman and V. Vovk</article-title>. <source>The Computer Journal</source> <volume>50</volume>(<issue>2</issue>) <fpage>164</fpage>–<lpage>172</lpage> (<year>2007</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1093/comjnl/bxl066" xlink:type="simple">https://doi.org/10.1093/comjnl/bxl066</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_090">
<label>[90]</label><mixed-citation publication-type="book"><string-name><surname>Shafer</surname>, <given-names>G.</given-names></string-name> and <string-name><surname>Vovk</surname>, <given-names>V.</given-names></string-name> <source>Game-Theoretic Foundations for Probability and Finance</source> <volume>455</volume>. <publisher-name>John Wiley &amp; Sons</publisher-name> (<year>2019</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1002/0471249696" xlink:type="simple">https://doi.org/10.1002/0471249696</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1852450">MR1852450</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_091">
<label>[91]</label><mixed-citation publication-type="book"><string-name><surname>Shalev-Shwartz</surname>, <given-names>S.</given-names></string-name> and <string-name><surname>Ben-David</surname>, <given-names>S.</given-names></string-name> <source>Understanding Machine Learning: From Theory to Algorithms</source>. <publisher-name>Cambridge University Press</publisher-name> (<year>2014</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3277164">MR3277164</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_092">
<label>[92]</label><mixed-citation publication-type="chapter"><string-name><surname>Simpson</surname>, <given-names>J.</given-names></string-name> and <string-name><surname>Weine</surname>, <given-names>E.</given-names></string-name> <chapter-title>“independence, n”</chapter-title>. In <source>Oxford English Dictionary</source>, <edition>2</edition>nd ed. <publisher-name>Oxford University Press</publisher-name> (<year>1989</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_093">
<label>[93]</label><mixed-citation publication-type="journal"><string-name><surname>Spohn</surname>, <given-names>W.</given-names></string-name> <article-title>Stochastic independence, causal independence, and shieldability</article-title>. <source>Journal of Philosophical Logic</source> <volume>9</volume>(<issue>1</issue>) <fpage>73</fpage>–<lpage>99</lpage> (<year>1980</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/BF00258078" xlink:type="simple">https://doi.org/10.1007/BF00258078</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0563250">MR0563250</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_094">
<label>[94]</label><mixed-citation publication-type="journal"><string-name><surname>Steinwart</surname>, <given-names>I.</given-names></string-name>, <string-name><surname>Hush</surname>, <given-names>D.</given-names></string-name> and <string-name><surname>Scovel</surname>, <given-names>C.</given-names></string-name> <article-title>Learning from dependent observations</article-title>. <source>Journal of Multivariate Analysis</source> <volume>100</volume>(<issue>1</issue>) <fpage>175</fpage>–<lpage>194</lpage> (<year>2009</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1016/j.jmva.2008.04.001" xlink:type="simple">https://doi.org/10.1016/j.jmva.2008.04.001</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2460486">MR2460486</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_095">
<label>[95]</label><mixed-citation publication-type="book"><string-name><surname>Stone</surname>, <given-names>P.</given-names></string-name> <source>The Luck of the Draw: The Role of Lotteries in Decision-Making</source>. <publisher-name>Oxford University Press</publisher-name> (<year>2011</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_096">
<label>[96]</label><mixed-citation publication-type="chapter"><string-name><surname>Tadaki</surname>, <given-names>K.</given-names></string-name> <chapter-title>An operational characterization of the notion of probability by algorithmic randomness</chapter-title>. In <source>Proceedings of the 37th Symposium on Information Theory and its Applications (SITA2014)</source> <volume>5</volume> <fpage>389</fpage>–<lpage>394</lpage> (<year>2014</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_097">
<label>[97]</label><mixed-citation publication-type="journal"><string-name><surname>Tajfel</surname>, <given-names>H.</given-names></string-name> <article-title>Social identity and intergroup behaviour</article-title>. <source>Social Science Information</source> <volume>13</volume>(<issue>2</issue>) <fpage>65</fpage>–<lpage>93</lpage> (<year>1974</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_098">
<label>[98]</label><mixed-citation publication-type="other"><string-name><surname>Tao</surname>, <given-names>T.</given-names></string-name> 275a, notes 2: Product measures and independence, October <year>2015</year>. <uri>https://terrytao.wordpress.com/2015/10/12/275a-notes-2-product-measures-and-independence/</uri></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_099">
<label>[99]</label><mixed-citation publication-type="journal"><string-name><surname>Uspenskii</surname>, <given-names>V. A.</given-names></string-name>, <string-name><surname>Semenov</surname>, <given-names>A. L.</given-names></string-name> and <string-name><surname>Shen</surname>, <given-names>A. Kh.</given-names></string-name> <article-title>Can an individual sequence of zeros and ones be random?</article-title> <source>Russian Mathematical Surveys</source> <volume>45</volume>(<issue>1</issue>) <fpage>121</fpage> (<year>1990</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1070/RM1990v045n01ABEH002321" xlink:type="simple">https://doi.org/10.1070/RM1990v045n01ABEH002321</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1050929">MR1050929</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_100">
<label>[100]</label><mixed-citation publication-type="journal"><string-name><surname>Van Lambalgen</surname>, <given-names>M.</given-names></string-name> <article-title>Von Mises’ definition of random sequences reconsidered</article-title>. <source>The Journal of Symbolic Logic</source> <volume>52</volume>(<issue>3</issue>) <fpage>725</fpage>–<lpage>755</lpage> (<year>1987</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.2307/2274360" xlink:type="simple">https://doi.org/10.2307/2274360</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0902987">MR0902987</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_101">
<label>[101]</label><mixed-citation publication-type="journal"><string-name><surname>Van Lambalgen</surname>, <given-names>M.</given-names></string-name> <article-title>The axiomatization of randomness</article-title>. <source>The Journal of Symbolic Logic</source> <volume>55</volume>(<issue>3</issue>) <fpage>1143</fpage>–<lpage>1167</lpage> (<year>1990</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.2307/2274480" xlink:type="simple">https://doi.org/10.2307/2274480</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1071321">MR1071321</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_102">
<label>[102]</label><mixed-citation publication-type="book"><string-name><surname>Ville</surname>, <given-names>J.</given-names></string-name> <source>Étude critique de la notion de collectif</source>. <publisher-name>Gauthier-Villars</publisher-name> (<year>1939</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3533075">MR3533075</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_103">
<label>[103]</label><mixed-citation publication-type="journal"><string-name><surname>Volchan</surname>, <given-names>S. B.</given-names></string-name> <article-title>What is a random sequence?</article-title> <source>The American Mathematical Monthly</source> <volume>109</volume>(<issue>1</issue>) <fpage>46</fpage>–<lpage>63</lpage> (<year>2002</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.2307/2695767" xlink:type="simple">https://doi.org/10.2307/2695767</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1903512">MR1903512</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_104">
<label>[104]</label><mixed-citation publication-type="journal"><string-name><surname>Von Collani</surname>, <given-names>E.</given-names></string-name> <article-title>A note on the concept of independence</article-title>. <source>Economic Quality Control</source> <volume>21</volume>(<issue>1</issue>) <fpage>155</fpage>–<lpage>164</lpage> (<year>2006</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_105">
<label>[105]</label><mixed-citation publication-type="journal"><string-name><surname>Von Mises</surname>, <given-names>R.</given-names></string-name> <article-title>Grundlagen der Wahrscheinlichkeitsrechnung</article-title>. <source>Mathematische Zeitschrift</source> <volume>5</volume>(<issue>1</issue>) <fpage>52</fpage>–<lpage>99</lpage> (<year>1919</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/BF01203155" xlink:type="simple">https://doi.org/10.1007/BF01203155</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1544374">MR1544374</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_106">
<label>[106]</label><mixed-citation publication-type="book"><string-name><surname>Von Mises</surname>, <given-names>R.</given-names></string-name> <source>Probability, Statistics, and Truth</source>. <publisher-name>Dover Publications, Inc.</publisher-name> (<year>1981</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0668875">MR0668875</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_107">
<label>[107]</label><mixed-citation publication-type="book"><string-name><surname>Von Mises</surname>, <given-names>R.</given-names></string-name> and <string-name><surname>Geiringer</surname>, <given-names>H.</given-names></string-name> <source>Mathematical Theory of Probability and Statistics</source>. <publisher-name>Academic Press</publisher-name> (<year>1964</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0178486">MR0178486</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_108">
<label>[108]</label><mixed-citation publication-type="journal"><string-name><surname>Vovk</surname>, <given-names>V. G.</given-names></string-name> <article-title>A logic of probability, with application to the foundations of statistics</article-title>. <source>Journal of the Royal Statistical Society: Series B (Methodological)</source> <volume>55</volume>(<issue>2</issue>) <fpage>317</fpage>–<lpage>341</lpage> (<year>1993</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1224399">MR1224399</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_109">
<label>[109]</label><mixed-citation publication-type="book"><string-name><surname>Wasserman</surname>, <given-names>L.</given-names></string-name> <source>All of Statistics: A Concise Course in Statistical Inference</source>. <publisher-name>Springer</publisher-name> (<year>2004</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/978-0-387-21736-9" xlink:type="simple">https://doi.org/10.1007/978-0-387-21736-9</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2055670">MR2055670</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds73_ref_110">
<label>[110]</label><mixed-citation publication-type="chapter"><string-name><surname>Wick</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Panda</surname>, <given-names>S.</given-names></string-name> and <string-name><surname>Tristan</surname>, <given-names>J.-B.</given-names></string-name> <chapter-title>Unlocking fairness: a trade-off revisited</chapter-title>. In <source>Proceedings of the 37th International Conference on Machine Learning</source> <volume>32</volume> (<year>2020</year>).</mixed-citation>
</ref>
<ref id="j_nejsds73_ref_111">
<label>[111]</label><mixed-citation publication-type="chapter"><string-name><surname>Williamson</surname>, <given-names>R. C.</given-names></string-name> and <string-name><surname>Menon</surname>, <given-names>A.</given-names></string-name> <chapter-title>Fairness risk measures</chapter-title>. In <source>International Conference on Machine Learning</source> <fpage>6786</fpage>–<lpage>6797</lpage>. <publisher-name>PMLR</publisher-name> (<year>2019</year>).</mixed-citation>
</ref>
</ref-list>
</back>
</article>
