<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.0 20120330//EN" "JATS-journalpublishing1.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">NEJSDS</journal-id>
<journal-title-group><journal-title>The New England Journal of Statistics in Data Science</journal-title></journal-title-group>
<issn pub-type="ppub">2693-7166</issn><issn-l>2693-7166</issn-l>
<publisher>
<publisher-name>New England Statistical Society</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">NEJSDS100</article-id>
<article-id pub-id-type="doi">10.51387/26-NEJSDS100</article-id>
<article-id pub-id-type="arxiv">2204.14121</article-id>
<article-categories><subj-group subj-group-type="heading">
<subject>Methodology Article</subject></subj-group><subj-group subj-group-type="area">
<subject>Statistical Methodology</subject></subj-group></article-categories>
<title-group>
<article-title>Inverse Probability Weighting: From Survey Sampling to Evidence Estimation</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0001-5991-5182</contrib-id>
<name><surname>Datta</surname><given-names>Jyotishka</given-names></name><email xlink:href="mailto:jyotishka@vt.edu">jyotishka@vt.edu</email><xref ref-type="aff" rid="j_nejsds100_aff_001"/><xref ref-type="corresp" rid="cor1">∗</xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Polson</surname><given-names>Nicholas</given-names></name><email xlink:href="mailto:ngp@chicagobooth.edu">ngp@chicagobooth.edu</email><xref ref-type="aff" rid="j_nejsds100_aff_002"/>
</contrib>
<aff id="j_nejsds100_aff_001">Jyotishka Datta is Associate Professor, Department of Statistics, <institution>Virginia Polytechnic Institute and State University</institution>, Blacksburg, VA 24061, <country>USA</country>. E-mail address: <email xlink:href="mailto:jyotishka@vt.edu">jyotishka@vt.edu</email></aff>
<aff id="j_nejsds100_aff_002">Nicholas G. Polson is Professor, Booth School of Business, <institution>University of Chicago</institution>, <country>USA</country>. E-mail address: <email xlink:href="mailto:ngp@chicagobooth.edu">ngp@chicagobooth.edu</email></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>∗</label>Corresponding author.</corresp>
</author-notes>
<pub-date pub-type="ppub"><year>2026</year></pub-date><pub-date pub-type="epub"><day>13</day><month>4</month><year>2026</year></pub-date><volume content-type="ahead-of-print">0</volume><issue>0</issue><fpage>1</fpage><lpage>13</lpage><history><date date-type="accepted"><day>3</day><month>2</month><year>2026</year></date></history>
<permissions><copyright-statement>© 2026 New England Statistical Society</copyright-statement><copyright-year>2026</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>Open access article under the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/">CC BY</ext-link> license.</license-p></license></permissions>
<abstract>
<p>We consider the class of inverse probability weight (IPW) estimators, including the popular Horvitz–Thompson and Hájek estimators used routinely in survey sampling, causal inference and for Bayesian computation. We focus on the ‘weak paradoxes’ for these estimators due to two counterexamples by <xref ref-type="bibr" rid="j_nejsds100_ref_001">Basu</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_001">1988</xref>) and <xref ref-type="bibr" rid="j_nejsds100_ref_054">Wasserman</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_054">2004</xref>) and investigate the two natural Bayesian answers to this problem: one based on binning and smoothing: a ‘Bayesian sieve’ and the other based on a conjugate hierarchical model that allows borrowing information via exchangeability. We compare the mean squared errors for the two Bayesian estimators with the IPW estimators for Wasserman’s example via simulation studies on a broad range of parameter configurations. We also prove posterior consistency for the Bayes estimators under missing-completely-at-random assumption and show that it requires fewer assumptions on the inclusion probabilities. We also revisit the connection between the different problems where improved or adaptive IPW estimators will be useful, including survey sampling, evidence estimation strategies such as Conditional Monte Carlo, Riemannian sum, Trapezoidal rules and vertical likelihood, as well as average treatment effect estimation in causal inference.</p>
</abstract>
<kwd-group>
<label>Keywords and phrases</label>
<kwd>Inverse probability weighting</kwd>
<kwd>Horvitz–Thompson</kwd>
<kwd>Hájek</kwd>
<kwd>Importance sampling</kwd>
<kwd>Stein phenomenon</kwd>
<kwd>Bias-variance trade-off</kwd>
<kwd>Evidence Estimation</kwd>
</kwd-group>
<funding-group><funding-statement>Dr. Datta acknowledges support from the National Science Foundation (NSF DMS-2015460 and NSF CAREER 2443282).</funding-statement></funding-group>
</article-meta>
</front>
<body>
<sec id="j_nejsds100_s_001">
<label>1</label>
<title>Introduction</title>
<p>Inverse probability weight (IPW) estimators have been used across statistical literature in diverse forms: in survey sampling, in designing importance sampling in Monte Carlo techniques and in the context of average treatment effect estimation in causal inference. In survey sampling, the goal is often to estimate population mean <italic>ψ</italic> from a finite sample (<inline-formula id="j_nejsds100_ineq_001"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${y_{1}},\dots ,{y_{n}})$]]></tex-math></alternatives></inline-formula>, and a common approach is to weigh each observation <inline-formula id="j_nejsds100_ineq_002"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${y_{i}}$]]></tex-math></alternatives></inline-formula> by a weight <inline-formula id="j_nejsds100_ineq_003"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">w</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${w_{i}}$]]></tex-math></alternatives></inline-formula> inversely related to their probability of inclusion (probability proportional to selection, or PPS). For example, common PPS estimators admit the form <inline-formula id="j_nejsds100_ineq_004"><alternatives><mml:math><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">w</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi></mml:math><tex-math><![CDATA[$\hat{\psi }={\textstyle\sum _{i\in s}}{w_{i}}{y_{i}}/n$]]></tex-math></alternatives></inline-formula> or a ‘ratio’ estimator: <inline-formula id="j_nejsds100_ineq_005"><alternatives><mml:math><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">w</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">w</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$\hat{\psi }={\textstyle\sum _{i\in s}}{w_{i}}{y_{i}}/{\textstyle\sum _{i\in s}}{w_{i}}$]]></tex-math></alternatives></inline-formula>, where <italic>s</italic> denotes the sample, and <inline-formula id="j_nejsds100_ineq_006"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">w</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${w_{i}}$]]></tex-math></alternatives></inline-formula>’s denote the sampling weights. These weights or probabilities of inclusion might be known, or unknown depending on whether their source is sampling design or non-response. Two of the most popular examples of such estimators are the Horvitz–Thompson estimator (<xref ref-type="bibr" rid="j_nejsds100_ref_018">Horvitz and Thompson</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_018">1952</xref>), which uses fixed weights to provide an unbiased estimate of the population mean, and the Hájek estimator (<xref ref-type="bibr" rid="j_nejsds100_ref_019">Hájek</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_019">1971</xref>), which normalizes the weights to improve stability at the cost of introducing slight bias.</p>
<p>Similarly, in importance sampling, the goal is to estimate the evidence <inline-formula id="j_nejsds100_ineq_007"><alternatives><mml:math>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo>=</mml:mo>
<mml:mo largeop="false" movablelimits="false">∫</mml:mo>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">d</mml:mi>
<mml:mi mathvariant="italic">F</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\psi ={\mathbb{E}_{F}}\{l(X)\}=\textstyle\int l(x)dF(x)$]]></tex-math></alternatives></inline-formula> where <italic>F</italic> might be difficult to sample from and a common technique is to estimate <italic>ψ</italic> by drawing samples from a candidate distribution <italic>G</italic> (with density <italic>g</italic>) and calculate: 
<disp-formula id="j_nejsds100_eq_001">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>≐</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">w</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \hat{\psi }={n^{-1}}{\sum \limits_{i=1}^{n}}l({x_{i}})f({x_{i}})/g({x_{i}})\doteq {n^{-1}}{\sum \limits_{i=1}^{n}}l({x_{i}})w({x_{i}}).\]]]></tex-math></alternatives>
</disp-formula> 
The candidate density <inline-formula id="j_nejsds100_ineq_008"><alternatives><mml:math>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mo>·</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$g(\cdot )$]]></tex-math></alternatives></inline-formula> is chosen such that the ‘weights’ <inline-formula id="j_nejsds100_ineq_009"><alternatives><mml:math>
<mml:mi mathvariant="italic">w</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$w({x_{i}})$]]></tex-math></alternatives></inline-formula> is nearly constant (<xref ref-type="bibr" rid="j_nejsds100_ref_011">Firth</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_011">2011</xref>). The ‘ratio’ estimator analog in importance sampling would correspond to <inline-formula id="j_nejsds100_ineq_010"><alternatives><mml:math>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">w</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mi mathvariant="italic">w</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${\textstyle\sum _{i=1}^{n}}l({x_{i}})w({x_{i}})/{\textstyle\sum _{i=1}^{n}}w({x_{i}})$]]></tex-math></alternatives></inline-formula> as developed by <xref ref-type="bibr" rid="j_nejsds100_ref_017">Hesterberg</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_017">1988</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_016">1995</xref>), and further developed in (<xref ref-type="bibr" rid="j_nejsds100_ref_011">Firth</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_011">2011</xref>). The unattainable ‘optimal’ choice of <inline-formula id="j_nejsds100_ineq_011"><alternatives><mml:math>
<mml:mi mathvariant="italic">w</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mo>·</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$w(\cdot )$]]></tex-math></alternatives></inline-formula> is of course <inline-formula id="j_nejsds100_ineq_012"><alternatives><mml:math>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mo>·</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$f(\cdot )$]]></tex-math></alternatives></inline-formula> itself, and a key insight in producing more accurate estimation is that self-normalization or biasing the sampler towards low probability regions can help. Methods such as nested sampling or vertical likelihood use a Lorenz curve re-ordering of summation to achieve this goal (<xref ref-type="bibr" rid="j_nejsds100_ref_049">Skilling</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_049">2006</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_005">Chopin and Robert</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_005">2010</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_037">Polson and Scott</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_037">2014</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_007">Datta and Polson</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_007">2025</xref>).</p>
<p>A third important setting where inverse probability weighting plays a central role is causal inference. In the potential outcomes framework, IPW estimators form the backbone of average treatment effect (ATE) estimation by reweighting observed outcomes according to treatment assignment probabilities (<xref ref-type="bibr" rid="j_nejsds100_ref_043">Rubin</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_043">1974</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_020">Imbens</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_020">2004</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_006">Cunningham</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_006">2021</xref>). We briefly discuss the ATE problem in subsection <xref rid="j_nejsds100_s_009">2.3</xref>. IPW estimators provide a common design-based principle whether adjusting for unequal inclusion probabilities in surveys, stabilizing evidence estimation in Bayesian computation, or addressing confounding in observational studies. The wide applicability of IPW motivates our reexamination of their weaknesses and possible Bayesian resolutions.</p>
<p>In this paper, we first contrast and compare the popular inverse probability weight estimators—Horvitz–Thompson and Hájek—and discuss their relative merits and demerits in light of two weak paradoxes, viz., Basu’s circus example (<xref ref-type="bibr" rid="j_nejsds100_ref_001">Basu</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_001">1988</xref>) and the Robins–Ritov–Wasserman example (<xref ref-type="bibr" rid="j_nejsds100_ref_042">Robins and Ritov</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_042">1997</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_054">Wasserman</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_054">2004</xref>). We then discuss a binning-and-smoothing estimator by <xref ref-type="bibr" rid="j_nejsds100_ref_012">Ghosh</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_012">2015</xref>) and a hierarchical Bayes estimator by <xref ref-type="bibr" rid="j_nejsds100_ref_026">Li</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_026">2010</xref>), in the light of the <xref ref-type="bibr" rid="j_nejsds100_ref_054">Wasserman</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_054">2004</xref>)’s example, and show how they can lead to a possible resolution. We also build connections between IPW estimators and evidence estimation techniques and argue that these innovative ideas from Monte Carlo methods can be exploited in designing IPW estimators to achieve a lower variance and higher stability. We argue that the issue of choice of weights – an optimal proposal <inline-formula id="j_nejsds100_ineq_013"><alternatives><mml:math>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mo>·</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$g(\cdot )$]]></tex-math></alternatives></inline-formula> in Monte Carlo or sampling weights – provides new insights on a long-standing controversy about a ‘weakness’ in Bayesian paradigm (e.g. <xref ref-type="bibr" rid="j_nejsds100_ref_054">Wasserman</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_054">2004</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_046">Sims</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_046">2010</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_012">Ghosh</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_012">2015</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_026">Li</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_026">2010</xref>).</p>
<p>There is a large, influential literature on statistical methods for improving survey-weighted estimates, particularly, in the context of complex survey designs. Our goal is not to attempt a review of these methodological advances in survey weight regularization or calibration, and we point the interested reader to the comprehensive review by <xref ref-type="bibr" rid="j_nejsds100_ref_004">Chen et al.</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_004">2017</xref>) and the references therein, e.g., <xref ref-type="bibr" rid="j_nejsds100_ref_015">Haziza and Beaumont</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_015">2017</xref>). <xref ref-type="bibr" rid="j_nejsds100_ref_015">Haziza and Beaumont</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_015">2017</xref>) provide a comprehensive review of methods for constructing survey weights, focusing on techniques to adjust for unequal probabilities of selection, nonresponse, and post-stratification to improve the representativeness and accuracy of survey estimates. <xref ref-type="bibr" rid="j_nejsds100_ref_004">Chen et al.</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_004">2017</xref>) discuss the limitations of basic design-based weights, derived from the inverse of inclusion probabilities, and propose modifications such as weight trimming, weight modeling, and incorporating weights into statistical models.</p>
<p>The structure of this article is as follows. In ğ<xref rid="j_nejsds100_s_002">2</xref>, we define the popular IPW estimators and recent developments, and their connections with importance sampling and ATE estimation. In ğ<xref rid="j_nejsds100_s_010">3</xref>, we discuss two popular examples of ‘weak paradoxes’ due to Basu, and Robins-Ritov-Wasserman. For the latter, we derive asymptotic properties of a Bayes estimator due to <xref ref-type="bibr" rid="j_nejsds100_ref_026">Li</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_026">2010</xref>) in ğ<xref rid="j_nejsds100_s_018">3.4</xref>. We compare different survey sampling estimators through a range of simulation studies in ğ<xref rid="j_nejsds100_s_019">4</xref>, to illustrate their relative merits and demerits. Finally, in ğ<xref rid="j_nejsds100_s_023">5</xref>, we discuss other areas of connection such as average treatment effect estimation in potential outcomes framework, and suggest a few possible future directions.</p>
</sec>
<sec id="j_nejsds100_s_002">
<label>2</label>
<title>Inverse Probability Weight Estimators</title>
<sec id="j_nejsds100_s_003">
<title>Horvitz–Thompson and Hájek Estimators</title>
<p>We begin by defining two widely used estimators in survey sampling: the Horvitz–Thompson (HT) estimator (<xref ref-type="bibr" rid="j_nejsds100_ref_018">Horvitz and Thompson</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_018">1952</xref>)<xref ref-type="fn" rid="j_nejsds100_fn_001">1</xref><fn id="j_nejsds100_fn_001"><label><sup>1</sup></label>
<p>Also called Narain-Horvitz-Thompson estimator after <xref ref-type="bibr" rid="j_nejsds100_ref_030">Narain</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_030">1951</xref>) by <xref ref-type="bibr" rid="j_nejsds100_ref_040">Rao et al.</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_040">1999</xref>); <xref ref-type="bibr" rid="j_nejsds100_ref_003">Chauvet</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_003">2014</xref>).</p></fn> and the Hájek estimator (<xref ref-type="bibr" rid="j_nejsds100_ref_019">Hájek</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_019">1971</xref>) briefly. Consider a finite population <inline-formula id="j_nejsds100_ineq_014"><alternatives><mml:math>
<mml:mi mathvariant="italic">U</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">N</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$U=\{1,2,\dots ,N\}$]]></tex-math></alternatives></inline-formula> with values <inline-formula id="j_nejsds100_ineq_015"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${Y_{k}}$]]></tex-math></alternatives></inline-formula> for each unit <inline-formula id="j_nejsds100_ineq_016"><alternatives><mml:math>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">U</mml:mi></mml:math><tex-math><![CDATA[$k\in U$]]></tex-math></alternatives></inline-formula>. A sample <inline-formula id="j_nejsds100_ineq_017"><alternatives><mml:math>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo stretchy="false">⊂</mml:mo>
<mml:mi mathvariant="italic">U</mml:mi></mml:math><tex-math><![CDATA[$s\subset U$]]></tex-math></alternatives></inline-formula> is drawn according to a sampling design with known inclusion probabilities <inline-formula id="j_nejsds100_ineq_018"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${p_{k}}=P(k\in s)$]]></tex-math></alternatives></inline-formula>.<xref ref-type="fn" rid="j_nejsds100_fn_002">2</xref><fn id="j_nejsds100_fn_002"><label><sup>2</sup></label>
<p>In a probability proportional to size (PPS) design, where the inclusion probabilities are proportional to an auxiliary variable <inline-formula id="j_nejsds100_ineq_019"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${x_{k}}$]]></tex-math></alternatives></inline-formula>, they take the form 
<disp-formula id="j_nejsds100_eq_002">
<label>(2.1)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">N</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {p_{k}}=\frac{n{x_{k}}}{{\textstyle\textstyle\sum _{i=1}^{N}}{x_{i}}}.\]]]></tex-math></alternatives>
</disp-formula>
</p></fn> The Horvitz–Thompson estimator for the population mean <inline-formula id="j_nejsds100_ineq_020"><alternatives><mml:math>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msub>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">U</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$\psi ={N^{-1}}{\textstyle\sum _{k\in U}}{Y_{k}}$]]></tex-math></alternatives></inline-formula> is then 
<disp-formula id="j_nejsds100_eq_003">
<label>(2.2)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">H</mml:mi>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">N</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
</mml:munder><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>.</mml:mo>
<mml:mspace width="1em"/>
<mml:mtext>(Horvitz–Thompson)</mml:mtext>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\hat{\psi }_{HT}}=\frac{1}{N}\sum \limits_{k\in s}\frac{{Y_{k}}}{{p_{k}}}.\hspace{1em}\text{(Horvitz--Thompson)}\]]]></tex-math></alternatives>
</disp-formula> 
The Hájek estimator is an alternative procedure that estimates the population sum as well as <italic>N</italic>, i.e., it normalizes the estimated sum by the estimated total: 
<disp-formula id="j_nejsds100_eq_004">
<label>(2.3)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mtext>Hájek</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>.</mml:mo>
<mml:mspace width="1em"/>
<mml:mtext>(Hájek)</mml:mtext>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\hat{\psi }_{\text{Hájek}}}=\frac{{\textstyle\sum _{k\in s}}{Y_{k}}/{p_{k}}}{{\textstyle\sum _{k\in s}}1/{p_{k}}}.\hspace{1em}\text{(Hájek)}\]]]></tex-math></alternatives>
</disp-formula> 
The HT estimator (<xref rid="j_nejsds100_eq_003">2.2</xref>) is an unbiased estimator of <italic>ψ</italic>, while the Hájek estimator (<xref rid="j_nejsds100_eq_004">2.3</xref>) is non-linear, and approximately unbiased estimator that does not require the knowledge of population size <italic>N</italic>, irrespective of whether <italic>N</italic> is known.</p>
</sec>
<sec id="j_nejsds100_s_004">
<title>Missing Data Framework and Adaptive Normalization</title>
<p>An equivalent framework arises in the context of missing data, <italic>i.e.,</italic> when subjects are observed with nonuniform probabilities. We follow the notations of <xref ref-type="bibr" rid="j_nejsds100_ref_023">Khan and Ugander</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_023">2023</xref>) to describe it here. Here our goal is to estimate the mean <italic>ψ</italic> of <inline-formula id="j_nejsds100_ineq_021"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${Y_{1}},\dots ,{Y_{n}}$]]></tex-math></alternatives></inline-formula>, with the observations missing at random. The Bernoulli indicators <inline-formula id="j_nejsds100_ineq_022"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi></mml:math><tex-math><![CDATA[${R_{k}},k=1,\dots ,n$]]></tex-math></alternatives></inline-formula> indicate whether <inline-formula id="j_nejsds100_ineq_023"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi></mml:math><tex-math><![CDATA[${Y_{k}},k=1,\dots ,n$]]></tex-math></alternatives></inline-formula> were observed or not. We assume <inline-formula id="j_nejsds100_ineq_024"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub><mml:mover>
<mml:mrow>
<mml:mo stretchy="false">∼</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">ind</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mover>
<mml:mtext>Bernoulli</mml:mtext>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${R_{k}}\stackrel{\mathrm{ind}}{\sim }\text{Bernoulli}({p_{k}})$]]></tex-math></alternatives></inline-formula> for <inline-formula id="j_nejsds100_ineq_025"><alternatives><mml:math>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi></mml:math><tex-math><![CDATA[$k=1,\dots ,n$]]></tex-math></alternatives></inline-formula>, which reflects non-response bias in sample surveys. Then, the Horvitz–Thompson and Hájek estimator of <inline-formula id="j_nejsds100_ineq_026"><alternatives><mml:math>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$\psi ={n^{-1}}{\textstyle\sum _{k=1}^{n}}{Y_{k}}$]]></tex-math></alternatives></inline-formula> are: <disp-formula-group id="j_nejsds100_dg_001">
<disp-formula id="j_nejsds100_eq_005">
<label>(2.4)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
<mml:mo>=</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:munderover><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="1em"/><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
<mml:mo>=</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:munderover><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}\hat{S}={\sum \limits_{k=1}^{n}}\frac{{Y_{k}}{R_{k}}}{{p_{k}}}& ,\hspace{1em}\hat{n}={\sum \limits_{k=1}^{n}}\frac{{R_{k}}}{{p_{k}}}\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
<disp-formula id="j_nejsds100_eq_006">
<label>(2.5)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">H</mml:mi>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mspace width="1em"/>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mtext>Hájek</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}{\hat{\psi }_{HT}}=\frac{\hat{S}}{n},& \hspace{1em}{\hat{\psi }_{\text{Hájek}}}=\frac{\hat{S}}{\hat{n}}.\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
</disp-formula-group></p>
<p>A notable generalization of the aforementioned estimators is the Trotter–Tukey estimator (<xref ref-type="bibr" rid="j_nejsds100_ref_051">Trotter and Tukey</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_051">1956</xref>). Recently, <xref ref-type="bibr" rid="j_nejsds100_ref_023">Khan and Ugander</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_023">2023</xref>) rediscovered the Trotter–Tukey estimator as the adaptive normalization (AN) idea by letting the data choose the tuning parameter <italic>λ</italic>. <disp-formula-group id="j_nejsds100_dg_002">
<disp-formula id="j_nejsds100_eq_007">
<label>(2.6)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">λ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">λ</mml:mi><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="0.2778em"/>
<mml:mi mathvariant="italic">λ</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="double-struck">R</mml:mi>
<mml:mspace width="1em"/>
<mml:mtext>(Trotter–Tukey)</mml:mtext>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}{\hat{\psi }_{TT}}& =\frac{\hat{S}}{(1-\lambda )n+\lambda \hat{n}},\hspace{0.2778em}\lambda \in \mathbb{R}\hspace{1em}\text{(Trotter--Tukey)}\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
<disp-formula id="j_nejsds100_eq_008">
<label>(2.7)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mi mathvariant="italic">N</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>+</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>.</mml:mo>
<mml:mspace width="0.2778em"/>
<mml:mspace width="1em"/>
<mml:mtext>(Adaptive Normalization)</mml:mtext>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}{\hat{\psi }_{AN}}& =\frac{\hat{S}}{(1-\hat{\lambda })n+\hat{\lambda }\hat{n}}.\hspace{0.2778em}\hspace{1em}\text{(Adaptive Normalization)}\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
</disp-formula-group> Furthermore, <xref ref-type="bibr" rid="j_nejsds100_ref_023">Khan and Ugander</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_023">2023</xref>) show that the ‘adaptive normalization’ idea dating back to <xref ref-type="bibr" rid="j_nejsds100_ref_051">Trotter and Tukey</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_051">1956</xref>) leads to a lower asymptotic variance than both while generalizing these estimators, in particular, <inline-formula id="j_nejsds100_ineq_027"><alternatives><mml:math>
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mi mathvariant="italic">N</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\mathbb{V}({\hat{\psi }_{AN}})$]]></tex-math></alternatives></inline-formula> is lower than both <inline-formula id="j_nejsds100_ineq_028"><alternatives><mml:math>
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">H</mml:mi>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\mathbb{V}({\hat{\psi }_{HT}})$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds100_ineq_029"><alternatives><mml:math>
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mtext>Hájek</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\mathbb{V}({\hat{\psi }_{\text{Hájek}}})$]]></tex-math></alternatives></inline-formula>. <xref ref-type="bibr" rid="j_nejsds100_ref_023">Khan and Ugander</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_023">2023</xref>) also provide empirical evidence that their adaptive normalization scheme leads to a lower mean squared error of IPW estimators in different application areas in average treatment effect estimation and policy learning.</p>
</sec>
<sec id="j_nejsds100_s_005">
<label>2.1</label>
<title>Horvitz-Thompson vs. Hájek Estimator</title>
<p>Horvitz–Thompson estimators possess many desirable properties: they are unbiased, admissible and consistent. <xref ref-type="bibr" rid="j_nejsds100_ref_039">Ramakrishnan</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_039">1973</xref>) provides a simple proof that the HT estimator is admissible in a class of all unbiased estimators of a finite population total. <xref ref-type="bibr" rid="j_nejsds100_ref_021">Isaki and Fuller</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_021">1982</xref>) proves that they achieve consistency under suitable conditions, such as, specifically requiring boundedness of population and weight sequences, bounded product of inclusion probabilities and second moments, and a variance that is <inline-formula id="j_nejsds100_ineq_030"><alternatives><mml:math>
<mml:mi mathvariant="italic">O</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$O({n_{t}^{-1}})$]]></tex-math></alternatives></inline-formula>. <xref ref-type="bibr" rid="j_nejsds100_ref_008">Delevoye and Sävje</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_008">2020</xref>) provides conditions such as boundedness of moments for the outcome and inclusion probabilities and weak-design dependence. <xref ref-type="bibr" rid="j_nejsds100_ref_008">Delevoye and Sävje</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_008">2020</xref>) also point out that these conditions might be violated in ill-behaved setting with heavy-tailed outcomes or skewed sampling designs, common in natural experiments, where the practitioner has little control over the design.</p>
<p>It is worth noting here that IPW estimators, in particular the HT estimator, can be looked at as a weighted estimator, where the weights are not model-based, but design-based, as pointed out by several authors from <xref ref-type="bibr" rid="j_nejsds100_ref_031">Neyman</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_031">1934</xref>) to <xref ref-type="bibr" rid="j_nejsds100_ref_028">Little</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_028">2008</xref>). For example, in the HT estimator (<xref rid="j_nejsds100_eq_003">2.2</xref>), the <italic>k</italic>th unit is assigned a weight proportional to the inverse of the selection probability as it ‘represents the <inline-formula id="j_nejsds100_ineq_031"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$1/{p_{k}}$]]></tex-math></alternatives></inline-formula> units of the population’.</p>
<sec id="j_nejsds100_s_006">
<label>2.1.1</label>
<title>When Is the H’ajek Estimator Preferable?</title>
<p>We follow <xref ref-type="bibr" rid="j_nejsds100_ref_044">Särndal et al.</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_044">2003</xref>, ch. 5) for this discussion. First, note that the two estimators, HT and Hájek, can produce identical results under certain designs. However, if the population total <italic>N</italic> is unknown, only the Hájek estimator (<xref rid="j_nejsds100_eq_004">2.3</xref>) can be used. On the other hand, if <italic>N</italic> is known—as is typically the case in the finite population estimation problem—the Hájek estimator “<italic>is usually the better estimator, despite estimation of an a priori known quantity</italic>” (<xref ref-type="bibr" rid="j_nejsds100_ref_044">Särndal et al.</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_044">2003</xref>). <xref ref-type="bibr" rid="j_nejsds100_ref_044">Särndal et al.</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_044">2003</xref>) supports this by identifying three situations where the weighted mean estimator (Hájek) outperforms the simple unbiased estimator (HT). In particular, <xref ref-type="bibr" rid="j_nejsds100_ref_044">Särndal et al.</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_044">2003</xref>, p. 183) shows that the <inline-formula id="j_nejsds100_ineq_032"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mtext>Hájek</mml:mtext>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\hat{\psi }_{\text{Hájek}}}$]]></tex-math></alternatives></inline-formula> estimator achieves lower variance than <inline-formula id="j_nejsds100_ineq_033"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mtext>HT</mml:mtext>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\hat{\psi }_{\text{HT}}}$]]></tex-math></alternatives></inline-formula> in each of these cases:</p>
<list>
<list-item id="j_nejsds100_li_001">
<label>(i)</label>
<p>When the values of <inline-formula id="j_nejsds100_ineq_034"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${y_{k}}$]]></tex-math></alternatives></inline-formula> are closer to the population mean <italic>ψ</italic>, a special case being fixed <inline-formula id="j_nejsds100_ineq_035"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi></mml:math><tex-math><![CDATA[${y_{k}}=c$]]></tex-math></alternatives></inline-formula> for <inline-formula id="j_nejsds100_ineq_036"><alternatives><mml:math>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">N</mml:mi></mml:math><tex-math><![CDATA[$k=1,\dots ,N$]]></tex-math></alternatives></inline-formula>.</p>
</list-item>
<list-item id="j_nejsds100_li_002">
<label>(ii)</label>
<p>When the sample size <inline-formula id="j_nejsds100_ineq_037"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${n_{s}}$]]></tex-math></alternatives></inline-formula> is variable with equal inclusion probabilities, e.g. the case where <inline-formula id="j_nejsds100_ineq_038"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi></mml:math><tex-math><![CDATA[${y_{k}}=c$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds100_ineq_039"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{k}}={p_{0}}$]]></tex-math></alternatives></inline-formula> for <inline-formula id="j_nejsds100_ineq_040"><alternatives><mml:math>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">N</mml:mi></mml:math><tex-math><![CDATA[$k=1,\dots ,N$]]></tex-math></alternatives></inline-formula>, but the sample sizes vary.</p>
</list-item>
<list-item id="j_nejsds100_li_003">
<label>(iii)</label>
<p>Finally, when the inclusion probabilities <inline-formula id="j_nejsds100_ineq_041"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{k}}$]]></tex-math></alternatives></inline-formula> are negatively correlated with the <inline-formula id="j_nejsds100_ineq_042"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${y_{k}}$]]></tex-math></alternatives></inline-formula> values, with large <inline-formula id="j_nejsds100_ineq_043"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${y_{k}}$]]></tex-math></alternatives></inline-formula> values corresponding to small <inline-formula id="j_nejsds100_ineq_044"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{k}}$]]></tex-math></alternatives></inline-formula> values and vice versa. It is easy to see that the Hájek estimator is adaptable to high fluctuations in <inline-formula id="j_nejsds100_ineq_045"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${y_{k}}/{p_{k}}$]]></tex-math></alternatives></inline-formula> values that HT will suffer from, leading to high variability. <xref ref-type="bibr" rid="j_nejsds100_ref_001">Basu</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_001">1988</xref>) pointed this out in his famous circus example, recounted in ğ<xref rid="j_nejsds100_s_011">3.1</xref>.</p>
</list-item>
</list>
<p>Despite being popular choices, both the Horvitz–Thompson (HT) (<xref ref-type="bibr" rid="j_nejsds100_ref_018">Horvitz and Thompson</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_018">1952</xref>), and the Hájek estimator (<xref ref-type="bibr" rid="j_nejsds100_ref_019">Hájek</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_019">1971</xref>) have generated debates and controversies in the recent past. For example, the Horvitz–Thompson estimator where we encounter randomly missing observations and a very high-dimensional parameter space is presented as evidence of weakness of Bayesian method in this problem, in an example due to <xref ref-type="bibr" rid="j_nejsds100_ref_054">Wasserman</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_054">2004</xref>), based on <xref ref-type="bibr" rid="j_nejsds100_ref_042">Robins and Ritov</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_042">1997</xref>). In response, a few authors (<xref ref-type="bibr" rid="j_nejsds100_ref_046">Sims</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_046">2010</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_012">Ghosh</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_012">2015</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_026">Li</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_026">2010</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_014">Harmeling and Touissant</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_014">2007</xref>) have tried to provide a Bayesian answer by constructing Bayesian estimators that achieve a lower variance than the HT estimator at the cost of admitting a small (and in some cases, vanishing) bias: another example of the bias-variance trade-off.</p>
<p><bold>Stein’s Paradox:</bold> Here, we argue that this phenomenon is another example of Stein’s paradox (<xref ref-type="bibr" rid="j_nejsds100_ref_022">James and Stein</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_022">1961</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_010">Efron and Morris</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_010">1977</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_050">Stigler</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_050">1990</xref>) observed in a different context. The original Stein’s paradox showed the inadmissibility of the ordinary maximum likelihood estimator which is both unbiased and minimax for normal means in dimensions more than 3 by shrinking each component towards the origin (or a pre-fixed value), thereby borrowing strength from each other. The Stein’s paradox and the James–Stein shrinkage estimator has been truly transformative in statistics in both establishing Empirical Bayes’ success in high-dimensional inference and inspiring both frequentist shrinkage estimators and Bayesian shrinkage priors in such problems. We argue that the simple HT estimator can be improved upon by borrowing strength by moving from an independence framework to an exchangeable one, exhibiting Stein’s shrinkage phenomenon and effects of regularization.</p>
<p>Before we discuss the weak paradoxes, we briefly discuss how the connection between IPW estimators and Monte Carlo sampling, in particular importance sampling, and various improvements such as Riemann sums (<xref ref-type="bibr" rid="j_nejsds100_ref_033">Philippe</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_033">1997</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_034">Philippe and Robert</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_034">2001</xref>) or vertical likelihood integration that applies ‘binning and smoothing’ using a score-function heurism to choose the weight function (<xref ref-type="bibr" rid="j_nejsds100_ref_037">Polson and Scott</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_037">2014</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_029">Madrid-Padilla et al.</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_029">2018</xref>). We then show how the idea of binning and smoothing also improves the HT estimator (<xref ref-type="bibr" rid="j_nejsds100_ref_012">Ghosh</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_012">2015</xref>) in the apparent weakness example due to <xref ref-type="bibr" rid="j_nejsds100_ref_054">Wasserman</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_054">2004</xref>). We discuss these connections briefly in the next subsection and point out the similarities between estimators employed in survey sampling and Monte Carlo integration and their connections with statistical mechanics.</p>
</sec>
</sec>
<sec id="j_nejsds100_s_007">
<label>2.2</label>
<title>Importance Sampling</title>
<p>One way to look at this connection between sampling strategies and integral approximation is to represent the former as a missing data problem. Suppose that our goal is to estimate: 
<disp-formula id="j_nejsds100_eq_009">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo>=</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∫</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">d</mml:mi>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \psi ={\int _{0}^{1}}y(x)dx,\]]]></tex-math></alternatives>
</disp-formula> 
which can be thought of as a limiting value of <inline-formula id="j_nejsds100_ineq_046"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\psi _{n}}={n^{-1}}{\textstyle\sum _{i=1}^{n}}{y_{i}}$]]></tex-math></alternatives></inline-formula> as <inline-formula id="j_nejsds100_ineq_047"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi></mml:math><tex-math><![CDATA[$n\to \infty $]]></tex-math></alternatives></inline-formula>, and can be approximated up to any degree of precision by <inline-formula id="j_nejsds100_ineq_048"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">N</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">N</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\psi _{N}}={N^{-1}}{\textstyle\sum _{i=1}^{N}}{y_{i}}$]]></tex-math></alternatives></inline-formula> for a sufficiently large <italic>N</italic>. Given a random sample of size <inline-formula id="j_nejsds100_ineq_049"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">≪</mml:mo>
<mml:mi mathvariant="italic">N</mml:mi></mml:math><tex-math><![CDATA[$n\ll N$]]></tex-math></alternatives></inline-formula>, with sample probabilities <italic>π</italic> attached to <inline-formula id="j_nejsds100_ineq_050"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${y_{i}}$]]></tex-math></alternatives></inline-formula>, we can view estimating <italic>θ</italic> as a problem of estimating the population quantity <inline-formula id="j_nejsds100_ineq_051"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">N</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\psi _{N}}$]]></tex-math></alternatives></inline-formula> by <inline-formula id="j_nejsds100_ineq_052"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\psi _{n}}$]]></tex-math></alternatives></inline-formula>. The usual importance sampling estimator in this case is akin to using <inline-formula id="j_nejsds100_ineq_053"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">π</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">π</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">]</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">ψ</mml:mi></mml:math><tex-math><![CDATA[${\mathbb{E}_{\pi }}[{\textstyle\sum _{i=1}^{n}}{y_{i}}/{\pi _{i}}]=\psi $]]></tex-math></alternatives></inline-formula>, the usual HT estimator. Just like the HT estimator, the variance of importance sampling could blow up for poor choices of <inline-formula id="j_nejsds100_ineq_054"><alternatives><mml:math>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mo>·</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$g(\cdot )$]]></tex-math></alternatives></inline-formula>, while the bias may shrink to zero. As we show below, these connections have been exploited by several authors to propose alternative importance sampling strategies.</p>
<p>Given a sample <inline-formula id="j_nejsds100_ineq_055"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({y_{1}},\dots ,{y_{n}})$]]></tex-math></alternatives></inline-formula>, from either the density <italic>f</italic> corresponding to <italic>F</italic> itself or a suitable proposal density <italic>g</italic>, the usual importance sampling estimates <italic>ψ</italic> by the empirical average: 
<disp-formula id="j_nejsds100_eq_010">
<label>(2.8)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">I</mml:mi>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:munderover><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\hat{\psi }_{IS}}=\frac{1}{n}{\sum \limits_{i=1}^{n}}\frac{l({y_{i}})f({y_{i}})}{g({y_{i}})}.\]]]></tex-math></alternatives>
</disp-formula> 
If the underlying probability measure is easy to sample from, <italic>i.e.,</italic> if we can afford <inline-formula id="j_nejsds100_ineq_056"><alternatives><mml:math>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">g</mml:mi></mml:math><tex-math><![CDATA[$f=g$]]></tex-math></alternatives></inline-formula>, the above will reduce to the empirical average <inline-formula id="j_nejsds100_ineq_057"><alternatives><mml:math><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\hat{\psi }={n^{-1}}{\textstyle\sum _{i=1}^{n}}l({y_{i}})$]]></tex-math></alternatives></inline-formula>, which by the Law of Large Numbers, converge to the true value of <italic>ψ</italic> at <inline-formula id="j_nejsds100_ineq_058"><alternatives><mml:math>
<mml:mi mathvariant="italic">O</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$O({n^{-1}})$]]></tex-math></alternatives></inline-formula> rate. It is worth pointing out that replacing the empirical average for the naïve importance sampling estimator in (<xref rid="j_nejsds100_eq_010">2.8</xref>) by a Riemann sum estimator provides a remarkable improvement in stability and convergence as demonstrated in (<xref ref-type="bibr" rid="j_nejsds100_ref_033">Philippe</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_033">1997</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_034">Philippe and Robert</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_034">2001</xref>) or (<xref ref-type="bibr" rid="j_nejsds100_ref_055">Yakowitz et al.</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_055">1978</xref>). Recently, <xref ref-type="bibr" rid="j_nejsds100_ref_007">Datta and Polson</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_007">2025</xref>) showed that combining Riemann sum estimators with nested sampling ideas can yield sharper convergence rates (up to <inline-formula id="j_nejsds100_ineq_059"><alternatives><mml:math>
<mml:mi mathvariant="italic">O</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$O({n^{-4}})$]]></tex-math></alternatives></inline-formula>) for marginal likelihood estimation in Bayesian problems. We refer the reader to <xref ref-type="bibr" rid="j_nejsds100_ref_007">Datta and Polson</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_007">2025</xref>) for a more detailed discussion of vertical likelihood, and how it connects importance sampling with nested inference. This framework also offers a unifying perspective on several sampling-based approaches that aim to improve the stability of likelihood-weighted estimators.</p>
<sec id="j_nejsds100_s_008">
<label>2.2.1</label>
<title>A Bayesian Perspective</title>
<p>We conclude the discussion of evidence estimation with a pertinent and profound point raised by <xref ref-type="bibr" rid="j_nejsds100_ref_009">Diaconis</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_009">1988</xref>) in support of Bayesian approaches for probabilistic numerical integration. <xref ref-type="bibr" rid="j_nejsds100_ref_009">Diaconis</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_009">1988</xref>) starts with a algebraic function: 
<disp-formula id="j_nejsds100_eq_011">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mo movablelimits="false">exp</mml:mo>
<mml:mfenced separators="" open="{" close="}">
<mml:mrow>
<mml:mo movablelimits="false">cosh</mml:mo>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>2</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>+</mml:mo>
<mml:mo movablelimits="false">cos</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>+</mml:mo>
<mml:mo movablelimits="false">sin</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="1em"/>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mo>:</mml:mo>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo>
<mml:mo stretchy="false">↦</mml:mo>
<mml:mi mathvariant="double-struck">R</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ f(x)=\exp \left\{\cosh \left(\frac{x+2{x^{2}}+\cos x}{3+\sin {x^{3}}}\right)\right\},\hspace{1em}f:[0,1]\mapsto \mathbb{R},\]]]></tex-math></alternatives>
</disp-formula> 
for which we want <inline-formula id="j_nejsds100_ineq_060"><alternatives><mml:math>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∫</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">d</mml:mi>
<mml:mi mathvariant="italic">x</mml:mi></mml:math><tex-math><![CDATA[${\textstyle\int _{0}^{1}}f(x)dx$]]></tex-math></alternatives></inline-formula> and asks “<italic>What does it mean to ‘know’ a function?.</italic>” In such a situation, where we might know a few properties of <italic>f</italic> and not others, a Bayesian approach seems natural, where one starts with a prior on <inline-formula id="j_nejsds100_ineq_061"><alternatives><mml:math>
<mml:mi mathvariant="script">C</mml:mi>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$\mathcal{C}[0,1]$]]></tex-math></alternatives></inline-formula>, the space of continuous functions, and estimate the integral using Bayes’ rule. <xref ref-type="bibr" rid="j_nejsds100_ref_009">Diaconis</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_009">1988</xref>) shows that Brownian motion, the ‘easiest’ prior on <inline-formula id="j_nejsds100_ineq_062"><alternatives><mml:math>
<mml:mi mathvariant="script">C</mml:mi>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$\mathcal{C}[0,1]$]]></tex-math></alternatives></inline-formula>, yields a linear interpolation leading to the trapezoid rule. <xref ref-type="bibr" rid="j_nejsds100_ref_009">Diaconis</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_009">1988</xref>) then shows that a host of well-known methods can be recovered as Bayesian estimates. This key question is revisited in (<xref ref-type="bibr" rid="j_nejsds100_ref_032">Owen</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_032">2019</xref>) where he asks what does it mean to know the error <inline-formula id="j_nejsds100_ineq_063"><alternatives><mml:math>
<mml:mi mathvariant="normal">Δ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo stretchy="false">|</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo stretchy="false">|</mml:mo></mml:math><tex-math><![CDATA[$\Delta =|\hat{\psi }-\psi |$]]></tex-math></alternatives></inline-formula> and discusses advantages of Bayesian approaches over classical methods. <xref ref-type="bibr" rid="j_nejsds100_ref_032">Owen</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_032">2019</xref>) argues in favour of Bayes methods in difficult problems where extreme cost of function evaluation or skewness or heavy-tailed properties of <inline-formula id="j_nejsds100_ineq_064"><alternatives><mml:math>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$l(x)$]]></tex-math></alternatives></inline-formula> or unavailability of central limit theorem exposes the weakness of classical methods.</p>
</sec>
</sec>
<sec id="j_nejsds100_s_009">
<label>2.3</label>
<title>Average Treatment Effect Estimation</title>
<p>A natural application of IPW estimator is average treatment effect (ATE) estimation in potential outcomes causal inference framework as pointed out in (<xref ref-type="bibr" rid="j_nejsds100_ref_008">Delevoye and Sävje</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_008">2020</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_023">Khan and Ugander</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_023">2023</xref>). We refer the readers to the excellent references in (<xref ref-type="bibr" rid="j_nejsds100_ref_006">Cunningham</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_006">2021</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_043">Rubin</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_043">1974</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_020">Imbens</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_020">2004</xref>) for in-depth coverage and historical backgrounds. Here we measure the difference in potential outcomes observed over time points <inline-formula id="j_nejsds100_ineq_065"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${t_{2}}\gt {t_{1}}$]]></tex-math></alternatives></inline-formula>, where only one of the two potential outcomes are observed for any unit. Using <xref ref-type="bibr" rid="j_nejsds100_ref_023">Khan and Ugander</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_023">2023</xref>)’s notation: we have the triplets <inline-formula id="j_nejsds100_ineq_066"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({Y_{k}}(0),{Y_{k}}(1),{p_{k}})$]]></tex-math></alternatives></inline-formula> for <inline-formula id="j_nejsds100_ineq_067"><alternatives><mml:math>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi></mml:math><tex-math><![CDATA[$k=1,\dots ,n$]]></tex-math></alternatives></inline-formula>, where we observe <inline-formula id="j_nejsds100_ineq_068"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">I</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${Y_{k}}({I_{k}})$]]></tex-math></alternatives></inline-formula> for <inline-formula id="j_nejsds100_ineq_069"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">I</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∼</mml:mo>
<mml:mtext>Bernoulli</mml:mtext>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${I_{k}}\sim \text{Bernoulli}({p_{k}})$]]></tex-math></alternatives></inline-formula>. and the parameter of interest is <inline-formula id="j_nejsds100_ineq_070"><alternatives><mml:math>
<mml:mi mathvariant="italic">A</mml:mi>
<mml:mi mathvariant="italic">T</mml:mi>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">τ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$ATE=\tau =\mathbb{E}[{Y_{k}}(1)-{Y_{k}}(0)]$]]></tex-math></alternatives></inline-formula> from the observations available to us. The IPW estimators are employed here to estimate the two population means <inline-formula id="j_nejsds100_ineq_071"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[${\psi _{1}}=\mathbb{E}[{Y_{k}}(1)]$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds100_ineq_072"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[${\psi _{0}}=\mathbb{E}[{Y_{k}}(0)]$]]></tex-math></alternatives></inline-formula> separately and estimating <italic>τ</italic> by <inline-formula id="j_nejsds100_ineq_073"><alternatives><mml:math><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">τ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
<mml:mo>=</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
<mml:mo>−</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover></mml:math><tex-math><![CDATA[$\hat{\tau }=\hat{{\psi _{1}}}-\hat{{\psi _{0}}}$]]></tex-math></alternatives></inline-formula>. Estimating the population means <inline-formula id="j_nejsds100_ineq_074"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\psi _{j}}$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_nejsds100_ineq_075"><alternatives><mml:math>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$j=1,2$]]></tex-math></alternatives></inline-formula> turns ATE into a survey sampling problem and one can use all the estimators: HT, Hájek or the various improvements such as the adaptive IPW estimator by <xref ref-type="bibr" rid="j_nejsds100_ref_023">Khan and Ugander</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_023">2023</xref>) based on the Trotter–Tukey idea (<xref ref-type="bibr" rid="j_nejsds100_ref_051">Trotter and Tukey</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_051">1956</xref>) here. <xref ref-type="bibr" rid="j_nejsds100_ref_023">Khan and Ugander</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_023">2023</xref>) further develop an adaptive estimator by minimizing the variance of between-group differences and show that both the separate AIPW and the joint AIPW estimators achieve lower mean squared errors compared to the usual HT and Hájek.</p>
</sec>
</sec>
<sec id="j_nejsds100_s_010">
<label>3</label>
<title>Weak Paradoxes in Ratio Estimation Problem</title>
<sec id="j_nejsds100_s_011">
<label>3.1</label>
<title><xref ref-type="bibr" rid="j_nejsds100_ref_001">Basu</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_001">1988</xref>)’s Example</title><statement id="j_nejsds100_stat_001"><label>Example 1.</label>
<p><bold>Basu’s circus example:</bold> The famous circus example, due to (<xref ref-type="bibr" rid="j_nejsds100_ref_001">Basu</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_001">1988</xref>), shows that the HT estimator (<xref rid="j_nejsds100_eq_003">2.2</xref>) could lead to absurd estimates in some situations. Here, we imagine a circus owner, trying to estimate the total weight of his 50 elephants (say <italic>Y</italic>), picks a representative elephant from his herd (<italic>Sambo</italic>) and multiply his weight by 50. But, then a circus statistician, appalled by this estimate, devices a plan where <italic>Sambo</italic> is picked with probability <inline-formula id="j_nejsds100_ineq_076"><alternatives><mml:math>
<mml:mn>99</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>100</mml:mn></mml:math><tex-math><![CDATA[$99/100$]]></tex-math></alternatives></inline-formula> and each of the remaining with <inline-formula id="j_nejsds100_ineq_077"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>100</mml:mn></mml:math><tex-math><![CDATA[$1/100$]]></tex-math></alternatives></inline-formula>. Unfortunately, now the HT estimate is <inline-formula id="j_nejsds100_ineq_078"><alternatives><mml:math>
<mml:mtext>Sambo’s weight</mml:mtext>
<mml:mo>×</mml:mo>
<mml:mn>100</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>99</mml:mn></mml:math><tex-math><![CDATA[$\text{Sambo's weight}\times 100/99$]]></tex-math></alternatives></inline-formula>, a serious underestimate, and moreover, if the owner picks <italic>Jumbo</italic>, the biggest in the herd, the HT estimate is <inline-formula id="j_nejsds100_ineq_079"><alternatives><mml:math>
<mml:mtext>Jumbo’s weight</mml:mtext>
<mml:mo>×</mml:mo>
<mml:mn>4900</mml:mn></mml:math><tex-math><![CDATA[$\text{Jumbo's weight}\times 4900$]]></tex-math></alternatives></inline-formula>, an absurdity.</p></statement>
<p>Discussing <xref ref-type="bibr" rid="j_nejsds100_ref_001">Basu</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_001">1988</xref>), <xref ref-type="bibr" rid="j_nejsds100_ref_019">Hájek</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_019">1971</xref>) points out that the HT estimator’s <italic>“usefulness is increased in connection with ratio estimation”</italic> and proposed an estimator in presence of auxiliary information <inline-formula id="j_nejsds100_ineq_080"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${A_{k}}$]]></tex-math></alternatives></inline-formula>, related to <inline-formula id="j_nejsds100_ineq_081"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${Y_{k}}$]]></tex-math></alternatives></inline-formula> and with known total: 
<disp-formula id="j_nejsds100_eq_012">
<label>(3.1)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mtext>Hájek</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>×</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msubsup><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msubsup><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="0.2778em"/>
<mml:mtext>where</mml:mtext>
<mml:mspace width="2.5pt"/>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∝</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\hat{Y}_{\text{Hájek}}}={\sum \limits_{k=1}^{n}}{A_{k}}\times \frac{{\textstyle\textstyle\sum _{k=1}^{n}}\frac{{Y_{k}}}{{p_{k}}}}{{\textstyle\textstyle\sum _{k=1}^{n}}\frac{{A_{k}}}{{p_{k}}}},\hspace{0.2778em}\text{where}\hspace{2.5pt}{Y_{k}}\propto {A_{k}},k=1,\dots ,n,\]]]></tex-math></alternatives>
</disp-formula> 
which would not be affected in this example like the standard HT estimator. The Hájek IPW estimate for the population mean in (<xref rid="j_nejsds100_eq_004">2.3</xref>) can be derived from the special case <inline-formula id="j_nejsds100_ineq_082"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">≡</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>∀</mml:mo>
<mml:mi mathvariant="italic">k</mml:mi></mml:math><tex-math><![CDATA[${A_{k}}\equiv 1,\forall k$]]></tex-math></alternatives></inline-formula>.</p>
<p>Although the circus example was intended to be a pathological example, it provides at least two useful insights. First, weighted estimators can lead to nonsensical answers despite having nice large sample properties (<xref ref-type="bibr" rid="j_nejsds100_ref_028">Little</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_028">2008</xref>). Second, problems of similar nature occur in importance sampling or Monte Carlo estimation of the marginal likelihood where empirical averages or unbiased estimator could have high or infinite variance (<xref ref-type="bibr" rid="j_nejsds100_ref_026">Li</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_026">2010</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_038">Raftery et al.</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_038">2006</xref>). As pointed out in (<xref ref-type="bibr" rid="j_nejsds100_ref_007">Datta and Polson</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_007">2025</xref>), strategies like Riemann sum (<xref ref-type="bibr" rid="j_nejsds100_ref_033">Philippe</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_033">1997</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_034">Philippe and Robert</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_034">2001</xref>), or careful choice of weight functions like vertical likelihood (<xref ref-type="bibr" rid="j_nejsds100_ref_037">Polson and Scott</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_037">2014</xref>) or nested sampling (<xref ref-type="bibr" rid="j_nejsds100_ref_049">Skilling</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_049">2006</xref>), or using ratio estimators (<xref ref-type="bibr" rid="j_nejsds100_ref_011">Firth</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_011">2011</xref>), or adaptive normalization (<xref ref-type="bibr" rid="j_nejsds100_ref_023">Khan and Ugander</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_023">2023</xref>) can resolve these issues. In particular, the same trick of ordering the draws and applying a Riemann-sum type approach is the key to avoid falling into examples like Basu’s circus.</p>
</sec>
<sec id="j_nejsds100_s_012">
<label>3.2</label>
<title><xref ref-type="bibr" rid="j_nejsds100_ref_054">Wasserman</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_054">2004</xref>)’s Example</title>
<p>We first re-state the example which is itself a simplification of an example from (<xref ref-type="bibr" rid="j_nejsds100_ref_042">Robins and Ritov</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_042">1997</xref>), in a similar spirit. We consider IID samples <inline-formula id="j_nejsds100_ineq_083"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({Y_{i}},{X_{i}},{R_{i}})$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_nejsds100_ineq_084"><alternatives><mml:math>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi></mml:math><tex-math><![CDATA[$i=1,\dots ,B$]]></tex-math></alternatives></inline-formula> such that <inline-formula id="j_nejsds100_ineq_085"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${Y_{i}}$]]></tex-math></alternatives></inline-formula>’s are generated as a mixture of Bernoulli distributions with individual parameters indexed by the component label <inline-formula id="j_nejsds100_ineq_086"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{i}}$]]></tex-math></alternatives></inline-formula>, and the ‘missingness’ indicator <inline-formula id="j_nejsds100_ineq_087"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${R_{i}}$]]></tex-math></alternatives></inline-formula> denoting whether <inline-formula id="j_nejsds100_ineq_088"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${Y_{i}}$]]></tex-math></alternatives></inline-formula> was observed or not. Let the ‘success’ probabilities associated to <inline-formula id="j_nejsds100_ineq_089"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${R_{i}}$]]></tex-math></alternatives></inline-formula> be known constants <inline-formula id="j_nejsds100_ineq_090"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{{X_{i}}}}$]]></tex-math></alternatives></inline-formula> satisfying: 
<disp-formula id="j_nejsds100_eq_013">
<label>(3.2)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mi mathvariant="italic">δ</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">δ</mml:mi>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="0.2778em"/>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ 0\lt \delta \le {p_{j}}\le 1-\delta \lt 1,\hspace{0.2778em}j=1,\dots ,B.\]]]></tex-math></alternatives>
</disp-formula> 
Note that this strong condition on <inline-formula id="j_nejsds100_ineq_091"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{j}}$]]></tex-math></alternatives></inline-formula>’s in (<xref rid="j_nejsds100_eq_013">3.2</xref>) ensures that the HT estimator will not lead to absurd answers like Basu’s paradox stated earlier. The hierarchical model for each draw <inline-formula id="j_nejsds100_ineq_092"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({Y_{i}},{X_{i}},{R_{i}})$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_nejsds100_ineq_093"><alternatives><mml:math>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi></mml:math><tex-math><![CDATA[$i=1,\dots ,n$]]></tex-math></alternatives></inline-formula>, is: <disp-formula-group id="j_nejsds100_dg_003">
<disp-formula id="j_nejsds100_eq_014">
<label>(3.3)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo stretchy="false">∼</mml:mo>
<mml:mi mathvariant="script">U</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}{X_{i}}& \sim \mathcal{U}(1,\dots ,B)\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
<disp-formula id="j_nejsds100_eq_015">
<label>(3.4)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo stretchy="false">∣</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">]</mml:mo>
<mml:mo stretchy="false">∼</mml:mo>
<mml:mtext>Bernoulli</mml:mtext>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}[{R_{i}}& \mid {X_{i}}={x_{i}}]\sim \text{Bernoulli}({p_{{x_{i}}}})\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
<disp-formula id="j_nejsds100_eq_016">
<label>(3.5)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo stretchy="false">∣</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">]</mml:mo>
<mml:mo stretchy="false">∼</mml:mo>
<mml:mfenced separators="" open="{" close="">
<mml:mrow>
<mml:mtable equalrows="false" columnlines="none" equalcolumns="false" columnalign="left">
<mml:mtr>
<mml:mtd class="array">
<mml:mtext>unobserved</mml:mtext>
<mml:mspace width="0.2778em"/>
<mml:mtext>if</mml:mtext>
<mml:mspace width="0.2778em"/>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
<mml:mspace width="1em"/>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="array">
<mml:mtext>Bernoulli</mml:mtext>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mspace width="0.2778em"/>
<mml:mtext>if</mml:mtext>
<mml:mspace width="0.2778em"/>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>.</mml:mo>
<mml:mspace width="1em"/>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}[{Y_{i}}& \mid {R_{i}},{X_{i}}={x_{i}}]\sim \left\{\begin{array}{l}\text{unobserved}\hspace{0.2778em}\text{if}\hspace{0.2778em}{R_{i}}=0\hspace{1em}\\ {} \text{Bernoulli}({\theta _{{x_{i}}}})\hspace{0.2778em}\text{if}\hspace{0.2778em}{R_{i}}=1.\hspace{1em}\end{array}\right.\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
</disp-formula-group> The parameter of interest is the average <inline-formula id="j_nejsds100_ineq_094"><alternatives><mml:math>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$\psi =(1/B){\textstyle\sum _{b=1}^{B}}{\theta _{b}}$]]></tex-math></alternatives></inline-formula>. <xref ref-type="bibr" rid="j_nejsds100_ref_054">Wasserman</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_054">2004</xref>) argues that since the likelihood has little information on most <inline-formula id="j_nejsds100_ineq_095"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{j}}$]]></tex-math></alternatives></inline-formula>’s and the known constants <italic>B</italic> and <inline-formula id="j_nejsds100_ineq_096"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{j}}$]]></tex-math></alternatives></inline-formula>’s drop from the likelihood, Bayes’ estimates for <italic>ψ</italic> are going to be poor. On the other hand, the HT estimate: 
<disp-formula id="j_nejsds100_eq_017">
<label>(3.6)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">H</mml:mi>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:munderover><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\hat{\psi }_{HT}}=\frac{1}{n}{\sum \limits_{i=1}^{n}}\frac{{R_{i}}{Y_{i}}}{{p_{{X_{i}}}}}\]]]></tex-math></alternatives>
</disp-formula> 
is easily seen to be unbiased, given <inline-formula id="j_nejsds100_ineq_097"><alternatives><mml:math>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∣</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>·</mml:mo>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∣</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∣</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">ψ</mml:mi></mml:math><tex-math><![CDATA[$\mathbb{E}({R_{i}}{Y_{i}}/{p_{{X_{i}}}})=\mathbb{E}\{\mathbb{E}({R_{i}}\mid {Y_{i}},{X_{i}})\cdot \mathbb{E}({Y_{i}}\mid {X_{i}})/{p_{{X_{i}}}}\}=\mathbb{E}(\mathbb{E}({Y_{i}}\mid {X_{i}}))=\psi $]]></tex-math></alternatives></inline-formula> since <inline-formula id="j_nejsds100_ineq_098"><alternatives><mml:math>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∣</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$\mathbb{E}({R_{i}}\mid {Y_{i}},{X_{i}})={p_{{X_{i}}}}$]]></tex-math></alternatives></inline-formula> by construction. The HT estimate will also satisfy <inline-formula id="j_nejsds100_ineq_099"><alternatives><mml:math>
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">H</mml:mi>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">δ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[$\mathbb{V}({\hat{\psi }_{HT}})\le 1/n{\delta ^{2}}$]]></tex-math></alternatives></inline-formula>, using Hoeffding’s inequality. We would like to note here that the <italic>assumption</italic> (<xref rid="j_nejsds100_eq_013">3.2</xref>) that the selection probabilities are between <italic>δ</italic> and <inline-formula id="j_nejsds100_ineq_100"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">δ</mml:mi></mml:math><tex-math><![CDATA[$1-\delta $]]></tex-math></alternatives></inline-formula> is exploited explicitly in this asymptotic result, <italic>i.e.,</italic> the HT estimator might have infinite variance as <italic>δ</italic> goes to zero, making the assumptions crucial.</p>
<p>This example of apparent weakness in Bayesian paradigm and the concluding remarks by <xref ref-type="bibr" rid="j_nejsds100_ref_054">Wasserman</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_054">2004</xref>)<xref ref-type="fn" rid="j_nejsds100_fn_003">3</xref><fn id="j_nejsds100_fn_003"><label><sup>3</sup></label>
<p>“<italic>Bayesians are slaves to the likelihood function. When the likelihood goes awry, so will Bayesian inference.</italic>” (<xref ref-type="bibr" rid="j_nejsds100_ref_054">Wasserman</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_054">2004</xref>, pp. 189)</p></fn> has since been a source of debate and elicited response from Bayesian community (<xref ref-type="bibr" rid="j_nejsds100_ref_026">Li</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_026">2010</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_046">Sims</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_046">2010</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_014">Harmeling and Touissant</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_014">2007</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_012">Ghosh</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_012">2015</xref>) which we review briefly below.</p>
</sec>
<sec id="j_nejsds100_s_013">
<label>3.3</label>
<title>Bayesian Solutions for <xref ref-type="bibr" rid="j_nejsds100_ref_054">Wasserman</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_054">2004</xref>)’s Problem</title>
<p>We review the existing Bayesian resolutions for the Robins-Ritov-Wasserman problem using Bayesian ideas and argue that, in some special cases, the improvement in accuracy of Bayes’ estimate is attained via exploiting the ‘borrowing strength’ phenomenon in Stein’s shrinkage.</p>
<sec id="j_nejsds100_s_014">
<label>3.3.1</label>
<title>Full Bayes Solution (<xref ref-type="bibr" rid="j_nejsds100_ref_026">Li</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_026">2010</xref>)</title>
<p><xref ref-type="bibr" rid="j_nejsds100_ref_026">Li</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_026">2010</xref>) provides a simple Bayes estimator by assuming that <inline-formula id="j_nejsds100_ineq_101"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{1}},\dots ,{\theta _{B}}$]]></tex-math></alternatives></inline-formula> are exchangeable, not independent (see Fig. <xref rid="j_nejsds100_fig_001">1</xref>). <xref ref-type="bibr" rid="j_nejsds100_ref_026">Li</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_026">2010</xref>) estimates <italic>ψ</italic> by augmenting (<xref rid="j_nejsds100_eq_016">3.5</xref>) by with Beta priors for <inline-formula id="j_nejsds100_ineq_102"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{1}},\dots ,{\theta _{B}}$]]></tex-math></alternatives></inline-formula> and a further hyperprior on the mean of these Beta priors, as follows: <disp-formula-group id="j_nejsds100_dg_004">
<disp-formula id="j_nejsds100_eq_018">
<label>(3.7)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∣</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mover>
<mml:mrow>
<mml:mo stretchy="false">∼</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">ind</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mover>
<mml:mtext>Beta</mml:mtext>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}{\theta _{b}}\mid {\alpha _{T}},\psi & \stackrel{\mathrm{ind}}{\sim }\text{Beta}({\alpha _{T}}\psi ,{\alpha _{T}}(1-{\psi _{0}}))\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
<disp-formula id="j_nejsds100_eq_019">
<label>(3.8)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∣</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo stretchy="false">∼</mml:mo>
<mml:mtext>Beta</mml:mtext>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>×</mml:mo>
<mml:mi mathvariant="italic">π</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}(\psi ,{\alpha _{T}}\mid {\alpha _{F}})& \sim \text{Beta}({\alpha _{F}},{\alpha _{F}})\times \pi ({\alpha _{T}}),\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
</disp-formula-group> where, <italic>ψ</italic> is the mean of <italic>θ</italic>, and the shape parameters <inline-formula id="j_nejsds100_ineq_103"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\alpha _{T}}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds100_ineq_104"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\alpha _{F}}$]]></tex-math></alternatives></inline-formula> control the width of range of <inline-formula id="j_nejsds100_ineq_105"><alternatives><mml:math>
<mml:mi mathvariant="bold-italic">θ</mml:mi></mml:math><tex-math><![CDATA[$\boldsymbol{\theta }$]]></tex-math></alternatives></inline-formula> and the concentration of <italic>ψ</italic>.</p>
<fig id="j_nejsds100_fig_001">
<label>Figure 1</label>
<caption>
<p>Hierarchical model for the Robins-Ritov-Wasserman problem.</p>
</caption>
<graphic xlink:href="nejsds100_g001.jpg"/>
</fig>
<p>Using this model, <xref ref-type="bibr" rid="j_nejsds100_ref_026">Li</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_026">2010</xref>) derives a posterior mean estimate of <italic>ψ</italic> as: 
<disp-formula id="j_nejsds100_eq_020">
<label>(3.9)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mn>2</mml:mn>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\hat{\psi }_{Li}}=\frac{{\textstyle\textstyle\sum _{i=1}^{n}}{R_{i}}{Y_{i}}+{\alpha _{F}}}{{\textstyle\textstyle\sum _{i=1}^{n}}{R_{i}}+2{\alpha _{F}}}\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>Through extensive numerical simulation, <xref ref-type="bibr" rid="j_nejsds100_ref_026">Li</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_026">2010</xref>) shows that this estimator achieves a smaller mean squared error compared to the HT estimator in <xref ref-type="bibr" rid="j_nejsds100_ref_054">Wasserman</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_054">2004</xref>). <xref ref-type="bibr" rid="j_nejsds100_ref_026">Li</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_026">2010</xref>) also argues that the variance of HT estimator, given by: 
<disp-formula id="j_nejsds100_eq_021">
<label>(3.10)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">H</mml:mi>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mfenced separators="" open="{" close="}">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:munderover><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>−</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \mathbb{V}({\hat{\psi }_{HT}})=\frac{1}{n}\left\{\frac{1}{B}{\sum \limits_{b=1}^{B}}\frac{{\theta _{b}}}{{p_{b}}}-{\psi ^{2}}\right\},\]]]></tex-math></alternatives>
</disp-formula> 
will lead to inflated variance for small <inline-formula id="j_nejsds100_ineq_106"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{b}}$]]></tex-math></alternatives></inline-formula> values, necessitating the bounds <inline-formula id="j_nejsds100_ineq_107"><alternatives><mml:math>
<mml:mi mathvariant="italic">δ</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">δ</mml:mi></mml:math><tex-math><![CDATA[$\delta \le {p_{j}}\le 1-\delta $]]></tex-math></alternatives></inline-formula> in (<xref rid="j_nejsds100_eq_013">3.2</xref>), but the variance for Li’s Bayes estimator remains unaffected and does not need this extra restriction. On the other hand, Li’s estimator (<xref rid="j_nejsds100_eq_020">3.9</xref>) is prone to bias if the missingness mechanism <inline-formula id="j_nejsds100_ineq_108"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{X}}$]]></tex-math></alternatives></inline-formula> and the parameter for observed outcomes <inline-formula id="j_nejsds100_ineq_109"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{X}}$]]></tex-math></alternatives></inline-formula> are correlated, <italic>i.e.,</italic> if the missingness is missing-at-random (MAR) and not missing-completely-at-random (MCAR). We shall formalize this via the derived approximation for the mean squared error of the Bayes estimator in our theoretical properties section. <xref ref-type="bibr" rid="j_nejsds100_ref_027">Linero</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_027">2024</xref>) argues that the ‘seemingly innocuous’ priors like the one used by <xref ref-type="bibr" rid="j_nejsds100_ref_026">Li</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_026">2010</xref>) encode what is effectively a-priori knowledge that the amount of selection bias is minimal.</p>
</sec>
<sec id="j_nejsds100_s_015">
<label>3.3.2</label>
<title>Binning and Smoothing (<xref ref-type="bibr" rid="j_nejsds100_ref_012">Ghosh</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_012">2015</xref>)</title>
<p><xref ref-type="bibr" rid="j_nejsds100_ref_012">Ghosh</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_012">2015</xref>) provides another estimator by reducing the dimension of <inline-formula id="j_nejsds100_ineq_110"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({p_{1}},\dots ,{p_{B}})$]]></tex-math></alternatives></inline-formula> by clubbing them into <italic>k</italic> (<inline-formula id="j_nejsds100_ineq_111"><alternatives><mml:math>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo stretchy="false">≪</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi></mml:math><tex-math><![CDATA[$k\ll B$]]></tex-math></alternatives></inline-formula>) groups by utilizing the boundedness assumption (<xref rid="j_nejsds100_eq_013">3.2</xref>). Based on this idea, we define the general binned-smoothed estimator as follows.</p><statement id="j_nejsds100_stat_002"><label>Definition 2.</label>
<p>Given a fixed <inline-formula id="j_nejsds100_ineq_112"><alternatives><mml:math>
<mml:mi mathvariant="italic">δ</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\delta \in (0,1)$]]></tex-math></alternatives></inline-formula> satisfying (<xref rid="j_nejsds100_eq_013">3.2</xref>), we divide the whole range <inline-formula id="j_nejsds100_ineq_113"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">δ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">δ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">∋</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi></mml:math><tex-math><![CDATA[$(\delta ,1-\delta )\ni {p_{b}},b=1,\dots ,B$]]></tex-math></alternatives></inline-formula> into <italic>k</italic> sub-intervals <inline-formula id="j_nejsds100_ineq_114"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">δ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">δ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">δ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">δ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">δ</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$\{{\delta _{0}}=\delta ,{\delta _{1}},\dots ,{\delta _{k}}=1-\delta \}$]]></tex-math></alternatives></inline-formula> such that <inline-formula id="j_nejsds100_ineq_115"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">δ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">δ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mn>2</mml:mn>
<mml:mi mathvariant="italic">δ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${\delta _{i+1}}-{\delta _{i}}=(1-2\delta )/(k-1)$]]></tex-math></alternatives></inline-formula>. We order the observed <inline-formula id="j_nejsds100_ineq_116"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{{X_{i}}}}$]]></tex-math></alternatives></inline-formula>’s into increasing order and define <inline-formula id="j_nejsds100_ineq_117"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">˜</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\tilde{p}_{j}}$]]></tex-math></alternatives></inline-formula> to be the mean of the <inline-formula id="j_nejsds100_ineq_118"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{{X_{i}}}}$]]></tex-math></alternatives></inline-formula> values in the <italic>j</italic>th partition consisting of <inline-formula id="j_nejsds100_ineq_119"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${n_{j}}$]]></tex-math></alternatives></inline-formula> points, i.e. 
<disp-formula id="j_nejsds100_eq_022">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">˜</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mfenced separators="" open="{" close="">
<mml:mrow>
<mml:mtable columnspacing="10.0pt" equalrows="false" columnlines="none" equalcolumns="false" columnalign="left left">
<mml:mtr>
<mml:mtd class="array">
<mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:msub>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>:</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">δ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">δ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd class="array">
<mml:mtext>if</mml:mtext>
<mml:mspace width="0.2778em"/>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="array">
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">δ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">δ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>2</mml:mn>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd class="array">
<mml:mtext>otherwise</mml:mtext>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\tilde{p}_{j}}=\left\{\begin{array}{l@{\hskip10.0pt}l}\frac{1}{{n_{j}}}{\textstyle\sum _{i:{p_{{X_{i}}}}\in ({\delta _{j}},{\delta _{j+1}})}}{p_{{X_{i}}}},\hspace{1em}& \text{if}\hspace{0.2778em}{n_{j}}\gt 0\\ {} ({\delta _{j}}+{\delta _{j+1}})/2\hspace{1em}& \text{otherwise}\end{array}\right.\]]]></tex-math></alternatives>
</disp-formula> 
Then, the <italic>binned-smoothed</italic> estimator, based on (<xref ref-type="bibr" rid="j_nejsds100_ref_012">Ghosh</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_012">2015</xref>) is: 
<disp-formula id="j_nejsds100_eq_023">
<label>(3.11)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">H</mml:mi>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:munderover><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">˜</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>.</mml:mo>
<mml:mspace width="0.2778em"/>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\hat{\psi }_{BS:HT}}={\sum \limits_{j=1}^{k}}\frac{{n_{j}}}{n}\frac{1}{{n_{j}}}\frac{({\textstyle\textstyle\sum _{i=1}^{{n_{j}}}}{R_{i}}{Y_{i}})}{{\tilde{p}_{j}}}.\hspace{0.2778em}\]]]></tex-math></alternatives>
</disp-formula> 
where <inline-formula id="j_nejsds100_ineq_120"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${n_{j}}$]]></tex-math></alternatives></inline-formula> is the number of <inline-formula id="j_nejsds100_ineq_121"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{{X_{i}}}}$]]></tex-math></alternatives></inline-formula>’s falling into the <italic>j</italic>th class. Note that, if the sample size <inline-formula id="j_nejsds100_ineq_122"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">≫</mml:mo>
<mml:mi mathvariant="italic">k</mml:mi></mml:math><tex-math><![CDATA[$n\gg k$]]></tex-math></alternatives></inline-formula>, the number of partitions, the number of <inline-formula id="j_nejsds100_ineq_123"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{{X_{i}}}}$]]></tex-math></alternatives></inline-formula>’s in any interval, i.e. <inline-formula id="j_nejsds100_ineq_124"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${n_{j}}$]]></tex-math></alternatives></inline-formula> values would typically be non-zero, and in the unlikely case <inline-formula id="j_nejsds100_ineq_125"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[${n_{j}}=0$]]></tex-math></alternatives></inline-formula> for any <inline-formula id="j_nejsds100_ineq_126"><alternatives><mml:math>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">k</mml:mi></mml:math><tex-math><![CDATA[$j=1,\dots ,k$]]></tex-math></alternatives></inline-formula>, the corresponding term will not contribute to the numerator in (<xref rid="j_nejsds100_eq_023">3.11</xref>).</p></statement>
<p><xref ref-type="bibr" rid="j_nejsds100_ref_012">Ghosh</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_012">2015</xref>)’s estimator performs at least as well as the HT estimator in small simulation studies. The idea of ‘binning and smoothing’ can be applied to other inverse probability weighted estimators such as Hájek (<xref rid="j_nejsds100_eq_004">2.3</xref>). The binned-smoothed Hájek estimator would take the form: 
<disp-formula id="j_nejsds100_eq_024">
<label>(3.12)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">H</mml:mi>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mi mathvariant="italic">e</mml:mi>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">˜</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">}</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">˜</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">}</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="0.2778em"/>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\hat{\psi }_{BS:Hajek}}=\frac{{\textstyle\textstyle\sum _{j=1}^{k}}\{({\textstyle\textstyle\sum _{i=1}^{{n_{j}}}}{R_{i}}{Y_{i}})/{\tilde{p}_{j}}\}}{{\textstyle\textstyle\sum _{j=1}^{k}}\{({\textstyle\textstyle\sum _{i=1}^{{n_{j}}}}{R_{i}})/{\tilde{p}_{j}}\}},\hspace{0.2778em}\]]]></tex-math></alternatives>
</disp-formula> 
where <inline-formula id="j_nejsds100_ineq_127"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${n_{j}}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds100_ineq_128"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">˜</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\tilde{p}_{j}}$]]></tex-math></alternatives></inline-formula>’s are as used in (<xref rid="j_nejsds100_eq_023">3.11</xref>) before.</p>
</sec>
<sec id="j_nejsds100_s_016">
<label>3.3.3</label>
<title>Bayesian Sieve (<xref ref-type="bibr" rid="j_nejsds100_ref_048">Sims</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_048">2012</xref>)</title>
<p>We note that the idea of grouping the probabilities attached to the missingness indicators <inline-formula id="j_nejsds100_ineq_129"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${R_{i}}$]]></tex-math></alternatives></inline-formula>’s, <italic>i.e.,</italic> clubbing <inline-formula id="j_nejsds100_ineq_130"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{{X_{i}}}}$]]></tex-math></alternatives></inline-formula>’s into <inline-formula id="j_nejsds100_ineq_131"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{j}}$]]></tex-math></alternatives></inline-formula>’s was also proposed in (<xref ref-type="bibr" rid="j_nejsds100_ref_048">Sims</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_048">2012</xref>). <xref ref-type="bibr" rid="j_nejsds100_ref_048">Sims</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_048">2012</xref>) argues that estimators better than the HT estimator (<xref rid="j_nejsds100_eq_017">3.6</xref>) can be constructed using Bayesian approach if one acknowledges the existence of infinite dimensional unknown parameter in the model, e.g. assuming <inline-formula id="j_nejsds100_ineq_132"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>:</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo stretchy="false">↦</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$p(b):1,\dots ,B\mapsto (0,1)$]]></tex-math></alternatives></inline-formula> is known but not the conditional distribution of <inline-formula id="j_nejsds100_ineq_133"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo stretchy="false">∣</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[\theta \mid p]$]]></tex-math></alternatives></inline-formula>. Sims suggested to break up the range of the known <inline-formula id="j_nejsds100_ineq_134"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mo>·</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$p(\cdot )$]]></tex-math></alternatives></inline-formula> into <italic>k</italic> segments, such that <inline-formula id="j_nejsds100_ineq_135"><alternatives><mml:math>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\theta (b)$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds100_ineq_136"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$p(b)$]]></tex-math></alternatives></inline-formula> are independent in each segment, and estimate the unknown <inline-formula id="j_nejsds100_ineq_137"><alternatives><mml:math>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\theta (b)$]]></tex-math></alternatives></inline-formula> via a step function, with a constant for each segment. To complete Sim’s Bayesian sieve, one needs to estimate the probability distribution on <inline-formula id="j_nejsds100_ineq_138"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{j}}$]]></tex-math></alternatives></inline-formula>’s induced by the partition, which will converge to a Dirichlet distribution for a large sample, making analytical calculation for the posterior mean of <italic>ψ</italic> feasible.</p>
</sec>
<sec id="j_nejsds100_s_017">
<label>3.3.4</label>
<title>Gaussian Likelihood (<xref ref-type="bibr" rid="j_nejsds100_ref_014">Harmeling and Touissant</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_014">2007</xref>)</title>
<p><xref ref-type="bibr" rid="j_nejsds100_ref_014">Harmeling and Touissant</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_014">2007</xref>) consider a slightly modified version of the Wasserman’s model (<xref rid="j_nejsds100_eq_016">3.5</xref>). Instead of a Bernoulli, they consider: 
<disp-formula id="j_nejsds100_eq_025">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∼</mml:mo>
<mml:mi mathvariant="script">U</mml:mi>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="0.2778em"/>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∣</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="normal">Θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∼</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="0.2778em"/>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">≤</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {X_{i}}\sim \mathcal{U}\{1,\dots ,B\},\hspace{0.2778em}{Y_{i}}\mid {\Theta _{{X_{i}}}}\sim \mathcal{N}({\theta _{{X_{i}}}},1),\hspace{0.2778em}1\le i\le n,\]]]></tex-math></alternatives>
</disp-formula> 
and <inline-formula id="j_nejsds100_ineq_139"><alternatives><mml:math>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">μ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\mathcal{N}(\mu ,1)$]]></tex-math></alternatives></inline-formula> prior for each <inline-formula id="j_nejsds100_ineq_140"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{b}}$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_nejsds100_ineq_141"><alternatives><mml:math>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi></mml:math><tex-math><![CDATA[$b=1,\dots ,B$]]></tex-math></alternatives></inline-formula>, and a <inline-formula id="j_nejsds100_ineq_142"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">σ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${\mathcal{N}_{\mu }}(0,\sigma )$]]></tex-math></alternatives></inline-formula> hyper-prior on <italic>μ</italic>. The maximum likelihood estimate and the posterior mean Bayes’ estimates for <inline-formula id="j_nejsds100_ineq_143"><alternatives><mml:math>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi></mml:math><tex-math><![CDATA[$B\to \infty $]]></tex-math></alternatives></inline-formula> are given by 
<disp-formula id="j_nejsds100_eq_026">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mtext>MLE</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" maxsize="2.03em" minsize="2.03em">(</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" maxsize="2.03em" minsize="2.03em">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" maxsize="2.03em" minsize="2.03em">(</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" maxsize="2.03em" minsize="2.03em">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="1em"/>
<mml:mtext>and</mml:mtext>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mtext>Bayes</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" maxsize="2.03em" minsize="2.03em">(</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">σ</mml:mi>
<mml:mo>+</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" maxsize="2.03em" minsize="2.03em">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" maxsize="2.03em" minsize="2.03em">(</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" maxsize="2.03em" minsize="2.03em">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}{\hat{\psi }_{\text{MLE}}}& ={\bigg({\sum \limits_{i=1}^{n}}{R_{i}}\bigg)^{-1}}\bigg({\sum \limits_{i=1}^{n}}{R_{i}}{Y_{i}}\bigg),\hspace{1em}\text{and}\\ {} {\hat{\psi }_{\text{Bayes}}}& ={\bigg(2/\sigma +{\sum \limits_{i=1}^{n}}{R_{i}}\bigg)^{-1}}\bigg({\sum \limits_{i=1}^{n}}{R_{i}}{Y_{i}}\bigg),\end{aligned}\]]]></tex-math></alternatives>
</disp-formula> 
respectively. Although these estimators are not directly comparable to (<xref rid="j_nejsds100_eq_020">3.9</xref>), <xref ref-type="bibr" rid="j_nejsds100_ref_014">Harmeling and Touissant</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_014">2007</xref>)’s likelihood-based estimators also achieve a lower MSE compared to the HT estimator, as expected.</p>
<p>We also note that the idea of binning and smoothing <inline-formula id="j_nejsds100_ineq_144"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{j}}$]]></tex-math></alternatives></inline-formula>’s is also connected to importance sampling ideas in subsection <xref rid="j_nejsds100_s_007">2.2</xref>, in particular, the trapezoidal rule (also called weighted Monte Carlo) by <xref ref-type="bibr" rid="j_nejsds100_ref_055">Yakowitz et al.</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_055">1978</xref>), or its generalizations, called the Riemann summation approach <xref ref-type="bibr" rid="j_nejsds100_ref_033">Philippe</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_033">1997</xref>); <xref ref-type="bibr" rid="j_nejsds100_ref_034">Philippe and Robert</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_034">2001</xref>). These trapezoidal rules based on ordered samples reduces the variance drastically and achieves faster convergence and better stability. Along these lines, a later development, called the nested sampling approach by <xref ref-type="bibr" rid="j_nejsds100_ref_049">Skilling</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_049">2006</xref>) also relies on “dividing the unit prior mass into tiny elements, and sorting them by likelihood.” For a comprehensive discussion of the different strategies for variance reduction for the evidence estimation problem, we refer the readers to (<xref ref-type="bibr" rid="j_nejsds100_ref_037">Polson and Scott</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_037">2014</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_007">Datta and Polson</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_007">2025</xref>).</p>
</sec>
</sec>
<sec id="j_nejsds100_s_018">
<label>3.4</label>
<title>Properties of Li’s Bayesian Estimator</title>
<p>While a closed form analytical expression for the variance of the Li’s estimator (<xref rid="j_nejsds100_eq_020">3.9</xref>) is difficult, we can use the linearization strategy aka the delta theorem to derive an approximation for the mean and variance of the Bayes estimator (<xref rid="j_nejsds100_eq_020">3.9</xref>). Doing so, we provide a proof and a closer look at an assertion in <xref ref-type="bibr" rid="j_nejsds100_ref_041">Ritov et al.</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_041">2014</xref>) that Li’s estimator is consistent only if <inline-formula id="j_nejsds100_ineq_145"><alternatives><mml:math>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo stretchy="false">∣</mml:mo>
<mml:mi mathvariant="italic">R</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\mathbb{E}(Y\mid R=1)=\mathbb{E}(Y)$]]></tex-math></alternatives></inline-formula>, and illustrate this phenomenon via simulation studies.</p>
<p>Recall that for a pair of random variables <italic>X</italic>, <italic>Y</italic> with mean <inline-formula id="j_nejsds100_ineq_146"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mu _{X}}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds100_ineq_147"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mu _{Y}}$]]></tex-math></alternatives></inline-formula> and variances <inline-formula id="j_nejsds100_ineq_148"><alternatives><mml:math>
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\mathbb{V}(X)$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_nejsds100_ineq_149"><alternatives><mml:math>
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\mathbb{V}(Y)$]]></tex-math></alternatives></inline-formula> and covariance <inline-formula id="j_nejsds100_ineq_150"><alternatives><mml:math>
<mml:mi mathvariant="normal">Cov</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\mathrm{Cov}(X,Y)$]]></tex-math></alternatives></inline-formula>, we have the following approximations: <disp-formula-group id="j_nejsds100_dg_005">
<disp-formula id="j_nejsds100_eq_027">
<label>(3.13)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo stretchy="false">≈</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}\mathbb{E}(X/Y)& \approx {\mu _{X}}/{\mu _{Y}}\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
<disp-formula id="j_nejsds100_eq_028">
<label>(3.14)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo stretchy="false">≈</mml:mo>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>+</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>−</mml:mo>
<mml:mn>2</mml:mn><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">Cov</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}\mathbb{V}(X/Y)& \approx \left(\frac{{\mu _{X}}}{{\mu _{Y}}}\right)\left(\frac{\mathbb{V}(X)}{{\mu _{X}^{2}}}+\frac{\mathbb{V}(Y)}{{\mu _{Y}^{2}}}-2\frac{\mathrm{Cov}(X,Y)}{{\mu _{X}}{\mu _{Y}}}\right)\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
</disp-formula-group></p>
<p>To calculate the approximate mean and variance for the Li’s posterior mean estimator (<xref rid="j_nejsds100_eq_020">3.9</xref>), we first calculate the mean and variances for the numerator and denominator separately as follows. Recall once again that the posterior mean estimator is given by <inline-formula id="j_nejsds100_ineq_151"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${\hat{\psi }_{Li}}=({\textstyle\sum _{i}}{R_{i}}{Y_{i}}+1)/({\textstyle\sum _{i}}{R_{i}}+2)$]]></tex-math></alternatives></inline-formula> assuming <inline-formula id="j_nejsds100_ineq_152"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[${\alpha _{F}}=1$]]></tex-math></alternatives></inline-formula>, i.e., a <inline-formula id="j_nejsds100_ineq_153"><alternatives><mml:math>
<mml:mtext>Beta</mml:mtext>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\text{Beta}(1,1)$]]></tex-math></alternatives></inline-formula> or Uniform prior on <italic>ψ</italic> but the exact choice of shape parameter <inline-formula id="j_nejsds100_ineq_154"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\alpha _{F}}$]]></tex-math></alternatives></inline-formula> does not influence the asymptotic nature of the variance. For notational convenience, we define the following quantities: 
<disp-formula id="j_nejsds100_eq_029">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="left">
<mml:mtr>
<mml:mtd>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>≐</mml:mo>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="1em"/><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="1em"/><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>−</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>=</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>−</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{array}{l}\displaystyle \bar{\theta }=\frac{1}{B}{\sum \limits_{b=1}^{B}}{\theta _{b}}\doteq \psi ,\hspace{1em}\bar{p}=\frac{1}{B}{\sum \limits_{b=1}^{B}}{p_{b}},\hspace{1em}\bar{\theta .p}=\frac{1}{B}{\sum \limits_{b=1}^{B}}{\theta _{b}}{p_{b}},\\ {} \displaystyle {\sigma _{\theta ,p}}=\bar{\theta .p}-\bar{p}\bar{\theta }=\bar{\theta .p}-\bar{p}\psi .\end{array}\]]]></tex-math></alternatives>
</disp-formula> 
First, the expectation for the numerator and denominator follows from applying iterated expectations as: <disp-formula-group id="j_nejsds100_dg_006">
<disp-formula id="j_nejsds100_eq_030">
<label>(3.15)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" maxsize="2.03em" minsize="2.03em">(</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:munder>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" maxsize="2.03em" minsize="2.03em">)</mml:mo>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" maxsize="2.03em" minsize="2.03em">{</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∣</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∣</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo fence="true" maxsize="2.03em" minsize="2.03em">}</mml:mo>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" maxsize="2.03em" minsize="2.03em">{</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" maxsize="2.03em" minsize="2.03em">}</mml:mo>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mi mathvariant="italic">ψ</mml:mi><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
<mml:mspace width="0.2778em"/>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}\mathbb{E}\bigg(\sum \limits_{i}{R_{i}}{Y_{i}}+1\bigg)& ={\mathbb{E}_{X}}\bigg\{{\sum \limits_{i=1}^{n}}\mathbb{E}({R_{i}}\mid {X_{i}})\mathbb{E}({Y_{i}}\mid {X_{i}})\bigg\}+1\\ {} & ={\mathbb{E}_{X}}\bigg\{{\sum \limits_{i=1}^{n}}{\theta _{{X_{i}}}}{p_{{X_{i}}}}\bigg\}+1\\ {} & =n\bar{\theta .p}+1=n\psi \bar{p}+n{\sigma _{\theta ,p}}+1\hspace{0.2778em}\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
<disp-formula id="j_nejsds100_eq_031">
<label>(3.16)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" maxsize="2.03em" minsize="2.03em">(</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:munder>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo>+</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal" fence="true" maxsize="2.03em" minsize="2.03em">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>+</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}\mathbb{E}\bigg(\sum \limits_{i}{R_{i}}& +2\bigg)=n\bar{p}+2.\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
</disp-formula-group> Hence, the expectation for the Bayes’ estimator can be approximated by taking the ratio of right hand sides from (<xref rid="j_nejsds100_eq_030">3.15</xref>) and (<xref rid="j_nejsds100_eq_031">3.16</xref>) as: 
<disp-formula id="j_nejsds100_eq_032">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="eqnarray-1">
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">≈</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mi mathvariant="italic">ψ</mml:mi><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>+</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="0.2778em"/>
<mml:mtext>as</mml:mtext>
<mml:mspace width="0.2778em"/>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \mathbb{E}({\hat{\psi }_{Li}})\approx \frac{n\psi \bar{p}+n{\sigma _{\theta ,p}}+1}{n\bar{p}+2}\to \psi +{\sigma _{\theta ,p}}/\bar{p},\hspace{0.2778em}\text{as}\hspace{0.2778em}n\to \infty ,\]]]></tex-math></alternatives>
</disp-formula> 
showing the <inline-formula id="j_nejsds100_ineq_155"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\hat{\psi }_{Li}}$]]></tex-math></alternatives></inline-formula> is asymptotically unbiased if <inline-formula id="j_nejsds100_ineq_156"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[${\sigma _{\theta ,p}}=0$]]></tex-math></alternatives></inline-formula>, i.e., if <inline-formula id="j_nejsds100_ineq_157"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">⊥</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∣</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi></mml:math><tex-math><![CDATA[${\theta _{X}}\perp {p_{X}}\mid X$]]></tex-math></alternatives></inline-formula>, or equivalently if <inline-formula id="j_nejsds100_ineq_158"><alternatives><mml:math>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo stretchy="false">∣</mml:mo>
<mml:mi mathvariant="italic">R</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\mathbb{E}(Y\mid R=1)=\mathbb{E}(Y)$]]></tex-math></alternatives></inline-formula>, as claimed by (<xref ref-type="bibr" rid="j_nejsds100_ref_041">Ritov et al.</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_041">2014</xref>). Now, the variances for the numerator and denominator can be calculated using the conditional variance identity. The expressions are given below. 
<disp-formula id="j_nejsds100_eq_033">
<label>(3.17)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" maxsize="2.03em" minsize="2.03em">(</mml:mo>
<mml:mspace width="-0.1667em"/>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:munder>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mspace width="0.1667em"/>
<mml:mo>+</mml:mo>
<mml:mspace width="0.1667em"/>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal" fence="true" maxsize="2.03em" minsize="2.03em">)</mml:mo>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mspace width="0.1667em"/>
<mml:mo>=</mml:mo>
<mml:mspace width="0.1667em"/>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mspace width="0.1667em"/>
<mml:mo>=</mml:mo>
<mml:mspace width="0.1667em"/>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∣</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo fence="true" stretchy="false">]</mml:mo>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∣</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo fence="true" stretchy="false">]</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mspace width="0.1667em"/>
<mml:mo>=</mml:mo>
<mml:mspace width="0.1667em"/>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mspace width="0.1667em"/>
<mml:mo>+</mml:mo>
<mml:mspace width="0.1667em"/>
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo fence="true" stretchy="false">]</mml:mo>
<mml:mspace width="0.1667em"/>
<mml:mo>=</mml:mo>
<mml:mspace width="0.1667em"/>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mspace width="0.1667em"/>
<mml:mo>−</mml:mo>
<mml:mspace width="0.1667em"/>
<mml:msup>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}\mathbb{V}\bigg(\hspace{-0.1667em}\sum \limits_{i}{R_{i}}\hspace{0.1667em}+\hspace{0.1667em}2\bigg)& \hspace{0.1667em}=\hspace{0.1667em}\sum \limits_{i}\mathbb{V}({R_{i}})\\ {} & \hspace{0.1667em}=\hspace{0.1667em}\sum \limits_{i}\mathbb{E}[\mathbb{V}({R_{i}}\mid {X_{i}}))]+\mathbb{V}[\mathbb{E}({R_{i}}\mid {X_{i}})]\\ {} & \hspace{0.1667em}=\hspace{0.1667em}\sum \limits_{i}[\mathbb{E}({p_{{x_{i}}}}(1-{p_{{X_{i}}}}))\hspace{0.1667em}+\hspace{0.1667em}\mathbb{V}({p_{{X_{i}}}})]\hspace{0.1667em}=\hspace{0.1667em}n(\bar{p}\hspace{0.1667em}-\hspace{0.1667em}{\bar{p}^{2}}).\end{aligned}\]]]></tex-math></alternatives>
</disp-formula> 
Similarly, 
<disp-formula id="j_nejsds100_eq_034">
<label>(3.18)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" maxsize="2.03em" minsize="2.03em">(</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:munder>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" maxsize="2.03em" minsize="2.03em">)</mml:mo>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>−</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mspace width="1em"/>
<mml:mtext>where</mml:mtext>
<mml:mspace width="0.2778em"/><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}\mathbb{V}\bigg(\sum \limits_{i}{R_{i}}{Y_{i}}+1\bigg)& =\sum \limits_{i}\mathbb{V}({R_{i}}{Y_{i}})=n(\bar{\theta .p}-{\bar{\theta .p}^{2}}),\\ {} \hspace{1em}\text{where}\hspace{0.2778em}\bar{\theta .p}& =\frac{1}{B}{\sum \limits_{b=1}^{B}}{\theta _{b}}{p_{b}}.\end{aligned}\]]]></tex-math></alternatives>
</disp-formula> 
Finally, the covariance term is: 
<disp-formula id="j_nejsds100_eq_035">
<label>(3.19)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="normal">Cov</mml:mi>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:munder>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="0.2778em"/>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:munder>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:mi mathvariant="normal">Cov</mml:mi>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:munder>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="0.2778em"/>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:munder>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mi mathvariant="normal">Cov</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="0.2778em"/>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mspace width="2.5pt"/>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}\mathrm{Cov}\left(\sum \limits_{i}{R_{i}}{Y_{i}}+1,\hspace{0.2778em}\sum \limits_{i}{R_{i}}+2\right)& =\mathrm{Cov}\left(\sum \limits_{i}{R_{i}}{Y_{i}},\hspace{0.2778em}\sum \limits_{i}{R_{i}}\right)\\ {} & =\sum \limits_{i}\mathrm{Cov}({R_{i}}{Y_{i}},\hspace{0.2778em}{R_{i}})\\ {} & =n\bar{\theta .p}\hspace{2.5pt}(1-\bar{p}).\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>Putting everything together from (<xref rid="j_nejsds100_eq_030">3.15</xref>), (<xref rid="j_nejsds100_eq_031">3.16</xref>), (<xref rid="j_nejsds100_eq_034">3.18</xref>), (<xref rid="j_nejsds100_eq_033">3.17</xref>) and (<xref rid="j_nejsds100_eq_035">3.19</xref>), we get the following formula for approximate variance for the Bayes estimator as: 
<disp-formula id="j_nejsds100_eq_036">
<label>(3.20)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd class="multline"/>
<mml:mtd>
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">≈</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mfenced separators="" open="(" close=")">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>+</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>×</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="multline"/>
<mml:mtd>
<mml:mfenced separators="" open="[" close="]">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>−</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>+</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>−</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>+</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>−</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mi mathvariant="italic">n</mml:mi><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>+</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{array}{cc}& \displaystyle \mathbb{V}({\hat{\psi }_{Li}})\approx {\left(\frac{n\bar{\theta .p}+1}{n\bar{p}+2}\right)^{2}}\times \\ {} & \displaystyle \left[\frac{n(\bar{\theta .p}-{\bar{\theta .p}^{2}})}{{(n\bar{\theta .p}+1)^{2}}}+\frac{n(\bar{p}-{\bar{p}^{2}})}{{(n\bar{p}+2)^{2}}}-\frac{2n\bar{\theta .p}(1-\bar{p})}{(n\bar{p}+2)(n\bar{\theta .p}+1)}\right].\end{array}\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>A couple of immediate implications of the formula in (<xref rid="j_nejsds100_eq_036">3.20</xref>) are as follows. 
<list>
<list-item id="j_nejsds100_li_004">
<label>(i)</label>
<p>Li’s Bayesian estimator is consistent if <inline-formula id="j_nejsds100_ineq_159"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{{X_{i}}}}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds100_ineq_160"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{{X_{i}}}}$]]></tex-math></alternatives></inline-formula> are uncorrelated, as <inline-formula id="j_nejsds100_ineq_161"><alternatives><mml:math>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi mathvariant="italic">ψ</mml:mi></mml:math><tex-math><![CDATA[$\mathbb{E}({\hat{\psi }_{Li}})\to \psi $]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds100_ineq_162"><alternatives><mml:math>
<mml:mi mathvariant="double-struck">V</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">O</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">→</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[$\mathbb{V}({\hat{\psi }_{Li}})=O(1/n)\to 0$]]></tex-math></alternatives></inline-formula> as the sample size <inline-formula id="j_nejsds100_ineq_163"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi></mml:math><tex-math><![CDATA[$n\to \infty $]]></tex-math></alternatives></inline-formula>.</p>
</list-item>
<list-item id="j_nejsds100_li_005">
<label>(ii)</label>
<p>Unlike the variance of the HT estimator, given in (<xref rid="j_nejsds100_eq_021">3.10</xref>), the variance of the Li’s Bayes estimator does not have the individual <inline-formula id="j_nejsds100_ineq_164"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{b}}$]]></tex-math></alternatives></inline-formula> terms in the denominator and will not inflate for very small <inline-formula id="j_nejsds100_ineq_165"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{b}}$]]></tex-math></alternatives></inline-formula> values. In other words, the restriction (<xref rid="j_nejsds100_eq_013">3.2</xref>) is not needed for the Bayesian solution.</p>
</list-item>
</list>
</p>
</sec>
</sec>
<sec id="j_nejsds100_s_019">
<label>4</label>
<title>Simulation Examples</title>
<sec id="j_nejsds100_s_020">
<label>4.1</label>
<title>Comparing IPW Estimators for Wasserman’s Example</title>
<sec id="j_nejsds100_s_021">
<label>4.1.1</label>
<title>Missing Completely at Random</title>
<p>We compare the estimation performance for four different candidate estimators for the Robins-Ritov-Wasserman’s problem. The candidates are: the original Horvitz–Thompson (<xref rid="j_nejsds100_eq_003">2.2</xref>), Hájek (<xref rid="j_nejsds100_eq_004">2.3</xref>), the Li’s estimator, <italic>i.e.,</italic> the Bayes posterior mean under a Beta hyperprior (<xref rid="j_nejsds100_eq_020">3.9</xref>) and <xref ref-type="bibr" rid="j_nejsds100_ref_012">Ghosh</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_012">2015</xref>)’s binning and smoothing idea applied to the Hájek estimator. We choose the bounds for <inline-formula id="j_nejsds100_ineq_166"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{i}}$]]></tex-math></alternatives></inline-formula>’s <inline-formula id="j_nejsds100_ineq_167"><alternatives><mml:math>
<mml:mi mathvariant="italic">δ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0.01</mml:mn></mml:math><tex-math><![CDATA[$\delta =0.01$]]></tex-math></alternatives></inline-formula>, and parameter space dimension <inline-formula id="j_nejsds100_ineq_168"><alternatives><mml:math>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1000</mml:mn></mml:math><tex-math><![CDATA[$B=1000$]]></tex-math></alternatives></inline-formula>, <italic>i.e.,</italic> the <inline-formula id="j_nejsds100_ineq_169"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{j}}$]]></tex-math></alternatives></inline-formula>’s in our experiment are <italic>B</italic> equidistant grid-points in <inline-formula id="j_nejsds100_ineq_170"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">δ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">δ</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[\delta ,1-\delta ]$]]></tex-math></alternatives></inline-formula>. We vary the support of generative distribution for <inline-formula id="j_nejsds100_ineq_171"><alternatives><mml:math>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="script">U</mml:mi>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$\theta \in \mathcal{U}[a,b]$]]></tex-math></alternatives></inline-formula> to four different values, viz. <inline-formula id="j_nejsds100_ineq_172"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>0.1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.9</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>0.1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.4</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>0.35</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.65</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>0.6</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.9</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$[a,b]\in \{[0.1,0.9],[0.1,0.4],[0.35,0.65],[0.6,0.9]\}$]]></tex-math></alternatives></inline-formula>. Finally, we take sample size <inline-formula id="j_nejsds100_ineq_173"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>100</mml:mn></mml:math><tex-math><![CDATA[$n=100$]]></tex-math></alternatives></inline-formula>, as was originally intended in the RRW example to make it analogous to a high-dimensional problem with <inline-formula id="j_nejsds100_ineq_174"><alternatives><mml:math>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo stretchy="false">≫</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi></mml:math><tex-math><![CDATA[$B\gg n$]]></tex-math></alternatives></inline-formula>.</p>
<p>Table <xref rid="j_nejsds100_tab_001">1</xref> and Fig. <xref rid="j_nejsds100_fig_002">2</xref> shows the mean squared errors calculated over 100 replicates, and shows the following: (1) the Li’s estimator leads to the lowest mean squared error over replicates, (2) the Horvitz–Thompson estimator performs the worst across all situations considered and finally, (3) the Hájek estimator and the Binning-Smoothing estimator (<xref ref-type="bibr" rid="j_nejsds100_ref_012">Ghosh</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_012">2015</xref>) achieves very similar performance and generally occupies a middle ground between the Bayes’ and the Horvitz–Thompson in terms of achieved mean-squared error.</p>
<table-wrap id="j_nejsds100_tab_001">
<label>Table 1</label>
<caption>
<p>Mean squared error comparison between the four candiate estimators: Horvitz–Thompson (<xref rid="j_nejsds100_eq_003">2.2</xref>), Hájek (<xref rid="j_nejsds100_eq_004">2.3</xref>), the Li’s Bayesian estimator (<xref rid="j_nejsds100_eq_020">3.9</xref>) and <xref ref-type="bibr" rid="j_nejsds100_ref_012">Ghosh</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_012">2015</xref>)’s estimator for different values of <inline-formula id="j_nejsds100_ineq_175"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[a,b]$]]></tex-math></alternatives></inline-formula>, where <inline-formula id="j_nejsds100_ineq_176"><alternatives><mml:math>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$\theta \in [a,b]$]]></tex-math></alternatives></inline-formula>. The numbers on the table are reported after multiplying the MSE by <inline-formula id="j_nejsds100_ineq_177"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${10^{2}}$]]></tex-math></alternatives></inline-formula> for better comparison and the column winner are denoted by boldfaced entries.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: double; border-bottom: solid thin"/>
<td colspan="2" style="vertical-align: top; text-align: center; border-top: double; border-bottom: solid thin">[0.6, 0.9]</td>
<td colspan="2" style="vertical-align: top; text-align: center; border-top: double; border-bottom: solid thin">[0.1, 0.9]</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">method</td>
<td style="vertical-align: top; text-align: center">mean</td>
<td style="vertical-align: top; text-align: right">sd</td>
<td style="vertical-align: top; text-align: right">mean</td>
<td style="vertical-align: top; text-align: right">sd</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Bayes’ (Li’s)</td>
<td style="vertical-align: top; text-align: center"><bold>0.37198</bold></td>
<td style="vertical-align: top; text-align: right">0.05093</td>
<td style="vertical-align: top; text-align: right"><bold>0.47541</bold></td>
<td style="vertical-align: top; text-align: right">0.05980</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Binning-smoothing</td>
<td style="vertical-align: top; text-align: center">0.63991</td>
<td style="vertical-align: top; text-align: right">0.08583</td>
<td style="vertical-align: top; text-align: right">0.85225</td>
<td style="vertical-align: top; text-align: right">0.10204</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Hájek</td>
<td style="vertical-align: top; text-align: center">0.71787</td>
<td style="vertical-align: top; text-align: right">0.11643</td>
<td style="vertical-align: top; text-align: right">0.95824</td>
<td style="vertical-align: top; text-align: right">0.12818</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">HT</td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin">3.09007</td>
<td style="vertical-align: top; text-align: right; border-bottom: solid thin">0.83348</td>
<td style="vertical-align: top; text-align: right; border-bottom: solid thin">2.27093</td>
<td style="vertical-align: top; text-align: right; border-bottom: solid thin">0.72187</td>
</tr>
</tbody><tbody>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td colspan="2" style="vertical-align: top; text-align: center; border-bottom: solid thin">[0.1, 0.4]</td>
<td colspan="2" style="vertical-align: top; text-align: center; border-bottom: solid thin">[0.35, 0.65]</td>
</tr>
</tbody><tbody>
<tr>
<td style="vertical-align: top; text-align: left">method</td>
<td style="vertical-align: top; text-align: center">mean</td>
<td style="vertical-align: top; text-align: right">sd</td>
<td style="vertical-align: top; text-align: right">mean</td>
<td style="vertical-align: top; text-align: right">sd</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Bayes’ (Li’s)</td>
<td style="vertical-align: top; text-align: center"><bold>0.35743</bold></td>
<td style="vertical-align: top; text-align: right">0.04758</td>
<td style="vertical-align: top; text-align: right"><bold>0.47329</bold></td>
<td style="vertical-align: top; text-align: right">0.06272</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Binning-smoothing</td>
<td style="vertical-align: top; text-align: center">0.64379</td>
<td style="vertical-align: top; text-align: right">0.08114</td>
<td style="vertical-align: top; text-align: right">0.86459</td>
<td style="vertical-align: top; text-align: right">0.10523</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Hájek</td>
<td style="vertical-align: top; text-align: center">0.73526</td>
<td style="vertical-align: top; text-align: right">0.13494</td>
<td style="vertical-align: top; text-align: right">0.96710</td>
<td style="vertical-align: top; text-align: right">0.13523</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">HT</td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin">1.17543</td>
<td style="vertical-align: top; text-align: right; border-bottom: solid thin">0.51219</td>
<td style="vertical-align: top; text-align: right; border-bottom: solid thin">2.22213</td>
<td style="vertical-align: top; text-align: right; border-bottom: solid thin">0.69317</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="j_nejsds100_fig_002">
<label>Figure 2</label>
<caption>
<p>Mean square error comparison for Horvitz-Thompson, Hájek and the Li’s estimator (<xref rid="j_nejsds100_eq_020">3.9</xref>) for different generative distributions of <inline-formula id="j_nejsds100_ineq_178"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{1}},\dots ,{\theta _{B}}$]]></tex-math></alternatives></inline-formula>, and fixed values of <italic>n</italic>, <italic>B</italic> and <italic>δ</italic>. In each case, Bayes’ estimator has the lowest MSE, followed by Hájek, and HT has the largest MSE.</p>
</caption>
<graphic xlink:href="nejsds100_g002.jpg"/>
</fig>
<p>For a single replication with <inline-formula id="j_nejsds100_ineq_179"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>000</mml:mn></mml:math><tex-math><![CDATA[$1,000$]]></tex-math></alternatives></inline-formula> evaluations, Fig. <xref rid="j_nejsds100_fig_003">3</xref> shows the bias and variance for the four candidate estimators. Fig. <xref rid="j_nejsds100_fig_003">3</xref> demonstrates the nature of variance reduction by Bayes’ estimator and the Binning-smoothing idea in estimating <inline-formula id="j_nejsds100_ineq_180"><alternatives><mml:math>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="bold-italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\psi =\mathbb{E}(\boldsymbol{\theta })$]]></tex-math></alternatives></inline-formula> for different ranges of <italic>θ</italic>, without any significant increase in bias.</p>
<fig id="j_nejsds100_fig_003">
<label>Figure 3</label>
<caption>
<p>Histogram and Kernel Density plots showing the distribution of Horvitz–Thompson, Hájek and the Bayes estimators for different generative distributions of <inline-formula id="j_nejsds100_ineq_181"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{1}},\dots ,{\theta _{B}}$]]></tex-math></alternatives></inline-formula>, and fixed values of <italic>n</italic>, <italic>B</italic> and <italic>δ</italic>.</p>
</caption>
<graphic xlink:href="nejsds100_g003.jpg"/>
</fig>
</sec>
<sec id="j_nejsds100_s_022">
<label>4.1.2</label>
<title>Missing at Random</title>
<p>As discussed earlier, Li’s estimator can be biased when there is a non-zero correlation between <inline-formula id="j_nejsds100_ineq_182"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{X}}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds100_ineq_183"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{X}}$]]></tex-math></alternatives></inline-formula>, <italic>e.g.,</italic> when the missingness might arise due to confounding. We show an example here to illustrate the effect of this situation, and compare the four candidates. We take <inline-formula id="j_nejsds100_ineq_184"><alternatives><mml:math>
<mml:mi mathvariant="italic">δ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0.1</mml:mn></mml:math><tex-math><![CDATA[$\delta =0.1$]]></tex-math></alternatives></inline-formula>, <italic>i.e.,</italic> <inline-formula id="j_nejsds100_ineq_185"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{b}}$]]></tex-math></alternatives></inline-formula>’s to be equidistant points in <inline-formula id="j_nejsds100_ineq_186"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>0.1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.9</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[0.1,0.9]$]]></tex-math></alternatives></inline-formula>, and instead of <inline-formula id="j_nejsds100_ineq_187"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{b}}$]]></tex-math></alternatives></inline-formula>’s to be uniformly distributed in <inline-formula id="j_nejsds100_ineq_188"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[a,b]$]]></tex-math></alternatives></inline-formula>, we take: 
<disp-formula id="j_nejsds100_eq_037">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">γ</mml:mi>
<mml:mo>×</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">γ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>×</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">U</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="0.2778em"/>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mtext>where</mml:mtext>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mspace width="0.2778em"/>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">U</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∼</mml:mo>
<mml:mi mathvariant="script">U</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="0.2778em"/>
<mml:mi mathvariant="italic">γ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0.5</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}{\theta _{b}}& =\gamma \times {p_{b}}+(1-\gamma )\times {U_{b}},\hspace{0.2778em}\\ {} \text{where}& \hspace{0.2778em}{U_{b}}\sim \mathcal{U}(a,b),\hspace{0.2778em}\gamma =0.5,b=1,\dots ,B.\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>The remaining set-up is same as before. Clearly, in this case the true value of <inline-formula id="j_nejsds100_ineq_189"><alternatives><mml:math>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>0.5</mml:mn></mml:math><tex-math><![CDATA[$\psi =\mathbb{E}(\theta )=0.5$]]></tex-math></alternatives></inline-formula>. Figure <xref rid="j_nejsds100_fig_004">4</xref>a shows that the Li’s estimator has an upward bias, as expected, because of the positive correlation between <italic>θ</italic> and <italic>p</italic>, but the Hájek and Binned-Smoothed estimators are unaffected. On the other hand, the Binned-Smoothed is significantly better than all the other candidates (HT, Hájek and Li’s) in terms of MSE (Fig. <xref rid="j_nejsds100_fig_004">4</xref>b). This shows that the Li’s posterior mean estimator can be biased when there is confounding and one needs to be careful when using it. The binned-smoothed method, on the other hand, performs well irrespective of the correlation between <inline-formula id="j_nejsds100_ineq_190"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{X}}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds100_ineq_191"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{X}}$]]></tex-math></alternatives></inline-formula>, <italic>i.e.,</italic> for both MAR and MCAR-type missingness.</p>
<fig id="j_nejsds100_fig_004">
<label>Figure 4</label>
<caption>
<p>Histogram, Kernel density plots and MSE comparison for Horvitz–Thompson, Hájek and the Bayes estimators for the missing-at-random scenario.</p>
</caption>
<graphic xlink:href="nejsds100_g004.jpg"/>
</fig>
</sec>
</sec>
</sec>
<sec id="j_nejsds100_s_023">
<label>5</label>
<title>Discussion</title>
<p>Our main goal in this paper was to shed light on various aspects of the family of inverse probability weight estimators, discuss its weakness and strengths and highlight the connections between survey sampling and Monte Carlo integration via these popular tool pervasive in Statistical literature. For survey sampling, we consider some of the ‘weak paradoxes’ (<xref ref-type="bibr" rid="j_nejsds100_ref_012">Ghosh</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_012">2015</xref>) that arise from using inverse probability estimators using the well-known examples from <xref ref-type="bibr" rid="j_nejsds100_ref_054">Wasserman</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_054">2004</xref>) and <xref ref-type="bibr" rid="j_nejsds100_ref_001">Basu</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_001">1988</xref>), and review the merits and demerits of popular IPW estimators. In particular, we show that the hierarchical Bayes’ estimator due to (<xref ref-type="bibr" rid="j_nejsds100_ref_026">Li</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_026">2010</xref>) leads to robustness against pathological situations but admits bias in presence of confounding. We provide sufficient conditions for consistency of the Li’s posterior mean estimate in Wasserman’s example and show that, under certain conditions, both the Li’s estimator and the binning-smoothing idea (<xref ref-type="bibr" rid="j_nejsds100_ref_012">Ghosh</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_012">2015</xref>) achieves lower variation in mean squared errors. We then highlight the analogous tools in Monte Carlo integration, revisiting a few earlier works (e.g., <xref ref-type="bibr" rid="j_nejsds100_ref_017">Hesterberg</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_017">1988</xref>), where the Horvitz–Thompson estimator is likened to the naïve importance sampling, and much like survey sampling, self-normalized weights or using control variates lead to better estimators. We conclude with a brief discussion of applications in the context of causal inference, a key use of IPW estimators, and a few possible directions for future work.</p>
<p>A possible future direction, borrowing from (<xref ref-type="bibr" rid="j_nejsds100_ref_025">Kong et al.</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_025">2003</xref>) is to use the general semiparametric models and the associated estimators or computational algorithm for <inline-formula id="j_nejsds100_ineq_192"><alternatives><mml:math>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$k\gt 1$]]></tex-math></alternatives></inline-formula> in the context of survey sampling that would allow one to combine data from one or more surveys that use different but known inclusion probabilities. We conjecture that this might lead to improved estimators for average treatment effect estimation, which we have briefly described in section <xref rid="j_nejsds100_s_009">2.3</xref>.</p>
<p>In the context of causal inference, a second possible direction for future research is comparing a suitable modification of the Bayes estimator in (<xref rid="j_nejsds100_eq_020">3.9</xref>) for the ATE problem with that of the adaptive estimators developed by <xref ref-type="bibr" rid="j_nejsds100_ref_023">Khan and Ugander</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_023">2023</xref>) and both empirically and theoretically to investigate consistency properties. It will be also worthwhile to consider the semiparametric model in (<xref ref-type="bibr" rid="j_nejsds100_ref_025">Kong et al.</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_025">2003</xref>) for ATE and compare the results of simultaneous estimation using MLE with these candidates.</p>
<p>Finally, the problems presented in (<xref ref-type="bibr" rid="j_nejsds100_ref_054">Wasserman</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_054">2004</xref>) and discussed in (<xref ref-type="bibr" rid="j_nejsds100_ref_047">Sims</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_047">2007</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_046">2010</xref>) are inherently high-dimensional in nature, where sparsity is pervasive. There is a large and growing literature on promoting sparsity in causal inference, <italic>e.g.,</italic> in presence of a large number of confounders or baseline variables while estimating the effect of an exposure on an outcome. We refer the readers to <xref ref-type="bibr" rid="j_nejsds100_ref_045">Shortreed and Ertefaie</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_045">2017</xref>); <xref ref-type="bibr" rid="j_nejsds100_ref_053">Wang et al.</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_053">2012</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_052">2015</xref>); <xref ref-type="bibr" rid="j_nejsds100_ref_024">Kim et al.</xref> (<xref ref-type="bibr" rid="j_nejsds100_ref_024">2022</xref>) for recent advances in this area. In our simple focal example, the high-dimensional parameter <inline-formula id="j_nejsds100_ineq_193"><alternatives><mml:math>
<mml:mi mathvariant="bold-italic">θ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\boldsymbol{\theta }=({\theta _{1}},\dots ,{\theta _{B}})$]]></tex-math></alternatives></inline-formula> or <inline-formula id="j_nejsds100_ineq_194"><alternatives><mml:math>
<mml:mi mathvariant="bold">p</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\mathbf{p}=({p_{1}},\dots ,{p_{B}})$]]></tex-math></alternatives></inline-formula> could exhibit sparsity with a small <inline-formula id="j_nejsds100_ineq_195"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi>ℓ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\ell _{0}}$]]></tex-math></alternatives></inline-formula> norm or other notions of sparsity. To handle sparsity in higher dimensions while maintaining tail-robustness and accuracy, the state-of-the-art Bayesian solution would be augmenting (<xref rid="j_nejsds100_eq_016">3.5</xref>) with global-local shrinkage priors (<xref ref-type="bibr" rid="j_nejsds100_ref_035">Polson and Scott</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_035">2010</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_036">2012</xref>; <xref ref-type="bibr" rid="j_nejsds100_ref_002">Bhadra et al.</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_002">2019</xref>), <italic>i.e.,</italic> 
<disp-formula id="j_nejsds100_eq_038">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∣</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">ψ</mml:mi>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mover>
<mml:mrow>
<mml:mo stretchy="false">∼</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">ind</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mover>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>HS</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="0.2778em"/>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="0.2778em"/>
<mml:mi mathvariant="italic">τ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>HS</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>;</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">τ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">τ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo fence="true" stretchy="false">}</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∣</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo stretchy="false">∼</mml:mo>
<mml:mtext>Beta</mml:mtext>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>×</mml:mo>
<mml:mi mathvariant="italic">π</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mspace width="0.2778em"/>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">ψ</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}{\theta _{b}}\mid {\alpha _{T}},\psi & \stackrel{\mathrm{ind}}{\sim }{f_{\text{HS}}}(\theta ;{\alpha _{T}}\psi ,\hspace{0.2778em}{\alpha _{T}}(1-\psi ),\hspace{0.2778em}\tau )\\ {} {f_{\text{HS}}}(\theta ;a,b,\tau )& ={\theta ^{a-1}}{(1-\theta )^{b-1}}{\{1-(1-\theta ){\tau ^{2}}\}^{-(a+b)}},\\ {} (\psi ,{\alpha _{T}}\mid {\alpha _{F}})& \sim \text{Beta}({\alpha _{F}},{\alpha _{F}})\times \pi ({\alpha _{T}}),\\ {} & \hspace{0.2778em}\theta ,\psi \in [0,1].\end{aligned}\]]]></tex-math></alternatives>
</disp-formula> 
Here, as before, <italic>ψ</italic> is the mean of <italic>θ</italic>, <inline-formula id="j_nejsds100_ineq_196"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\alpha _{T}}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds100_ineq_197"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\alpha _{F}}$]]></tex-math></alternatives></inline-formula> are the and the shape parameters for <inline-formula id="j_nejsds100_ineq_198"><alternatives><mml:math>
<mml:mi mathvariant="bold-italic">θ</mml:mi></mml:math><tex-math><![CDATA[$\boldsymbol{\theta }$]]></tex-math></alternatives></inline-formula> and <italic>ψ</italic> respectively and <italic>τ</italic> is a global shrinkage parameter adjusting to sparsity. Such Bayesian regularization priors have been used successfully for treatment effect estimation with more control variates than observations (<xref ref-type="bibr" rid="j_nejsds100_ref_013">Hahn et al.</xref>, <xref ref-type="bibr" rid="j_nejsds100_ref_013">2018</xref>). Designing built-in sparsity priors that can also incorporate selection mechanism for confounder selection is an interesting problem that we plan to address in a future endeavor.</p>
</sec>
</body>
<back>
<ack id="j_nejsds100_ack_001">
<title>Acknowledgements</title>
<p>The first author was inspired to study this problem after attending a talk by Professor Christopher Sims, who passed away on March 14, 2026. We dedicate this paper to his memory.</p></ack>
<ref-list id="j_nejsds100_reflist_001">
<title>References</title>
<ref id="j_nejsds100_ref_001">
<mixed-citation publication-type="book"><string-name><surname>Basu</surname>, <given-names>D.</given-names></string-name> <source>Statistical information and likelihood: a collection of critical essays by Dr. D. Basu</source>. (<string-name><given-names>J. K.</given-names> <surname>Ghosh</surname></string-name>, ed.) <series>Lecture Notes in Statistics</series> <volume>45</volume>. <publisher-name>Springer Science &amp; Business Media</publisher-name> (<year>1988</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0953081">MR0953081</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_002">
<mixed-citation publication-type="journal"><string-name><surname>Bhadra</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Datta</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Polson</surname>, <given-names>N. G.</given-names></string-name> and <string-name><surname>Willard</surname>, <given-names>B.</given-names></string-name> <article-title>Lasso meets horseshoe</article-title>. <source>Statistical Science</source> <volume>34</volume>(<issue>3</issue>) <fpage>405</fpage>–<lpage>427</lpage> (<year>2019</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/19-STS700" xlink:type="simple">https://doi.org/10.1214/19-STS700</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4017521">MR4017521</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_003">
<mixed-citation publication-type="other"><string-name><surname>Chauvet</surname>, <given-names>G.</given-names></string-name> A note on the consistency of the narain-horvitz-thompson estimator (<year>2014</year>). <italic>arXiv preprint arXiv:</italic><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1412.2887"><italic>1412.2887</italic></ext-link>.</mixed-citation>
</ref>
<ref id="j_nejsds100_ref_004">
<mixed-citation publication-type="journal"><string-name><surname>Chen</surname>, <given-names>Q.</given-names></string-name>, <string-name><surname>Elliott</surname>, <given-names>M. R.</given-names></string-name>, <string-name><surname>Haziza</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Ghosh</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Little</surname>, <given-names>R. J.</given-names></string-name>, <string-name><surname>Sedransk</surname>, <given-names>J.</given-names></string-name> and <string-name><surname>Thompson</surname>, <given-names>M.</given-names></string-name> <article-title>Approaches to improving survey-weighted estimates</article-title>. <source>Statistical Science</source> <volume>32</volume>(<issue>2</issue>) <fpage>227</fpage>–<lpage>248</lpage> (<year>2017</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/17-STS609" xlink:type="simple">https://doi.org/10.1214/17-STS609</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3648957">MR3648957</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_005">
<mixed-citation publication-type="journal"><string-name><surname>Chopin</surname>, <given-names>N.</given-names></string-name> and <string-name><surname>Robert</surname>, <given-names>C. P.</given-names></string-name> <article-title>Properties of nested sampling</article-title>. <source>Biometrika</source> <volume>97</volume>(<issue>3</issue>) <fpage>741</fpage>–<lpage>755</lpage> (<year>2010</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1093/biomet/asq021" xlink:type="simple">https://doi.org/10.1093/biomet/asq021</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2672495">MR2672495</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_006">
<mixed-citation publication-type="chapter"><string-name><surname>Cunningham</surname>, <given-names>S.</given-names></string-name> <chapter-title>Causal inference</chapter-title>. In <source>Causal Inference</source>. <publisher-name>Yale University Press</publisher-name> (<year>2021</year>).</mixed-citation>
</ref>
<ref id="j_nejsds100_ref_007">
<mixed-citation publication-type="other"><string-name><surname>Datta</surname>, <given-names>J.</given-names></string-name> and <string-name><surname>Polson</surname>, <given-names>N. G.</given-names></string-name> Quantile importance sampling (<year>2025</year>). <italic>arXiv preprint arXiv:</italic><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2305.03158"><italic>2305.03158</italic></ext-link>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/25-bjps638" xlink:type="simple">https://doi.org/10.1214/25-bjps638</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=5007225">MR5007225</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_008">
<mixed-citation publication-type="journal"><string-name><surname>Delevoye</surname>, <given-names>A.</given-names></string-name> and <string-name><surname>Sävje</surname>, <given-names>F.</given-names></string-name> <article-title>Consistency of the horvitz–thompson estimator under general sampling and experimental designs</article-title>. <source>Journal of Statistical Planning and Inference</source> <volume>207</volume>. <fpage>190</fpage>–<lpage>197</lpage> (<year>2020</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1016/j.jspi.2019.12.002" xlink:type="simple">https://doi.org/10.1016/j.jspi.2019.12.002</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4066130">MR4066130</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_009">
<mixed-citation publication-type="journal"><string-name><surname>Diaconis</surname>, <given-names>P.</given-names></string-name> <article-title>Bayesian numerical analysis</article-title>. <source>Statistical decision theory and related topics IV</source> <volume>1</volume>. <fpage>163</fpage>–<lpage>175</lpage> (<year>1988</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0927099">MR0927099</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_010">
<mixed-citation publication-type="journal"><string-name><surname>Efron</surname>, <given-names>B.</given-names></string-name> and <string-name><surname>Morris</surname>, <given-names>C.</given-names></string-name> <article-title>Stein’s paradox in statistics</article-title>. <source>Scientific American</source> <volume>236</volume>(<issue>5</issue>) <fpage>119</fpage>–<lpage>127</lpage> (<year>1977</year>).</mixed-citation>
</ref>
<ref id="j_nejsds100_ref_011">
<mixed-citation publication-type="journal"><string-name><surname>Firth</surname>, <given-names>D.</given-names></string-name> <article-title>On improved estimation for importance sampling</article-title>. <source>Brazilian Journal of Probability and Statistics</source> <volume>25</volume>(<issue>3</issue>) <fpage>437</fpage>–<lpage>443</lpage> (<year>2011</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/11-BJPS155" xlink:type="simple">https://doi.org/10.1214/11-BJPS155</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2832895">MR2832895</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_012">
<mixed-citation publication-type="chapter"><string-name><surname>Ghosh</surname>, <given-names>J. K.</given-names></string-name> <chapter-title>Weak paradoxes and paradigms</chapter-title>. In <source>Statistical Paradigms: Recent Advances and Reconciliations</source> <fpage>3</fpage>–<lpage>12</lpage>. <publisher-name>World Scientific</publisher-name> (<year>2015</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1142/9789814343961_0001" xlink:type="simple">https://doi.org/10.1142/9789814343961_0001</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3308074">MR3308074</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_013">
<mixed-citation publication-type="journal"><string-name><surname>Hahn</surname>, <given-names>P. R.</given-names></string-name>, <string-name><surname>Carvalho</surname>, <given-names>C. M.</given-names></string-name>, <string-name><surname>Puelz</surname>, <given-names>D.</given-names></string-name> and <string-name><surname>He</surname>, <given-names>J.</given-names></string-name> <article-title>Regularization and confounding in linear regression for treatment effect estimation</article-title>. <source>Bayesian Analysis</source> <volume>13</volume>(<issue>1</issue>) <fpage>163</fpage>–<lpage>182</lpage> (<year>2018</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/16-BA1044" xlink:type="simple">https://doi.org/10.1214/16-BA1044</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3737947">MR3737947</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_014">
<mixed-citation publication-type="other"><string-name><surname>Harmeling</surname>, <given-names>S.</given-names></string-name> and <string-name><surname>Touissant</surname>, <given-names>M.</given-names></string-name> Bayesian estimators for robins-ritov’s problem. Technical report, (<year>2007</year>).</mixed-citation>
</ref>
<ref id="j_nejsds100_ref_015">
<mixed-citation publication-type="other"><string-name><surname>Haziza</surname>, <given-names>D.</given-names></string-name> and <string-name><surname>Beaumont</surname>, <given-names>J.-F.</given-names></string-name> (<year>2017</year>). Construction of weights in surveys: A review. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/16-STS608" xlink:type="simple">https://doi.org/10.1214/16-STS608</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3648956">MR3648956</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_016">
<mixed-citation publication-type="journal"><string-name><surname>Hesterberg</surname>, <given-names>T.</given-names></string-name> <article-title>Weighted average importance sampling and defensive mixture distributions</article-title>. <source>Technometrics</source> <volume>37</volume>(<issue>2</issue>) <fpage>185</fpage>–<lpage>194</lpage> (<year>1995</year>).</mixed-citation>
</ref>
<ref id="j_nejsds100_ref_017">
<mixed-citation publication-type="book"><string-name><surname>Hesterberg</surname>, <given-names>T. C.</given-names></string-name> <source>Advances in importance sampling</source>. <comment>PhD thesis</comment>, <publisher-name>Stanford University</publisher-name> (<year>1988</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2637036">MR2637036</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_018">
<mixed-citation publication-type="journal"><string-name><surname>Horvitz</surname>, <given-names>D. G.</given-names></string-name> and <string-name><surname>Thompson</surname>, <given-names>D. J.</given-names></string-name> <article-title>A generalization of sampling without replacement from a finite universe</article-title>. <source>Journal of the American statistical Association</source> <volume>47</volume>(<issue>260</issue>) <fpage>663</fpage>–<lpage>685</lpage> (<year>1952</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0053460">MR0053460</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_019">
<mixed-citation publication-type="chapter"><string-name><surname>Hájek</surname>, <given-names>J.</given-names></string-name> <chapter-title>Comment on “an essay on the logical foundations of survey sampling” by basu</chapter-title>. (<string-name><given-names>V. P.</given-names> <surname>Godambe</surname></string-name> and <string-name><given-names>D. A.</given-names> <surname>Sprott</surname></string-name>, eds.) In <source>Foundations of Statistical Inference</source> <fpage>236</fpage>–<lpage>242</lpage>. <publisher-name>Holt, Rinehart and Winston</publisher-name> (<year>1971</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0423625">MR0423625</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_020">
<mixed-citation publication-type="journal"><string-name><surname>Imbens</surname>, <given-names>G. W.</given-names></string-name> <article-title>Nonparametric estimation of average treatment effects under exogeneity: A review</article-title>. <source>Review of Economics and statistics</source> <volume>86</volume>(<issue>1</issue>) <fpage>4</fpage>–<lpage>29</lpage> (<year>2004</year>).</mixed-citation>
</ref>
<ref id="j_nejsds100_ref_021">
<mixed-citation publication-type="journal"><string-name><surname>Isaki</surname>, <given-names>C. T.</given-names></string-name> and <string-name><surname>Fuller</surname>, <given-names>W. A.</given-names></string-name> <article-title>Survey design under the regression superpopulation model</article-title>. <source>Journal of the American Statistical Association</source> <volume>77</volume>(<issue>377</issue>) <fpage>89</fpage>–<lpage>96</lpage> (<year>1982</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0648029">MR0648029</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_022">
<mixed-citation publication-type="chapter"><string-name><surname>James</surname>, <given-names>W.</given-names></string-name> and <string-name><surname>Stein</surname>, <given-names>C.</given-names></string-name> <chapter-title>Estimation with quadratic loss</chapter-title>. In <source>Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Contributions to the Theory of Statistics</source> <fpage>361</fpage>–<lpage>379</lpage>. <publisher-name>University of California Press</publisher-name> (<year>1961</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0133191">MR0133191</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_023">
<mixed-citation publication-type="journal"><string-name><surname>Khan</surname>, <given-names>S.</given-names></string-name> and <string-name><surname>Ugander</surname>, <given-names>J.</given-names></string-name> <article-title>Adaptive normalization for ipw estimation</article-title>. <source>Journal of Causal Inference</source> <volume>11</volume>(<issue>1</issue>) <elocation-id>20220019</elocation-id> (<year>2023</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1515/jci-2022-0019" xlink:type="simple">https://doi.org/10.1515/jci-2022-0019</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4545024">MR4545024</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_024">
<mixed-citation publication-type="other"><string-name><surname>Kim</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Tec</surname>, <given-names>M.</given-names></string-name> and <string-name><surname>Zigler</surname>, <given-names>C. M.</given-names></string-name> Bayesian nonparametric adjustment of confounding (<year>2022</year>). <italic>arXiv preprint arXiv:</italic><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2203.11798"><italic>2203.11798</italic></ext-link>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1111/biom.13833" xlink:type="simple">https://doi.org/10.1111/biom.13833</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4680719">MR4680719</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_025">
<mixed-citation publication-type="journal"><string-name><surname>Kong</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>McCullagh</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Meng</surname>, <given-names>X.-L.</given-names></string-name>, <string-name><surname>Nicolae</surname>, <given-names>D.</given-names></string-name> and <string-name><surname>Tan</surname>, <given-names>Z.</given-names></string-name> <article-title>A theory of statistical models for monte carlo integration</article-title>. <source>Journal of the Royal Statistical Society: Series B (Statistical Methodology)</source> <volume>65</volume>(<issue>3</issue>) <fpage>585</fpage>–<lpage>604</lpage> (<year>2003</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1111/1467-9868.00404" xlink:type="simple">https://doi.org/10.1111/1467-9868.00404</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1998624">MR1998624</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_026">
<mixed-citation publication-type="journal"><string-name><surname>Li</surname>, <given-names>L.</given-names></string-name> <article-title>Are bayesian inferences weak for wasserman’s example?</article-title> <source>Communications in Statistics—Simulation and Computationő</source> <volume>39</volume>(<issue>3</issue>) <fpage>655</fpage>–<lpage>667</lpage> (<year>2010</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/03610910903576540" xlink:type="simple">https://doi.org/10.1080/03610910903576540</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2784548">MR2784548</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_027">
<mixed-citation publication-type="journal"><string-name><surname>Linero</surname>, <given-names>A. R.</given-names></string-name> <article-title>In nonparametric and high-dimensional models, bayesian ignorability is an informative prior</article-title>. <source>Journal of the American Statistical Association</source> <volume>119</volume>(<issue>548</issue>) <fpage>2785</fpage>–<lpage>2798</lpage> (<year>2024</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/01621459.2023.2278202" xlink:type="simple">https://doi.org/10.1080/01621459.2023.2278202</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4833915">MR4833915</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_028">
<mixed-citation publication-type="journal"><string-name><surname>Little</surname>, <given-names>R. J.</given-names></string-name> <article-title>Weighting and prediction in sample surveys</article-title>. <source>Calcutta Statistical Association Bulletin</source> <volume>60</volume>(<issue>3–4</issue>) <fpage>147</fpage>–<lpage>167</lpage> (<year>2008</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1177/0008068320080301" xlink:type="simple">https://doi.org/10.1177/0008068320080301</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2553424">MR2553424</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_029">
<mixed-citation publication-type="journal"><string-name><surname>Madrid-Padilla</surname>, <given-names>O.-H.</given-names></string-name>, <string-name><surname>Polson</surname>, <given-names>N. G.</given-names></string-name> and <string-name><surname>Scott</surname>, <given-names>J.</given-names></string-name> <article-title>A deconvolution path for mixtures</article-title>. <source>Electronic Journal of Statistics</source> <volume>12</volume>(<issue>1</issue>) <fpage>1717</fpage>–<lpage>1751</lpage> (<year>2018</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/18-ejs1430" xlink:type="simple">https://doi.org/10.1214/18-ejs1430</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3806437">MR3806437</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_030">
<mixed-citation publication-type="journal"><string-name><surname>Narain</surname>, <given-names>R.</given-names></string-name> <article-title>On sampling without replacement with varying probabilities</article-title>. <source>Journal of the Indian Society of Agricultural Statistics</source> <volume>3</volume>(<issue>2</issue>) <fpage>169</fpage>–<lpage>175</lpage> (<year>1951</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0045354">MR0045354</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_031">
<mixed-citation publication-type="journal"><string-name><surname>Neyman</surname>, <given-names>J.</given-names></string-name> <article-title>On the two different aspects of the representative method: The method of stratified sampling and the method of purposive selection</article-title>. <source>Journal of the Royal Statistical Society Series A: Statistics in Society</source> <volume>97</volume>(<issue>4</issue>) <fpage>558</fpage>–<lpage>606</lpage> (<year>1934</year>).</mixed-citation>
</ref>
<ref id="j_nejsds100_ref_032">
<mixed-citation publication-type="journal"><string-name><surname>Owen</surname>, <given-names>A. B.</given-names></string-name> <article-title>Comment: Unreasonable effectiveness of monte carlo</article-title>. <source>Statistical Science</source> <volume>34</volume>(<issue>1</issue>) <fpage>29</fpage>–<lpage>33</lpage> (<year>2019</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/18-STS676" xlink:type="simple">https://doi.org/10.1214/18-STS676</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3938960">MR3938960</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_033">
<mixed-citation publication-type="journal"><string-name><surname>Philippe</surname>, <given-names>A.</given-names></string-name> <article-title>Processing simulation output by riemann sums</article-title>. <source>Journal of Statistical Computation and Simulation</source> <volume>59</volume>(<issue>4</issue>) <fpage>295</fpage>–<lpage>314</lpage> (<year>1997</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/00949659708811863" xlink:type="simple">https://doi.org/10.1080/00949659708811863</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1623186">MR1623186</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_034">
<mixed-citation publication-type="journal"><string-name><surname>Philippe</surname>, <given-names>A.</given-names></string-name> and <string-name><surname>Robert</surname>, <given-names>C. P.</given-names></string-name> <article-title>Riemann sums for mcmc estimation and convergence monitoring</article-title>. <source>Statistics and Computing</source> <volume>11</volume>(<issue>2</issue>) <fpage>103</fpage>–<lpage>115</lpage> (<year>2001</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1023/A:1008926514119" xlink:type="simple">https://doi.org/10.1023/A:1008926514119</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1837130">MR1837130</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_035">
<mixed-citation publication-type="journal"><string-name><surname>Polson</surname>, <given-names>N. G.</given-names></string-name> and <string-name><surname>Scott</surname>, <given-names>J. G.</given-names></string-name> <article-title>Shrink globally, act locally: Sparse Bayesian regularization and prediction</article-title>. <source>Bayesian Statistics</source> <volume>9</volume>. <fpage>501</fpage>–<lpage>538</lpage> (<year>2010</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1093/acprof:oso/9780199694587.003.0017" xlink:type="simple">https://doi.org/10.1093/acprof:oso/9780199694587.003.0017</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3204017">MR3204017</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_036">
<mixed-citation publication-type="journal"><string-name><surname>Polson</surname>, <given-names>N. G.</given-names></string-name> and <string-name><surname>Scott</surname>, <given-names>J. G.</given-names></string-name> <article-title>Local shrinkage rules, lévy processes and regularized regression</article-title>. <source>Journal of the Royal Statistical Society: Series B (Statistical Methodology)</source> <volume>74</volume>(<issue>2</issue>) <fpage>287</fpage>–<lpage>311</lpage> (<year>2012</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1111/j.1467-9868.2011.01015.x" xlink:type="simple">https://doi.org/10.1111/j.1467-9868.2011.01015.x</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2899864">MR2899864</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_037">
<mixed-citation publication-type="other"><string-name><surname>Polson</surname>, <given-names>N. G.</given-names></string-name> and <string-name><surname>Scott</surname>, <given-names>J. G.</given-names></string-name> (<year>2014</year>). Vertical-likelihood monte carlo. <italic>arXiv preprint arXiv:</italic><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1409.3601"><italic>1409.3601</italic></ext-link>.</mixed-citation>
</ref>
<ref id="j_nejsds100_ref_038">
<mixed-citation publication-type="other"><string-name><surname>Raftery</surname>, <given-names>A. E.</given-names></string-name>, <string-name><surname>Newton</surname>, <given-names>M. A.</given-names></string-name>, <string-name><surname>Satagopan</surname>, <given-names>J. M.</given-names></string-name> and <string-name><surname>Krivitsky</surname>, <given-names>P. N.</given-names></string-name> Estimating the integrated likelihood via posterior simulation using the harmonic mean identity (<year>2006</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2433201">MR2433201</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_039">
<mixed-citation publication-type="journal"><string-name><surname>Ramakrishnan</surname>, <given-names>M.</given-names></string-name> <article-title>An alternative proof of the admissibility of the horvitz-thompson estimator</article-title>. <source>The Annals of Statistics</source> <volume>1</volume>(<issue>3</issue>) <fpage>577</fpage>–<lpage>579</lpage> (<year>1973</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0408061">MR0408061</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_040">
<mixed-citation publication-type="journal"><string-name><surname>Rao</surname>, <given-names>J. N.</given-names></string-name>, <string-name><surname>Chaudhuri</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Eltinge</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Fay</surname>, <given-names>R. E.</given-names></string-name>, <string-name><surname>Ghosh</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Ghosh</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Lahiri</surname>, <given-names>P.</given-names></string-name> and <string-name><surname>Pfeffermann</surname>, <given-names>D.</given-names></string-name> <article-title>Some current trends in sample survey theory and methods (with discussion)</article-title>. <source>Sankhy: The Indian Journal of Statistics, Series B</source> <fpage>1</fpage>–<lpage>57</lpage> (<year>1999</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1720726">MR1720726</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_041">
<mixed-citation publication-type="journal"><string-name><surname>Ritov</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Bickel</surname>, <given-names>P. J.</given-names></string-name>, <string-name><surname>Gamst</surname>, <given-names>A. C.</given-names></string-name> and <string-name><surname>Kleijn</surname>, <given-names>B. J. K.</given-names></string-name> <article-title>The bayesian analysis of complex, high-dimensional models: Can it be coda?</article-title> <source>Statistical Science</source> <volume>29</volume>(<issue>4</issue>) <fpage>619</fpage>–<lpage>639</lpage> (<year>2014</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/14-STS483" xlink:type="simple">https://doi.org/10.1214/14-STS483</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3300362">MR3300362</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_042">
<mixed-citation publication-type="journal"><string-name><surname>Robins</surname>, <given-names>J. M.</given-names></string-name> and <string-name><surname>Ritov</surname>, <given-names>Y.</given-names></string-name> <article-title>Toward a curse of dimensionality appropriate (coda) asymptotic theory for semi-parametric models</article-title>. <source>Statistics in medicine</source> <volume>16</volume>(<issue>3</issue>) <fpage>285</fpage>–<lpage>319</lpage> (<year>1997</year>).</mixed-citation>
</ref>
<ref id="j_nejsds100_ref_043">
<mixed-citation publication-type="journal"><string-name><surname>Rubin</surname>, <given-names>D. B.</given-names></string-name> <article-title>Estimating causal effects of treatments in randomized and nonrandomized studies</article-title>. <source>Journal of educational Psychology</source> <volume>66</volume>(<issue>5</issue>) <fpage>688</fpage>–<lpage>701</lpage> (<year>1974</year>).</mixed-citation>
</ref>
<ref id="j_nejsds100_ref_044">
<mixed-citation publication-type="book"><string-name><surname>Särndal</surname>, <given-names>C.-E.</given-names></string-name>, <string-name><surname>Swensson</surname>, <given-names>B.</given-names></string-name> and <string-name><surname>Wretman</surname>, <given-names>J.</given-names></string-name> <source>Model assisted survey sampling</source>. <publisher-name>Springer Science &amp; Business Media</publisher-name> (<year>2003</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/978-1-4612-4378-6" xlink:type="simple">https://doi.org/10.1007/978-1-4612-4378-6</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1140409">MR1140409</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_045">
<mixed-citation publication-type="journal"><string-name><surname>Shortreed</surname>, <given-names>S. M.</given-names></string-name> and <string-name><surname>Ertefaie</surname>, <given-names>A.</given-names></string-name> <article-title>Outcome-adaptive lasso: variable selection for causal inference</article-title>. <source>Biometrics</source> <volume>73</volume>(<issue>4</issue>) <fpage>1111</fpage>–<lpage>1122</lpage> (<year>2017</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1111/biom.12679" xlink:type="simple">https://doi.org/10.1111/biom.12679</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3744525">MR3744525</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_046">
<mixed-citation publication-type="other"><string-name><surname>Sims</surname>, <given-names>C.</given-names></string-name> <italic>Understanding non-bayesians</italic>. Unpublished chapter, Department of Economics, Princeton University (<year>2010</year>).</mixed-citation>
</ref>
<ref id="j_nejsds100_ref_047">
<mixed-citation publication-type="other"><string-name><surname>Sims</surname>, <given-names>C. A.</given-names></string-name> <italic>Thinking about instrumental variables</italic>. manuscript, Department of Economics, Princeton University (<year>2007</year>) manuscript.</mixed-citation>
</ref>
<ref id="j_nejsds100_ref_048">
<mixed-citation publication-type="other"><string-name><surname>Sims</surname>, <given-names>C. A.</given-names></string-name> On an example of larry wasserman, round 2 (<year>2012</year>).</mixed-citation>
</ref>
<ref id="j_nejsds100_ref_049">
<mixed-citation publication-type="journal"><string-name><surname>Skilling</surname>, <given-names>J.</given-names></string-name> <article-title>Nested sampling for general bayesian computation</article-title>. <source>Bayesian analysis</source> <volume>1</volume>(<issue>4</issue>) <fpage>833</fpage>–<lpage>859</lpage> (<year>2006</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/06-BA127" xlink:type="simple">https://doi.org/10.1214/06-BA127</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2282208">MR2282208</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_050">
<mixed-citation publication-type="journal"><string-name><surname>Stigler</surname>, <given-names>S. M.</given-names></string-name> <article-title>The 1988 neyman memorial lecture: a galtonian perspective on shrinkage estimators</article-title>. <source>Statistical Science</source> <fpage>147</fpage>–<lpage>155</lpage> (<year>1990</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1054859">MR1054859</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_051">
<mixed-citation publication-type="chapter"><string-name><surname>Trotter</surname>, <given-names>H. F.</given-names></string-name> and <string-name><surname>Tukey</surname>, <given-names>H.</given-names></string-name> <chapter-title>Conditional monte carlo for normal samples</chapter-title>. In <source>Proc. Symp. on Monte Carlo Methods</source> <fpage>64</fpage>–<lpage>79</lpage>. <publisher-name>John Wiley and Sons</publisher-name> (<year>1956</year>). <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0079825">MR0079825</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_052">
<mixed-citation publication-type="journal"><string-name><surname>Wang</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Dominici</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Parmigiani</surname>, <given-names>G.</given-names></string-name> and <string-name><surname>Zigler</surname>, <given-names>C. M.</given-names></string-name> <article-title>Accounting for uncertainty in confounder and effect modifier selection when estimating average causal effects in generalized linear models</article-title>. <source>Biometrics</source> <volume>71</volume>(<issue>3</issue>) <fpage>654</fpage>–<lpage>665</lpage> (<year>2015</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1111/biom.12315" xlink:type="simple">https://doi.org/10.1111/biom.12315</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3402601">MR3402601</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_053">
<mixed-citation publication-type="journal"><string-name><surname>Wang</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Parmigiani</surname>, <given-names>G.</given-names></string-name> and <string-name><surname>Dominici</surname>, <given-names>F.</given-names></string-name> <article-title>Bayesian effect estimation accounting for adjustment uncertainty</article-title>. <source>Biometrics</source> <volume>68</volume>(<issue>3</issue>) <fpage>661</fpage>–<lpage>671</lpage> (<year>2012</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1111/j.1541-0420.2011.01731.x" xlink:type="simple">https://doi.org/10.1111/j.1541-0420.2011.01731.x</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3055168">MR3055168</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_054">
<mixed-citation publication-type="chapter"><string-name><surname>Wasserman</surname>, <given-names>L.</given-names></string-name> <chapter-title>Bayesian inference</chapter-title>. In <source>All of Statistics</source> <fpage>175</fpage>–<lpage>192</lpage>. <publisher-name>Springer</publisher-name> (<year>2004</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/978-0-387-21736-9" xlink:type="simple">https://doi.org/10.1007/978-0-387-21736-9</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2055670">MR2055670</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds100_ref_055">
<mixed-citation publication-type="journal"><string-name><surname>Yakowitz</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Krimmel</surname>, <given-names>J.</given-names></string-name> and <string-name><surname>Szidarovszky</surname>, <given-names>F.</given-names></string-name> <article-title>Weighted monte carlo integration</article-title>. <source>SIAM Journal on Numerical Analysis</source> <volume>15</volume>(<issue>6</issue>) <fpage>1289</fpage>–<lpage>1300</lpage> (<year>1978</year>). <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1137/0715088" xlink:type="simple">https://doi.org/10.1137/0715088</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0512700">MR0512700</ext-link></mixed-citation>
</ref>
</ref-list>
</back>
</article>
