<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.0 20120330//EN" "JATS-journalpublishing1.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">NEJSDS</journal-id>
<journal-title-group><journal-title>The New England Journal of Statistics in Data Science</journal-title></journal-title-group>
<issn pub-type="ppub">2693-7166</issn><issn-l>2693-7166</issn-l>
<publisher>
<publisher-name>New England Statistical Society</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">NEJSDS14</article-id>
<article-id pub-id-type="doi">10.51387/22-NEJSDS14</article-id>
<article-categories>
<subj-group subj-group-type="heading"><subject>Methodology Article</subject></subj-group>
<subj-group subj-group-type="area"><subject>Statistical Methodology</subject></subj-group>
</article-categories>
<title-group>
<article-title>Effect of Model Space Priors on Statistical Inference with Model Uncertainty</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Porwal</surname><given-names>Anupreet</given-names></name><email xlink:href="mailto:porwaa@uw.edu">porwaa@uw.edu</email><xref ref-type="aff" rid="j_nejsds14_aff_001"/><xref ref-type="corresp" rid="cor1">∗</xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Raftery</surname><given-names>Adrian E.</given-names></name><email xlink:href="mailto:raftery@uw.edu">raftery@uw.edu</email><xref ref-type="aff" rid="j_nejsds14_aff_002"/>
</contrib>
<aff id="j_nejsds14_aff_001">Department of Statistics, <institution>University of Washington</institution>, Seattle, WA 98195, <country>USA</country>. E-mail address: <email xlink:href="mailto:porwaa@uw.edu">porwaa@uw.edu</email></aff>
<aff id="j_nejsds14_aff_002">Department of Statistics, <institution>University of Washington</institution>, Seattle, WA 98195, <country>USA</country>. E-mail address: <email xlink:href="mailto:raftery@uw.edu">raftery@uw.edu</email></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>∗</label>Corresponding author.</corresp>
</author-notes>
<pub-date pub-type="ppub"><year>2023</year></pub-date><pub-date pub-type="epub"><day>16</day><month>11</month><year>2022</year></pub-date><volume>1</volume><issue>2</issue><fpage>149</fpage><lpage>158</lpage><supplementary-material id="S1" content-type="document" xlink:href="nejsds14_s001.pdf" mimetype="application" mime-subtype="pdf">
<caption>
<title>Supplementary Material</title>
<p>The supplementary material contains detailed summary results for each metric and dataset used in the study. It also contains a summary of data-generating models for each of the datasets.</p>
</caption>
</supplementary-material><history><date date-type="accepted"><day>18</day><month>10</month><year>2022</year></date></history>
<permissions><copyright-statement>© 2023 New England Statistical Society</copyright-statement><copyright-year>2023</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>Open access article under the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/">CC BY</ext-link> license.</license-p></license></permissions>
<abstract>
<p>Bayesian model averaging (BMA) provides a coherent way to account for model uncertainty in statistical inference tasks. BMA requires specification of model space priors and parameter space priors. In this article we focus on comparing different model space priors in the presence of model uncertainty. We consider eight reference model space priors used in the literature and three adaptive parameter priors recommended by Porwal and Raftery [<xref ref-type="bibr" rid="j_nejsds14_ref_037">37</xref>]. We assess the performance of these combinations of prior specifications for variable selection in linear regression models for the statistical tasks of parameter estimation, interval estimation, inference, point and interval prediction. We carry out an extensive simulation study based on 14 real datasets representing a range of situations encountered in practice. We found that beta-binomial model space priors specified in terms of the prior probability of model size performed best on average across various statistical tasks and datasets, outperforming priors that were uniform across models. Recently proposed complexity priors performed relatively poorly.</p>
</abstract>
<kwd-group>
<label>Keywords and phrases</label>
<kwd>Bayesian model averaging</kwd>
<kwd>Zellner’s <italic>g</italic>-prior</kwd>
<kwd>Model space prior</kwd>
<kwd>Beta-Binomial prior</kwd>
<kwd>Complexity prior</kwd>
<kwd>Model selection</kwd>
<kwd>Prediction</kwd>
</kwd-group>
<funding-group><award-group><funding-source xlink:href="https://doi.org/10.13039/100000071">NICHD</funding-source><award-id>R01 HD-070936</award-id></award-group><award-group><funding-source xlink:href="https://doi.org/10.13039/100007812">University of Washington</funding-source></award-group><funding-statement>This research was supported by NICHD grant R01 HD-070936, and by the Boeing International Professorship at the University of Washington. </funding-statement></funding-group>
</article-meta>
</front>
<body>
<sec id="j_nejsds14_s_001">
<label>1</label>
<title>Introduction</title>
<p>Analysis of data in the presence of model uncertainty is a critical problem in statistical modeling applications. Accounting for model uncertainty, rather than selecting a single statistical model, improves predictive performance and robustness in estimation and inference of model parameters [<xref ref-type="bibr" rid="j_nejsds14_ref_040">40</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_037">37</xref>].</p>
<p>One common instance of model uncertainty is that of variable selection in linear regression model. Given an <italic>n</italic>-dimensional continuous response variable, <inline-formula id="j_nejsds14_ineq_001"><alternatives><mml:math>
<mml:mi mathvariant="bold-italic">Y</mml:mi></mml:math><tex-math><![CDATA[$\boldsymbol{Y}$]]></tex-math></alternatives></inline-formula>, and a set of <italic>p</italic> potential predictor variables <inline-formula id="j_nejsds14_ineq_002"><alternatives><mml:math>
<mml:mi mathvariant="bold-italic">X</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">∈</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="script">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>×</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[$\boldsymbol{X}=({X_{1}},\dots ,{X_{p}})\in {\mathcal{R}^{n\times p}}$]]></tex-math></alternatives></inline-formula>, the aim is to do statistical analysis of the data when it is not known in advance which of the <inline-formula id="j_nejsds14_ineq_003"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${2^{p}}$]]></tex-math></alternatives></inline-formula> possible models is most appropriate. Consider a binary indexing vector <inline-formula id="j_nejsds14_ineq_004"><alternatives><mml:math>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">γ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">γ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">γ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\boldsymbol{\gamma }=({\gamma _{1}},{\gamma _{2}},\dots ,{\gamma _{p}})$]]></tex-math></alternatives></inline-formula> for the model space which indicates which explanatory variables are part of a model <inline-formula id="j_nejsds14_ineq_005"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mathcal{M}_{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula>. Under <inline-formula id="j_nejsds14_ineq_006"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mathcal{M}_{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula>, the linear regression model can then be expressed as: 
<disp-formula id="j_nejsds14_eq_001">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mn mathvariant="bold">1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold-italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="bold-italic">ϵ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\mathcal{M}_{\boldsymbol{\gamma }}}:\boldsymbol{Y}={\mathbf{1}_{n}}\alpha +{X_{\boldsymbol{\gamma }}}{\boldsymbol{\beta }_{\boldsymbol{\gamma }}}+\boldsymbol{\epsilon },\]]]></tex-math></alternatives>
</disp-formula> 
where <inline-formula id="j_nejsds14_ineq_007"><alternatives><mml:math>
<mml:mi mathvariant="bold-italic">ϵ</mml:mi>
<mml:mo stretchy="false">∼</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn mathvariant="bold">0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">I</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\boldsymbol{\epsilon }\sim \mathcal{N}(\mathbf{0},{\sigma ^{2}}{I_{n}})$]]></tex-math></alternatives></inline-formula>, and <inline-formula id="j_nejsds14_ineq_008"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula> is a <inline-formula id="j_nejsds14_ineq_009"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>×</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$n\times {p_{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula> matrix where each column is centered around its mean and <inline-formula id="j_nejsds14_ineq_010"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula> denotes the number of explanatory variables in the model <inline-formula id="j_nejsds14_ineq_011"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mathcal{M}_{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula>.</p>
<p>The Bayesian framework provides a straightforward way to account for model uncertainty by treating the model as a parameter itself, using Bayesian model averaging (BMA) [<xref ref-type="bibr" rid="j_nejsds14_ref_029">29</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_038">38</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_020">20</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_034">34</xref>]. BMA requires the specification of a model space prior and a parameter space prior. However, subjective elicitation of these priors is often not feasible, particularly when <italic>p</italic> is large, motivating the use of default reference priors.</p>
<p>Several default parameter prior choices have been proposed in the last thirty years (see Porwal and Raftery [<xref ref-type="bibr" rid="j_nejsds14_ref_037">37</xref>] for an overview) and several other comparisons of these methods have been carried out [<xref ref-type="bibr" rid="j_nejsds14_ref_004">4</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_008">8</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_011">11</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_013">13</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_015">15</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_032">32</xref>]. However, similar comparisons of default model space priors remain limited. In this article, our focus is on understanding the effect of model space priors on the statistical tasks of parameter estimation, interval estimation, statistical inference, point and interval prediction.</p>
<p>We compare combinations of three default parameter priors with eight choices of model space priors that have been advocated in the literature. These model space priors correspond to different flavors of Bayesian inference with: (i) fixed hyper-parameter choices, (ii) with Bayesian treatment of hyper-parameters, and (iii) estimation of hyper-parameters in an empirical Bayes (EB) manner. The comparison is carried out over an extensive simulation study closely based on 14 real datasets that span a wide range of practical data analysis situations.</p>
<p>The article is organized as follows. Section <xref rid="j_nejsds14_s_002">2</xref> provides a brief review of BMA and discusses in detail the parameter and model prior choices considered in this article. We discuss the simulation design, metrics and datasets used for comparison, and the results in Section <xref rid="j_nejsds14_s_006">3</xref>, followed by discussion and concluding remarks in Section <xref rid="j_nejsds14_s_010">4</xref>.</p>
</sec>
<sec id="j_nejsds14_s_002">
<label>2</label>
<title>Bayesian Model Averaging: A Review</title>
<p>Bayesian model averaging [<xref ref-type="bibr" rid="j_nejsds14_ref_024">24</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_028">28</xref>] provides a formal way to account for model uncertainty in statistical inference. Several reviews of the BMA literature are available [<xref ref-type="bibr" rid="j_nejsds14_ref_028">28</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_025">25</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_013">13</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_050">50</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_007">7</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_015">15</xref>]. The basic idea of BMA is that it uses prior probabilities for the models considered, and Bayes’ theorem to deal with model uncertainty by calculating their posterior probabilities.</p>
<p>Assuming that there is one true model among the set of <inline-formula id="j_nejsds14_ineq_012"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${2^{p}}$]]></tex-math></alternatives></inline-formula> candidate models, the posterior probability of a model <inline-formula id="j_nejsds14_ineq_013"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mathcal{M}_{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula> is 
<disp-formula id="j_nejsds14_eq_002">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold">′</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold">′</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ P({\mathcal{M}_{\boldsymbol{\gamma }}}|\boldsymbol{Y})=\frac{P(\boldsymbol{Y}|{\mathcal{M}_{\boldsymbol{\gamma }}})P({\mathcal{M}_{\boldsymbol{\gamma }}})}{{\textstyle\sum _{{\boldsymbol{\gamma }^{\mathbf{\prime }}}}}P(\boldsymbol{Y}|{\mathcal{M}_{{\boldsymbol{\gamma }^{\prime }}}})P({\mathcal{M}_{{\boldsymbol{\gamma }^{\mathbf{\prime }}}}})},\]]]></tex-math></alternatives>
</disp-formula> 
where <inline-formula id="j_nejsds14_ineq_014"><alternatives><mml:math>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$P({\mathcal{M}_{\boldsymbol{\gamma }}})$]]></tex-math></alternatives></inline-formula> is the prior model probability of <inline-formula id="j_nejsds14_ineq_015"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mathcal{M}_{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula>, and <inline-formula id="j_nejsds14_ineq_016"><alternatives><mml:math>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$P(\boldsymbol{Y}|{\mathcal{M}_{\boldsymbol{\gamma }}})$]]></tex-math></alternatives></inline-formula> is the marginal likelihood of the model after integrating out parameters with respect to the prior <inline-formula id="j_nejsds14_ineq_017"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">π</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\pi _{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula>, namely: 
<disp-formula id="j_nejsds14_eq_003">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr class="split-mtr">
<mml:mtd class="split-mtd">
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∫</mml:mo></mml:mstyle>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mn mathvariant="bold">1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mo>+</mml:mo>
</mml:mtd>
<mml:mtd class="split-mtd">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold-italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">I</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr class="split-mtr">
<mml:mtd class="split-mtd"/>
<mml:mtd class="split-mtd">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">π</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold-italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">d</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold-italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="italic">d</mml:mi>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mi mathvariant="italic">d</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}P(\boldsymbol{Y}|{\mathcal{M}_{\boldsymbol{\gamma }}})=\int \mathcal{N}(\boldsymbol{Y}|{\mathbf{1}_{n}}\alpha +& {X_{\boldsymbol{\gamma }}}{\boldsymbol{\beta }_{\boldsymbol{\gamma }}},{\sigma ^{2}}{I_{n}})\\ {} & {\pi _{\boldsymbol{\gamma }}}({\boldsymbol{\beta }_{\boldsymbol{\gamma }}},\alpha ,{\sigma ^{2}})d{\boldsymbol{\beta }_{\boldsymbol{\gamma }}}d\alpha d{\sigma ^{2}}.\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>Under BMA inference, we can express the predictive distribution of a quantity of interest, Δ, such as a parameter or an observable future quantity, as a weighted average of its predictive distributions under the different candidate models: 
<disp-formula id="j_nejsds14_eq_004">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="normal">Δ</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="normal">Δ</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ P(\Delta |\boldsymbol{Y})=\sum \limits_{\boldsymbol{\gamma }}P(\Delta |{\mathcal{M}_{\boldsymbol{\gamma }}})P({\mathcal{M}_{\boldsymbol{\gamma }}}|\boldsymbol{Y}),\]]]></tex-math></alternatives>
</disp-formula> 
where the posterior model probabilities <inline-formula id="j_nejsds14_ineq_018"><alternatives><mml:math>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$P({\mathcal{M}_{\boldsymbol{\gamma }}}|\boldsymbol{Y})$]]></tex-math></alternatives></inline-formula> serve as averaging weights. In the case where Δ is a regression coefficient, the resulting posterior distribution, <inline-formula id="j_nejsds14_ineq_019"><alternatives><mml:math>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="normal">Δ</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$P(\Delta |\boldsymbol{Y})$]]></tex-math></alternatives></inline-formula>, is a mixture of a point mass at 0 and a continuous density.</p>
<p>BMA has several desirable theoretical properties [<xref ref-type="bibr" rid="j_nejsds14_ref_039">39</xref>]. When choosing between two models, one of which is nested within the other, choosing the one with the higher posterior probability minimizes the total error rate (sum of Type I and Type II error probabilities); BMA point estimators and predictions minimize mean squared error (MSE); BMA estimation and prediction intervals are calibrated; and BMA predictive distributions have optimal performance in the log score sense.</p>
<p>The next subsection discusses the choice of parameter and model space priors that need to be specified by the user when implementing BMA.</p>
<sec id="j_nejsds14_s_003">
<label>2.1</label>
<title>Choice of Parameter Priors</title>
<p>Despite the wide adoption of Bayesian methods in linear models, prior elicitation for linear models is still an open problem. The parameter prior distribution <inline-formula id="j_nejsds14_ineq_020"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">π</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\pi _{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula> can be expressed as 
<disp-formula id="j_nejsds14_eq_005">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">π</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold-italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">π</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold-italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">π</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\pi _{\boldsymbol{\gamma }}}({\boldsymbol{\beta }_{\boldsymbol{\gamma }}},\alpha ,{\sigma ^{2}})={\pi _{\boldsymbol{\gamma }}}({\boldsymbol{\beta }_{\boldsymbol{\gamma }}}|\alpha ,{\sigma ^{2}}){\pi _{\boldsymbol{\gamma }}}(\alpha ,{\sigma ^{2}}).\]]]></tex-math></alternatives>
</disp-formula> 
A standard Jeffreys’ prior is often used for the intercept and error variance, which are often common to all models considered, i.e. <inline-formula id="j_nejsds14_ineq_021"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">π</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">∝</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${\pi _{\boldsymbol{\gamma }}}(\alpha ,{\sigma ^{2}})\propto {\sigma ^{-2}}$]]></tex-math></alternatives></inline-formula>. One of the most popular priors used for the model parameters <inline-formula id="j_nejsds14_ineq_022"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold-italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\boldsymbol{\beta }_{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula> is Zellner’s g-prior [<xref ref-type="bibr" rid="j_nejsds14_ref_053">53</xref>], namely
<disp-formula id="j_nejsds14_eq_006">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold-italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∼</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn mathvariant="bold">0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\boldsymbol{\beta }_{\boldsymbol{\gamma }}}|{\sigma ^{2}},{\mathcal{M}_{\boldsymbol{\gamma }}}\sim \mathcal{N}(\mathbf{0},g{\sigma ^{2}}{({X_{\boldsymbol{\gamma }}^{T}}{X_{\boldsymbol{\gamma }}})^{-1}}).\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>This is popular because of its computational efficiency in evaluating marginal likelihoods and performing model search. It is also attractive because of its intuitive interpretation arising from analysis of a conceptual sample generated using the same design matrix <inline-formula id="j_nejsds14_ineq_023"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold-italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\boldsymbol{X}_{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula> as in the data at hand. It is a special case of spike-and-lab family with the slab density given by the Normal density above and the spike being a point mass at zero. Also, <italic>g</italic>-priors are appealing in variable selection problems since they require the user to specify only value (or hyper-prior) for the scalar hyper-parameter <italic>g</italic>. This controls the prior variance of the parameters; the effective prior sample size is <inline-formula id="j_nejsds14_ineq_024"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">g</mml:mi></mml:math><tex-math><![CDATA[$n/g$]]></tex-math></alternatives></inline-formula>. Several choices of <italic>g</italic> have been proposed [<xref ref-type="bibr" rid="j_nejsds14_ref_006">6</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_013">13</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_019">19</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_023">23</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_028">28</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_032">32</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_048">48</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_054">54</xref>].</p>
<p>Based on an extensive simulation study, Porwal and Raftery [<xref ref-type="bibr" rid="j_nejsds14_ref_037">37</xref>] found that in comparing parameter priors for BMA, three adaptive <italic>g</italic>-priors performed the best among many popular choices across the statistical tasks of parameter estimation, interval estimation, model inference, point prediction and interval prediction. In what follows, we shall focus only on these three parameter prior choices, namely: 
<list>
<list-item id="j_nejsds14_li_001">
<label>•</label>
<p><inline-formula id="j_nejsds14_ineq_025"><alternatives><mml:math>
<mml:mi mathvariant="bold-italic">g</mml:mi>
<mml:mi mathvariant="bold">=</mml:mi>
<mml:msqrt>
<mml:mrow>
<mml:mi mathvariant="bold-italic">n</mml:mi>
</mml:mrow>
</mml:msqrt></mml:math><tex-math><![CDATA[$\boldsymbol{g}\mathbf{=}\sqrt{\boldsymbol{n}}$]]></tex-math></alternatives></inline-formula>: First proposed by [<xref ref-type="bibr" rid="j_nejsds14_ref_013">13</xref>], it corresponds to a prior sample size equal to <inline-formula id="j_nejsds14_ineq_026"><alternatives><mml:math>
<mml:msqrt>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msqrt></mml:math><tex-math><![CDATA[$\sqrt{n}$]]></tex-math></alternatives></inline-formula> and has been found to work well in high dimensional settings [<xref ref-type="bibr" rid="j_nejsds14_ref_052">52</xref>]. The complexity penalty for a model using this specification is effectively half that in the BIC [<xref ref-type="bibr" rid="j_nejsds14_ref_037">37</xref>].</p>
</list-item>
<list-item id="j_nejsds14_li_002">
<label>•</label>
<p><bold>EB-local</bold>: An alternative to fixing <italic>g</italic> to a pre-specified value is to instead estimate it from the data in an empirical Bayes (EB) manner. The local EB approach estimates a different <italic>g</italic> for each model. Let <inline-formula id="j_nejsds14_ineq_027"><alternatives><mml:math>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$P(\boldsymbol{Y}|{\mathcal{M}_{\boldsymbol{\gamma }}},g)$]]></tex-math></alternatives></inline-formula> denotes the marginal likelihood of the data under a <italic>g</italic>-prior. Then 
<disp-formula id="j_nejsds14_eq_007">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">g</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">arg max</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo stretchy="false">≥</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:munder>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\hat{g}_{\boldsymbol{\gamma }}}=\underset{g\ge 0}{\operatorname{arg\,max}}P(\boldsymbol{Y}|{\mathcal{M}_{\boldsymbol{\gamma }}},g).\]]]></tex-math></alternatives>
</disp-formula> 
For a linear model, Hansen and Yu [<xref ref-type="bibr" rid="j_nejsds14_ref_023">23</xref>] showed that it reduces to <inline-formula id="j_nejsds14_ineq_028"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">g</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mo movablelimits="false">max</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[${\hat{g}_{\boldsymbol{\gamma }}}=\max \{{F_{\boldsymbol{\gamma }}}-1,0\}$]]></tex-math></alternatives></inline-formula> where <inline-formula id="j_nejsds14_ineq_029"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${F_{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula> is the F statistic for testing <inline-formula id="j_nejsds14_ineq_030"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold-italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn mathvariant="bold">0</mml:mn></mml:math><tex-math><![CDATA[${\boldsymbol{\beta }_{\boldsymbol{\gamma }}}=\mathbf{0}$]]></tex-math></alternatives></inline-formula>.</p>
</list-item>
<list-item id="j_nejsds14_li_003">
<label>•</label>
<p><bold>Hyper-</bold><italic>g</italic>: A natural Bayesian way to account for uncertainty about the scale parameter <italic>g</italic> is to specify a hyper-prior for <italic>g</italic> and perform fully Bayesian inference. Liang et al [<xref ref-type="bibr" rid="j_nejsds14_ref_032">32</xref>] proposed the hyper-<italic>g</italic> prior 
<disp-formula id="j_nejsds14_eq_008">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="italic">π</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \pi (g)=\frac{a-2}{2}{(1+g)^{-a/2}},\]]]></tex-math></alternatives>
</disp-formula>
</p>
</list-item>
</list> 
which is proper for <inline-formula id="j_nejsds14_ineq_031"><alternatives><mml:math>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$a\gt 2$]]></tex-math></alternatives></inline-formula>. Liang et al [<xref ref-type="bibr" rid="j_nejsds14_ref_032">32</xref>] recommended <inline-formula id="j_nejsds14_ineq_032"><alternatives><mml:math>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>3</mml:mn></mml:math><tex-math><![CDATA[$a=3$]]></tex-math></alternatives></inline-formula> as a default choice for the hyper-<italic>g</italic> prior. One advantage of using a hyper-<italic>g</italic> prior is that the posterior distribution of <italic>g</italic> given the model <inline-formula id="j_nejsds14_ineq_033"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mathcal{M}_{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula> is available in closed form, simplifying Bayesian inference.</p>
<p>In terms of theoretical properties, all three priors are model-selection consistent [<xref ref-type="bibr" rid="j_nejsds14_ref_032">32</xref>], except when the true model is the null model. This means that if the true model, denoted by <inline-formula id="j_nejsds14_ineq_034"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mathcal{M}_{{\boldsymbol{\gamma }_{T}}}}$]]></tex-math></alternatives></inline-formula> belongs to the model space, then the posterior probability of the true model, <inline-formula id="j_nejsds14_ineq_035"><alternatives><mml:math>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">→</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$P({\mathcal{M}_{\boldsymbol{\gamma }}}={\mathcal{M}_{{\boldsymbol{\gamma }_{T}}}}|\boldsymbol{y})\to 1$]]></tex-math></alternatives></inline-formula> as the sample size <inline-formula id="j_nejsds14_ineq_036"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi>∞</mml:mi></mml:math><tex-math><![CDATA[$n\to \infty $]]></tex-math></alternatives></inline-formula>. None of the above priors suffers from Bartlett’s paradox [<xref ref-type="bibr" rid="j_nejsds14_ref_001">1</xref>]. BMA with <inline-formula id="j_nejsds14_ineq_037"><alternatives><mml:math>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo>=</mml:mo>
<mml:msqrt>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msqrt></mml:math><tex-math><![CDATA[$g=\sqrt{n}$]]></tex-math></alternatives></inline-formula> is subject to the “information paradox”, but it has been argued that information consistency is of little practical importance in real data applications [<xref ref-type="bibr" rid="j_nejsds14_ref_037">37</xref>]. The EB-local and Hyper-<italic>g</italic> BMA methods are not subject to the information paradox.</p>
</sec>
<sec id="j_nejsds14_s_004">
<label>2.2</label>
<title>Choice of Model Space Priors</title>
<p>Model space priors require specification of the prior probabilities of all models <inline-formula id="j_nejsds14_ineq_038"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mathcal{M}_{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula>, indexed by the binary variable inclusion vector <inline-formula id="j_nejsds14_ineq_039"><alternatives><mml:math>
<mml:mi mathvariant="bold-italic">γ</mml:mi></mml:math><tex-math><![CDATA[$\boldsymbol{\gamma }$]]></tex-math></alternatives></inline-formula>. A common approach is to consider the inclusion of each variable as an independent and exchangeable Bernoulli random variable with a common prior probability of inclusion <italic>θ</italic>, i.e. 
<disp-formula id="j_nejsds14_eq_009">
<label>(2.1)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∏</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">γ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">γ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msup>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ p({\mathcal{M}_{\boldsymbol{\gamma }}}|\theta )={\prod \limits_{i=1}^{p}}{\theta ^{{\gamma _{i}}}}{(1-\theta )^{1-{\gamma _{i}}}}={\theta ^{{p_{\boldsymbol{\gamma }}}}}{(1-\theta )^{p-{p_{\boldsymbol{\gamma }}}}},\]]]></tex-math></alternatives>
</disp-formula> 
where <italic>θ</italic> is the prior expected fraction of the <inline-formula id="j_nejsds14_ineq_040"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">j</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\beta _{j}}$]]></tex-math></alternatives></inline-formula>’s that are not zero and <inline-formula id="j_nejsds14_ineq_041"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">γ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{\boldsymbol{\gamma }}}={\textstyle\sum _{i=1}^{p}}{\gamma _{i}}$]]></tex-math></alternatives></inline-formula> is the total number of covariates included in the model <inline-formula id="j_nejsds14_ineq_042"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mathcal{M}_{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula>.</p>
<p>In the absence of prior information, a common choice is to set <inline-formula id="j_nejsds14_ineq_043"><alternatives><mml:math>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0.5</mml:mn></mml:math><tex-math><![CDATA[$\theta =0.5$]]></tex-math></alternatives></inline-formula>. This induces a uniform prior over all models with <inline-formula id="j_nejsds14_ineq_044"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[$p({\mathcal{M}_{\boldsymbol{\gamma }}})={2^{-p}}$]]></tex-math></alternatives></inline-formula> for all <inline-formula id="j_nejsds14_ineq_045"><alternatives><mml:math>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[$\boldsymbol{\gamma }\in {\{0,1\}^{p}}$]]></tex-math></alternatives></inline-formula>, where <italic>p</italic> is the total number of covariates considered. The expected prior model size under the uniform model prior is <inline-formula id="j_nejsds14_ineq_046"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$p/2$]]></tex-math></alternatives></inline-formula>. However, choosing <inline-formula id="j_nejsds14_ineq_047"><alternatives><mml:math>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0.5</mml:mn></mml:math><tex-math><![CDATA[$\theta =0.5$]]></tex-math></alternatives></inline-formula> does not provide any multiplicity control [<xref ref-type="bibr" rid="j_nejsds14_ref_019">19</xref>].</p>
<p>For a fixed value of <italic>θ</italic>, the above prior induces a binomial prior for model size <inline-formula id="j_nejsds14_ineq_048"><alternatives><mml:math>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo>=</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">γ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$S={\textstyle\sum _{i=1}^{p}}{\gamma _{i}}$]]></tex-math></alternatives></inline-formula> i.e. <inline-formula id="j_nejsds14_ineq_049"><alternatives><mml:math>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo stretchy="false">∼</mml:mo>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$S\sim Bin(p,\theta )$]]></tex-math></alternatives></inline-formula>, with prior mean <inline-formula id="j_nejsds14_ineq_050"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mi mathvariant="italic">θ</mml:mi></mml:math><tex-math><![CDATA[$p\theta $]]></tex-math></alternatives></inline-formula> and prior variance <inline-formula id="j_nejsds14_ineq_051"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$p\theta (1-\theta )$]]></tex-math></alternatives></inline-formula>. Another way to specify <italic>θ</italic> is by using the researcher’s prior belief about expected model size <inline-formula id="j_nejsds14_ineq_052"><alternatives><mml:math>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$E[S]$]]></tex-math></alternatives></inline-formula>. Sala-i-Martin et al [<xref ref-type="bibr" rid="j_nejsds14_ref_046">46</xref>] (hereafter SDM) recommended a prior expected model size of 7 based on their growth regression analysis. Similar priors for expected model size have also been proposed elsewhere [<xref ref-type="bibr" rid="j_nejsds14_ref_045">45</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_030">30</xref>]. Hence, we can define an SDM version of the above prior with <inline-formula id="j_nejsds14_ineq_053"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mi mathvariant="italic">D</mml:mi>
<mml:mi mathvariant="italic">M</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>7</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi></mml:math><tex-math><![CDATA[${\theta _{SDM}}=7/p$]]></tex-math></alternatives></inline-formula>.</p>
<p>Any fixed choice of <italic>θ</italic> can lead to rather informative priors on model size <inline-formula id="j_nejsds14_ineq_054"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${p_{\boldsymbol{\gamma }}}$]]></tex-math></alternatives></inline-formula>. One way to address this issue is by estimating <italic>θ</italic> from the data using an empirical Bayes (EB) approach [<xref ref-type="bibr" rid="j_nejsds14_ref_019">19</xref>]. The EB approach involves maximizing the marginal likelihood of the data given <italic>θ</italic>: 
<disp-formula id="j_nejsds14_eq_010">
<label>(2.2)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">arg max</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo>
</mml:mrow>
</mml:munder>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\hat{\theta }_{EB}}=\underset{\theta \in [0,1]}{\operatorname{arg\,max}}\sum \limits_{\boldsymbol{\gamma }}P(\boldsymbol{Y}|{\mathcal{M}_{\boldsymbol{\gamma }}})P({\mathcal{M}_{\boldsymbol{\gamma }}}|\theta ).\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>However, maximization of (<xref rid="j_nejsds14_eq_010">2.2</xref>) can be computationally challenging, especially when <italic>p</italic> is large since the sum is over all models. Moreover, when <italic>p</italic> is large, marginal likelihood evaluation for all models is not feasible and the sum is approximated based on the models explored by MCMC. To optimize the marginal likelihood in (<xref rid="j_nejsds14_eq_010">2.2</xref>), we implement Algorithm <xref rid="j_nejsds14_fig_001">1</xref>, iterating between fitting the BMA approach to find likely models given <italic>θ</italic> and solving (<xref rid="j_nejsds14_eq_010">2.2</xref>) using the fitted models to find a new <italic>θ</italic>:</p>
<fig id="j_nejsds14_fig_001">
<label>Algorithm 1</label>
<caption>
<p>EB optimisation algorithm for <inline-formula id="j_nejsds14_ineq_055"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\hat{\theta }_{EB}}$]]></tex-math></alternatives></inline-formula>.</p>
</caption>
<graphic xlink:href="nejsds14_g001.jpg"/>
</fig>
<p>An alternative way to reduce the sensitivity of the posterior distribution to prior assumptions is to use hierarchical modeling and specify a weak hyper-prior for <italic>θ</italic>. One choice of such a hyper-prior is a Beta distribution, <inline-formula id="j_nejsds14_ineq_056"><alternatives><mml:math>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo stretchy="false">∼</mml:mo>
<mml:mtext>Beta</mml:mtext>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\theta \sim \text{Beta}(a,b)$]]></tex-math></alternatives></inline-formula>. Marginalizing out <italic>θ</italic> in (<xref rid="j_nejsds14_eq_009">2.1</xref>), gives 
<disp-formula id="j_nejsds14_eq_011">
<label>(2.3)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∫</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">d</mml:mi>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}P({\mathcal{M}_{\boldsymbol{\gamma }}}|a,b)& ={\int _{0}^{1}}p({\mathcal{M}_{\boldsymbol{\gamma }}}|\theta )p(\theta )d\theta \\ {} & =\frac{B({p_{\boldsymbol{\gamma }}}+a,p-{p_{\boldsymbol{\gamma }}}+b)}{B(a,b)},\end{aligned}\]]]></tex-math></alternatives>
</disp-formula> 
where <inline-formula id="j_nejsds14_ineq_057"><alternatives><mml:math>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">Γ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="normal">Γ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">Γ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>+</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle></mml:math><tex-math><![CDATA[$B({a^{\prime }},{b^{\prime }})=\frac{\Gamma ({a^{\prime }})\Gamma ({b^{\prime }})}{\Gamma ({a^{\prime }}+{b^{\prime }})}$]]></tex-math></alternatives></inline-formula> is the Beta function. It thus induces a Beta-Binomial<inline-formula id="j_nejsds14_ineq_058"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(a,b)$]]></tex-math></alternatives></inline-formula> prior on the model size <italic>S</italic> with probability mass function 
<disp-formula id="j_nejsds14_eq_012">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mfenced separators="" open="(" close=")">
<mml:mfrac linethickness="0">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mfenced>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {P_{S}}(s)=\left(\genfrac{}{}{0pt}{}{p}{s}\right)\frac{B(s+a,p-s+b)}{B(a,b)}.\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>Under a uniform prior on <italic>θ</italic>, i.e. when <inline-formula id="j_nejsds14_ineq_059"><alternatives><mml:math>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$a=b=1$]]></tex-math></alternatives></inline-formula>, (<xref rid="j_nejsds14_eq_011">2.3</xref>) simplifies to <inline-formula id="j_nejsds14_ineq_060"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:msup>
<mml:mrow>
<mml:mfenced separators="" open="(" close=")">
<mml:mfrac linethickness="0">
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[$p({\mathcal{M}_{\boldsymbol{\gamma }}})=\frac{1}{p+1}{\left(\genfrac{}{}{0pt}{}{p}{{p_{\boldsymbol{\gamma }}}}\right)^{-1}}$]]></tex-math></alternatives></inline-formula>. This is a combination of a uniform prior over model size with a uniform prior over the models of same size given the model size.</p>
<p>Under a Beta-Binomial (BB) prior, the prior expected model size is <inline-formula id="j_nejsds14_ineq_061"><alternatives><mml:math>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mi mathvariant="italic">p</mml:mi></mml:math><tex-math><![CDATA[$E[S]=\frac{a}{a+b}p$]]></tex-math></alternatives></inline-formula>. Similarly to a Bernoulli prior, we can elicit the prior in terms of the prior expected model size <inline-formula id="j_nejsds14_ineq_062"><alternatives><mml:math>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$E[S]$]]></tex-math></alternatives></inline-formula>. To facilitate prior elicitation, we fix <inline-formula id="j_nejsds14_ineq_063"><alternatives><mml:math>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$a=1$]]></tex-math></alternatives></inline-formula>. We can then define an SDM version of the BB prior (BB-SDM) with an expected prior model size, such as <inline-formula id="j_nejsds14_ineq_064"><alternatives><mml:math>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>7</mml:mn></mml:math><tex-math><![CDATA[$E[S]=7$]]></tex-math></alternatives></inline-formula>, by setting <inline-formula id="j_nejsds14_ineq_065"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mi mathvariant="italic">D</mml:mi>
<mml:mi mathvariant="italic">M</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[${b_{SDM}}=\frac{p}{E[S]}-1$]]></tex-math></alternatives></inline-formula>. Note that SDM themselves [<xref ref-type="bibr" rid="j_nejsds14_ref_046">46</xref>] did not use a Beta-Binomial prior on models, but only a Bernouilli prior.</p>
<fig id="j_nejsds14_fig_002">
<label>Figure 1</label>
<caption>
<p>Prior model size distribution for the Boston Housing and Nutrimouse datasets.</p>
</caption>
<graphic xlink:href="nejsds14_g002.jpg"/>
</fig>
<p>Alternatively, we can use an EB approach to learn <italic>b</italic> from the data. This can be done by maximizing the marginal likelihood given <italic>b</italic>, namely 
<disp-formula id="j_nejsds14_eq_013">
<label>(2.4)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false">arg max</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi>∞</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:munder>
<mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {\hat{b}_{EB}}=\underset{b\in (0,\infty )}{\operatorname{arg\,max}}\sum \limits_{\boldsymbol{\gamma }}P(\boldsymbol{Y}|{\mathcal{M}_{\boldsymbol{\gamma }}})P({\mathcal{M}_{\boldsymbol{\gamma }}}|a=1,b).\]]]></tex-math></alternatives>
</disp-formula> 
We find the optimal value, <inline-formula id="j_nejsds14_ineq_066"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\hat{b}_{EB}}$]]></tex-math></alternatives></inline-formula>, using Algorithm <xref rid="j_nejsds14_fig_003">2</xref>.</p>
<fig id="j_nejsds14_fig_003">
<label>Algorithm 2</label>
<caption>
<p>EB optimisation algorithm for <inline-formula id="j_nejsds14_ineq_067"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\hat{b}_{EB}}$]]></tex-math></alternatives></inline-formula>.</p>
</caption>
<graphic xlink:href="nejsds14_g003.jpg"/>
</fig>
<p>For Zellner’s <italic>g</italic>-prior, we require the model size to be no larger than the number of regression coefficients that can be identified from the data, so that <inline-formula id="j_nejsds14_ineq_068"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[${p_{\boldsymbol{\gamma }}}\lt n-2$]]></tex-math></alternatives></inline-formula>. Thus for higher dimensional datasets <inline-formula id="j_nejsds14_ineq_069"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(p\gt n)$]]></tex-math></alternatives></inline-formula>, we require that <inline-formula id="j_nejsds14_ineq_070"><alternatives><mml:math>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[$P({\mathcal{M}_{\boldsymbol{\gamma }}})=0$]]></tex-math></alternatives></inline-formula> for all models with model size greater than <inline-formula id="j_nejsds14_ineq_071"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$n-2$]]></tex-math></alternatives></inline-formula>. Hence, we use truncated versions of the priors defined in (<xref rid="j_nejsds14_eq_009">2.1</xref>) and (<xref rid="j_nejsds14_eq_011">2.3</xref>), namely <disp-formula-group id="j_nejsds14_dg_001">
<disp-formula id="j_nejsds14_eq_014">
<label>(2.5)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo stretchy="false">∝</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msup>
<mml:mn mathvariant="double-struck">1</mml:mn>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}p({\mathcal{M}_{\boldsymbol{\gamma }}}|\theta )& \propto {\theta ^{{p_{\boldsymbol{\gamma }}}}}{(1-\theta )^{p-{p_{\boldsymbol{\gamma }}}}}\mathbb{1}\{{p_{\boldsymbol{\gamma }}}\lt n-2\},\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
<disp-formula id="j_nejsds14_eq_015">
<label>(2.6)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
<mml:mtd class="align-even">
<mml:mo stretchy="false">∝</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">B</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mn mathvariant="double-struck">1</mml:mn>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}P({\mathcal{M}_{\boldsymbol{\gamma }}}|a,b)& \propto \frac{B({p_{\boldsymbol{\gamma }}}+a,p-{p_{\boldsymbol{\gamma }}}+b)}{B(a,b)}\mathbb{1}\{{p_{\boldsymbol{\gamma }}}\lt n-2\}.\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
</disp-formula-group></p>
<p>Castillo et al [<xref ref-type="bibr" rid="j_nejsds14_ref_003">3</xref>] introduced complexity priors, also known in the literature as diffusing [<xref ref-type="bibr" rid="j_nejsds14_ref_035">35</xref>] or power priors [<xref ref-type="bibr" rid="j_nejsds14_ref_005">5</xref>]. Here the marginal probability of inclusion of any variable decays at the rate <inline-formula id="j_nejsds14_ineq_072"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">κ</mml:mi>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${p^{-\kappa }}$]]></tex-math></alternatives></inline-formula> for some <inline-formula id="j_nejsds14_ineq_073"><alternatives><mml:math>
<mml:mi mathvariant="italic">κ</mml:mi>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[$\kappa \gt 0$]]></tex-math></alternatives></inline-formula>, where <italic>p</italic> is the total number of possible covariates. This specifies a vanishing prior probability of large models and leads to a faster rate of rejection of spurious parameters, at the cost of slower rates of detection of active parameters [<xref ref-type="bibr" rid="j_nejsds14_ref_044">44</xref>]. Similar priors have also been used elsewhere [<xref ref-type="bibr" rid="j_nejsds14_ref_051">51</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_035">35</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_042">42</xref>].</p>
<p>The complexity prior is defined as 
<disp-formula id="j_nejsds14_eq_016">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">∝</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">κ</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mn mathvariant="double-struck">1</mml:mn>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:mo stretchy="false">≤</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ p({\mathcal{M}_{\boldsymbol{\gamma }}})\propto {p^{-\kappa |\boldsymbol{\gamma }|}}\mathbb{1}\{|\boldsymbol{\gamma }|\le {s_{0}}\},\]]]></tex-math></alternatives>
</disp-formula> 
where <inline-formula id="j_nejsds14_ineq_074"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${s_{0}}$]]></tex-math></alternatives></inline-formula> is a pre-specified integer specifying the maximum number of important covariates and <inline-formula id="j_nejsds14_ineq_075"><alternatives><mml:math>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="bold-italic">γ</mml:mi>
<mml:mo stretchy="false">|</mml:mo></mml:math><tex-math><![CDATA[$|\boldsymbol{\gamma }|$]]></tex-math></alternatives></inline-formula> is the model size. In the absence of external information, we set <inline-formula id="j_nejsds14_ineq_076"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mo movablelimits="false">min</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[${s_{0}}=\min \{n-2,p\}$]]></tex-math></alternatives></inline-formula>. This prior is implemented in the BAS package as <monospace>tr.power.prior(kappa,trunc)</monospace>. We implement Complexity priors with <inline-formula id="j_nejsds14_ineq_077"><alternatives><mml:math>
<mml:mi mathvariant="italic">κ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$\kappa =1$]]></tex-math></alternatives></inline-formula> [<xref ref-type="bibr" rid="j_nejsds14_ref_043">43</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_044">44</xref>], and with <inline-formula id="j_nejsds14_ineq_078"><alternatives><mml:math>
<mml:mi mathvariant="italic">κ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$\kappa =2$]]></tex-math></alternatives></inline-formula> which is the default choice in the <monospace>BAS</monospace> package.</p>
</sec>
<sec id="j_nejsds14_s_005">
<label>2.3</label>
<title>Model Space Priors – A Graphical Illustration</title>
<p>To illustrate the effect of different model space priors, we use two datasets from our analysis: Boston Housing <inline-formula id="j_nejsds14_ineq_079"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>506</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>103</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(n=506,p=103)$]]></tex-math></alternatives></inline-formula> and Nutrimouse <inline-formula id="j_nejsds14_ineq_080"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>40</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>120</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(n=40,p=120)$]]></tex-math></alternatives></inline-formula> (Figure <xref rid="j_nejsds14_fig_002">1</xref>). The solid lines show the independent Bernoulli prior from (<xref rid="j_nejsds14_eq_009">2.1</xref>), while the dashed lines represent the Beta-Binomial prior in (<xref rid="j_nejsds14_eq_011">2.3</xref>) and the dash-dotted lines illustrate Complexity priors. For the Nutrimouse dataset <inline-formula id="j_nejsds14_ineq_081"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(p\gt n)$]]></tex-math></alternatives></inline-formula>, we use the truncated versions as discussed above. The colors group different flavors of methods: (i) Uniform versions with <inline-formula id="j_nejsds14_ineq_082"><alternatives><mml:math>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0.5</mml:mn></mml:math><tex-math><![CDATA[$\theta =0.5$]]></tex-math></alternatives></inline-formula> or <inline-formula id="j_nejsds14_ineq_083"><alternatives><mml:math>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$b=1$]]></tex-math></alternatives></inline-formula> (blue), (ii) SDM versions with expected prior model size 7 (red), (iii) EB versions with <italic>θ</italic> or <italic>b</italic> learned from the data (green), (iv) Complexity prior with <inline-formula id="j_nejsds14_ineq_084"><alternatives><mml:math>
<mml:mi mathvariant="italic">κ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$\kappa =1$]]></tex-math></alternatives></inline-formula> (orange), and (v) Complexity prior with <inline-formula id="j_nejsds14_ineq_085"><alternatives><mml:math>
<mml:mi mathvariant="italic">κ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$\kappa =2$]]></tex-math></alternatives></inline-formula> (brown).</p>
<p>The Bernoulli model space priors are very concentrated around their mean, <inline-formula id="j_nejsds14_ineq_086"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mi mathvariant="italic">θ</mml:mi></mml:math><tex-math><![CDATA[$p\theta $]]></tex-math></alternatives></inline-formula>. The complexity priors are concentrated around smaller model sizes with a mode at 0. The BB priors, on the other hand, are more diffuse, implying more prior uncertainty about model size. For the Nutrimouse dataset, all the model space priors assign zero probability to any model with size greater than <inline-formula id="j_nejsds14_ineq_087"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>38</mml:mn></mml:math><tex-math><![CDATA[$(n-2)=38$]]></tex-math></alternatives></inline-formula>. Among the Bernoulli versions, <inline-formula id="j_nejsds14_ineq_088"><alternatives><mml:math>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0.5</mml:mn></mml:math><tex-math><![CDATA[$\theta =0.5$]]></tex-math></alternatives></inline-formula> implies a prior mode around <inline-formula id="j_nejsds14_ineq_089"><alternatives><mml:math>
<mml:mo movablelimits="false">min</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$\min \{p/2,n-2\}$]]></tex-math></alternatives></inline-formula> while <inline-formula id="j_nejsds14_ineq_090"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mi mathvariant="italic">D</mml:mi>
<mml:mi mathvariant="italic">M</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{SDM}}$]]></tex-math></alternatives></inline-formula> has a prior model size of 7 (the same as the prior mean). The prior model size distribution induced by the Ber<inline-formula id="j_nejsds14_ineq_091"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({\theta _{EB}})$]]></tex-math></alternatives></inline-formula> prior adapts based on the data, with a prior mode between <inline-formula id="j_nejsds14_ineq_092"><alternatives><mml:math>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0.5</mml:mn></mml:math><tex-math><![CDATA[$\theta =0.5$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds14_ineq_093"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mi mathvariant="italic">D</mml:mi>
<mml:mi mathvariant="italic">M</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{SDM}}$]]></tex-math></alternatives></inline-formula> for the Boston Housing dataset, while having the lowest prior mode among the Bernoulli priors considered for the Nutrimouse dataset. The BB<inline-formula id="j_nejsds14_ineq_094"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(1,1)$]]></tex-math></alternatives></inline-formula> prior corresponds to a uniform prior over model size. The BB<inline-formula id="j_nejsds14_ineq_095"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mi mathvariant="italic">D</mml:mi>
<mml:mi mathvariant="italic">M</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(1,{\theta _{SDM}})$]]></tex-math></alternatives></inline-formula> and BB<inline-formula id="j_nejsds14_ineq_096"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(1,{\theta _{EB}})$]]></tex-math></alternatives></inline-formula> priors both induce a model size distribution with prior mode at zero.</p>
<table-wrap id="j_nejsds14_tab_001">
<label>Table 1</label>
<caption>
<p>Summary of prior moments of model size <italic>S</italic> under different model space priors and <monospace>BAS</monospace> code to implement them.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: center; border-top: double; border-bottom: solid thin">Model prior</td>
<td style="vertical-align: top; text-align: center; border-top: double; border-bottom: solid thin">E[S]</td>
<td style="vertical-align: top; text-align: center; border-top: double; border-bottom: solid thin">Var(S)</td>
<td style="vertical-align: top; text-align: center; border-top: double; border-bottom: solid thin"><monospace>BAS code</monospace></td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: center">Bernoulli<inline-formula id="j_nejsds14_ineq_097"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\theta )$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: center"><inline-formula id="j_nejsds14_ineq_098"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mi mathvariant="italic">θ</mml:mi></mml:math><tex-math><![CDATA[$p\theta $]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: center"><inline-formula id="j_nejsds14_ineq_099"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$p\theta (1-\theta )$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: center"><monospace>bas.lm(</monospace><inline-formula id="j_nejsds14_ineq_100"><alternatives><mml:math>
<mml:mo>…</mml:mo>
<mml:mspace width="0.1667em"/></mml:math><tex-math><![CDATA[$\dots \hspace{0.1667em}$]]></tex-math></alternatives></inline-formula><monospace>, model.prior=bernoulli(probs))</monospace></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center">Beta-Binomial<inline-formula id="j_nejsds14_ineq_101"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(a,b)$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: center"><inline-formula id="j_nejsds14_ineq_102"><alternatives><mml:math><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mi mathvariant="italic">a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle></mml:math><tex-math><![CDATA[$\frac{pa}{a+b}$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: center"><inline-formula id="j_nejsds14_ineq_103"><alternatives><mml:math><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle></mml:math><tex-math><![CDATA[$\frac{pab(a+b+p)}{{(a+b)^{2}}(a+b+1)}$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: center"><monospace>bas.lm(</monospace><inline-formula id="j_nejsds14_ineq_104"><alternatives><mml:math>
<mml:mo>…</mml:mo>
<mml:mspace width="0.1667em"/></mml:math><tex-math><![CDATA[$\dots \hspace{0.1667em}$]]></tex-math></alternatives></inline-formula><monospace>, model.prior=beta.binomial(alpha,beta))</monospace></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin">Complexity<inline-formula id="j_nejsds14_ineq_105"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">κ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\kappa )$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin">–</td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin">–</td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin"><monospace>bas.lm(</monospace><inline-formula id="j_nejsds14_ineq_106"><alternatives><mml:math>
<mml:mo>…</mml:mo>
<mml:mspace width="0.1667em"/></mml:math><tex-math><![CDATA[$\dots \hspace{0.1667em}$]]></tex-math></alternatives></inline-formula><monospace>, model.prior=tr.power.prior(kappa,trunc))</monospace></td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="j_nejsds14_s_006">
<label>3</label>
<title>Numerical Comparison</title>
<p>We investigate the performance of different model space priors and parameter prior combinations using an extensive simulation study based closely on real datasets. We evaluate the effect of prior choices for the statistical tasks of parameter point and interval estimation, inference, point and interval prediction, and computation time.</p>
<p>All the parameter and model space prior combinations were implemented using the <monospace>BAS</monospace> R package [<xref ref-type="bibr" rid="j_nejsds14_ref_005">5</xref>] with skeleton code shown in Table <xref rid="j_nejsds14_tab_001">1</xref>. A combination of the MC<sup>3</sup> Metropolis-Hastings algorithm for sampling from the posterior distribution of models [<xref ref-type="bibr" rid="j_nejsds14_ref_040">40</xref>], along with a random swap between a currently included and a currently excluded variable is used for model space exploration. This is implemented by setting the option <monospace>method=”MCMC”</monospace> in the <monospace>bas.lm()</monospace> function. We used a default of 10,000 MCMC iterations for all methods.</p>
<p>For the EB methods, we used Algorithm <xref rid="j_nejsds14_fig_001">1</xref> (or <xref rid="j_nejsds14_fig_003">2</xref>) to find the optimal <inline-formula id="j_nejsds14_ineq_107"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\theta _{EB}}$]]></tex-math></alternatives></inline-formula> (or <inline-formula id="j_nejsds14_ineq_108"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${b_{EB}}$]]></tex-math></alternatives></inline-formula>) before fitting a BAS model with the estimated hyperparameter value. For higher dimensional datasets <inline-formula id="j_nejsds14_ineq_109"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(p\gt n)$]]></tex-math></alternatives></inline-formula>, a truncated version of the Beta-Binomial prior (<xref rid="j_nejsds14_eq_015">2.6</xref>) was implemented by setting the option <monospace>model.prior=”tr.beta.binomial(alpha,beta,trunc=n-2)”</monospace> in BAS. Similarly, a truncated version of complexity prior is implemented in the BAS package. A truncated version of the Bernoulli prior (<xref rid="j_nejsds14_eq_014">2.5</xref>) is not available in BAS. We implemented it by (i) implementing <monospace>bas.lm()</monospace> with <monospace>tr.beta.binomial(1,1,trunc=n-2)</monospace>, and then (ii) using importance sampling to calculate updated posterior model probabilities with weights proportional to the ratio of the prior model space densities in (<xref rid="j_nejsds14_eq_014">2.5</xref>) and (<xref rid="j_nejsds14_eq_015">2.6</xref>).</p>
<sec id="j_nejsds14_s_007">
<label>3.1</label>
<title>Datasets</title>
<p>We based our analysis on 14 publicly available datasets, of which six are available from <ext-link ext-link-type="uri" xlink:href="https://archive.ics.uci.edu/ml/index.php">UCI machine learning repository</ext-link> and the others are examples available in the literature. The sample size and number of candidate variables along with the data sources for the different datasets are listed in Table <xref rid="j_nejsds14_tab_002">2</xref>. These include the classical statistical setting with <inline-formula id="j_nejsds14_ineq_110"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi></mml:math><tex-math><![CDATA[$n\gt p$]]></tex-math></alternatives></inline-formula>, high dimensional datasets with <inline-formula id="j_nejsds14_ineq_111"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi></mml:math><tex-math><![CDATA[$p\gt n$]]></tex-math></alternatives></inline-formula>, and intermediate settings where <inline-formula id="j_nejsds14_ineq_112"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo stretchy="false">≈</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi></mml:math><tex-math><![CDATA[$n\approx p$]]></tex-math></alternatives></inline-formula>. For each dataset, continuous covariates are standardized to have zero mean and variance 1 and the response variable is centered to have zero mean. Details of the choice of datasets and additional pre-processing can be found in [<xref ref-type="bibr" rid="j_nejsds14_ref_037">37</xref>].</p>
<table-wrap id="j_nejsds14_tab_002">
<label>Table 2</label>
<caption>
<p>Datasets used in the study.</p>
</caption> 
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: center; border-top: double; border-bottom: solid thin; border-left: solid thin; border-right: solid thin">Dataset Name</td>
<td style="vertical-align: top; text-align: center; border-top: double; border-bottom: solid thin; border-right: solid thin">Sample size (N)</td>
<td style="vertical-align: top; text-align: center; border-top: double; border-bottom: solid thin; border-right: solid thin">Covariates (p)</td>
<td style="vertical-align: top; text-align: center; border-top: double; border-bottom: solid thin; border-right: solid thin">Source</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: center; border-left: solid thin; border-right: solid thin">College</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">777</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">14</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin"><monospace>ISLR</monospace> [<xref ref-type="bibr" rid="j_nejsds14_ref_027">27</xref>]</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-left: solid thin; border-right: solid thin">Bias Correction-Tmax</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">7590</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">21</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">UCI ML repository</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-left: solid thin; border-right: solid thin">Bias Correction-Tmin</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">7590</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">21</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">UCI ML repository</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-left: solid thin; border-right: solid thin">SML2010</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">1373</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">22</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">UCI ML repository</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-left: solid thin; border-right: solid thin">Bike sharing-daily</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">731</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">28</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">UCI ML repository</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-left: solid thin; border-right: solid thin">Bike sharing-hourly</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">17379</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">32</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">UCI ML repository</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-left: solid thin; border-right: solid thin">Superconductivity</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">21263</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">81</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">UCI ML repository</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-left: solid thin; border-right: solid thin">Diabetes</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">442</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">64</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin"><monospace>spikeslab</monospace> [<xref ref-type="bibr" rid="j_nejsds14_ref_026">26</xref>]</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-left: solid thin; border-right: solid thin">Ozone</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">330</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">44</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin"><monospace>gss</monospace> [<xref ref-type="bibr" rid="j_nejsds14_ref_022">22</xref>]</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-left: solid thin; border-right: solid thin">Boston housing</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">506</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">103</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin"><monospace>mlbench</monospace> [<xref ref-type="bibr" rid="j_nejsds14_ref_036">36</xref>]</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-left: solid thin; border-right: solid thin">NIR</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">166</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">225</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin"><monospace>chemometrics</monospace> [<xref ref-type="bibr" rid="j_nejsds14_ref_014">14</xref>]</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-left: solid thin; border-right: solid thin">Nutrimouse</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">40</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">120</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin"><monospace>mixOmics</monospace> [<xref ref-type="bibr" rid="j_nejsds14_ref_041">41</xref>]</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-left: solid thin; border-right: solid thin">Multidrug</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">60</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin">853</td>
<td style="vertical-align: top; text-align: center; border-right: solid thin"><monospace>mixOmics</monospace> [<xref ref-type="bibr" rid="j_nejsds14_ref_041">41</xref>]</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin; border-left: solid thin; border-right: solid thin">Liver toxicity</td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin; border-right: solid thin">64</td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin; border-right: solid thin">3116</td>
<td style="vertical-align: top; text-align: center; border-bottom: solid thin; border-right: solid thin"><monospace>mixOmics</monospace> [<xref ref-type="bibr" rid="j_nejsds14_ref_041">41</xref>]</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="j_nejsds14_s_008">
<label>3.2</label>
<title>Simulation Design</title>
<p>For each dataset, we selected a data generating model that closely approximates the real dataset. We carry out all-subsets regression for datasets with <inline-formula id="j_nejsds14_ineq_113"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mn>30</mml:mn></mml:math><tex-math><![CDATA[$p\lt 30$]]></tex-math></alternatives></inline-formula> using the <monospace>leaps</monospace> package [<xref ref-type="bibr" rid="j_nejsds14_ref_033">33</xref>] in R. We then selected the largest model with all variables significant at the 0.05 level. For datasets with <inline-formula id="j_nejsds14_ineq_114"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mn>30</mml:mn></mml:math><tex-math><![CDATA[$p\gt 30$]]></tex-math></alternatives></inline-formula>, all subsets regression is not computationally feasible. For these datasets, we obtained a filtered list of variables using iterative sure independence screening [<xref ref-type="bibr" rid="j_nejsds14_ref_012">12</xref>]. If the filtered list contained more than 30 variables, we selected the top 30 variables with the highest <inline-formula id="j_nejsds14_ineq_115"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${R^{2}}$]]></tex-math></alternatives></inline-formula> values under univariate regression. All-subsets regression was then applied to the filtered list of covariates to obtain the data generating model for our study. A summary of the data generating model used for each dataset can be found in the supplementary materials.</p>
<p>We used the data generating model and parametric bootstrapping to generate 100 bootstrapped datasets with the same design matrix <inline-formula id="j_nejsds14_ineq_116"><alternatives><mml:math>
<mml:mi mathvariant="bold-italic">X</mml:mi></mml:math><tex-math><![CDATA[$\boldsymbol{X}$]]></tex-math></alternatives></inline-formula> and different simulated response vectors <inline-formula id="j_nejsds14_ineq_117"><alternatives><mml:math>
<mml:mi mathvariant="bold-italic">Y</mml:mi></mml:math><tex-math><![CDATA[$\boldsymbol{Y}$]]></tex-math></alternatives></inline-formula>. Each of the resulting simulated datasets had the same design matrix and error distribution as the real dataset on it was based.</p>
<p>We compared the performance of different parameter and model space prior combinations on these simulated datasets using the following metrics:</p>
<list>
<list-item id="j_nejsds14_li_004">
<label>•</label>
<p><bold>PointEst</bold>: We use root mean squared error (RMSE) as a metric to evaluate the parameter point estimation performance of different combinations. RMSE is evaluated as 
<disp-formula id="j_nejsds14_eq_017">
<label>(3.1)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mi mathvariant="italic">R</mml:mi>
<mml:mi mathvariant="italic">M</mml:mi>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mo>=</mml:mo>
<mml:msqrt>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">D</mml:mi>
<mml:mi mathvariant="italic">G</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msqrt>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ RMSE=\sqrt{\frac{1}{p}{\sum \limits_{i=1}^{p}}{({\beta _{i,DG}}-{\hat{\beta }_{i}})^{2}}},\]]]></tex-math></alternatives>
</disp-formula> 
where <inline-formula id="j_nejsds14_ineq_118"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">D</mml:mi>
<mml:mi mathvariant="italic">G</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi></mml:math><tex-math><![CDATA[${\beta _{i,DG}},i=1,\dots ,p$]]></tex-math></alternatives></inline-formula> denote the coefficients in the data generating model, and <inline-formula id="j_nejsds14_ineq_119"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi></mml:math><tex-math><![CDATA[${\hat{\beta }_{i}},i=1,\dots ,p$]]></tex-math></alternatives></inline-formula> denote the posterior means of the coefficients.</p>
</list-item>
<list-item id="j_nejsds14_li_005">
<label>•</label>
<p><bold>IntEst</bold>: The interval score (IS) [<xref ref-type="bibr" rid="j_nejsds14_ref_021">21</xref>] evaluates the performance of interval estimators in terms of both their coverage and width. It is the sum of two terms, the first of which rewards narrow intervals while the second rewards accurate coverage. For a variable <italic>z</italic>, the IS is 
<disp-formula id="j_nejsds14_eq_018">
<label>(3.2)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr class="split-mtr">
<mml:mtd class="split-mtd">
<mml:mi mathvariant="italic">I</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">u</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">u</mml:mi>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>+</mml:mo>
</mml:mtd>
<mml:mtd class="split-mtd">
<mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mn mathvariant="double-struck">1</mml:mn>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mi mathvariant="italic">l</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr class="split-mtr">
<mml:mtd class="split-mtd"/>
<mml:mtd class="split-mtd">
<mml:mo>+</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">u</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mn mathvariant="double-struck">1</mml:mn>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:mi mathvariant="italic">u</mml:mi>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \begin{aligned}{}I{S_{\alpha }}(l,u,z)=(u-l)+& \frac{2}{\alpha }(l-z)\mathbb{1}\{z\lt l\}\\ {} & +\frac{2}{\alpha }(z-u)\mathbb{1}\{u\lt z\},\end{aligned}\]]]></tex-math></alternatives>
</disp-formula> 
where <italic>l</italic> and <italic>u</italic> denote the upper and lower bounds of the <inline-formula id="j_nejsds14_ineq_120"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>×</mml:mo>
<mml:mn>100</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$(1-\alpha )\times 100\% $]]></tex-math></alternatives></inline-formula> posterior interval of <italic>z</italic>. The Mean Interval Score (MIS) is the average of the IS values for the quantities being estimated. In order to assess the quality of the interval estimation, we compute the Mean Interval score (MIS) for the coefficients and calculate the average MIS across coefficients for each of the datasets. We use <inline-formula id="j_nejsds14_ineq_121"><alternatives><mml:math>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0.05</mml:mn></mml:math><tex-math><![CDATA[$\alpha =0.05$]]></tex-math></alternatives></inline-formula>.</p>
</list-item>
<list-item id="j_nejsds14_li_006">
<label>•</label>
<p><bold>Inference</bold>: We calculate the area under the precision recall curve (AUPRC) using the posterior inclusion probabilities of the covariates to evaluate the model selection performance of different combinations of priors. We assess the quality of the resulting inference using (<inline-formula id="j_nejsds14_ineq_122"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo></mml:math><tex-math><![CDATA[$1-$]]></tex-math></alternatives></inline-formula>AUPRC) as our metric, where a lower value is better.</p>
</list-item>
</list>
<p>We also compared methods based on their out-of-sample predictive performance. We divided each dataset into 100 random 75%–25% train-test splits. We trained the methods on the training data and used the test data to assess the predictive performance using the metrics described below:</p>
<list>
<list-item id="j_nejsds14_li_007">
<label>•</label>
<p><bold>Prediction</bold>: We calculate <inline-formula id="j_nejsds14_ineq_123"><alternatives><mml:math>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mi mathvariant="italic">e</mml:mi>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup></mml:math><tex-math><![CDATA[${R_{test}^{2}}$]]></tex-math></alternatives></inline-formula> to evaluate accuracy of point predictions as follows: 
<disp-formula id="j_nejsds14_eq_019">
<label>(3.3)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mi mathvariant="italic">e</mml:mi>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mi mathvariant="italic">e</mml:mi>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>−</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∑</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mi mathvariant="italic">e</mml:mi>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mi mathvariant="italic">r</mml:mi>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {R_{test}^{2}}=1-\frac{{\textstyle\sum _{i\in test}}{({y_{i}}-\hat{{y_{i}}})^{2}}}{{\textstyle\sum _{i\in test}}{({y_{i}}-{\bar{y}_{train}})^{2}}},\]]]></tex-math></alternatives>
</disp-formula> 
where <inline-formula id="j_nejsds14_ineq_124"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mi mathvariant="italic">e</mml:mi>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo fence="true" stretchy="false">}</mml:mo></mml:math><tex-math><![CDATA[$\{{y_{i}}:i\in test\}$]]></tex-math></alternatives></inline-formula> denotes the response variable of the test set, <inline-formula id="j_nejsds14_ineq_125"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\hat{y}_{i}}$]]></tex-math></alternatives></inline-formula> denotes the corresponding predictions, and <inline-formula id="j_nejsds14_ineq_126"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mi mathvariant="italic">r</mml:mi>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\bar{y}_{train}}$]]></tex-math></alternatives></inline-formula> denotes the mean of the response variable in the training set.</p>
</list-item>
<list-item id="j_nejsds14_li_008">
<label>•</label>
<p><bold>IntPred</bold>: To assess the quality of the prediction intervals, we calculate the interval score using (<xref rid="j_nejsds14_eq_018">3.2</xref>) for each of the test set observations. Here, <italic>l</italic> and <italic>u</italic> represent the lower and upper bounds of the <inline-formula id="j_nejsds14_ineq_127"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>×</mml:mo>
<mml:mn>100</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$(1-\alpha )\times 100\% $]]></tex-math></alternatives></inline-formula> posterior predictive interval for the test observation. We calculate the mean interval score (MIS), averaging IS over test set observations for each of the train-test splits. A lower MIS corresponds to a better interval forecast.</p>
</list-item>
</list>
<p>We also recorded the average size of the sampled models for each dataset and the average CPU time (in seconds) to carry out BMA for one bootstrapped dataset.</p>
</sec>
<sec id="j_nejsds14_s_009">
<label>3.3</label>
<title>Results</title>
<p>The results are shown in Table <xref rid="j_nejsds14_tab_003">3</xref>. We used the combination of the <italic>g</italic>-prior with <inline-formula id="j_nejsds14_ineq_128"><alternatives><mml:math>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo>=</mml:mo>
<mml:msqrt>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msqrt></mml:math><tex-math><![CDATA[$g=\sqrt{n}$]]></tex-math></alternatives></inline-formula> as the parameter prior and the Beta-Binomial<inline-formula id="j_nejsds14_ineq_129"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(1,1)$]]></tex-math></alternatives></inline-formula> model space prior as the reference. Note that the <italic>g</italic>-prior with <inline-formula id="j_nejsds14_ineq_130"><alternatives><mml:math>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo>=</mml:mo>
<mml:msqrt>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msqrt></mml:math><tex-math><![CDATA[$g=\sqrt{n}$]]></tex-math></alternatives></inline-formula> was found to the best parameter prior by [<xref ref-type="bibr" rid="j_nejsds14_ref_037">37</xref>]. Metrics for all other combinations were calculated relative to the reference metric, and averaged across datasets. Detailed results of performance metrics for the simulation studies based on each of the 14 datasets can be found in the Supplementary materials. The “Score” column contains the average of the scores for PointEst, IntEst, Inference, Prediction and IntPred under each method. We used the Score column to rank the methods.</p>
<p>For each metric, we color the methods based on their performance relative to the reference metric. A method is colored green if it performed similarly or better than the reference method, yellow if it performed somewhat worse, and orange if it performed substantially worse.</p>
<p>For all choices of parameter prior, Beta-Binomial<inline-formula id="j_nejsds14_ineq_131"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(1,1)$]]></tex-math></alternatives></inline-formula> was the top scoring model space prior. The three Beta-Binomial versions with <inline-formula id="j_nejsds14_ineq_132"><alternatives><mml:math>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo>=</mml:mo>
<mml:msqrt>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msqrt></mml:math><tex-math><![CDATA[$g=\sqrt{n}$]]></tex-math></alternatives></inline-formula> were the top three methods across statistical tasks. The Complexity priors with <inline-formula id="j_nejsds14_ineq_133"><alternatives><mml:math>
<mml:mi mathvariant="italic">κ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$\kappa =1$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_nejsds14_ineq_134"><alternatives><mml:math>
<mml:mi mathvariant="italic">κ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$\kappa =2$]]></tex-math></alternatives></inline-formula> were the worst performing model space priors. The uniform prior denoted by Ber<inline-formula id="j_nejsds14_ineq_135"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0.5</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\theta =0.5)$]]></tex-math></alternatives></inline-formula> also performed less well than the Beta-Binomial priors. This ranking of methods was consistent across different performance metrics.</p>
<p>Most parameter and model prior combinations selected sparser models than the <inline-formula id="j_nejsds14_ineq_136"><alternatives><mml:math>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo>=</mml:mo>
<mml:msqrt>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msqrt></mml:math><tex-math><![CDATA[$g=\sqrt{n}$]]></tex-math></alternatives></inline-formula> and BB<inline-formula id="j_nejsds14_ineq_137"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(1,1)$]]></tex-math></alternatives></inline-formula> prior combination, with the exception of methods involving the uniform model space prior. The complexity priors selected very sparse models compared to our baseline, which may be explained by the strong sparsity induced by the prior. This may also explain the poor performance of the complexity priors across statistical tasks. Notably, the rankings of the prior combinations are similar for the different tasks. In particular, the rankings for prediction are consistent with those for point estimation and parameter inference, with a correlation of 0.77 between scores for point estimation and point prediction.</p>
<p>We also note that the EB model space priors tended to outperform the corresponding SDM model space priors when combined with the Hyper-<italic>g</italic> and EB-local parameter priors. However, the results with the EB model space priors took longer to computer on average because of the optimisation procedure. The hyper-<italic>g</italic> parameter priors are the slowest due to the integral calculations required in the posterior computation. In general, the Beta-Binomial priors performed better than the Bernoulli and complexity priors.</p>
<table-wrap id="j_nejsds14_tab_003">
<label>Table 3</label>
<caption>
<p>Performance of different parameter prior and model space prior combinations for inference in linear regression under model uncertainty: “PointEst” is the RMSE for point estimation, “IntEst” is the Mean Interval Score (MIS) for interval estimation, “Inference” is the 1- area under the precision-recall curve (AUPRC), “Prediction” is the RMSE for point prediction, while “IntPred” is the MIS for interval prediction. “N vars” is the average number of variables used for the task. All metrics are standardized to equal 1 for the <inline-formula id="j_nejsds14_ineq_138"><alternatives><mml:math>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo>=</mml:mo>
<mml:msqrt>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msqrt></mml:math><tex-math><![CDATA[$g=\sqrt{n}$]]></tex-math></alternatives></inline-formula> with BB<inline-formula id="j_nejsds14_ineq_139"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(1,1)$]]></tex-math></alternatives></inline-formula> prior on model space. For each column, lower value is better.</p>
</caption>
<graphic xlink:href="nejsds14_g004.jpg"/>
</table-wrap>
</sec>
</sec>
<sec id="j_nejsds14_s_010">
<label>4</label>
<title>Discussion</title>
<p>We have compared BMA techniques with different choices of model space priors and parameter priors using an empirical study based closely on real datasets. We found that the Beta-Binomial<inline-formula id="j_nejsds14_ineq_140"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(1,1)$]]></tex-math></alternatives></inline-formula> model space prior performed the best across various statistical tasks and choices of parameter priors. We found that the hierarchical model space priors with a hyper-prior on the prior inclusion probability <italic>θ</italic> was more diffuse and led to more efficient exploration of the model space. Fixed choices of <italic>θ</italic> led to worse performance across statistical tasks and were often quite concentrated. Complexity priors that induce high sparsity on model complexity performed worst among all the methods considered.</p>
<p>We are not the first to compare model space priors in the presence of model uncertainty. Past comparisons have either focused on a subset of the model priors discussed here, or evaluated BMA methods for only a subset of the statistical tasks considered here. In several cases, they also tended to use simulation designs that are at best loosely related to empirical data observed in practice.</p>
<p>Ley and Steel [<xref ref-type="bibr" rid="j_nejsds14_ref_031">31</xref>] evaluated the effect of different model priors on model selection performance using three real economic growth regressions datasets. However, they used only two fixed choices of g-priors: the Unit Information prior (UIP) with <inline-formula id="j_nejsds14_ineq_141"><alternatives><mml:math>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi></mml:math><tex-math><![CDATA[$g=n$]]></tex-math></alternatives></inline-formula> [<xref ref-type="bibr" rid="j_nejsds14_ref_028">28</xref>] and the risk inflation criterion (RIC) with <inline-formula id="j_nejsds14_ineq_142"><alternatives><mml:math>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[$g={p^{2}}$]]></tex-math></alternatives></inline-formula> [<xref ref-type="bibr" rid="j_nejsds14_ref_016">16</xref>], motivated by the simulation study of [<xref ref-type="bibr" rid="j_nejsds14_ref_013">13</xref>]. Porwal and Raftery [<xref ref-type="bibr" rid="j_nejsds14_ref_037">37</xref>] found both of these parameter prior choices to be outperformed by the parameter priors used in this study. Also, their comparison was based only on tall datasets <inline-formula id="j_nejsds14_ineq_143"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(n\gt p)$]]></tex-math></alternatives></inline-formula> and their comparison of methods was limited to the statistical tasks of inference and probabilistic prediction using the log-predictive score. They also did not consider EB versions of the Binomial<inline-formula id="j_nejsds14_ineq_144"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(p,\theta )$]]></tex-math></alternatives></inline-formula> and Beta-Binomial<inline-formula id="j_nejsds14_ineq_145"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(1,b)$]]></tex-math></alternatives></inline-formula> model space priors and complexity priors. Like Ley and Steel, we found that random <italic>θ</italic> versions (or Beta-Binomial versions) performed better since the hierarchical prior is less sensitive to the choice of prior model size <inline-formula id="j_nejsds14_ineq_146"><alternatives><mml:math>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$E[S]$]]></tex-math></alternatives></inline-formula>. Similarly, they found that priors specified by a fixed <italic>θ</italic> tended to be quite informative, casting doubt on their appropriateness as default reference priors.</p>
<p>Scott and Berger [<xref ref-type="bibr" rid="j_nejsds14_ref_047">47</xref>] discussed the multiplicity correction effect of a subset of the model space priors discussed here, specifically Ber<inline-formula id="j_nejsds14_ineq_147"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0.5</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\theta =0.5)$]]></tex-math></alternatives></inline-formula>, Ber<inline-formula id="j_nejsds14_ineq_148"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mi mathvariant="italic">B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({\theta _{EB}})$]]></tex-math></alternatives></inline-formula> and BB<inline-formula id="j_nejsds14_ineq_149"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(1,1)$]]></tex-math></alternatives></inline-formula>. They used a non-empirical simulation design, and did not compare methods based on the statistical tasks discussed here. Eicher et al [<xref ref-type="bibr" rid="j_nejsds14_ref_011">11</xref>] compared 12 parameter priors (of which <inline-formula id="j_nejsds14_ineq_150"><alternatives><mml:math>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mo>=</mml:mo>
<mml:msqrt>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msqrt></mml:math><tex-math><![CDATA[$g=\sqrt{n}$]]></tex-math></alternatives></inline-formula> is common with ours) along with two fixed model priors: Uniform model priors with <inline-formula id="j_nejsds14_ineq_151"><alternatives><mml:math>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0.5</mml:mn></mml:math><tex-math><![CDATA[$\theta =0.5$]]></tex-math></alternatives></inline-formula> and Ber<inline-formula id="j_nejsds14_ineq_152"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mi mathvariant="italic">D</mml:mi>
<mml:mi mathvariant="italic">M</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({\theta _{SDM}})$]]></tex-math></alternatives></inline-formula> with a prior expected model size of 7. The comparison was based on non-empirical simulation studies and one real growth regression dataset using predictive performance and inference measures. They found that the UIP with a uniform model prior performed better than Ber<inline-formula id="j_nejsds14_ineq_153"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mi mathvariant="italic">D</mml:mi>
<mml:mi mathvariant="italic">M</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({\theta _{SDM}})$]]></tex-math></alternatives></inline-formula> on the three statistical tasks common with ours. In contrast, we found that Ber<inline-formula id="j_nejsds14_ineq_154"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mi mathvariant="italic">D</mml:mi>
<mml:mi mathvariant="italic">M</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$({\theta _{SDM}})$]]></tex-math></alternatives></inline-formula> was ranked higher than Uniform model priors for all our three preferred parameter priors across the statistical tasks considered.</p>
<p>We found the complexity priors [<xref ref-type="bibr" rid="j_nejsds14_ref_003">3</xref>] to perform relatively poorly. At first sight, this seems to be in conflict with the theoretical results of Castillo et al [<xref ref-type="bibr" rid="j_nejsds14_ref_003">3</xref>], who showed that under certain assumptions the posterior distribution contracts optimally to recover an unknown sparse parameter vector and gives optimal predictions. However, their theoretical results assume that the data are generated from a spike and slab prior with the Laplace distribution as the slab density, and that the error variance <inline-formula id="j_nejsds14_ineq_155"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${\sigma ^{2}}$]]></tex-math></alternatives></inline-formula> is known, which rarely holds in practice. Also, Rossell [<xref ref-type="bibr" rid="j_nejsds14_ref_044">44</xref>] argued that complexity priors can introduce very strong sparsity a priori, and showed empirically that when the true model is not sparse, complexity priors may perform suboptimally for finite <italic>n</italic>. This is consistent with our results.</p>
<p>We have focused attention on independent model priors, i.e. priors in which the inclusion of each variable is statistically independent of that of the other variables. However, non-independent default priors have been proposed as well. George [<xref ref-type="bibr" rid="j_nejsds14_ref_017">17</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_018">18</xref>] proposed dilution priors which dilute the prior model probability within subsets of similar models with highly correlated predictors. There is also research designing dependent model priors based on domain knowledge [<xref ref-type="bibr" rid="j_nejsds14_ref_002">2</xref>, <xref ref-type="bibr" rid="j_nejsds14_ref_010">10</xref>]. Dellaportas et al [<xref ref-type="bibr" rid="j_nejsds14_ref_009">9</xref>] proposed a joint specification of the prior distribution across models so that the sensitivity of posterior model probabilities to the dispersion of prior distributions for the parameters of individual models (Lindley’s paradox) is diminished. Villa and Walker [<xref ref-type="bibr" rid="j_nejsds14_ref_049">49</xref>] assigned prior mass to models on the basis of their <italic>worth</italic>, based on the KL-divergence between densities under different models. However, all of these dependent model space priors lead to increased computational complexity and have been shown to work only when <italic>p</italic> is relatively small. They have also not yet been implemented in publicly available software.</p>
</sec>
</body>
<back>
<ack id="j_nejsds14_ack_001">
<title>Acknowledgements</title>
<p>We thank Abel Rodriguez for helpful discussions.</p></ack>
<ref-list id="j_nejsds14_reflist_001">
<title>References</title>
<ref id="j_nejsds14_ref_001">
<label>[1]</label><mixed-citation publication-type="journal"> <string-name><surname>Bartlett</surname>, <given-names>M. S.</given-names></string-name> (<year>1957</year>). <article-title>A Comment on D. V. Lindley’s Statistical Paradox</article-title>. <source>Biometrika</source> <volume>44</volume> <fpage>533</fpage>–<lpage>534</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1093/biomet/52.3-4.507" xlink:type="simple">https://doi.org/10.1093/biomet/52.3-4.507</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0207142">MR0207142</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_002">
<label>[2]</label><mixed-citation publication-type="other"> <string-name><surname>Brock</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Durlauf</surname>, <given-names>S. N.</given-names></string-name> and <string-name><surname>West</surname>, <given-names>K. D.</given-names></string-name> (2003). <italic>Policy evaluation in uncertain economic environments</italic>. National Bureau of Economic Research Cambridge, Mass., USA.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_003">
<label>[3]</label><mixed-citation publication-type="journal"> <string-name><surname>Castillo</surname>, <given-names>I.</given-names></string-name>, <string-name><surname>Schmidt-Hieber</surname>, <given-names>J.</given-names></string-name> and <string-name><surname>Van der Vaart</surname>, <given-names>A.</given-names></string-name> (<year>2015</year>). <article-title>Bayesian linear regression with sparse priors</article-title>. <source>The Annals of Statistics</source> <volume>43</volume>(<issue>5</issue>) <fpage>1986</fpage>–<lpage>2018</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/15-AOS1334" xlink:type="simple">https://doi.org/10.1214/15-AOS1334</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3375874">MR3375874</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_004">
<label>[4]</label><mixed-citation publication-type="journal"> <string-name><surname>Celeux</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>El Anbari</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Marin</surname>, <given-names>J. q. M.</given-names></string-name> and <string-name><surname>Robert</surname>, <given-names>C. P.</given-names></string-name> (<year>2012</year>). <article-title>Regularization in Regression: Comparing Bayesian and Frequentist Methods in a Poorly Informative Situation</article-title>. <source>Bayesian Analysis</source> <volume>7</volume> <fpage>477</fpage>–<lpage>502</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/12-BA716" xlink:type="simple">https://doi.org/10.1214/12-BA716</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2934959">MR2934959</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_005">
<label>[5]</label><mixed-citation publication-type="other"> <string-name><surname>Clyde</surname>, <given-names>M.</given-names></string-name> (2020). BAS: Bayesian Variable Selection and Model Averaging using Bayesian Adaptive Sampling. R package version 1.5.5.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_006">
<label>[6]</label><mixed-citation publication-type="journal"> <string-name><surname>Clyde</surname>, <given-names>M.</given-names></string-name> and <string-name><surname>George</surname>, <given-names>E. I.</given-names></string-name> (<year>2000</year>). <article-title>Flexible empirical Bayes estimation for wavelets</article-title>. <source>Journal of the Royal Statistical Society: Series B, Statistical Methodology</source> <volume>62</volume>(<issue>4</issue>) <fpage>681</fpage>–<lpage>698</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1111/1467-9868.00257" xlink:type="simple">https://doi.org/10.1111/1467-9868.00257</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1796285">MR1796285</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_007">
<label>[7]</label><mixed-citation publication-type="journal"> <string-name><surname>Clyde</surname>, <given-names>M.</given-names></string-name> and <string-name><surname>George</surname>, <given-names>E. I.</given-names></string-name> (<year>2004</year>). <article-title>Model uncertainty</article-title>. <source>Statistical Science</source> <volume>19</volume>(<issue>1</issue>) <fpage>81</fpage>–<lpage>94</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/088342304000000035" xlink:type="simple">https://doi.org/10.1214/088342304000000035</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2082148">MR2082148</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_008">
<label>[8]</label><mixed-citation publication-type="journal"> <string-name><surname>Deckers</surname>, <given-names>T.</given-names></string-name> and <string-name><surname>Hanck</surname>, <given-names>C.</given-names></string-name> (<year>2014</year>). <article-title>Variable Selection in Cross-Section Regressions: Comparisons and Extensions</article-title>. <source>Oxford Bulletin of Economics and Statistics</source> <volume>76</volume>(<issue>6</issue>) <fpage>841</fpage>–<lpage>873</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_009">
<label>[9]</label><mixed-citation publication-type="journal"> <string-name><surname>Dellaportas</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Forster</surname>, <given-names>J. J.</given-names></string-name> and <string-name><surname>Ntzoufras</surname>, <given-names>I.</given-names></string-name> (<year>2012</year>). <article-title>Joint specification of model space and parameter space prior distributions</article-title>. <source>Statistical Science</source> <volume>27</volume>(<issue>2</issue>) <fpage>232</fpage>–<lpage>246</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/11-STS369" xlink:type="simple">https://doi.org/10.1214/11-STS369</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2963994">MR2963994</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_010">
<label>[10]</label><mixed-citation publication-type="journal"> <string-name><surname>Durlauf</surname>, <given-names>S. N.</given-names></string-name>, <string-name><surname>Kourtellos</surname>, <given-names>A.</given-names></string-name> and <string-name><surname>Tan</surname>, <given-names>C. M.</given-names></string-name> (<year>2008</year>). <article-title>Are any growth theories robust?</article-title> <source>The Economic Journal</source> <volume>118</volume>(<issue>527</issue>) <fpage>329</fpage>–<lpage>346</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_011">
<label>[11]</label><mixed-citation publication-type="journal"> <string-name><surname>Eicher</surname>, <given-names>T. S.</given-names></string-name>, <string-name><surname>Papageorgiou</surname>, <given-names>C.</given-names></string-name> and <string-name><surname>Raftery</surname>, <given-names>A. E.</given-names></string-name> (<year>2011</year>). <article-title>Default priors and predictive performance in Bayesian model averaging, with application to growth determinants</article-title>. <source>Journal of Applied Econometrics</source> <volume>26</volume>(<issue>1</issue>) <fpage>30</fpage>–<lpage>55</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1002/jae.1112" xlink:type="simple">https://doi.org/10.1002/jae.1112</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2759908">MR2759908</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_012">
<label>[12]</label><mixed-citation publication-type="journal"> <string-name><surname>Fan</surname>, <given-names>J.</given-names></string-name> and <string-name><surname>Lv</surname>, <given-names>J.</given-names></string-name> (<year>2008</year>). <article-title>Sure independence screening for ultrahigh dimensional feature space</article-title>. <source>Journal of the Royal Statistical Society: Series B — Statistical Methodology</source> <volume>70</volume>(<issue>5</issue>) <fpage>849</fpage>–<lpage>911</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1111/j.1467-9868.2008.00674.x" xlink:type="simple">https://doi.org/10.1111/j.1467-9868.2008.00674.x</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2530322">MR2530322</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_013">
<label>[13]</label><mixed-citation publication-type="journal"> <string-name><surname>Fernández</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Ley</surname>, <given-names>E.</given-names></string-name> and <string-name><surname>Steel</surname>, <given-names>M. F. J.</given-names></string-name> (<year>2001</year>). <article-title>Benchmark priors for Bayesian model averaging</article-title>. <source>Journal of Econometrics</source> <volume>100</volume>(<issue>2</issue>) <fpage>381</fpage>–<lpage>427</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1016/S0304-4076(00)00076-2" xlink:type="simple">https://doi.org/10.1016/S0304-4076(00)00076-2</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1820410">MR1820410</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_014">
<label>[14]</label><mixed-citation publication-type="other"> <string-name><surname>Filzmoser</surname>, <given-names>P.</given-names></string-name> and <string-name><surname>Varmuza</surname>, <given-names>K.</given-names></string-name> (2017). chemometrics: Multivariate Statistical Analysis in Chemometrics. R package version 1.4.2. <uri>https://CRAN.R-project.org/package=chemometrics</uri>.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_015">
<label>[15]</label><mixed-citation publication-type="journal"> <string-name><surname>Forte</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Garcia-Donato</surname>, <given-names>G.</given-names></string-name> and <string-name><surname>Steel</surname>, <given-names>M. F. J.</given-names></string-name> (<year>2018</year>). <article-title>Methods and tools for Bayesian variable selection and model averaging in normal linear regression</article-title>. <source>International Statistical Review</source> <volume>86</volume>(<issue>2</issue>) <fpage>237</fpage>–<lpage>258</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1111/insr.12249" xlink:type="simple">https://doi.org/10.1111/insr.12249</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3852410">MR3852410</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_016">
<label>[16]</label><mixed-citation publication-type="journal"> <string-name><surname>Foster</surname>, <given-names>D. P.</given-names></string-name> and <string-name><surname>George</surname>, <given-names>E. I.</given-names></string-name> (<year>1994</year>). <article-title>The risk inflation criterion for multiple regression</article-title>. <source>Annals of Statistics</source> <volume>22</volume>(<issue>4</issue>) <fpage>1947</fpage>–<lpage>1975</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/aos/1176325766" xlink:type="simple">https://doi.org/10.1214/aos/1176325766</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1329177">MR1329177</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_017">
<label>[17]</label><mixed-citation publication-type="chapter"> <string-name><surname>George</surname>, <given-names>E.</given-names></string-name> (<year>1999</year>). <article-title>Discussion of “Model averaging and model search strategies” by M. Clyde</article-title>. In <source>Bayesian Statistics 6–Proceedings of the Sixth Valencia International Meeting</source>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1723497">MR1723497</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_018">
<label>[18]</label><mixed-citation publication-type="chapter"> <string-name><surname>George</surname>, <given-names>E. I.</given-names></string-name> (<year>2010</year>). <chapter-title>Dilution priors: Compensating for model space redundancy</chapter-title>. In <source>Borrowing Strength: Theory Powering Applications–A Festschrift for Lawrence D. Brown</source> <fpage>158</fpage>–<lpage>165</lpage> <publisher-name>Institute of Mathematical Statistics</publisher-name>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2798517">MR2798517</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_019">
<label>[19]</label><mixed-citation publication-type="journal"> <string-name><surname>George</surname>, <given-names>E. I.</given-names></string-name> and <string-name><surname>Foster</surname>, <given-names>D. P.</given-names></string-name> (<year>2000</year>). <article-title>Calibration and empirical Bayes variable selection</article-title>. <source>Biometrika</source> <volume>87</volume>(<issue>4</issue>) <fpage>731</fpage>–<lpage>747</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1093/biomet/87.4.731" xlink:type="simple">https://doi.org/10.1093/biomet/87.4.731</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1813972">MR1813972</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_020">
<label>[20]</label><mixed-citation publication-type="journal"> <string-name><surname>George</surname>, <given-names>E. I.</given-names></string-name> and <string-name><surname>McCulloch</surname>, <given-names>R. E.</given-names></string-name> (<year>1993</year>). <article-title>Variable selection via Gibbs sampling</article-title>. <source>Journal of the American Statistical Association</source> <volume>88</volume>(<issue>423</issue>) <fpage>881</fpage>–<lpage>889</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_021">
<label>[21]</label><mixed-citation publication-type="journal"> <string-name><surname>Gneiting</surname>, <given-names>T.</given-names></string-name> and <string-name><surname>Raftery</surname>, <given-names>A. E.</given-names></string-name> (<year>2007</year>). <article-title>Strictly proper scoring rules, prediction, and estimation</article-title>. <source>Journal of the American Statistical Association</source> <volume>102</volume>(<issue>477</issue>) <fpage>359</fpage>–<lpage>378</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1198/016214506000001437" xlink:type="simple">https://doi.org/10.1198/016214506000001437</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2345548">MR2345548</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_022">
<label>[22]</label><mixed-citation publication-type="journal"> <string-name><surname>Gu</surname>, <given-names>C.</given-names></string-name> (<year>2014</year>). <article-title>Smoothing Spline ANOVA Models: R Package gss</article-title>. <source>Journal of Statistical Software</source> <volume>58</volume>(<issue>5</issue>) <fpage>1</fpage>–<lpage>25</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_023">
<label>[23]</label><mixed-citation publication-type="journal"> <string-name><surname>Hansen</surname>, <given-names>M. H.</given-names></string-name> and <string-name><surname>Yu</surname>, <given-names>B.</given-names></string-name> (<year>2003</year>). <article-title>Minimum description length model selection criteria for generalized linear models</article-title>. <source>Lecture Notes-Monograph Series</source> <volume>40</volume> <fpage>145</fpage>–<lpage>163</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/lnms/1215091140" xlink:type="simple">https://doi.org/10.1214/lnms/1215091140</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2004337">MR2004337</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_024">
<label>[24]</label><mixed-citation publication-type="journal"> <string-name><surname>Hoeting</surname>, <given-names>J. A.</given-names></string-name>, <string-name><surname>Raftery</surname>, <given-names>A. E.</given-names></string-name> and <string-name><surname>Madigan</surname>, <given-names>D.</given-names></string-name> (<year>2002</year>). <article-title>Bayesian variable and transformation selection in linear regression</article-title>. <source>Journal of Computational and Graphical Statistics</source> <volume>11</volume>(<issue>3</issue>) <fpage>485</fpage>–<lpage>507</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1198/106186002501" xlink:type="simple">https://doi.org/10.1198/106186002501</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1938444">MR1938444</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_025">
<label>[25]</label><mixed-citation publication-type="journal"> <string-name><surname>Hoeting</surname>, <given-names>J. A.</given-names></string-name>, <string-name><surname>Madigan</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Raftery</surname>, <given-names>A. E.</given-names></string-name> and <string-name><surname>Volinsky</surname>, <given-names>C. T.</given-names></string-name> (<year>1999</year>). <article-title>Bayesian model averaging: a tutorial</article-title>. <source>Statistical Science</source> <volume>14</volume> <fpage>382</fpage>–<lpage>417</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/ss/1009212519" xlink:type="simple">https://doi.org/10.1214/ss/1009212519</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1765176">MR1765176</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_026">
<label>[26]</label><mixed-citation publication-type="other"> <string-name><surname>Ishwaran</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Rao</surname>, <given-names>J. S.</given-names></string-name> and <string-name><surname>Kogalur</surname>, <given-names>U. B.</given-names></string-name> (2013). spikeslab: Prediction and variable selection using spike and slab regression. R package version 1.1.5. <uri>http://cran.r-project.org/web/packages/spikeslab/</uri>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/21-ejp733" xlink:type="simple">https://doi.org/10.1214/21-ejp733</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4366222">MR4366222</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_027">
<label>[27]</label><mixed-citation publication-type="other"> <string-name><surname>James</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Witten</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Hastie</surname>, <given-names>T.</given-names></string-name> and <string-name><surname>Tibshirani</surname>, <given-names>R.</given-names></string-name> (2017). ISLR: Data for an Introduction to Statistical Learning with Applications in R. R package version 1.2. <uri>https://CRAN.R-project.org/package=ISLR</uri>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/978-1-0716-1418-1" xlink:type="simple">https://doi.org/10.1007/978-1-0716-1418-1</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4309209">MR4309209</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_028">
<label>[28]</label><mixed-citation publication-type="journal"> <string-name><surname>Kass</surname>, <given-names>R. E.</given-names></string-name> and <string-name><surname>Raftery</surname>, <given-names>A. E.</given-names></string-name> (<year>1995</year>). <article-title>Bayes factors</article-title>. <source>Journal of the American Statistical Association</source> <volume>90</volume>(<issue>430</issue>) <fpage>773</fpage>–<lpage>795</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/01621459.1995.10476572" xlink:type="simple">https://doi.org/10.1080/01621459.1995.10476572</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3363402">MR3363402</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_029">
<label>[29]</label><mixed-citation publication-type="book"> <string-name><surname>Leamer</surname>, <given-names>E. E.</given-names></string-name> (<year>1978</year>) <source>Specification Searches: Ad hoc Inference with Nonexperimental Data</source> <volume>53</volume>. <publisher-name>Wiley</publisher-name>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0471118">MR0471118</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_030">
<label>[30]</label><mixed-citation publication-type="journal"> <string-name><surname>Levine</surname>, <given-names>R.</given-names></string-name> and <string-name><surname>Renelt</surname>, <given-names>D.</given-names></string-name> (<year>1992</year>). <article-title>A sensitivity analysis of cross-country growth regressions</article-title>. <italic>The American economic review</italic> <fpage>942</fpage>–<lpage>963</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_031">
<label>[31]</label><mixed-citation publication-type="journal"> <string-name><surname>Ley</surname>, <given-names>E.</given-names></string-name> and <string-name><surname>Steel</surname>, <given-names>M. F.</given-names></string-name> (<year>2009</year>). <article-title>On the effect of prior assumptions in Bayesian model averaging with applications to growth regression</article-title>. <source>Journal of applied econometrics</source> <volume>24</volume>(<issue>4</issue>) <fpage>651</fpage>–<lpage>674</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1002/jae.1057" xlink:type="simple">https://doi.org/10.1002/jae.1057</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2675199">MR2675199</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_032">
<label>[32]</label><mixed-citation publication-type="journal"> <string-name><surname>Liang</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Paulo</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Molina</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Clyde</surname>, <given-names>M. A.</given-names></string-name> and <string-name><surname>Berger</surname>, <given-names>J. O.</given-names></string-name> (<year>2008</year>). <article-title>Mixtures of g priors for Bayesian variable selection</article-title>. <source>Journal of the American Statistical Association</source> <volume>103</volume> <fpage>410</fpage>–<lpage>423</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1198/016214507000001337" xlink:type="simple">https://doi.org/10.1198/016214507000001337</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2420243">MR2420243</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_033">
<label>[33]</label><mixed-citation publication-type="other"> <string-name><surname>Lumley</surname>, <given-names>T.</given-names></string-name> (2020). leaps: Regression Subset Selection. R package version 3.1. <uri>https://CRAN.R-project.org/package=leaps</uri>.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_034">
<label>[34]</label><mixed-citation publication-type="journal"> <string-name><surname>Madigan</surname>, <given-names>D.</given-names></string-name> and <string-name><surname>Raftery</surname>, <given-names>A. E.</given-names></string-name> (<year>1994</year>). <article-title>Model selection and accounting for model uncertainty in graphical models using Occam’s window</article-title>. <source>Journal of the American Statistical Association</source> <volume>89</volume>(<issue>428</issue>) <fpage>1535</fpage>–<lpage>1546</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_035">
<label>[35]</label><mixed-citation publication-type="journal"> <string-name><surname>Narisetty</surname>, <given-names>N. N.</given-names></string-name> and <string-name><surname>He</surname>, <given-names>X.</given-names></string-name> (<year>2014</year>). <article-title>Bayesian variable selection with shrinking and diffusing priors</article-title>. <source>The Annals of Statistics</source> <volume>42</volume>(<issue>2</issue>) <fpage>789</fpage>–<lpage>817</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/14-AOS1207" xlink:type="simple">https://doi.org/10.1214/14-AOS1207</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3210987">MR3210987</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_036">
<label>[36]</label><mixed-citation publication-type="other"> <string-name><surname>Newman</surname>, <given-names>D. J.</given-names></string-name>, <string-name><surname>Hettich</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Blake</surname>, <given-names>C. L.</given-names></string-name> and <string-name><surname>Merz</surname>, <given-names>C. J.</given-names></string-name> (1998). <italic>UCI Repository of machine learning databases</italic>. <uri>http://www.ics.uci.edu/~mlearn/MLRepository.html</uri>.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_037">
<label>[37]</label><mixed-citation publication-type="journal"> <string-name><surname>Porwal</surname>, <given-names>A.</given-names></string-name> and <string-name><surname>Raftery</surname>, <given-names>A. E.</given-names></string-name> (<year>2022</year>). <article-title>Comparing methods for statistical inference with model uncertainty</article-title>. <source>Proceedings of the National Academy of Sciences</source> <volume>119</volume>(<issue>16</issue>) <fpage>2120737119</fpage>.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_038">
<label>[38]</label><mixed-citation publication-type="other"> <string-name><surname>Raftery</surname>, <given-names>A. E.</given-names></string-name> (1988). Approximate Bayes factors for generalized linear models. Technical Report No. <elocation-id>121</elocation-id>, Department of Statistics, University of Washington. <uri>https://stat.uw.edu/sites/default/files/files/reports/1988/tr121.pdf</uri>.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_039">
<label>[39]</label><mixed-citation publication-type="journal"> <string-name><surname>Raftery</surname>, <given-names>A. E.</given-names></string-name> and <string-name><surname>Zheng</surname>, <given-names>Y.</given-names></string-name> (<year>2003</year>). <article-title>Discussion: Performance of Bayesian model averaging</article-title>. <source>Journal of the American Statistical Association</source> <volume>98</volume>(<issue>464</issue>) <fpage>931</fpage>–<lpage>938</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_040">
<label>[40]</label><mixed-citation publication-type="journal"> <string-name><surname>Raftery</surname>, <given-names>A. E.</given-names></string-name>, <string-name><surname>Madigan</surname>, <given-names>D.</given-names></string-name> and <string-name><surname>Hoeting</surname>, <given-names>J. A.</given-names></string-name> (<year>1997</year>). <article-title>Bayesian model averaging for linear regression models</article-title>. <source>Journal of the American Statistical Association</source> <volume>92</volume>(<issue>437</issue>) <fpage>179</fpage>–<lpage>191</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.2307/2291462" xlink:type="simple">https://doi.org/10.2307/2291462</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1436107">MR1436107</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_041">
<label>[41]</label><mixed-citation publication-type="journal"> <string-name><surname>Rohart</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Gautier</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Singh</surname>, <given-names>A.</given-names></string-name> and <string-name><surname>Le Cao</surname>, <given-names>K. q. A.</given-names></string-name> (<year>2017</year>). <article-title>mixOmics: An R package for ’omics feature selection and multiple data integration</article-title>. <source>PLoS Computational Biology</source> <volume>13</volume>(<issue>11</issue>) <fpage>1005752</fpage>.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_042">
<label>[42]</label><mixed-citation publication-type="journal"> <string-name><surname>Rossell</surname>, <given-names>D.</given-names></string-name> (<year>2021</year>). <article-title>Concentration of posterior model probabilities and normalized l0 criteria</article-title>. <source>Bayesian Analysis</source> <volume>1</volume>(<issue>1</issue>) <fpage>1</fpage>–<lpage>27</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/21-ba1262" xlink:type="simple">https://doi.org/10.1214/21-ba1262</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4483231">MR4483231</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_043">
<label>[43]</label><mixed-citation publication-type="journal"> <string-name><surname>Rossell</surname>, <given-names>D.</given-names></string-name> and <string-name><surname>Rubio</surname>, <given-names>F. J.</given-names></string-name> (<year>2018</year>). <article-title>Tractable Bayesian Variable Selection: Beyond Normality</article-title>. <source>Journal of the American Statistical Association</source>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/01621459.2017.1371025" xlink:type="simple">https://doi.org/10.1080/01621459.2017.1371025</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3902243">MR3902243</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_044">
<label>[44]</label><mixed-citation publication-type="journal"> <string-name><surname>Rossell</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Abril</surname>, <given-names>O.</given-names></string-name> and <string-name><surname>Bhattacharya</surname>, <given-names>A.</given-names></string-name> (<year>2021</year>). <article-title>Approximate Laplace approximations for scalable model selection</article-title>. <source>Journal of the Royal Statistical Society: Series B (Statistical Methodology)</source> <volume>83</volume>(<issue>4</issue>) <fpage>853</fpage>–<lpage>879</lpage>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4320004">MR4320004</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_045">
<label>[45]</label><mixed-citation publication-type="other"> <string-name><surname>Sala-i-Martin</surname>, <given-names>X.</given-names></string-name> (1997). <italic>I just ran four million regressions</italic>. National Bureau of Economic Research Cambridge, Mass., USA.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_046">
<label>[46]</label><mixed-citation publication-type="journal"> <string-name><surname>Sala-I-Martin</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Doppelhofer</surname>, <given-names>G.</given-names></string-name> and <string-name><surname>Miller</surname>, <given-names>R. I.</given-names></string-name> (<year>2004</year>). <article-title>Determinants of long-term growth: A Bayesian averaging of classical estimates (BACE) approach</article-title>. <source>American economic review</source> <volume>94</volume>(<issue>4</issue>) <fpage>813</fpage>–<lpage>835</lpage>.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_047">
<label>[47]</label><mixed-citation publication-type="journal"> <string-name><surname>Scott</surname>, <given-names>J. G.</given-names></string-name> and <string-name><surname>Berger</surname>, <given-names>J. O.</given-names></string-name> (<year>2010</year>). <article-title>Bayes and empirical-Bayes multiplicity adjustment in the variable-selection problem</article-title>. <source>The Annals of Statistics</source> <fpage>2587</fpage>–<lpage>2619</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/10-AOS792" xlink:type="simple">https://doi.org/10.1214/10-AOS792</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=2722450">MR2722450</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_048">
<label>[48]</label><mixed-citation publication-type="journal"> <string-name><surname>van Zwet</surname>, <given-names>E.</given-names></string-name> (<year>2019</year>). <article-title>A default prior for regression coefficients</article-title>. <source>Statistical Methods in Medical Research</source> <volume>28</volume>(<issue>12</issue>) <fpage>3799</fpage>–<lpage>3807</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1177/0962280218817792" xlink:type="simple">https://doi.org/10.1177/0962280218817792</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=4003623">MR4003623</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_049">
<label>[49]</label><mixed-citation publication-type="journal"> <string-name><surname>Villa</surname>, <given-names>C.</given-names></string-name> and <string-name><surname>Walker</surname>, <given-names>S.</given-names></string-name> (<year>2015</year>). <article-title>An objective Bayesian criterion to determine model prior probabilities</article-title>. <source>Scandinavian Journal of Statistics</source> <volume>42</volume>(<issue>4</issue>) <fpage>947</fpage>–<lpage>966</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1111/sjos.12145" xlink:type="simple">https://doi.org/10.1111/sjos.12145</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3426304">MR3426304</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_050">
<label>[50]</label><mixed-citation publication-type="journal"> <string-name><surname>Wasserman</surname>, <given-names>L.</given-names></string-name> (<year>2000</year>). <article-title>Bayesian model selection and model averaging</article-title>. <source>Journal of Mathematical Psychology</source> <volume>44</volume>(<issue>1</issue>) <fpage>92</fpage>–<lpage>107</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1006/jmps.1999.1278" xlink:type="simple">https://doi.org/10.1006/jmps.1999.1278</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=1770003">MR1770003</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_051">
<label>[51]</label><mixed-citation publication-type="journal"> <string-name><surname>Yang</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Wainwright</surname>, <given-names>M. J.</given-names></string-name> and <string-name><surname>Jordan</surname>, <given-names>M. I.</given-names></string-name> (<year>2016</year>). <article-title>On the computational complexity of high-dimensional Bayesian variable selection</article-title>. <source>The Annals of Statistics</source> <volume>44</volume>(<issue>6</issue>) <fpage>2497</fpage>–<lpage>2532</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1214/15-AOS1417" xlink:type="simple">https://doi.org/10.1214/15-AOS1417</ext-link>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=3576552">MR3576552</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_052">
<label>[52]</label><mixed-citation publication-type="journal"> <string-name><surname>Young</surname>, <given-names>W. C.</given-names></string-name>, <string-name><surname>Raftery</surname>, <given-names>A. E.</given-names></string-name> and <string-name><surname>Yeung</surname>, <given-names>K. Y.</given-names></string-name> (<year>2014</year>). <article-title>Fast Bayesian inference for gene regulatory networks using ScanBMA</article-title>. <source>BMC Systems Biology</source> <volume>8</volume>(<issue>1</issue>) <fpage>47</fpage>.</mixed-citation>
</ref>
<ref id="j_nejsds14_ref_053">
<label>[53]</label><mixed-citation publication-type="chapter"> <string-name><surname>Zellner</surname>, <given-names>A.</given-names></string-name> (<year>1986</year>). <chapter-title>On assessing prior distributions and Bayesian regression analysis with g-prior distributions</chapter-title>. In <source>Bayesian Inference and Decision Techniques</source> <volume>6</volume>. <ext-link ext-link-type="uri" xlink:href="https://mathscinet.ams.org/mathscinet-getitem?mr=0881437">MR0881437</ext-link></mixed-citation>
</ref>
<ref id="j_nejsds14_ref_054">
<label>[54]</label><mixed-citation publication-type="journal"> <string-name><surname>Zellner</surname>, <given-names>A.</given-names></string-name> and <string-name><surname>Siow</surname>, <given-names>A.</given-names></string-name> (<year>1980</year>). <article-title>Posterior odds ratios for selected regression hypotheses</article-title>. <source>Trabajos de Estadística y de Investigaciów Operativa</source> <volume>31</volume>(<issue>1</issue>) <fpage>585</fpage>–<lpage>603</lpage>.</mixed-citation>
</ref>
</ref-list>
</back>
</article>
