The New England Journal of Statistics in Data Science logo


  • Help
Login Register

  1. Home
  2. Issues
  3. Volume 1, Issue 3 (2023)
  4. Construction of Supersaturated Designs w ...

Construction of Supersaturated Designs with Small Coherence for Variable Selection
Volume 1, Issue 3 (2023), pp. 323–333
Youran Qi   Peter Chien  

Authors

 
Placeholder
https://doi.org/10.51387/23-NEJSDS34
Pub. online: 5 June 2023      Type: Methodology Article      Open accessOpen Access
Area: Machine Learning and Data Mining

Accepted
19 April 2023
Published
5 June 2023

Abstract

The supersaturated design is often used to discover important factors in an experiment with a large number of factors and a small number of runs. We propose a method for constructing supersaturated designs with small coherence. Such designs are useful for variable selection methods such as the Lasso. Examples are provided to illustrate the proposed method.

1 Introduction

Modern experiments can involve a large number of factors. Assuming the effect sparsity principle [3], an experiment with high-dimensional inputs only has a small number of significant factors. Two-level supersaturated designs are often used to identify important factors in screening experiments. Data from such a design can be modeled by a linear model [33] given by
(1.1)
\[ \boldsymbol{y}=\mathbf{X}\boldsymbol{\beta }+\boldsymbol{\epsilon },\]
where X is the $n\times p$ design matrix at two levels $-1$ and $+1$, denoted by − and +, $\boldsymbol{y}$ is the response vector of the n observations, $\boldsymbol{\beta }$ is the regression vector of p coefficients, the error vector $\boldsymbol{\epsilon }$ follows the multivariate normal distribution ${N_{n}}(\mathbf{0},{\sigma ^{2}}\mathbf{I})$ and I is the $n\times n$ identity matrix. In X, the first column is the intercept with all +’s and the remaining columns represent the $p-1$ factors. We call X balanced if all its factors consist of an equal number of − and +. If $p=n$, X is a saturated design since all the degrees of freedom are used to estimate $\boldsymbol{\beta }$. If $p\gt n$, X is a supersaturated design since there are not enough degrees of freedom to estimate all components of $\boldsymbol{\beta }$. Construction of supersaturated designs that can accommodate a large number of factors relative to the number of runs has been gaining more and more attention [5, 16, 30, 26]. Popular criteria for comparing supersaturated designs include minimax [2], $E({s^{2}})$ [18, 27] and mean square correlation [8].
It is increasingly common to use modern variable selection methods to analyze data from a supersaturated design and identify important factors [18, 32, 17]. For example, when the number of variables is greater than the number of observations, Osborne, Presnell and Turlach [22] advised to use the Lasso [28] for an initial selection, and then conduct a best subset selection (or similar method) on the variables selected by the Lasso. We focus on discussing and demonstrating our method for the Lasso given its popularity. The proposed designs also work with other penalized regression methods. For variable selection methods such as the Lasso, Zhao and Yu [34] showed that their performance depends more critically on the worst case column correlation than the average column correlation of the design matrix. More details about this point can be found in Appendix B.
Motivated by the importance of controlling the worst case column correlation, we propose a new method to construct supersaturated designs according to the coherence criterion [7], which measures the worst case column correlation. The coherence of an $n\times p$ matrix X is
\[ \mu (\mathbf{X})=\underset{1\le i\lt j\le p}{\max }\frac{|\langle {\boldsymbol{x}_{i}},{\boldsymbol{x}_{j}}\rangle |}{\| {\boldsymbol{x}_{i}}\| \| {\boldsymbol{x}_{j}}\| },\]
where the subscript i denotes the ith column, $\langle \cdot ,\cdot \rangle $ is the dot product and $\| \cdot \| $ is the ${L_{2}}$-norm. Whenever there is no confusion, hereinafter we will drop the symbol X in $\mu (\mathbf{X})$. Clearly, $0\le \mu \le 1$. The smaller μ is, the closer the matrix is to orthogonal. If the columns of X are centered, μ is reduced to the maximum absolute column correlation of X, which is equivalent to the correlation criterion used in [19]. Note that
\[ \mu (\mathbf{X})=\underset{1\le i\lt j\le p}{\max }\frac{|{\boldsymbol{x}_{i}^{\top }}{\boldsymbol{x}_{j}}|}{n}=\frac{{s_{\text{max}}}}{n},\]
where ${s_{\text{max}}}={\max _{1\le i\lt j\le p}}|{\boldsymbol{x}_{i}^{\top }}{\boldsymbol{x}_{j}}|$ is the value used in the minimax criterion [2]. Different from ${s_{\text{max}}}$, $\mu (\mathbf{X})$ reflects the sample size n in its definition. Because the sample size matters in many data analysis procedures, we will construct designs under $\mu (\mathbf{X})$ instead of ${s_{\text{max}}}$ (we will also compare designs with different sample sizes in Section 3.1).
Our construction method allows some columns of the design matrix to be unbalanced. Factor balance in a supersaturated design guarantees an accurate estimate of the intercept but unbalance provides other significant benefits in constructing designs. Being unbalanced provides flexibility in controlling the column correlations between every pair of the main effects to achieve a lower average column correlation [29]. Sacrificing the precision of the estimate of the intercept can better estimate the main effects. In addition, allowing unbalance enables us to construct designs with more flexible sizes.
The remainder of the article will unfold as follows. In Section 2, we detail our construction method. In Section 3, we use several examples to demonstrate the advantage of the constructed designs against some benchmark designs for a variable selection problem. We conclude the paper and provide some discussions in Section 4.

2 The Construction Method

2.1 Notation and Definitions

Throughout, “#Balance” denotes the number of balanced columns of a design. Let ${\mathbf{1}_{n}}$ denote the n-dimensional unit vector with $+1$’s. Let I denote the identity matrix. For a matrix A with entries in $\{-1,+1\}$ and an even number of rows, let ${\mathbf{A}^{\ast }}$ denote a matrix obtained by changing the signs of entries in all odd rows and ${\mathbf{A}^{\ast \ast }}$ denote a matrix obtained by changing the signs of entries in all even rows, where ∗ and $\ast \ast $ are two matrix operators. Note that ${(-\mathbf{A})^{\ast \ast }}={\mathbf{A}^{\ast }}$. An $n\times n$ Hadamard matrix H is a matrix with entries in $\{-1,+1\}$ and ${\mathbf{H}^{\top }}\mathbf{H}=n\mathbf{I}$, where n can only be 1, 2 or a multiple of 4.

2.2 Construction Steps

Our construction method expands a $6m\times p$ supersaturated design ${\mathbf{D}_{0}}$ with $\mu \le 1/3$ to a $12m\times 4p$ design with $\mu \le 1/3$, where $m,p$ are positive integers with $6m\lt p$. Partition ${\mathbf{D}_{0}}$ as
(2.1)
\[ {\mathbf{D}_{0}}=\left[\begin{array}{r}\mathbf{U}\\ {} \mathbf{L}\end{array}\right],\]
where U and L have $4m$ and $2m$ rows, respectively. The proposed method has three steps.
  • • Step 1: Use two copies of ${\mathbf{D}_{0}}$ to obtain a matrix
    nejsds34_g001.jpg
  • • Step 2: Expand ${\mathbf{D}_{1}}$ to obtain a matrix
    nejsds34_g002.jpg
  • • Step 3: Expand ${\mathbf{D}_{2}}$ to obtain a matrix
    nejsds34_g003.jpg
If ${\mathbf{D}_{1}}$ is viewed as a block matrix with four rows, Step 2 applies the operator ∗ to the second row and changes the signs of all entries in the fourth row to obtain ${\mathbf{D}_{2}}$. If ${\mathbf{D}_{2}}$ is viewed as a block matrix with four rows and two columns, Step 3 applies the operator $\ast \ast $ to the first and fourth rows and applies the operator ∗ to the third row to obtain ${\mathbf{D}_{3}}$. It can be proved that the coherence of ${\mathbf{D}_{1}}$, ${\mathbf{D}_{2}}$ and ${\mathbf{D}_{3}}$ is no greater than $1/3$ by Lemmas 1, 2, 3 and Theorem 1 in Appendix A.
Since the design ${\mathbf{D}_{3}}$ from the construction is a $12m\times 4p$ design with $\mu \le 1/3$, repeating the above procedure multiple times with ${\mathbf{D}_{0}}$ being ${\mathbf{D}_{3}}$ produces a class of designs with $({2^{k}}\times 6m)$ rows, $({4^{k}}\times p)$ columns and $\mu \le 1/3$ for every positive integer k.
The only requirement for ${\mathbf{D}_{0}}$ is to be a $6m\times p$ two-level supersaturated design with $\mu \le 1/3$. For a given pair of m and p, different choices of ${\mathbf{D}_{0}}$ yield different forms of ${\mathbf{D}_{3}}$ but all of them have guaranteed small coherence. As suggested by Chen and Lin [6] and Liu, Ruan and Dean [20], a supersaturated design with coherence of $1/3$ can identify most of the important factors.

2.3 Examples

We now provide several examples for the proposed method with different choices of ${\mathbf{D}_{0}}$. In the following examples, we show the first several designs constructed by the proposed method and choose ${\mathbf{D}_{0}}$ to have as many columns as possible.
Example 1.
Let ${\mathbf{D}_{0}}$ be a $6\times 16$ design given by
\[ \left[\begin{array}{r@{\hskip10.0pt}r@{\hskip10.0pt}r@{\hskip10.0pt}r@{\hskip10.0pt}r@{\hskip10.0pt}r@{\hskip10.0pt}r@{\hskip10.0pt}r@{\hskip10.0pt}r@{\hskip10.0pt}r@{\hskip10.0pt}r@{\hskip10.0pt}r@{\hskip10.0pt}r@{\hskip10.0pt}r@{\hskip10.0pt}r@{\hskip10.0pt}r}+& +& +& +& +& +& +& +& +& +& +& +& +& +& +& +\\ {} +& +& +& +& -& -& -& -& -& -& -& -& +& +& +& +\\ {} +& +& -& -& +& +& -& -& +& +& -& -& +& +& -& -\\ {} +& +& -& -& +& +& -& -& -& -& +& +& -& -& +& +\\ {} +& -& +& -& +& -& +& -& +& -& +& -& +& -& +& -\\ {} +& -& +& -& -& +& -& +& +& -& +& -& -& +& -& +\end{array}\right],\]
where − and + denote $-1$ and $+1$, respectively, and its coherence is $1/3$. The coherence of the above design attains the lower bound derived in [31]. The sizes, coherence values and numbers of the balanced columns of the first several designs constructed by the proposed method are given in Table 1.
Table 1
Designs Constructed from a $6\times 16$ ${\mathbf{D}_{0}}$.
Size μ #Balance
$12\times 64\phantom{xx}$ 1/3 24
$24\times 256\phantom{x}$ 1/3 168
$48\times 1024$ 1/3 840
$96\times 4096$ 1/3 3720
$192\times 16384$ 1/3 15624
Example 2.
Let nejsds34_g004.jpg
where ${\mathbf{D}_{01}}$ is the $24\times 23$ Plackett and Burman design [23], ${\mathbf{D}_{02}}$ is formed by taking the two-order interaction terms of ${\mathbf{D}_{01}}$ and ${\mathbf{D}_{03}}$ is formed by taking the three-order interaction terms of ${\mathbf{D}_{01}}$. Then ${\mathbf{D}_{0}}$ is a $24\times 2048$ design with $\mu =1/3$. The first 277 columns of this design form the design in [32]. The sizes, coherence values and numbers of the balanced columns of the first several designs constructed by the proposed method are given in Table 2.
Table 2
Designs Constructed from a $24\times 2048$ ${\mathbf{D}_{0}}$.
Size μ #Balance
$48\times 8192\phantom{x}$ 1/3 2968
$96\times 32768$ 1/3 18616
$192\times 131072$ 1/3 99064
Example 3.
Let ${\mathbf{D}_{0}}$ be the $30\times 59$ design from Lin [18], where an additional column ${\mathbf{1}_{30}}$ is added as the first column. This design has $\mu =0.2$. The sizes, coherence values and numbers of the balanced columns of the first several designs constructed by the proposed method are given in Table 3. This example illustrates that our method can use a design ${\mathbf{D}_{0}}$ with $\mu \lt 1/3$.
Table 3
Designs Constructed from a $30\times 59$ ${\mathbf{D}_{0}}$.
Size μ #Balance
$60\times 236\phantom{x}$ 1/3 101
$120\times 944\phantom{xx}$ 1/3 550
$240\times 3776\phantom{x}$ 1/3 2864
$480\times 15104$ 1/3 13156
$960\times 60416$ 1/3 56396
Whenever necessary, one can select some columns of the designs constructed for use. Selecting a subset of columns could remain coherence the same or further decrease it, but will never increase coherence. For example, for a given p, first, one can select as many balanced columns as possible. If there are no sufficient balanced columns, then one can select some unbalanced columns to achieve p columns. In addition, one may also select the columns according to an additional criterion such as $E({s^{2}})$.

2.4 Generalization

We now generalize the proposed method to obtain a supersaturated design with μ smaller than $1/3$. Recall that our method partitions ${\mathbf{D}_{0}}$ into two parts. The upper part U has $6mr$ rows and the lower part L has $6m(1-r)$ rows with a partition ratio $r=2/3$. We generalize the original construction by using a different partition ratio and stopping at Step 2. Suppose that ${\mathbf{D}_{0}}$ is an $n\times p$ two-level supersaturated design with $\mu \le t/n$ for $t\in \{2,4,\dots ,n/2\}$ and an even n. Partition ${\mathbf{D}_{0}}$ as in (2.1), where U has $2t$ rows and L has $n-2t$ rows with $r=2t/n$. Define ${\mathbf{D}_{2}}$ as in (2.2). Then ${\mathbf{D}_{2}}$ is the supersaturated design constructed by our generalized method. It can be proved that the coherence of ${\mathbf{D}_{2}}$ constructed here is no greater than $t/n$ by Theorem 2 in Appendix A.
The generalization indicates the possibility of expanding any two-level supersaturated design with $\mu \le 1/2$ and an even number of runs while retaining its coherence. For the original construction, coherence must be no greater than $1/3$ and the initial design must have $6m$ runs but the initial design with p columns can be expanded to a design with $4p$ columns and $\mu \le 1/3$. The generalization only requires that the coherence is no greater than $1/2$ and the initial design has an even number of runs but the initial design with p columns can only be expanded to a design with $2p$ columns and the same coherence. In addition, the number of balanced columns can be fewer than the number of runs for the designs constructed by the generalization. Here is an example.
Example 4.
Let ${\mathbf{D}_{0}}$ be the $30\times 59$ design in Example 3 with $\mu =0.2$. The sizes, coherence values and numbers of the balanced columns of the first several designs constructed by the generalization are given in Table 4.
Table 4
Designs Constructed from a $30\times 59$ ${\mathbf{D}_{0}}$ by the Generalization.
Size μ #Balance
$60\times 118\phantom{}$ 0.2 76
$120\times 236\phantom{..}$ 0.2 112
$240\times 472\phantom{x}$ 0.2 184
$\phantom{.}480\times 944\phantom{x.}$ 0.2 328
$960\times 1888$ 0.2 616

3 Simulation Study

In this section, we compare the proposed designs with four popular classes of supersaturated designs: Lin’s designs [18], Wu’s designs [32], the Bayesian D-optimal supersaturated designs (Jones, Lin, and Nachtsheim 2008) and the $UE({s^{2}})$-optimal designs [15] for the Lasso problem by simulations. These designs are denoted by “LIN”, “WU”, “BAYES” and “JM” respectively and our proposed designs are denoted by “Proposed”.
Lin’s designs are constructed from the Plackett and Burman design [23]. According to Bulutoglu and Cheng [4], Lin’s designs are $E({s^{2}})$-optimal among all the balanced two-level supersaturated designs. Wu’s designs can be obtained by appending interaction columns to a Hadamard matrix. Bulutoglu and Cheng [4] showed that in certain cases, Wu’s designs are $E({s^{2}})$-optimal among all the balanced two-level supersaturated designs. The Bayesian D-optimal supersaturated design is obtained by the coordinate exchange algorithm [9, 14]. Any n rows of a $p\times p$ Hadamard matrix form a type of $UE({s^{2}})$-optimal design [15, 5].
We simulate data from the linear model in (1.1) with different designs and active coefficients, where the error vector $\boldsymbol{\epsilon }$ follows the standard multivariate normal distribution ${N_{n}}(\mathbf{0},\mathbf{I})$. We use the Lasso to select the active factors and fix the intercept term ${\beta _{1}}$ as active. We repeat the above data generation and variable selection procedure $N=300$ times. We use the “cv.glmnet” function [12] in the R software [24] and calculate $\hat{\boldsymbol{\beta }}$ with “s=lambda.min”.
We conduct simulations using the eight active coefficients settings in Table 5 and denote them as Case 1, $\dots \hspace{0.1667em}$, Case 8, respectively. We use “#Active” to denote the number of active factors. For each group of active coefficients, we consider both the case where they take large values, and the case where they take small values. For example, for ${\beta _{1}}$, ${\beta _{21}}$ and ${\beta _{23}}$, we consider both the case where they take values of 1, 5 and 10, and the case where they take values of 0.1, 1 and 1.5. We let the regression coefficients of all inactive factors be zero and do not write them in Table 5. We neither use a single setting of active coefficients for all the comparisons, nor use all settings in Table 5 for each comparison, because of the following reasons. First, the designs under consideration have different numbers of columns and some designs do not have enough columns for a given setting, which makes the setting not applicable. Second, in practice, the more factors we are considering, the more active factors there tends to be, so we use settings with more active coefficients for designs with large numbers of columns, and settings with fewer active coefficients for designs with small numbers of columns.
Table 5
Active Coefficients Settings.
Case #Active Active Coefficients and Their Values
1 2 ${\beta _{1\phantom{00}}}=1\phantom{.0}\phantom{x}{\beta _{21\phantom{0}}}=5\phantom{.0}\phantom{x}{\beta _{23\phantom{0}}}=10\phantom{.}$
2 2 ${\beta _{1\phantom{00}}}=0.1\phantom{}\phantom{x}{\beta _{21\phantom{0}}}=1\phantom{.0}\phantom{x}{\beta _{23\phantom{0}}}=1.5\phantom{}$
3 4 ${\beta _{1\phantom{00}}}=1\phantom{.0}\phantom{x}{\beta _{6\phantom{00}}}=5\phantom{.0}\phantom{x}{\beta _{8\phantom{00}}}=10\phantom{.}\phantom{x}{\beta _{9\phantom{00}}}=10\phantom{.}\phantom{x}{\beta _{26\phantom{0}}}=15\phantom{.}$
4 4 ${\beta _{1\phantom{00}}}=0.1\phantom{}\phantom{x}{\beta _{6\phantom{00}}}=0.6\phantom{}\phantom{x}{\beta _{8\phantom{00}}}=1.7\phantom{}\phantom{x}{\beta _{9\phantom{00}}}=0.9\phantom{}\phantom{x}{\beta _{26\phantom{0}}}=1\phantom{.0}$
5 4 ${\beta _{1\phantom{00}}}=1\phantom{.0}\phantom{x}{\beta _{19\phantom{0}}}=5\phantom{.0}\phantom{x}{\beta _{59\phantom{0}}}=14\phantom{.}\phantom{x}{\beta _{143\phantom{}}}=10\phantom{.}\phantom{x}{\beta _{214\phantom{}}}=13\phantom{.}$
6 4 ${\beta _{1\phantom{00}}}=0.1\phantom{}\phantom{x}{\beta _{19\phantom{0}}}=1.3\phantom{}\phantom{x}{\beta _{59\phantom{0}}}=0.9\phantom{}\phantom{x}{\beta _{143\phantom{}}}=1\phantom{.0}\phantom{x}{\beta _{214\phantom{}}}=1.2\phantom{}$
7 6 ${\beta _{1\phantom{00}}}=1\phantom{.0}\phantom{x}{\beta _{38\phantom{0}}}=5\phantom{.0}\phantom{x}{\beta _{111\phantom{}}}=5\phantom{.0}\phantom{x}{\beta _{122\phantom{}}}=9\phantom{.0}\phantom{x}{\beta _{123\phantom{}}}=9\phantom{.0}\phantom{x}{\beta _{147\phantom{}}}=9\phantom{.0}\phantom{x}{\beta _{151\phantom{}}}=11\phantom{.}$
8 6 ${\beta _{1\phantom{00}}}=0.1\phantom{}\phantom{x}{\beta _{38\phantom{0}}}=1.5\phantom{}\phantom{x}{\beta _{111\phantom{}}}=1.5\phantom{}\phantom{x}{\beta _{122\phantom{}}}=1.9\phantom{}\phantom{x}{\beta _{123\phantom{}}}=1.9\phantom{}\phantom{x}{\beta _{147\phantom{}}}=1.9\phantom{}\phantom{x}{\beta _{151\phantom{}}}=2.8\phantom{}$
Table 6
Comparison with Lin’s Design.
Case #Active Design Size μ $E({s^{2}})$ #Balance AFDR AMR MSE EME
1 2 Proposed $12\times 27$ 0.33 9.70 24 0.49 0.00 0.64 7.22
LIN $14\times 27$ 0.43 7.84 26 0.59 0.00 0.93 9.18
2 2 Proposed $12\times 27$ 0.33 9.70 24 0.51 0.02 0.74 9.15
LIN $14\times 27$ 0.43 7.84 26 0.59 0.08 1.13 12.51
3 4 Proposed $12\times 27$ 0.33 9.70 24 0.25 0.18 73.58 537.69
LIN $14\times 27$ 0.43 7.84 26 0.40 0.28 230.66 794.73
4 4 Proposed $12\times 27$ 0.33 9.70 24 0.49 0.29 2.15 18.52
LIN $14\times 27$ 0.43 7.84 26 0.64 0.53 4.19 19.79
For each design in our simulations, we will show its size, coherence, and number of balanced columns. Although $E({s^{2}})$ is not our focused criterion, we will also show the $E({s^{2}})$ of each design for reference. We use the following four criteria to compare variable selection and model fitting accuracy of each design:
  • 1. Average False Discovery Rate (AFDR) ${\textstyle\sum _{i=1}^{N}}{\text{FDR}_{i}}\big/N$, where
    \[ {\text{FDR}_{i}}=\frac{\text{the number of falsely discovered coefficients}}{\text{the number of discovered coefficients}}\]
    if the discovered model is not a null model, and ${\text{FDR}_{i}}=1$ if the discovered model is a null model.
  • 2. Average Miss Rate (AMR) ${\textstyle\sum _{i=1}^{N}}{\text{MR}_{i}}\big/N$, where
    \[ {\text{MR}_{i}}=\frac{\text{the number of undiscovered active coefficients}}{\text{the number of active coefficients}}.\]
  • 3. Mean Squared Error (MSE) ${\textstyle\sum _{i=1}^{N}}\| \boldsymbol{\beta }-{\hat{\boldsymbol{\beta }}^{(i)}}{\| ^{2}}\big/N$ to estimate $E(\| \boldsymbol{\beta }-\hat{\boldsymbol{\beta }}{\| ^{2}})$.
  • 4. Expected Model Error (EME) ${\textstyle\sum _{i=1}^{N}}\| \mathbf{X}\boldsymbol{\beta }-\mathbf{X}{\hat{\boldsymbol{\beta }}^{(i)}}{\| ^{2}}\big/N$ to estimate $E(\| \mathbf{X}\boldsymbol{\beta }-\mathbf{X}\hat{\boldsymbol{\beta }}{\| ^{2}})$.
Smaller values of AFDR, AMR, MSE and EME are desirable. Note that AFDR and AMR are computed based on the number of discovered or active coefficients, not the number of discovered or active factors.

3.1 Comparison with Lin’s Design

We obtain the $12\times 27$ proposed design by selecting the intercept, the 24 balanced columns and the first two unbalanced columns of the $12\times 64$ design in Example 1. We use Lin’s $14\times 27$ design directly. The result is shown in Table 6, where and in other tables below, the smaller values of the four criteria are in boldface. Table 6 indicates that the proposed design outperforms Lin’s design in terms of variable selection and parameter estimation with the Lasso. In addition, the proposed design has fewer runs than Lin’s design.
To verify that this result is not due to some better property of the selected active factors in one design versus another, we also conducted a follow-up comparison with randomly selected active factors. More specifically, for a given setting in Table 5, we randomly select an #Active number of active factors with the same values given in the setting, and repeat 10 times to get 10 sets of active factors. Then, for each set of randomly selected active factors, we repeat the above data generation and variable selection procedure $N=30$ times for the proposed design and Lin’s design. The final criteria are averaged over the criteria from the 10 sets of active factors, and they can be found in Table 7. According to Table 7, the proposed design still generally outperforms Lin’s design, even with randomly selected active factors. To save computational cost and keep our paper concise, we will omit such a verification for the following comparisons.
Table 7
Comparison with Lin’s Design with Randomly Selected Active Factors.
Case #Active Design Size μ $E({s^{2}})$ #Balance AFDR AMR MSE EME
1 2 Proposed $12\times 27$ 0.33 9.70 24 0.61 0.00 5.01 9.02
LIN $14\times 27$ 0.43 7.84 26 0.61 0.00 1.59 11.58
2 2 Proposed $12\times 27$ 0.33 9.70 24 0.58 0.09 1.34 10.83
LIN $14\times 27$ 0.43 7.84 26 0.59 0.14 1.36 12.93
3 4 Proposed $12\times 27$ 0.33 9.70 24 0.47 0.00 51.62 27.23
LIN $14\times 27$ 0.43 7.84 26 0.41 0.04 33.89 135.42
4 4 Proposed $12\times 27$ 0.33 9.70 24 0.55 0.15 2.11 13.23
LIN $14\times 27$ 0.43 7.84 26 0.45 0.27 2.13 18.56
Table 8
Comparison with Wu’s Designs.
Case #Active Design Size μ $E({s^{2}})$ #Balance AFDR AMR MSE EME
1 2 Proposed $12\times 64$ 0.33 9.90 24 0.65 0.00 0.95 8.41
WU $12\times 64$ 0.33 11.06 63 0.54 0.00 0.89 9.03
2 2 Proposed $12\times 64$ 0.33 9.90 24 0.66 0.05 1.20 12.46
WU $12\times 64$ 0.33 11.06 63 0.56 0.04 0.98 11.22
__________________ _____________________ _ ________________________ ________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________ ________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________ ________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________ ________________________ ___________________ _________________ ________________ _________________
3 4 Proposed $24\times 256$ 0.33 21.84 168 0.41 0.00 0.96 19.96
WU $24\times 256$ 0.33 23.00 255 0.65 0.00 2.08 23.25
4 4 Proposed $24\times 256$ 0.33 21.84 168 0.67 0.04 0.94 18.20
WU $24\times 256$ 0.33 23.00 255 0.68 0.14 1.76 24.90
__________________ _______________________ __________________________ __________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________ __________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________ __________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________ __________________________ _____________________ ___________________ __________________ ___________________
7 6 Proposed $48\times 300$ 0.33 93.05 299 0.62 0.00 1.76 27.15
WU $48\times 300$ 0.33 41.99 299 0.75 0.00 2.52 37.94
8 6 Proposed $48\times 300$ 0.33 93.05 299 0.71 0.00 1.74 21.65
WU $48\times 300$ 0.33 41.99 299 0.77 0.01 2.58 44.09

3.2 Comparison with Wu’s Designs

We use the $12\times 64$, $24\times 256$ proposed designs in Example 1 directly. We obtain the $48\times 300$ proposed design by selecting the intercept and the first 299 balanced columns of the $48\times 8192$ design in Example 2. We obtain Wu’s $12\times 64$, $24\times 256$ and $48\times 300$ designs by selecting the first 64 columns of Wu’s $12\times 67$ design, the first 256 columns of Wu’s $24\times 277$ design and the first 300 columns of Wu’s $48\times 1129$ design, respectively. The result shown in Table 8 indicates that the proposed designs are generally better than Wu’s designs in terms of variable selection and parameter estimation with the Lasso. In almost all cases, the proposed designs have better performance while in the first two cases, Wu’s designs perform relatively better.

3.3 Comparison with the Bayesian D-Optimal Supersaturated Designs

We use the $12\times 64$ proposed design in Example 1 directly. We obtain the $24\times 253$ proposed design by selecting the intercept, the 168 balanced columns and the first 84 unbalanced columns of the $24\times 256$ design in Example 1. We first search a $12\times 64$ and a $24\times 256$ Bayesian D-optimal supersaturated design by the JMPő software [13] with 500 random starting designs. We use the $12\times 64$ design directly. We remove the 22th, 134th and 209th columns of the $24\times 256$ design to obtain the $24\times 253$ design used in the simulations since the original searched $24\times 256$ Bayesian D-optimal supersaturated design has $\mu =0.83$. The result shown in Table 9 indicates that the proposed designs generally outperform the Bayesian D-optimal supersaturated designs in terms of variable selection and parameter estimation with the Lasso.
Table 9
Comparison with the Bayesian D-Optimal Supersaturated Designs.
Case #Active Design Size μ $E({s^{2}})$ #Balance AFDR AMR MSE EME
1 2 Proposed $12\times 64$ 0.33 9.90 24 0.65 0.00 0.95 8.41
BAYES $12\times 64$ 0.67 11.04 63 0.76 0.06 45.02 76.31
2 2 Proposed $12\times 64$ 0.33 9.90 24 0.66 0.05 1.20 12.46
BAYES $12\times 64$ 0.67 11.04 63 0.78 0.45 3.20 16.31
3 4 Proposed $12\times 64$ 0.33 9.90 24 0.30 0.00 1.83 18.04
BAYES $12\times 64$ 0.67 11.04 63 0.66 0.16 240.96 535.21
4 4 Proposed $12\times 64$ 0.33 9.90 24 0.65 0.06 1.48 12.77
BAYES $12\times 64$ 0.67 11.04 63 0.63 0.46 4.61 30.40
__________________ _______________________ __________________________ __________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________ __________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________ __________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________ __________________________ _____________________ ___________________ _____________________ _____________________
5 4 Proposed $24\times 253$ 0.33 21.84 168 0.60 0.00 1.61 21.59
BAYES $24\times 253$ 0.67 22.91 252 0.67 0.00 3.62 27.99
6 4 Proposed $24\times 253$ 0.33 21.84 168 0.70 0.05 1.85 29.70
BAYES $24\times 253$ 0.67 22.91 252 0.61 0.28 3.36 50.61
7 6 Proposed $24\times 253$ 0.33 21.84 168 0.54 0.00 4.24 38.37
BAYES $24\times 253$ 0.67 22.91 252 0.67 0.28 162.70 934.85
8 6 Proposed $24\times 253$ 0.33 21.84 168 0.72 0.01 3.49 25.34
BAYES $24\times 253$ 0.67 22.91 252 0.70 0.44 16.57 158.04
Table 10
Comparison with the $UE({s^{2}})$-Optimal Designs.
Case #Active Design Size μ $E({s^{2}})$ #Balance MAFDR MAMR MMSE MEME
1 2 Proposed $12\times 64$ 0.33 9.90 24 0.67 0.00 0.87 7.99
JM $12\times 64$ 0.83 9.90 16 0.61 0.00 2.53 11.02
2 2 Proposed $12\times 64$ 0.33 9.90 24 0.67 0.03 1.04 10.38
JM $12\times 64$ 0.83 9.90 16 0.66 0.26 2.35 15.26
3 4 Proposed $12\times 64$ 0.33 9.90 24 0.30 0.00 2.03 21.28
JM $12\times 64$ 0.83 9.90 16 0.62 0.45 351.16 956.38
4 4 Proposed $12\times 64$ 0.33 9.90 24 0.59 0.09 1.44 13.18
JM $12\times 64$ 0.83 9.90 16 0.66 0.55 4.47 24.95
__________________ _______________________ __________________________ __________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________ __________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________ __________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________ __________________________ _________________________ _______________________ ______________________ ________________________
5 4 Proposed $24\times 256$ 0.33 21.84 168 0.58 0.00 1.68 26.33
JM $24\times 256$ 0.58 21.84 44 0.69 0.00 66.86 61.37
6 4 Proposed $24\times 256$ 0.33 21.84 168 0.72 0.02 1.42 20.00
JM $24\times 256$ 0.58 21.84 44 0.73 0.31 3.73 41.41
7 6 Proposed $24\times 256$ 0.33 21.84 168 0.71 0.30 260.42 398.54
JM $24\times 256$ 0.58 21.84 44 0.72 0.32 294.96 1336.26
8 6 Proposed $24\times 256$ 0.33 21.84 168 0.62 0.35 13.62 110.92
JM $24\times 256$ 0.58 21.84 44 0.72 0.43 18.54 147.44

3.4 Comparison with the $UE({s^{2}})$-Optimal Designs

According to Jones and Majumdar [15], for given n and p, there are a large number of $UE({s^{2}})$-optimal designs. In practical scenarios, people may randomly generate only one $UE({s^{2}})$-optimal design and use it to screen important factors. In our comparison, we want to mimic such a scenario, i.e., we want to know whether people will have a higher chance ($\gt 0.5$) of getting a worse design (compared with our design) if they generate a $UE({s^{2}})$-optimal design in this way. Therefore, we randomly generate 500 $UE({s^{2}})$-optimal designs and for each of them, repeat the simulation procedure $N=30$ times (rather than $N=300$ in comparisons with other competing designs). We take the median values for the four criteria introduced in the beginning of the section over the 500 randomly generated designs and compare these median values with the corresponding criteria of the proposed design. If the AFDR of a proposed design is less than the median AFDR (MAFDR) of all the 500 $UE({s^{2}})$-optimal designs, it means the AFDR of the proposed design is less than that of the $UE({s^{2}})$-optimal design with more than half probability. The coherence, $E({s^{2}})$ and number of balanced columns in the table are also the corresponding medians. In addition, it is possible that a generated $UE({s^{2}})$-optimal design has $\mu =1$ and we will avoid using these designs in the simulations, because in practice if people get a bad design with $\mu =1$, they will probably regenerate another one.
We use the $12\times 64$ and $24\times 256$ proposed designs in Example 1 directly. We obtain the $UE({s^{2}})$-optimal designs by randomly sampling 12 rows of a $64\times 64$ Hadamard matrix 500 times and 24 rows of a $256\times 256$ Hadamard matrix 500 times, respectively. The result shown in Table 10 indicates that with more than half probability, the proposed designs outperform the $UE({s^{2}})$-optimal designs in terms of variable selection and parameter estimation with the Lasso. Furthermore, one may want to see the distributions of the criteria of 500 $UE({s^{2}})$-optimal designs. We take Case 8 for an example and show the distributions of the criteria in Figure 1.
nejsds34_g005.jpg
Figure 1
Comparison between the $UE({s^{2}})$-optimal design and the proposed design in Case 8. The solid curves are the kernel density estimates of the probability density functions of $UE({s^{2}})$-optimal design’s criteria and the solid lines are the corresponding median values. The dashed lines are the values of the proposed design’s criteria.

3.5 Summary of Comparisons

To help people better understand how the proposed design performs compared to competing designs with the same size and under the same active coefficient setting, we collect the comparisons between designs with exactly the same size and active coefficient setting all together in Table 11. Lin’s design is not of size $12\times 64$, so it is not included in the table (it is actually impossible to construct a design of this size with Lin’s method).
Table 11
Comparison Between Designs with the Same Size and Active Coefficients.
Case #Active Design Size μ $E({s^{2}})$ #Balance AFDR AMR MSE EME
1 2 Proposed $12\times 64$ 0.33 9.90 24 0.65 0.00 0.95 8.41
WU $12\times 64$ 0.33 11.06 63 0.54 0.00 0.89 9.03
BAYES $12\times 64$ 0.67 11.04 63 0.76 0.06 45.02 76.31
JM $12\times 64$ 0.83 9.90 16 0.61 0.00 2.53 11.02
2 2 Proposed $12\times 64$ 0.33 9.90 24 0.66 0.05 1.20 12.46
WU $12\times 64$ 0.33 11.06 63 0.56 0.04 0.98 11.22
BAYES $12\times 64$ 0.67 11.04 63 0.78 0.45 3.20 16.31
JM $12\times 64$ 0.83 9.90 16 0.66 0.26 2.35 15.26
According to Table 11, the proposed design significantly outperforms the Bayesian D-optimal design and the $UE({s^{2}})$-optimal design. In this scenario, Wu’s design performs slightly better than the proposed design, but there are also scenarios where the proposed design performs better than Wu’s design (see Table 8).
In summary, for Wu’s design, the proposed design is generally comparable to it, because of the same coherence they have. For Lin’s design, the Bayesian D-optimal design and the $UE({s^{2}})$-optimal design, the proposed design generally outperforms them, except that it sometimes tends to select more factors and gives a larger AFDR. However, the proposed design always gives a smaller AMR when compared to these three designs. In practice, a small AMR is more important than a small AFDR, because missing active factors is a more serious problem than falsely discovering inactive factors. This is because we can conduct follow-up experiments to screen out spurious factors, but we can never detect active factors that were already removed in the screening phase [21].

4 Conclusion and Discussion

We have proposed a method for constructing supersaturated designs with small coherence. The constructed designs are allowed to be unbalanced to achieve more flexible sample sizes. The proposed method uses direct constructions and it entails no intensive computing even for large p. Since the proposed method can efficiently construct a supersaturated design with a large number of columns, it can be applied to high-dimensional variable selection problems in marketing, biology, engineering and other areas.
Here are possible directions for future work. First, the proposed method can expand a $6m\times p$ design ${\mathbf{D}_{0}}$ with $\mu \le 1/3$ to a larger design with $\mu \le 1/3$. One may be interested in finding a special ${\mathbf{D}_{0}}$ such that the expanded design has coherence strictly smaller than $1/3$ or using a different construction method to reduce the coherence upper bound. Second, beyond the Lasso, the proposed design can be applicable to other variable selection and high-dimensional problems where controlling coherence of the design matrix is important. Moreover, many new powerful variable selection techniques have emerged. For example, Fan and Lv [10] and Fan, Feng and Song [11] proposed the sure independence screening and nonparametric independence screening for ultrahigh dimensional variable selection. Another example is the best subset regression. Shen, Pan, Zhu and Zhou [25] provided theoretical support for the use of best subset regression. Bertsimas, King and Mazumder [1] proposed a mixed integer optimization approach for the best subset selection problem, where they showed that coherence plays a role in parameter specification of their approach. In future work, our proposed designs will be applied to other penalized variable selection methods with results reported elsewhere.

Appendix A Proof for the Upper Bound of Coherence

Lemma 1.
For any matrix with levels − and +, if the signs of all the entries in a row are changed simultaneously, then the dot product of every two columns remains the same. Thus, the coherence of the matrix remains the same.
Lemma 2.
The coherence of ${\mathbf{D}_{1}}$ constructed in Section 2.2 is no greater than $1/3$.
Proof.
Throughout, for two columns ${\boldsymbol{d}_{i}}$ and ${\boldsymbol{d}_{j}}$ of a two-level design D with n rows, we call $|{\boldsymbol{d}_{i}^{\top }}{\boldsymbol{d}_{j}}|/n$ their absolute column correlation. Since ${\mathbf{D}_{1}}$ is obtained by two copies of ${\mathbf{D}_{0}}$, the absolute column correlations of any two columns of ${\mathbf{D}_{1}}$ are the same as those of ${\mathbf{D}_{0}}$. Thus, the coherence of ${\mathbf{D}_{1}}$ equals that of ${\mathbf{D}_{0}}$, which is no greater than $1/3$.  □
Lemma 3.
The coherence of ${\mathbf{D}_{2}}$ constructed in Section 2.2 is no greater than $1/3$.
Proof.
Let ${\boldsymbol{d}^{(i)}}$ denote the ith column of ${\mathbf{D}_{2}}$ and ${d_{j}^{(i)}}$ denote the jth entry of ${\boldsymbol{d}^{(i)}}$. Put nejsds34_g006.jpg
where ${\mathbf{C}_{1}}$ and ${\mathbf{C}_{2}}$ are $12m\times p$ matrices.
Pick two arbitrary columns of ${\mathbf{D}_{2}}$.
  • (i) Suppose that both of them are from ${\mathbf{C}_{1}}$ or ${\mathbf{C}_{2}}$. By Lemma 2, $\mu ({\mathbf{C}_{1}})$ is ≤ $1/3$. By Lemmas 1 and 2, $\mu ({\mathbf{C}_{2}})$ is ≤ $1/3$. Thus, the absolute column correlation of the two columns is ≤ $1/3$.
  • (ii) Suppose that one of the two columns is the ith column ${\boldsymbol{d}^{(i)}}$ of ${\mathbf{C}_{1}}$ and the other column is the jth column ${\boldsymbol{d}^{(p+j)}}$ of ${\mathbf{C}_{2}}$ with $1\le i,j\le p$. Partition the two columns as
    \[ \left[\begin{array}{c@{\hskip10.0pt}c}{\boldsymbol{d}^{(i)}}& {\boldsymbol{d}^{(p+j)}}\end{array}\right]=\left[\begin{array}{c}\mathbf{A}\\ {} \mathbf{B}\end{array}\right],\]
    where A and B have $8m$ and $4m$ rows, respectively. By Step 2 in Section 2.2,
    nejsds34_g007.jpg Note that A consists of $2m$ blocks with 4 rows and 2 columns
    \[ \left[\begin{array}{c@{\hskip10.0pt}c}\phantom{-}{d_{2l-1}^{(i)}}& \phantom{-}{d_{2l-1}^{(j)}}\\ {} {d_{2l}^{(i)}}& {d_{2l}^{(j)}}\\ {} \phantom{-}{d_{2l-1}^{(i)}}& -{d_{2l-1}^{(j)}}\\ {} {d_{2l}^{(i)}}& {d_{2l}^{(j)}}\end{array}\right],\hspace{1em}l=1,\dots ,2m.\]
    Since the absolute dot products of the two columns in these blocks are ≤ 2, the absolute dot product of the two columns in A is ≤ $4m$. Similarly, B consists of m blocks with 4 rows and 2 columns
    \[ \left[\begin{array}{c@{\hskip10.0pt}c}\phantom{-}{d_{8m+2l-1}^{(i)}}& \phantom{-}{d_{8m+2l-1}^{(j)}}\\ {} {d_{8m+2l}^{(i)}}& \phantom{-}{d_{8m+2l}^{(j)}}\phantom{-1}\\ {} \phantom{-}{d_{8m+2l-1}^{(i)}}& -{d_{8m+2l-1}^{(j)}}\\ {} {d_{8m+2l}^{(i)}}& -{d_{8m+2l}^{(j)}}\phantom{-1}\end{array}\right],\hspace{1em}l=1,\dots ,m.\]
    Since the absolute dot products of the two columns in these blocks are 0, the absolute dot product of the two columns in B is 0. Thus, the absolute dot product of ${\boldsymbol{d}^{(i)}}$ and ${\boldsymbol{d}^{(p+j)}}$ is ≤ $4m$ and their absolute column correlation is ≤ $4m/12m=1/3$.
Combining $(\mathrm{i})$ and $(\mathrm{ii})$ proves that $\mu ({\mathbf{D}_{2}})$ is ≤ $1/3$.  □
Theorem 1.
The coherence of ${\mathbf{D}_{3}}$ constructed in Section 2.2 is no greater than $1/3$.
Proof.
Let ${\boldsymbol{d}^{(i)}}$ denote the ith column of ${\mathbf{D}_{3}}$ and ${d_{j}^{(i)}}$ denote the jth entry of ${\boldsymbol{d}^{(i)}}$. Put nejsds34_g008.jpg
where ${\mathbf{C}_{1}}$, ${\mathbf{C}_{2}}$, ${\mathbf{C}_{3}}$ and ${\mathbf{C}_{4}}$ are $12m\times p$ matrices.
Pick two arbitrary columns of ${\mathbf{D}_{3}}$.
  • (i) Suppose that both of them are from $[{\mathbf{C}_{1}},{\mathbf{C}_{2}}]$ or $[{\mathbf{C}_{3}},{\mathbf{C}_{4}}]$. By Lemma 3, the coherence of $[{\mathbf{C}_{1}},{\mathbf{C}_{2}}]$ is ≤ $1/3$. By Lemmas 1 and 3, the coherence of $[{\mathbf{C}_{3}},{\mathbf{C}_{4}}]$ is ≤ $1/3$. Thus, the absolute column correlation of these two columns is ≤ $1/3$.
  • (ii) Suppose that one of the two columns is the ith column ${\boldsymbol{d}^{(i)}}$ of ${\mathbf{C}_{1}}$ and the other column is the jth column ${\boldsymbol{d}^{(2p+j)}}$ of ${\mathbf{C}_{3}}$ with $1\le i,j\le p$. Partition these columns as
    \[ \left[\begin{array}{c@{\hskip10.0pt}c}{\boldsymbol{d}^{(i)}}& {\boldsymbol{d}^{(2p+j)}}\end{array}\right]=\left[\begin{array}{c}\mathbf{A}\\ {} \mathbf{B}\end{array}\right],\]
    where A and B have $8m$ and $4m$ rows, respectively. By Step 3 in Section 2.2,
    nejsds34_g009.jpg Note that A consists of $2m$ blocks with 4 rows and 2 columns
    \[ \left[\begin{array}{c@{\hskip10.0pt}c}\phantom{-}{d_{2l-1}^{(i)}}& \phantom{-}{d_{2l-1}^{(j)}}\\ {} {d_{2l}^{(i)}}& -{d_{2l}^{(j)}}\phantom{-}\\ {} \phantom{-}{d_{2l-1}^{(i)}}& \phantom{-}{d_{2l-1}^{(j)}}\\ {} {d_{2l}^{(i)}}& \phantom{-}{d_{2l}^{(j)}}\phantom{-}\end{array}\right],\hspace{1em}l=1,\dots ,2m.\]
    Since the absolute dot products of the two columns in these blocks are ≤ 2, the absolute dot product of the two columns in A is ≤ $4m$. Similarly, B consists of m blocks with 4 rows and 2 columns
    \[ \left[\begin{array}{c@{\hskip10.0pt}c}\phantom{-}{d_{8m+2l-1}^{(i)}}& -{d_{8m+2l-1}^{(j)}}\\ {} {d_{8m+2l}^{(i)}}& \phantom{-}{d_{8m+2l}^{(j)}}\phantom{-}\\ {} \phantom{-}{d_{8m+2l-1}^{(i)}}& \phantom{-.}{d_{8m+2l-1}^{(j)}}\\ {} {d_{8m+2l}^{(i)}}& -{d_{8m+2l}^{(j)}}\phantom{-}\end{array}\right],\hspace{1em}l=1,\dots ,m.\]
    Since the absolute dot products of the two columns in these blocks are 0, the absolute dot product of the two columns in B is 0. Thus, the absolute dot product of ${\boldsymbol{d}^{(i)}}$ and ${\boldsymbol{d}^{(2p+j)}}$ is ≤ $4m$ and their absolute column correlation is ≤ $4m/12m=1/3$.
  • (iii) Suppose that one of the two columns is the ith column ${\boldsymbol{d}^{(p+i)}}$ of ${\mathbf{C}_{2}}$ and the other column is the jth column ${\boldsymbol{d}^{(3p+j)}}$ of ${C_{4}}$ with $1\le i,j\le p$. This case can be proved similarly to $(\mathrm{ii})$.
  • (iv) Suppose that one of the two columns is the ith column ${\boldsymbol{d}^{(i)}}$ of ${\mathbf{C}_{1}}$ and the other column is the jth column ${\boldsymbol{d}^{(3p+j)}}$ of ${\mathbf{C}_{4}}$ with $1\le i,j\le p$. Partition these columns as
    \[ \left[\begin{array}{c@{\hskip10.0pt}c}{\boldsymbol{d}^{(i)}}& {\boldsymbol{d}^{(3p+j)}}\end{array}\right]=\left[\begin{array}{c}\mathbf{A}\\ {} \mathbf{B}\end{array}\right],\]
    where A and B have $8m$ and $4m$ rows, respectively. By Step 3 in Section 2.2,
    nejsds34_g010.jpg Note that A consists of $2m$ blocks with 4 rows and 2 columns
    \[ \left[\begin{array}{c@{\hskip10.0pt}c}\phantom{-}{d_{2l-1}^{(i)}}& \phantom{-}{d_{2l-1}^{(j)}}\\ {} {d_{2l}^{(i)}}& -{d_{2l}^{(j)}}\phantom{-}\\ {} \phantom{-}{d_{2l-1}^{(i)}}& -{d_{2l-1}^{(j)}}\\ {} {d_{2l}^{(i)}}& \phantom{-}{d_{2l}^{(j)}}\phantom{-}\end{array}\right],\hspace{1em}l=1,\dots ,2m.\]
    Since the absolute dot products of the two columns in these blocks are 0, the absolute dot product of the two columns in A is 0. Similarly, B consists of m blocks with 4 rows and 2 columns
    \[ \left[\begin{array}{c@{\hskip10.0pt}c}\phantom{-}{d_{8m+2l-1}^{(i)}}& -{d_{8m+2l-1}^{(j)}}\\ {} {d_{8m+2l}^{(i)}}& \phantom{-}{d_{8m+2l}^{(j)}}\phantom{-}\\ {} \phantom{-}{d_{8m+2l-1}^{(i)}}& -{d_{8m+2l-1}^{(j)}}\\ {} {d_{8m+2l}^{(i)}}& \phantom{-}{d_{8m+2l}^{(j)}}\phantom{-}\end{array}\right],\hspace{1em}l=1,\dots ,m.\]
    The absolute dot product of the two columns in the lth block is
    \[ 2|-{d_{8m+2l-1}^{(i)}}{d_{8m+2l-1}^{(j)}}+{d_{8m+2l}^{(i)}}{d_{8m+2l}^{(j)}}|,\]
    which is either 0 or 4. This implies that the absolute dot product of the two columns in B is ≤ $4m$. Thus, the absolute dot product of ${\boldsymbol{d}^{(i)}}$ and ${\boldsymbol{d}^{(3p+j)}}$ is ≤ $4m$ and their absolute column correlation is ≤ $4m/12m=1/3$.
  • (v) Suppose that one of the two columns is the ith column ${\boldsymbol{d}^{(p+i)}}$ of ${\mathbf{C}_{2}}$ and the other column is the jth column ${\boldsymbol{d}^{(2p+j)}}$ of ${\mathbf{C}_{3}}$ with $1\le i,j\le p$. This case can be proved similarly to $(\mathrm{iv})$.
Combining $(\mathrm{i})$, $(\mathrm{ii})$, $(\mathrm{iii})$, $(\mathrm{iv})$ and $(\mathrm{v})$ proves that $\mu ({\mathbf{D}_{3}})$ is ≤ $1/3$.  □
Theorem 2.
The coherence of ${\mathbf{D}_{2}}$ constructed in Section 2.4 is no greater than $t/n$.
Proof.
To prove this theorem, we use the same notation from the proof of Lemma 3, and pick two arbitrary columns of ${\mathbf{D}_{2}}$.
  • (i) Suppose that both of the two columns are from ${\mathbf{C}_{1}}$ or ${\mathbf{C}_{2}}$. By a similar argument to Lemma 2, the coherence of ${\mathbf{C}_{1}}$ is ≤ $t/n$. Further, by Lemma 1, the coherence of ${\mathbf{C}_{2}}$ is ≤ $t/n$. Thus, the absolute column correlation of these two columns is ≤ $t/n$.
  • (ii) Suppose that one of the two columns is from ${\mathbf{C}_{1}}$ and the other column is from ${\mathbf{C}_{2}}$. Since the matrices A and B have $4t$ and $2n$−$4t$ rows respectively, both of which are multiples of 4, a similar argument to Lemma 3 (ii) still works. Hence, the absolute column correlation of these two columns is ≤ $(4t/4\times 2)/2n=t/n$.
Combining (i) and ($\mathrm{ii}$) proves that the coherence of ${\mathbf{D}_{2}}$ is ≤ $t/n$.  □

Appendix B Coherence and Model Selection Consistency of the Lasso

Zhao and Yu [34] studied the model selection consistency of the Lasso, i.e., the ability of the Lasso to exactly identify all the active factors from a large number of factors as the sample size n gets large. Assuming the model in (1.1), they defined the Strong Irrepresentable Condition, which requires the existence of a positive constant vector $\boldsymbol{\eta }$ such that
\[ |{\mathbf{C}_{21}}{\mathbf{C}_{11}^{-1}}\text{sign}({\boldsymbol{\beta }_{(1)}})|\le \mathbf{1}-\boldsymbol{\eta },\]
where ${\mathbf{C}_{21}}=\mathbf{X}{(2)^{\top }}\mathbf{X}(1)/n$, ${\mathbf{C}_{11}}=\mathbf{X}{(1)^{\top }}\mathbf{X}(1)/n$ is invertible, $\mathbf{X}(1)$ and $\mathbf{X}(2)$ are the submatrices consisting of the columns of X corresponding to the active and inactive factors respectively, the element-wise function $\text{sign}(\cdot )$ maps zero to zero, positive numbers to $+1$ and negative numbers to $-1$, ${\boldsymbol{\beta }_{(1)}}$ is the vector of coefficients of active factors, $\mathbf{1}$ is a vector of $+1$’s, and the inequality holds element-wise. They showed that for fixed p and $\boldsymbol{\beta }$, under regularity conditions
\[\begin{aligned}{}& \underset{n\to +\infty }{\lim }\frac{1}{n}{\mathbf{X}^{\top }}\mathbf{X}=\mathbf{C},\mathbf{C}\hspace{2.5pt}\text{is a positive definite matrix},\\ {} & \underset{n\to +\infty }{\lim }\frac{1}{n}\underset{1\le i\le n}{\max }{\boldsymbol{x}_{(i)}^{\top }}{\boldsymbol{x}_{(i)}}=0,\end{aligned}\]
where ${\boldsymbol{x}_{(i)}}$ is the ith row of X, if the Strong Irrepresentable Condition holds, then there exists a ${\lambda _{n}}=f(n)$ such that
\[ \underset{n\to +\infty }{\lim }P(\text{sign}(\hat{\boldsymbol{\beta }}({\lambda _{n}}))=\text{sign}(\boldsymbol{\beta }))=1,\]
where $\hat{\boldsymbol{\beta }}({\lambda _{n}})$ is the Lasso estimate. Further, in Corollary 2 of their paper, they showed that if the worst case column correlation is no greater than $c/(2\| \boldsymbol{\beta }{\| _{0}}-1)$ for a constant $0\le c\lt 1$, then the Strong Irrepresentable Condition holds. In sum, a small worst case column correlation implies the Strong Irrepresentable Condition, which in turn guarantees the model selection consistency of the Lasso.

References

[1] 
Bertsimas, D., King, A. and Mazumder, R. Best subset selection via a modern optimization lens. The Annals of Statistics 44(2) 813–852 (2016). https://doi.org/10.1214/15-AOS1388. MR3476618
[2] 
Booth, K. H. and Cox, D. R. Some systematic supersaturated designs. Technometrics 4(4) 489–495 (1962). https://doi.org/10.2307/1266285. MR0184369
[3] 
Box, G. E. and Meyer, R. D. An analysis for unreplicated fractional factorials. Technometrics 28(1) 11–18 (1986). https://doi.org/10.2307/1269599. MR0824728
[4] 
Bulutoglu, D. A. and Cheng, C. S. Hidden projection properties of some nonregular fractional factorial designs and their applications. The Annals of Statistics 31(3) 1012–1026 (2003). https://doi.org/10.1214/aos/1056562472. MR1994740
[5] 
Chai, F. S., Das, A., Singh, R. and Stufken, J. Discriminating between superior $UE({s^{2}})$-optimal supersaturated designs. Statistics and Applications 18(2) (2020).
[6] 
Chen, J. and Lin, D. K. On the identifiability of a supersaturated design. Journal of Statistical Planning and Inference 72(1–2) 99–107 (1998). https://doi.org/10.1016/S0378-3758(98)00025-1. MR1655185
[7] 
Davenport, M. A., Duarte, M. F., Eldar, Y. C. and Kutyniok, G. Introduction to Compressed Sensing. Citeseer (2012). https://doi.org/10.1017/CBO9780511794308.002. MR2963166
[8] 
Deng, L. Y. and Lin, D. K. Criteria for supersaturated designs. In Proceedings of the Section on Physical and Engineering Sciences 124–128. American Statistical Association (1994).
[9] 
DuMouchel, W. and Jones, B. A simple Bayesian modification of D-optimal designs to reduce dependence on an assumed model. Technometrics 36(1) 37–47 (1994).
[10] 
Fan, J. and Lv, J. Sure independence screening for ultrahigh dimensional feature space. Journal of the Royal Statistical Society, Series B 70(5) 849–911 (2008). https://doi.org/10.1111/j.1467-9868.2008.00674.x. MR2530322
[11] 
Fan, J., Feng, Y. and Song, R. Nonparametric independence screening in sparse ultra-high-dimensional additive models. Journal of the American Statistical Association 106(494) 544–557 (2011). https://doi.org/10.1198/jasa.2011.tm09779. MR2847969
[12] 
Friedman, J., Hastie, T. and Tibshirani, R. Regularization paths for generalized linear models via coordinate descent. Journal of Statistical Software 33(1) 1–22 (2010).
[13] 
JMP (version 11). SAS Institute Inc., Cary, NC (1989–2007).
[14] 
Jones, B., Lin, D. K. and Nachtsheim, C. J. Bayesian D-optimal supersaturated designs. Journal of Statistical Planning and Inference 138(1) 86–92 (2008). https://doi.org/10.1016/j.jspi.2007.05.021. MR2369616
[15] 
Jones, B. and Majumdar, D. Optimal supersaturated designs. Journal of the American Statistical Association 109(508) 1592–1600 (2014). https://doi.org/10.1080/01621459.2014.938810. MR3293612
[16] 
Jones, B., Lekivetz, R., Majumdar, D., Nachtsheim, C. J. and Stallrich, J. W. Construction, properties, and analysis of group-orthogonal supersaturated designs. Technometrics 62(3) 403–414 (2020). https://doi.org/10.1080/00401706.2019.1654926. MR4125505
[17] 
Li, R. and Lin, D. K. Data analysis in supersaturated designs. Statistics & Probability Letters 59(2) 135–144 (2002). https://doi.org/10.1016/S0167-7152(02)00140-2. MR1927525
[18] 
Lin, D. K. A new class of supersaturated designs. Technometrics 35(1) 28–31 (1993).
[19] 
Lin, D. K. Generating systematic supersaturated designs. Technometrics 37(2) 213–225 (1995). https://doi.org/10.2307/1269765. MR1365722
[20] 
Liu, Y., Ruan, S. and Dean, A. M. Construction and analysis of $E({s^{2}})$ efficient supersaturated designs. Journal of Statistical Planning and Inference 137(5) 1516–1529 (2007). https://doi.org/10.1016/j.jspi.2006.09.001. MR2303773
[21] 
Marley, C. J. and Woods, D. C. A comparison of design and model selection methods for supersaturated experiments. Computational Statistics & Data Analysis 54(12) 3158–3167 (2010). https://doi.org/10.1016/j.csda.2010.02.017. MR2727742
[22] 
Osborne, M. R., Presnell, B. and Turlach, B. A. Knot selection for regression splines via the lasso. Computing Science and Statistics. 44–49 (1998).
[23] 
Plackett, R. L. and Burman, J. P. The design of optimum multifactorial experiments. Biometrika 33(4) 305–325 (1946). https://doi.org/10.1093/biomet/33.4.305. MR0016624
[24] 
R Core Team. R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria (2017). https://www.R-project.org/.
[25] 
Shen, X., Pan, W., Zhu, Y. and Zhou, H. On constrained and regularized high-dimensional regression. Annals of the Institute of Statistical Mathematics 65(5) 807–832 (2013). https://doi.org/10.1007/s10463-012-0396-3. MR3105798
[26] 
Singh, R. and Stufken, J. Selection of two-level supersaturated designs for main effects models. Technometrics 65 1–9 (2022). https://doi.org/10.1080/00401706.2022.2102080. MR4543063
[27] 
Tang, B. and Wu, C. F. J. A method for constructing supersaturated designs and its $E({s^{2}})$ optimality. Canadian Journal of Statistics 25(2) 191–201 (1997). https://doi.org/10.2307/3315731. MR1463319
[28] 
Tibshirani, R. Regression shrinkage and selection via the Lasso. Journal of the Royal Statistical Society, Series B 58 267–288 (1996). MR1379242
[29] 
Weese, M. L., Smucker, B. J. and Edwards, D. J. Searching for powerful supersaturated designs. Journal of Quality Technology 47(1) 66–84 (2015).
[30] 
Weese, M. L., Stallrich, J. W., Smucker, B. J. and Edwards, D. J. Strategies for supersaturated screening: Group orthogonal and constrained var (s) designs. Technometrics 63(4) 443–455 (2021). https://doi.org/10.1080/00401706.2020.1850529. MR4331445
[31] 
Welch, L. R. Lower bounds on the maximum cross correlation of signals (corresp.). IEEE Transactions on Information Theory 20(3) 397–399 (1974).
[32] 
Wu, C. F. J. Construction of supersaturated designs through partially aliased interactions. Biometrika 80(3) 661–669 (1993). https://doi.org/10.1093/biomet/80.3.661. MR1248029
[33] 
Wu, C. F. J. and Hamada, M. S. Experiments: planning, analysis, and optimization 2nd ed. John Wiley & Sons, Hoboken, NJ (2009). MR2583259
[34] 
Zhao, P. and Yu, B. On model selection consistency of Lasso. Journal of Machine Learning Research 7 2541–2563 (2006). MR2274449
Exit Reading PDF XML


Table of contents
  • 1 Introduction
  • 2 The Construction Method
  • 3 Simulation Study
  • 4 Conclusion and Discussion
  • Appendix A Proof for the Upper Bound of Coherence
  • Appendix B Coherence and Model Selection Consistency of the Lasso
  • References

Export citation

Copy and paste formatted citation
Placeholder

Download citation in file


Share


RSS

The New England Journal of Statistics in Data Science

  • ISSN: 2693-7166
  • Copyright © 2021 New England Statistical Society

About

  • About journal

For contributors

  • Submit
  • OA Policy
  • Become a Peer-reviewer
Powered by PubliMill  •  Privacy policy