Next Article in Journal
A Blow-Up Criterion for 3D Compressible Isentropic Magnetohydrodynamic Equations with Vacuum
Next Article in Special Issue
A Modified Cure Rate Model Based on a Piecewise Distribution with Application to Lobular Carcinoma Data
Previous Article in Journal
Application of Fractional Differential Equations for Modeling Bacteria Migration in Porous Medium
Previous Article in Special Issue
Sensitivity Analysis on Hyperprior Distribution of the Variance Components of Hierarchical Bayesian Spatiotemporal Disease Mapping
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Nonparametric Additive Regression for High-Dimensional Group Testing Data

1
School of Mathematics and Statistics, Guangxi Normal University, Guilin 541004, China
2
School of Mathematics, Hohai University, Nanjing 210098, China
*
Authors to whom correspondence should be addressed.
Mathematics 2024, 12(5), 686; https://doi.org/10.3390/math12050686
Submission received: 12 January 2024 / Revised: 20 February 2024 / Accepted: 21 February 2024 / Published: 27 February 2024
(This article belongs to the Special Issue Advances in Biostatistics and Applications)

Abstract

:
Group testing has been verified as a cost-effective and time-efficient approach, where the individual samples are pooled with a predefined group size for subsequent testing. Recent research has explored the integration of covariate information to improve the modeling of the group testing data. While existing works for high-dimensional data primarily focus on parametric models, this study considers a more flexible generalized nonparametric additive model. Nonlinear components are approximated using B-splines and model estimation under the sparsity assumption is derived employing group lasso. Theoretical results demonstrate that our method selects the true model with a high probability and provides consistent estimates. Numerical studies are conducted to illustrate the good performance of our proposed method, using both simulated and real data.

1. Introduction

Group testing is a proven cost-effective and time-efficient approach. In group testing, individual samples are pooled with a predefined group size and subsequently tested. Only those pools yielding positive results require additional tests, including individual tests. Dorfman [1] introduced group testing to screen syphilis among soldiers. Afterward, it has been widely applied to detect infected samples in a large population, including a variety of infectious diseases such as chlamydia, gonorrhea [2,3,4], HIV, the hepatitis B virus [5], West Nile virus [6], SARS-CoV-2 [7], and COVID-19 [8]. Group testing is also used in many other fields such as genetics [9], blood safety [10], drug discovery [11], and hyopneumoniae [12]. In particular, the efficiency of this approach is remarkable, especially when the prevalence of the disease is low. Gregory et al. [13] noted that the State Hygienic Laboratory (SHL) in Iowa saved approximately $3.1 million from 2009 to 2014 in screening female subjects for chlamydia. Verougstraete et al. [4] evaluated the feasibility of CT and NG molecular testing in a large number of female sex workers and showed that pool testing compared with single-site samples reduced the cost by 35 % .
Given the rapid advancement of digital technologies, it is possible to access extensive covariates, encompassing personal details such as age and occupation, electronic medical records, genetic data, and environmental variables. In recent literature on group testing, covariate information has been explored for estimation and case identification [14,15,16]. Vansteelandt et al. [17] proposed estimating individual infection probabilities by utilizing the test results obtained from initial pools. Xie [18] developed a general regression methodology that relates the group testing responses to individual covariate information. Black et al. [19] improved the Halving group testing procedure by taking into account population heterogeneity. It is evident that covariant information improves the estimate of individual risk probability, further reduces the test number, and lowers the cost [20].
Most existing research on group testing by incorporating covariate information has been developed under a parametric regression model, such as the logistic model. Zhang et al. [2] addressed the goal of case identification by incorporating retesting results. McMahan et al. [21] developed a Bayesian framework for group testing with covariate information. Gregory et al. [13] used the adaptive elastic net for group testing with high-dimensional covariates. Lin et al. [3] considered variable selection for multiple-infection group testing data. Alternatively, semiparametric or nonparametric regression has also received attention, but it mainly focuses on low-dimensional group testing. Yuan et al. [22] considered semiparametric isotonic regression models for the simultaneous estimation of the conditional probability curve and covariate effects, in which an expectation–maximization (EM) algorithm is used to overcome the computational challenge. Delaigle et al. considered nonparametric regression methods for group testing data [23,24,25]. Liu et al. [26] developed a Bayesian generalized additive regression approach to model the individual-level probability of disease. However, rare work has been conducted on high-dimensional covariates within the semiparametric or nonparametric regression framework for group testing.
In a standard group testing procedure, test results are collected at every stage, and high-dimensional covariate information is obtained during the initial enrollment of specimens. To maximize the utility of this extensive data, we propose an integrated modeling framework that employs test results from every phase of the group testing algorithm and combines them with high-dimensional covariate information. We employ a generalized nonparametric additive model framework to analyze the heterogeneous high-dimensional group testing data. The unknown nonlinear components are approximated with B-splines [27] and estimated by the group lasso method [28]. Four commonly used group testing algorithms are considered: master pool, Dorfman, Halving, and array testing [1,29,30,31]. To enhance computational efficiency, we utilize the EM algorithm in conjunction with the stochastic gradient descent algorithm. Theoretical results demonstrate that our method selects the true model with high probability and provides consistent estimates for components.
The remaining text is organized as follows. In Section 2, we establish our estimation method under the generalized nonparametric additive model through the B-spline approximation, along with the corresponding algorithm. Subsequently, the theoretical properties of the proposed method are investigated and proved. In Section 3, we derive the conditional expectation utilized in the EM algorithm to enhance computational efficiency. In Section 4 and Section 5, we conduct extensive simulations and make some real applications. The numerical result shows the good performance of our method. We conclude our methods in Section 6. The technical details of the proofs are presented in Appendix A.

2. Main Results

2.1. Notation

Throughout this work, we define R d as the set of d-dimensional real number vectors. We denote λ min ( A ) and λ max ( A ) as the minimum and maximum eigenvalues of matrix A , respectively. Furthermore, we employ [ m ] to define the set { 1 , , m } . The notation · represents the sup norm. Let ϵ n = o ( 1 ) indicate that ϵ n 0 . We use E ( · ) to denote the expectation concerning the underlying probability space. The symbol d denotes the convergence in distribution. We denote a n b n to signify the relationship b n a n = o ( 1 ) . Let f 2 = E ( f 2 ) and f n 2 = E n ( f 2 ) with E n represent the empirical measures. The superscript “T” indicates the transpose of a matrix or a vector.

2.2. Model

Suppose that n individuals are recruited to participate in testing. We denote by Y ˜ i the true status of the i-th individual, indicating if it is infected or not ( Y ˜ i = 1 or 0) for i [ n ] . Then Y ˜ i follows a Bernoulli distribution with a risk probability of p i . Simultaneously, we gather and express individual information as X i = ( X i 1 , , X i q n ) T R q n for the i t h individual, where X i has a dimension of q n . Assuming that Y ˜ i follows the generalized additive model (GAM) given X i ,
P ( Y ˜ i = 1 | X i ) = g ( μ + j = 1 q n f j ( X i j ) ) ,
where the link function is logit, g ( t ) = e t 1 + e t , μ represents the intercept, and  f j ( j [ q n ] ) is an unknown smooth nonlinear function [27,32,33]. The corresponding true values are denoted as μ * and f j * , respectively. The dimension of covariates X is notably large and increases alongside n. Typically, there is an assumption of model sparsity, which implies that only a subset of the covariates X correlates with the response variable. Let the true model be represented as Ω * = { j [ q n ] : f j * ( · ) 0 } , where its size is denoted by s 0 . Without loss of generality, we assume that the unknown additive component f j * is centered, meaning that E [ f j * ( X j ) ] = 0 .
In group testing, the true status of an individual is not directly observable; instead, tests are performed at the group level. Moreover, for groups flagged as high-risk due to positive results, additional tests, including individual tests, may be conducted, as individuals within these groups are more likely to be infected [34]. A typical group testing algorithm usually contains two steps. First, partition the specimens into J groups, and perform tests for each group. Second, conduct additional tests for the groups that tested positive in the first step. We denote the J groups as { P j , 1 } j = 1 J with group sizes { k j } j = 1 J , satisfying j = 1 J P j , 1 = { 1 , , n } . We define P j = ( P j , 1 , , P j , T j ) , where P j , 1 denotes the initial j t h group, and  P j , l denotes the subsequent subgroups or individuals within this group for l = 2 , , T j , with  T j representing the total number of tests associated with the j t h group. In Figure 1, we present a typical example of employing the Halving algorithm with a group size of 6, where the number is T j = 6 . Correspondingly, let Z j encompass all the test results associated with P j , 1 , including subgroup and individual tests, denoted by Z j = ( Z j , 1 , , Z j , T j ) .
Considering measurement error, we denote sensitivity and specificity by S e and S p , respectively. Sensitivity is the probability that a test correctly identifies a positive sample as positive, while specificity is the probability that a test correctly identifies a negative sample as negative. We assume these values are known and do not depend on the group size. We denote Z ˜ j = ( Z ˜ j , 1 , , Z ˜ j , T j ) as the true status corresponding to Z j , where Z ˜ j , l = max i P j , l Y ˜ i . Given the true status Z ˜ j , l , the outcome Z j , l follows a Bernoulli distribution with probability S e Z ˜ j , l ( 1 S p ) 1 Z ˜ j , l for l [ T j ] and j [ J ] . They are conditionally independent given their true status. Given X 1 , , X n , the individual statuses Y ˜ 1 , , Y ˜ n are conditionally independent. Similarly, the group statuses { Z ˜ j , l , l [ T j ] } are conditionally independent, given { Y ˜ i , i P j , 1 } . These assumptions are commonly made in the literature [13,16]. Let Y ˜ = ( Y ˜ 1 , , Y ˜ n ) and X = ( X 1 , , X n ) T . Then, the log-likelihood function of Y ˜ is expressed as follows:
ln P ( Y ˜ | X ) = i = 1 n [ Y ˜ i ln ( p i ) + ( 1 Y ˜ i ) ln ( 1 p i ) ] ,
where p i = P ( Y ˜ i = 1 | X i ) = g ( μ + j = 1 q n f j ( X i j ) ) . We note that the individual true status Y ˜ is unknown, whereas the group-level test results Z = { Z j , j [ J ] } are observed. The corresponding log-likelihood function is as follows:
P ( Z | X ) = Y ˜ { 0 , 1 } n P ( Z | Y ˜ ) P ( Y ˜ | X ) ,
where the conditional probability P ( Z | Y ˜ ) is formulated as follows:
P ( Z | Y ˜ ) = j = 1 J P ( Z j | Y ˜ ) = j = 1 J l = 1 T j { S e Z ˜ j , l ( 1 S p ) 1 Z ˜ j , l } Z j , l { ( 1 S e ) Z ˜ j , l S p 1 Z ˜ j , l } 1 Z j , l .
With the above likelihood function, our goal is to estimate the true model Ω * and obtain an accurate estimate of the risk probability through the model (1). As model (1) is a nonparametric additive model, it prompts the use of the B-spline method, known for its effectiveness in modeling such nonparametric frameworks.
Let S n denote the space of polynomial splines of degree l 1 , and let ϕ j k ( · ) ,   k [ m n ] denote the standardized B-spline basis functions satisfying ϕ j k 1 . Then for any f n j S n , there exist coefficients β j k , k [ m n ] , such that
f n j ( x ) = k = 1 m n β j k ϕ j k ( x ) .
Using f n j ( x ) allows us to approximate the nonlinear function f j ( x ) . Let β j = ( β j 1 , , β j m n ) T and Φ i j = ( ϕ j 1 ( X i j ) , , ϕ j m n ( X i j ) ) T . Then μ + j = 1 q n f j ( X i j ) could be approximated by Φ i T β , where Φ i = ( 1 , Φ i 1 T , , Φ i q n T ) T and β = ( β 0 , β 1 T , , β q n T ) T includes the intercept β 0 . Therefore, the risk probability p i = g ( μ + j = 1 q n f j ( X i j ) ) is approximated by p i β = g ( Φ i T β ) for i [ n ] . Consequently, the corresponding likelihood function is
P β ( Z | X ) = Y ˜ { 0 , 1 } n P ( Z | Y ˜ ) P β ( Y ˜ | X ) ,
with P β represents the probability measure with respect to the parameter β and
ln P β ( Y ˜ | X ) = i = 1 n [ Y ˜ i ln ( p i β 1 p i β ) + ln ( 1 p i β ) ] .
Given the high dimensionality of X , we propose maximizing P β ( Z | X ) using the group lasso penalty to select the important components and obtain the estimator:
β ^ = arg max β { ln P β ( Z | X ) n λ j = 1 q n β j 2 } ,
where λ is the tuning parameter, and the intercept β 0 is not penalized. The model Ω * is estimated by Ω ^ = { j [ q n ] : β ^ j 2 0 } , and the nonlinear components are estimated by f ^ n j ( x ) = k = 1 m n β ^ j k ϕ j k ( x ) , j [ q n ] .

2.3. The Estimation Procedure with EM Algorithm

The individual status Y ˜ is latent and not directly observable, posing a challenge in maximizing (5) to obtain the estimator. To address this issue, we employ the EM algorithm, which iterates with two steps: the E-step and the M-step. For the ( t + 1 ) t h iteration, these two steps are carried out as follows: in the E-step, we calculate the conditional expectation of the complete log-likelihood function of Y ˜ , given the observed data Z and the t t h iteration β ( t ) , denoted as follows: E [ ln f ( Y ˜ | β ) | Z , β ( t ) ] . In the M-step, we maximize the regularized function S ( t ) ( β ) = E [ ln f ( Y ˜ | β ) | Z , β ( t ) ] n λ s = 1 q n β s 2 , and update the parameter with β ( t + 1 ) . Our method is summarized in Algorithm 1.
Algorithm 1 Regularized GAM for group testing.
Input: The observed outcomes Z , covariates X , and the initial value β ( 0 ) .
Compute:
Step 1 (E-step). Given the parameter at the t t h iteration, β ( t ) , we calculate the conditional expectation E [ ln f ( Y ˜ | β ) | Z , β ( t ) ] .
Step 2 (M-step). We update the iterative parameter β ( t + 1 ) by maximizing the objective function S ( t ) ( β ) = E i = 1 n [ Y ˜ i ln ( p i β 1 p i β ) + ln ( 1 p i β ) ] | Z , β ( t ) n λ s = 1 q n β s 2 .
Step 3. We repeat the E-step and M-step until the parameters converge.
Output: The estimate β ^ , f ^ n j , j [ q n ] and Ω ^ .
We should note that the conditional expectation in the E-step exhibits variability across different group testing methods. For a comprehensive understanding, specific details are presented in Section 3 to facilitate the implementation of the EM algorithm.

2.4. The Consistency of the Estimate

In this section, we establish the theoretical results of our method. We commence by presenting essential conditions that are required to establish our main results. These conditions are commonly employed in B-spline methodologies. Within Conditions D and E, a set, U m , is introduced and defined as follows: the results { Z 1 , , Z J } are categorized into M classes, with U m representing the index set to which Z j belongs to the m-th category for m [ M ] .
A .
The nonparametric components { f j * } j = 1 q n belong to the function class F , where the r t h derivative f ( r ) exists and is Lipschitz of order α ,
F = { f ( · ) : | f ( r ) ( s ) f ( r ) ( t ) | K | s t | α , s , t [ a , b ] }
for some positive constant K, where r is a non-negative integer, [ a , b ] is the support of the covariate, and α ( 0 , 1 ] , such that d = r + α 0.5 .
B .
E f j * = 0 , and the marginal density function g j of X j adheres to 0 < K 1 g j ( X j ) K 2 < on [ a , b ] for 1 j p , where K 1 and K 2 are constants.
C .
min j Ω * { f j * 2 } q n n 2 d 1 2 ( 2 d + 1 ) , with q n = o ( n 3 4 1 2 ( 2 d + 1 ) ) .
D .
| U m | / J γ m ( 0 , 1 ) as J .
E .
I m ( β * ) = 1 | U m | j U m 2 log P β ( Z j | X ) β β T | β = β * is positive-definite and possesses a minimum eigenvalue larger than τ m for any m [ M ] .
Conditions A–C are commonly employed for nonparametric additive models, with similar assumptions made by Huang et al. [35]. When the dimension q n is of the order O ( 1 ) , Condition C is less stringent than Condition (A1) as described in Huang et al. [35]. Conditions D and E are presumed to pertain to the high-dimensional group testing data, guaranteeing an adequate dataset size for all categories of testing results, as employed in work by Gregory et al. [13].
Let the coefficients β * represent the population-level parameter used to approximate the nonparametric function, given by
β * = arg min β E Z P β ( Z | X ) .
Under mild conditions, our penalized estimate β ^ converges to β * , as demonstrated in the following theorem:
Theorem 1. 
Given Conditions A–E, m n = O ( n 1 2 d + 1 ) and λ = o ( q n 1 / 2 n γ / 2 / n ) , we have
β ^ β * 2 = O p ( q n 1 / 2 n ( 1 γ ) / 2 ) ,
where γ = 1 2 d + 1 .
This theorem reveals that the difference between β * and its estimate converges to zero in probability, achieving a convergence rate of O p ( q n 1 / 2 n ( 1 γ ) / 2 ) . According to Huang et al. [35], we set m n = O ( n 1 2 d + 1 ) to achieve the optimal convergence rate. In what follows, we derive the convergence rate between f j * and its estimate f ^ n j = Φ j T β ^ . Consequently, we validate the model selection property of our method.
Theorem 2. 
Given the conditions in Theorem 1, we have
(1) 
f ^ n j f j * 2 = O p ( q n n 2 d 1 2 ( 2 d + 1 ) ) .
(2) 
P ( Ω * Ω ^ ) 1 as n .
The above theorem demonstrates the convergence of the estimator f ^ n j to f j * , which quantifies the importance of the j t h component. According to Huang et al. [35], we set m n = O ( n 1 2 d + 1 ) to achieve the optimal convergence rate. We also establish the sure screening property of the set Ω ^ , encompassing the true model Ω * with a probability approaching 1.

3. The Conditional Expectation

In this section, we present the derivations of the conditional expectations for Y ˜ i based on the observed status of groups, a crucial component for the EM algorithm. These expectations are denoted by w i ( t ) = E [ Y ˜ i | Z 1 , , Z J , μ ( t ) , β ( t ) ] . Given that Y ˜ i follows a Bernoulli distribution and is independent, we express w i ( t ) as w i ( t ) = P ( Y ˜ i = 1 | Z 1 , , Z J , μ ( t ) , β ( t ) ) . This conditional expectation is derived based on the group testing algorithm we implemented. Detailed derivations are provided for three group testing methods: the master pool, Dorfman’s algorithm, and Halving. However, the derivation for the array method is exceedingly complex and has been omitted in this context for brevity.

3.1. Master Pool

With the master pool method, specimens are divided into non-overlapping groups, with each individual exclusively assigned to one of these groups. Testing is performed on each group, without requiring additional tests. Then we have the following:
ω i ( t ) = P ( Y ˜ i = 1 | Z j , 1 = 0 ) = ( 1 S e ) · p i B ( t ) / ( 1 P ( Z j , 1 = 1 ) ) if Z j , 1 = 0 , P ( Y ˜ i = 1 | Z j , 1 = 1 ) = S e · p i B ( t ) / P ( Z j , 1 = 1 ) if Z j , 1 = 1 ,
where P ( Z j , 1 = 1 ) = S e + ( 1 S e S p ) i P j ( 1 p i B ( t ) ) .

3.2. Dorfman

In Dorfman’s algorithm, specimens are partitioned into non-overlapping groups, similar to the master pool method, and group tests are conducted. Subsequently, for those groups testing positive, individual tests are conducted. When Z j , 1 = 0 , the conditional expectation mirrors that of master pool testing, specifically ω i ( t ) = P ( Y ˜ i = 1 | Z j , 1 = 0 ) = ( 1 S e ) · p i B ( t ) / ( 1 P ( Z j , 1 = 1 ) ) . When the j t h group tests positive, Z j , 1 = 1 , individual tests are performed subsequently. Let Z j , s : t = ( Z j , s , , Z j , t ) and Z j is short for Z 1 : T j . Then we have the following:
ω i ( t ) = P ( Y ˜ i = 1 | Z j , 1 = 1 , Z j , 2 : T j ) = P ( Y ˜ i = 1 , Z j , 1 = 1 , Z j , 2 : T j ) P ( Z j , 1 = 1 , Z j , 2 : T j ) .
The denominator is expressed as follows:
P ( Z j , 1 = 1 , Z j , 2 : T j ) = Z ˜ j P ( Z j , 1 = 1 , Z j , 2 : T j | Z ˜ j ) P ( Z ˜ j ) = Z ˜ j S e Z ˜ j , 1 ( 1 S p ) 1 Z ˜ j , 1 P ( Z j , 2 : T j | Z ˜ j , 2 : T j ) P ( Z ˜ j ) ,
where Z ˜ j = ( Z ˜ j , 1 , , Z ˜ j , T j ) and Z ˜ j , 2 : T j = ( Z ˜ j , 2 , , Z ˜ j , T j ) . Without loss of generalization, we assume that Y ˜ i represents the true status of the first individual of the group, with similar calculations applied to other cases. Subsequently, the numerator is as follows:
P ( Y ˜ i = 1 , Z j , 1 = 1 , Z j , 2 : T j ) = P ( Z j , 1 = 1 , Z j , 2 , Z j , 3 : T j | Y ˜ i = 1 ) P ( Y ˜ i = 1 ) = S e S e Z j , 2 ( 1 S e ) 1 Z j , 2 P ( Z j , 3 : T j ) p i B ( t )

3.3. Halving

With the Halving algorithm, the testing procedure is performed in three stages. The first stage is identical to the master pool method. In the second stage, groups that test positive are evenly split into two subgroups. In the third stage, these subgroups are tested, and positive ones undergo further individual tests. When Z j , 1 = 0 , the conditional expectation aligns with master pool testing, with ω i ( t ) = P ( Y ˜ i = 1 | Z j , 1 = 0 ) = ( 1 S e ) · p i B ( t ) / ( 1 P ( Z j , 1 = 1 ) ) . Assuming the j t h group tests positive with Z j , 1 = 1 , sequential tests are further carried out. These tests can unfold in any of the following scenarios:
(1)
Z j = ( 1 , 0 , 0 ) signifies that the group containing the i t h specimen initially tested positive in the first stage, followed by a negative result in a subsequent subgroup test; ultimately, the i t h specimen is concluded with a negative outcome. Then, we have the following:
ω i ( t ) = P ( Y ˜ i = 1 | Z j , 1 : 3 = ( 1 , 0 , 0 ) ) = P ( Y ˜ i = 1 , Z j , 1 : 3 = ( 1 , 0 , 0 ) ) P ( Z j , 1 : 3 = ( 1 , 0 , 0 ) ) ,
where the denominator is as follows:
P ( Z j , 1 : 3 = ( 1 , 0 , 0 ) ) = Z ˜ j P ( Z j , 1 : 3 = ( 1 , 0 , 0 ) | Z ˜ j , 1 : 3 ) P ( Z ˜ j , 1 : 3 ) = Z ˜ j S e Z ˜ j , 1 ( 1 S p ) 1 Z ˜ j , 1 ( 1 S e ) Z ˜ j , 2 + Z ˜ j , 3 S p 2 Z ˜ j , 2 Z ˜ j , 3 P ( Z ˜ j , 1 : 3 ) .
For simplicity, we assume that Y ˜ i is the first individual specimen P j , 2 . The derivation is similar to other cases. Then, the numerator is as follows:
P ( Y ˜ i = 1 , Z j , 1 : 3 = ( 1 , 0 , 0 ) ) = P ( Z j , 1 : 3 = ( 1 , 0 , 0 ) | Y ˜ i = 1 ) P ( Y ˜ i = 1 ) = S e ( 1 S e ) P ( Z j , 3 = 0 ) p i B ( t ) .
(2)
For Z j = ( 1 , 1 , 0 ) , we have the following:
ω i ( t ) = P ( Y ˜ i = 1 | Z j , 1 : 3 = ( 1 , 1 , 0 ) , Z j , 4 : T j ) = P ( Y ˜ i = 1 , Z j , 1 : 3 = ( 1 , 1 , 0 ) , Z j , 4 : T j ) P ( Z j , 1 : 3 = ( 1 , 1 , 0 ) , Z j , 4 : T j ) ,
where the denominator is as follows:
P ( Z j , 1 : 3 = ( 1 , 1 , 0 ) , Z j , 4 : T j ) = Z ˜ j P ( Z j , 1 : 3 = ( 1 , 1 , 0 ) , Z j , 4 : T j | Z ˜ j ) P ( Z ˜ j ) = Z ˜ j S e Z ˜ j , 1 + Z ˜ j , 2 ( 1 S p ) 2 Z ˜ j , 1 Z ˜ j , 2 ( 1 S e ) Z ˜ j , 3 S p 1 Z ˜ j , 3 P ( Z j , 4 : T j | Z ˜ j , 4 : T j ) P ( Z ˜ j ) .
For the numerator, we consider the two situations where the i t h specimen is located. If Y ˜ i is the first individual in the first subgroup, correspondingly, Z ˜ j , 4 = 1 , then the numerator is as follows:
P ( Y ˜ i = 1 , Z j , 1 : 3 = ( 1 , 1 , 0 ) , Z j , 4 : T j ) = P ( Z j , 1 : 3 = ( 1 , 1 , 0 ) , Z j , 4 : T j | Y ˜ i = 1 ) P ( Y ˜ i = 1 ) = S e 2 + Z j , 4 ( 1 S e ) 1 Z j , 4 P ( Z j , 3 = 0 ) P ( Z j , 5 : T j ) p i B ( t ) .
Otherwise, if Y ˜ i corresponds to the first individual in the second subgroup, then the numerator is as follows:
P ( Y ˜ i = 1 , Z j , 1 : 3 = ( 1 , 1 , 0 ) , Z j , 4 : T j ) = P ( Z j , 1 : 3 = ( 1 , 1 , 0 ) , Z j , 4 : T j | Y ˜ i = 1 ) P ( Y ˜ i = 1 ) = S e ( 1 S e ) P ( Z j , 2 = 1 ) P ( Z j , 4 : T j ) p i B ( t )
(3)
For Z j = ( 1 , 0 , 1 ) , we have the following:
ω i ( t ) = P ( Y ˜ i = 1 | Z j , 1 : 3 = ( 1 , 0 , 1 ) , Z j , 4 : T j ) = P ( Y ˜ i = 1 , Z j , 1 : 3 = ( 1 , 0 , 1 ) , Z j , 4 : T j ) P ( Z j , 1 : 3 = ( 1 , 0 , 1 ) , Z j , 4 : T j ) ,
where the denominator is as follows:
P ( Z j , 1 : 3 = ( 1 , 0 , 1 ) , Z j , 4 : T j ) = Z ˜ j P ( Z j , 1 : 3 = ( 1 , 0 , 1 ) , Z j , 4 : T j | Z ˜ j ) P ( Z ˜ j ) = Z ˜ j S e Z ˜ j , 1 + Z ˜ j , 3 ( 1 S p ) 2 Z ˜ j , 1 Z ˜ j , 3 ( 1 S e ) Z ˜ j , 2 S p 1 Z ˜ j , 2 P ( Z j , 4 : T j | Z ˜ j , 4 : T j ) P ( Z ˜ j ) .
Similarly, we derive the numerator according to the location of the i t h specimen. If it is assigned in the first subgroup with Z ˜ j , 4 = 1 , then the numerator is as follows:
P ( Y ˜ i = 1 , Z j , 1 : 3 = ( 1 , 0 , 1 ) , Z j , 4 : T j ) = P ( Z j , 1 : 3 = ( 1 , 0 , 1 ) , Z j , 4 : T j | Y ˜ i = 1 ) P ( Y ˜ i = 1 ) = S e ( 1 S e ) P ( Z P j , 3 = 1 ) P ( Z j , 4 : T j ) p i B ( t ) .
If it is assigned in the second subgroup, then the numerator is as follows:
P ( Y ˜ i = 1 , Z j , 1 : 3 = ( 1 , 1 , 0 ) , Z j , 4 : T j ) = P ( Z j , 1 : 3 = ( 1 , 1 , 0 ) , Z j , 4 : T j | Y ˜ i = 1 ) P ( Y ˜ i = 1 ) = S e 2 + Z j , 4 ( 1 S e ) 1 Z j , 4 P ( Z j , 2 = 0 ) P ( Z j , 5 : T j ) p i B ( t ) .
(4)
For Z j = ( 1 , 1 , 1 ) , we have the following:
ω i ( t ) = P ( Y ˜ i = 1 | Z j , 1 : 3 = ( 1 , 1 , 1 ) , Z j , 4 : T j ) = P ( Y ˜ i = 1 , Z j , 1 : 3 = ( 1 , 1 , 1 ) , Z j , 4 : T j ) P ( Z j , 1 : 3 = ( 1 , 1 , 1 ) , Z j , 4 : T j ) ,
where the denominator is as follows:
P ( Z j , 1 : 3 = ( 1 , 1 , 1 ) , Z j , 4 : T j ) = Z ˜ j P ( Z j , 1 : 3 = ( 1 , 1 , 1 ) , Z j , 4 : T j | Z ˜ j ) P ( Z ˜ j ) = Z ˜ j k = 1 3 S e Z ˜ j , k ( 1 S p ) 1 Z ˜ j , k P ( Z j , 4 : T j | Z ˜ j , 4 : T j ) P ( Z ˜ j ) .
In this case, the location of the i t h specimen in the subgroups will not affect the conditional expectation. Therefore, assuming it is the first individual in the first subgroup. Given Y ˜ i = 1 , then we have Z ˜ j , 1 = Z ˜ j , 2 = Z ˜ j , 4 = 1 , and, therefore, the numerator is derived as follows:
P ( Y ˜ i = 1 , Z j , 1 : 3 = ( 1 , 1 , 1 ) , Z j , 4 : T j ) = P ( Z j , 1 : 3 = ( 1 , 1 , 1 ) , Z j , 4 : T j | Y ˜ i = 1 , ) P ( Y ˜ i = 1 ) = S e 2 + Z j , 4 ( 1 S e ) 1 Z j , 4 P ( Z j , 2 = 1 ) P ( Z j , 5 : T j ) p i B ( t ) .

4. Simulation

In this section, we conduct numerical studies to demonstrate the performance of Algorithm 1 and assess the effectiveness of modeling group testing data within the framework of the generalized additive model. We adapt the methodology outlined by Liu et al. [26] to generate group testing data with high-dimensional covariates. We employ the following additive models:
Example 1. 
( n = 1000 and 500, q n = 50 , s 0 = 4 ). The model is P ( Y ˜ i = 1 | X i ) = g ( μ + f 1 ( X i 1 ) + f 2 ( X i 2 ) + f 3 ( X i 3 ) + f 4 ( X i 4 ) ) , where μ = 4 , f 1 ( x ) = 2 x , f 2 ( x ) = ( 2 x 1 4 ) 2 1.4 , f 3 ( x ) = 3 s i n ( x ) + c o s ( x 2 ) 5 6 , f 4 ( x ) = 3 e ( ( x + 1.5 ) 2 / 0.8 ) 1 4 and f j = 0 , j = 5 , , q n . The covariates are generated from the uniform distribution U ( 1 , 1 ) . In this case, the average prevalence rate is about 0.112 . Figure 2 presents the true components and their corresponding estimators in a simulation run.
Example 2. 
( n = 500 , q n = 50 and 100, s 0 = 4 , ρ = 0.2 ). In contrast to Example 1, Example 2 incorporates a correlation structure within the data while utilizing the same functions for data generation. The covariates ( X 1 , , X n ) T are generated as
X j = W j + t U 1 + t 2 , j [ q n ] ,
where W 1 , , W q n and U are i.i.d. samples from Uniform ( 1 , 1 ) , and t = ρ 1 ρ . In this case, the disease prevalence rate is about 0.064 .
Example 3. 
( n = 1000 , q n = 50 and 100, s 0 = 3 , ρ = 0.2 ). The model is defined as P ( Y ˜ i = 1 | X i ) = g ( μ + f 1 ( X i 1 ) + f 2 ( X i 2 ) + f 3 ( X i 3 ) ) , with μ = 3.5 , f 1 ( x ) = 2 x , f 2 ( x ) = ( 2 x 1 4 ) 2 1.1 , f 3 ( x ) = 3 s i n ( π 4 x ) + c o s 2 ( π 4 x ) 0.9 and f j = 0 , j = 4 , , q n . The covariates are simulated as in Example 2, and the average disease prevalence rate can be approximated as 0.063 .
Example 4. 
( n = 500 , q n = 50 and 100, s 0 = 3 , ρ = 0.2 ). Following Example 3, we employ the same function for data generation and conduct simulation experiments with varied values of q n while reducing the sample size to 500. The individual average prevalence rate approximates around 0.10 .
We implement our method using four group testing algorithms: master pool testing (master pool), Dorfman, Halving, and array. For the master pool, individuals are grouped into non-overlapping pools with a size of 4, conducting tests without any subsequent individual tests. For Dorfman and Halving testing, the group size is fixed at 4. In the case of array testing, individuals are arranged into 4 × 4 arrays. The sensitivity and specificity are set at 0.98 .
Let | | f ^ n j | | n 2 = 1 n j = 1 n f ^ n j 2 ( X i j ) = β ^ j T ( E n Φ j Φ j T ) β ^ j denote the effect of a component. The estimators β ^ are obtained using Algorithm 1. Since we utilize stochastic gradient descent to optimize the regularized objective function, it is possible that the coefficients of non-significant components may not precisely equal zero. As illustrated in Figure 2, a variable is deemed significant and selected if f ^ n j n deviates significantly from zero. In contrast, for other non-significant variables, such as x 5 and x 6 depicted in Figure 2, their functional estimates closely approximate zero.
To assess the performance of Algorithm 1 in variable selection and model estimation, we conduct experiments across the four examples, repeating the procedure 100 times for each example. The initial parameters for each experiment follow a standard normal distribution. We calculate the number of true positives (TPs) and false positives (FPs) to evaluate variable selection attributes. Additionally, we employ the mean prediction error (PE) to measure the overall prediction error of the model, defined as PE = E X [ ( f ^ ( X ) f ( X ) ) 2 ] . Table 1 reports the average and corresponding standard deviations of TP, FP, and PE. The results demonstrate that Algorithm 1 effectively identifies significant variables and exhibits high robustness. In Table 1, we set a threshold of τ = 0.05 to obtain the estimate M ^ , employing stochastic gradient descent to optimize the objective. Notably, only a few cases are observed where variables that are not significant are mistakenly identified.
To further validate the performance of Algorithm 1, we utilize two metrics, namely PEfj to signify the prediction error of each component, and PEnon to represent the average prediction error of non-significant variables. In Table 2 and Table 3, we present the outcomes of four models using four group testing methods. Notably, the prediction error (PEfj) associated with each significant variable is relatively small, which serves as an indication that the B-spline successfully captures the unknown nonlinear functions using Algorithm 1. These findings ascertain Algorithm 1’s outstanding capability in variable selection, enabling it to precisely identify important variables while minimizing the probability of mistakenly selecting non-significant variables.
Among the four different group testing methods, the Dorfman and array methods demonstrate superior performance, compared to the master pool and Halving methods. This characteristic verifies the fact that Dorfman employs sufficient individual repeated testing and array utilizes more group-level tests. In contrast, Halving adopts a sub-group retest, and subsequently, the individual test, while the master pool only uses the initial pools and does not involve any retesting.

5. Application

In our study, we perform regression analysis using the data set from the 1999–2004 National Health and Nutrition Examination Survey (NHANES). NHANES is a continuous cross-sectional survey conducted in the United States, which utilizes a probabilistic sampling method to select participants. The survey collects demographic information, health history, and behavioral data through interviews with participants and their families. The dataset contains 5515 observations ( n = 5515 ) and classifies individuals as diabetics or non-diabetic patients. Following Yu et al. [36], we select a subset of covariates for regression analysis and subsequently divide the dataset into a training set ( n = 4412 ) and a test set ( n = 1103 ). The subset contains 14 variables that are potentially associated with the risk of diabetes: education ( X 1 ), household income ( X 2 ), physical activity ( X 3 ), hypertension ( X 4 ), family history ( X 5 ), sex ( X 6 ), race and ethnicity ( X 7 ), alcohol use ( X 8 ), BMI ( X 9 ), height ( X 10 ), waist circumference ( X 11 ), weight ( X 12 ), age ( X 13 ), and smoking ( X 14 ). To exam the performance of our method for high-dimensional data, we add non-significant variables that follow the standard normal distribution, increasing the dimension to q n = 500 .
To perform the group testing algorithm, we choose Dorfman’s algorithm with a group size of k = 2 , 5 , 8 , setting the sensitivity to S e = 0.95 and the specificity to S p = 0.98 . Based on all available risk factors collected from each individual, we use the following model to describe an individual’s diabetes status:
P ( Y ˜ i = 1 | X i ) = g ( β 0 + X i 1 T β 1 + + X i 7 T β 7 + j = 8 q n f j ( X i j ) ) ,
where the categorical variables X 1 to X 7 are transformed using one-hot encoding, represented by X i j for j = 1 , , 7 . Additionally, nonlinear functions are applied to variables X 8 through X 14 to capture their relationship with the diabetes status more accurately. The unknown function, f j ( · ) , in (6) allows variables to have a non-linear influence on the state of diabetes. We assume that the link function g ( · ) is a logistic function. For comparison, we also report the results using the individual data ( k = 1 ) and a generalized linear regression model (GLM) based on the original data. To demonstrate the performance of our algorithm, we report the following statistics: accuracy (ACC), positive predictive value ( P P V = T P T P + F P ), negative predictive value ( N P V = T N T N + F N ), and recall ( R e c a l l = T P T P + F N ), where TP, FP, TN, and FN represent the numbers of true positives, false positives, true negatives, and false negatives, respectively. The results are presented in Table 4.
Our algorithm exhibits an accuracy of approximately 83 % , while the GLM model demonstrates an accuracy of 81.50 % when correctly classifying individuals as diabetics or non-diabetics. Our algorithm also has better performance than directly using GLM on the other three measures. It is expected that individual tests ( k = 1 ) achieve the highest precision. However, our group testing algorithms with k > 1 perform competitively while saving a considerable number of tests compared with individual tests. For continuous variables from X 8 to X 500 , our algorithm selects the variables X 10 , X 11 , and X 13 . The other variables have little or no impact on the response. The estimates for the functions f 10 , f 11 , and f 13 are shown in Figure 3, along with the estimates obtained using GLM. It is noteworthy that our method demonstrates a slightly higher accuracy compared to the GLM model, indicating its potential advantages in capturing more complex relationships and improving predictive capabilities.

6. Concluding Remarks

In this study, we propose a generalized nonparametric additive model to characterize high-dimensional group testing data. We mainly use the B-spline to approximate the unknown nonparametric function. Theoretical results demonstrate that our method selects the true model with high probability and provides consistent estimates for components. Our work extends the existing works from a generalized linear model to a more flexible generalized nonparametric model for high-dimensional data.
Consideration of the measurement error in a test kit is essential for practical applications. The unobservable nature of an individual’s true status, treated as a latent variable, prompts us to use the EM algorithm to maximize the regularized log-likelihood function for high-dimensional data. The E-step of the EM algorithm varies according to the structures of different group testing procedures and needs to be calculated separately to speed up the algorithm.
Another critical issue for group testing is case identification, whose efficiency could be improved with the help of covariates. In future work, we will study the improvement of our method for case identification, and implement an informative retest procedure to accurately identify the infections and further reduce costs.

Author Contributions

Methodology, W.X. and X.Z.; writing—original draft preparation, X.Z. and J.D.; validation, J.D. and J.Z.; formal analysis, X.Z., J.Z. and J.D.; writing—review and editing, J.D. and W.X.; supervision, W.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (grant nos. 11801102, 12361055), the Fundamental Research Funds for the Central Universities (423131), and Guangxi Natural Science Foundation (2021GXNSFAA220054).

Data Availability Statement

Data source is provided within the manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

In this appendix, we first present a fact and a lemma that have been established in the literature. Then we provide detailed proofs of Theorems 1 and 2.
Fact A1 
(Zhou et al. [37]). Under Conditions A and B, there exist some positive constants, C 1 and C 2 , such that we have the following:
C 1 m n 1 λ min ( E ( Φ Φ T ) ) λ max ( E ( Φ Φ T ) ) C 2 m n 1 .
Lemma A1 
(Huang et al. [35]). Under Condition B, for any j [ q n ] , there exists an f n j S n , satisfying the following:
f n j f j * 2 = O p ( m n d + m n 1 2 n 1 2 ) .
In particular, if we choose m n = O ( n 1 2 d + 1 ) , then
f n j f j * 2 = O p ( m n d ) = O p ( n d 2 d + 1 ) .
Proof of Theorem 1. 
Note that the objective function is as follows:
S n ( β ) = n ( β ) n λ P ( β ) ,
where n ( β ) = j = 1 J log P β ( Z j | X ) is the log-likelihood function and n λ P ( β ) is a penalty with a tuning parameter λ . We first prove that the estimate β ^ has a convergence rate of β ^ β * = O p ( q n n γ / 2 n ) with γ = 1 2 d + 1 . Let α n = q n n γ / 2 n , and let u denote a q n + 1 dimensional vector. To verify that the convergence rate is α n , we only need to prove that there exists some positive constant C, such that
P Sup u = C S n ( β * + α n u ) S n ( β * ) < 0 1 .
Taking the Taylor expansion, we have the following:
S n ( β * + α n u ) S n ( β * ) = n ( β * + α n u ) n ( β * ) n λ j = 1 q n P ( β j * + α n u j ) P ( β j * ) = α n u T n ( β * ) β + 1 2 α n 2 u T 2 n ( β * ) β β T u + 1 6 α n 3 l , j , k 3 n ( β w ) β l β j β k u l u j u k n λ j = 1 q n P ( β j * + α n u j ) P ( β j * ) = I 1 + I 2 + I 3 I 4 ,
where β w is an intermediate value between β * and β * + α n u . In what follows, we calculate the value of I 1 , I 2 , I 3 , and I 4 , separately.
(1)
Since I 1 = α n u T n ( β * ) β , its first-order derivative is as follows:
n ( β * ) β = j = 1 J log P β * ( Z j | X ) β = n m = 1 M ( | U m | n ) 1 2 S m ,
where S m = 1 | U m | j U m log P β * ( Z j | X ) β and P β * is a probability that depends on the parameter β * . Recall that P j = ( P j , 1 , , P j , T j ) , and P j , 1 represents the original group with size k j . Since Z j only depends on the true status of individual y P j , 1 { 0 , 1 } k j , we have the following:
P β * ( Z j | X ) = y P j , 1 P ( Z j | Y ˜ P j , 1 = y P j , 1 ) P β * ( Y ˜ P j , 1 = y P j , 1 | X P j , 1 ) .
Recall that P β * ( Y ˜ P j , 1 = y P j , 1 | X P j , 1 ) = s P j , 1 P β * ( Y ˜ s = y s | X s ) . We denote the true risk probability by P * ( Y ˜ s = y s | X s ) = e ( μ * + j = 1 q n f j * ( X s j ) ) y s 1 + e μ * + j = 1 q n f j * ( X s j ) . Then we have the following:
| P β * ( Y ˜ s = y s | X s ) P * ( Y ˜ s = y s | X s ) | = e Φ s T β * y s 1 + e Φ s T β * e ( μ * + j = 1 q n f j * ( X s j ) ) y s 1 + e μ * + j = 1 q n f j * ( X s j ) | Φ s T β * ( μ * + j = 1 q n f j * ( X s j ) ) | .
Note that the logistic model can be viewed as a generalized linear model exp ( y i η i b ( η i ) ) , where b ( η i ) = log ( 1 + exp ( η i ) ) and η i = Φ i T β . We denote the population-level parameter, η i * = Φ i T β * , then we have the following:
S m = 1 | U m | j U m 1 P β * ( Z j | X ) P β * ( Z j | X ) β = 1 | U m | j U m y P j , 1 P ( Z j | Y ˜ P j , 1 = y P j , 1 ) exp P j , 1 ( y l η l * b ( η l * ) ) P j , 1 ( y l b ( η l * ) ) Φ l y P j , 1 P ( Z j | Y ˜ P j , 1 = y P j , 1 ) exp P j , 1 ( y l b ( η l * ) ) = 1 U m j U m y P j , 1 P ( Z j | Y ˜ P j , 1 = y P j , 1 ) exp P j , 1 y l η l * b ( η l * ) P j , 1 ( y l b ( η l * ) ) Φ l y P j , 1 P ( Z j | Y ˜ P j , 1 = y P j , 1 ) exp P j , 1 ( y l η l * b ( η l * ) ) = 1 U m j U m y P j , 1 P β * ( Y ˜ P j , 1 = y P j , 1 | Z j ) · P j , 1 ( y l b ( η l * ) ) Φ l ,
where the last equality comes from the following conditional probability:
P β * ( Y ˜ P j , 1 = y P j , 1 | Z j ) = P ( Z j | Y ˜ P j , 1 = y P j , 1 ) exp P j , 1 ( y l η l * b ( η l * ) ) y P j , 1 P ( Z j | Y ˜ P j , 1 = y P j , 1 ) exp P j , 1 ( y l η l * b ( η l * ) ) .
Let ρ l * = μ * + j = 1 q n f j * ( X l , j ) denote the true model, then we have the following:
y l η l * b ( η l * ) = y l ρ l * b ( ρ l * ) + y l ( η l * ρ l * ) b ( η l * ) b ( ρ l * ) = y l ρ l * b ( ρ l * ) + ( y l b ( ρ l * ) ) ( η l * ρ l * ) ϖ l ,
where ϖ l = b ( η l * ) b ( ρ l * ) b ( ρ l * ) ( η l * ρ l * ) . Then,
P j , 1 [ y l η l * b ( η l * ) ] = P j , 1 [ y l ρ l * b ( ρ l * ) ] + P j , 1 ( y l b ( ρ l * ) ) ( η l * ρ l * ) P j , 1 ϖ l .
Let δ l = ( y l b ( ρ l * ) ) ( η l * ρ l * ) . Combining with the fact that ϖ l is independent with the value y P j , 1 , we have the following:
S m = 1 | U m | j U m y P j , 1 P ( Z j | Y ˜ P j , 1 = y P j , 1 ) exp P j , 1 ( y l b ( ρ l * ) ) ρ l * + P j , 1 δ l P j , 1 y l b ( ρ l * ) + b ( η , ω ) ( ρ l * η l * ) Φ l y P j , 1 P ( Z j | Y ˜ P j , 1 = y P j , 1 ) exp P j , 1 ( y l b ( ρ l * ) ) ρ l * + P j , 1 δ l
where η , ω lies between ρ l * and η l * . Note that b ( η ) = e η 1 + e η and b ( η ) = e η ( 1 + e η ) 2 both belonging to the interval ( 0 , 1 ) . Recall that η l * represents a B-spline function that approximates ρ l * . By Lemma A1, we have P j , 1 δ l = o p ( 1 ) for a finite group size and
S m = 1 U m j U m y P j , 1 P ( Z j | Y ˜ P j , 1 = y P j , 1 ) exp P j , 1 ( y l b ( ρ l * ) ) ρ l * + o p ( 1 ) P j , 1 ( y l b ( ρ l * ) ) Φ l + P j , 1 ( b ( ρ l * ) b ( η l * ) ) Φ l y P j , 1 P ( Z j | Y ˜ P j , 1 = y P j , 1 ) exp P j , 1 ( y l b ( ρ l * ) ) ρ l * + o p ( 1 ) .
Since [ b ( ρ l * ) b ( η l * ) ] Φ l is independent with the latent variable y Q l , we have the following:
S m = 1 U m j U m y P j , 1 P ( Z j | Y ˜ P j , 1 = y P j , 1 ) exp P j , 1 ( y l b ( ρ l * ) ) ρ l * + o p ( 1 ) P j , 1 ( y l b ( ρ l * ) ) Φ l y P j , 1 P ( Z j | Y ˜ P j , 1 = y P j , 1 ) exp P j , 1 ( y l b ( ρ l * ) ) ρ l * + o p ( 1 ) + 1 U m j U m P j , 1 b ( ρ l * ) b ( η l * ) Φ l .
Combining with (A2), we have the following:
S m = 1 | U m | j U m y P j , 1 P * ( Y ˜ P j , 1 = y P j , 1 | Z j ) · P j , 1 ( y l b ( ρ l * ) ) Φ l ( 1 + o p ( 1 ) ) + 1 | U m | j U m y P j , 1 P * ( Y ˜ P j , 1 = y P j , 1 | Z j ) · P j , 1 ( b ( ρ l * ) b ( η l * ) ) Φ l ( 1 + o p ( 1 ) ) = 1 | U m | j U m I m , j ( 1 + o p ( 1 ) ) + 1 | U m | j U m II m , j ( 1 + o p ( 1 ) ) .
By taking the expectation over I m , j , we find that
E Z j I m , j = E Z j y P j , 1 P * ( Y ˜ P j , 1 = y P j , 1 | Z j ) · P j , 1 ( y l b ( ρ l * ) ) Φ l = E Z j E ( Y ˜ P j , 1 | Z j ) P j , 1 ( y l b ( ρ l * ) ) Φ l = 0 .
By the fact that P j , j [ m ] is non-overlapped and Condition D, we have 1 | U m | j U m I m , j d N ( 0 , V m ) with V m = E Z j I m , j 2 . Additionally, by Lemma A1 and Fact A1, we have the following:
| 1 | U m | j U m II m , j | = | 1 | U m | P j , 1 ( b ( ρ * ) b ( η l * ) ) Φ l | = O p ( | U m | n d 2 d + 1 ) = O p ( n 1 2 ( 2 d + 1 ) ) .
So, S m is dominated by 1 | U m | j U m II m , j . It follows that S m = O p ( n 1 2 ( 2 d + 1 ) ) and n ( β * ) β = O P ( n · n 1 2 ( 2 d + 1 ) ) = O p ( n d + 1 2 d + 1 ) . Since γ = 1 2 d + 1 , we have the following: n ( β * ) β = O P ( n γ + 1 2 ) and
I 1 = α n u T n ( β * ) β = O p ( α n u 2 q n n γ / 2 n ) = O p ( n α n 2 u 2 ) .
(2)
We continue to derive I 2 = 1 2 α n 2 u T 2 n ( β * ) β β T u , in which
2 n ( β * ) β β T = n m = 1 M | U m | n 1 | U m | j U m 2 log P β * ( Z j | X ) β β T .
By Condition D and the law of large numbers, we have
1 n 2 n ( β * ) β β T m = 1 M γ m I m ( β * ) .
By Condition E, I m ( β * ) is positive-definite, with a minimum eigenvalue exceeding τ m for any m [ M ] . Let τ ̲ = min m [ M ] { τ m } . Then, for a sufficiently large n, we have
I 2 1 4 n τ ̲ α n 2 u 2 .
(3)
In the following, we derive the bound for I 3 = 1 6 α n 3 l , j , k 3 n ( β w ) β l β j β k u l u j u k , where
1 n 3 n ( β w ) β l β j β k = m = 1 M | U m | n 1 | U m | j U m 3 n j ( β w ) β l β j β k ,
and n j ( · ) denotes the log-likelihood function for Z j . Note that
3 n j ( β w ) β l β j β k = P β ( Z j | X ) 3 3 P β ( Z j | X ) β l β j β k + 2 P β ( Z j | X ) 3 i { , j , k } P β ( Z j | X ) β i P β ( Z j | X ) 2 2 P β ( Z j | X ) β l β j P β ( Z j | X ) β k P β ( Z j | X ) 2 2 P β ( Z j | X ) β l β k P β ( Z j | X ) β j P β ( Z j | X ) 2 2 P β ( Z j | X ) β j β k P β ( Z j | X ) β l .
According to the proof of Theorem 1 by Gregory et al. [13], we have the following results:
| P β ( Z j | X ) β l | 2 i P j | Φ i l | , | 2 P β ( Z j | X ) β l β j | 4 i 1 , i 2 P j | Φ i 1 Φ i 2 j | , | 2 P β ( Z j | X ) β l β j | 12 i 1 , i 2 , i 3 P j | Φ i 1 Φ i 2 j Φ i 2 k | .
By the fact that covariates are bounded, we have the following:
1 | U m | j U m log P β ( Z j | X , β w ) β l β j β k = O p ( 1 ) .
It follows that
, j , k = 1 q n + 1 u l u j u k j = 1 J 3 n j ( β w ) β l β j β k = O p ( q n 3 / 2 u 3 ) .
By Condition C with q n = o ( n 3 / 4 γ / 2 ) , we have the following: I 3 = o ( n α n 2 ) .
(4)
Further, by the triangle inequality, we have
| n λ j = 1 q n ( β j * + α n u j β j * ) | n λ j = 1 q n α n u j .
Given the condition λ = o ( q n 1 / 2 n γ / 2 / n ) , we find that
I 4 = n λ j = 1 q n P ( β j * + α n u j ) P ( β j * ) = o p ( n α n 2 ) .
Combining the above results, for a sufficiently large u , we observe that the value of S n ( β * + α n u ) S n ( β * ) is dominated by I 2 , which is negative. It implies the existence of a positive constant C, such that
P Sup u = C S n ( β * + α n u ) S n ( β * ) < 0 1 , n .
Then, the proof is finished.
Proof of Theorem 2. 
(1).
By the property of the B-spline approximation [38], there exist some positive constants c ˜ 1 and c ˜ 2 , such that
c ˜ 1 m n 1 β ^ j β j * 2 f ^ n j f n j * 2 2 c ˜ 2 m n 1 β ^ j β j * 2 ,
where f n j * = Φ j T β j * . Then we have the following:
f ^ n j f n j * 2 = O p ( q n n 2 d 1 2 ( 2 d + 1 ) ) .
Combining with the fact f n j * f j * 2 = O p ( m n d ) , we have f ^ n j f j * 2 = O p ( q n n 2 d 1 2 ( 2 d + 1 ) ) .
(2).
If there exists j Ω * , such that β ^ j = 0 , then f ^ n j = 0 . Combining with the above result (1), we have 0 f j * 2 = O p ( q n n 2 d 1 2 ( 2 d + 1 ) ) . It contradicts Condition C. Then, the proof is finished.

References

  1. Dorfman, R. The detection of defective members of large populations. Ann. Math. Stat. 1943, 14, 436–440. [Google Scholar] [CrossRef]
  2. Zhang, B.; Bilder, C.R.; Tebbs, J.M. Group testing regression model estimation when case identification is a goal. Biom. J. 2013, 55, 173–189. [Google Scholar] [CrossRef]
  3. Lin, J.; Wang, D.; Zheng, Q. Regression analysis and variable selection for two-stage multiple-infection group testing data. Stat. Med. 2019, 38, 4519–4533. [Google Scholar] [CrossRef]
  4. Verougstraete, N.; Verbeke, V.; De Cannière, A.S.; Simons, C.; Padalko, E.; Coorevits, L. To pool or not to pool? Screening of Chlamydia trachomatis and Neisseria gonorrhoeae in female sex workers: Pooled versus single-site testing. Sex. Transm. Infect. 2020, 96, 417–421. [Google Scholar] [CrossRef] [PubMed]
  5. Stramer, S.L.; Notari, E.P.; Krysztof, D.E.; Dodd, R.Y. Hepatitis B virus testing by minipool nucleic acid testing: Does it improve blood safety? Transfusion 2013, 53, 2449–2458. [Google Scholar] [CrossRef]
  6. Busch, M.P.; Caglioti, S.; Robertson, E.F.; McAuley, J.D.; Tobler, L.H.; Kamel, H.; Linnen, J.M.; Shyamala, V.; Tomasulo, P.; Kleinman, S.H. Screening the blood supply for West Nile virus RNA by nucleic acid amplification testing. N. Engl. J. Med. 2005, 353, 460–467. [Google Scholar] [CrossRef]
  7. Mutesa, L.; Ndishimye, P.; Butera, Y.; Souopgui, J.; Uwineza, A.; Rutayisire, R.; Ndoricimpaye, E.L.; Musoni, E.; Rujeni, N.; Nyatanyi, T.; et al. A pooled testing strategy for identifying SARS-CoV-2 at low prevalence. Nature 2021, 589, 276–280. [Google Scholar] [CrossRef] [PubMed]
  8. Bish, D.R.; Bish, E.K.; El-Hajj, H.; Aprahamian, H. A robust pooled testing approach to expand COVID-19 screening capacity. PLoS ONE 2021, 16, e0246285. [Google Scholar] [CrossRef]
  9. Gastwirth, J.L. The efficiency of pooling in the detection of rare mutations. Am. J. Hum. Genet. 2000, 67, 1036–1039. [Google Scholar] [CrossRef] [PubMed]
  10. Okasha, H.; Baddour, M.; Elsawy, M.; Sadek, N.; Eltoweissy, M.; Gouda, A.; Abdelkhalek, O.; Meheissen, M. Optimization of pooling technique for hepatitis C virus nucleic acid testing (NAT) in blood banks. Hepat. Mon. 2020, 20, e99571. [Google Scholar] [CrossRef]
  11. Hughes-Oliver, J.M. Pooling experiments for blood screening and drug discovery. In Screening: Methods for Experimentation in Industry, Drug Discovery, and Genetics; Springer: New York, NY, USA, 2006; pp. 48–68. [Google Scholar]
  12. Sponheim, A.; Munoz-Zanzi, C.; Fano, E.; Polson, D.; Pieters, M. Pooled-sample testing for detection of Mycoplasma hyopneumoniae during late experimental infection as a diagnostic tool for a herd eradication program. Prev. Vet. Med. 2021, 189, 105313. [Google Scholar] [CrossRef]
  13. Gregory, K.B.; Wang, D.; McMahan, C.S. Adaptive elastic net for group testing. Biometrics 2019, 75, 13–23. [Google Scholar] [CrossRef]
  14. Chen, P.; Tebbs, J.M.; Bilder, C.R. Group testing regression models with fixed and random effects. Biometrics 2009, 65, 1270–1278. [Google Scholar] [CrossRef] [PubMed]
  15. McMahan, C.S.; Tebbs, J.M.; Bilder, C.R. Regression models for group testing data with pool dilution effects. Biostatistics 2013, 14, 284–298. [Google Scholar] [CrossRef]
  16. Wang, D.; McMahan, C.; Gallagher, C.; Kulasekera, K. Semiparametric group testing regression models. Biometrika 2014, 101, 587–598. [Google Scholar] [CrossRef]
  17. Vansteelandt, S.; Goetghebeur, E.; Verstraeten, T. Regression models for disease prevalence with diagnostic tests on pools of serum samples. Biometrics 2000, 56, 1126–1133. [Google Scholar] [CrossRef] [PubMed]
  18. Xie, M. Regression analysis of group testing samples. Stat. Med. 2001, 20, 1957–1969. [Google Scholar] [CrossRef] [PubMed]
  19. Black, M.S.; Bilder, C.R.; Tebbs, J.M. Group testing in heterogeneous populations by using halving algorithms. J. R. Stat. Soc. Ser. C Appl. Stat. 2012, 61, 277–290. [Google Scholar] [CrossRef] [PubMed]
  20. Bilder, C.R.; Tebbs, J.M.; Chen, P. Informative retesting. J. Am. Stat. Assoc. 2010, 105, 942–955. [Google Scholar] [CrossRef]
  21. McMahan, C.S.; Tebbs, J.M.; Hanson, T.E.; Bilder, C.R. Bayesian regression for group testing data. Biometrics 2017, 73, 1443–1452. [Google Scholar] [CrossRef] [PubMed]
  22. Yuan, A.; Piao, J.; Ning, J.; Qin, J. Semiparametric isotonic regression modelling and estimation for group testing data. Can. J. Stat. 2021, 49, 659–677. [Google Scholar] [CrossRef] [PubMed]
  23. Delaigle, A.; Meister, A. Nonparametric regression analysis for group testing data. J. Am. Stat. Assoc. 2011, 106, 640–650. [Google Scholar] [CrossRef]
  24. Delaigle, A.; Hall, P. Nonparametric regression with homogeneous group testing data. Ann. Stat. 2012, 40, 131–158. [Google Scholar] [CrossRef]
  25. Delaigle, A.; Hall, P.; Wishart, J. New approaches to nonparametric and semiparametric regression for univariate and multivariate group testing data. Biometrika 2014, 101, 567–585. [Google Scholar] [CrossRef]
  26. Liu, Y.; McMahan, C.S.; Tebbs, J.M.; Gallagher, C.M.; Bilder, C.R. Generalized additive regression for group testing data. Biostatistics 2021, 22, 873–889. [Google Scholar] [CrossRef]
  27. Yoshida, T.; Naito, K. Asymptotics for penalised splines in generalised additive models. J. Nonparametr. Stat. 2014, 26, 269–289. [Google Scholar] [CrossRef]
  28. Yuan, M.; Lin, Y. Model selection and estimation in regression with grouped variables. J. R. Stat. Soc. Ser. B Stat. Methodol. 2006, 68, 49–67. [Google Scholar] [CrossRef]
  29. Litvak, E.; Tu, X.M.; Pagano, M. Screening for the presence of a disease by pooling sera samples. J. Am. Stat. Assoc. 1994, 89, 424–434. [Google Scholar] [CrossRef]
  30. Kim, H.Y.; Hudgens, M.G.; Dreyfuss, J.M.; Westreich, D.J.; Pilcher, C.D. Comparison of group testing algorithms for case identification in the presence of test error. Biometrics 2007, 63, 1152–1163. [Google Scholar] [CrossRef]
  31. Xiong, W.; Ding, J.; He, Y.; Li, Q. Improved matrix pooling. Stat. Methods Med. Res. 2019, 28, 211–222. [Google Scholar] [CrossRef]
  32. Stone, C.J. Additive regression and other nonparametric models. Ann. Stat. 1985, 13, 689–705. [Google Scholar] [CrossRef]
  33. Hastie, T.; Tibshirani, R. Generalized additive models: Some applications. J. Am. Stat. Assoc. 1987, 82, 371–386. [Google Scholar] [CrossRef]
  34. Xiong, W.; Ding, J.; Zhang, W.; Liu, A.; Li, Q. Nested Group Testing Procedure. Commun. Math. Stat. 2023, 11, 663–693. [Google Scholar] [CrossRef]
  35. Huang, J.; Horowitz, J.L.; Wei, F. Variable selection in nonparametric additive models. Ann. Stat. 2010, 38, 2282. [Google Scholar] [CrossRef]
  36. Yu, W.; Liu, T.; Valdez, R.; Gwinn, M.; Khoury, M.J. Application of support vector machine modeling for prediction of common diseases: The case of diabetes and pre-diabetes. BMC Med. Inform. Decis. Mak. 2010, 10, 16. [Google Scholar] [CrossRef] [PubMed]
  37. Zhou, S.; Shen, X.; Wolfe, D. Local asymptotics for regression splines and confidence regions. Ann. Stat. 1998, 26, 1760–1782. [Google Scholar]
  38. De Boor, C.; De Boor, C. A Practical Guide to Splines; Springer: New York, NY, USA, 1978; Volume 27. [Google Scholar]
Figure 1. An illustration of the test procedure using the Halving algorithm, where filled circles denote positive samples, marked with a “+”, and hollow circles denote negative samples, marked with a “−”.
Figure 1. An illustration of the test procedure using the Halving algorithm, where filled circles denote positive samples, marked with a “+”, and hollow circles denote negative samples, marked with a “−”.
Mathematics 12 00686 g001
Figure 2. The estimators for the functional components in Example 1, simulated using the master pool group testing algorithm. The solid line represents the estimation function, f ^ j , and the dashed line represents the true function, f j .
Figure 2. The estimators for the functional components in Example 1, simulated using the master pool group testing algorithm. The solid line represents the estimation function, f ^ j , and the dashed line represents the true function, f j .
Mathematics 12 00686 g002
Figure 3. The estimate of the non-linear functions using our method, along with the coefficients’ estimate obtained using GLM.
Figure 3. The estimate of the non-linear functions using our method, along with the coefficients’ estimate obtained using GLM.
Mathematics 12 00686 g003
Table 1. Overall performance of our method under four group testing algorithms.
Table 1. Overall performance of our method under four group testing algorithms.
ModelSettingProcedurePETPFP
Example 1
( q n = 50 , s 0 = 4 )
n = 1000 Master Pool0.82 (0.12)3.92 (0.27)3.09 (0.29)
Halving0.83 (0.11)3.96 (0.20)2.98 (0.14)
Dorfman0.76 (0.01)4.00 (0.00)1.98 (0.15)
Array0.80 (0.01)3.99 (0.10)2.20 (0.45)
n = 500 Master Pool1.57 (0.51)3.83 (0.38)3.22 (1.45)
Halving1.45 (0.55)3.89 (0.31)3.18 (1.42)
Dorfman0.95 (0.01)3.90 (0.30)2.94 (0.34)
Array0.71 (0.01)4.00 (0.00)1.01 (0.17)
Example 2
( n = 500 , s 0 = 4 )
q n = 50 Master Pool1.67 (0.32)3.74 (0.44)3.94 (1.78)
Halving1.32 (1.01)3.71 (0.48)3.75 (1.67)
Dorfman1.19 (0.01)3.88 (0.36)1.94 (0.23)
Array1.10 (0.01)3.92 (0.31)1.93 (0.26)
q n = 100 Master Pool1.72 (0.43)3.65 (0.58)3.21 (1.71)
Halving1.52 (0.28)3.57 (0.59)2.72 (1.54)
Dorfman1.45 (0.01)3.64 (0.57)2.02 (1.41)
Array1.29 (0.01)3.85 (0.48)2.48 (1.72)
Example 3
( n = 1000 , s 0 = 3 )
q n = 50 Master Pool1.61 (0.08)2.89 (0.45)3.97 (0.33)
Halving1.64 (0.24)2.85 (0.48)3.51 (0.88)
Dorfman0.60 (0.01)2.91 (0.35)1.73 (0.44)
Array0.54 (0.01)3.00 (0.00)1.87 (0.34)
q n = 100 Master Pool0.92 (0.01)2.89 (0.40)0.00 (0.00)
Halving1.23 (0.15)2.74 (0.50)1.91 (0.29)
Dorfman0.85 (0.01)2.84 (0.44)0.97 (0.17)
Array0.95 (0.08)2.92 (0.37)2.01 (0.18)
Example 4
( n = 500 , s 0 = 3 )
q n = 50 Master Pool1.47 (0.35)2.78 (0.48)3.13 (1.01)
Halving1.48 (0.21)2.69 (0.53)3.16 (1.36)
Dorfman0.81 (0.02)2.93 (0.36)0.44 (0.50)
Array0.86 (0.01)2.91 (0.38)0.46 (1.11)
q n = 100 Master Pool1.52 (0.16)2.69 (0.53)3.17 (1.54)
Halving1.46 (0.28)2.73 (0.51)3.55 (1.55)
Dorfman1.31 (0.02)2.88 (0.40)2.96 (1.97)
Array1.10 (0.01)2.90 (0.39)1.98 (0.83)
Table 2. The estimates of the functions using Example 1 and Example 2.
Table 2. The estimates of the functions using Example 1 and Example 2.
ModelSettingProcedurePEf1PEf2PEf3PEf4PEnon
Example 1
( q n = 50 )
n = 1000 Master Pool0.04 (0.03)0.17 (0.10)0.18 (0.09)0.15 (0.05)0.02 (0.01)
Halving0.05 (0.04)0.16 (0.11)0.17 (0.08)0.14 (0.05)0.01 (0.01)
Dorfman0.03 (0.01)0.06 (0.02)0.08 (0.01)0.11 (0.01)0.02 (0.01)
Array0.04 (0.01)0.06 (0.03)0.09 (0.02)0.12 (0.02)0.02 (0.01)
n = 500 Master Pool0.05 (0.06)0.40 (0.14)0.41 (0.17)0.12 (0.04)0.01 (0.01)
Halving0.04 (0.03)0.23 (0.09)0.31 (0.08)0.10 (0.04)0.02 (0.01)
Dorfman0.03 (0.01)0.17 (0.03)0.17 (0.02)0.07 (0.01)0.01 (0.01)
Array0.01 (0.01)0.16 (0.02)0.24 (0.01)0.05 (0.02)0.01 (0.01)
Example 2
( n = 500 )
q n = 50 Master Pool0.32 (0.06)0.22 (0.06)0.44 (0.17)0.14 (0.03)0.02 (0.01)
Halving0.14 (0.14)0.28 (0.30)0.27 (0.22)0.14 (0.08)0.01 (0.01)
Dorfman0.19 (0.01)0.21 (0.04)0.39 (0.01)0.04 (0.01)0.02 (0.01)
Array0.24 (0.01)0.20 (0.04)0.51 (0.01)0.06 (0.01)0.01 (0.01)
q n = 100 Master Pool0.31 (0.17)0.14 (0.11)0.20 (0.19)0.09 (0.04)0.01 (0.01)
Halving0.36 (0.20)0.13 (0.12)0.18 (0.08)0.09 (0.03)0.02 (0.01)
Dorfman0.05 (0.01)0.12 (0.03)0.08 (0.01)0.06 (0.01)0.01 (0.01)
Array0.05 (0.01)0.15 (0.07)0.09 (0.02)0.07 (0.01)0.01 (0.01)
Table 3. The specific results of Example 3 and Example 4.
Table 3. The specific results of Example 3 and Example 4.
ModelSettingProcedurePEf1PEf2PEf3PEnon
Example 3
( n = 1000 )
q n = 50 Master Pool0.09 (0.02)0.19 (0.26)0.07 (0.24)0.04 (0.02)
Halving0.06 (0.03)0.17 (0.26)0.06 (0.25)0.05 (0.01)
Dorfman0.01 (0.02)0.10 (0.01)0.02 (0.02)0.02 (0.01)
Array0.01 (0.01)0.07 (0.01)0.01 (0.01)0.01 (0.01)
q n = 100 Master Pool0.04 (0.02)0.25 (0.09)0.03 (0.01)0.01 (0.01)
Halving0.02 (0.01)0.13 (0.03)0.06 (0.02)0.01 (0.01)
Dorfman0.03 (0.01)0.24 (0.02)0.04 (0.01)0.02 (0.01)
Array0.03 (0.01)0.12 (0.01)0.03 (0.01)0.01 (0.01)
Example 4
( n = 500 )
q n = 50 Master Pool0.23 (0.14)0.41 (0.21)0.06 (0.07)0.01 (0.01)
Halving0.19 (0.07)0.51 (0.17)0.06 (0.02)0.02 (0.01)
Dorfman0.19 (0.01)0.13 (0.01)0.02 (0.01)0.01 (0.01)
Array0.14 (0.01)0.12 (0.01)0.01 (0.01)0.02 (0.01)
q n = 100 Master Pool0.21 (0.02)0.10 (0.01)0.27 (0.05)0.02 (0.01)
Halving0.22 (0.05)0.08 (0.02)0.32 (0.10)0.01 (0.01)
Dorfman0.06 (0.01)0.15 (0.01)0.13 (0.02)0.01 (0.01)
Array0.09 (0.01)0.16 (0.01)0.26 (0.01)0.01 (0.01)
Table 4. The performance of the models; GLM is q n = 14 in a low dimension, while in our method, q n = 500.
Table 4. The performance of the models; GLM is q n = 14 in a low dimension, while in our method, q n = 500.
MetricsOur MethodGLM
k = 1k = 2k = 5k = 8
ACC0.8296 0.8250 0.8268 0.8277 0.8150 
PPV0.5800 0.5703 0.5761 0.5766 0.5472 
NPV0.9027 0.8993 0.8977 0.9006 0.8952 
Recall0.6360 0.6228 0.6140 0.6272 0.6096 
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zuo, X.; Ding, J.; Zhang, J.; Xiong, W. Nonparametric Additive Regression for High-Dimensional Group Testing Data. Mathematics 2024, 12, 686. https://doi.org/10.3390/math12050686

AMA Style

Zuo X, Ding J, Zhang J, Xiong W. Nonparametric Additive Regression for High-Dimensional Group Testing Data. Mathematics. 2024; 12(5):686. https://doi.org/10.3390/math12050686

Chicago/Turabian Style

Zuo, Xinlei, Juan Ding, Junjian Zhang, and Wenjun Xiong. 2024. "Nonparametric Additive Regression for High-Dimensional Group Testing Data" Mathematics 12, no. 5: 686. https://doi.org/10.3390/math12050686

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop