1. Introduction
Examining controversial attitudes and behaviors via survey often necessitates a careful approach. Rather than affirmatively responding to questions about sensitive or self-incriminating behavior, survey respondents may instead withhold information or provide inaccurate answers. To mitigate the bias that these actions introduce into survey data, Ref. [
1] developed the Randomized Response Technique (RRT) to restore the assurance of confidentiality. Through the deliberate introduction of randomness into survey procedures, randomized response models safeguard respondent privacy and yield more accurate data.
The first binary RRT model was proposed by [
1]. The Warner model prompts participants to answer sensitive questions in direct or indirect form. This method utilizes binary sensitive variables that preserve respondent privacy. Since then, many other models have been proposed, including the model by [
2]. The Greenberg model asks respondents to answer either a direct sensitive question or an unrelated question. In 2023, Ref. [
3] proposed a mixture model that combines the Warner and Greenberg models, treating the two models as special cases of the larger model. This paper also investigated the impact of untruthfulness on binary RRT models, but these models have not accounted for the effect of measurement error. In this paper, we aim to investigate the impact of measurement error and untruthfulness on binary RRT models.
The effect of measurement error has been studied in quantitative RRT models by many authors including [
4,
5]. Ref. [
6] explored the effect of measurement error on the [
1] binary model. This model and the mixture model proposed by [
3] are discussed in detail in
Section 2. We apply the effect of measurement error explored by [
6] onto the Lovig mixture model in
Section 3. In this paper, we aim to study the impact of measurement error in the Lovig mixture model in comparison to the models of [
1] and the [
2].
Certain models may be better suited against the effects of measurement error. In
Section 4, we test our mixture model that accounts for measurement error by introducing measurement error in a numerical study. This study computes several levels of the error that is introduced into the model by the measurement error,
, for different choices of
p and
q. This design compares the mixture model to the Warner and Greenberg models since these two are considered special cases of the Lovig mixture model (
and
, respectively). This investigation relies on our estimation of measurement error
m, which uses a secondary question described in
Section 3.3.2.
In the discussion of privacy in
Section 5, we propose a new measure of privacy for binary RRT models offered by the model. This method uses binary logistic regression to compute the odds ratio (OR) between true participant responses and recorded responses from the model as a measure of predictability. A higher OR value indicates higher individual predictability of the presence of the sensitive trait as the recorded response changes from “No” to “Yes”, thereby offering lower privacy. This privacy measure is estimated in the presence of measurement error and participant untruthfulness.
2. Previous Models
Several RRT models form the building blocks for the proposed model in
Section 3. The Warner model was the first RRT model introduced in [
1]. In this model, participants are presented with either the sensitive question or an indirect version of the sensitive question using a randomization device with known probabilities. The key aspect of RRT models is that the interviewer remains unaware of whether the respondent is answering the direct or the indirect question, ensuring confidentiality. Similarly, the Greenberg model involves respondents responding to either a sensitive question or an entirely unrelated question.
Prior to [
7], all binary RRT models assumed that respondents provide truthful responses. In this work, the authors accounted for a lack of trust and demonstrated that not accounting for untruthfulness leads to poor estimates in the Greenberg model. Later, Ref. [
3] demonstrated the negative impact that untruthfulness has on the efficiency of the Greenberg, Warner, and Lovig mixture binary models. The method of accounting for untruthfulness proposed by [
7] added a node to switch the respondent’s answer with probability
only when the respondent is in the sensitive group with probability
. This design ensures that the respondent’s mistrust of the model will only occur at questions that the respondent finds sensitive. This conceptual framework of accounting for lack of trust was utilized to develop a more efficient and private model.
The Warner model has been shown to offer the highest privacy protection to respondents, and the best efficiency is found in the Greenberg model [
3]. The mixture model proposed in [
3] that combines these two models asks participants an indirect sensitive question, a direct sensitive question, or an unrelated question, each with known likelihoods. This model is discussed and depicted in
Section 2. The approach in [
3] offered an opportunity to compare privacy and efficiency between the Warner model, the Greenberg model, and a mixture of the two accounting for lack of trust. To fairly compare the three models, this work uses a modified version of the unified measure proposed by [
8], which proposed unified measure
. This novel model that accounts for untruthfulness may be improved to also account for measurement error.
Accounting for the effect of measurement error in the model of [
1] was first proposed in [
6]. This model represents the first work exploring measurement error in binary RRT models. This work developed a unique method of defining measurement error in the context of binary RRT models and examined the estimation of the prevalence of sensitive characteristics under measurement error in some cases. The authors found that, in most cases, the measurement error introduced a non-negligible bias into the efficiency offered by the Warner model. This work includes estimators for measurement error and the sensitive trait accounting for measurement error. The approach in this paper examines the same effect of measurement error in the more comprehensive mixture model by [
3].
Mixture Model Accounting for Lack of Trust by [3]
The [
3] mixture model offers a trichotomy that randomly chooses from the direct question, the indirect question, or the unrelated question. This model is best shown by the flow diagram given in
Figure 1 below:
This work found that respondent lack of trust decreases model efficiency and increases privacy when truthfulness remains unaccounted for. Note that the Warner and Greenberg models may be considered as special cases of the Lovig mixture model as pictured above. Comparing the three models, the Greenberg model outperforms other models in terms of efficiency, the Warner model excels in privacy protection, and the Lovig mixture model emerges as the best model in terms of the unified measure .
The Lovig mixture model outperforms both the Greenberg and the Warner models. It exhibits a lower MSE than the Warner model, better privacy protection than the Greenberg model, and significantly better unified measure
than both of the basic models in most recommended cases. This model was groundbreaking in investigating and accounting for the effect of untruthfulness in binary RRT models. The proposed model in this work is an extension of the Lovig model in
Figure 1 to account for measurement error.
3. Proposed Model
We hypothesize that measurement error imposes a significant bias on model efficiency. This section proposes a model that accounts for such effects by estimating it using a secondary question. We acknowledge that this will not account for every type of measurement error. The appropriate methodological guidelines should still be followed to reduce the risk of measurement error confounding study results. These guidelines may include types of survey modes, response formats, respondent training, and response biases.
This paper proposes a model that introduces the effect of measurement error as in [
6] onto the binary RRT mixture model that accounts for lack of trust proposed by [
3]. Following the work of [
7] and utilizing the model proposed by [
3], a respondent in the sensitive group (with probability
) will switch his/her answer due to mistrust with probability
. So far, this design constitutes the Lovig mixture model shown in
Figure 1. To account for measurement error, the model is designed such that each recorded response is switched with probability
m. Nodes that switch the recorded response with rate
m are added on each terminal node because we assume that the measurement error discussed has an equal chance of occurring at each branch.
After demonstrating that poor estimates occur when measurement error is not accounted for in
Section 3.2, we outline a method for estimating the prevalence of measurement error
m and trust rates
A in
Section 3.3 using secondary questions. To begin this discussion, our proposed mixture model is best shown by
Figure 2 below:
For this model, we use the following notation:
;
;
;
;
;
;
;
m = probability that the participant’s recorded response was switched due to measurement error;
3.1. Estimating Trust Parameter A with a Greenberg Model
Before estimating the proportion of the sensitive trait, we must estimate the trust parameter
A. This study uses an initial question to estimate
A using a Greenberg model similar to the approach in [
3]. Since the Greenberg model is the most efficient model and privacy is not prioritized in this initial question, the Greenberg model is the ideal choice to estimate untruthfulness. For this question, let there be the following:
= proportion of the direct question used in a Greenberg model to estimate truthfulness;
= probability of a “Yes” response in this Greenberg model;
= proportion of people who would answer “Yes” to the unrelated question in the Greenberg model.
Following the approach from [
3], we use the question, “Do you trust the model?”, for Question 1 with probability
and an unrelated question with probability
. This leads to the equations:
where
is the proportion of “Yes” responses in the sample of Greenberg responses. Then,
Since
, this question design provides an unbiased estimator for respondent trust prevalence
A.
Now, we discuss the efficiency of the proposed model using the estimator
. In
Section 3.2, we compute the efficiency of the proposed model that does not account for measurement error to test the behavior of the bias. After demonstrating its impact, this initial question that estimates untruthfulness is used as Question 1 in the proposed model in
Section 3.3.
3.2. Proposed Model: Not Accounting for Measurement Error
3.2.1. Efficiency
In this section, we build an estimator for estimator
for the sensitive trait that does not account for measurement error using the probability of a “Yes” response. The probability of a “Yes” response is given by
If the researcher ignores measurement error and erroneously assumes
, our model becomes naive to measurement error. Let
be the probability of a “Yes” response in this naive case. Then, we may find
using (2) as follows:
From this, we have the estimator for the sensitive trait under the naive case:
where
is the proportion of “Yes” responses in the sample assuming random sampling with replacement is given by
Since
is a function of random variables
and
, we use a first-order Taylor’s expansion for
given by
where
,
,
, and
. Then, we have the expansion:
From (
4), we can easily verify
where
is given by (
1) and
is given by (
3). From (5), we may observe the bias that is introduced when measurement error is not accounted for. Bias for this naive approach is given by:
3.2.2. Simulation Results
To simulate randomized response trials, we used the NumPy module in Python to randomly generate survey results based on specified parameters for , , p, q, A, and m. Estimators were then calculated using these generated data. We tabulated the data across trials via the Pandas software library in Python and created visualizations with the Matplotlib plotting library.
The simulations in
Table 1 demonstrate the simulated effect of the bias from measurement error using the Lovig model in
Section 3.2.1. We used MSEs to compare different models. The MSE has been used for this purpose in all major RRT papers including the seminal papers [
1,
2]. Observing the
, which encompasses the variance and bias of
as in (6), the error rates increase in this estimator of
as significant levels of
A and
m are introduced. This positive bias causing poor estimates is especially prevalent when measurement error is introduced. Low levels
do not cause noticeable changes, but a 10% rate of measurement error causes approximately a 5% error rate in the sensitive trait. A similar positive bias may be observed when untruthfulness is introduced, but this is not as severe.
3.3. Proposed Model: Accounting for Measurement Error
Following [
3], we first estimate
A, then the probability of measurement error (
m), and finally, the sensitive trait
.
3.3.1. Approach
Let Question 1 in this approach be the question, “Do you trust the model?”, using the Greenberg model outlined in
Section 3.1. This provides an estimator for
A. This is followed by Questions 2 and 3 to estimate measurement error
m in the model and the prevalence of the sensitive trait
, respectively, using the proposed mixture model.
3.3.2. Estimating Measurement Error Using a Secondary Question
Section 3.2 demonstrates that not accounting for measurement error leads to inaccurate estimates. Following [
6], we estimate the parameter for measurement error
using our model with an additional modified question that ensures a known sensitivity probability.
For example, this test question could be, “Are you a robot?” Such a question ensures a known sensitive probability of for human respondents. Note that truthfulness A does not have to be accounted for in this question since sensitivity is always zero. Researchers must design a question for this estimator that satisfies these two requirements. For this secondary design, we may derive the probability of a “Yes” response in this rigged question using (2), where .
Question 2: (with secondary question) “Are you a robot?”
Using (
7), we can set
Since
, (
8) provides an unbiased estimator of
m. We may then find the MSE given by
3.3.3. Estimating Proportion of Sensitive Trait
Now that we have estimates for
m and
A, Question 3 uses the proposed mixture model from
Figure 2 to ask a direct sensitive, indirect sensitive, or unrelated question. We use the full probability of a “Yes” response
to this question to derive the estimator
.
Question 3: (with mixture model) “Do you have the sensitive trait?”
We use the following Taylor’s approximation to make use of the estimator
:
where
,
,
,
,
, and
. Then,
is approximated by:
From (11), we may note that
where
is given by (
3),
is given by (
9), and
is given by (
1).
3.3.4. Simulation Results
We observed that the estimates for
were more accurate when we accounted for measurement error. The estimates in
Table 2 contrast the poor results shown in
Table 1, where measurement error was ignored. It is now clear that neglecting to account for measurement error in binary RRT models results in inaccurate estimators.
Table 2 simulates the estimators discussed.
The simulation in
Table 2 offers several insights. Our estimators
,
,
, and
are statistically close to their theoretical values, indicating that these are good estimators. Most significantly, the
column confirms that our estimator of
using the proposed model is accurate as untruthfulness and measurement error are introduced. The Greenberg model offers the greatest efficiency indicated by the lowest
rates, and the Warner model offers the worst efficiency. The Lovig mixture model is between these values for each level of
. We will note in
Section 5 that, when privacy protection is also factored in, the mixture model will offer the best performance.
4. Comparison of Measurement Error between Models
Observing the effect of measurement error across models will aid in the researcher’s choice of the model design. Since it is always desired to minimize the impact of measurement error in the chosen model, we discuss the MSE of m for different choices of p and q. To do this, we compare the effect of measurement error between the Lovig, Warner, and Greenberg models.
Each model is differentially affected by measurement error since it is estimated using a secondary question outlined in
Section 3.3.2. Recall that the Warner and Greenberg models are special cases of the Lovig mixture model, where
and
, respectively. The estimator for
m in (
8) is unbiased for all models, so we will compare the measurement error between the models using the respective MSEs.
Numerical Comparison
The simulation results so far have fixed the level of
p and parameterized
q, but the choice of
p has an effect on the model’s performance. The numerical discussion provided in
Figure 3 provides four levels of
p to observe the behavior of
.
Asymptotic behavior is expected around when since we have terms in the denominators of these estimators for . This indicates that should never be chosen close to 0.5 to avoid poor estimators.
When
, the Warner model introduces the highest error in terms of measurement error, while the Greenberg model offers the least error due to measurement error for all considered values of
m. Since this error function is monotonically increasing, the Lovig model offers moderate error as the bold values in
Table 3 demonstrate. When
, it is imperative that the parameter
q be carefully chosen for the Lovig model to avoid large values of
. In this case, the comparative rank of error due to measurement error between models is similar to when
. This is seen in
Table 3 with the
values provided below. This table also shows that
is a good estimator.
Both
Figure 3 and
Table 3 show that the choice of
p and
q has a more significant impact than the level of
m on the amount of error introduced into the model from
. This fact underscores the importance of a proper choice of model parameters since certain models are better suited for accounting for measurement error. To provide a complete recommendation to a researcher designing an RRT study, a discussion on privacy and a unified measure is required. The choice of
involves considering trade-offs between efficiency, privacy protection, and measurement error. This is provided in
Section 6.
5. Privacy of Mixture Model
We provide a brief overview of how privacy is measured in previous binary RRT models and then propose our new privacy measure.
5.1. Previous Work
Model efficiency is not the only performance basis by which researchers should design their studies; respondent privacy is just as important. Without response privacy, respondents may refuse to respond or provide an untruthful response.
5.1.1. Privacy Measure
Ref. [
9] provided a measure of privacy loss as described below. Let:
where
Privacy protection,
introduced by [
10] is defined as
5.1.2. Unified Measure
Ref. [
3] proposed a unified measure of privacy and efficiency using the following metric:
where
a and
b are weights based on the importance the researcher places on privacy and efficiency, respectively. Ref. [
3] assumed
, arguing equal importance to efficiency and privacy. We follow the same approach here.
5.1.3. Privacy of Proposed Model
The following estimators for privacy loss
are derived for the proposed mixture model accounting for measurement error from
Section 3.
5.1.4. Unified Measure of Proposed Model
Table 4 performs a simulation study that compares the three models by efficiency, privacy protection, and [
3]’s unified measure. Privacy is calculated using the traditional method as defined in (
13). The unified measure
is defined in (
14). Several key values are noted in bold.
This is the classical view of model privacy, but now, we pursue a new method of discussing privacy offered by binary RRT models. As expected, the
and
match well with their empirical estimates even in the presence of measurement error. We observed the same conclusions as in [
3] for efficiency, privacy protection, and
in the presence of
m: the Greenberg model is best in terms of efficiency; the Warner model is best in terms of privacy; the Lovig model is best in terms of
.
5.2. Proposed Measure of Privacy for Binary RRT Models
5.2.1. Description
The traditional method of privacy introduced by [
9] can be difficult to interpret. We propose using the odds ratio as a measure of the predictability of participants’ true responses from recorded model responses. The odds ratio (OR) is a statistical measure commonly used in binary logistic regression to quantify the association between two binary variables.
The odds ratio quantifies the change in the odds of the sensitive trait as the reported response changes from a 0 (“No”) to a 1 (“Yes”). Values greater than 1 indicate higher odds of the true response being “Yes”, suggesting better predictability. The odds ratio serves as a valuable metric to assess the predictability of the reported response data from the RRT model.
In our study, we hypothesize that a higher odds ratio corresponds to a lower level of privacy protection offered by the RRT model. Conversely, a lower odds ratio suggests a lower level of predictability and efficiency, but provides greater respondent privacy. We will investigate how different factors, such as the introduction of measurement error and trust in the chosen RRT model, affect the odds ratio.
5.2.2. Privacy Measure Simulation Results
In
Table 5, we simulate the odds ratio in all three models accounting for measurement error.
Table 5 provides several insights into the effect of privacy offered by the three models accounting for measurement error. We observed the same conclusions as in [
3] for privacy with this new approach: the Warner model offers the best privacy; the Greenberg model offers the least privacy; the Lovig model offers moderate to high privacy.
The odds ratio offers greater interpretability for a choice of
compared to
from (
13). For example, a mixture model was conducted where
were chosen as
and
and
and
. The estimated odds ratio for this model would be interpreted by making the claim that the odds of the true answer being “Yes” given that the reported answer was “Yes” is 11.63-times greater than the odds of the true answer being “No”. This contrasts the traditional definition of privacy protection
, where its definition is a conditional probability that may only be interpreted using relative levels between models. While the traditional definition of privacy protection is useful for defining the unified measure
, the logistic regression coefficients hold interpretation value. This new definition for privacy requires no adjustments to the unified measure’s definition.
6. Discussion
The researcher’s choice of the ideal model must take into account measurement error and the unified measure. It has been shown that the effect of measurement error on both model efficiency and privacy is significant. To account for this, the unified measure accounts for measurement error since it uses the updated estimators in this paper. Since is in the equation for , discounts the performance for choices of where increases. Therefore, we recommend utilizing for designing a binary RRT study accounting for measurement error.
After fixing the sample size and preliminary question parameters, for all reasonable possibilities of untruthfulness, measurement error, and the sensitive trait, the Lovig model for measurement error (with parameters
sufficiently far from 0.5) optimizes the unified measure
. This is preserved when the ratio between privacy protection and efficiency is parameterized using
. There are two regions of
that locally optimize
: when
and
, as demonstrated in
Table 4, and when
and
. The former was chosen in the tables because its unified measure is always greater than the latter. Unified measure values for these cases become closer when
(indicating prioritization of privacy protection above efficiency) or under higher rates of untruthfulness and measurement error. All options reduce the overall
values. Researchers should, therefore, choose their parameters in this region so that
is maximized regardless of untruthfulness, measurement error, or sensitive trait levels.
The Lovig mixture model for measurement error outperforms both the Greenberg and the Warner models in terms of the MSE and unified measure in most cases in the presence of untruthfulness and measurement error. The researcher must choose model parameters that optimize the unified measure while ensuring a high rate of participant cooperation. To this point, [
3] noted that the choice from three questions in the mixture model helps improve the respondent cooperation as compared to when they have a choice of two questions, as in the Warner and Greenberg models.
7. Concluding Remarks
The proposed mixture model is the recommended RRT model for collecting sensitive data with the best performance and flexibility for privacy and efficiency compared to existing models. This model accounts for measurement error and untruthful responses using secondary questions to estimate their prevalence. The choice of model parameters significantly impacts both privacy and efficiency. A new logistic regression method is proposed to compare the models comprehensively concerning privacy. The adaptability for conditions of untruthfulness and measurement error allows researchers to choose the most suitable model for their specific needs.
While this paper has outlined a promising design for addressing measurement error within binary RRT models and introduced an innovative privacy measure, several considerations warrant acknowledgment. The real-world implementation of the outlined methods might encounter challenges, particularly with the proposed three-question design. The complexity of this design necessitates a more extensive explanation to participants. This may lead to increased abandonment when combined with sensitive questions, impacting data collection and accuracy. Future research and practical application should address these concerns to ensure the effectiveness and feasibility of these proposed procedures in diverse settings.
This study leverages the comprehensive [
3] mixture model to account for the effects of measurement error. The measurement error’s impact on efficiency and privacy has not been discussed in the published literature on binary RRT models. This development will help to improve the efficacy of future studies that make use of binary RRT models.
Author Contributions
Conceptualization, B.M., V.P., S.G. and S.K.; Methodology, B.M., V.P., S.G. and S.K.; Software, B.M. and V.P.; Formal analysis, B.M., V.P. and S.G.; Resources, S.G.; Writing—original draft, B.M.; Writing—review & editing, B.M., V.P., S.G. and S.K.; Visualization, B.M. and V.P.; Supervision, S.G. and S.K.; Project administration, S.G.; Funding acquisition, S.G. All authors have read and agreed to the published version of the manuscript.
Funding
We would like to thank the National Science Foundation, under grant No. DMS-2244160 for supporting this research.
Data Availability Statement
Acknowledgments
The authors would like to express their deep appreciation to the reviewers for their careful reading of the initial submission and helpful comments, which helped to improve the presentation.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Warner, S.L. Randomized response: A survey technique for eliminating evasive answer bias. J. Am. Stat. Assoc. 1965, 60, 63–69. [Google Scholar] [CrossRef]
- Greenberg, B.G.; Abul-Ela, A.L.A.; Simmons, W.R.; Horvitz, D.G. The unrelated question randomized response model: Theoretical framework. J. Am. Stat. Assoc. 1969, 64, 520–539. [Google Scholar] [CrossRef]
- Lovig, M.; Khalil, S.; Rahman, S.; Sapra, P.; Gupta, S. A mixture binary RRT model with a unified measure of privacy and efficiency. Commun. Stat.-Simul. Comput. 2023, 52, 2727–2737. [Google Scholar] [CrossRef]
- Kumar, S.; Kour, S.P. The joint influence of estimation of sensitive variable under measurement error and non-response using ORRT models. J. Stat. Comput. Simul. 2022, 92, 3583–3604. [Google Scholar] [CrossRef]
- Tiwari, K.K.; Bhougal, S.; Kumar, S.; Rather, K.U.I. Using Randomized Response to Estimate the Population Mean of a Sensitive Variable under the Influence of Measurement Error. J. Stat. Theory Pract. 2022, 16, 28. [Google Scholar] [CrossRef]
- McCance, W.; Gupta, S.; Khalil, S.; Shou, W. Binary Randomized Response Technique (RRT) Models Under Measurement Error. Commun. Stat. Simul. Comput. 2024, 1–8. [Google Scholar] [CrossRef]
- Young, A.; Gupta, S.; Parks, R. A binary unrelated-question rrt model accounting for untruthful responding. Involve J. Math. 2019, 12, 1163–1173. [Google Scholar] [CrossRef]
- Gupta, S.; Mehta, S.; Shabbir, J.; Khalil, S. A unified measure of respondent privacy and model efficiency in quantitative RRT models. J. Stat. Theory Pract. 2018, 12, 506–511. [Google Scholar] [CrossRef]
- Lanke, J. On the degree of protection in randomized interviews. In International Statistical Review/Revue Internationale de Statistique; International Statistical Institute: The Hague, The Netherlands, 1976; Volume 44, No. 2; pp. 197–203. [Google Scholar]
- Fligner, M.A.; Policello, G.E.; Singh, J. A comparison of two randomized response survey methods with consideration for the level of respondent protection. Commun. Stat. Theory Methods 1977, 6, 1511–1524. [Google Scholar] [CrossRef]
| Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).