 Research article
 Open Access
 Published:
Comparison of robustness to outliers between robust poisson models and logbinomial models when estimating relative risks for common binary outcomes: a simulation study
BMC Medical Research Methodology volume 14, Article number: 82 (2014)
Abstract
Background
To estimate relative risks or risk ratios for common binary outcomes, the most popular modelbased methods are the robust (also known as modified) Poisson and the logbinomial regression. Of the two methods, it is believed that the logbinomial regression yields more efficient estimators because it is maximum likelihood based, while the robust Poisson model may be less affected by outliers. Evidence to support the robustness of robust Poisson models in comparison with logbinomial models is very limited.
Methods
In this study a simulation was conducted to evaluate the performance of the two methods in several scenarios where outliers existed.
Results
The findings indicate that for data coming from a population where the relationship between the outcome and the covariate was in a simple form (e.g. loglinear), the two models yielded comparable biases and mean square errors. However, if the true relationship contained a higher order term, the robust Poisson models consistently outperformed the logbinomial models even when the level of contamination is low.
Conclusions
The robust Poisson models are more robust (or less sensitive) to outliers compared to the logbinomial models when estimating relative risks or risk ratios for common binary outcomes. Users should be aware of the limitations when choosing appropriate models to estimate relative risks or risk ratios.
Background
When the outcome of a study is binary, the most common method to estimate the effect is to calculate an odds ratio (OR) as an estimate of relative risk (RR) using a logistic regression. When the outcome prevalence is high (>10%), the OR can still be estimated by using the logistic regression model, but the OR is no longer an acceptable estimate for RR [1]. Interpreting the OR as being equivalent to the RR still occurs in medical research, leading to overstated effect in the study findings [2, 3]. The degree of overstatement depends on the outcome rate (e.g., disease prevalence). A higher outcome rate leads to higher degree of exaggeration. Zhang and Yu proposed a formula to convert the adjusted OR derived from the logistic regression model to a risk ratio in studies with common outcomes [4]. However, the method was noted by McNutt et al. to produce biases in both point estimates and confidence intervals (CIs) [5]. Miettinen suggested the doublingofcases method to estimate OR as an approximation of RR using logistic regression based on a modified dataset [6]. Schouten et al. improved the method by applying robust standard errors [7]. However, this method was mentioned by Skov et al. to produce prevalence greater than one [8]. Several modelbased methods have been proposed to estimate RR and its CI directly [9]. The most popular ones are the robust (also known as modified) Poisson model [10–12] and the logbinomial model [8, 11, 13]. The performance based on simulations seemed to be equally good between the logbinomial model and the robust Poisson model [3, 11, 12, 14] when sample sizes are reasonably large. Out of the two models, it was reported that the robust Poisson may be less affected by outliers compared to the logbinomial method [15]. However, the research in this area is very limited. The purpose of this study is to evaluate the performance of the two methods using simulation in several scenarios when outliers exist and to provide insight into the selection of the appropriate models.
Methods
Estimation methods
The Poisson regression uses a logarithm as the natural link function under the generalized linear model framework. When the outcome is common, the standard Poisson regression over estimates the variance for the measured effect [10–12]. The robust Poisson regression model uses the classical sandwich estimator under the generalized estimation equation (GEE) framework to correct the inflated variance (also known as overdispersion) in the standard Poisson regression. This correction can be achieved by using the REPEATED statement in SAS Proc GENMOD [12] or the ROBUST option in STATA’s Poisson procedure [11]. The estimators based on the robust Poisson models are pseudolikelihood estimators.
The logbinomial regression approach models the probability of having the outcome (e.g., disease) based on the binomial distribution and logrithm of the probability as the link function in a generalized linear model [8, 13]. The logbinomial model attempts to find a MLE if it exists. MLEs are preferred estimators because they carry many good properties including higher efficiency compared to nonMLE estimators. The analyses can be performed using SAS’s GENMOD, R’s GLM or STATA’s GLM procedure. However, for many situations in which quantitative covariates exist, the MLE can be on the boundary of the parameter space (i.e. the predicted probability of the outcome equals to 1), leading to the difficulty of finding the MLE. To address the nonconvergence issue, a SAS macro called “COPY” was developed [16] and later enhanced [17] to increase the chance of finding an approximate MLE inside the interior of the parameter space using PROC GENMOD. The COPY method uses the results of the logbinomial regression if the model converges. When the logbinomial regression fails to converge, the COPY method creates C (usually a large number) copies of the original data and one copy of the original data with the outcome variable reversed (Y converted to 1Y) and then takes the modified data to generate the estimates [16]. The enhanced version of the COPY method simply generated virtual copies by applying weights of C and 1, respectively, to the original dataset and the dataset with the outcome variable reversed [17]. The choice of C impacts the accuracy and the efficiency of the estimates as well as the model convergence. A larger C produces results that are closer to MLEs yet leads to a higher level of difficulty in finding the MLEs due to failure of convergence [17].
Simulation methods
Design of uncontaminated datasets
Let X be a binary treatment/exposure variable (X = 1 for treatment/exposure and X = 0 for nontreatment/nonexposure) from a binomial distribution with the probability of X = 1 fixed at 50%. Let Y be a binary common outcome from a population with the probability of Y = 1 varying from 10%, 25% to 40%. Let Z be a continuous confounder following the Beta (6, 2) distribution. The Beta distribution offers a wide range of possible forms. The parameters chosen (α = 6 and β = 2) provided the distribution with a mean of approximately 0.75. If Z × 100 represents the ages of study subjects, they came from an elderly population with the average age being 75 years old. The relationship between X and Z is defined by the equation logit (p _{ x }) = a _{0} + a _{1} Z, where a _{1} = log(2) or log (4) to indicate moderate or strong association, respectively. The relationship between Y and X is loglinear with adjusted RR being 1.5 or 2. Z is designed as a linear or quadratic confounder.
The value of β _{1} in the models denotes the effect due to exposure on the logrelative risk scale (i.e. β _{1} = log (RR)). To simplify the design, we set β _{2} = α _{1}. There were a total of 24 scenarios (2 RRs, 3 outcomes, 2 levels of association, 2 types of confounders). A detailed description of the 24 scenarios is listed in “Additional file 1”.
Data generating process
We first generated the random variable Z following the Beta (6, 2) distribution for 1,500 subjects. Then we created the exposure variable X for each subject based on the subjectspecific probability of exposure. Finally, the outcome conditional on the exposure status and the covariate was randomly generated with the logbinomial distribution. For each of the 24 scenarios described above, the values of α _{0} and β _{0} were searched iteratively among all the possible values to guarantee that the expose and the outcome rates reached the designed level (50% for exposure and 10%, 25% or 40% for outcome). The α _{0} and β _{0} for each scenario can be found in Additional file 1. To study the differences between the two models in samples of moderate sizes, the same process was repeated to generate another set of data with n = 500.
Contaminating the datasets
For each of the 24 scenarios, the simulation datasets were contaminated by flipping the outcomes (Y becomes 1Y) of the records that had the extreme probabilities of having Y = 1 (i.e., either very low or very high p _{ y }). Models were assessed by using the original simulated datasets and the datasets contaminated at the following level.
Contamination rates:
0% – original datasets
2% – flipping the outcomes of records with p _{ y } at bottom or top 1%
5% – flipping the outcomes of records with p _{ y } at bottom or top 2.5%
Although it is not realistic to expect outliers to solely come from observations with very low or very high probabilities of having the outcome, outliers resulting from flipping the outcome are possible due to documentation or data entry errors. For example, “0” (“not having the disease”) may be erroneously entered into the study database as “1” (“the disease was present”). Compared to adding outliers from a distribution that is different from the underlying one, the flipping approach produced outliers that are more likely to be leverage points. In other words, they are more likely to impact the estimates.
Measures of model performance
For each scenario, the simulation process was repeated 1,000 times to estimate the relative bias, standard error and mean square error (MSE) for the logbinomial model and the robust Poisson model. The relative bias was calculated as the average of the 1,000 estimated RR in log scale minus the log of the true RR divided by the log of the true RR. Standard error was defined as the empirical standard error of the estimated RR in log scale over all 1,000 simulations. MSE was calculated by taking the sum of the squared bias in log scale and the variances.
Due to the nonconvergence issue in standard statistical software, we used the COPY method to generate the estimates of the logbinomial models. However, the accuracy of parameter estimates depends on the number of virtual copies. For this evaluation the number of copies were set to 1,000,000, the default of the COPY method [17]. All the datasets were generated and analyzed using SAS Version 9.2 [18].
Results
Relative bias
Table 1 and Figure 1 revealed the relative biases of the estimated RR in log scale from the two models in each of the 24 scenarios mentioned above when n = 1,500. As expected, both models accurately estimated β _{1} or log (RR) when the simulated databases were not contaminated (level of contamination = 0%). When the data were contaminated (level of contamination > 0%), the relative biases were all negative, indicating that the point estimates were biased towards null. For a fixed RR and an outcome rate, the absolute value of the relative biases increased quickly with the level of contamination. However, the pace of such an increase varied by the outcome rate. Scenarios with lower outcome rates seemed to be associated with more elevated absolute relative biases.
Models with linear confounder Z
When the models contained a linear confounder Z, the two models yielded comparable relative biases (Table 1, Figures 1A and B) except for a few scenarios in which the biases from the logbinomial models were slightly larger than those of the robust Poisson models. The largest differences occurred when the contamination rate was 5% and the outcome rate was 10%, in which the logbinomial models had 3.5%6.7% higher biases compared to the corresponding robust Poisson models.
Models with nonlinear confounder Z
With the nonlinear confounder Z, the robust Poisson model outperformed the logbinomial model when the data was contaminated (Table 1, Figure 1C and D). The difference between the two models was visible even when the level of contamination was only 2%. The magnitude of difference between the two models also varied with the outcome rate. A larger difference was associated with smaller outcomes.
Standard Error (SE)
When the simulated databases were not contaminated (level of contamination = 0%), the SEs of the two models were identical at the 2nd decimal point (Table 2). When the levels of contamination were greater than 0%, the estimated SEs remained comparable between the two models when the outcome rates were high (25% or 40%). However, when the outcome rate was low (10%), the SEs derived from the logbinomial models seemed to be slightly higher than those of robust Poisson in some of the scenarios.
Mean Square Error (MSE)
The same pattern was observed for the MSEs as those of relative biases (Table 3). When the models contained a linear confounder Z, the two models had comparable MSEs except when the outcome rate was 10% and the contamination reached 5%. When the models contained a nonlinear confounder Z, the two models diverged even when the degree of contamination was only 2% (Table 3).
When the simulation was conducted based on samples of moderate sizes (n = 500), the same pattern was observed. As expected, the SEs based on the small samples were larger compared to those derived from large samples (n = 1,500). The results summarized from samples of size 500 were included in “Additional file 2”.
Discussion
In this study, evaluation was performed on the statistical properties of the two most popular modelbased approaches to estimate RR for common binary outcomes in various scenarios when outliers existed. The results suggest that for data coming from a population in which the true relationship between the outcome and the covariate is not in a simple form (e.g., containing a higher order term), the robust Poisson model consistently outperforms the logbinomial model even when the level of contamination is low (e.g., 2%). Statistical software utilizes iterative weighted least squares (IWLS) approach or variations of IWLS to find MLEs for generalized linear models. For logbinomial models, the weights used by the IWLS approach contain the term 1/(1p), where p = exp (X ^{T} β) with a range from 0 to 1 [19]. Lumley et al. pointed out that the MLE of a logbinomial model is likely to be too sensitive to outliers because a very large p (referred to as μ by the authors) has a large influence on the weights, even though the sum of the covariate values are still bounded [15]. In our study both the MLE, generated by the logbinomial models and the pseudolikelihood estimators, produced by the robust Poisson models, were deteriorated when outliers were introduced. However, the level of deterioration differed when the relationships between the confounder and the outcome was not in a simple form, possibly due to the bigger “μ”s yielded by the logbinomial model when the higherorder term of Z was added into the model.
Deddens and Pertersen compared the logbinomial and robust Poisson models by using three reallife examples [20]. Out of the three examples, two produced different point estimates and SEs. In one of these two examples, the difference was nearly two folds for both point estimates and SEs for one of the covariates. The authors concluded that “the decision on which method to use is very important” [20]; however, since the truth was unknown, it is unlikely to tell that between the logbinomial model and the robust Poisson model, which one can yield estimates that are closer to the truth. In one of the two examples in which differences between the two models were observed, the model contained a higherorder (quadratic) term. It is unclear whether or not the complexity of the model degenerated the performance of the models, especially for the logbinomial model.
Of the two methods, the logbinomial method is generally preferred due to the fact that the MLEs estimated by the logbinomial models are more efficient compared to the pseudolikelihood estimators used by the robust Poisson models ([10], page 2300). Spiegelman and Hertzmark recommend using the logbinomial models over the robust Poisson models when convergence is not an issue [21]. Very small differences were observed in a simulation study with a sample size of 100 and a single independent variable with a uniform distribution [14]. When data perfectly follows a logbinomial distribution (i.e., without outliers), the current study did not observe any difference in either biases or variances between the logbinomial and the robust Poisson models for large (n = 1,500) and moderate (n = 500) sample sizes. It appears that the gain in efficiency is beneficial to logbinomial models only for samples of small sizes.
It is not a surprise to observe negative biases when the simulation datasets were contaminated, because flipping the outcomes of the records that have the very low or very high probabilities tend to weaken the associations between the exposure and the outcome leading the associations towards null. The observation of more elevated biases when outcome rate = 10% compared to those of 25% and 40% comes with no surprise either since the impact of flipping on the estimates is expected to be more significant for scenarios with more extreme outcomes (close to 0% or 100%).
Robust methods to detect outliers especially high leverage points for logistic regression models are available in popular statistical software packages [22–24]. However, similar approaches for logbinomial models are not yet available in commercial software packages although the adoption of the diagnostic statistics from those of logistic regression models were demonstrated and applied in a real life example [25]. Efforts to develop goodness of fit tests resulted in reasonable type I errors yet low to moderate power [25]. For this reason, the robust Poisson model seems to be a more attractive choice due to its capability of providing more robust results when outliers are undetected.
For the COPY method, the accuracy of parameter estimates depends on the number of virtual copies. Peterson and Deddens pointed out that “with 10,000 copies the results were generally accurate to three decimal places” in their scenarios [17]. Therefore, the number of virtual copies we used (1,000,000) should provide accuracies that are high enough for our evaluation. The number of virtual copies should be carefully chosen, because a high number of virtual copies may result in failure of convergence.
Occasionally, failure of convergence remains to be an issue for logbinomial models even if the COPY method is applied. Peterson and Deddens [14] included a continuous exposure variable (referred to as the continuous covariate by the authors) when applying the COPY method in the simulation and reported a range of convergence between 70% and 100%. In this study where C, the number of the virtue copies, was set to be 1,000,000, the COPY method converged in all 1,000 simulations in all the 48 scenarios with the linear confounder, and in 30 of out 48 scenarios with the nonlinear confounder. In the 18 scenarios in which the COPY method did not converge in all 1,000 simulations (failed in 1 or more simulations), the converge rates ranged from 0.983 to 0.999 (median 0.996). If the COPY method fails to converge and the maximum likelihoodbased estimators are desired, one can choose the NonLinear Programming (NLP) procedure in SAS [26]. The NLP method is computationally expensive. However, it does not encounter any convergence issues.
Given the lack of robustness of logbinomial models, the authors recommend using robust Poisson models to estimate RR when there are continuous covariates in the model, especially when the covariates are not in a simple linear form. Due to the concern of lack of efficiency for the robust Poisson models for small samples, logbinomial may still be the choice when the sample size is small.
A potential limitation of this study is that complex forms between the confounder and the outcome were generated by a quadratic equation. It is not clear whether or not the findings can be generalized to other complex situations. In addition, all of the outliers generated occurred to records with very low or very high probabilities and such outliers are more likely to be leverage points. The impact of the outliers generated by this study could be more significant compared to that of another study with outliers that were differently distributed.
In summary, the current study revealed the evidence to support the robustness of the robust Poisson model in various scenarios. Further research should focus on the model misspecification due to deviations of underlying probabilities. It is desirable for future studies to develop methods to identify leverage points and efficient goodnessoffit test for logbinomial models.
Conclusions
The robust Poisson models are more robust to outliers compared to the logbinomial models when estimating relative risks or risk ratios for common binary outcomes. Users should be aware of the limitations when choosing appropriate models to estimate relative risks or risk ratios.
Abbreviations
 OR:

Odds ratio
 RR:

Relative risk
 MLE:

Maximum likelihood estimator
 GEE:

Generalized estimation equation
 SE:

Standard error
 MSE:

Mean square error.
References
 1.
Greenland S: Interpretation and choice of effect measures in epidemiologic analyses. Am J Epidemiol. 1987, 125 (5): 761768.
 2.
Agrawal D: Inappropriate interpretation of the odds ratio: oddly not that uncommon. Pediatrics. 2005, 116: 16121613.
 3.
Mirjam JK, Cessie SL, Algra A, Vandenbroucke JP: Overestimation of risk ratios by odds ratios in trials and cohort studies: alternatives to logistic regression. Can Med Assoc J. 2012, 184 (8): 895899. 10.1503/cmaj.101715.
 4.
Zhang J, Yu KF: What’s relative risk? A method of correcting the odds ratio in cohort studies of common outcomes. JAMA. 1998, 280: 16901691. 10.1001/jama.280.19.1690.
 5.
McNutt LA, Wu C, Xue X, Hafner JP: Estimating the relative risk in cohort studies and clinical trials of common outcomes. Am J Epidemiol. 2003, 157 (10): 940943. 10.1093/aje/kwg074.
 6.
Miettinen O: Design options in epidemiologic research. An update. Scand J Work Environ Health. 1982, 8 (Suppl 1): 714.
 7.
Schouten EG, Dekker JM, Kok FJ, Le Cessie S, Van Houwelingen HC, Pool J, Vanderbroucke JP: Risk ratio and rate ratio estimation in casecohort designs: hypertension and cardiovascular mortality. Stat Med. 1993, 12: 17331745. 10.1002/sim.4780121808.
 8.
Skov T, Deddens J, Petersen MR, Endahl L: Prevalence proportion ratios: estimation and hypothesis testing. Int J Epidemiol. 1998, 27: 9195. 10.1093/ije/27.1.91.
 9.
Greenland S: Modelbased estimation of relative risks and other epidemiologic measures in studies of common outcomes and in case–control studies. Am J Epidemiol. 2004, 160 (4): 301305. 10.1093/aje/kwh221.
 10.
Stijnen T, Van Houwelingen HC: Relative risk, risk difference and rate difference models for sparse stratified data: a pseudo likelihood approach. Stat Med. 1993, 12 (24): 22852303. 10.1002/sim.4780122406.
 11.
Barros AJ, Hirakata VN: Alternatives for logistic regression in crosssectional studies an empirical comparison of models that directly estimate the prevalence ratio. BMC Med Res Method. 2003, 3 (21): 113.
 12.
Zou G: A modified poisson regression approach to prospective studies with binary data. Am J Epidemiol. 2004, 159 (7): 702706. 10.1093/aje/kwh090.
 13.
Wacholder S: Binomial regression in GLIM: estimating risk ratios and risk differences. Am J Epidemiol. 1986, 123: 174184.
 14.
Petersen MR, Deddens JA: A comparison of two methods for estimating prevalence ratios. BMC Med Res Method. 2008, 8 (1): 910.1186/1471228889.
 15.
Lumley T, Kronmal RA, Ma S: Relative risk regression in medical research: models, contrasts, estimators and algorithms. UW Biostatistics Working Paper Series 293. Seattle, WA: University of Washington, http://www.bepress.com/uwbiostat/paper293. Accessed June 20, 2014
 16.
Deddens JA, Petersen MR, Lei X: In Proceedings at the SAS Users Group International Proceedings: March 30 SUGI28. Estimation of Prevalence Ratios When PROC GENMOD Does not Converge. 2003, Washington: Seattle
 17.
Petersen MR, Deddens JA: A revised SAS macro for maximum likelihood estimation of prevalence ratios using the COPY method [letter]. Occup Environ Med. 2009, 66 (9): 639
 18.
Institute SAS: Software Version 9.2. 2008, Cary, North Carolina: SAS Institute
 19.
McCullagh P, Nelder J: Generalized Linear Models. 1989, New York: John Wiley & Sons, Secondth
 20.
Deddens JA, Petersen MR: Approaches for estimating prevalence ratios. Occup Environ Med. 2008, 65: 501506. 10.1136/oem.2007.034777.
 21.
Spiegelman D, Hertzmark E: Easy SAS calculations for risk or prevalence ratios and differences. Am J Epidemiol. 2005, 162: 199200. 10.1093/aje/kwi188.
 22.
Hosmer DW, Lemeshow S: Applied Logistic Regression. 2000, New York: John Wiley & Sons, 2
 23.
Imon AHMR, Hadi AS: Identification of multiple high leverage points in logistic regression. J Appl Stat. 2013, 40 (12): 26012616. 10.1080/02664763.2013.822057.
 24.
Syaiba BA, Habshah M: Robust logistic diagnostic for the identification of high leverage points in logistic regression model. J Appl Sci. 2010, 10 (23): 20423050.
 25.
Blizzard L, Hosmer DW: Parameter estimation and goodnessoffit in log binomial regression. Biometrical J. 2006, 48 (1): 522. 10.1002/bimj.200410165.
 26.
Yu B, Wang Z: Estimating relative risks for common outcome using PROC NLP. Comput Meth Prog Bio. 2008, 90 (2): 179186.
Prepublication history
The prepublication history for this paper can be accessed here:http://0www.biomedcentral.com.brum.beds.ac.uk/14712288/14/82/prepub
Acknowledgements
We would like to thank Bonnie Li, M.S. for the programming assistance.
Author information
Affiliations
Corresponding author
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
WC conceived and carried out the study, and drafted the manuscript. JS participated in the design, data generation and interpretation of the analyses. LQ participated in the design, simulation and interpretation of the analyses. SA participated in the design and provided guidance. All the authors read and approved the final manuscript.
Authors’ original submitted files for images
Below are the links to the authors’ original submitted files for images.
Rights and permissions
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly credited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
Chen, W., Shi, J., Qian, L. et al. Comparison of robustness to outliers between robust poisson models and logbinomial models when estimating relative risks for common binary outcomes: a simulation study. BMC Med Res Methodol 14, 82 (2014). https://0doiorg.brum.beds.ac.uk/10.1186/147122881482
Received:
Accepted:
Published:
DOI: https://0doiorg.brum.beds.ac.uk/10.1186/147122881482
Keywords
 Relative risk
 Risk ratio
 Logbinomial regression
 Robust poisson regression
 Outliers
 Common binary outcomes