Prospective evaluation of direct approach with a tablet device as a strategy to enhance survey study participant response rate
© Parker et al.; licensee BioMed Central Ltd. 2012
Received: 4 July 2012
Accepted: 29 October 2012
Published: 31 October 2012
Investigators conduct survey studies for a variety of reasons. Poor participant response rates are common, however, and may limit the generalizability and utility of results. The objective of this study was to determine whether direct approach with a tablet device enhances survey study participant response rate and to assess participants’ experiences with this mode of survey administration.
An interventional study nested within a single center survey study was conducted at McMaster Children’s Hospital. The primary outcome was the ability to achieve of a survey study response rate of 70% or greater. Eligible participants received 3 email invitations (Week 0, 2, 4) to complete a web-based (Survey Monkey) survey. The study protocol included plans for a two-week follow-up phase (Phase 2) where non-responders were approached by a research assistant and invited to complete an iPad-based version of the survey. The Phase 1 response rate was 48.7% (56/115). Phase 2 effectively recruited reluctant responders, increasing the overall response rate to 72.2% (83/115). On a 7-point Likert scale, reluctant responders highly rated their enjoyment (mean 6.0, sd 0.83 [95% CI: 5.7-6.3]) and ease of use (mean 6.7, sd 0.47 [95% CI: 6.5-6.9]) completing the survey using the iPad. Reasons endorsed for Phase 2 participation included: direct approach (81%), immediate survey access (62%), and the novelty of completing a tablet-based survey (54%). Most reluctant responders (89%) indicated that a tablet-based survey is their preferred method of survey completion.
Use of a tablet-based version of the survey was effective in recruiting reluctant responders and this group reported positive experiences with this mode of survey administration.
KeywordsResearch design Data collection Computers Handheld Bias (epidemiology)
Survey-based research is commonly undertaken to gain knowledge and understanding of the attitudes, preferences and self-reported practices of individuals within a population . From a clinical research perspective, survey data can provide valuable information to investigators prior to undertaking further study in a particular area. Results obtained from a well-conducted survey study can inform investigators on a number of important feasibility considerations, including level of interest in a topic, current practice patterns, intervention(s) acceptable to practicing clinicians, and willingness to enroll participants into a study .
A major hurdle in conducting survey-based research is obtaining an adequate response rate from among those invited to participate. A good response rate, considered 70% or higher, is important to minimize the risk of selection bias [3, 4]. Guarding against selection bias is important to increase confidence that the study results are generalizable to the entire population surveyed . Previous work has demonstrated that survey non-respondents may differ from respondents in important ways . Investigators should therefore use all reasonable means to encourage as many eligible subjects as possible to participate in any survey study conducted.
Researchers frequently opt for a survey study design involving postal and/or electronic distribution of self-administered surveys for reasons of cost and convenience [7, 8]. Factors known to impact on survey study response rates include survey length, appearance, pre-notification, inclusion of monetary or non-monetary incentives, academic sponsorship, and follow-up . Beyond strategies previously demonstrated to improve survey study response rate, investigators may consider using other creative means to engage and recruit eligible participants [10, 11]. In this paper we describe use of a tablet device to enhance participant response rate in a single center survey study involving a self-administered questionnaire.
A single center survey study was conducted at McMaster University Medical Centre (Hamilton, Canada) between January and May 2012. The purpose of the survey study was to assess health care providers’ attitudes and beliefs regarding pediatric fluid resuscitation practices. The sampling frame included all nursing and physician staff of the Pediatric Emergency Department (PED) and Pediatric Critical Care Unit (PCCU) of McMaster Children’s hospital (n = 115). We report here the results of a nested study, which prospectively evaluated the effectiveness of direct approach with a tablet-based version of the survey to improve the response rate and participants’ experiences with this approach to survey administration.
For the primary outcome of this nested study, we sought to determine whether use of a recruitment strategy involving a tablet-based version of the survey was effective in improving the response rate to the desirable 70%-plus level within a 2-week time frame. Secondary outcomes related to the assessment of respondents’ enjoyment, ease of use, and experience of any technical problems while using the tablet to complete the survey. We also planned to examine differences in participant characteristics according to study phase and to compare the characteristics of responders and non-responders.
The Faculty of Health Sciences/Hamilton Health Sciences Research Ethics Board approved this study and all procedures were conducted in accordance with the Tri-council Policy Statement . The full study protocol called for the survey to be administered in two distinct phases. In Phase 1, eligible participants were invited to complete the survey at 0, 2, and 4 weeks via an email invitation that included a hyperlink to a web-based (Survey Monkey) version of the survey. An electronic version of the study information and consent sheet was provided as an attachment along with the study invitation email. The email invitation advised eligible participants of an incentive (chance to win a coffee card of $25-value). Participants were instructed to complete the survey only once. The protocol included provisions to proceed with a second phase of survey distribution should a response rate of 70% or more not be achieved by 1 month following the final email inviting survey participation.
In Phase 2 a pre-notification email was sent to all individuals in the sampling frame advising that a research assistant would be visiting the PED and PCCU over a 2-week period to offer a final opportunity to complete the survey. Potentially eligible subjects were approached in person by a research assistant and offered a paper copy of the information and consent sheet inviting study participation. Those approached were advised of the participation incentive, which was the same as that offered during Phase 1. Individuals were considered potentially eligible subjects if they were present and/or working in the relevant clinical environment and appeared available to be approached. Interested persons were screened for eligibility and consent was obtained from qualifying individuals who had not previously completed the survey. Eligible subjects were informed that they were being invited to complete an iPad-based version of the survey.
The research assistant presented consenting participants with an Apple iPad2 tablet device, trademark of Apple Inc. (32 GB with Wi-Fi, Apple model A1395), to facilitate completion of the tablet-based version of the survey. The iPad-based version of the survey was created using a web-based program, iSURVEY (iSurveySoft, Wellington, NZ), and administered using an iPad application by the same name. This survey program differs from the Survey-Monkey program in that iSURVEY facilitates transfer of the survey onto the iPad with appropriate formatting. This application also permits access to and completion of the survey when the iPad is not connected to the Internet, allowing the survey to be iPad-based rather than web-based. Completed surveys are stored on the iPad until an Internet connection is reestablished, at which time, completed surveys are uploaded to the company website. Completed surveys located (temporarily) on the iPad cannot be accessed once completed and are fully secure. This program allows for many surveys to be completed using an iPad while not connected to the Internet and then subsequently uploaded at a later time that is convenient. This significantly increases the flexibility of survey administration, as the iPad is small and portable. Further details regarding the nature of this application can be found on the company website, https://www.isurveysoft.com.
The Phase 2 version of the survey was modified in that 5 questions eliciting participants’ experience completing the iPad-based version of the survey were included following the original slate of 15 questions. On a 7-point Likert scale, participants were asked to rate their level of enjoyment (1 = no enjoyment; 4 = neutral; 7 = it was an amazing experience) and ease of use (1 = it was extremely difficult; 4 = neutral; 7 it was extremely easy) completing the iPad-based version of the survey. A free text question also asked whether respondents experienced any technical problems while attempting to complete the survey. Given that our Phase 2 respondents were reluctant responders, we included questions asking why they elected to participate at this time but not when previous invitations had been sent. We also asked them to report their preferred method of survey completion.
The primary outcome analysis consisted of determination of achievement of a 70% response rate based on the calculated proportion of surveys completed. Descriptive data regarding incomplete surveys and item response rate are reported. Binary logistic regression was used to obtain the odds ratio point estimate for submission of incomplete surveys in Phase 1 as well as the odds ratio point estimate for missing item responses. Participant characteristics and survey data are summarized using mean (standard deviation) for continuous variables and count (percent) for categorical variables, with 95% confidence intervals (CI) included where appropriate. Binary logistic regression was used to generate univariate odds ratio point estimates and 95% confidence intervals for participant characteristics, with Phase 1 survey completion as the dependent variable. We planned a priori to include significant variables into a multivariable model to determine independent effects, if appropriate. In all cases, significance was determined at the p < 0.05 level. Collected survey data were analyzed using IBM SPSS statistics, Version 20 (IBM, Armonk, New York, USA). Data related to non-respondent characteristics were analyzed using an on-line calculator for 2 x 2 contingency tables .
Survey response rate by study phase
Survey Study Phase
Number of Participant Responses Received
Cumulative Number of Participant Responses Received and Response Rate (%)
Participant characteristics by survey study phase
Phase 1 N(%)
Phase 2 N(%)
Odds Ratio Point Estimate [95% CI]
Staff Physician or Fellow
Primary Area of Practice
Pediatric Emergency Medicine
Pediatric Critical Care Medicine
Part Time or Occasional Staff
Years of Experience in Pediatric Emergency of Pediatric Critical Care Medicine
Less than 2 years
10 years or more
Pediatric Resuscitation Experience*
None or Minimal experience
Experienced or Very Experienced
The limited data available regarding non-respondents reveals that these individuals were more likely to work as part time or occasional staff (17/34 (50.0%) vs. 18/81 (22.2%), odds ratio point estimate 3.5 [95% CI: 1.49-8.21]. Physicians were no less likely to be non-responders when compared to nurses (7/21 (33.3%) vs. 26/94 (27.7%), odds ratio 1.31, [95% CI: 0.42-3.99]) and the odds of non-response was not significantly influenced by primary area of practice (19/51 (37.3%) for PED vs. 14/64 (21.9%) for PCCU, odds ratio 2.12 [95% CI: 0.87-5.23]).
Reasons endorsed by reluctant responders regarding why they elected to participate during Phase 2
Factors Influencing Study Participation
Participants Selecting Item N (%)
Being approached/invited in person was important
Ability to immediately access the survey was important
The novelty of completing the survey using the iPad
The fact that internet access was not required
I tend to delete any email that invites survey study participation
Preferred mode of survey administration of reluctant responders
Survey administration method
Number of respondents selecting N (%)
Hard Copy (Pen and Paper)
Internet/Web-based survey e.g. Survey Monkey
Tablet Device e.g. iPad
Smart Phone Device e.g. iPhone
Telephone (Verbal) Interview
Face-to-Face (In person, verbal) Interview
An ongoing challenge for investigators is to find effective ways to engage potential study participants in survey-based research. Recent data suggests that survey response rates are in fact deteriorating , and that surveys of physicians in particular are at risk for low response rates [5, 15–18]. While monetary incentives appear to be effective in increasing response rates [9, 17, 18], this can add substantially to study costs depending on the value of the incentive and the number of participants surveyed. In this single center survey study of physician and nursing staff we demonstrate that direct approach with a tablet device is an effective strategy to increase participant response rate into the desirable 70%-plus range and that this mode of survey administration is well received by reluctant responders.
To our knowledge, this is the first report describing use of a tablet device as a tool to enhance participant response rate in a survey study of health care professionals. A search of OVID MEDLINE (R) 1946 to June Week 1, 2012, under the MeSH heading “Computers, Handheld” revealed 1783 results, with 64 unique citations also mapping to the “Data Collection” MeSH heading. Published reports describe a variety of types and uses of Personal Digital Assistants (PDAs) for the purpose of data collection including use in remote fieldwork [19–23], household surveys [22, 24], patient data diaries [25–28], observational research [29–32], quality assurance [33, 34], and laboratory data collection . Good agreement has been found between paper-based and PDA-based survey data [36–38], with PDA-based methods yielding comparatively greater data accuracy and less missing data [36, 39].
In the present study, several findings are of interest in addition to the primary outcome result. First, the resuscitation experience of respondents significantly differed by study phase, suggesting that individuals with greater confidence in the survey subject matter were more likely to participate when initially invited. Secondly, there was a trend toward increased recruitment of physicians in Phase 1 of the study. While non-significant, the odds ratio point estimate was more than seven, with a wide confidence interval owing to our small sample size. It is possible that participating physicians may have been more inclined to complete the survey during Phase 1 due to their level of experience and comfort with the subject matter. Conversely, the research assistant may have had more difficulty accessing physicians or felt less comfortable approaching them to invite study participation during Phase 2. Lastly, direct approach, immediate survey access, and the novelty of this survey strategy were endorsed by a majority of reluctant responders as aiding in their recruitment. Given that most reluctant responders indicated that a tablet-based mode of survey completion is their preference, this strategy should be further evaluated in the context of a larger study of health care providers.
It will be important to clarify in future work that use of this strategy actually improves representativeness and does not exacerbate response bias. Direct approach with a tablet device resulted in the recruitment of units that otherwise would have been missing, and our data indicate that some differences in participant characteristics existed between the two study phases. This, however, does not answer the question of whether the data that would have been missing (or remains missing) is missing at random (MAR) or not, which is the major determinant of whether survey results are representative of the total population surveyed [40, 41]. From a practical perspective, there is no way to determine whether data is MAR or not unless follow up data from nonrespondents can be obtained . Without complete follow up data from nonrespondents, it is difficult to determine whether any resulting changes to the outcome analysis results of interest represent a movement toward the “truth” or away from it. Further complicating matters is the fact that missingness alone is not the only potential source of bias in survey results. Obtaining survey data that is inaccurate may be just as problematic .
There are several limitations to this study. First, our sample size was small and this study was conducted at a single site. While sufficient to demonstrate proof-of-concept, the small sample size limited study power. Secondly, our findings may not be generalizable to other settings. It is possible that this type of protocol is neither feasible nor cost efficient in larger survey studies where multiple tablet devices (and research assistants) would be required. No additional cost outlay was required in our study as the iPad tablet had already been purchased for research use and a student volunteer functioned as the research assistant. Thirdly, based on our data it is not possible to tease out the impact of the tablet independent from the effects attributable to direct approach as these occurred together. Finally, the fact that non-respondents were more likely to work as part-time or occasional staff is important to note. It is likely that making contact with these individuals in the workplace setting was less probable, limiting the opportunity to invite their participation through direct approach.
We conclude that use of a strategy involving direct approach with a tablet-based survey is effective in enhancing participant response rate to the desirable 70%-plus level. Further study to assess whether this improved recruitment increases overall representativeness is warranted. Reluctant responders reported positive experiences with this mode of survey administration and no technical issues were encountered.
MP is an Assistant Professor of Pediatrics at McMaster University and an Adjunct Clinical Assistant Professor of Pediatrics at the University of Toronto. AM is currently an MSc student in Biomedical Engineering at McMaster University. SU is currently a BSc student in Life Sciences at McMaster University.
MP currently holds research funding in the form of McMaster (MAC) new faculty research start-up grant and as the sponsor for a McMaster Children’s Hospital Resident Research Award. Funds from the McMaster (MAC) new faculty research start-up grant were used to support this work. MP currently receives salary support from McMaster Children’s Hospital and McMaster University. Costs related to use of the iSURVEY program for one month is normally $89.00 USD. The fee for use of this program was waived due to the intent to use this for academic research purposes related to use of mobile technology use in the academic research setting. AM has no funding to declare. SU has no funding to declare. No funding body or institution was involved in the design, collection, analysis or interpretation of the data or in the writing of the manuscript.
Pediatric Emergency Department
Pediatric Critical Care Unit
Personal Digital Assistant
Credit: iPad is a trademark of Apple Inc., registered in the U.S. and other countries.
Credit: iSURVEY is a copyrighted software program of iSurveySoft, Wellington, New Zealand.
Disclaimer: BMC Research Notes is an independent publication and has not been authorized, sponsored, or otherwise approved by Apple Inc.
Disclaimer: None of the authors hold any financial interests in the company iSurveySoft or the iSURVEY program utilized in this study.
- Dilman D: Mail and telephone surveys: The total design method. 1978, Wiley-Interscience, New York
- Burns KE, Duffett M, Kho ME, Meade MO, Adhikari NK, Sinuff T, Cook DJ: A guide for the design and conduct of self-administered surveys of clinicians. CMAJ. 2008, 179 (3): 245-252. 10.1503/cmaj.080372.PubMedPubMed CentralView Article
- Henry R, Zivick J: Principles of survey research. Fam Pract Res J. 1986, 5 (3): 145-157.PubMed
- Rubenfeld G: Surveys: an introduction. Respir Care. 2004, 49 (10): 1181-1185.PubMed
- McColl E, Jacoby A, Thomas L, Soutter J, Bamford C, Steen N, Thomas R, Harvey E, Garratt A, Bond J: Design and use of questionnaires: a review of best practices applicable to surveys of health care staff and patients. Health Technol Assess. 2001, 5 (31): 1-256.PubMedView Article
- Etter J, Perneger T: Analysis of non-response bias in a mailed health survey. J Clin Epidemiol. 1997, 50 (10): 1123-1128. 10.1016/S0895-4356(97)00166-2.PubMedView Article
- Edwards P, Roberts I, Clarke M, DiGuiseppi C, Pratap S, Wentz R, Kwan I: Increasing response rates to postal questionnaires: systematic review. BMJ. 2002, 324: 1183-1191. 10.1136/bmj.324.7347.1183.PubMedPubMed CentralView Article
- Dilman D: Mail and internet surveys: The total design method. 2007, Wiley, New York, 2
- Edwards P, Roberts I, Clarke M, DiGuiseppi C, Wentz R, Kwan I, Cooper R, Felix L, Pratap S: Methods to increase response to postal and electronic questionnaires (Review). Cochrane Database Syst Rev. 2009, Art. No.: MR000008-10.1002/14651858. pub4, 3
- Ziegenfuss JY, Burmeister K, James K, Haas L, Tilburt J, Beebe TJ: Getting physicians to open the survey: Little evidence that an envelope teaser increases response rates. BMC Med Res Methodol. 2012, 12 (1): 41-10.1186/1471-2288-12-41.PubMedPubMed CentralView Article
- Olsen F, Abelsen B, Olsen JA: Improving response rate and quality of survey data with a scratch lottery ticket incentive. BMC Med Res Methodol. 2012, 12 (1): 52-10.1186/1471-2288-12-52.PubMedPubMed CentralView Article
- Tri-Council Policy Statement: Canadian Institutes of Health Research, Natural Sciences and Engineering Research Council of Canada. Ethical Conduct for Research Involving Humans. 2011, Social Sciences and Humanities Research Council of Canada, Canada, 2
- Interactive Statistics page: 2-way contingency table analysis.http://statpages.org/ctab2x2.html,
- Groves R: Nonresponse rates and nonresponse bias in household surveys. Public Opin Q. 2006, 70 (5): 646-675. 10.1093/poq/nfl033.View Article
- Asch D, Jedrziewski M, Christakis N: Response rates to mail surveys published in medical journals. J Clin Epidemiol. 1997, 50 (10): 1129-1136. 10.1016/S0895-4356(97)00126-1.PubMedView Article
- Grava-Gubins I, Scott S: Effect of various methodologic strategies. Can Fam Physician. 2008, 54: 1424-1430.PubMedPubMed Central
- Thorpe C, Ryan B, McLean SL, Burt A, Stewart M, Brown JB, Reid GJ, Harris S: How to obtain excellent response rates when surveying physicians. Fam Pract. 2009, 26 (1): 65-68.PubMedView Article
- VanGeest J, Johnson T, Welch V: Methodologies for improving response rates in surveys of physicians: a systematic review. Eval Health Prof. 2007, 30 (4): 303-321. 10.1177/0163278707307899.PubMedView Article
- Onono MA, Carraher N, Cohen RC, Bukusi EA, Turan JM: Use of personal digital assistants for data collection in a multi-site AIDS stigma study in rural south Nyanza, Kenya. Afr Health Sci. 2011, 11 (3): 464-473.PubMedPubMed Central
- Kelly GC, Hii J, Batarii W, Donald W, Hale E, Nausien J, Pontifex S, Vallely A, Tanner M, Clements A: Modern geographical reconnaissance of target populations in malaria elimination zones. Malar J. 2010, 9: 289-10.1186/1475-2875-9-289.PubMedPubMed CentralView Article
- Yu P, de Courten M, Pan E, Galea G, Pryor J: The development and evaluation of a PDA-based method for public health surveillance data collection in developing countries. Int J Med Inform. 2009, 78 (8): 532-542. 10.1016/j.ijmedinf.2009.03.002.PubMedView Article
- Byass P, Hounton S, Ouedraogo M, Some H, Diallo I, Fottrell E, Emmelin A, Meda N: Direct data capture using hand-held computers in rural Burkina Faso: experiences, benefits and lessons learnt. Trop Med Int Health. 2008, 13 (Suppl 1): 25-30.PubMedView Article
- Acosta JA, Hatzigeorgiou C, Smith LS: Developing a trauma registry in a forward deployed military hospital: Preliminary report. J Trauma. 2006, 61 (2): 256-260. 10.1097/01.ta.0000230280.69849.97.PubMedView Article
- Vanden Eng JL, Wolkon A, Frolov AS, Terlouw DJ, Eliades MJ, Morgah K, Takpa V, Dare A, Sodahlon YK, Doumanou Y: Use of handheld computers with global positioning systems for probability sampling and data entry in household surveys. Am J Trop Med Hyg. 2007, 77 (2): 393-399.PubMed
- Luckmann R, Vidal A: Design of a handheld electronic pain, treatment and activity diary. J Biomed Inform. 2010, 43 (5 Suppl): S32-36.PubMedView Article
- Bray P, Bundy AC, Ryan MM, North KN: Feasibility of a computerized method to measure quality of “everyday” life in children with neuromuscular disorders. Phys Occup Ther Pediatr. 2010, 30 (1): 43-53. 10.3109/01942630903294687.PubMedView Article
- Langan SM, Williams HC: The use of electronic diaries in dermatology research. Dermatology. 2009, 218 (4): 344-346. 10.1159/000203991.PubMedView Article
- Marceau LD, Link C, Jamison RN, Carolan S: Electronic diaries as a tool to improve pain management: is there any evidence?. Pain Med. 2007, 8 (Suppl 3): S101-109.PubMedView Article
- Hlady CS, Severson MA, Segre AM, Polgreen PM: A mobile handheld computing application for recording hand hygiene observations. Infect Control Hosp Epidemiol. 2010, 31 (9): 975-977. 10.1086/655834.PubMedPubMed CentralView Article
- Tejani N, Dresselhaus TR, Weinger MB: Development of a hand-held computer platform for real-time behavioral assessment of physicians and nurses. J Biomed Inform. 2010, 43 (1): 75-80. 10.1016/j.jbi.2009.08.011.PubMedView Article
- Guryan J, Jacob B, Klopfer E, Groff J: Using technology to explore social networks and mechanisms underlying peer effects in classrooms. Dev Psychol. 2008, 44 (2): 355-364.PubMedView Article
- Asaro PV: Synchronized time-motion study in the emergency department using a handheld computer application. Stud Health Technol Inform. 2004, 107 (Pt 1): 701-705.PubMed
- Fu Q, Xue Z, Zhu J, Fors U, Klein G: Anaesthesia record system on handheld computers–pilot experience and uses for quality control and clinical guidelines. Comput Methods Programs in Biomed. 2005, 77 (2): 155-163. 10.1016/j.cmpb.2004.09.003.View Article
- Benson K: Electronic Data Collection Tools for Quality Improvement: Antibiotic Utilization Project. AMIA Annu Symp Proc. 2003, 2003: 788-PMCID: PMC1480294PubMed Central
- Blaya JA, Gomez W, Rodriguez P, Fraser H: Cost and implementation analysis of a personal digital assistant system for laboratory data collection. Int J Tuberc Lung Dis. 2008, 12 (8): 921-927.PubMed
- Bernabe-Ortiz A, Curioso WH, Gonzales MA, Evangelista W, Castagnetto JM, Carcamo CP, Hughes JP, Garcia PJ, Garnett GP, Holmes KK: Handheld computers for self-administered sensitive data collection: a comparative study in Peru. BMC Med Inform Decis Mak. 2008, 8: 11-10.1186/1472-6947-8-11.PubMedPubMed CentralView Article
- Gwaltney CJ, Shields AL, Shiffman S: Equivalence of electronic and paper-and-pencil administration of patient-reported outcome measures: a meta-analytic review. Value Health. 2008, 11 (2): 322-333. 10.1111/j.1524-4733.2007.00231.x.PubMedView Article
- Fletcher LA, Erickson DJ, Toomey TL, Wagenaar AC: Handheld computers. A feasible alternative to paper forms for field data collection. Eval Rev. 2003, 27 (2): 165-178. 10.1177/0193841X02250527.PubMedView Article
- Galliher JM, Stewart TV, Pathak PK, Werner JJ, Dickinson LM, Hickner JM: Data collection outcomes comparing paper forms with PDA forms in an office-based patient survey. Ann Fam Med. 2008, 6 (2): 154-160. 10.1370/afm.762.PubMedPubMed CentralView Article
- Rubin DB: Inference and missing data. Biometrika. 1976, 63 (3): 581-592. 10.1093/biomet/63.3.581.View Article
- Schafer JL, Graham JW: Missing data: Our view of the state of the art. Psychol Methods. 2002, 7 (2): 147-177.PubMedView Article
- Egleston BL, Miller SM, Meropol NJ: The impact of misclassification due to survey response fatigue on estimation and identifiability of treatment effects. Stat Med. 2011, 30 (30): 3560-3572. 10.1002/sim.4377.PubMedPubMed CentralView Article
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.