EconPapers    
Economics at your fingertips  
 

The Effect of Item Screeners on the Quality of Patient Survey Data

Hector Rodriguez (), Ted Glahn, Angela Li, William Rogers and Dana Safran

The Patient: Patient-Centered Outcomes Research, 2009, vol. 2, issue 2, 135-141

Abstract: Background: The use of item screeners is viewed as an essential feature of quality survey design because only respondents who are ‘qualified’ to answer questions that apply to a subset of the sample are directed to answer. However, empirical evidence supporting this view is scant. Objective: This study compares data quality resulting from the administration of ambulatory care experience measures that use item screeners versus tailored ‘not applicable’ options in response scales. Methods: Patients from the practices of 367 primary care physicians in 65 medical groups were randomly assigned to receive one of two versions of a well validated ambulatory care experience survey. Respondents (n=2240) represent random samples of active established patients from participating physicians’ panels. The ‘screener’ survey version included item screeners for five test items and the ‘no screener’ version included tailored ‘not applicable’ options in response scales instead of using screeners. The main outcomes measures were data quality resulting from the two item versions, including the mean item scores, the level of missing values, outgoing patient sample sizes needed to achieve adequate medical group-level reliability, and the relative ranking of medical groups. Results: Mean survey item scores generally did not differ by version. There were consistently fewer respondents to the ‘screener’ versions than ‘no screener’ versions. However, because the ‘screener’ versions improved measurement precision, smaller outgoing patient samples were needed to achieve adequate medical group-level reliability for four of the five items than for the ‘no screener’ version. The relative ranking of medical groups did not differ by item version. Conclusion: Screeners appear to reduce noise by ensuring that respondents who are not ‘qualified’ to answer a question are screened out instead of providing unreliable responses. The increased precision resulting from ‘screener’ versions appears to more than offset the higher item non-response rates compared with ‘no screener’ versions. Copyright Adis Data Information BV 2009

Date: 2009
References: View complete reference list from CitEc
Citations:

Downloads: (external link)
http://hdl.handle.net/10.2165/01312067-200902020-00009 (text/html)
Access to full text is restricted to subscribers.

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:spr:patien:v:2:y:2009:i:2:p:135-141

Ordering information: This journal article can be ordered from
http://www.springer.com/economics/journal/40271

DOI: 10.2165/01312067-200902020-00009

Access Statistics for this article

The Patient: Patient-Centered Outcomes Research is currently edited by Christopher I. Carswell

More articles in The Patient: Patient-Centered Outcomes Research from Springer, International Academy of Health Preference Research
Bibliographic data for series maintained by Sonal Shukla () and Springer Nature Abstracting and Indexing ().

 
Page updated 2025-03-20
Handle: RePEc:spr:patien:v:2:y:2009:i:2:p:135-141