1
|
Kutscher T, Eid M. Psychometric benefits of self-chosen rating scales over given rating scales. Behav Res Methods 2024:10.3758/s13428-024-02429-w. [PMID: 38710987 DOI: 10.3758/s13428-024-02429-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 04/08/2024] [Indexed: 05/08/2024]
Abstract
Rating scales are susceptible to response styles that undermine the scale quality. Optimizing a rating scale can tailor it to individuals' cognitive abilities, thereby preventing the occurrence of response styles related to a suboptimal response format. However, the discrimination ability of individuals in a sample may vary, suggesting that different rating scales may be appropriate for different individuals. This study aims to examine (1) whether response styles can be avoided when individuals are allowed to choose a rating scale and (2) whether the psychometric properties of self-chosen rating scales improve compared to given rating scales. To address these objectives, data from the flourishing scale were used as an illustrative example. MTurk workers from Amazon's Mechanical Turk platform (N = 7042) completed an eight-item flourishing scale twice: (1) using a randomly assigned four-, six-, or 11-point rating scale, and (2) using a self-chosen rating scale. Applying the restrictive mixed generalized partial credit model (rmGPCM) allowed examination of category use across the conditions. Correlations with external variables were calculated to assess the effects of the rating scales on criterion validity. The results revealed consistent use of self-chosen rating scales, with approximately equal proportions of the three response styles. Ordinary response behavior was observed in 55-58% of individuals, which was an increase of 12-15% compared to assigned rating scales. The self-chosen rating scales also exhibited superior psychometric properties. The implications of these findings are discussed.
Collapse
Affiliation(s)
- Tanja Kutscher
- Leibniz Institute for Educational Trajectories, Department Research Data Center, Methods Development | Scaling and Test Design, Wilhelmsplatz 3, 96047, Bamberg, Germany.
| | - Michael Eid
- Department of Psychology, Division of Methods and Evaluation, Freie Universitaet Berlin, Habelschwerdter Allee 45, 14195, Berlin, Germany
| |
Collapse
|
2
|
Merhof V, Meiser T. Dynamic Response Strategies: Accounting for Response Process Heterogeneity in IRTree Decision Nodes. PSYCHOMETRIKA 2023; 88:1354-1380. [PMID: 36746887 PMCID: PMC10656330 DOI: 10.1007/s11336-023-09901-0] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/15/2022] [Indexed: 06/18/2023]
Abstract
It is essential to control self-reported trait measurements for response style effects to ensure a valid interpretation of estimates. Traditional psychometric models facilitating such control consider item responses as the result of two kinds of response processes-based on the substantive trait, or based on response styles-and they assume that both of these processes have a constant influence across the items of a questionnaire. However, this homogeneity over items is not always given, for instance, if the respondents' motivation declines throughout the questionnaire so that heuristic responding driven by response styles may gradually take over from cognitively effortful trait-based responding. The present study proposes two dynamic IRTree models, which account for systematic continuous changes and additional random fluctuations of response strategies, by defining item position-dependent trait and response style effects. Simulation analyses demonstrate that the proposed models accurately capture dynamic trajectories of response processes, as well as reliably detect the absence of dynamics, that is, identify constant response strategies. The continuous version of the dynamic model formalizes the underlying response strategies in a parsimonious way and is highly suitable as a cognitive model for investigating response strategy changes over items. The extended model with random fluctuations of strategies can adapt more closely to the item-specific effects of different response processes and thus is a well-fitting model with high flexibility. By using an empirical data set, the benefits of the proposed dynamic approaches over traditional IRTree models are illustrated under realistic conditions.
Collapse
Affiliation(s)
- Viola Merhof
- Department of Psychology, University of Mannheim, L 13 15, 68161, Mannheim, Germany.
| | - Thorsten Meiser
- Department of Psychology, University of Mannheim, L 13 15, 68161, Mannheim, Germany
| |
Collapse
|
3
|
Gonzalez K, Portocarrero FF, Ekema ML. Disposition activation during organizational change: A meta‐analysis. PERSONNEL PSYCHOLOGY 2022. [DOI: 10.1111/peps.12513] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
Affiliation(s)
- Katerina Gonzalez
- Department of Management and Entrepreneurship Sawyer Business School Suffolk University MA USA
| | | | - Michael Luma Ekema
- Narendra P. Loomba Department of Management Zicklin School of Business Baruch College City University of New York NY USA
| |
Collapse
|
4
|
A Novel Partial Credit Extension Using Varying Thresholds to Account for Response Tendencies. JOURNAL OF EDUCATIONAL MEASUREMENT 2020. [DOI: 10.1111/jedm.12268] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
5
|
Kutscher T, Eid M, Crayen C. Sample Size Requirements for Applying Mixed Polytomous Item Response Models: Results of a Monte Carlo Simulation Study. Front Psychol 2019; 10:2494. [PMID: 31798490 PMCID: PMC6863808 DOI: 10.3389/fpsyg.2019.02494] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/19/2019] [Accepted: 10/22/2019] [Indexed: 12/03/2022] Open
Abstract
Mixture models of item response theory (IRT) can be used to detect inappropriate category use. Data collected by panel surveys where attitudes and traits are typically assessed by short scales with many response categories are prone to response styles indicating inappropriate category use. However, the application of mixed IRT models to this data type can be challenging because of many threshold parameters within items. Up to now, there is very limited knowledge about the sample size required for an appropriate performance of estimation methods as well as goodness-of-fit criteria of mixed IRT models in this case. The present Monte Carlo simulation study examined these issues for two mixed IRT models [the restricted mixed generalized partial credit model (rmGPCM) and the mixed partial credit model (mPCM)]. The population parameters of the simulation study were taken from a real application to survey data which is challenging (a 5-item scale with an 11-point rating scale, and three latent classes). Additional data conditions (e.g., long tests, a reduced number of response categories, and a simple latent mixture) were included in this simulation study to improve the generalizability of the results. Under this challenging data condition, for each model, data were generated based on varying sample sizes (from 500 to 5,000 observations with a 500-step). For the additional conditions, only three sample sizes (consisting of 1,000, 2,500, and 4,500 observations) were examined. The effect of sample size on estimation problems and accuracy of parameter and standard error estimates were evaluated. Results show that the two mixed IRT models require at least 2,500 observations to provide accurate parameter and standard error estimates under the challenging data condition. The rmGPCM produces more estimation problems than the more parsimonious mPCM, mostly because of the sparse tables arising due to many response categories. These models exhibit similar trends of estimation accuracy across sample sizes. Under the additional conditions, no estimation problems are observed. Both models perform well with a smaller sample size when long tests were used or a true latent mixture includes two classes. For model selection, the AIC3 and the SABIC are the most reliable information criteria.
Collapse
Affiliation(s)
- Tanja Kutscher
- Department of Education and Psychology, Freie Universitaet Berlin, Berlin, Germany.,Department of Data Center and Method Development, Leibniz Institute for Educational Trajectories, Bamberg, Germany
| | - Michael Eid
- Department of Education and Psychology, Freie Universitaet Berlin, Berlin, Germany
| | - Claudia Crayen
- Department of Education and Psychology, Freie Universitaet Berlin, Berlin, Germany
| |
Collapse
|
6
|
Takagishi M, van de Velden M, Yadohisa H. Clustering preference data in the presence of response-style bias. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2019; 72:401-425. [PMID: 31049942 DOI: 10.1111/bmsp.12170] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/04/2018] [Revised: 02/22/2019] [Indexed: 06/09/2023]
Abstract
Preference data, such as Likert scale data, are often obtained in questionnaire-based surveys. Clustering respondents based on survey items is useful for discovering latent structures. However, cluster analysis of preference data may be affected by response styles, that is, a respondent's systematic response tendencies irrespective of the item content. For example, some respondents may tend to select ratings at the ends of the scale, which is called an 'extreme response style'. A cluster of respondents with an extreme response style can be mistakenly identified as a content-based cluster. To address this problem, we propose a novel method of clustering respondents based on their indicated preferences for a set of items while correcting for response-style bias. We first introduce a new framework to detect, and correct for, response styles by generalizing the definition of response styles used in constrained dual scaling. We then simultaneously correct for response styles and perform a cluster analysis based on the corrected preference data. A simulation study shows that the proposed method yields better clustering accuracy than the existing methods do. We apply the method to empirical data from four different countries concerning social values.
Collapse
Affiliation(s)
| | | | - Hiroshi Yadohisa
- Facluty of Culture and Information Science, Doshisha University, Japan
| |
Collapse
|
7
|
Khorramdel L, von Davier M, Pokropek A. Combining mixture distribution and multidimensional IRTree models for the measurement of extreme response styles. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2019; 72:538-559. [PMID: 31385610 DOI: 10.1111/bmsp.12179] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/29/2018] [Revised: 04/21/2019] [Indexed: 05/10/2023]
Abstract
Personality constructs, attitudes and other non-cognitive variables are often measured using rating or Likert-type scales, which does not come without problems. Especially in low-stakes assessments, respondents may produce biased responses due to response styles (RS) that reduce the validity and comparability of the measurement. Detecting and correcting RS is not always straightforward because not all respondents show RS and the ones who do may not do so to the same extent or in the same direction. The present study proposes the combination of a multidimensional IRTree model with a mixture distribution item response theory model and illustrates the application of the approach using data from the Programme for the International Assessment of Adult Competencies (PIAAC). This joint approach allows for the differentiation between different latent classes of respondents who show different RS behaviours and respondents who show RS versus respondents who give (largely) unbiased responses. We illustrate the application of the approach by examining extreme RS and show how the resulting latent classes can be further examined using external variables and process data from computer-based assessments to develop a better understanding of response behaviour and RS.
Collapse
|
8
|
Meiser T, Plieninger H, Henninger M. IRTree models with ordinal and multidimensional decision nodes for response styles and trait-based rating responses. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2019; 72:501-516. [PMID: 30756379 DOI: 10.1111/bmsp.12158] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/12/2018] [Revised: 11/14/2018] [Indexed: 05/10/2023]
Abstract
IRTree models decompose observed rating responses into sequences of theory-based decision nodes, and they provide a flexible framework for analysing trait-related judgements and response styles. However, most previous applications of IRTree models have been limited to binary decision nodes that reflect qualitatively distinct and unidimensional judgement processes. The present research extends the family of IRTree models for the analysis of response styles to ordinal judgement processes for polytomous decisions and to multidimensional parametrizations of decision nodes. The integration of ordinal judgement processes overcomes the limitation to binary nodes, and it allows researchers to test whether decisions reflect qualitatively distinct response processes or gradual steps on a joint latent continuum. The extension to multidimensional node models enables researchers to specify multiple judgement processes that simultaneously affect the decision between competing response options. Empirical applications highlight the roles of extreme and midpoint response style in rating judgements and show that judgement processes are moderated by different response formats. Model applications with multidimensional decision nodes reveal that decisions among rating categories are jointly informed by trait-related processes and response styles.
Collapse
|
9
|
Böckenholt U, Meiser T. Response style analysis with threshold and multi-process IRT models: A review and tutorial. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2017; 70:159-181. [PMID: 28130934 DOI: 10.1111/bmsp.12086] [Citation(s) in RCA: 44] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/09/2016] [Revised: 10/27/2016] [Indexed: 05/13/2023]
Abstract
Two different item response theory model frameworks have been proposed for the assessment and control of response styles in rating data. According to one framework, response styles can be assessed by analysing threshold parameters in Rasch models for ordinal data and in mixture-distribution extensions of such models. A different framework is provided by multi-process item response tree models, which can be used to disentangle response processes that are related to the substantive traits and response tendencies elicited by the response scale. In this tutorial, the two approaches are reviewed, illustrated with an empirical data set of the two-dimensional 'Personal Need for Structure' construct, and compared in terms of multiple criteria. Mplus is used as a software framework for (mixed) polytomous Rasch models and item response tree models as well as for demonstrating how parsimonious model variants can be specified to test assumptions on the structure of response styles and attitude strength. Although both frameworks are shown to account for response styles, they differ on the quantitative criteria of model selection, practical aspects of model estimation, and conceptual issues of representing response styles as continuous and multidimensional sources of individual differences in psychological assessment.
Collapse
|
10
|
Kutscher T, Crayen C, Eid M. Using a Mixed IRT Model to Assess the Scale Usage in the Measurement of Job Satisfaction. Front Psychol 2017; 7:1998. [PMID: 28101067 PMCID: PMC5209345 DOI: 10.3389/fpsyg.2016.01998] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2016] [Accepted: 12/12/2016] [Indexed: 11/13/2022] Open
Abstract
This study investigated the adequacy of a rating scale with a large number of response categories that is often used in panel surveys for assessing diverse aspects of job satisfaction. An inappropriate scale usage is indicative of overstraining respondents and of diminished psychometric scale quality. The mixed Item Response Theory (IRT) approach for polytomous data allows exploring heterogeneous patterns of inappropriate scale usage in form of avoided categories and response styles. In this study, panel data of employees (n = 7036) on five aspects of job satisfaction measured on an 11-point rating scale within the “Household, Income and Labor Dynamics in Australia” (wave 2001) were analyzed. A three-class solution of the restricted mixed generalized partial credit model fit the data best. The results showed that in no class the 11-point scale was appropriately used but that the number of categories used was reduced in all three classes. Respondents of the large class (40%) appropriately differentiate between up to six categories. The two smaller classes (33 and 27%) avoid even more categories and show some kind of extreme response style. Furthermore, classes differ in socio-demographic and job-related factors. In conclusion, a two- to six-point scale without the middle point might be more adequate for assessing job satisfaction.
Collapse
Affiliation(s)
- Tanja Kutscher
- Department of Education and Psychology, Freie Universitaet Berlin Berlin, Germany
| | - Claudia Crayen
- Department of Education and Psychology, Freie Universitaet Berlin Berlin, Germany
| | - Michael Eid
- Department of Education and Psychology, Freie Universitaet Berlin Berlin, Germany
| |
Collapse
|
11
|
Plieninger H. Mountain or Molehill? A Simulation Study on the Impact of Response Styles. EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT 2017; 77:32-53. [PMID: 29795902 PMCID: PMC5965522 DOI: 10.1177/0013164416636655] [Citation(s) in RCA: 23] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/15/2023]
Abstract
Even though there is an increasing interest in response styles, the field lacks a systematic investigation of the bias that response styles potentially cause. Therefore, a simulation was carried out to study this phenomenon with a focus on applied settings (reliability, validity, scale scores). The influence of acquiescence and extreme response style was investigated, and independent variables were, for example, the number of reverse-keyed items. Data were generated from a multidimensional item response model. The results indicated that response styles may bias findings based on self-report data and that this bias may be substantial if the attribute of interest is correlated with response style. However, in the absence of such correlations, bias was generally very small, especially for extreme response style and if acquiescence was controlled for by reverse-keyed items. An empirical example was used to illustrate and validate the simulations. In summary, it is concluded that the threat of response styles may be smaller than feared.
Collapse
|
12
|
Ziegler M. “F*** You, I Won’t Do What You Told Me!” – Response Biases as Threats to Psychological Assessment. EUROPEAN JOURNAL OF PSYCHOLOGICAL ASSESSMENT 2015. [DOI: 10.1027/1015-5759/a000292] [Citation(s) in RCA: 18] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/31/2022]
|
13
|
Jasper F, Nater UM, Hiller W, Ehlert U, Fischer S, Witthöft M. Rasch scalability of the somatosensory amplification scale: a mixture distribution approach. J Psychosom Res 2013; 74:469-78. [PMID: 23731743 DOI: 10.1016/j.jpsychores.2013.02.006] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/04/2012] [Revised: 01/23/2013] [Accepted: 02/12/2013] [Indexed: 11/28/2022]
Abstract
OBJECTIVE Somatosensory amplification refers to a person's tendency to experience somatic sensations as inappropriately intense and involves hypervigilance concerning bodily sensations. We applied the Somatosensory Amplification Scale (SSAS) in an Internet sample of young adults (N=3031) to test whether the SSAS is Rasch scalable. METHODS We applied mixture distribution extensions of the partial credit and rating scale models to identify possible subgroups that use the response set of the SSAS in different ways. RESULTS A partial credit model, with two latent classes, showed a superior fit to all other models. Still, one of the SSAS items had to be removed because it showed severe underfit. Overall fit of the remaining items was acceptable, although the differentiation between at least two of the five item categories was questionable in both classes. Class 1 was characterized by a higher SSAS sum score, female gender, more somatic complaints, more anxiety, more psychosocial stress, and slightly higher depressiveness. Further exploratory analyses showed that the three mid categories of the SSAS can be collapsed without a large loss of information. CONCLUSIONS Our results show that a shortened version of the SSAS is Rasch scalable but also reveal that there is a lot of room for further improvements of the scale. Based on our results, Item 1 should be removed from the scale and a reduction of the number of response categories is probably warranted.
Collapse
Affiliation(s)
- Fabian Jasper
- Department of Clinical Psychology, Johannes Gutenberg University, Germany.
| | | | | | | | | | | |
Collapse
|
14
|
Parmak M, Mylle JJC, Euwema MC. Personality and the perception of situation structure in a military environment: seeking sensation versus structure as a soldier. JOURNAL OF APPLIED SOCIAL PSYCHOLOGY 2013. [DOI: 10.1111/jasp.12067] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
Affiliation(s)
- Merle Parmak
- Applied Research Centre; Estonian National Defence College
| | | | - Martin C. Euwema
- Centre for Organizational and Personnel Psychology; Catholic University ; Leuven
| |
Collapse
|
15
|
Parmak M, Euwema MC, Mylle JJC. Changes in Sensation Seeking and Need for Structure Before and After a Combat Deployment. MILITARY PSYCHOLOGY 2012. [DOI: 10.1080/08995605.2012.742843] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2022]
Affiliation(s)
- Merle Parmak
- a Applied Research Center, Estonian National Defence College , Tartu , Estonia
| | - Martin C. Euwema
- b Work, Organizational and Personnel Psychology, Catholic University Leuven , Leuven , Belgium
| | - Jacques J. C. Mylle
- c Behavioural Sciences Department, Royal Military Academy , Brussels , Belgium
| |
Collapse
|
16
|
Preinerstorfer D, Formann AK. Parameter recovery and model selection in mixed Rasch models. THE BRITISH JOURNAL OF MATHEMATICAL AND STATISTICAL PSYCHOLOGY 2012; 65:251-262. [PMID: 21675964 DOI: 10.1111/j.2044-8317.2011.02020.x] [Citation(s) in RCA: 20] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
This study examines the precision of conditional maximum likelihood estimates and the quality of model selection methods based on information criteria (AIC and BIC) in mixed Rasch models. The design of the Monte Carlo simulation study included four test lengths (10, 15, 25, 40), three sample sizes (500, 1000, 2500), two simulated mixture conditions (one and two groups), and population homogeneity (equally sized subgroups) or heterogeneity (one subgroup three times larger than the other). The results show that both increasing sample size and increasing number of items lead to higher accuracy; medium-range parameters were estimated more precisely than extreme ones; and the accuracy was higher in homogeneous populations. The minimum-BIC method leads to almost perfect results and is more reliable than AIC-based model selection. The results are compared to findings by Li, Cohen, Kim, and Cho (2009) and practical guidelines are provided.
Collapse
Affiliation(s)
- David Preinerstorfer
- Department of Statistics and Operations Research, University of Vienna, Austria.
| | | |
Collapse
|
17
|
Meiser T, Hewstone M. Contingency learning and stereotype formation: Illusory and spurious correlations revisited. EUROPEAN REVIEW OF SOCIAL PSYCHOLOGY 2010. [DOI: 10.1080/10463283.2010.543308] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
18
|
Schweizer K. Judging a Journal by the Impact Factor: Is It Appropriate and Fair for Assessment Journals? EUROPEAN JOURNAL OF PSYCHOLOGICAL ASSESSMENT 2010. [DOI: 10.1027/1015-5759/a000031] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
|