- Research note
- Open Access
Item response theory analysis of the Utrecht Work Engagement Scale for Students (UWES-S) using a sample of Japanese university and college students majoring medical science, nursing, and natural science
BMC Research Notes volume 10, Article number: 528 (2017)
The Utrecht Work Engagement Scale for Students has been used internationally to assess students’ academic engagement, but it has not been analyzed via item response theory. The purpose of this study was to conduct an item response theory analysis of the Japanese version of the Utrecht Work Engagement Scale for Students translated by authors. Using a two-parameter model and Samejima’s graded response model, difficulty and discrimination parameters were estimated after confirming the factor structure of the scale.
The 14 items on the scale were analyzed with a sample of 3214 university and college students majoring medical science, nursing, or natural science in Japan. The preliminary parameter estimation was conducted with the two parameter model, and indicated that three items should be removed because there were outlier parameters. Final parameter estimation was conducted using the survived 11 items, and indicated that all difficulty and discrimination parameters were acceptable. The test information curve suggested that the scale better assesses higher engagement than average engagement. The estimated parameters provide a basis for future comparative studies. The results also suggested that a 7-point Likert scale is too broad; thus, the scaling should be modified to fewer graded scaling structure.
Students’ academic engagement has been studied across multiple disciplines [1, 2], including students’ behavioral norms, emotional experiences, and cognitive factors [1,2,3]. Empirical studies indicate that there is a positive relationships between students’ engagement and their performance [4,5,6]. These studies have widely used the Utrecht Work Engagement Scale for Students (UWES-S) [7,8,9,10]. However, the UWES-S has not been translated into Japanese, nor has it been validated. Moreover, few studies have been conducted on the use of the UWES-S. Therefore, the aim of the current study was to construct the Japanese version of the UWES-S via item response theory.
Participants and procedures
The data for the current study were obtained from previous research on the relationships among students’ engagement, burnout, and other related variables. We used a convenient sampling method. Seven Japanese universities and five colleges were invited to participate. Of the 3401 students contacted, 3280 returned the questionnaires. The participating university students were majoring in medical science, nursing, and natural science. One university and two colleges were national institutions, and six universities and three colleges were private. Students were informed orally and in writing on the front page of the questionnaire that participation was voluntary, and that participation refusal would not negatively impact them. All students were assured that their responses were anonymous. Passive consent was used herein, and participants were informed that by submitting their questionnaire, they were consenting to participate in the study. Before or after class, participants were instructed to complete the questionnaire, which included a cover sheet asking for their age, sex, grade, and other relevant questions. Cases with no answers were excluded. The total number of participants included was 3214 students. This research was reviewed and approved by the ethics committee of Chubu University.
The UWES-S  was used in the current study (Additional file 1: Appendix). The 14-item version of the UWES-S was selected over the newer 17-item version of the UWES-S  owing to how it has been used internationally. Moreover, the preliminary manual of the UWES-S  states that the default three factor model of the 17-item version does not fit to the data well (N = 572, Chi square = 59.99, CFI = 0.85, RMSEA = 0.08). It was ultimately converted to an 11-item version (N = 572, Chi square = 92.75, CFI = 0.95, RMSEA = 0.07). The 11-item version is similar to the 14-item version used herein. The original version of the 14-item scale had three factors: vigor (5 items; e.g., “When I’m studying, I feel mentally strong”), dedication (4 items; e.g., “I find my studies to be full of meaning and purpose”), and absorption (5 items; e.g., “Time flies when I’m studying”) . All items are scored on a 7-point Likert scale ranging from 0 = “never” to 6 = “always”.
The Japanese version of the UWES-S used in this study was constructed by using the back-translation technique.
A computer program randomly divided the sample into two groups. Sample 1 included 1607 participants (male = 922, female = 618, others were gender unknown; mean age = 19.85, SD of the age = 2.12); sample 2 included 1607 participants (male = 930; female = 671, others were gender unknown; mean age = 19.79, SD age = 2.11).
First, sample 1 was used for preliminary analysis to examine if item response theory analysis can be applied to the UWES-S. During this stage, the polyserial correlation coefficients were calculated, and an item with a coefficient under 0.20 was removed. Next, confirmatory factor analyses were conducted for the default three-factor structure and for a one-factor structure. The default, three-factor model was the same model that was presented by Schaufeli et al. . The one-factor model was the model where only one latent factor influenced all observed variables. If the one-factor model fit to the data well, item response theory analysis could be conducted without considering the local factors generated by the bi-factor exploratory factor analysis. This process could be followed because this structure is conventionally regarded as one-factor structure. However, in a study by Wefald and Downey, the one-factor model, the two-factor model, and the default three-factor model did not demonstrate satisfactory fit to the data; the authors suggested that the one-factor model was the most parsimonious of the three . Therefore, we implemented a bi-factor exploratory factor analysis as the second best option. In this model, one additional latent factor was installed and the local factors were explored with the data. This additional latent factor is called a general factor. The general factor influences all observed variables, but is not correlated with the local factors that influence the observed variables. If the factor structure with the general factor was confirmed, item response theory analysis could be conducted. Item response theory analysis was conducted using Samejima’s graded response modeling  with a two-parameter model. At the end of this stage, items indicating outlier values for each parameter were removed from subsequent analysis.
At the second analytical stage, the resulting items from the first stage were examined using sample 2. The parameters were estimated again and the test information curve represented the amount of test information to consider the characteristics of the scale. It was theoretically described by I (θ) as follows:
where θ is the latent trait measured by the scale (i.e., ability parameter), a is the discrimination parameter of each item, P j (θ) is item characteristic function, and Q j (θ) is calculated by 1 − P j (θ), and D is 1.7. P j (θ) is calculated as follows:
where b is the difficulty parameter of each item.
All estimation was calculated by the maximum likelihood estimation. To evaluate the fit of each model to the data, we adopted the following indices: (1) the Chi square statistic, (2) the comparative fit index (CFI) , (3) the Tucker –Lewis index (TLI) , (4) the root mean square error of approximation (RMSEA) , and (5) the Akaike information criterion (AIC) . Previous studies indicated that values for CFI and TLI greater than 0.90 indicate acceptable model-data fit [8, 9]. For RMSEA, values less than 0.08 indicate a satisfactory fit, while those greater than 0.10 signify that the model should be rejected . The internal reliability was evaluated by McDonald’s omega coefficient .
Items with difficulty parameters more than an absolute value of 6.8 and discrimination parameters out of the range 0.34–3.4 were regarded as outlier values. Analyses were run via the statistical software R, version 3.3.0. The packages used for the analysis are “psych,” “lavaan,” “ltm,” “polycor,” and their related packages. The logistic model in the “ltm” package defines the measurement factor “D” as “1.0”.
The descriptive statistics showed that the item with the lowest mean was item 12 (“When I get up in the morning, I feel like going to class.”) and the item with the highest mean was item 2 (“I find my studies to be full of meaning and purpose.”). The polyserial correlation coefficients ranged from 0.59 to 0.82.
Confirmatory factor analyses showed that the default three-factor model was not satisfactorily supported, and that its fit indices were not necessarily better than those of the one-factor model (one factor model: Chi square = 1721.67, degree of freedom = 77, CFI = 0.86, TLI = 0.84, RMSEA = 0.11, AIC = 70,997.50; default three-factor model: Chi square = 1675.07, degree of freedom = 74, CFI = 0.86, TLI = 0.83, RMSEA = 0.11, AIC = 70,956.90). A bi-factor exploratory factor analysis was conducted, thereby indicating that the model-fit was better than both the one-factor model and the default model (Chi square = 663.99, degree of freedom = 52, RMSEA = 0.086). In this model, three local factors were generated. The model shared some aspects of the default model, but the relations between the factors and items were different. Items 3 and 4 equally loaded on the general factor and the local factor. Other items loaded more on the general factor than the local factors. From the results of the factor analyses, the scale can be analyzed via item response analysis.
The result of the graded response modeling analysis were shown in Table 1; these results indicate that the discrimination parameters of three items were outliers. Therefore, we removed these items in the subsequent analyses.
Next, data from sample 2 were used to estimate parameters of the surviving 11 items. The polyserial correlation coefficients ranged from 0.589 to 0.824. The mean difficulty parameters ranged from − 1.542 to 2.857, and the mean of the discrimination parameters was 1.862 (Table 2).
The test information curve indicated that these items provide information on the latent trait from − 1.5 to 2.2 (Fig. 1). The omega coefficient of the surviving 11 items was 0.91, suggesting that there was an acceptable internal reliability in this scale.
In the current study, items on the Utrecht Work Engagement Scale for Students were analyzed in the item response theory paradigm. When conducting the item response theory analysis, the one-factor structure required confirmation, and our results suggested that a general factor existed. This was the second best solution to meet the goal of the study: to confirm the one-factor structure of the UWES-S. These results corresponded with Schaufeli et al. where intercorrelations between the default three factors were high, ranging from 0.71 to 0.94 .
The parameter analysis indicated that difficulty parameters for each item were generally high, thereby suggesting that participants tended to rate these items from “0” to “2”. The discrimination parameters were also high. Therefore, the items tended to sharply discriminate the degree of engagement. Items with outlier discrimination parameters were items 5, 7, and 9. These items may have confused participants because they ask participants about the perceived strength of one’s positive attitude toward studying; however, the participants were required to select the frequency of these items. The words and phrases of the items—“strong” (item 1), “bursting with energy” (item 7), “strong and vigorous” (item 9), and “enthusiastic” (item 8)—may be difficult to judge in terms of frequency.
Moreover, the difficulty parameters of a rating of 6 versus a rating ranging from 0 to 5 were extremely high. Only a small portion of participants selected a “6”. Thus, it appears that the 7-point structure for this item gave participants too many choices, as suggested by prior research on the UWES-17 .
The test information curve suggested that the current 11-item version of the UWES-S provides the most accurate information from θ = 0.4 to 2.0. Consequently, this test is suitable for those students who engage more in academic activity than average. This is reflected in the fact that the difficulty parameters of the 11 items were relatively high. Therefore, fewer rating options may be preferable for assessing lower degrees of engagement.
We recommend that the current 11-item version of the UWES-S be used for assessing Japanese students’ academic engagement. However, it is important to note that the Japanese version of the UWES-S should not be shortened to 11 items. Specifically, the items removed herein could be retained by truncating the Likert scale, or refining the translation of the items. Therefore, the 11 items and the removed items require further empirical investigation.
In this study, the 2002 version of the UWES-S was used, not the 2003 version. The difference between the two versions is small; however, additional research is needed to replicate the current findings with the newer 17-item version of the UWES-S. Despite these limitations, the item parameters estimated herein provide useful information for future studies on students majoring in the medical science, nursing, and natural science.
Akaike’s information criterion
comparative fit index
root mean square error of approximation
Utrecht Work Engagement Scale for Students
Pekrun R, Linnenbrink-Garcia L. Academic emotions and student engagement. In: Handbook of research on student engagement. Berlin: Springer; 2012. p. 259–82.
Trowler V. Student engagement literature review. The higher education academy. 2010; p. 1–15.
Morgan GL. Improving student engagement. Current issues in education, vol. 14. 2008; p. 1–33.
Zepke N, Leach L. Improving student engagement: ten proposals for action. Act Learn High Educ. 2010;11:167–77.
Carini RM, Kuh GD, Klein SP. Student engagement and student learning: testing the linkages. Res High Educ. 2006;47:1–32.
Pike GR, Kuh GD, Massa-McKinley RC. First-year students’ employment, engagement, and academic achievement: untangling the relationship between work and grades. J Stud Aff Res Pract. 2009;45:560–83.
Schaufeli WB, Bakker AB. UWES Utrecht work engagement scale preliminarty manual. J Occup Health Psychol. 2003;58.
Schaufeli WB, Martinez IM, Pinto AM, Salanova M, Bakker AB. Burnout and engagement in university students: a cross-national study. J Cross Cult Psychol. 2002;33:464–81.
Zhang Y, Gan Y, Cham H. Perfectionism, academic burnout and engagement among Chinese college students: a structural equation modeling analysis. Pers Individ Dif. 2007;43:1529–40.
Bigna JJR, Fonkoue L, Tchatcho MFF, Dongmo CN, Soh DM, Um JLLN, et al. Association of academic performance of premedical students to satisfaction and engagement in a short training program: a cross sectional study presenting gender differences. BMC Res Notes. 2014;7:105.
Schaufeli WB, Salanova M, Gonzlez-Roma VA, Bakker AB. The measurement of engagement and burnout: a two sample confirmatory factor analytic approach. J Happiness Stud. 2002;3:71–92.
Wefald AJ, Downey RG. Construct dimensionality of engagement and its relation with satisfaction. J Psychol. 2009;143:91–112.
Samejima F. Graded response model. Am J Commun Psychol. 1997;35:85–100.
Goffin RD. A comparison of two new indices for the assessment of fit of structural equation models. Multivar Behav Res. 1993;28:205–14.
Tucker LR, Lewis C. A reliability coefficient for maximum likelihood factor analysis. Psychometrika. 1973;38:1–10.
Browne MW, Cudeck R. Single sample cross-validation indices for covariance structures. Multivar Behav Res. 1989;24:445–55.
Wagenmakers E-J, Farrell S. AIC model selection using Akaike weights. Psychon Bull Rev. 2004;11:192–6.
Dunn TJ, Baguley T, Brunsden V. From alpha to omega: a practical solution to the pervasive problem of internal consistency estimation. Br J Psychol. 2014;105:399–412.
De Bruin DP, Hill C, Henn CM, Muller K-P. Dimensionality of the UWES-17: AN item response modelling analysis. SA J Ind Psychol. 2013;39:1–8.
Study design: TT, KS, HI, and NK; data collection: KS; data analysis: NK and TT;, manuscript writing: TT, KS, and HI; manuscript revision: TT, HI, and NK. All authors read and approved the final manuscript.
The authors would like to thank the participating students and faculty for their cooperation in this study.
We also thank Edanz Group (http://www.edanzediting.com/ac) for editing a draft of this manuscript.
The authors declare that they have no competing interests.
Availability of data and materials
Per the decision of the ethics committee at the Chubu University, the datasets generated and analyzed in the current study are not publicly available. However, the datasets are available from the corresponding author upon reasonable request.
Consent for publication
Ethics approval and consent to participate
Ethical approval was obtained from the Ethics Committee of the Chubu University (Approval Number 240048). The committee approved that written informed consent forms were not used in the study. The nature, benefits, and risk of participating in the current study were explained to the participants in detail. The participants were told orally and in writing that consent was obtained by their submission of the questionnaires.
The research presented in the current study was not funded by a grant from any funding agency in the public, commercial, or not-for-profit sectors.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Tsubakita, T., Shimazaki, K., Ito, H. et al. Item response theory analysis of the Utrecht Work Engagement Scale for Students (UWES-S) using a sample of Japanese university and college students majoring medical science, nursing, and natural science. BMC Res Notes 10, 528 (2017). https://doi.org/10.1186/s13104-017-2839-7
- Academic engagement
- The Utrecht Work Engagement Scale for Students
- Item response theory
- University students