- Research article
- Open Access
Linguistic validation of the Alberta Context Tool and two measures of research use, for German residential long term care
BMC Research Notes volume 7, Article number: 67 (2014)
To study the association between organizational context and research utilization in German residential long term care (LTC), we translated three Canadian assessment instruments: the Alberta Context Tool (ACT), Estabrooks’ Kinds of Research Utilization (RU) items and the Conceptual Research Utilization Scale. Target groups for the tools were health care aides (HCAs), registered nurses (RNs), allied health professionals (AHPs), clinical specialists and care managers. Through a cognitive debriefing process, we assessed response processes validity–an initial stage of validity, necessary before more advanced validity assessment.
We included 39 participants (16 HCAs, 5 RNs, 7 AHPs, 5 specialists and 6 managers) from five residential LTC facilities. We created lists of questionnaire items containing problematic items plus items randomly selected from the pool of remaining items. After participants completed the questionnaires, we conducted individual semi-structured cognitive interviews using verbal probing. We asked participants to reflect on their answers for list items in detail. Participants’ answers were compared to concept maps defining the instrument concepts in detail. If at least two participants gave answers not matching concept map definitions, items were revised and re-tested with new target group participants.
Cognitive debriefings started with HCAs. Based on the first round, we modified 4 of 58 ACT items, 1 ACT item stem and all 8 items of the RU tools. All items were understood by participants after another two rounds. We included revised HCA ACT items in the questionnaires for the other provider groups. In the RU tools for the other provider groups, we used different wording than the HCA version, as was done in the original English instruments. Only one cognitive debriefing round was needed with each of the other provider groups.
Cognitive debriefing is essential to detect and respond to problematic instrument items, particularly when translating instruments for heterogeneous, less well educated provider groups such as HCAs. Cognitive debriefing is an important step in research tool development and a vital component of establishing response process validity evidence. Publishing cognitive debriefing results helps researchers to determine potentially critical elements of the translated tools and assists with interpreting scores.
Substantial evidence suggests that residential long term care (LTC) providers’ use of best practices is sub-optimal in Germany [1–7]. Research implementation is complex, challenging and hard to manage [8–10]. Organizational context (i.e., “the environment or setting in which the proposed change is to be implemented”  (p. 150), or–more generally–“the environment or setting in which people receive health care services”  (p. 96)) has been argued to be of vital importance in these processes [9, 13–22]. However, research implementation and its influencing factors are not well understood in the residential LTC setting [23–26].
We lack instruments in German that a) capture reliable and valid scores on organizational context and research utilization in residential LTC institutions, and b) can be used with various provider groups in this setting. We thus translated three Canadian tools into German (see  for details): the Alberta Context Tool (ACT) [28–30], the Estabrooks’ Kinds of Research Utilization (RU) items [31, 32] (residential LTC version ) and the Conceptual Research Utilization (CRU) Scale . These tools have been widely used to investigate health care providers’ utilization of research in their daily work and its association with organizational context [18, 33–35]. The psychometric properties of the three instruments have been evaluated, providing substantial evidence for appropriate acceptability, reliability and validity. See  for the ACT pediatric acute care version, based on RN responses;  for the ACT residential LTC version, based on health care aide (HCA) responses;  for an overview of studies assessing the psychometric properties of the Estabrooks’ Kinds of RU items; and  for the CRU Scale, based on HCA responses.
The ACT contains 10 concepts of organizational context: (1) leadership, (2) culture, (3) evaluation (feedback processes), (4) social capital, (5) informal interactions, (6) formal interactions, (7) structural and electronic resources, (8) organizational slack (staff), (9) organizational slack (space) and (10) organizational slack (time) . Three versions (acute care–adult hospitals and pediatrics, residential LTC, community health) are available, containing forms for six provider groups (HCA, registered nurses (RN), allied health professionals (AHP), practice specialists, managers, physicians). The English original was translated into four languages (Dutch, Swedish, Mandarin Chinese, French) . In our study we translated the HCA, RN, AHP, specialist, and manager forms of the ACT LTC version into German .
The Estabrooks’ Kinds of RU tool [31, 32] comprises four items, each of them reflecting a particular kind of research utilization: (1) instrumental (i.e., using observable research-based practices when caring for residents), (2) conceptual (i.e., thinking about research-based knowledge and then using it to inform clinical decision making), (3) persuasive (i.e., using research findings to win an argument or make a case to someone), and (4) overall (i.e., using any kind of research findings, in any kind of way, in any aspect of work). Items ask care providers how often they used research in the described way. In the HCA questionnaires, the conceptual RU item is not included. The CRU Scale is a one-concept, five-item tool, asking care providers how often best practice knowledge e.g., gave them new knowledge or changed their mind .
The aim and challenge of the translation process is to ensure validity of scores obtained with the translated instruments. We need to maintain the quality of source instruments (ensuring equivalence of source and target versions) and, simultaneously, ensure that translated instruments are appropriate for the target audiences (meeting adaptation needs) [37, 38]. In a previous publication, we reported on the translation process, the challenges and the strategies chosen to deal with challenges . In this paper we, report on the cognitive debriefing–a linguistic validation procedure to “assess the clarity, intelligibility, appropriateness, and cultural relevance of the target language version to the target population”  (p. 47). This is a critical step in translating assessment instruments, as it examines how the target audience responds to translated items and whether they understand them as intended by the tool developers [40, 41].
The evidence provided by cognitive debriefing corresponds to response process validity evidence, as defined by the standards for educational and psychological testing  (hereafter referred to as “the standards”). These standards are regarded as best practice in psychometric testing  and they guided our understanding of validity. In contrast to approaches which suggest that there are different types of validity (e.g., construct or criterion validity), the standards regard validity as a “unitary concept”  (p. 11) for which different sources of evidence are available: (1) tool content, (2) response processes, (3) internal structure and (4) relations to other variables. Validity then is “the degree to which all the accumulated evidence supports the intended interpretation of test scores for the proposed purpose”  (p. 11). Content evidence is obtained if tool items represent the construct(s) the tool intends to measure. We obtained initial content validity evidence through an expert panel step in the translation process . Response process evidence indicates whether the test participants understand tool items as intended by the tool developers. Internal structure evidence refers to the associations between the tool items and components and their conformity with the proposed construct(s). Relations to other variables evidence is supported if the tool items are related (or not related) to concepts to which they are theorized to be related (or not related). In this article we present our cognitive debriefing results, providing information about response process validity evidence of the translated tools.
Cognitive debriefing was one of the last steps in the translation process (Figure 1, step 8, see  for details). Translation guidelines state that cognitive debriefing is an important step in instrument translation, but guidelines differ in the suggested methods and the level of detail of the instructions [38, 44]. We identified a need for additional information to design this step, as the guidelines we used in our translation process [38, 44] did not cover all methodological questions that arose. We based our design on Willis’  comprehensive overview of cognitive interviewing methods. Table 1 shows the critical issues in designing the cognitive debriefing process, potential options and our decisions.
The study was approved by the Ethics Committee of the Medical Faculty, Martin-Luther-University Halle-Wittenberg, Germany (reference number: 2011–39).
Cognitive debriefing is a method to pretest and validate a survey questionnaire, involving qualitative interview methods and thus typically including small samples of 5 to 15 target group participants [39, 41]. Our care provider sample consisted of 39 participants from a convenience sample of five residential LTC facilities: 16 HCAs, 5 RNs, 7 AHPs, 5 specialists and 6 managers. Ten HCAs participated in the initial cognitive debriefing. We modified the translated instruments based on the findings from this round and tested them again in a second cognitive debriefing round with three other HCAs. As this step led to further item modifications, we carried out a third cognitive debriefing round with another three HCAs. Only one cognitive debriefing round was needed for the other provider groups. The inclusion and exclusion criteria for care providers are shown in Table 2.
Cognitive debriefing focusses on studying “the cognitive processes that respondents use to answer survey questions; in particular their comprehension, recall, decisions and judgement, and response processes”  (p. 6; emphases in the original). In order “to detect a wide range of problems in survey questionnaires”  (p. 6), samples should include participants with a broad spectrum of characteristics, known or expected to influence item understanding [39, 41]. Variation, achieved by judgement sampling (i.e., actively selecting the most productive sample with regard to those essential characteristics) , is therefore more critical for cognitive debriefing samples than statistical representativeness [39, 41]. Characteristics known to be important include: age, sex, level of education, and socio-economic background [39, 41]. In addition, Squires et al.  found that HCAs whose native language was not English responded differently to the Items of the CRU Scale than English native speakers – underscoring the significance of ethnicity and native language for item understanding. The managers of the participating facilities were asked to identify eligible staff members and ask them if they would like to participate. In order to reflect the heterogeneity of care providers in the residential LTC setting, we sampled the participants according to characteristics potentially influencing their abilities to understand tool items (Table 3).
We began with the HCAs, asking their managers to identify HCAs who were eligible and willing to participate. First, we identified one HCA with a combination of criteria that we assumed would reflect a low ability to understand tool items: (1) native language not German but moderate German language skills, (2) little job experience (between three and six months), (3) low general education level, and (4) no HCA training. Next, we identified one person with a combination of criteria that we expected would maximize their ability to understand tool items: (1) native language German, (2) extensive job experience (>10 years), (3) high general education level, and (4) HCA training. Finally, we included eight persons with combinations of criteria somewhere between those of the first two HCAs. In rounds two and three of the HCA cognitive debriefing, we used the same procedure to identify new participants: identify one person with a criteria combination unfavourable for item understanding, one with a criteria combination optimal for item understanding, and one in between. RNs, AHPs, specialists and managers were sampled similarly.
Due to staff time constraints in residential LTC it was impossible to test all items with all participants. We thus selected a list of items for each participant before their interview and data collection. Participants completed the whole questionnaire (including all ACT and RU items), but only selected items were discussed in their interviews. Six ACT items (three items from the feedback sub-scale, two slack (time) items and one slack (space) item) and all Estabrooks’ Kinds of RU items had been difficult to translate and had been extensively discussed in the expert focus groups and back translation reviews. We thus included them in all item lists. The remaining HCA ACT and CRU items were randomly allocated to the lists of the 10 HCAs participating in the initial cognitive debriefing round, until a) each item was assigned to at least one list and b) each of those lists contained 20 items. Our approach of randomized item selection is an adaptation of the proceeding described by Schuman  as part of the random probe technique. In second and third HCA cognitive debriefing rounds we included only the items revised in the previous round. In the cognitive debriefing sessions with the other provider groups the items were not selected randomly. In all lists for participants from other provider groups we included the six ACT and Estabrooks’ Kinds of RU items as above, as well as ACT items that were problematic in the HCA cognitive debriefings and all ACT items that differed between the previously translated version and the one to be tested. Finally, all items from the CRU Scale were added. This resulted in lists containing 26 items for RNs, 25 items for allied health professionals, 24 items for specialists and 28 items for managers.
First, the researcher explained the procedure to the participant and asked for informed consent. Participants who were willing to participate completed the questionnaire. Subsequently, the researcher reviewed questionnaire responses for missing items or mistakes (e.g., items ticked twice or items ticked although they should have been skipped according to skip patterns). Participants were asked whether they found some items hard to understand or to answer, and how they rated the clarity of the questionnaire design. Problematic items were added to the predefined item lists if not already included.
After participants completed the questionnaire, we conducted individual cognitive debriefings. In the interviews, participant understanding of the items was assessed using verbal probing–a qualitative, semi-structured interview method. The interviewer stimulated participant reflections on the meanings of questionnaire items or the backgrounds of their answers to questionnaire items by asking specific types of questions–so-called cognitive probes . Willis  discusses six kinds of probes, which we adapted. Each type of probe is illustrated in Table 4 with an example question based on one of the ACT items. That item asked participants to what extent they agree or disagree that they are a member of a supportive team. They could select one of five answers on a Likert scale: strongly disagree, disagree, neither agree nor disagree, agree or strongly agree.
We followed Willis’  (p. 95) recommendation to maintain “a flexible approach to probe construction”. Before the interviews, we developed example questions for all six probes relating to each of the included tool items. We did not decide definitely which kind of probe to use with which item. As Willis  (p. 95) states, “the most interesting and productive forms of probing often develop through the course of the interview, as the product of the particular relationship between the interviewer, subject, and survey questionnaire”. Therefore, the interviewer used an interview guideline with example questions, but was free to choose the kind of probe and to ask questions other than the ones pre-formulated, depending on the tool item and communication situation. Participants responded to the probes in their own words with open ended statements. The interviews were recorded with an electronic voice recorder.
Recorded interviews were transcribed and interview texts were reduced by MH using a qualitative content analysis technique called “summarizing content analysis” [47, 48]. Text segments constituting a unit of meaning (typically sentences) and referring to participants’ responses to the cognitive probes were identified. Components not related to the core content (such as repetitions or embellishing elements) were removed and the remaining sentence was reduced to a concise statement by paraphrasing it. These statements were compared to concept maps, designed by the instrument developers, which define each concept in detail. Responses to each of the probed items were evaluated by MH and MB as to whether they matched the relating concept map definition.
Revision of items and further cognitive debriefing rounds
Items were revised if answers from at least two participants did not match the relating concept map definition. The revised wording was then tested in another cognitive debriefing round. As Figure 2 shows, three rounds were required for the HCA forms before the participants understood all items as intended. In the other translations (RN, AHP, specialist, manager), participants understood all items in the initial cognitive debriefing.
Thirty-nine providers from five non-profit nursing homes in the “Metropolregion Rhein-Neckar” (in the south-west of Germany) participated. The median number of beds per facility was 163 (range = 82 to 217). The mean age of the participants was 40.26 years (SD = 10.58). Of the nine non-native German speakers, six specified Polish as their native language, two Russian, and one Spanish. Their mean number of years speaking German was 13.56 (range = 2.00 to 39.00, SD = 12.76). Further details of participants’ socio-demographic characteristics are given in Table 5.
HCA cognitive debriefing
An item was classified as problematic if the answers from at least two participants did not match the intended item meaning. Table 6 provides an example of an answer that matched well with the intended item meaning, Table 7 contains an example of a problematic item.
In the first example, the participant understood that the question related to a private room, used to talk about resident care or best practices. The second example shows that the participant only focussed on the best practices part of the question. She understood neither the concept of balancing nor the concept of productivity as the counterpart to be balanced with best practice.
Results of the first round of HCA cognitive debriefing demonstrated that 11 of the 58 ACT items and all 8 RU items were not understood by at least two of the participants. Examples of these items and their revisions are presented in Additional file 1.
Based on these results, 4 of the 11 problematic ACT items and all RU items were modified. We decided not to modify the wording of the formal interactions item (continuing education (…) outside this nursing home). Two participants did not read carefully enough and thought about education in their facility. Wording modification would not have resolved that issue. We also did not modify the six feedback items, which ask the participants how often they get formal information about care quality and how this information is handled in their facility. The problem was not caused by the items themselves, but rather by the preceding stem, which we therefore changed. Participants did not think about overarching information for their entire unit or facility (such as falls rate), but rather about information related to individual residents. Therefore the answers to all feedback items were incorrect. Changes to the stem aimed for more clarity about which kind of information was required for the items. In the English version, the RU items ask the HCAs how often they used best practices in specific ways in their daily routine. In German there is no corresponding word for best practice. Germans often use the English term, but the HCAs were not familiar with this. As a result, we chose to describe this principle and to provide examples for the sake of clarity (see  for further details). The changed wordings are presented in Additional file 1.
After analysis of the round two data, 2 of the 4 modified ACT items were still problematic: culture item 3 and structural and electronic resources item 5 (Additional file 1). In addition, the six ACT feedback items and all 8 RU items were still problematic. In order to have the participants focus on the intended kind of information, we changed the feedback stem again by (1) introducing the term “statistics”, (2) emphasizing the important passages even more, and (3) adding a sentence that explained what was not meant by this item (i.e., individual resident information). The other two problematic ACT items and the RU items were also further modified. We described the best practice term by the wording “knowledge of how to provide the best possible care quality” (German: “Fachwissen über optimale Pflege und Betreuung”). In round three, all items were understood as intended.
Cognitive debriefing with the other provider groups
In the German questionnaires for the other provider groups, we adopted the new item wordings that we had developed during the HCA cognitive debriefing sessions for all items, which had the same wording in the original Canadian tools for both the HCA version and the versions for the other providers: the two Culture items, Feedback stem and all six items, and the Structural and Electronic Resources item (Additional file 1). The English wording of the other items (ACT Time item and the RU tools items) differs in the Canadian tools between the HCA questionnaires and the questionnaires for the other provider groups. The term “best practice” is used for the HCAs, while “clinical knowledge” (ACT Time item) or “research use“ (RU tools) was chosen for the other provider groups. We retained this difference in our translation and used the German wording “klinische Erkenntnisse” (clinical knowledge, ACT time item) and “Anwendung wissenschaftlicher Erkenntnisse” (use of scientific knowledge, RU tools) for the regulated providers. In the subsequent cognitive debriefings with RNs, AHPs, specialists and managers, we included all items of the two RU tools, all items that were problematic in the translation process, all items that were problematic in the HCA cognitive debriefing and all items whose wordings differed between the previously translated version and the one to be tested (see Methods section for details). All participants understood each of the included items as intended. Thus, only one cognitive debriefing round was required for each of these provider groups. An example answer of each provider group to the instrumental research use item is provided in Table 8.
The role of organizational context must be understood to improve research implementation in residential LTC, but that understanding is still lacking [23–25]. Researchers need robust assessment tools to study organizational context [49, 50]. We could find no German assessment tool that a) specifically and validly assessed modifiable organizational context factors that are asserted to influence research implementation in residential LTC, and b) could be used with various residential LTC provider groups. Thus, we translated three Canadian tools  from English into German. Cognitive debriefing is an essential step to assess response process validity of the translated tools [40–42].
Finding an appropriate German wording for items asking HCAs about best practices (all 8 RU items, ACT culture item 3, and ACT time item 3) was the major challenge for us, particularly in the RU tools translations. While “research” is the wording of choice in the regulated provider RU tool versions (RNs, AHPs, specialists, managers), “best practice” is used in the HCA forms, as this terminology is commonly used and better understood by English-speaking HCAs [30, 33, 51, 52]. German has no equivalent for this term and German HCAs didn’t understand the English term when it was directly adopted in German. Furthermore, German HCAs had difficulty understanding the terms “research”, “research knowledge”, “scientific knowledge” or “evidence”. They often found it hard to imagine what kind of research knowledge might be available and relevant for their practice. Like Canadian and Swedish residential LTC providers , they tended to discuss barriers to RU rather than RU itself. However, they agreed that some kind of knowledge is important for their practice–either obtained by experience or by asking colleagues. This is consistent with findings that even RNs prefer informal, interactive or experience-based knowledge sources to formal ones such as journals or text books [53–55].
Nevertheless, we were highly motivated to find wording that HCAs understood. In Germany, about 40% of the staff providing direct care in residential LTC are HCAs (i.e., staff with one year of HCA training, brief training of a few weeks or months, or no elder-care-related training at all) . They give feeding assistance, mobilize residents, turn them to prevent pressure ulcers, provide oral health care, interact with persons with dementia, etc. All of these tasks can pose safety risks to residents if carried out improperly. We thus believe that it is crucial to know how HCAs rate their use of best practice. Very few RU studies have included HCAs as yet [33, 57–59]. The rigorous translation process  and the cognitive debriefing in particular were important to create robust tools. The cognitive debriefing helped us to detect problematic items that would have undermined the validity of the tools’ scores if unmodified. In our context, it was essential to avoid specific instrument terms like “research”, “scientific”, or “best practice” in favour of clear simple terms and explanations of these concepts. The translation of “best practice” to “knowledge of how to provide the best possible care quality” (German: “… Fachwissen über optimale Pflege und Betreuung …”) worked best and facilitated HCAs’ comprehension and understanding of those items.
We expected that the problems discussed above would occur mainly with the HCAs, and that the regulated providers would understand the more technical wording. Therefore, we did not adopt the HCA wording “best practice” for items referring to “research use” in the regulated provider versions. Revised versions of the remaining problematic HCA items, which we assumed would also be problematic in the other provider groups (i.e., the individual versus unit/facility level problem of the ACT Feedback section), were adopted in the regulated provider forms. Working on the items and modifying them until the HCAs understood them was certainly time consuming but necessary and ultimately fruitful. All items understood by the HCAs were subsequently understood by the other provider groups, indicating no further need for modifications.
The cognitive debriefing step helped us prepare the instruments for our larger field testing study. We are evaluating the psychometric properties of the translated instruments and the relationship between organizational context and research implementation in German residential LTC facilities. Qualitative findings will assist us as well as future researchers in interpreting the instrument scores and identifying and understanding potential problems. However, additional validity evidence sources (i.e., internal structure and relations to other variables) need to be examined. Currently, we are evaluating the translated tools with this focus in a larger sample.
Some limitations of cognitive interviewing need to be considered. Generally, cognitive interviewing tends to underestimate problems because:
persons who volunteer to participate in cognitive interviews are more ready to spend time thinking about the items, are often better educated and are more confident in being able to understand the questionnaire.
it is a testing situation in which participants work to perform well and are “patient and forgiving”  (p. 226).
In addition, we did not test all items with each participant, a condition that we accepted for both better feasibility of the cognitive debriefing and participants’ compliance. Thus, we tested some problematic items with all participants and distributed all other items randomly among participants, ensuring that no item remained untested and avoiding selection bias. Finally, due to the qualitative design the sample size was relatively small (although appropriate for cognitive debriefing purposes). Facilities and care providers participating in our study are therefore not statistically representative of the German facility and care provider population. Although we could find evidence for response process validity of the final questionnaires, these results cannot be generalized. Validity needs further investigation in larger samples, using rigorous statistical methods. Nevertheless, our cognitive debriefings detected a variety of problems and helped to minimize them, although other problems may remain.
Cognitive debriefing is essential in translating instruments as an early step in instrument validation. It provides information about response process validity evidence and helps translators to detect and respond to problems. Translating tools intended to assess HCA use of research is challenging. HCAs are not trained to find and use research on their own and they are not familiar with the related terminology. However, assessing their use of best practice is important because they provide hands-on care that may risk the safety of residents if not provided properly. Cognitive debriefing is important to assess whether HCAs understand the chosen wording of tool items, in order to validly assess their rating of best practice use. Publishing cognitive debriefing results helps researchers anticipate and plan for potential challenges, determine potentially critical elements of the translated tools and interpret the resulting scores.
Alberta context tool
Allied health professional
Conceptual research use
Health care aide
Long term care
Treusch Y, Jerosch D, Majic T, Heinz A, Gutzmann H, Rapp MA: How can we provide better services for demented nursing home residents suffering from apathy?. Psychiatr Prax. 2010, 37: 84-88. 10.1055/s-0029-1223472.
Majic T, Pluta JP, Mell T, Aichberger MC, Treusch Y, Gutzmann H, Heinz A, Rapp MA: The pharmacotherapy of neuropsychiatric symptoms of dementia: a cross-sectional study in 18 homes for the elderly in Berlin. Dtsch Ärztebl Int. 2010, 107: 320-327.
Seidl U, Lueken U, Volker L, Re S, Becker S, Kruse A, Schröder J: Non-cognitive symptoms and psychopharmacological treatment in demented nursing home residents. Fortschr Neurol Psychiatr. 2007, 75: 720-727. 10.1055/s-2007-959211.
Bredthauer D, Becker C, Eichner B, Koczy P, Nikolaus T: Factors relating to the use of physical restraints in psychogeriatric care: a paradigm for elder abuse. Z Gerontol Geriat. 2005, 38: 10-18. 10.1007/s00391-005-0285-y.
Bartholomeyczik S, Reuther S, Luft L, van Nie N, Meijers J, Schols J, Halfens R: Prevalence of malnutrition, interventions and quality indicators in German nursing homes: first results of a nationwide pilot study. Gesundheitswesen. 2010, 72: 868-874. 10.1055/s-0029-1246150.
Hardenacke D, Bartholomeyczik S, Halek M: Implementation and evaluation of “Understanding Diagnostics” on the example of the light house project InDemA. Pflege Gesellschaft. 2011, 16: 101-115.
Kuske B, Luck T, Hanns S, Matschinger H, Angermeyer MC, Behrens J, Riedel-Heller SG: Training in dementia care: a cluster-randomized controlled trial of a training program for nursing home staff in Germany. Int Psychogeriatr. 2009, 21: 295-308. 10.1017/S1041610208008387.
Sterns S, Miller SC, Allen S: The complexity of implementing culture change practices in nursing homes. J Am Med Dir Assoc. 2010, 11: 511-518. 10.1016/j.jamda.2009.11.002.
Greenhalgh T, Glenn R, Bate P, Macfarlane F, Kyriakidou O: Diffusion of innovations in health service organisations: a systematic literature review. 2005, Massachusetts, Oxford, Victoria: Blackwell
Kitson AL: The need for systems change: reflections on knowledge translation and organizational change. J Adv Nurs. 2009, 65: 217-228. 10.1111/j.1365-2648.2008.04864.x.
Kitson A, Harvey G, McCormack B: Enabling the implementation of evidence based practice: a conceptual framework. Qual Health Care. 1998, 7: 149-158. 10.1136/qshc.7.3.149.
McCormack B, Kitson A, Harvey G, Rycroft-Malone J, Titchen A, Seers K: Getting evidence into practice: the meaning of ‘context’. J Adv Nurs. 2002, 38: 94-104. 10.1046/j.1365-2648.2002.02150.x.
Knowledge to Action? Evidence-Based Health Care in Context. Edited by: Dopson S, Fitzgerald L. 2005, New York: Oxford University Press
Meijers JMM, Janssen MAP, Cummings GG, Wallin L, Estabrooks CA, Halfens RYG: Assessing the relationships between contextual factors and research utilization in nursing: systematic literature review. J Adv Nurs. 2006, 55: 622-635. 10.1111/j.1365-2648.2006.03954.x.
Rycroft-Malone J: The PARIHS framework: a framework for guiding the implementation of evidence-based practice. J Nurs Care Qual. 2004, 19: 297-304. 10.1097/00001786-200410000-00002.
Estabrooks CA, Kenny DJ, Adewale AJ, Cummings GG, Mallidou AA: A comparison of research utilization among nurses working in Canadian civilian and United States Army healthcare settings. Res Nurs Health. 2007, 30: 282-296. 10.1002/nur.20218.
Estabrooks CA, Scott SD, Squires JE, Stevens B, O’Brien-Pallas L, Watt-Watson J, Profetto-McGrath J, McGilton K, Golden-Biddle K, Lander J, et al: Patterns of research utilization on patient care units. Implement Sci. 2008, 3: 31-10.1186/1748-5908-3-31.
Estabrooks CA, Squires JE, Cummings GG, Teare GF, Norton PG: Study protocol for the translating research in elder care (TREC): building context - an organizational monitoring program in long-term care project (project one). Implement Sci. 2009, 4: 52-10.1186/1748-5908-4-52.
Cummings GG, Hutchinson AM, Scott SD, Norton PG, Estabrooks CA: The relationship between characteristics of context and research utilization in a pediatric setting. BMC Health Serv Res. 2010, 10: 168-10.1186/1472-6963-10-168.
Estabrooks CA, Squires JE, Hayduk LA, Cummings GG, Norton PG: Advancing the argument for validity of the Alberta context tool with healthcare aides in residential long-term care. BMC Med Res Methodol. 2011, 11: 107-10.1186/1471-2288-11-107.
Estabrooks CA, Squires JE, Hutchinson AM, Scott S, Cummings GG, Kang SH, Midodzi WK, Stevens B: Assessment of variation in the Alberta context tool: the contribution of unit level contextual factors and specialty in Canadian pediatric acute care settings. BMC Health Serv Res. 2011, 11: 251-10.1186/1472-6963-11-251.
Aarons GA, Glisson C, Green PD, Hoagwood K, Kelleher KJ, Landsverk JA: The organizational social context of mental health services and clinician attitudes toward evidence-based practice: a United States national study. Implement Sci. 2012, 7: 56-10.1186/1748-5908-7-56.
Masso M, McCarthy G: Literature review to identify factors that support implementation of evidence-based practice in residential aged care. Int J Evid Based Healthc. 2009, 7: 145-156. 10.1111/j.1744-1609.2009.00132.x.
Boström A-M, Slaughter SE, Chojecki D, Estabrooks CA: What do we know about knowledge translation in the care of older adults? A scoping review. J Am Med Dir Assoc. 2012, 13: 210-219. 10.1016/j.jamda.2010.12.004.
Estabrooks CA, Hutchinson AM, Squires JE, Birdsell J, Cummings GG, Degner L, Morgan D, Norton PG: Translating research in elder care: an introduction to a study protocol series. Implement Sci. 2009, 4: 51-10.1186/1748-5908-4-51.
Rahman AN, Applebaum RA, Schnelle JF, Simmons SF: Translating research into practice in nursing homes: can we close the gap?. Gerontologist. 2012, 52: 597-606. 10.1093/geront/gnr157.
Hoben M, Mahler C, Bär M, Berger S, Squires JE, Estabrooks CA, Behrens J: German translation of the Alberta Context Tool and two measures of research use: methods, challenges and lessons learned. BMC Health Serv Res. 2013, 13: 478-10.1186/1472-6963-13-478.
Estabrooks CA, Squires JE, Adachi A-M, Kong L, Norton PG: Utilization of health research in acute care settings in Alberta (ACUTE). Technical Report No. 08-01-TR, April 2008. 2008, Edmonton: University of Alberta, Faculty of Nursing
Hutchinson AM, Kong L, Adachi A-M, Estabrooks CA, Stevens B: Context and research use in the care of children: a pilot study. Technical report no. 08-03-TR, October 2008. 2008, Edmonton: University of Alberta, Faculty of Nursing
Squires JE, Kong L, Brooker S, Mitchell A, Sales AE, Estabrooks CA: Examining the role of context in alzheimer care centers: a pilot study. Technical report no. 08-04-TR, June 2009. 2009, Edmonton: University of Alberta, Faculty of Nursing
Estabrooks CA: Modeling the individual determinants of research utilization. West J Nurs Res. 1999, 21: 758-772. 10.1177/01939459922044171.
Estabrooks CA: The conceptual structure of research utilization. Res Nurs Health. 1999, 22: 203-216. 10.1002/(SICI)1098-240X(199906)22:3<203::AID-NUR3>3.0.CO;2-9.
Squires JE, Estabrooks CA, Newburn-Cook CV, Gierl M: Validation of the conceptual research utilization scale: an application of the standards for educational and psychological testing in healthcare. BMC Health Serv Res. 2011, 11: 107-10.1186/1472-6963-11-107.
Squires JE, Hayduk L, Hutchinson AM, Cranley LA, Gierl M, Cummings GG, Norton PG, Estabrooks CA: A protocol for advanced psychometric assessment of surveys. Nurs Res Pract. 2013, 2013: 156782-
Squires JE, Estabrooks CA, O’Rourke HM, Gustavsson P, Newburn-Cook CV, Wallin L: A systematic review of the psychometric properties of self-report research utilization measures used in healthcare. Implement Sci. 2011, 6: 83-10.1186/1748-5908-6-83.
Estabrooks CA, Squires JE, Cummings GG, Birdsell JM, Norton PG: Development and assessment of the Alberta Context Tool. BMC Health Serv Res. 2009, 9: 234-10.1186/1472-6963-9-234.
Maneesriwongul W, Dixon JK: Instrument translation process: a methods review. J Adv Nurs. 2004, 48: 175-186. 10.1111/j.1365-2648.2004.03185.x.
Acquadro C, Conway K, Hareendran A, Aaronson N: Literature review of methods to translate health-related quality of life questionnaires for use in multinational clinical trials. Value Health. 2008, 11: 509-521. 10.1111/j.1524-4733.2007.00292.x.
Mear I, Giroudet C: Linguistic validation procedures. Linguistic Validation Manual for Health Outcome Assessments. Edited by: Acquadro C, Conway K, Giroudet C, Mear I. 2012, Lyon: MAPI Institute, 33-117. 2
Linguistic Validation Manual for Health Outcome Assessments. Edited by: Acquadro C, Conway K, Giroudet C, Mear I. 2012, Lyon: MAPI Institute, 2
Willis GB: Cognitive interviewing: A tool for improving questionnaire design. 2005, Thousand Oaks: Sage
American Educational Research Association (AERA), American Psychological Association (APA), National Council on Measurement in Education (NCME): Standards for educational and psychological testing. 1999, Washington, DC: AERA, 5
Streiner DL, Norman GR: Health Measurement Scales: A Practical Guide to Their Development and Use. 2008, Oxford: Oxford University Press, 4
Wild D, Grove A, Martin M, Eremenco S, McElroy S, Verjee-Lorenz A, Erikson P: Principles of good practice for the translation and cultural adaptation process for Patient-Reported Outcomes (PRO) measures: report of the ISPOR task force for translation and cultural adaptation. Value Health. 2005, 8: 94-104. 10.1111/j.1524-4733.2005.04054.x.
Marshall MN: Sampling for qualitative research. Fam Pract. 1996, 13: 522-525. 10.1093/fampra/13.6.522.
Schuman H: The random probe: a technique for evaluating validity of closed questions. Am Sociol Rev. 1966, 31: 218-222. 10.2307/2090907.
Mayring P: Qualitative Inhaltsanalyse: Grundlagen und Techniken. 2010, Beltz: Weinheim, Basel, 11
Mayring P: Qualitative content analysis. Forum Qual Soc Res. 2000, 1: 2-
Chaudoir SR, Dugan AG, Barr CH: Measuring factors affecting implementation of health innovations: a systematic review of structural, organizational, provider, patient, and innovation level measures. Implement Sci. 2013, 8: 22-10.1186/1748-5908-8-22.
Proctor EK, Brownson RC: Measurement issues in dissemination and implementation research. Dissemination and implementation research in health: Translating science to practice. Edited by: Brownson RC, Colditz GA, Proctor EK. 2012, Oxford: Oxford University Press, 261-280.
Estabrooks CA, Squires JE, Strandberg E, Nilsson-Kajermo K, Scott SD, Profetto-McGrath J, Harley D, Wallin L: Towards better measures of research utilization: a collaborative study in Canada and Sweden. J Adv Nurs. 2011, 67: 1705-1718. 10.1111/j.1365-2648.2011.05610.x.
Squires JE, Adachi AM, Estabrooks CA: Developing a valid and reliable measure of research utilization: measurement technical report. Report no. 08-02-TR, November 2008. 2008, Edmonton: University of Alberta, Faculty of Nursing
Estabrooks CA, Chong H, Brigidear K, Profetto-McGrath J: Profiling Canadian nurses’ preferred knowledge sources for clinical practice. Can J Nurs Res. 2005, 37: 118-140.
Estabrooks CA, Rutakumwa W, O’Leary KA, Profetto-McGrath J, Milner M, Levers MJ, Scott-Findlay S: Sources of practice knowledge among nurses. Qual Health Res. 2005, 15: 460-476. 10.1177/1049732304273702.
Spenceley SM, O’Leary KA, Chizawsky LL, Ross AJ, Estabrooks CA: Sources of information used by nurses to inform practice: an integrative review. Int J Nurs Stud. 2008, 45: 954-970.
Statistisches Bundesamt: Pflegestatistik 2011: Pflege im Rahmen der Pflegeversicherung – Deutschlandergebnisse. 2013, Statistisches Bundesamt: Wiesbaden
Boström AM, Wallin L, Nordström G: Research use in the care of older people: a survey among healthcare staff. Int J Older People Nurs. 2006, 1: 131-140. 10.1111/j.1748-3743.2006.00014.x.
Boström AM, Wallin L, Nordström G: Evidence-based practice and determinants of research use in elderly care in Sweden. J Eval Clin Pract. 2007, 13: 665-673. 10.1111/j.1365-2753.2007.00807.x.
Connor N: The relationship between organizational culture and research utilization practices among nursing home departmental staff. PhD Thesis. 2007, Dalhousie University
This project was funded by the Robert Bosch Stiftung Stuttgart within the Graduate Program Dementia (GPD), located at the Network Aging Research (NAR), University of Heidelberg. The authors would like to thank the NAR for providing infra-structure and support. Particularly we want to express our gratitude to Dr. Konrad T. Beyreuther, the director of the NAR, for his valuable contribution to this project. Finally we want to thank the care providers who participated in the cognitive debriefings and the participating facilities for all their support.
The authors declare that they have no competing interests.
MH led the development of the translation process and cognitive debriefing design, led the research project, carried out the cognitive debriefings, and drafted this manuscript. MB contributed to the design and analysis of the cognitive debriefings, and was a member of the translation team (second forward translator). CM assisted with the development of the translation process and cognitive debriefing design and performed the first back translation. SB was member of the translation team (second back translation). CAE and JES developed the ACT and the two RU measures, approved the translation of the tools, did the back translation reviews and provided advice and recommendations regarding the translation process and cognitive debriefing design and the translations. AK and JB were the co-supervisor and primary supervisor, respectively, of the entire project. All authors contributed to the research design and contributed to, read, and approved the final version of this manuscript.
Electronic supplementary material
About this article
Cite this article
Hoben, M., Bär, M., Mahler, C. et al. Linguistic validation of the Alberta Context Tool and two measures of research use, for German residential long term care. BMC Res Notes 7, 67 (2014). https://doi.org/10.1186/1756-0500-7-67
- Cognitive debriefing
- Alberta context tool
- Estabrooks’ kind of research utilization items
- Conceptual research utilization scale
- Organizational context
- Research utilization
- Residential long term care