Skip to main content
  • Research article
  • Open access
  • Published:

Development of a measure of model fidelity for mental health Crisis Resolution Teams

Abstract

Background

Crisis Resolution Teams (CRTs) provide short-term intensive home treatment to people experiencing mental health crisis. Trial evidence suggests CRTs can be effective at reducing hospital admissions and increasing satisfaction with acute care. When scaled up to national level however, CRT implementation and outcomes have been variable. We aimed to develop and test a fidelity scale to assess adherence to a model of best practice for CRTs, based on best available evidence.

Methods

A concept mapping process was used to develop a CRT fidelity scale. Participants (n = 68) from a range of stakeholder groups prioritised and grouped statements (n = 72) about important components of the CRT model, generated from a literature review, national survey and qualitative interviews. These data were analysed using Ariadne software and the resultant cluster solution informed item selection for a CRT fidelity scale. Operational criteria and scoring anchor points were developed for each item. The CORE CRT fidelity scale was then piloted in 75 CRTs in the UK to assess the range of scores achieved and feasibility for use in a 1-day fidelity review process. Trained reviewers (n = 16) rated CRT service fidelity in a vignette exercise to test the scale’s inter-rater reliability.

Results

There were high levels of agreement within and between stakeholder groups regarding the most important components of the CRT model. A 39-item measure of CRT model fidelity was developed. Piloting indicated that the scale was feasible for use to assess CRT model fidelity and had good face validity. The wide range of item scores and total scores across CRT services in the pilot demonstrate the measure can distinguish lower and higher fidelity services. Moderately good inter-rater reliability was found, with an estimated correlation between individual ratings of 0.65 (95% CI: 0.54 to 0.76).

Conclusions

The CORE CRT Fidelity Scale has been developed through a rigorous and systematic process. Promising initial testing indicates its value in assessing adherence to a model of CRT best practice and to support service improvement monitoring and planning. Further research is required to establish its psychometric properties and international applicability.

Peer Review reports

Background

Fidelity measures are tools to assess the implementation of intervention or programme models [1], and as such, can help address the major challenge for mental health services of translating scientific knowledge into patient benefit [2]. Development of fidelity measures for complex interventions in mental health services has been advocated not only as a means to define an intervention and measure services’ adherence to the model specified, but also to support service improvement [1]. The US Evidence-Based Practice Program demonstrated that a service improvement initiative involving fidelity measurement as a key component led to successful implementation of five different evidence-based practices in a majority of services, in a large-scale, national programme [3]. Fidelity scales become credible measures of service quality when higher fidelity scores have been shown to be associated with better services outcomes, as for instance with a fidelity scale measuring evidence-based supported employment [4]. Fidelity scales have been developed for complex mental health services, such as Assertive Community Treatment [5] but there is no existing fidelity scale for Crisis Resolution Teams.

Crisis Resolution Teams (CRTs) provide short-term, intensive home treatment to people experiencing a mental health crisis, with the aim of averting hospital admission wherever possible, or supporting people to return home as promptly as possible following an acute admission [6]. The CRT model has not been highly specified in the literature, leading to diverse approaches to implementing these services. Key characteristics of CRTs recommended in government and expert guidance are that: CRTs should provide an easy access, rapid response, 24 h service; should be multi-disciplinary and able to provide medical, psychological and social interventions; should help facilitate prompt discharge from acute wards; and should fulfil a “gatekeeping” function of assessing all patients before admission to acute wards and considering home treatment as an alternative to admission wherever possible [7, 8]. CRTs have been implemented on the largest scale in the UK, where they were mandated in England by the NHS Plan in 2000 [9]. They also form part of national mental health policy in Norway [10] and have been implemented regionally in a number of countries including Australia and the Netherlands.

CRTs are one form of home-based crisis intervention. A recent systematic review from the Cochrane Collaboration concluded that home-based crisis intervention can be an effective alternative to hospital admission [11]. This review included only one randomised trial of a UK CRT service [12], which found that a CRT reduced hospital admissions and inpatient bed use, and increased service users’ satisfaction with acute care. Similarly positive results have been found in non-randomised studies [13, 14]. However, the potential benefits of CRT services suggested in research studies have not been fully translated into practice. Two analyses from a UK nationwide study using routine hospital admissions data reached conflicting conclusions about whether there is any association between the introduction of CRTs to a local area and a reduction in bed use [15, 16]. Rates of compulsory inpatient admissions in the UK have risen over the last decade despite CRT implementation [17, 18]. Dissatisfaction from service users with CRT care has also been reported in recent national reports [19, 20]. These findings may reflect the incomplete and inconsistent CRT implementation in the UK: a national survey by Onyett and colleagues [21] found that only 40% of CRTs considered themselves fully implemented as intended, with wide variation in teams’ organisation and service delivery. Similar variation in CRT services’ characteristics was found in a more recent UK survey [22]. In Norway too, emerging evidence suggests that CRTs are providing a less intensive, less frequently home-based service, to a less acutely unwell client group than originally intended, with a consequent diminished impact on averting hospital admissions [10]. A recent systematic review [23] found little empirical evidence about the critical ingredients of CRTs, but found that there are indications from qualitative research, surveys and guidelines about which aspects of CRT service delivery and organisation are considered important or helpful by stakeholders.

The lack of a clearly specified CRT model, and the suggestion that potential benefits of CRTs are not being consistently achieved when services are scaled up to a national level, indicate the need for a rigorously defined and well-validated CRT fidelity scale. In the absence of a clearly prescribed theoretical model or sufficient empirical evidence about the critical ingredients of an intervention or service model, stakeholders’ views regarding best practice may also inform the development of fidelity criteria. Structured approaches used in fidelity scale development to elicit stakeholder opinion have included a Delphi process [24] and concept mapping [25]. Stakeholder groups often include researchers, program leaders, practitioners, and service users [26, 27]. In this study, we aimed to systematically develop a fidelity scale for CRTs; to test the feasibility and utility of the scale in practice settings; and to conduct a preliminary exploration of its psychometric properties. This work was undertaken as part of a larger research programme on implementation of CRTs, the CORE Study [28]. The CORE Study programme as a whole aimed to develop evidence to inform effective CRT implementation. It involved: i) developing evidence regarding the optimal CRT model from a systematic literature review, a national survey, and interviews with a range of stakeholders; ii) development and testing of a measure to assess model fidelity in CRTs (the work reported in this paper); and iii) development and testing in a cluster randomised trial of a package of service improvement resources designed to enhance model fidelity and improve outcomes in CRT teams. Further information about the CORE Study as a whole is available from the study website [29] and the trial protocol for the service improvement programme trial [30].

Methods

Development of the CRT fidelity scale consisted of three steps: construction of the scale; piloting and refinement; and exploration of its psychometric properties.

Construction of the fidelity scale

Concept mapping was proposed by [31] as a structured process to facilitate group participation in developing conceptual frameworks to guide evaluation. We used a concept mapping process to construct the CRT fidelity scale, following the six stages described by Trochim of: i) developing the focus for conceptualisation; ii) generating statements; iii) group participation in conceptualising (grouping) and prioritising statements; iv) representing these statements in a concept map; v) interpreting the map; vi) utilising the map.

  1. i.

    Developing the focus for conceptualisation: Potential characteristics of CRT resources, organisation and service delivery for inclusion in a fidelity scale were identified from three sources: a literature review of quantitative and qualitative studies and guidelines relating to CRT implementation [23]; a UK national survey of CRT managers, reporting description of teams’ organisation and service delivery and managers’ views on priorities for effective CRT implementation [22]; and over 100 interviews and focus groups with CRT stakeholders (CRT service users, carers, staff and managers; and other stakeholders from organisations which refer to or work with CRTs) conducted for the CORE study [28]. The list of potential CRT fidelity characteristics was also confirmed by the results from a similar survey of the 56 CRT managers in Norway and a qualitative study of experiences of service users, carers, team members and collaborating services in Norway [32].

  2. ii.

    Generating statements: From a “longlist” of potential components of a CRT model generated in stage 1, a group of CRT stakeholders (n = 10), comprising clinicians and academic researchers, including service user-researchers, were asked to develop a set of fewer than 100 statements specifying potential components of a CRT fidelity scale. Participants completed this task as a group exercise, through discussion and manual sorting of cards (each with a statement on it). At this stage, participants were asked to de-duplicate, collapse or combine conceptually related statements rather than to judge the relative importance of statements or exclude any distinct areas of CRT organisation or practice included in the statements.

  3. iii.

    Group participation in conceptualising statements: Stakeholder involvement was sought through concept mapping meetings. Participants were invited to one of four meetings, held in London, UK (n = 2), Northampton, UK (n = 1) or Oslo, Norway (n = 1). Participants who could not attend a meeting could complete the concept mapping tasks individually and return their results to the research team. Participants were sought from the following six CRT stakeholder groups: i) service users; ii) family and friends supporting service users (carers); iii) CRT staff; iv) other mental health staff (including senior managers and staff from mental health services which work with CRTs, including acute wards and Community Mental Health Teams; v) staff from voluntary sector organisations providing support to people with mental health problems; and vi) academic researchers involved in acute care research. Participants were convenience sampled from service user and carer research groups, clinical professional networks and a clinicians’ advisory group already assembled for the CORE Study, UK and Norwegian clinical and research networks, and via a large national mental health charity in England (MIND).

    Participants were given two sets of cards, each set with a set of statements relating to service organisation or delivery in CRTs. Participants were also provided with an accompanying sheet with brief information clarifying the meaning of each statement and presenting a rationale for its inclusion (based on development work from stage 1, and indicating whether any empirical evidence, policy guidance or evidence of stakeholders’ views supported its inclusion). First, participants were asked to group cards together into a minimum of two groups, according to the participant’s view of how the statements best fit together, and to name each group (the conceptualisation task). Secondly, participants were asked to sort the cards into five equal-sized groups, identifying those viewed as most important, next most important, down to least important for delivering an effective CRT service (the prioritisation task). Participants completed the tasks individually, without discussion. Research staff were present to help explain the tasks where necessary and record results from each participant.

  4. iv.

    Representation of statements in a concept map: Participants’ data from the concept mapping exercises were entered into a specialist concept mapping software programme “Ariadne” [33]. Ariadne generates outputs regarding the mean importance ratings for each statement (for all participants and participant groups, with each statement scoring on a scale 1–5 for each participant, based on their prioritisation of items). Using principal component analysis and cluster analysis, it generates a series of concept maps, identifying how statements can best be grouped together in cluster solutions ranging from 2–20 clusters. The concept maps are shaped based on the participants’ conceptualisation of statements.

  5. v.

    Interpretation of concept maps: A stakeholder group (n = 8) including CRT service users, carers, mental health staff and academics reviewed the 19 cluster solutions generated by Ariadne. With reference to the statements grouped within each cluster for each solution, group members were asked to: a) Select the cluster solution with greatest conceptual coherence; and b) name each cluster (referring back to participants’ naming of groups in the concept mapping exercise). A consensus was sought through discussion.

  6. vi.

    Utilising the concept map: The chosen cluster solution, i.e. the final concept map, was then used as a basis for developing the CRT fidelity measure. Statements representing each concept map cluster (i.e. each CRT conceptual domain) were included in the fidelity scale. Decisions regarding the number of statements from each cluster to be included in the scale were guided by the mean importance score overall for statements within each cluster, as well as the number of statements the cluster contained. Statements with higher mean importance scores within each cluster were prioritised for inclusion. The mean importance scores from each participant group were also inspected for each cluster and (with participants regrouped into three broader groups – service users and carers, mental health staff, others), compared using bivariate tests for individual items: additional consideration was given to items prioritised by any respondent group.

    Once statements for inclusion in a fidelity measure had been selected, the research team transformed these essential components into items useful for a CRT fidelity scale, by developing operational definitions and scoring anchor points for each item. This was achieved through an iterative process of reviews of evidence on CRT functioning and, where possible, evidence which supported the dosing of interventions. The resultant fidelity items allowed adherence to each item to be scored on a five-point scale, in keeping with other well-established fidelity measures [34, 35]. At each stage, further stakeholder consultation was sought from: concept mapping participants, advisory working groups of service user and carer researchers attached to the CORE Study, clinical networks (e.g. the Royal College of Psychiatrists CRT network) and available CRT experts. Views were sought regarding: whether the criteria are valid – i.e. do they belong in a model of best practice for CRTs, whether the criteria are attainable in routine service settings; and whether criteria could be reliably measured during an audit process.

Piloting

To pilot the fidelity scale and test its feasibility as a measure of model adherence for CRTs, a 1-day fidelity review process was developed. This followed the processes for assessing fidelity developed for established scales [34, 35]. Reviews were carried out by a team of three reviewers, who visited a CRT service for a full day. To prepare for the fidelity review, a member of the review team contacted the CRT management prior to the visit, explaining the purpose, identifying documents the review team would need, and setting up the schedule. Review teams included at least one mental health clinician and one service user or carer, as well as one member of the study research team. Reviews involved: interviews with multiple groups (the CRT manager and staff team, managers of other services which work closely with the CRT, 6 CRT service users and 6 carers); a review of anonymized case records for the 10 most recent, consecutively discharged, service users; and review of service policies, records and routinely collected data. Reviewers collected evidence using during the review day using interview schedules and checklists provided to them, then met to share information and collectively score each fidelity item at the end of the review day. A draft fidelity review report was then sent to the CRT manager, seeking clarification on any outstanding issues and offering an opportunity for the CRT manager to query any scores and provide further evidence if available, before a finalised report and score were provided. A half-day training programme and electronic training materials were provided in advance to all reviewers, in which interview schedules, checklists and scoring guidance for use during reviews were provided.

Piloting of the fidelity scale and review process was planned in two stages; an initial pilot of four teams, then a larger programme of reviews in 75 teams. The protocol for this programme of reviews was confirmed by the London Camden and Islington Research Ethics Committee as meeting criteria for audit rather than research [36], so approvals from participating NHS organisations only were required. Study researchers contacted CRT managers about willingness to participate in a fidelity review. Where the CRT manager was interested, the NHS Trust Research and Development Department was contacted: any registration or application processes required by Trusts to conduct the fidelity review as an audit were complied with. For each review, the manager of the participating CRT service invited other staff, service users and carers to take part in the review and the review team provided an information sheet for potential service user and carer participants. The review team also confirmed the team’s willingness to take part in the review, and provided all participants an opportunity to ask questions on the review day.

The review team sought feedback from the manager of all participating CRT teams regarding the acceptability of the review process, clarity and validity of the fidelity scale criteria, and whether scoring anchor points for individual items were set appropriately. The research team monitored the range of scores being achieved for each item, to identify potential floor or ceiling effects in scale items. A review of the scale was planned after 50 reviews, to allow use of later reviews to test a Version 2 of the scale if necessary.

Psychometric properties

Three properties of the CRT fidelity scale were explored. Feedback from managers of CRT services undergoing a fidelity review was used to assess the scale’s face validity. The range of scores generated in the 75-team survey, for total and item scores, was used to assess for restricted range (e.g., floor or ceiling effects). Reviewers’ inter-rater reliability was tested using an extended vignette. Researchers (JF, KF, DL) developed mock interview transcripts and audit records for a fidelity review, informed by records from actual reviews. Reviewers from the 75-team survey were then invited to score each of the scale’s 39 fidelity items with reference to the mock records, as an individual exercise. Reviewers’ scores for each item were entered into Stata Statistical Software version 13 for analysis. Absolute agreement intra class correlations were calculated, based on a two-way random effects model [37].

Results

Construction of the fidelity scale

An initial list of 232 statements relating to recommended characteristics of CRT services was collapsed to 72 statements regarding good practice in CRTs’ organisation and service delivery. In total, there were 68 participants in the concept mapping exercises, including representatives of all six stakeholder groups, and participants from the UK and Norway, the Netherlands and Australia. Participants’ characteristics are summarised in Table 1.

Table 1 Characteristics of participants in the concept mapping exercise (N = 68)

Participants grouped statements into between 2–12 groups. Mean and median importance scores for each statement are provided in the data supplement (Additional file 1: Table DS1). A four-cluster solution was selected as the most conceptually coherent model (Fig. 1). The stakeholder working group (n = 8) named the clusters as: referrals and access; content and delivery of care; staffing and team procedures; timing and location of care.

Fig. 1
figure 1

CORE CRT concept map: the chosen 4-cluster solution

The three broadly defined stakeholder groups (service users and carers; mental health staff; others) showed high agreement in the importance ratings of most items. Bivariate tests found significant differences between stakeholder groups in mean importance ratings for 15 (21%) of 72 statements. All statements rated very important (mean score above 4) by any of the three groups, or moderately important (mean score above 2.5) by more than one stakeholder group were included in the scale, either as a distinct item, or among item scoring criteria. Additional file 2: Table DS2 provides further information and an explanation of how discrepancies between stakeholder groups’ ratings were accounted for in development of the fidelity scale.

Thirty nine statements were selected for inclusion in the CRT Fidelity Scale. Table 2 reports how the internal structure of the fidelity scale relates to the clusters and their mean importance rating from the concept map. Statements representing each of the four clusters were included in the fidelity scale: the relative contribution to the total scale and score of each cluster was guided by the number of statements, and the mean importance rating for statements, within each cluster. Further consultation (within the study team, with study working groups of service users and carers, with concept mapping participants, and with other available CRT experts) following the concept mapping process, resulted in slightly greater representation in the scale of items relating to the content of care than was indicated by the concept map.

Table 2 The relationship of the CRT fidelity scale structure to concept map clusters

Piloting

During initial piloting in four CRT services we refined scoring criteria and guidance to make the wording clearer. This resultant revised version of the fidelity scale (“Version 1”) was then piloted in 75 CRTs in England, Scotland and Wales, serving a range of inner city, suburban and more rural areas, and catchment area populations ranging from about 100,000 to 400,000. The 75 CRTs surveyed included 70 in England – about a third of CRT teams within England. Reviews were completed successfully and a report and score generated for all 75 teams which agreed to take part in the survey. Feedback from CRT managers was that the fidelity reviews were acceptable and the reports helpful, but that the preparation required by the team was time-consuming and at the upper limit of what was manageable. Fidelity reviews required long days (typically 8–10 h), with scoring on occasion being completed by reviewers after the scheduled finish time.

The information required by reviewers was not always all available on the review day: in particular, six service users and carers were not always contactable or available. Wherever possible, reviewers sought to obtain missing information later (for instance by returning to a service for a second day, or conducting interviews by phone). Services were only scored on the information provided: some teams may therefore have scored lower than they might have. CRT managers also reported that for some items, low scores may have reflected poor documentation in case notes of care provided, rather than a lack of care itself.

This 75-team survey yielded a range of total fidelity scores of 73–151 (maximum possible range 39 – 195). At item level, the maximum range of scores (1–5) was obtained for 33 items; 4 items (items 2, 4, 15 and 39) yielded a range of scores from 2–5; and 2 items (items 16 and 24) yielded a range of scores from 1–4. Thirty one items yielded a median score in the range 2–4. For three items, a majority of teams received the minimum score of 1 (item 17 (51% scored 1), item 24 (80%) and item 37 (56%)). For five items, a majority of teams received the maximum score of 5 (items 19 (51% scored 5), 26 (53%), 27 (63%), 32 (56%) and 39 (56%)). The median score and range of scores for each item are reported in Table 3.

Table 3 CORE CRT Fidelity Scale - item median scores and range of scores from pilot (n = 75)

The research team reviewed Version 1 of the fidelity scale following 50 reviews. Additional file 3: Table DS3 summarises the changes made to the scale following this review and their rationale. Changes were made to items and scoring criteria in response to: i) feedback from CRT managers and teams during the survey about the validity or items; ii) feedback from reviewers about the clarity of items or how possible it was to retrieve the required information during a review day; iii) floor or ceiling effects observed from the range and variance of scores for individual items which limited their ability to distinguish higher and lower fidelity services; and iv) further consultation with stakeholders, including review of new government and expert guidance for CRTs published during the survey. All 39-items were retained. New criteria were added to four items, and criteria were dropped, amended or combined in an additional 15 items.

The refined, Version 2 of the CRT Fidelity Scale was used alongside Version 1 in the last 9 reviews of the 75-team survey (reviewers scored both versions). Teams’ total scores were a mean of 2 points higher on V2 compared to V1 of the scale (range 0–6 points). For each of the two items with most marked floor effects in the pilot (items 16 and 24), three teams increased their score by one point in V2 compared to V1, and six teams’ scores were unchanged. The effect of the changes to Version 2 of the scale on teams’ overall scores was therefore modest. The feedback from reviewers and participating services was that the changes increased the clarity of the scale and retained good face validity.

Psychometric properties

Seventeen reviewers participated in the inter-rater reliability exercise using Version 2 of the CRT fidelity scale and an extended case note vignette. Sixteen participants provided complete data which were used in analysis. These included service user and carer, clinician, and non-clinical researcher reviewers.

The estimated correlation between individual ratings was 0.65 (95 CI: 0.54 to 0.76), indicating reasonably high similarity between ratings within an item. The estimated intra-class correlation (ICC) between ratings averaged over the 16 raters was very high, 0.97 (0.95 to 0.98). Consistency of agreement ICCs produced nearly identical values to those above; the average consistency of agreement ICC (0.97, above) is equivalent to Cronbach’s alpha.

Mean scores and the standard deviation were calculated for each item to establish which items showed the most variability (high SD), and are provided in Additional file 4: Table DS4. It is unclear, however, from this single vignette, whether these results reflect the inherent reliability of scale items, or the clarity of the specific information within this vignette.

Discussion

Main findings

This paper describes the development of the first measure of model fidelity for Crisis Resolution Teams: the CORE CRT Fidelity Scale. The CRT model specified within the scale is based on best available evidence and stakeholders’ priorities for CRT service delivery and organisation. The development of the scale has followed a structured and transparent concept mapping process. The 75-team pilot shows that the resulting CRT Fidelity Scale and review process have good feasibility and acceptability for use in CRTs. Results from the pilot show that the scale can distinguish higher and lower fidelity services overall, and that most items also generate a fairly balanced spread of scores across items’ five point scoring range. A minority of items where high scores were obtained by either very few or very many teams were retained with minimal changes for two reasons: first, further consultation with stakeholders confirmed that these items accurately describe important components of the CRT service model; second, a national survey of CRTs undertaken as development work for the fidelity scale [22], and the fact that at least some teams in the 75-team pilot scored good or excellent fidelity, both suggest that the scoring criteria are attainable in practice.

A 9-team pilot of version 2 of the CORE CRT Fidelity Scale suggests it generates only modest changes to total and item scores compared to version 1, but increases the clarity of the scale. The inter-rater reliability testing of V2 of the CORE CRT Fidelity Scale indicates promising initial psychometric properties.

Limitations

Two limitations of the scale development work reported here are: the scope of consultation in the development of the scale; and the extent of psychometric testing conducted.

First, regarding consultation, we aimed to include all stakeholder groups’ views at all stages of the development of the scale through: i) a thorough review of available evidence, including qualitative research, surveys and guidance as well as empirical studies at the statement-generating stage; ii) inclusion of service user, clinical staff and other stakeholder representatives at statement selection, concept mapping and concept mapping cluster solution-stages; and iii) consultation with service users, carers and clinicians at the fidelity item development stage through research forums and national networks. However, participants in our concept mapping exercise were pragmatically recruited: clinicians slightly outnumbered service users and carers, and some important stakeholder groups (e.g. General Practitioners, service commissioners, emergency services staff (e.g. police and ambulance) were not represented. Most of those who participated in consultations were UK stakeholders: while a small number of participants from the Netherlands, Norway and Australia contributed to the concept mapping exercise, how far the CRT fidelity scale reflects the priorities of stakeholders in other countries, and is suitable for use outside the UK, remains to be investigated.

Second, more extensive testing of the psychometric properties of the CRT Fidelity Scale is desirable: for example, assessing the test-retest reliability of the scale. This was not attempted in our study. The demands and time consumed by a fidelity review day and preparation for it are considerable for participating CRT services. Expecting CRT teams to participate in a second review soon after the first, which would provide no additional benefit to the service, was therefore not feasible.

Exploring inter-rater reliability in vivo was also not attempted. During a fidelity review, the three reviewers all collect evidence from different sources, then meet at the end of the day to share information and agree fidelity scores. Because no single reviewer holds all the required information until scores are discussed, individual reviewers could not assess services’ fidelity independently, to allow comparison of scores.

The vignette exercise used to assess inter-rater reliability drew on information collected from actual reviews (anonymised, and combined from multiple services), giving the exercise a high degree of realism. It only tested the reliability of reviewers’ ratings from the available evidence however, rather than how consistently reviewers gather evidence during a review day. Moreover, conclusions about the inter-rater reliability of the CRT Fidelity Scale, when based on ratings from a single vignette, can only be viewed as provisional. Arguably however, the vignette exercise provided a harder test of inter-rater reliability than an actual review, in that participants in the exercise made their scores entirely independently. In a real fidelity review, by contrast, three reviewers discuss the evidence collected and how to score a service, and refer back to the CRT manager or others to collect more information where necessary. In this light, the moderately good inter-rater reliability of the scale indicated by the vignette exercise, is promising, and suggests the reviewer training and scoring guidance within the scale are adequate to allow the scale to be used reliably.

The rigorous initial evidence gathering, and positive feedback from stakeholders and CRT teams which participated in piloting, demonstrate that the CRT Fidelity Scale has good face validity. The concept mapping grouping exercise, and cluster structure underpinning the scale, also afford it good content validity. However, the criterion validity of the scale, i.e. the relationship between a high fidelity score and outcomes for CRT services, has yet to be explored. This step is critical to establishing the utility of a fidelity scale [38]. Establishing criterion validity for the scale as well as for individual items is of particular importance, given the lack of empirical evidence about critical ingredients of CRT services [23] which was available to inform scale development during the development of statements about CRT best practice or stakeholders’ prioritisation of statements in the concept mapping process.

Clinical implications

The successfully completed 75-team pilot suggests that the fidelity scale and accompanying audit process, involving clinician and service user or carer-reviewers, is feasible and acceptable to CRT services. Experience from piloting suggests three features of the review process can help to maximise the reliability of fidelity reviews and scoring, and thus enhance the potential clinical utility of the scale. First, reviewing teams should include at least three reviewers, to manage the workload of a review day, provide a range of expertise to inform the evidence review, and to help the reliability of the review scoring process by moderating any outlying views of individual reviewers. Second, using the interview guides and checklists developed for reviewers aids consistent collection and recording of information with which to score a team using the scale. Third, a right of reply by the CRT team manager to an initial draft of the fidelity report and scores can allow any additional evidence to be provided, or misunderstandings by the reviewing team to be corrected.

The CRT Fidelity Scale can be a useful tool for service planners, managers and commissioners in three ways. First, it provides clear specification of the CRT model, in more detail than has been previously provided in statutory or expert guidance [7, 39]. This can guide service planners in setting commissioning specifications for Crisis Resolution Teams, and guide CRT managers and staff in setting service improvement goals. Second, a fidelity review using the CRT Fidelity Scale, which provides services with a score and feedback on each of the scale’s 39 items, can help CRT teams recognise their existing strengths and areas where service development is required. It can thus help teams to make service improvement plans and assess the impact of quality improvement initiatives. We have demonstrated that an external audit is feasible, acceptable, and, preliminary evidence suggests, adequately reliable. The CRT Fidelity Scale could in principle also be used for internal audit within health organisations, or self-audit by teams (although the reliability of scores when the scale is used this way is yet to be tested). Third, at a local or national level, the CRT Fidelity Scale can be used to generate benchmarking data about CRTs’ model fidelity, against which individual teams’ model fidelity can be assessed, or with which changes to service provision across a region or country over time could be explored. Our 75-team survey demonstrates that such a large-scale use of the CRT Fidelity Scale can be achieved. (All 75 reviews in our study were completed in less than 1 year.)

The value of the CRT fidelity scale to service planners and policy-makers is demonstrated by the speed with which it has been promoted and used by expert bodies and policy-making organisations nationally in the UK. The CRT Fidelity Scale is advertised as a “national inspiration” in the Crisis Care Concordat campaign led by NHS England [40]. Benchmarking data from our 75-team survey have been used by the Care Quality Commission, the body responsible for quality inspections of health and social care services in England [20], and the Royal College of Psychiatrists [41] in recent reports presenting recommendations for CRT care.

Research implications

An important next step in developing the CRT Fidelity Scale is investigating its criterion validity. Fidelity scales assess elements of service or intervention structure and process: these can only be legitimately used as measures of service quality if a positive relationship to outcomes has been demonstrated [42]. Investigating the relationship of teams’ CRT Fidelity Scale score to important outcomes like service users’ experience of and satisfaction with services, recovery following CRT care, and inpatient admission rates and bed use across a catchment area, is of high research interest. An initial exploration of the criterion validity of the CRT Fidelity Scale will be carried out in connection with a current trial of a CRT service improvement programme, also conducted as part of the CORE Study [30].

A further question is whether the CRT Fidelity Scale can be an internationally useful measure, as other Fidelity Scales have become [34]. There is a need to explore its feasibility and utility in non-UK contexts. Norway, as the other country where CRTs have been implemented nationally, is an obvious site for further testing of the scale.

While a fidelity review may in itself help services plan service improvement, there is also a pressing clinical need to develop and test resources to help CRT services meet the priorities of stakeholders, and achieve high model fidelity, with the aim of improving service outcomes. The development of a program manual, which can then be used in the training and supervision of programme staff, is advocated as a key tool in supporting high fidelity implementation of evidence based practices [43]. Development of a CRT manual is required, informed by the CORE CRT Fidelity Scale. The CORE CRT Service Improvement Programme Trial [30] will test a package of service improvement resources in a cluster randomised trial: 15 CRT teams will receive the package of resources over a 1-year period; 10 teams will act as controls. Fidelity reviews will be used to support and evaluate the service improvement intervention.

The development of the CRT Fidelity Scale also has two implications regarding research methods. First, concept mapping [29] proved to be a useful method of developing and defining the CRT service model. It allowed the views of several stakeholder groups to contribute to developing the model, and provided a transparent basis for making decisions about which of many competing elements of CRT services to prioritise for inclusion in a fidelity scale. Second, the development and piloting of the CRT fidelity scale provides a model for patient and public involvement in research and clinical audit, as recommended to improve the quality and credibility of research [44].

Conclusion

Crisis Resolution Teams are a complex intervention, with evidence of effectiveness from trials [12], but for which implementation has been variable [21, 22], and which appear to have been less effective than hoped for when scaled up to national level [16, 18]. Implementation has been identified as the major barrier to translating scientific knowledge into patient benefit [2]. The CORE CRT Fidelity Scale can support implementation of CRTs by providing clear and detailed specification of the CRT model, a means to assess CRT services’ adherence to this model, and a means to support and focus CRT service improvement initiatives at the levels of practice and policy. It fills an unmet need, in the previous absence of a fidelity scale for CRTs, and can help in the overarching goal of improving support for people experiencing a mental health crisis.

Abbreviations

CRT:

Crisis Resolution Team

ICC:

intra-class correlation

NHS:

National Health Service

References

  1. Bond GR, Evans L, Salyers MP, Williams J, Kim HK. Measurement of fidelity in psychiatric rehabilitation. Ment Health Serv Res. 2000;2:75–87.

    Article  CAS  PubMed  Google Scholar 

  2. Tansella M, Thornicroft G. Implementation science: understanding the translation of evidence into practice. Br J Psychiatry. 2009;195:283–5.

    Article  PubMed  Google Scholar 

  3. McHugo GJ, Drake RE, Whitley R, Bond GR, Campbell K, Rapp CA, et al. Fidelity outcomes in the National Implementing Evidence-Based Practices Project. Psychiatr Serv. 2007;58:1279–84.

    Article  PubMed  Google Scholar 

  4. Bond GR, Peterson AE, Becker DR, Drake RE. Validating the revised Individual Placement and Support Fidelity Scale (IPS-25). Psychiatr Serv. 2012;63:758–63.

    Article  PubMed  Google Scholar 

  5. Monroe-DeVita M, Teague GB, Moser LL. The TMACT: A new tool for measuring fidelity to assertive community treatment. J Am Psychiatr Nurses Assoc. 2011;17:17–29.

    Article  PubMed  Google Scholar 

  6. Glover G, Johnson S. The crisis resolution team model: recent developments and dissemination. In: Johnson S, Needle J, Bindman J, Thornicroft G, editors. Crisis Resolution and Home Treatment in Mental Health. Cambridge: Cambridge University Press; 2008. p. p23–35.

    Chapter  Google Scholar 

  7. Department of Health “The Mental Health Policy Implementation Guide” Department of Health. 2001. http://webarchive.nationalarchives.gov.uk/20130107105354/http://www.dh.gov.uk/prod_consum_dh/groups/dh_digitalassets/@dh/@en/documents/digitalasset/dh_4058960.pdf. Accessed 23 May 2016

  8. Johnson S, Needle J. Crisis Resolution Teams: rationale and core model. In: Johnson S, Needle J, Bindman J, Thornicroft G, editors. Crisis resolution and home treatment in mental health. Cambridge: Cambridge University Press; 2008. p. p67–84.

    Chapter  Google Scholar 

  9. Department of Health “The NHS Plan: a plan for investment, a plan for reform” UK Department of Health. 2000 Web resource http://webarchive.nationalarchives.gov.uk/20130107105354/http://www.dh.gov.uk/prod_consum_dh/groups/dh_digitalassets/@dh/@en/@ps/documents/digitalasset/dh_118522.pdf. Accessed 23 May 2016.

  10. Hasselberg N, Grawe RW, Johnson S, Ruud T. Treatment and outcomes of crisis resolution teams: a prospective multicentre study. BMC Psychiatry. 2011;11:183.

    Article  PubMed  PubMed Central  Google Scholar 

  11. Murphy S, Irving C, Adams C, Driver R. “Crisis intervention for people with severe mental illnesses” Cochrane Database of Systematic Reviews Issue 5. 2012. Art. No.: CD001087. DOI:10.1002/14651858.CD001087.pub4

  12. Johnson S, Nolan F, Pilling S, Sandor A, Hoult J, McKenzie N, White IR, Thompson M, Bebbington P. “Randomised controlled trial of acute mental health care by a crisis resolution team: the north Islington crisis study” British Medical Journal. 2005a;331(7517):599.

  13. Johnson S, Nolan F, Hoult J, White IR, Bebbington P, Sandor A, et al. “Outcomes of crises before and after introduction of a Crisis Resolution Team. British Journal of Psychiatry. 2005b; 187 p68-75.

  14. Jethwa K, Galappathie N, Hewson P. Effects of a crisis resolution and home treatment team on in-patient admissions. Psychiatr Bull. 2007;31:170–2.

    Article  Google Scholar 

  15. Glover G, Arts G, Babu KS. Crisis resolution/home treatment teams and psychiatric admission rates in England. Br J Psychiatry. 2006;189:441–5.

    Article  PubMed  Google Scholar 

  16. Jacobs R, Barrenho E. The Impact of Crisis Resolution and Home Treatment Teams on Psychiatric Admissions in England. J Mental Health Policy Econ. 2011;14:S13.

    Google Scholar 

  17. Keown P, Weich S, Bhui K, Scott J. Association between provision of mental illness beds and rate of involuntary admissions in the NHS 1988–2008: ecological study. BMJ. 2011;343:d3736.

    Article  PubMed  PubMed Central  Google Scholar 

  18. Health and Social Care Information Centre. Inpatients formally detained in hospitals under the Mental Health Act 1983, and patients subject to supervised community treatment: Annual report, England, 2013/14.2014. http://www.hscic.gov.uk/catalogue/PUB15812/inp-det-m-h-a-1983-sup-com-eng-13-14-rep.pdf. Accessed 23 May 2016.

  19. MIND “Listening to experience: An independent inquiry into acute and crisis mental healthcare” 2011; MIND, London.

  20. Care Quality Commission “Right here right now: people’s experiences of help, care and support during a mental health crisis” 2015 Care Quality Commission, London http://www.cqc.org.uk/sites/default/files/20150630_righthere_mhcrisiscare_full.pdf. Accessed 23 May 2016

  21. Onyett S, Linde K, Glover G, Loyd S, Bradley S, Middleton H. Implementation of crisis resolution/home treatment teams in England: national survey 2005–2006. Psychiatr Bull. 2008;32:374–7.

    Article  Google Scholar 

  22. Lloyd-Evans B and Johnson S. “Crisis Resolution Teams. How are they performing?” May/June 2014. Mental Health Today. https://www.mentalhealthtoday.co.uk/crisis_resolution_teams__how_are_they_performing_25769813430.aspx. Accessed 20 Nov 2015.

  23. Wheeler C, Lloyd-Evans B, Churchard A, Fitzgerald C, Fullarton K, Mosse L, et al. Implementation of the Crisis Resolution Team model in adult mental health settings: a systematic review. BMC Psychiatry. 2015;15:74.

    Article  PubMed  PubMed Central  Google Scholar 

  24. Addington DE, McKenzie E, Norman R, Wang JL, Bond GR. Identification of essential evidence-based components of first episode psychosis services. Psychiatr Serv. 2013;64:452–7.

    Article  PubMed  Google Scholar 

  25. Armstrong N, Steffen J. The recovery promotion fidelity scale: assessing the organisational promotion of recovery. Community Ment Health J. 2009;45(3):163–70.

    Article  PubMed  Google Scholar 

  26. Schaedle RW, Epstein I. Specifying intensive case management: a multiple stakeholder approach. Ment Health Serv Res. 2000;2:95–105.

    Article  CAS  PubMed  Google Scholar 

  27. Holter MC, Mowbray CT, Bellamy CD, MacFarlane P, Dukarski J. Critical ingredients of consumer run services: results of a national survey. Community Ment Health J. 2004;40:47–63.

    Article  PubMed  Google Scholar 

  28. UK Clinical Research Network Study Portfolio (web resource) http://public.ukcrn.org.uk/Search/StudyDetail.aspx?StudyID=9937. Accessed 20 Nov 2015.

  29. The Core Study. University College London (web resource) https://www.ucl.ac.uk/core-study. Accessed 22 Oct 2016.

  30. Lloyd-Evans B, Fullarton K, Lamb D, JohnstonE, Onyett S, Osborn D et al. “The CORE Service Improvement Programme for mental health crisis resolution teams: study protocol for a cluster-randomised controlled trial” BMC Trials 2016; 17:158 DOI 10.1186/s13063-016-1283-7

  31. Trochim W. An introduction to concept mapping for planning and evaluation. Eval Program Plann. 1989;12(1):1–16.

    Article  Google Scholar 

  32. Ruud T, Karlsson B, Klevan T, Hasselberg N Crisis Resolution Teams in mental health services: Practice and experiences. [Ambulante akutteam i psykisk helsevern: Praksis og erfaringer.] 2015; Report, Akershus universitetssykehus, Lørenskog.) https://www.usn.no/getfile.php/usn.no/filer/forskning/Hva%20forsker%20vi%20p%C3%A5/Helse%20og%20velferd/Psykisk%20helse%20og%20rus/Ambulante%20akutteam%20i%20psykisk%20helsevern%20Praksis%20og%20erfaringer.pdf. Accessed 23 May 2016.

  33. Severens P. Handbook for Concept Mapping, National Centre for Mental Health. Amsterdam: Talcott; 1995.

    Google Scholar 

  34. Becker DR, Swanson S, Reese SL, Bond GR, McLeman BM. Evidence-based supported employment fidelity review manual. 3rd ed. Lebanon: NH: Dartmouth Psychiatric Research Center; 2015.

    Google Scholar 

  35. Teague GB, Bond GR, Drake RE. Program fidelity in assertive community treatment: development and use of a measure. Am J Orthopsychiatry. 1998;68:216–32.

    Article  CAS  PubMed  Google Scholar 

  36. Health Research Authority “Defining research: NRES guidance to help you decide if your project requires review by an ethics committee” HRA 2016. http://www.hra.nhs.uk/documents/2016/06/defining-research.pdf. [Accessed 30 Nov 2016].

  37. McGraw KO, Wong SP. Forming inferences about some intraclass correlation coefficients. Psychol Methods. 1996;1:30–46.

    Article  Google Scholar 

  38. Bond GR, Becker DR, Drake RE. Measurement of fidelity of implementation of evidence-based practices: case example of the IPS fidelity scale. Clin Psychol. 2011;18:126–41.

    Google Scholar 

  39. Joint Commissioning Panel for Mental Health “Guidance for commissioners of acute care – Inpatient and crisis home treatment” 2013 http://www.jcpmh.info/wp-content/uploads/jcpmh-acutecare-guide.pdf. Accessed 22 Dec 2014.

  40. Mental Health Crisis Care Concordat “Improving outcomes for people experiencing a mental health crisis. HM Government” 2014. https://www.nao.org.uk/wp-content/uploads/2007/12/07085.pdf. Accessed 20 Nov 2015.

  41. Crisp N, Smith G and Nicholson K (Eds.). “Old Problems, New Solutions – Improving Acute Psychiatric Care for Adults in England (The Commission on Acute Adult Psychiatric Care, 2016)” 2016 http://www.cqc.org.uk/sites/default/files/20150630_righthere_mhcrisiscare_full.pdf. Accessed 23 May 2016.

  42. Donabedian A. Evaluating the quality of medical care: 1966. Millbank Q. 2005;83(4):691–729.

    Article  Google Scholar 

  43. Mowbray CT, Holter MC, Teague GB, Bybee D. Fidelity criteria: development, measurement, and validation. Am J Eval. 2003;24(3):315–40.

    Article  Google Scholar 

  44. Involve “Briefing notes for researchers: public involvement in NHS, public health and social care research” 2012; National Institute for Health Research, London http://www.invo.org.uk/wp-content/uploads/2012/04/INVOLVEBriefingNotesApr2012.pdf. Accessed 23 May 2016.

Download references

Acknowledgements

We are grateful for the time and help of all the research and clinical staff, and the members of the CORE Service User and Carer Advisory Group who contributed to the concept mapping process and advised on development of the CORE CRT Fidelity Scale. [CORE Advisory Group members: Mary Clarke, Stephanie Driver, Pauline Edwards, Kathleen Fraser-Jackson, Jackie Hardy, Kate Holmes, Lyn Kent, Jacqui Lynskey, Brendan Macken, Mary Plant, Tony Rivett, Jo Shenton, Geoff Stone, Janice Skinner, Jules Tennick, Gen Wallace]

We are grateful to Peter Severens for providing access to the Ariadne concept mapping software and User Handbook.

Our colleague, Steve Onyett, died suddenly in September 2015. Steve contributed to the design and development of the CORE CRT Fidelity Scale and the whole CORE Study research programme. His contribution to the study and his warmth and generosity as a colleague are greatly missed.

Funding

This paper presents independent research funded by the National Institute for Health Research (NIHR) under its Programme Grants for Applied Research programme (Reference Number: RP-PG-0109-10078). The views expressed are those of the author and not necessarily those of the NHS, the NIHR or the Department of Health.

Availability of data and materials

Two datasets supporting the conclusions of this paper are provided as Additional Files.

Additional file 5: Table DS5: Concept mapping participant prioritisation data: CORE Study data set.

Additional file 6: Table DS6: Inter-rater reliability testing participant data: CORE Study data set.

The CORE CRT Fidelity Scale and CRT Fidelity Review guidance are available via the CORE Study website: www.ucl.ac.uk/core-study.

Authors’ contributions

SJ leads the CORE Study, the research programme for which the work in this paper was conducted. BLE, RG, DO, FN, CH, OM, GA, NM and SO are co-applicants on the CORE Study and also contributed to the study design. BLE, GB, TR, RG, DO, FN, CH, OM, NG, KK, NM, SO, DL, SF, EB, BP, AS, DH, KF, JF and SJ all contributed to developing the CORE CRT Fidelity Scale. BLE, GB, RG, DO, FN, CH, OM, NG, KK, SO, DL, SF, EB, BP, AS, DH, KF, JF and SJ all contributed to developing and piloting the CORE CRT Fidelity Review process. JF led on developing the mock review documents used in the inter-rater reliability testing, with help from KF, DL, BLE and SJ. AI provided expert advice on concept mapping methods, and data management and analysis using Ariadne concept mapping software. GA conducted the statistical analysis of inter-rater reliability testing data. BLE led on writing the paper; all authors contributed to the final manuscript. All authors read and approved the final manuscript.

Competing interests

The authors declare that they have no competing interests.

Consent for publication

Not applicable.

Ethics approval and consent to participate

The CRT fidelity reviews described in this paper were approved as meeting the criteria for audit by the London Camden and Kings Cross Research Ethics Committee on 12th August 2013. Approvals for each fidelity review were obtained from participating NHS Trusts and CRT services, in advance. Consent was obtained from individual participants in fidelity reviews, who were approached initially via the CRT service manager or staff.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Brynmor Lloyd-Evans.

Additional information

Steve Onyett died in September 2015

This article has been updated to correct a spelling mistake in author affiliation 3.

Additional files

Additional file 1: Table DS1.

CORE CRT concept mapping: participant ratings – summary scores. (DOCX 21 kb)

Additional file 2: Table DS2.

CORE CRT concept mapping: statements with significant between-group differences in importance ratings. (DOCX 16 kb)

Additional file 3: Table DS3.

Changes to the CORE CRT Fidelity Scale post-piloting. (DOCX 16 kb)

Additional file 4: Table DS4.

Scoring of fidelity items in the CORE CRT Fidelity Scale inter-rater reliability exercise. (DOCX 21 kb)

Additional file 5: Table DS5.

Concept mapping participant prioritisation data: CORE Study data set. (XLSX 27 kb)

Additional file 6: Table DS6.

Inter-rater reliability testing participant data: CORE Study data set. (XLSX 11 kb)

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Lloyd-Evans, B., Bond, G.R., Ruud, T. et al. Development of a measure of model fidelity for mental health Crisis Resolution Teams. BMC Psychiatry 16, 427 (2016). https://doi.org/10.1186/s12888-016-1139-4

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12888-016-1139-4

Keywords