Skip to main content

Evaluating the CYP-IAPT transformation of child and adolescent mental health services in Cambridgeshire, UK: a qualitative implementation study

Abstract

Background

The Children and Young People’s Improving Access to Psychological Therapies (CYP-IAPT) programme was introduced to transform Child and Adolescent Mental Health Services (CAMHS) across England. The programme comprised a set of principles that local CAMHS partnerships were expected to operationalise and embed with the aim of increasing access to services and improving the quality of care. This study explored how the implementation of the CYP-IAPT programme was executed and experienced by CAMHS professionals in the county of Cambridgeshire (UK), and the extent to which the CYP-IAPT principles were perceived to be successfully embedded into everyday practice.

Methods

We analysed 275 documents relating to the CYP-IAPT programme issued between 2011 and 2015. We also conducted a thematic analysis of 20 qualitative interviews, undertaken at two time points, with professionals from three CAMHS teams in Cambridgeshire. Analysis was informed by implementation science frameworks.

Results

Document analysis suggested that the CYP-IAPT programme was initially not clearly defined and lacked guidance on how to operationalise key programme principles and apply them in everyday practice. There was also a degree of programme evolution over time, which made it difficult for local stakeholders to understand the scope and aims of CYP-IAPT. Interviews with staff showed low coherent understanding of the programme, variable levels of investment among stakeholders and difficulties in collaborative working. Barriers and facilitators to programme implementation were identified at individual, service and strategic levels. These in turn impacted the local implementation efforts and sustainability of the programme in Cambridgeshire.

Conclusions

We identified factors relating to programme design and national and local implementation planning, as well as features of inner and outer context, which impacted on the delivery and sustainability of the programme. These findings can be drawn upon to inform the development and delivery of other local and national quality improvement (QI) initiatives relating to children and young people’s mental health.

Peer Review reports

Background

The Children and Young People’s Improving Access to Psychological Therapies (CYP-IAPT) quality improvement (QI) programme was initiated in 2011 to fulfil a new vision of Child and Adolescent Mental Health Services (CAMHS) in England [1]. The underlying rationale of the initiative was that system-wide change could be achieved through embedding key principles into routine practice. The foundational principles of the programme focused on improving access to services, increasing service user participation, raising awareness of mental health issues, improving quality of care through evidence-based practice (EBP) and enhancing accountability through use of routine outcome monitoring (ROMs) [2, 3].

Implementation of the CYP-IAPT programme was supported by the use of Quality Improvement Collaboratives (QIC), which brought together teams from multiple organisations to share knowledge and learn about best practice with the aim of improving service quality. This was facilitated at national and regional levels by linking higher education institutions with local partnerships (consisting of statutory, third sector service providers and commissioners) to form ‘Learning Collaboratives’ which would provide a mechanism for implementing new QI practices. It was expected that these inter-organisational partnerships would transform service provision in their local area by developing regional implementation plans in collaboration with service users to reflect local needs and priorities, and by training existing staff in several evidence-based therapies based on the national CYP-IAPT training curricula [4]. However, despite the QIC model being widely used in health care, the evidence for this approach is equivocal [5, 6]. Whilst some studies find that QICs confer some benefits to patient care and outcomes [7, 8], others find there is insufficient evidence that collaboratives improve outcomes or processes [6]. Arguably, for a QIC to be effective, it is vital to create cohesive networks to facilitate care coordination and involve key individuals who act as liaisons to connect disparate groups [9]. Furthermore, organisational readiness [10] and buy-in from senior management are perceived to be key determinants of success and sustainability of QICs [11,12,13].

Evaluation of the CYP-IAPT programme implementation has been limited; at a national level, a full evaluation of service and user outcomes was not possible due to large amounts of missing data [14]. Furthermore, there has been little evaluation of how the implementation of CYP-IAPT has been experienced and embedded at a local level. One exception is a rapid internal audit of 12 partnership sites conducted by programme developers [15], which highlighted a range of barriers and facilitators of service transformation including the individual characteristics of practitioners and the particular implementation process chosen by a local service.

Good process evaluation is needed to determine the effectiveness and sustainability of (any) programme implementation in real world settings [5, 16]. Furthermore, formal evaluation provides an opportunity to capture ‘how to’ knowledge that tends to be lost over time [17]. There is merit in undertaking a local evaluation because it provides an in-depth account of how the local context influenced implementation and is a way of retaining tacit knowledge from stakeholders who have been involved in multiple QI efforts. This information gives a measure of success or lack of and can aid replication and optimisation of the intervention [5].

There have been considerable efforts to describe the process of moving an intervention from the research environment or its pilot phase to widespread use [18]. Proctor et al. [19] presented a heuristic model to guide implementation research in the context of mental health services, which distinguishes between the evidence-based programme or practices, implementation strategies, and implementation, service, and client outcomes. According to this model, implementation outcomes relate to issues such as the feasibility, acceptability and reach of the programme being implemented and are distinct from service level outcomes such as effectiveness and safety, and client outcomes such as satisfaction.

The aim of the current study was to examine the process through which CYP-IAPT was implemented in an early pilot site: the Cambridgeshire and Peterborough Foundation Trust (CPFT) which is part of the English National Health Service (NHS). It explored the extent to which the CYP-IAPT principles have become embedded in everyday practice and identified perceived barriers and facilitators to local implementation. In line with Proctor’s model of implementation research, we sought to (i) describe the implementation strategies offered by programme developers and those developed and used by local implementers and (ii) evaluate implementation outcomes, specifically feasibility, fidelity, acceptability and sustainability. We used qualitative methods to explore stakeholder descriptions of the implementation process and associated outcomes drawing on normalisation process theory (NPT) [20,21,22] to explore the social processes underpinning implementation, and the Consolidated Framework for Implementation Research (CFIR) [23] to identify specific factors that impacted on the implementation process.

Methods

We undertook a document analysis to describe the intended CYP-IAPT implementation model as articulated by programme developers and to describe the local activities and strategies used to embed the CYP-IAPT principles into routine practice in Cambridgeshire. We conducted interviews with CAMHS staff at two time points (early and late in the implementation process) to explore their views and experiences of the programme implementation. Reporting followed the Standards for Reporting Implementation Studies (StaRI) [24] and Consolidated Criteria for Reporting Qualitative Studies (COREQ) guidelines (additional file 1) [25].

Data sources and sample

Document analysis

Documents relating to the implementation and delivery of the CYP-IAPT initiative between January 2011 and December 2015 were identified through targeted online searches and by contacting key personnel involved in the programme. The start date marks the beginning of the initiative in Cambridgeshire concurrent with Wave 1 of the national CYP-IAPT programme. A total of 609 published and unpublished documents were identified including reports, technical documents, web content, briefings, and meeting minutes. Of these, 334 were excluded because they were duplicates or outside of the date range. A total of 275 documents were included in the analysis (Table 1). Documents were categorised in Excel by year of publication, level of document (National, Collaborative and Partnership/Trust), issuer and document type.

Table 1 Summary of data sources

Interviews

We conducted 20 in-depth semi-structured interviews with professionals from three CAMHS teams within the Cambridgeshire and Peterborough Foundation Trust. Interviews were carried out at two time points: early implementation stage (T1) and full implementation stage (T2). Participants at both time points were purposively sampled to ensure a range of stakeholders would be represented (implementation team, managers and frontline staff). Table 1 details the number of interviews with each type of stakeholder. Participants’ years in service ranged between 3 and 16 years (mean 10 years) and no participants dropped out of the study.

Interview schedules were informed by the NPT framework [20,21,22] and specifically tailored for each stakeholder group (e.g. additional file 2). Participants were invited for interview by email, and a snowball sampling approach was used to recruit additional participants. The majority of interviews were conducted face to face by one researcher (A-MB) in a clinical setting; three were telephone interviews. Participants were given an information sheet prior to interview and taken through the consent process. Interviews lasted about an hour, were audio recorded, transcribed verbatim and anonymised. Research and Development approval was granted from the Cambridgeshire and Peterborough Foundation Trust.

Analysis

Documents

Documents were imported to NVivo V11 and the framework approach applied [26]. An initial framework was developed by the research team drawing on the research questions and key CYP-IAPT principles (accessibility, participation, awareness, EBP, accountability). Documents were systematically indexed and coded inductively and deductively by an experienced coder (MV). Framework matrices were produced by another member of the team (A-MB) which further validated the first round of coding and enabled the organisation, reduction and interpretation of the data along a 5-year timeline.

Interviews

Thematic analysis [27] of the interview data began with three members of the research team (A-MB, AH, EH) independently reading and becoming familiar with a subset of transcripts and identifying preliminary themes. The team met to review the inductive codes and to develop a working codeframe. All transcripts were coded in NVivo V12 by an experienced qualitative researcher (A-MB) and reviewed by the team to finalise overarching themes. We drew on NPT which is widely used to explore the processes by which an intervention becomes embedded into everyday practice [20,21,22]. Themes were mapped onto the four main NPT components: coherence (making sense of the intervention), cognitive participation (buy-in/investment), collective action (work done) and reflexive monitoring (appraisal of the benefits of the intervention).

Additionally, we drew on the CFIR [23] to identify the barriers and facilitators to implementation [28], [29]. We mapped inductive and deductive themes from combined interview and document data to the CFIR constructs. NPT and CFIR are both well-operationalised and have been used in conjunction to explore the implementation of QI initiatives [30].

Results

We present our findings in four sections: (1) ‘CYP-IAPT implementation guidance’, (2) ‘Local implementation process’, (3) ‘Local stakeholders’ implementation experiences’ and (4) ‘Barriers and facilitators to implementation’.

CYP-IAPT implementation guidance

Analysis of national and collaborative documents told the story of a programme that was evolving over time. Initially, there seemed to be a lack of clarity in the documents regarding the articulation of the core components. For example, an early document listed only four CYP-IAPT principles [31]; in another key document, the principles [32] were not clearly defined; and in a later document [33], six principles were listed. Across the national and collaborative documents, not all principles received equal attention. Whilst 62% of the thematic coding referred to EBP or ROMs, there were fewer references to the principles of participation (16%), access (4%) and awareness (1%).

Programme developers did not provide an implementation plan, and over the 5-year period, there was little practical guidance in the documents as to how to achieve the aims of the programme. However, this was addressed by a document published, more than 4 years into the programme, which gave an overview of practical steps to embedding the specific principles into practice [34]. The document followed the six stages of implementation proposed by Fixsen et al. [35]. It was the only comprehensive material relating to implementation that we could find.

The lack of specific guidance relating to implementation strategies seems to reflect a deliberate attempt by developers to avoid ‘implementing from the top’ [36]. Instead, Learning Collaboratives in five regions were established as the key implementation mechanism for QI [32], based on the assumption that local areas could learn from their neighbours with respect to moving key aspects of the programme into practice. There was an expectation that partnerships would share experience and competencies of implementation within and across Collaboratives. Additionally, it was envisaged that CYP-IAPT trainees would cascade EBP to colleagues in their service and act as ‘change agents’, thus accelerating the transformation and facilitating whole service culture change. However, this champion role was not formally specified and no guidance was given how to cascade knowledge or influence colleagues’ practices.

Local implementation process

In the early exploration phase, local activity centred on reaching consensus about the programme’s aims and objectives. An Implementation Lead was appointed to facilitate communication between the local setting and the Collaborative, and a Steering Board established to address the infrastructure needed to capture data and the mechanisms needed to release staff for training. A CYP-IAPT project group was formed to oversee the day-to-day delivery of the programme.

Echoing the emphasis of the national documents, much of the initial local implementation activity focused on embedding the use of EBP and ROMs. A Data Manager worked with teams and individuals to ensure that the rationale for data collection was understood, and data were collected and used to inform practice. At full operation, activities for implementing EBP included interviewing staff for training, providing training on supervision and using ROMs in everyday practice. A high level of local activity increased children and young people’s engagement in and feedback about services, including the set-up of a user participation group. A Participation Coordinator was appointed and many activities were undertaken to make clinic environments less intimidating and young people were consulted in relation to patient facing information.

The analysis of documents across a 5-year period identified a number of challenges to longer term sustainability of the programme. The national vision of sharing data across the partnership did not fit with the local governance structures and by the second year of full implementation, the attendance at the Steering Board had waned. Importantly, there were ongoing difficulties with inputting clinical outcomes data into the information technology (IT) system which led to clinicians not using it and created a considerable time burden for the Data Manager who needed to input them manually. Three years into the project, as several people fulfilling key roles left, meetings and in-house training stopped and the Steering Board was disbanded.

Local stakeholders’ implementation experiences

The interview data shows how the programme was understood and experienced by CAMHS professionals. The key themes were mapped onto the four main NPT constructs (coherence, cognitive participation, collective action and reflexive monitoring) (Table 2). Within each of the constructs, we describe how participants’ perceptions, attitudes and experiences changed over the course of implementation. Example quotes are labelled by interview time point (T1, T2) and type of participant. There was an alignment in the views of participants across the three CAMHS teams so findings were merged.

Table 2 NPT construct

Theme 1: Lack of clarity about the CYP-IAPT programme (coherence)

At the early implementation stage, there was a lack of clarity about the scope and aims of the CYP-IAPT programme. Many staff found it difficult to differentiate the programme from other service transformation models or initiatives with similar names. Participants described how the programme was introduced at a time when the local service was in a state of flux and ‘upheaval’, and staff with many years of service expressed apathy and ‘change fatigue’ generated by multiple service change initiatives. At the time, concerns related to the data from outcome measures being used to appraise staff performance. Criticisms related to the lack of information about how CYP-IAPT could improve the service or what their role would be in delivering it.

I felt we weren’t at all well informed on it. We weren’t even told that was a task that we were meant to be doing until…I just learned a bit on the job. (T2, Frontline staff)

A common misconception early on was that CYP-IAPT would be a completely new service delivered by a separate specialist team rather than a system-wide service improvement, a view reinforced by the fact that only a limited number of clinicians were sent on the training courses.

Maybe there was a bit of a perception that it was previously a separate service rather than it being a general transformation and what everybody should be doing. (T2, Implementation Team)

In the follow-up interviews, there was a strong consensus that the core principles in the CYP-IAPT model (awareness, accessible, accountability, EBP, participation) had been given different levels of prominence during programme implementation.

…they [principles] haven’t all had the same impact as each other, haven’t had an equal impact. (T2, Service Manager)

The principle of accountability (i.e. embedding ROMs into everyday practice) was seen to be the defining principle of the programme, which was closely interconnected with the principle of EBP and the participation of young people in planning and reviewing their treatment (see Fig. 1). However, the engagement of young people in the design of services was seen as a separate and unrelated activity from the overarching principle of participation. Meanwhile, as illustrated in Fig. 1, there was a general perception that the principles of access and awareness were not part of the core programme.

Fig. 1
figure 1

Depiction of participants’ perceived view of the CYP-IAPT principles at full implementation

Theme 2: Variable levels of stakeholder investment in CYP-IAPT (cognitive participation)

A prominent feature of participants’ accounts was the variable levels of investment from key stakeholders which influenced local implementation efforts. Members of the implementation team had been highly invested in the programme and were credited for driving forward implementation plans. In particular, two champions (Data Manager and Participation Coordinator) were instrumental in embedding ROMs and engaging young people in the re-design of services. At that time, children and young people were more visible and their feedback gave rise to tangible improvements to the service.

…we had a very enthusiastic young service users’ Coordinator, and [name] was great and provided people to join interview panels. (T2, Frontline staff)

Variable levels of investment among frontline staff was largely attributed to the practitioner’s background. For some therapists, the principles ‘didn’t really fit with their philosophy’ and consequently they did not prioritise using them in practice. The general perception was that Psychologists were more invested in CYP-IAPT since the principles were more aligned with their training and skill sets.

…we were all clinicians. We all knew what the problems were and the language of CYP-IAPT in many ways spoke to us. (T2, Implementation Team)

Participants reported that a lack of buy-in from senior management constrained the local team’s abilities to initiate the organisational changes needed to deliver the project. Indeed, later stage interviews confirmed that senior managers stopped attending Steering Board meetings altogether where their input and authority was needed to make crucial decisions about funding and IT infrastructure. This contributed to a loss of momentum and gradually, CYP-IAPT disappeared from the agenda at local clinic meetings and the steering board eventually disbanded.

I think the steering boards, I know they were put on hold for a while because we couldn’t get attendance from senior management level to actually run them so we had to put them on hold really, which was really frustrating. (T2, Frontline staff)

Theme 3: Work and resources needed to implement CYP-IAPT (collective work)

A key part of the initiative was a training programme to enhance staff skills and knowledge to deliver evidence-based therapies. In the early implementation phase, the first cohort of trainees were highly critical of the training, reporting that courses had been rolled out in haste and were thus poorly prepared and disorganised.

My impression of the training process was that we were very much a guinea-pig year… I actually had a really poor experience of the training (T1, Frontline staff)

Perceptions of training improved over time with later trainees being more positive about the content and delivery and some reporting an increase in their confidence to deliver evidence-based treatments. Staff valued being given protected time to attend courses and the opportunity to learn from, and network with, practitioners from other teams.

I think what CYP-IAPT has done for me personally and professionally it’s given me lots of links and network opportunities with other services so that I can sort of talk to people about their practice and what they do in their teams and take that back to my team (T2, Frontline staff)

Although funding was provided to release staff for training, this created challenges for managers who were often required to fill training places at short notice, leaving the service understaffed in a climate of high service demand.

So you’d get an email to say that there was training and that did cause problems with staffing because we were pretty well down to the bones at the time. I remember we did grumble about that… I think sometimes we felt that our voices weren’t being heard. (T2, Service Manager)

A consistent theme was that the technological infrastructure to support the delivery of the programme was inadequate. Participants faced multiple technical problems collecting patient outcome data, and it took valuable time to carry out simple tasks. For example, they were unable to produce graphs which young people valued to visualise their progress and consequently many practitioners resorted to plotting graphs by hand. Another major obstacle was that the IT system was not aligned with other systems and this prohibited data from being automatically migrated to the national database, a requirement set for CYP-IAPT sites.

I mean this is just so typical of the NHS that things weren't joined up. So you have a separate data collecting system for CYP-IAPT to what was on offer for the rest of the service. And that's clunky. (T2, Implementation Team)

Attempts to resolve technical issues were counterproductive because the IT developers lacked an understanding of reporting clinical outcome data.

…we were still trying to convince the developer to embed the measures, and they really struggled with understanding clinical thresholds. So, they were producing graphs that were not really useful. (T2, Implementation Team)

A salient feature of participants’ accounts related to the collaborative work between different partnership organisations who were tasked with implementing the programme. Initially, it was seen as a positive to have a range of stakeholders ‘around the table’, but tensions arose due to conflicting organisational goals and priorities. Over time, it emerged that partner organisations (commissioners, statutory and third sector service providers) were not aligned in terms of their organisational structures, technical systems or working norms, and differences in the language used about mental health conditions and treatment were a barrier to effective communication. Despite the initially well-attended meetings, partner organisations were often represented by different people, which slowed progress and diffused responsibility.

…too many people were invited to the meetings, nobody felt they were a key part of it. Nobody kind of owned it, if you know what I mean, it was too dispersed. (T2, Frontline Staff)

Theme 4: Evaluating and embedding CYP-IAPT (reflexive monitoring)

Interviews explored whether the CYP-IAPT principles had become embedded into routine practice. An accepted view was that the principle of access had not been implemented locally due to concerns that ‘we would be flooded’ and it would overwhelm an already overstretched service. Nonetheless, there was a strong consensus that during the project, outcome measures had become embedded and were the ‘normal day to day way of doing things’ and participation activities had successfully placed young people at the forefront of shaping services. Although some staff had initially been sceptical about involving young people in service design, this view altered over time and staff later commented that the service increasingly recognised the value of service user feedback.

…it just really did feel at the time that young people were much more visible, not just a patient but a people who helped to shape what we were doing. (T2, Implementation Team)

The momentum of CYP-IAPT was difficult to sustain, and as noted above, gradually the programme stopped being discussed at team meetings. This was compounded by a high staff turnover; new staff joining the service were less aware of the principles, and consequently, some of the changes in practice began to disappear. Participants reported a lack of feedback from the national team to appraise whether local efforts had fulfilled the vision of the programme by improving services, and there were no clear metrics offered in order that teams could undertake this type of audit for themselves.

I think people would be quite interested to know where it is, six years on… there are always new agendas and priorities being set, some have a shelf-life and finish without anyone ever saying that they’ve finished…I think there’s a general sense about CYP-IAPT, that maybe that’s what might have happened. (T2, Manager)

Barriers and facilitators to implementation

Using the CFIR [23] as a guide, we identified a number of factors from the combined document and interview data that influenced the implementation and sustainability of CYP-IAPT (summarised in Table 3) and described at strategic, service and individual levels.

Table 3 Barriers and facilitators to implementation of CYP-IAPT

National funding was a key facilitator for implementing CYP IAPT; however, the funding was time limited and this meant backfill, training and dedicated staff roles could not continue. At a strategic level, there was no implementation strategy and therefore a lack of readiness for implementation; the training course was rolled out before it was ready, and staff had restricted access to information. Moreover, a lack of measurable objectives or markers of success relating to the programme made it difficult for local staff to adequately evaluate the initiative’s progress. It was also unclear how the data was used, if at all, to drive service improvements.

At a service level, the internal implementation leaders facilitated local strategies and activities and provided valuable links between CAMHS teams and other partnership organisations. However, a lack of engagement from senior management hindered local implementation efforts and access to resources. Specifically, lack of investment in the IT system was a restriction. Limited capacity to record and report outcome data reduced the quality of service evaluation.

A high turnover of staff led to a loss of skills and know-how. A lack of ongoing training meant new staff were not formally introduced to CYP-IAPT. This curtailed the extent to which the CYP-IAPT principles remained embedded in the service over time. Additionally, competing priorities between the partnership organisations were barriers to effective co-operation.

The key facilitator to CYP IAPT implementation was highly invested staff. Many held the view that the CYP-IAPT model provided a relative advantage over previous ways of working, although this was highly dependent on practitioners’ norms and values. Dedicated staff played an important role in championing ROMs and young people’s participation in service design. This influence was lost, however, when these members of staff moved on. The main challenge evidenced at an individual level, particularly in relation to embedding EBP, ROMs and participation, was the lack of adequate technical infrastructure and support.

Discussion

Moving evidence-based practice into healthcare settings is a complex and challenging process [37,38,39]. The CYP-IAPT programme was introduced to change practices and embed a new way of working in CAMHS. Whilst the programme has been widely praised and integrated into subsequent mental health policy [40], limited evaluation at national and local levels means its success in driving service improvement and ultimately in improving mental health outcomes is unclear, and its value debated [36, 41, 42]. The purpose of this study was to explore how CYP-IAPT was delivered and experienced over time by local teams in Cambridgeshire including barriers and facilitators to implementation, with a view to informing decisions about how transformation efforts are sustained in mental health settings and how future QI initiatives should be delivered to maximise success. Our results, discussed in detail below, indicate that a lack of preparedness for implementation at national and local levels, a lack of senior management buy-in and difficulties with establishing an adequate IT infrastructure impacted on perceptions of acceptability, feasibility and sustainability. Moreover, differential emphasis by developers of the clinical principles comprising the programme contributed to a lack of fidelity, whereby only three of the five principles were implemented in practice.

Lack of preparation for implementation at all levels

A fundamental problem was the lack of operationalisation of the implementation model by developers (see below), coupled with poor preparation and organisational readiness for change at a national level. As set out in numerous implementation frameworks [18, 35], these are all essential prerequisites for successful QI implementation. A sense that the programme was not quite ready for delivery emerged from both the document and interview analyses. The training programme was rolled out in haste and was not fully prepared for the first cohort of trainees (which included staff from Cambridgeshire), even though this could have been an effective way to establish engagement and a consistent understanding by staff. At the partnership level, the lack of senior management buy-in highlights the importance of properly laying the foundations to convince stakeholders that the programme will provide an effective solution to an existing problem [43]. In the preparation stage, there should be a ‘pre-emptive consideration’ of which stakeholder groups will (or will not) buy-in to the QI programme, so that conflicts can be resolved early on and align the needs and priorities of clinical staff and managers [44]. Furthermore, the poor IT infrastructure in Cambridgeshire was a major barrier to implementation and may be reflective of a bigger issue relating to a lack of early preparation at a national level to implement the programme. For instance, there appeared to be no overarching technical strategy to ensure systems could adequately collect data for effective monitoring of the programme.

Proper preparation requires investment in data collection and monitoring systems, yet the local system was not fit for purpose and thus created an extra burden for staff requiring them to put in high levels of effort to collect, analyse and report data. This is particularly important given the requirement that CAMHS submit outcome data to the NHS Digital Mental Health Services Data Set [45]. Currently, most clinicians use paper versions of outcome measures since organisational barriers to electronic collection, analysis and reporting still remain today.

The programme was not clearly defined or understood

Our findings showed there was lack of clarity among local stakeholders regarding the scope of the programme, which is a crucial element for achieving successful implementation [46]. There was consensus that the programme was predominantly about embedding outcome measures and EBP into practice. This was reflected through national programme documents, which revealed that the principles were sometimes poorly articulated and to some extent differentially emphasised over time. This finding is resonant with other studies that have shown behaviour change interventions to be often poorly described [47] and therefore poorly understood [48]. A well-operationalised programme requires a clear description of the core components [49]. Research suggests there is a need to fully and precisely describe implementation strategies to ensure ‘reproducibility’ of the intervention and its effects [49]. Whilst the morphing and flexing of the programme appears to be a deliberate strategy to provide a flexible and phased approach to QI [36], it could be difficult to draw a boundary around what the programme actually was at any given time, and by extension difficult for partnerships to determine their progress in the context of poorly specified goals. This was addressed 3 years into the delivery of the programme via the publication of a document that could be used to guide transformation efforts [50].

Lack of implementation guidance

There was a lack of practical and evidence-based implementation guidance to inform local teams as to how to embed the principles with everyday practice. The practice of providing implementers with an architecture (a plan of what needs to be transformed) rather than a blueprint (detailed plan of what and how) for transformation is consistent with a place-based approach to implementation, whereby differentiated models of service delivery will be required in different places in order to achieve equitable outcomes [51,52,53,54]. This model allows implementers the flexibility to use whatever strategies are locally appropriate, in order to achieve a stated goal, and avoids a top-down prescriptive approach which can undermine buy-in of local partners [55]. However, this approach assumes that those implementing the programme have knowledge of the range of effective strategies that could be used to implement change and the ways in which these can be tailored to local context. In Cambridgeshire, the local implementation team devised their own strategies which points to a lack of effectiveness of the QIC approach, which did not provide the local team with guidance and support as anticipated [5, 6, 56]. To some extent, this may have been because Cambridgeshire was part of the early pilot, but it highlights the need for early adopter sites to receive additional support and clear guidance regarding implementation strategies.

Variable investment from practitioners

Low levels of programme coherence can lead to low cognitive participation and collective action [22] and change fatigue on the back of multiple transformations [57,58,59,60]. It was clear that there was high investment from practitioners whose therapeutic backgrounds (usually Psychologists) were perceived to be aligned with the ethos of CYP-IAPT; the previous internal audit conducted by programme developers found that embedding EBT was ‘easier for some disciplines than others’ [15]. Studies have shown that the quality of implementation can be enhanced when stakeholders value an intervention and perceive that changes will improve current practice [39, 61]. However, this runs the risk, as demonstrated here, of alienating some professionals who do not feel their therapeutic background aligns with the programme and is exacerbated when a training programme is only available for a subset of professionals. This can have the unintended consequences of some staff turning against a QI programme and creating a ‘them and us’ culture within a team [43, 62].

The importance of key individuals and relationships

Consistent with the previous internal audit [15], our findings highlighted that the local implementation was facilitated by key individuals or champions. However, we found that their efforts were hampered by the lack of engagement from senior management, which impeded the mobilisation of essential resources. Progress stalled when key people left, and we found evidence that the CYP-IAPT principles were beginning to ‘wash out’ without continual focus on sustaining the programme. Collective focus and engagement of individuals at all levels is necessary for transformational changes to be sustained, as is succession planning to ensure turnover does not undermine progress [63, 64].

We also found that collaborative efforts between partnership organisations faced a number of challenges to implementation, which led to an over-reliance on individual members [9]. Although the collaboration of multiple agencies can facilitate implementation efforts, they also run the risk of problems such as ‘collaborative inertia’ [65], ‘free rider’ [66] and ‘social loafing’ effects [67]. Building relationships between different organisations requires time and effort, and maintaining focus can be challenging with shifting organisational priorities [43].

Strengths and limitations

We used extensive database and lateral searches to identify all the documents issued between 2011 and 2015, and although we are confident in our search strategy, it is possible that we have missed some relevant documents. The experiences of the participants who were interviewed for this study may not be representative of others working in CAMHS within Cambridgeshire or elsewhere in the UK. We did not gather the views of members of the national team for crosschecking. The retrospective nature of this study means that we were unable to observe implementation activities as they occurred and therefore we have relied on the recollection provided by stakeholders. However, the interviews at Time 1 provided some verification of Time 2 accounts. We explored implementation outcomes as defined by Proctor et al. [19], although we did not gather objective data regarding programme fidelity or costs. Further, we were not able to evaluate the extent to which the initiative changed outcomes for services or patients. Notwithstanding, these findings are valuable for enhancing our understanding of the local context in order to inform other local transformation initiatives in young people’s mental health.

Conclusions and implications for future transformation efforts

Despite efforts by local teams, we identified several barriers to implementation, which prevented the CYP-IAPT principles from being integrated into everyday practice. As indicated by quality improvement frameworks, before QI programmes are rolled out, time should be given to assessing readiness for change at both strategic and service levels and for preparing the organisational infrastructure to support the changes. This early stage of preparation is vital for laying the foundations necessary to engage stakeholders at all levels of the organisation. From the outset, there should be clarity about what the programme is and what it is trying to achieve, rather than a trickle of information over time with which it is difficult to keep up. The flexibility of a place-based approach to implementation is essential, but without training and support, it is unrealistic to expect local teams to have the requisite knowledge of evidence-based implementation strategies to do this. Early adopter sites, as Cambridgeshire was in this instance, are often in unchartered territory and may need extra support and training to undertake local implementation. A toolkit of implementation strategies along with support from an implementation expert could assist local teams to develop implementation strategies that are adapted to and reflective of local needs and priorities. The study findings are relevant for on-going national QI programmes directed at young people’s mental health interventions such as the mental health support teams into schools.

Availability of data and materials

Consent was obtained from participants for the use of anonymised quotes in order to support the findings in this manuscript.

Abbreviations

CAMHS:

Child and Adolescent Mental Health Services

CFIR:

Consolidated Framework for Implementation Research

CPFT:

Cambridge and Peterborough Foundation Trust

CYP-IAPT:

Children and Young People’s Improving Access to Psychological Therapies

EBP:

Evidence-based practice

EBT:

Evidence-based treatment

IT:

Information Technology

NHS:

National Health Service

NPT:

Normalisation Process Theory

QIC:

Quality Improvement Collaborative

ROMs:

Routine Outcome Measures

References

  1. Department of Health. Talking therapies: a four-year plan of action. A supporting document to No health without mental health: a cross- government mental health outcomes strategy for people of all ages. 2011;:1–28.http://www.dh.gov.uk/en/Publicationsandstatistics/Publications/PublicationsPolicyAndGuidance/DH_123759.

  2. Shafran R, Fonagy P, Pugh K, et al. Transformation of mental health services for children and young people in England. In: Beidas, R. S., & Kendall PC, ed. Dissemination and implementation of evidence-based practices in child and adolescent mental health. Oxford University Press. 2014. 1–41.

  3. London and South East CYP-IAPT Learning Collaborative. What is CYP IAPT? https://cypiapt.com/cyp-iapt-principles/.

  4. Improving access to psychological therapies. Children and young people’s improving access to psychological therapies programme: national curriculum for core, cognitive behavioural therapy, parenting training (3–10 year olds), systemic family practice, interpersonal psychotherapy for adolescents, Sup. 2013;:1–145.http://www.cypiapt.org/docs/CYP_Curriculum_December_2013.pdf.

  5. Hulscher MEJL, Laurant MGH, Grol RPTM. Process evaluation on quality improvement interventions. Qual Saf Health Care. 2003;12:40–6. https://doi.org/10.1136/QHC.12.1.40.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  6. Schouten L, Hulscher M, van Everdingen J, et al. Evidence for the impact of quality improvement collaboratives: Systematic review. BMJ. 2008;336:1491–4. https://doi.org/10.1136/bmj.39570.749884.BE.

    Article  PubMed  PubMed Central  Google Scholar 

  7. Newton PJ, Davidson PM, Halcomb EJ, et al. An introduction to the collaborative methodology and its potential use for the management of heart failure. J Cardiovasc Nurs. 2006;21:161–8. https://doi.org/10.1097/00005082-200605000-00002.

    Article  PubMed  Google Scholar 

  8. Power M, Tyrrell PJ, Rudd AG, et al. Did a quality improvement collaborative make stroke care better? A cluster randomized trial. Implement Sci. 2014;9:40. https://doi.org/10.1186/1748-5908-9-40.

    Article  PubMed  Google Scholar 

  9. Cunningham FC, Ranmuthugala G, Plumb J, et al. Health professional networks as a vector for improving healthcare quality and safety: a systematic review. BMJ Qual Saf. 2012;21:239–49. https://doi.org/10.1136/bmjqs-2011-000187.

    Article  PubMed  Google Scholar 

  10. Pinto A, Benn J, Burnett S, et al. Predictors of the perceived impact of a patient safety collaborative: an exploratory study. Int J Qual Heal Care. 2011;23:173–81. https://doi.org/10.1093/intqhc/mzq089.

    Article  Google Scholar 

  11. Versteeg MH, Laurant MGH, Franx GC, et al. Factors associated with the impact of quality improvement collaboratives in mental healthcare: An exploratory study. Implement Sci. 2012;7:3–13. https://doi.org/10.1186/1748-5908-7-1.

    Article  Google Scholar 

  12. de Silva D. Improvement collaboratives in health care: Evidence scan. 2014. https://doi.org/10.1521/aeap.2010.22.1.1.

  13. Parand A, Benn J, Burnett S, et al. Strategies for sustaining a quality improvement collaborative and its patient safety gains. Int J Qual Heal Care. 2012;24:380–90. https://doi.org/10.1093/intqhc/mzs030.

    Article  Google Scholar 

  14. Fonagy P, Pugh K, O’Herlihy A. The Children and Young People’s Improving Access to Psychological Therapies (CYP IAPT) programme in England. In: David Skuse HB and LD, ed. Child Psychology and Psychiatry: Frameworks for Clinical Training and Practice. Chichester, UK: : John Wiley & Sons, Ltd 2017. 429–35. doi:10.1002/9781119170235.ch48.

  15. Edbrooke-Childs J, Calderon A. Wolpert M, et al. Rapid Internal Audit: Children and Young People’s Improving Access to Psychological Therapies; 2015.

    Google Scholar 

  16. Aarons GA, Palinkas LA. Implementation of evidence-based practice in child welfare: service provider perspectives. Adm Policy Ment Heal Ment Heal Serv Res. 2007;34:411–9. https://doi.org/10.1007/s10488-007-0121-3.

    Article  Google Scholar 

  17. Dixon-Woods M, Professor R, Research HS, et al. Does quality improvement improve quality? https://www.repository.cam.ac.uk/bitstream/handle/1810/260182/Dixon-Woods_et_al-2016-Future_Hospital_Journal-AM.pdf?sequence = 1 ().

  18. Meyers DC, Durlak JA, Wandersman A. The quality implementation framework: a synthesis of critical steps in the implementation process. Am J Community Psychol. 2012;50:462–80. https://doi.org/10.1007/s10464-012-9522-x.

    Article  PubMed  Google Scholar 

  19. Proctor EK, Landsverk J, Aarons G, et al. Implementation research in mental health services: an emerging science with conceptual, methodological, and training challenges. Adm Policy Ment Heal Ment Heal Serv Res. 2009;36:24–34. https://doi.org/10.1007/s10488-008-0197-4.

    Article  Google Scholar 

  20. May C, Finch T. Implementing, embedding, and integrating practices: an outline of normalization process theory. Sociology. 2009;43:535–54. https://doi.org/10.1177/0038038509103208.

    Article  Google Scholar 

  21. McEvoy R, Ballini L, Maltoni S, et al. A qualitative systematic review of studies using the normalization process theory to research implementation processes. Implement Sci. 2014;9:2. https://doi.org/10.1186/1748-5908-9-2.

    Article  PubMed  PubMed Central  Google Scholar 

  22. Murray E, Treweek S, Pope C, et al. Normalisation process theory: a framework for developing, evaluating and implementing complex interventions. BMC Med. 2010;8:63. https://doi.org/10.1186/1741-7015-8-63.

    Article  PubMed  PubMed Central  Google Scholar 

  23. Damschroder LJ, Aron DC, Keith RE, et al. Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement Sci. 2009;4:50. https://doi.org/10.1186/1748-5908-4-50.

    Article  PubMed  PubMed Central  Google Scholar 

  24. Pinnock H, Barwick M, Carpenter CR, et al. Standards for reporting implementation studies ( StaRI ): explanation and elaboration document. Published Online First. 2017. https://doi.org/10.1136/bmjopen-2016-013318.

  25. Tong A, Sainsbury P, Craig J. Consolidated criteria for reporting qualitative research (COREQ): a 32-item checklist for interviews and focus groups. Int J Qual Heal Care. 2007;19:349–57. https://doi.org/10.1093/intqhc/mzm042.

    Article  Google Scholar 

  26. Richie J, Spencer L. Qualitative data analysis for applied policy research. Analysing qualitative data. In: Bryman A, Burgess B, editors. 1994.

  27. Braun V, Clarke V. Using thematic analysis in psychology. In: Qualitative Research in Psychology. 2006:77–101.

  28. Kirk MA, Kelley C, Yankey N, et al. A systematic review of the use of the Consolidated Framework for Implementation Research. Implement Sci. 2015;11:72. 10.1186/s13012-016-0437-z.

    Article  Google Scholar 

  29. Damschroder LJ, Lowery JC. Evaluation of a large-scale weight management program using the consolidated framework for implementation research (CFIR). Implement Sci. 2013;8:51. https://doi.org/10.1186/1748-5908-8-51.

    Article  PubMed  PubMed Central  Google Scholar 

  30. Connell LA, McMahon NE, Harris JE, et al. A formative evaluation of the implementation of an upper limb stroke rehabilitation intervention in clinical practice: a qualitative interview study. Implement Sci. 2014;9:90. https://doi.org/10.1186/s13012-014-0090-3.

    Article  PubMed  PubMed Central  Google Scholar 

  31. Ivings K, Barnes S. Children and young people’s project: getting ready for CYP IAPT; 2012.

    Google Scholar 

  32. Children and young people’s programme: key facts briefing. 2013. http://www.iapt.nhs.uk/silo/files/cyp-iapt-key-facts-july-2013-.pdf.

  33. London & South East CYP IAPT Learning Collaborative. CYPIAPT and Children & Young Peoples Mental Health Service Transformation: an essential guide for commissioners. 2015.

  34. Byrne N, Law D. A step-by-step guide to implementing CYP IAPT; 2016.

    Google Scholar 

  35. Fixsen D, Naoom S, Blase K, et al. Implementation research: a synthesis of the literature. 2005. https://nirn.fpg.unc.edu/sites/nirn.fpg.unc.edu/files/resources/NIRN-MonographFull-01-2005.pdf ().

    Google Scholar 

  36. Fonagy P, Clark DM. Update on the Improving Access to Psychological Therapies programme in England: Commentary on … Children and Young People’s Improving Access to Psychological Therapies. BJPsych Bull. 2015;39:248–51. 10.1192/pb.bp.115.052282.

    Article  PubMed  PubMed Central  Google Scholar 

  37. Backer TE. The failure of success: Challenges of disseminating effective substance abuse prevention programs. J Community Psychol 2000;28:363–73. doi:10.1002/(SICI)1520-6629(200005)28:3 < 363::AID-JCOP10 > 3.0.CO;2-T.

  38. Bond GR, Drake RE, McHugo GJ, et al. Strategies for improving fidelity in the national evidence-based practices project. Res Soc Work Pract. 2009;19:569–81. https://doi.org/10.1177/1049731509335531.

    Article  Google Scholar 

  39. Weiner BJ. A theory of organizational readiness for change. Implement Sci. 2009;4:67. https://doi.org/10.1186/1748-5908-4-67.

    Article  PubMed  PubMed Central  Google Scholar 

  40. NHS England. Future in mind. Promoting, protecting and improving our children and young people’s mental health and wellbeing. London, UK: 2015.

  41. Timimi S. Children and Young People’s Improving Access to Psychological Therapies: inspiring innovation or more of the same? BJPsych Bull. 2015;39:57–60. https://doi.org/10.1192/pb.bp.114.047118.

    Article  PubMed  PubMed Central  Google Scholar 

  42. Swaile CJ. Fair criticism also needs to be based on evidence. BJPsych Bull. 2015;39:261–2. https://doi.org/10.1192/pb.39.5.261a.

    Article  PubMed  PubMed Central  Google Scholar 

  43. Dixon-Woods M, McNicol S, Martin G. Ten challenges in improving quality in healthcare: lessons from the Health Foundation’s programme evaluations and relevant literature. BMJ Qual Saf. 2012;21:876–84. https://doi.org/10.1136/bmjqs-2011-000760.

    Article  PubMed  PubMed Central  Google Scholar 

  44. Pannick S, Sevdalis N, Athanasiou T. Beyond clinical engagement: a pragmatic model for quality improvement interventions, aligning clinical and managerial priorities. BMJ Qual Saf. 2016;25:716–25. https://doi.org/10.1136/bmjqs-2015-004453.

    Article  PubMed  Google Scholar 

  45. Mental Health Services Data Set - NHS Digital. https://digital.nhs.uk/data-and-information/data-collections-and-data-sets/data-sets/mental-health-services-data-set (accessed 20 Apr 2020).

  46. Sutton E, Herbert G, Burden S, et al. Using the Normalization Process Theory to qualitatively explore sense-making in implementation of the Enhanced Recovery After Surgery programme: It’s not rocket science&quot. PLoS One. 2018;13:e0195890. https://doi.org/10.1371/journal.pone.0195890.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  47. Michie S, Fixsen D, Grimshaw JM, et al. Specifying and reporting complex behaviour change interventions: the need for a scientific method. Implement Sci. 2009;4:40. https://doi.org/10.1186/1748-5908-4-40.

    Article  PubMed  PubMed Central  Google Scholar 

  48. Alexander JA, Hearld LR. Methods and metrics challenges of delivery-system research. Implement Sci. 2012;7:15. https://doi.org/10.1186/1748-5908-7-15.

    Article  PubMed  PubMed Central  Google Scholar 

  49. Proctor EK, Powell BJ, McMillen JC. Implementation strategies: recommendations for specifying and reporting. Implement Sci. 2013;8:139. https://doi.org/10.1186/1748-5908-8-139.

    Article  PubMed  PubMed Central  Google Scholar 

  50. CYP IAPT. CYP IAPT Principles in CAMH services values and standards: delivering with and delivering well. Published Online First: 2014.https://www.england.nhs.uk/wp-content/uploads/2014/12/delvr-with-delvrng-well.pdf.

  51. Mulley A, Coulter A, Wolpert M, et al. New approaches to measurement and management for high integrity health systems. BMJ. 2017;356:j1401. https://doi.org/10.1136/bmj.j1401.

    Article  PubMed  Google Scholar 

  52. Raballo A, Poletti M, McGorry P. Architecture of change: rethinking child and adolescent mental health. Lancet Psychiatry. 2017;4:656–8. https://doi.org/10.1016/S2215-0366(17)30280-8.

    Article  PubMed  Google Scholar 

  53. Ham C, Alderwick H. Place-based systems of care: A way forward for the NHS in England. King’s Fund. 2015.

  54. Stiris PT. A consensus on the improvement of community and primary care services for children , adolescents and their families in Europe. 2016;:1–12.

  55. Dixon-Woods M, Leslie M, Tarrant C, et al. Explaining Matching Michigan: an ethnographic study of a patient safety program. Implement Sci. 2013;8:70. https://doi.org/10.1186/1748-5908-8-70.

    Article  PubMed  PubMed Central  Google Scholar 

  56. Schouten LJ, Hoppener P, van den Brandt PA, et al. Completeness of cancer registration in Limburg, The Netherlands. Int J Epidemiol 1993;22:369–76.http://ovidsp.ovid.com/ovidweb.cgi?T = JS&PAGE = reference&D = med3&NEWS=N&AN = 8359950.

  57. Brown M, Cregan C. Organizational change cynicism: The role of employee involvement. Hum Resour Manage. 2008;47:667–86. https://doi.org/10.1002/hrm.20239.

    Article  Google Scholar 

  58. Buchanan DA, Fitzgerald L, Fitzgerald L. Improvement evaporation: why do successful changes decay? 2006;:54–72. doi:10.4324/9780203030653-12.

  59. Torppa CB, Smith KL. Organizational change management: a test of the effectiveness of a communication plan. Commun Res Reports. 2011;28:62–73. https://doi.org/10.1080/08824096.2011.541364.

    Article  Google Scholar 

  60. McMillan K, Perron A. Nurses Amidst Change. Policy, Polit Nurs Pract. 2013;14:26–32. https://doi.org/10.1177/1527154413481811.

    Article  Google Scholar 

  61. Domitrovich CE, Bradshaw CP, Poduska JM, et al. Maximizing the Implementation Quality of Evidence-Based Preventive Interventions in Schools: A Conceptual Framework. Adv Sch Ment Health Promot. 2008;1:6–28. https://doi.org/10.1080/1754730X.2008.9715730.

    Article  PubMed  PubMed Central  Google Scholar 

  62. The Health Foundation. Safer Patients Initiative Phase One: Mixed-method Evaluation of a Large-Scale Organisational Intervention to Improve Patient Safety in Four UK Hospitals. 2011. https://www.health.org.uk/sites/default/files/2018-10/safer_patients_initiative_phase_1.pdf.

  63. Best A, Greenhalgh T, Lewis S, et al. Large-system transformation in health care: A realist review. Milbank Q. 2012;90:421–56. https://doi.org/10.1111/j.1468-0009.2012.00670.x.

    Article  PubMed  PubMed Central  Google Scholar 

  64. Dougall D, Lewis M, Ross S. Transformational change in health and care: Reports from the field | The King’s Fund: King’s Fund Published Online First; 2018.https://www.kingsfund.org.uk/sites/default/files/2018-05/Transformational_change_Kings_Fund_May_2018_0.pdf%0A. https://www.kingsfund.org.uk/publications/transformational-change-health-care.

  65. Kalu KN. All That Glitters. Adm Soc. 2013;45:420–42. https://doi.org/10.1177/0095399712453828.

    Article  Google Scholar 

  66. Carter P, Ozieranski P, McNicol S, et al. How collaborative are quality improvement collaboratives: a qualitative study in stroke care. Implement Sci. 2014;9:32. https://doi.org/10.1186/1748-5908-9-32.

    Article  PubMed  PubMed Central  Google Scholar 

  67. Karau SJ, Williams KD. Social loafing: A meta-analytic review and theoretical integration. J Pers Soc Psychol. 1993;65(4):681.

    Article  Google Scholar 

Download references

Acknowledgements

We would like to thank all the participants who gave up their time to be interviewed for the study.

Funding

This is a summary of research funded by the National Institute of Health Research (NIHR) Applied Research Collaboration East of England (ARC EoE) programme.

The views expressed are those of the authors and not necessarily those of the NIHR, NHS or Department of Health and Social Care.

Author information

Authors and Affiliations

Authors

Contributions

EH and AH wrote the protocol. MV and A-MB were responsible for data management and document analysis. A-MB carried out the interviews. A-MB, EH and AH analysed the interview data and wrote the first draft of the manuscript. All authors contributed to and approved the final manuscript.

Corresponding author

Correspondence to Anne-Marie Burn.

Ethics declarations

Ethics approval and consent to participate

A favourable Research & Development approval was granted from Cambridgeshire and Peterborough Foundation Trust. All participants who were interviewed provided written consent.

Consent for publication

Not applicable.

Competing interests

AH and MV worked as part of the local CYP-IAPT team.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary information

Additional file 1.

COREQ checklist.

Additional file 2.

Interview schedule.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Burn, AM., Vainre, M., Humphrey, A. et al. Evaluating the CYP-IAPT transformation of child and adolescent mental health services in Cambridgeshire, UK: a qualitative implementation study. Implement Sci Commun 1, 89 (2020). https://doi.org/10.1186/s43058-020-00078-6

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s43058-020-00078-6

Keywords