Skip to main content
  • Research article
  • Open access
  • Published:

Process evaluation results of a cluster randomised controlled childhood obesity prevention trial: the WAVES study



Increasing prevalence of childhood obesity and its related consequences emphasises the importance of developing and evaluating interventions aimed at prevention. The importance of process evaluation in health intervention research is increasingly recognised, assessing implementation and participant response, and how these may relate to intervention success or failure. A comprehensive process evaluation was designed and undertaken for the West Midlands ActiVe lifestyle and healthy Eating in School children (WAVES) study that tested the effectiveness of an obesity prevention programme for children aged 6-7 years, delivered in 24 UK schools. The four intervention components were: additional daily school-time physical activity (PA); cooking workshops for children and parents; Villa Vitality (VV), a 6-week healthy lifestyle promotion programme run by a local football club; and signposting to local PA opportunities.


Data relating to six dimensions (Fidelity, Reach, Recruitment, Quality, Participant Responsiveness, Context) were collected via questionnaires, logbooks, direct observations, focus groups and interviews. Multiple data collection methods allowed for data triangulation and validation of methods, comparing research observations with teacher records. The 6-stage WAVES study model ((i) Data collection, (ii) Collation, (iii) Tabulation, (iv) Score allocation and discussion, (v) Consultation, (vi) Final score allocation) was developed to guide the collection, assimilation and analysis of process evaluation data. Two researchers independently allocated school scores on a 5-point Likert scale for each process evaluation dimension. Researchers then discussed school score allocations and reached a consensus. Schools were ranked by total score, and grouped to reflect low, medium or high intervention implementation.


The intervention was predominantly well-implemented and well-received by teachers, parents and children. The PA component was identified as the most challenging, VV the least. Median implementation score across schools was 56/75 (IQR, 51.0 - 60.8). Agreement between teacher logbooks and researcher observations was generally high, the main discrepancies occurred in session duration reporting where in some cases teachers’ estimations tended to be higher than researchers’.


The WAVES study model provides a rigorous and replicable approach to undertaking and analysing a multi-component process evaluation. Challenges to implementing school-based obesity prevention interventions have been identified which can be used to inform future trials.

Trial registration

ISRCTN97000586. 19 May 2010.

Peer Review reports


The rapid increase in childhood obesity over a relatively short time period, with its associated adverse health and social consequences, is a serious challenge to public health [1]. Interventions aimed at preventing the upward trend in obesity prevalence have been developed and evaluated, with varying success [2]. The importance of process evaluation in public health intervention research is increasingly recognised [3]. Assessing whether interventions are delivered as intended, and factors affecting implementation, allows researchers to add context to the interpretation of intervention outcomes, and policy makers to optimise future implementation of interventions. The release of the Medical Research Council (MRC) framework - Process evaluation of complex interventions [3] - provides much-needed guidance for a structured approach to undertaking evaluations of health-related interventions, by considering the key elements and how they interrelate. To move beyond simple documentation of intervention delivery and inform implementation and mechanisms of impact, a clear understanding is needed of how the intervention is delivered, the level of implementation and the context within which it is delivered [4].

Whilst reporting of process evaluation in health research is improving, a lack of uniformity in approach remains [5, 6]. Process measures have often focused on a limited number of dimensions, such as reach, dose and fidelity [7], or just dose alone [8]. Similarly, data collection methods limited to questionnaires [9, 10], survey data [11], or qualitative data [5, 12], provide no opportunity for data triangulation. The MRC guidance has increased recognition of the importance of using multiple methods for process evaluation data collection [13,14,15] (usually a combination of observations, logbooks, questionnaires and qualitative methods [6]), however guidance for amalgamation and analysis of such data is lacking.

This paper describes the findings of a multi-method process evaluation undertaken in the WAVES study, including the approach for data synthesis. The WAVES study is a cluster randomised controlled trial testing the clinical and cost-effectiveness of an obesity prevention intervention in a sample of 54 primary schools in the West Midlands, United Kingdom (UK). The intervention is multifaceted, and designed to be delivered by teachers/external organisations, therefore the monitoring of implementation is essential. The intervention programme, designed for children aged 6–7 years (Year 2), aims to prevent obesity by targeting schools and families to encourage increased physical activity levels and improved dietary intake among their children. A prior systematic review of behavioural interventions to prevent childhood obesity reported small positive changes in target behaviours in school based programmes but stressed the need for longer term evaluation [2]. The intervention and its evaluation were informed by developmental and feasibility work (Birmingham Healthy Eating and Active lifestyle for Children study: BEACHeS [16]). Full details are presented in the protocol paper [17] and briefly summarised in Table 1.

Table 1 A summary of the WAVES study intervention components implemented with Year 2 (aged 6-7 years) in primary schools

The WAVES study process evaluation methods (described in detail elsewhere [18]) were based on frameworks and guidance by Linnen and Steckler (2002) [6], and Dane and Schneider (1998) [19]. Although developed prior to the MRC guidance document [3] on comparison the WAVES study process evaluation corresponds well. Contextual factors influencing both implementation and pathways to impact (through qualitative [20, 21], survey, and researcher experience data) were explored, and intervention implementation measured using a variety of direct and indirect methods.

The aim of this paper is twofold: 1) to demonstrate a replicable method of process evaluation data synthesis for use by other complex health intervention researchers, and 2) to present the results of the WAVES study process evaluation, demonstrating how the intervention was delivered and received.


Within the WAVES study, 26 schools across the West Midlands were randomised to receive the intervention (13 schools in 2011/12 and 13 in 2012/13), with a further 28 schools allocated as control. The intervention was intended for children in school year 2 (age 6-7 years). Process evaluation of the intervention involved six stages: (i) Data collection, (ii) Collation, (iii) Tabulation, (iv) Score allocation and discussion, (v) Consultation, and (vi) Final score allocation (see Fig. 1). As recommended by the MRC guidance [3], all data were analysed before trial outcomes were available to minimise the risk of bias in interpretation.

Fig. 1
figure 1

The ‘WAVES model’ for analysis of process evaluation data

Stage 1: Data collection

The development of data collection methods, the process evaluation dimensions of intervention delivery assessed, and the information collected to make the assessment (including a rationale for those used) have been detailed previously [18]. In brief: teachers of each class in each school were asked to complete logbooks (a daily logbook for PA collected once a term, a logbook to accompany each of the three CW’s and one for the whole VV programme) for the various intervention activities as well as a summary questionnaire. Trained researchers undertook direct observations of intervention delivery (every class was observed delivering both of their selected PA packages at least once each term, one of the three cooking workshops and at least one of the three Villa Vitality sessions). Interviews/focus groups were conducted with teachers, children and parents. Schools were purposively selected to take part in the qualitative aspect of the process evaluation. This sampling was used to ensure inclusion of parents, children and teachers from a range of schools, diverse in terms of geographical location, ethnic mix of pupils, school size, deprivation (indicated by free school meal entitlement), and the degree to which the intervention was implemented (as indicated by the other process evaluation methods). Informed consent was obtained from all participants involved in an interview or focus group. The researcher observation targets were: PA - once/term/class/activity, CWs - at least one of the three workshops during the intervention year; VV - one of the three sessions during the intervention year. Researchers also kept a diary of their experiences of intervention implementation throughout intervention delivery.

Use of multiple data collection methods ensured data were collected for each intervention component across all process evaluation dimensions, allowed for a cross check between data sources and enabled triangulation of the data to create an accurate and holistic picture of intervention implementation and response.

Stages 2 and 3: Collation and tabulation

The key information from each data source (observations, questionnaire, logbooks etc.) was collated by school separately for each included process evaluation dimension and tabulated for the three main intervention components (Physical Activity, Cooking Workshop, Villa Vitality). An example of this table for the PA component is provided as an Additional file 1: (Table S1). When data relating to the same session were available from more than one source a check for consistency of reporting was undertaken. Agreement between data sources was generally good but for one school where large discrepancies were identified, information collected directly by researchers took precedence. Qualitative data, obtained from a sample of schools, were also used to better understand how each intervention component was delivered and the influence of any contextual factors on delivery. The signposting component was not considered in this process as all schools confirmed distribution of the supplied information sheets to the children - thus delivery did not vary between schools. Participant response to the signposting was assessed through the qualitative data collection with children and parents [20], and teachers [21].

Stage 4: Intervention implementation score allocation and discussion

The tabulated evidence for each individual intervention component was used independently by two researchers (TLG and JLC) to allocate school-specific scores using a five point scale ranging from one (very low) to five (very high). A score was allocated to each of the five process evaluation dimensions: fidelity/adherence, reach/dose/exposure, recruitment, quality, and participant responsiveness. Information on context and programme differentiation influences all the above dimensions of process, and was considered in all scoring allocations. Schools could achieve a maximum score of 25 per intervention component (PA, CW, VV) and an overall score of 75.

To maximise consistency, score allocation was an iterative process. Four schools were selected at random and two researchers (JLC and TLG) independently allocated scores. The scores were then discussed to reach a consensus. This was repeated until all schools had been included in the process. Schools were then ranked by total score and reviewed by both researchers to check for anomalies. Data were not universally available from all sources for each activity at every school, however, due to the use of multiple data collection methods, information from at least one source was available for each activity at all schools. Where data were missing from one or more sources scores were allocated based on the information that was available. Where schools had more than one class, the classes were considered individually and then the scores were averaged across all classes to identify the overall school score. Examples of what counted as high or low implementation are presented in Table 2.

Table 2 Examples of high and low implementation scores

Stage 5: Consultation

Five members of the WAVES study research team with a working knowledge of intervention delivery were asked to independently score six randomly-selected schools following the same process as used by TLG and JLC. Their scores were compared with those obtained by TLG and JLC. All component and process evaluation specific scores (score range = 1 to 5) were within one point of those initially allocated. Finally, the five researchers were asked to review the school rankings for all scores sorted by total score to consider whether, based on their experience, the order of the schools was appropriate. TLG and JLC then revisited the score allocations, specifically reviewing where differences occurred between original scores and those allocated by the wider research team. Following discussion a consensus was reached, and the wider team had another chance to review scores for all schools and provide further comment. This was an iterative process which continued until all scores were agreed by the wider team.

Stage 6: Final score allocation

To define three levels of intervention implementation, schools were divided into tertiles based on their ranking; low, medium and high, represented by score ranges of 0-51, 52-58 and 59-75, respectively. These cut-offs were used to calculate proportionally that scores of 0-17, 18-19 and 20-25 reflected low, medium and high levels of implementation for each intervention component.


Twenty four schools with a total of 38 classes implemented the WAVES study intervention. Twenty six schools were randomised to the intervention arm of the trial, but unforeseen school circumstances meant that two were unable to deliver the intervention (although they agreed to participate in follow up child measurement data collection). An overview of data availability by class is provided in Table 3. Out of the 24 schools, two failed to return any paperwork (logbooks or questionnaires). Researcher observation targets (not outlined in the table) were all achieved.

Table 3 A summary of WAVES study process evaluation data availability by class (except where specified otherwise)

Data triangulation and cross checking

Where process evaluation dimension data were available for an intervention component from multiple sources, there was broad agreement in terms of fidelity achieved across the different sources. The main variation was observed in reported duration of PA and CW sessions, where higher values were obtained from teacher provided logbook information than researcher observation data (CW: average duration of 85 compared with 60 min, based on data from 19 schools; PA: mean difference of 1.3 min (SD = 5), based on 61 matched data points across 16 schools). With the exception of one school (where PA logbook data were discounted due to high levels of disagreement with its matched observation information), all other cross checking of data between these two sources (with acknowledgement of marginal reporting errors) suggested logbooks provide a generally fair estimate of schools’ activities The quality and volume of information obtained varied by school. However, the advantage of using several methods of data collection allowed sufficient information gathering to build a comprehensive picture of intervention implementation in each school. The positive findings from cross checking where several data sources were available gave confidence that in schools where limited data were available it was still likely to be a fair representation of intervention implementation in that school. The qualitative data collection provided an additional source of information to support that shown through other data collection methods.

Intervention implementation scores and levels of fidelity

Total intervention implementation scores ranged from 35 to 68 (out of a maximum of 75) with a median score of 56 (IQR: 51.0 – 60.8). The scores for each school are presented in Table 4, alongside the distribution of schools classified as achieving ‘high, medium or low’ for each main intervention component. Overall, there was little variance in implementation scores between classes within the same schools. When level of implementation fidelity achieved was explored by school characteristics (school size, free school meal eligibility, ethnic mix), no significant differences were observed.

Table 4 Fidelity scores for all schools included in the WAVES study intervention

Process evaluation by intervention component

The following section describes each intervention component separately to provide further detail on the implementation of the WAVES study intervention. Although findings from the teacher interviews [21] and the separate child and parent focus groups [20] are reported in detail elsewhere, this section includes key findings from the qualitative element of the process evaluation where relevant (discussed in the text and illustrated using direct quotes presented in Table 5).

Table 5 Illustrative quotes from the qualitative work undertaken as part of the process evaluation

Physical activity

Four schools (17%) were classed as having high implementation fidelity for the PA component delivery and 13 (54%) as having low fidelity. Based on logbook and teacher questionnaire data, delivery frequency for the PA component was available for 19 schools and showed that almost three quarters delivered this component on at least 4 days per week (just over half met the daily delivery target); 10% provided the PA component on one or less days per week. In terms of duration, a daily average of 17.5 min (12.5 min short of the 30 min target) was achieved on days when PA delivery took place. However when days on which delivery did not take place were taken into account the daily average fell to 12 min of additional activity.

Researcher observations identified better child skill levels, familiarity with activities, and a smoother transition back to classroom work for classes where the extra PA was timetabled a regular slot/slots within the school day compared with when teachers adopted a more ad hoc approach to delivery. Of the teachers interviewed, those who had embedded the component into their daily timetable (Table 5, quote 1) were more positive than those who fitted it in ‘as and when’ at varying points in the school day (Table 5, quote 2). The diary of researcher experiences also suggested that the latter group tended to have a generally less structured daily routine and more challenging child behaviours in the classroom. Teachers reported that fitting in daily PA, and in particular achieving the 30 min target was challenging (Table 5, quote 3).

Cooking workshops

High or medium implementation fidelity was achieved for the CW component by 15 schools (63%). In four of these schools delivery was undertaken by external delivery staff trained by the WAVES study research team; two schools advised that they had insufficient staff to run the sessions, and two provided delivery of the first workshop which was deemed unsatisfactory (incorrect nutrition messages being delivered) by the researchers observing, thus subsequent workshops were delivered by external staff. Fifteen schools delivered the three planned workshops, six delivered only two workshops, and three schools delivered just one.

There was generally good agreement between logbooks and observations for the CW element of the intervention. The main discrepancy was in the workshop duration. The matched records are described earlier, but when all records were considered a similar pattern was observed; the average workshop duration reported across the logbooks was 87 min, whereas from all observations (n = 31) the average duration was 58 min (range 35-100). Parents were invited to the cooking workshops in most schools, however researcher observation data show wide variation in the proportion of children with a parent attending (mean: 41% (SD 15%); range: 2 to 67%). Those parents who did attend were positive about the format and content of workshops (Table 5, quote 4-5), as were the children and teachers (Table 5, quotes 6-9). Some parents reported behaviour changes at home based on the messages delivered in the workshops [19].

Villa vitality

All schools that delivered the intervention completed Villa Vitality (VV). Although VV was mainly delivered by Aston Villa Football Club (AVFC) staff, implementation fidelity still varied by teacher involvement in the sessions, encouragement of children, and classroom delivery of the school project and weekly challenges. Although the majority of schools (n = 17, 71%) achieved a high level of implementation fidelity there were still five schools (21%) who only managed to achieve a low level. Reasons for this identified from qualitative work and researcher observations were primarily due to teachers’ lack of engagement with the homework and classroom activities set by the programme. For example, part of the programme asked teachers to set a weekly challenge for the children in the 6-week period between the two visits to AVFC. Some teachers created a board display and set up a star chart for the children to mark off their challenges whilst others failed to hand out the challenges to the children. There was positive feedback regarding VV from both child focus groups and teacher interviews, identifying it to be a highlight of the intervention programme (Table 5, quotes 10-11). Teachers also thought it helped to bring together the nutrition and PA aspects of the intervention (Table 5, quote 12).


Qualitative data revealed teachers to be unsure about the impact of the signposting sheets (Table 5, quotes 13 a&b) and parent recollections were vague or non-existent (Table 5, quote 14). A few parents discussed barriers preventing them from following the included advice (Table 5, quote 15) and there were no reports of behaviour changes made based on this element of the intervention.

Key factors found to influence intervention delivery and pathways to impact

The key contextual factors affecting implementation were: (i) internal and external pressure on schools to focus on academic attainment, resulting in teachers perceiving a lack of time to accommodate intervention components in the school day, (ii) teachers’ own attitudes and motivation to deliver the intervention (identified through the qualitative data and researcher observations/experiences) and (iii) the degree of existing infrastructure and support within the school for health promoting activities. Pathways to impact within families were influenced by level of parental engagement with the school, the consistency of messages from and degree of influence of the teachers, and the pre-existing knowledge and lifestyles of families.


The results of the WAVES study process evaluation provide detailed information on intervention implementation, and a replicable method for analysing process data from health intervention research. Inter-component differences in fidelity were evident, seemingly driven by required teacher workload and the enthusiasm and support from senior staff. We found inter-school variation in delivery of the WAVES study intervention programme, although overall there was good fidelity of implementation in most schools.

Recently, several extensive process evaluations which have used multiple methods for data collection, similar to the WAVES study, have been undertaken [13, 15, 22]. However, reporting tends to focus on the findings of each method (e.g. reporting questionnaire data or observation data) in isolation followed by an overview of what this meant for overall implementation. The findings of these studies provide useful information in helping future researchers learn from the experiences of intervention delivery; however the confined approach to data collection and synthesis limits interpretation. In this study, the triangulation and integration of data sources increases the validity of the findings. It enables a complete picture of implementation and participant response to be synthesised, and identifies variation between clusters. The generation of an overall implementation score also allows for intervention implementation to be considered in relation to the trial outcomes, in line with recommendations in the recent MRC process evaluation guidance [3].

The implementation findings specific to the WAVES study are also useful to help inform future intervention in the research field. Schools are often considered a key setting for the delivery of health interventions as they provide a teaching and learning environment alongside eating and PA opportunities [2, 23]. However in our study many teachers reported finding it challenging to deliver an intervention in addition to their teaching responsibilities. Individual teachers’ beliefs in the importance of the intervention’s overall objective (prevention of childhood obesity through the encouragement of healthy lifestyle behaviours) was found to have a positive impact on implementation fidelity, particularly when they perceived healthy behaviours as central to children’s development and learning.

The daily school-time delivery of PA was the most challenging intervention component for teachers despite the activity packages offered being easy to implement in the classroom setting, flexible to deliver, and teachers having a choice of packages. However, it was the component which placed the most burden on teachers, as it was a daily activity. Most schools achieved at least some additional physical activity, and it may have been different if the intervention was only focused on this one component rather than also having the additional activities schools were asked to incorporate. The findings support the need for leadership within schools to encourage regular inclusion of additional PA, particularly as there is evidence to suggest that moderate-vigorous physical activity (MVPA) may be positively associated with, or at least does not negatively impact, academic attainment [24,25,26].

The importance of PA, for health and the development of basic movement skills, warrants continued efforts to try to learn from experiences such as those of the WAVES study to help address difficulties in delivery and identify ways in which PA can be incorporated into the primary school day. In the UK, much media attention has been given to a recent initiative - ‘the daily mile’ [27], an intervention whereby all children attending school run outside for 15 min every day, a simple concept that is reportedly easy to deliver may be a more user friendly approach for the teachers compared to the options provided by the WAVES study intervention. Recommendations from the WAVES study experience would be to: (i) encourage teachers to understand the central importance of PA to child development, aiming to improve enthusiasm for delivery, (ii) enable schools and class teachers individually to identify the best way to ensure PA is routinely timetabled every day, (iii) allow teachers adequate time to consider their competing demands and plan delivery to suit their individual class needs, and (iv) provide training and support for teachers to help them feel confident with delivery.

Current national policy in the UK stipulates that schools must teach physical education but there is no guidance on the minimum amount of time that schools need to dedicate to it. Although headteachers see healthy lifestyles as an important part of development of the whole child [23, 28] it is hard for them to give such aspects of child development as much importance as academic achievement due to the present external pressures placed on schools. This is a similar finding to that reported in the results of the Active for Life Year 5 process evaluation – a key reason for teachers failing to adhere to intervention elements was pressure to focus on literacy, numeracy and academic attainment [15].

The CW and VV components of the intervention were relatively well received and delivered. However, as for the PA component, the overarching limiting factor in optimal delivery was time. Although VV achieved the most promising levels of implementation fidelity, it has significant cost implications. In addition, despite teachers being positive about the CWs and reporting that the materials and session plans made them easy to deliver, they indicated that, due to the logistics of organising the sessions, continued delivery in future years would require particularly motivated staff. Cooking skills have since been included in the National Curriculum for all UK schools [29] which is a positive step towards incorporating interventions such as this one. The signposting sheets were resource intensive to produce. This, together with no evidence of their impact on families, suggests that this element should not be included in future school based interventions.

In stakeholder consultations undertaken as part of the development work for the WAVES study intervention, family involvement through activities aiming to improve practical skills in addition to knowledge was identified as a priority [16]. Systematic review evidence also supports the importance of involving family members [2]. The WAVES study tried to involve families through school-specific signposting sheets, parental invitation to the cooking workshops and the healthy challenges element of VV. Unfortunately, the former had little or no impact, and although there was positive feedback regarding CWs from the parent focus groups, attendance rates were often low (mean parental attendance was 41%). However, pre-existing parent-school relationships heavily influenced the level of parental engagement, again highlighting the important contextual influences on intervention implementation. The level of involvement of parents with the VV healthy challenges was dependent on the teacher’s approach to delivery of the weekly challenges. Further research to determine how schools can better engage parents with health promotion initiatives would be valuable for both schools and intervention developers.

Limitations of the WAVES study process evaluation need to be considered. Process evaluation of a multifaceted intervention programme is inevitably a balance between comprehensive and detailed data collection and the resulting participant burden. The latter was a strong driving force during the development of our data collection tools, and in general completion rates of 60 to 70% were achieved, although lower rates were achieved for some items, especially PA teacher logbooks. However it is promising that: most returned logbooks were well completed; the cross check of data between matched logbook and observation time points revealed good consistency; and despite some short session durations, observation data indicated that most CWs covered key content and activities.

To ensure blinding of researchers to trial arm allocation, randomisation of schools was delayed until after baseline measurement completion resulting in very limited time (the last two weeks of the summer term) to introduce the class teacher to the intervention programme, a step we have previously highlighted as critical. This process was further hindered, as despite best efforts to involve class teachers as early as possible in the recruitment stage, it was clear in some schools that the first time they were aware of the expectation for them to undertake intervention delivery during the subsequent school year was at the introductory visit by the research team. Both of these would have resulted in insufficient planning and preparation time for teachers and are likely to have negatively impacted overall implementation fidelity of the programme. Another factor likely to have negatively influenced both quality of intervention delivery and process evaluation data return rates is that the intervention year in half the schools (2011-12) coincided with two events for which schools took on many additional activities (The Queen’s Diamond Jubilee and the London 2012 Olympics).

The possible impact that direct observation of teachers undertaking intervention activities may have had on quality of delivery must be acknowledged. The intention was to arrive at schools unannounced; however this approach was poorly received by schools and also resulted in wasted researcher time due to last minute rescheduling of planned activities (e.g. researchers arriving to find the children were at swimming lessons or school play practice). This meant that subsequent researcher visits were prearranged and as such prior teacher knowledge of session observation may have influenced implementation. Although it is important to acknowledge this as a potential limitation, in reality both the teachers’ and children’s proficiency with the routines provided a good indication of implementation consistency. The observation checklists were tested until inter-rate reliability was achieved, but by their nature the completion and rating of them is subjective.

In the current study it was appropriate to consider intervention implementation at school level as there was limited variation in implementation scores between classes at the same school. In future studies if there was a greater inter-class variation in implementation fidelity between classes at the same school, it may be important to consider implementation fidelity by class rather than by school to avoid the possible masking of such differences.

The use of qualitative data is time intensive both in collection and analysis. However, due to the nature of the WAVES study intervention it provided a useful insight into a school based obesity prevention programme, providing key recommendations for future delivery. It also supported the other methods of data collection and gave a clearer picture of intervention implementation in the schools in which interviews and focus groups were conducted.

Despite the limitations, the WAVES study process evaluation was comprehensive and provides a unique approach to working with process data. The methods allowed for data triangulation and cross checking of data sources. Drawing on multiple sources of evidence allowed for the generation of a score that can be used in analysis of the main trial outcomes. The approach to data analysis was rigorous and several steps were taken to try to minimise the effect of subjectivity in the scoring process. Researchers scored schools independently, and consensus was sought from the wider research team. The WAVES study model (Fig. 1) is replicable and could be applied to process evaluations from many different aspects of health intervention research. This paper reports on the analysis of process evaluation data, providing a level of detail which is rarely reported in the process evaluation literature [6]. Following the MRC recommendation for analysis of process data [3] we present data which meets recommendation by (i) providing information on fidelity, dose and reach for the intervention, (ii) detailing variation in implementation between schools (iii) using thematic analysis to analyse the qualitative data (iv) integrating both qualitative and quantitative data sources to provide an overall indicator of intervention implementation, and (v) completing all analyses before analysis of the main trials outcomes. The WAVES study was undertaken in the West Midlands, UK, − a region that is socioeconomically, ethnically and culturally diverse. The school selection process ensured an over-representation of schools with a higher proportion of South Asian or Black pupils by using a randomly ordered, weighted random sampling procedure from amongst 970 eligible state maintained schools. Randomisation of schools to the control or intervention arm used a statistical procedure to minimise inter-arm imbalance in relation to school size, free school meal eligibility (as an indicator of deprivation) and proportion of pupils of South Asian, Black and White ethnicities. Additionally, as reported earlier, schools from the intervention arm were purposively sampled for inclusion in the interviews/focus groups to ensure representation from a diverse range of schools. These processes helped to improve the generalisability of the findings across different UK locations and the findings from the intervention delivery should be useful to other researchers working in the field.


We have presented a unique, rigorous and replicable approach to the analysis of process evaluation data, demonstrating how it can provide insight into intervention implementation, allow for analysis of main trial results by implementation, and test assumptions about pathways to impact. The findings identified challenges that need to be addressed both in the design of future interventions and in the future direction of national policy to optimise their implementation.


  1. Ng M, Fleming T, Robinson M, Thomson B, Graetz N, Margono C, et al. Global, regional, and national prevalence of overweight and obesity in children and adults during 1980–2013: a systematic analysis for the global burden of disease study 2013. Lancet. 2014;384(9945):766–81.

    Article  PubMed  PubMed Central  Google Scholar 

  2. Waters E, de Silva SA, Hall B, Brown T, Campbell K, Gao Y, et al. Interventions for preventing obesity in children (review). Cochrane collaboration. 2011;12:1–212.

    Google Scholar 

  3. Moore GF, Audrey S, Barker M, Bond L, Bonell C, Hardeman W, et al. Process evaluation of complex interventions: Medical Research Council guidance. BMJ. 2015;350

  4. Baranowski T, Stables G. Process evaluations of the 5-a-day projects. Health Education and Behavior. 2000;27(2):157–66.

    Article  CAS  PubMed  Google Scholar 

  5. Grant A, Treweek S, Dreischulte T, Foy R, Guthrie B. Process evaluations for cluster-randomised trials of complex interventions: a proposed framework for design and reporting. Trials. 2013;14(1):15.

    Article  PubMed  PubMed Central  Google Scholar 

  6. Linnan LE, Steckler AE. Process evaluation for public health interventions and research. San Franciso: Jossey-Bass; 2002.

    Google Scholar 

  7. Robbins LB, Pfeiffer KA, Wesolek SM, Lo Y-J. Process evaluation for a school-based physical activity intervention for 6th-and 7th-grade boys: reach, dose, and fidelity. Evaluation and Program Planning. 2014;42:21–31.

    Article  PubMed  Google Scholar 

  8. Salmon J, Ball K, Crawford D, Booth M, Telford A, Hume C, et al. Reducing sedentary behaviour and increasing physical activity among 10-year-old children: overview and process evaluation of the ‘switch-Play’intervention. Health Promot Int. 2005;20(1):7–17.

    Article  PubMed  Google Scholar 

  9. Christian MS, Evans CE, Ransley JK, Greenwood DC, Thomas JD, Cade JE. Process evaluation of a cluster randomised controlled trial of a school-based fruit and vegetable intervention: project tomato. Public Health Nutr. 2012;15(03):459–65.

    Article  PubMed  Google Scholar 

  10. van Nassau F, Singh AS, Hoekstra T, van Mechelen W, Brug J, Chinapaw MJ. Implemented or not implemented? Process evaluation of the school-based obesity prevention program DOiT and associations with program effectiveness. Health Educ Res. 2016;31(2):220–33.

    Article  PubMed  Google Scholar 

  11. Pettigrew S, Borys JM, du Plessis HR, Walter L, Huang TT, Levi J, et al. Process evaluation outcomes from a global child obesity prevention intervention. BMC Public Health. 2014;14(1):1.

    Article  Google Scholar 

  12. Jogova M, Song JE-S, Campbell AC, Warbuton D, Warshawski T, Chanoine J-P. Process evaluation of the living green, healthy and thrifty (LiGHT) web-based child obesity management program: combining health promotion with ecology and economy. Can J Diabetes. 2013;37(2):72–81.

    Article  PubMed  Google Scholar 

  13. Sebire S, Edwards M, Kesten J, May T, Banfield K, Bird E, et al. Process evaluation of the Bristol girls dance project. BMC Public Health. 2016;16(1):1.

    Article  Google Scholar 

  14. Van den Branden S, Van den Broucke S, Leroy R, Declerck D, Hoppenbrouwers K. Evaluating the implementation fidelity of a multicomponent intervention for oral health promotion in preschool children. Prev Sci. 2015;16(1):1–10.

    Article  PubMed  Google Scholar 

  15. Campbell R, Rawlins E, Wells S, Kipping RR, Chittleborough CR, Peters TJ, et al. Intervention fidelity in a school-based diet and physical activity intervention in the UK: Active for life year 5. Int J Behav Nutr Phys Act. 2015;12(1):1.

    Article  Google Scholar 

  16. Pallan M, Parry J, Cheng KK, Adab P. Development of a childhood obesity prevention programme with a focus on UK south Asian communities. Prev Med. 2013;57(6):948–54.

    Article  PubMed  PubMed Central  Google Scholar 

  17. Adab P, Pallan MJ, Lancashire ER, Hemming K, Frew E, Griffin T, et al. A cluster-randomised controlled trial to assess the effectiveness and cost-effectiveness of a childhood obesity prevention programme delivered through schools, targeting 6–7 year old children: the WAVES study protocol. BMC Public Health. 2015;15(1):1.

    Article  Google Scholar 

  18. Griffin TL, Pallan MJ, Clarke JL, Lancashire ER, Lyon A, Parry JM, et al. Process evaluation design in a cluster randomised controlled childhood obesity prevention trial: the WAVES study. Int J Behav Nutr Phys Act 2014;11(1):1-12.

  19. Dane AV, Schneider BH. Program integrity in primary and early secondary prevention: are implementation effects out of control? Clin Psychol Rev. 1998;18(1):23–45.

    Article  CAS  PubMed  Google Scholar 

  20. Clarke JL, Griffin TL, Lancashire ER, Adab P, Parry JM, Pallan MJ. Parent and child perceptions of school-based obesity prevention in England: a qualitative study. BMC Public Health. 2015;15(1):1.

    Article  Google Scholar 

  21. Griffin TL, Clarke JL, Lancashire ER, Pallan MJ, Passmore S, Adab P. Teacher experiences of delivering an obesity prevention programme (the WAVES study intervention) in a primary school setting. Health Educ J. 2015;74(6):655–67.

    Article  Google Scholar 

  22. Gibson CA, Smith BK, DuBose KD, Greene JL, Bailey BW, Williams SL, et al. Physical activity across the curriculum: year one process evaluation results. Int J Behav Nutr Phys Act. 2008;5(1):36.

    Article  PubMed  PubMed Central  Google Scholar 

  23. Clarke JL, Pallan MJ, Lancashire ER, Adab P. Obesity prevention in English primary schools: headteacher perspectives. Health Promot Int. 2015;32(3):490–9.

    Google Scholar 

  24. Booth J, Leary S, Joinson C, Ness A, Tomporowski P, Boyle J, et al. Associations between objectively measured physical activity and academic attainment in adolescents from a UK cohort. Br J Sports Med. 2014;48:265–20.

    Article  CAS  PubMed  Google Scholar 

  25. Rasberry CN, Lee SM, Robin L, Laris B, Russell LA, Coyle KK, et al. The association between school-based physical activity, including physical education, and academic performance: a systematic review of the literature. Prev Med. 2011;52:S10–20.

    Article  PubMed  Google Scholar 

  26. Trudeau F, Shephard RJ. Physical education, school physical activity, school sports and academic performance. Int J Behav Nutr Phys Act. 2008;5(1):1.

    Article  Google Scholar 

  27. The Daily Mile. The Daily Mile Foundation. 2015. Accessed 14 Oct 2016.

  28. Howard-Drake EJ, Halliday V. Exploring primary school headteachers' perspectives on the barriers and facilitators of preventing childhood obesity. Journal of Public Health. 2015;38(1):44–52.

    Article  PubMed  Google Scholar 

  29. Department for Education. National curriculum. 2014. Accessed 14 Oct 2016.

Download references


With thanks to all teachers, parents and children involved in the WAVES study.

Trial steering committee: Peymané Adab, John Bennett, Kelvin Jordan (chair), Karla Hemming, Louise Longworth, Peter Whincup.

REC number: 10/H1202/69.

WAVES study trial investigators: University of Birmingham: Peymane Adab (Professor of Public Health and Chief Investigator), Tim Barrett (Professor of Paediatrics), KK Cheng (Professor of Public Health and Primary Care) Amanda Daley (Reader in Behavioural Medicine), Jonathan J Deeks (Professor of Biostatistics), Joan L Duda (Professor of Sport and Exercise Psychology), Emma Frew (Reader in Health Economics), Paramjit Gill (Clinical Reader in Primary Care Research), Karla Hemming (Senior Lecturer in Medical Statistics), Miranda J Pallan (Senior Clinical Lecturer in Public Health), Jayne Parry (Professor of Policy and Public Health); University of Cambridge, Cambridge MRC Epidemiology Unit / Norwegian School of Sport Sciences: Ulf Ekelund (Professor of Physical Activity Epidemiology and Public Health/Senior Investigator Scientist); University of Leeds: Janet E Cade (Professor of Nutritional Epidemiology and Public Health); The University of Edinburgh: Raj Bhopal (Bruce and John Usher Chair in Public Health); Birmingham Community Healthcare NHS Trust: Eleanor McGee (Public Health Nutrition Lead); Birmingham Services for Education: Sandra Passmore (Education Advisor).


This project was funded by the National Institute for Health Research (NIHR) Health Technology Assessment programme (project number 06/85/11). The views and opinions expressed therein are those of the authors and do not necessarily reflect those of the HTA programme, NIHR, NHS or the Department of Health.

Availability of data and materials

The datasets used and analysed during the current study available from the corresponding author on reasonable request.

Author information

Authors and Affiliations




All WAVES Trial Co-investigators contributed to the development of the design for the WAVES trial and all authors to the methodology for process evaluation. TLG, JLC, MJP, ERL and PA designed the detailed methods for data collection. TLG operationalised and oversaw the process evaluation. TLG and JLC led on the data analysis and developed the analysis methodology described in this manuscript. TLG wrote the first draft with substantial contribution from JLC, MJP, ERL and PA. All authors contributed to critical revisions of the paper. The final manuscript was read and approved by all authors. We confirm that the authors declare no conflicts of interest.

Corresponding authors

Correspondence to E. R. Lancashire or P. Adab.

Ethics declarations

Ethics approval and consent to participate

Ethical approval for the WAVES study was obtained from the National Research Ethics Service Committee West Midlands, The Black Country (10/H1202/69, 25th November 2010). Informed written consent was attained for all participants involved in an interview or focus group.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Additional file

Additional file 1: Table S1.

Example of WAVES study process evaluation data used to inform the scores achieved by each school for the physical activity intervention component. Example data collected from various sources to inform scoring for the physical activity component of the WAVES study intervention. (XLSX 13 kb)

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (, which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Griffin, T.L., Clarke, J.L., Lancashire, E.R. et al. Process evaluation results of a cluster randomised controlled childhood obesity prevention trial: the WAVES study. BMC Public Health 17, 681 (2017).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: