- Research article
- Open Access
- Open Peer Review
Implementation of the Good School Toolkit in Uganda: a quantitative process evaluation of a successful violence prevention program
© The Author(s). 2018
- Received: 3 March 2017
- Accepted: 13 April 2018
- Published: 9 May 2018
The Good School Toolkit, a complex behavioural intervention designed by Raising Voices a Ugandan NGO, reduced past week physical violence from school staff to primary students by an average of 42% in a recent randomised controlled trial. This process evaluation quantitatively examines what was implemented across the twenty-one intervention schools, variations in school prevalence of violence after the intervention, factors that influence exposure to the intervention and factors associated with students’ experience of physical violence from staff at study endline.
Implementation measures were captured prospectively in the twenty-one intervention schools over four school terms from 2012 to 2014 and Toolkit exposure captured in the student (n = 1921) and staff (n = 286) endline cross-sectional surveys in 2014. Implementation measures and the prevalence of violence are summarised across schools and are assessed for correlation using Spearman’s Rank Correlation Coefficient. Regression models are used to explore individual factors associated with Toolkit exposure and with physical violence at endline.
School prevalence of past week physical violence from staff against students ranged from 7% to 65% across schools at endline. Schools with higher mean levels of teacher Toolkit exposure had larger decreases in violence during the study. Students in schools categorised as implementing a ‘low’ number of program school-led activities reported less exposure to the Toolkit. Higher student Toolkit exposure was associated with decreased odds of experiencing physical violence from staff (OR: 0.76, 95%CI: 0.67-0.86, p-value< 0.001). Girls, students reporting poorer mental health and students in a lower grade were less exposed to the toolkit. After the intervention, and when adjusting for individual Toolkit exposure, some students remained at increased risk of experiencing violence from staff, including, girls, students reporting poorer mental health, students who experienced other violence and those reporting difficulty with self-care.
Our results suggest that increasing students and teachers exposure to the Good School Toolkit within schools has the potential to bring about further reductions in violence. Effectiveness of the Toolkit may be increased by further targeting and supporting teachers’ engagement with girls and students with mental health difficulties.
The trial is registered at clinicaltrials.gov, NCT01678846, August 24th 2012.
- Process evaluation
- Corporal punishment
In Uganda, physical punishment in schools has been banned since 1997, and became illegal in May 2016. Despite this, physical punishment persists as normal practice in primary schools. In one study conducted in 2012, over half of school children reported experiencing physical violence from staff in the last week and 8% sought treatment for injury from a healthcare provider . This high level of violence in schools is not unique to Uganda. Recent national prevalence studies have shown that 40% of 13-17 year olds report being punched, kicked or whipped by a teacher in the last week in Kenya and in Tanzania 50% reported experiencing physical violence from a teacher when they were under 18 year of age [2, 3].
The Good School Toolkit developed by Raising Voices, a Uganda-based Non Government Organisation (NGO), is one of the very few rigorously evaluated interventions designed to reduce physical violence from school staff to students. The Toolkit is a violence prevention behavioural intervention that aims to change school operational culture. We recently conducted a trial to assess effectiveness as part of the Good Schools Study. The trial results showed a 42% reduction in relative risk of students experiencing physical violence in the last week from staff (corresponding to an odds ratio: 0.40, 95% CI: 0.26 to 0.64, p < 0.001) . However, even with this highly effective intervention, 31% of students in the intervention schools had experienced physical violence from staff in the last week after intervention delivery. This may be due to variation in delivery of the intervention by Raising Voices, school-led Toolkit implementation by school staff, adoption of the intervention by schools, or due to the characteristics of schools or composition of students within schools.
Exploring reasons for variation in intervention impact in complex interventions such as the Good School Toolkit is important to inform intervention development, adaptations, program monitoring, cost-effective implementation and scale-up . Qualitative and quantitative methods for evaluation bring complementary insights into what and how interventions are delivered and received. Quantitative evaluation can not only describe what was implemented but also explore dose response and how delivery and reach may vary across contexts or participant characteristics . Quantitative evaluation can therefore add valuable insight into how implementation is associated with effectiveness, and can highlight inequalities to inform intervention future development. However, quantitative process evaluations measuring implementation are relatively rare, with a limited number of health related behaviour change interventions reporting on how implementation is related to outcomes [7–10].
Here we present a quantitative process evaluation of the Good School Toolkit intervention in Uganda that focuses on the Toolkit implementation and explaining variations in effectiveness across schools. In associated papers we present the study protocol , main study results , qualitative findings on pathways of change  and an economic evaluation of the Good School Toolkit .
Overview of main trial
The Good Schools Study consists of a cluster randomised controlled trial, a qualitative evaluation, an economic evaluation, and the process evaluation presented here. A cross-sectional baseline survey was conducted in June 2012, and endline survey during June 2014 in 42 primary schools in Luwero District, Uganda. Luwero is a large district with urban trading centres and rural sub-districts. Using school enrolment lists, 151 eligible primary schools were identified and grouped into strata according to the sex ratio of their students. Forty-two primary schools were randomly selected proportional to the stratum size, and all agreed to participate. Stratified randomisation was carried out after the baseline survey, with 21 schools receiving the Good School Toolkit and 21 forming a waitlisted control group. All 21 intervention schools completed the Toolkit intervention, which took place over 18 months (corresponding to four school terms) between September 2012 and May 2014. At endline, 92% of the sampled students were interviewed and 91% of all staff interviewed. The study was approved by the London School of Hygiene and Tropical Medicine Ethics Committee (6183) and the Uganda National Council for Science and Technology (SS2520). Our protocol and main trial results, that include details on ethics and consent procedures for all participants in the Good School Study, as well as child protection referral procedures, are published elsewhere [4, 11, 14].
The Good School Toolkit is publicly available at www.raisingvoices.org. It consists of a six-step process that involves implementation of a series of about 60 activities described in manuals and supporting leaflets and posters. Activities are coordinated by two lead teacher ‘protagonists’ and two student representatives in each school; some activities involve outreach to parents and the surrounding community. Schools receive one-on-one support visits and phone calls from Raising Voices staff. The intervention uses the Transtheoretical behaviour change model  and involves the application of behaviour change techniques shown to be effective in other fields and in other violence prevention interventions, including setting a goal, making an action plan, and providing social support . A description of the Good School Toolkit intervention and a summary of the Toolkit six-step process are presented in Additional file 1.
Process evaluation design
Drawing broadly on the Grant et al. 2013 process evaluation framework  we describe the overall implementation of the Toolkit in terms of four implementation components: Raising Voices delivery of the intervention to schools, school-led implementation of Toolkit activities, adoption of Toolkit elements in schools and the reach of the intervention to students and teachers at school. A pre-specified set of process data were prospectively captured in the 21 intervention schools during the 18-month implementation of the Toolkit and at the study end-line survey in 2014. All the measures were developed and pilot tested in the field before use and are described with data quality summarised in Additional file 2.
Delivery of intervention to schools by Raising Voices was measured using data routinely collected by Raising Voices Programme Officers as part of the intervention. Each of four officers directly supported five or six schools, and a Program Manager provided oversight and audit visits to the schools. All interactions with the schools—including technical support visits, group trainings and telephone calls—were systematically documented by each program officer termly. For analysis, the ‘delivery’ variable used was the number of Raising Voices technical support visits per school.
School-led implementation of Toolkit activities was measured using termly ‘action plans’ routinely completed by schools as part of the intervention, and standardised activity monitoring forms distributed to intervention schools collected only for the Good Schools Study. Schools were asked to complete action plans detailing all Toolkit activities they would conduct at the start of each term, and then to record the details of each activity they actually completed on a monitoring form (one monitoring form per activity). The number of planned and completed school-led activities reported were measured throughout implementation and used as the implementation process measures.
Adoption of Toolkit elements by schools was tracked by an independent ‘Study Process Monitor’ who was hired specifically to collect data on implementation of the intervention for the Good Schools Study. Once every term, she asked the Head Teacher or teacher protagonist a standard set of questions about Toolkit structures present in each school, a sub-set of which were verified by direct observation (e.g., the presence of a Good School mural). The number of Toolkit elements present in school were measured using data from the last term of implementation and used as the adoption process measure. This multi-item measure was constructed based on fourteen observations (count 0-14, Cronbach alpha 0.72 that are listed in Additional file 2).
Other factors were measured using items from the baseline and endline surveys, described elsewhere [1, 4]. Violence exposure (in students) and use (in staff) was measured using an adapted version of the International Society for the Prevention of Child Abuse and Neglect Child Abuse Screening Tool-Child Institutional (ICAST) . Mental health difficulties were measured using the Strengths and Difficulties Questionnaire (SDQ) . Both these measures have been widely used and have established data on aspects of validity and reliability in different contexts [20–25]. All questions were adapted and piloted in the local population prior to use in the survey.
The Study Process Monitor was responsible for all process data and tracked termly submission of data from schools and Raising Voices program team, as well as collation and storage of paper copies. Raising Voices intervention delivery data were entered on to standardised excel entry sheets by programme staff, copies of the school Good School term action plans were collected from schools by Program Officers and number of activities listed on the action plans captured on to an excel tracking form by Study Process Monitor and all other process data were collected on paper and double entered on to Epidata. The Study Manager oversaw data entry and the study London based Data Manager ran all data comparison and double entry cleaning reports. The study endline survey data, including Toolkit exposure, were captured on programmed tablet computers with algorithms designed to eliminate erroneous skips. All further data management and analysis was performed in Stata/IC 13.1.
The sample size calculation for the main trial is describe in the main trial paper . All process evaluation analysis is exploratory and therefore no sample size calculation relating to process outcomes are specified for these analyses. All analysis was conducted using Stata/IC 13.1. Endline survey data from the twenty-one intervention schools for both students (n = 1921) and staff (n = 286) were used for analysis. Missing data from baseline and endline surveys were very low (less than 1% for all measures), however there were more missing data for routine implementation measures collected by schools. Data quality issues, including missing data for process measures, are fully described in Table 2.1 in Additional file 2.
Implementation process and endline violence in schools’
To describe the overall implementation of the Toolkit in schools, we calculated school level mean or median values of each measure, along with standard deviations or inter-quartile ranges. To describe the variation in levels of violence in the intervention schools, we calculated the mean percentage of students who experienced any physical violence from staff in the last week, at baseline and endline, along with 95% Confidence intervals.
Factors associated with students’ toolkit exposure
Are students attending schools that implement a higher number of school-led Toolkit activities more exposed to the Toolkit?
Unadjusted mean increase in student individual Toolkit exposure
Model a) Students in schools reporting low, medium or high number of planned school-led Toolkit activities, over four terms of implementation:
Number of students in regression model
n = 1921
Students in schools reporting:
Low number Toolkit activities planned
Medium number Toolkit activities planned
High number Toolkit activities planned
Model b) Students in schools reporting low, medium or high number of completed school-led activities, over four implementation terms:
Number of students in regression model
n = 1921
Students in schools reporting:
Low number of completed Toolkit activities
Medium number of completed Toolkit activities
High number of completed Toolkit activities
Student factors associated with level of student Toolkit exposure
Unadjusted mean difference in student Toolkit exposure (95% CI)
Adjusted mean difference in student Toolkit exposure (95% CI)
Number of students in model
n = 1921
n = 1914 (a)
Student in current school for full Toolkit implementation period
School grade 2014:
0.05 (−0.03 to 0.14)
0.04 (− 0.05 to 0.12)
0.30 (0.20 to 0.39)
0.23 (0.14 to 0.32)
Poorer mental health
−0.42 (− 0.56 to − 0.27)
−0.36 (− 0.51 to − 0.20)
−0.18 (− 0.26 to − 0.10)
−0.18 (− 0.25 to − 0.10)
Three or more meals eaten yesterday
0.10 (0.03 to 0.18)
0.07 (− 0.01 to 0.15)
Absent from school one or more days in last week
−0.12 (− 0.20 to − 0.02)
−0.09 (− 0.18 to − 0.01)
Any self-reported functional difficulty
0.03 (− 0.05 to 0.12)
0.09 (0.00 to 0.18)
Any other violence experienced in the last year(b)
− 0.10 (− 0.18 to − 0.02)
−0.05 (− 0.13 to 0.03)
Factors associated with physical violence at endline
Are school-level process measures correlated with school prevalence of physical violence against students’ from staff at endline and change in prevalence over the Toolkit implementation period
School endline physical violence prevalence (a)
Change in school physical violence prevalence (b)
School-level process measures
Number of schools
Delivery of intervention
Total number of Raising Voices support visits to school.
Total number of planned school-led activities.
Total number completed school-led activities reported.
Toolkit structural elements observed in place, last term of school implementation (count: 0-14).
School mean of student Toolkit exposure (count: 0-10)
School mean of staff Toolkit exposure (count: 0-11).
School mean of teacher Toolkit exposure (count: 0-11).
Student Toolkit exposure measures associations with physical violence from staff experienced by students in the last week
Student Toolkit exposure measures
Physical violence from staff in last week (students self-reports)
Unadjusted Odds Ratio (95% CI)
Total students in models
n = 1921
Exposure to Toolkit (total factor score)
Toolkit exposure factor groupings:
Toolkit exposure and self-reported use of physical violence against students in the last week and last term, presented for all school staff and restricted to teaching staff only
Staff and teachers Toolkit exposure measures
Physical violence use against students
In the last week (staff self-reports)
Physical violence use against students
In the last term (staff self-reports)
Unadjusted Odds Ratio (95% CI)
Unadjusted Odds Ratio (95% CI)
Total staff in model
n = 283
n = 283
Staff Toolkit exposure (count 0-11)
Total teachers in model
n = 246
n = 246
Teachers Toolkit exposure (count 0-11)
Toolkit implementation: Delivery, school-led implementation, adoption and reach
All twenty-one intervention schools completed the six steps of the Toolkit intervention (Additional file 1), although Raising Voices program officers reported that the intensity and quality of each step varied between schools. School implementation measures are summarised in Fig. 2. On average Raising Voices delivery of support visits took place twice per school term and this was similar between schools. The number of planned and reported school-led activities implemented throughout the whole 18 months (four school terms) implementation period varied between schools, ranging from 18 to 52 and 6 to 36 respectively. Little variation in schools reach to students was seen with schools’ average student Toolkit exposure at endline ranging between 8 and 9 out of 10 exposure questions. School averages of staff Toolkit exposure varied between 8 and 11 out of 11 questions.
Variation in the prevalence of physical violence in intervention schools at endline
Factors associated with students’ Toolkit exposure
As expected, student exposure to the Toolkit at endline was associated with attending a school that had reported a higher number of school-led Toolkit activities implemented over the intervention period. Students in schools with a `high’ or `medium’ number of activities planned had a higher mean exposure score compared to those students in schools with a `low’ number planned. A similar association is seen for students attending schools that report more of the Toolkit activities completed, with students in these schools having a higher mean exposure score compared to those in `low’ implementing schools (Table 1).
When investigating individual student’s characteristics and Toolkit exposure, students that had been in the current school for the full implementation period and those students in a higher school grade, had increased odds of exposure to the Toolkit. Girls were much less likely to be exposed than boys. Children experiencing more mental health difficulties were the group of students least likely to be exposed (Table 2).
School and individual factors associated with physical violence from staff in schools’ at endline
Schools where teachers reported more exposure to the Toolkit had larger decreases in prevalence of school violence between baseline and endline (Table 3). One unexpected observation was that schools with an increased number of Toolkit elements present during the last term of implementation, had a smaller decrease in violence over the implementation period. No other school process measures were significantly correlated with school level violence at endline or decrease in prevalence of violence over the implementation period.
Table 4 shows that students with increased exposure to the Toolkit have a 24% reduction in odds of experiencing physical violence irrespective of which intervention school they attend. In terms of exposure to specific Toolkit processes, participation in Good School `active groups’, `classroom rules’, and Toolkit `materials’ were each independently associated with reduced odds of experiencing violence, whereas awareness of Toolkit `tools’ did not show a significant association.
Table 5 shows that teachers who were more exposed to the Toolkit reported less use of physical violence against students over the last week and last term, although this was only statistically significant over the past school term period.
Student factors associated with experience of physical violence from staff at endline
Unadjusted Physical violence from staff towards students.
Adjusted Physical violence from staff towards students.
Number of students in model
n = 1921
n = 1920
Exposure to Toolkit (factor score)
Self-reported functional difficulty with memory and/or concentration
Self-reported functional difficulty with self-care (e.g. washing)
Eaten three or more meals yesterday
Any other violence experienced in the last year (2)
Poorer mental health (3)
Summary of main findings
The intervention delivery, in terms of technical support visits to schools from Raising Voices program staff, was similar across schools. In contrast, we observed large variation between schools in Toolkit implementation and adoption measured by number of school-led Toolkit activities planned and reported as completed by schools and in the number of Toolkit structural elements observed in place in each school. Despite variation in levels of school-led implementation, at the end of the implementation period we found high levels of Toolkit reach as measured by school mean aggregates of students and staff exposure to the Toolkit, with relatively little variation across schools. Regardless which school they attended, girls, lower grade students and students with mental health difficulties were less exposed to the intervention. In terms of intervention outcomes, although there was a large average reduction in physical violence, the reduction and prevalence at endline varied widely across schools, ranging from 7% to 65% prevalence. Even after the intervention and adjusting for individual Toolkit exposure, girls, students with mental health difficulties, who experienced other violence, had eaten fewer meals or had difficulties with self-care, were at higher risk of violence compared with other students.
Strengths and limitations
Like all studies, this process evaluation has strengths and limitations. We used a combination of data collected through a rigorous program of research, and routine monitoring data collected by schools themselves as they would do in the absence of formal research. Data on Toolkit activities planned and completed activities reported by schools were incomplete, and in some cases whole term data were not available from schools. Therefore, these measures may not reflect accurately the number of Toolkit activities in all of the schools. This highlights the need for simple tools and process data embedded in the programme implementation. Measures developed for this study had not been fully tested to determine validity and reliability, however all questions were piloted prior to use. Measures development, construction and data quality issues including estimated internal reliability are documented in Additional file 2. Staff and students responded to interviewer-administered questionnaires during baseline and endline surveys. Like all self-reported measures, there may be some social desirability bias in responses. For teachers, we would expect those more exposed to the Toolkit to report using less violence. However, we would expect a bias in the opposite direction for students, where students who are more exposed to the intervention report more violence experience, which could dilute the effect. Strengths of this evaluation include high student and staff response rate and triangulation of wide range of data sources, use of an independent study process monitor and prospective data collection of process measures.
Why did the intervention work better in some schools than others?
Taken together, our results suggest that increased exposure to the intervention was the main driver of larger intervention effects. This is true for both increased teacher exposure, and increased student exposure, which supports the Toolkit’s holistic model of engagement with multiple actors within a school to try to engender school-wide change. These results also highlight the importance of on-going training and activities to ensure newly transferred teachers and students are exposed to, and invested in, the Good School Toolkit intervention.
Counter-intuitively, schools with more Toolkit structural elements observed in place by our Study Process Monitor in the final term of the intervention implementation had smaller decreases in violence. This might be explained by a ‘last push’ in schools that were slower to implement. This may have resulted in more visible elements in the final term, without full engagement in the program of work required to sustain these elements or investment in the underlying change. None of the other implementation measures captured at the school level were associated with changes in school violence. We are, however, limited by our relatively small sample size of 21 schools.
In terms of exposure to specific Toolkit processes, participation in ‘active groups’, ‘classroom rules’, and Toolkit ‘materials’ were each independently associated with reduced violence, whereas awareness of ‘Tools’ alone was not. However, qualitatively we found that tools such as the ‘wall of fame’ were perceived positively by students and staff suggesting these are important in promoting reward and praise in schools . In summary, all Toolkit processes seem to be important to bring about change, a finding that supports the idea that multiple and repeated engagement with Toolkit ideas contributes to intervention effectiveness.
Two schools that represent unexpected outcomes are shown on Fig. 3: school number 21 had the highest prevalence of violence at endline and shows no change over time, and school number 18 had a significant increase in violence at endline compared to baseline. Through our program monitoring, we are aware that both of these schools had changes in staff during the implementation of the Good School Toolkit. Anecdotally, it is possible that new staff may be less invested in the program, or in some cases even reverse policy on corporal punishment or dismantle the school wide intervention—shifts Raising Voices has experienced in other schools following staff turnover. This highlights the importance of strong leadership and ownership for the programme to remain successful and stresses the need for early identification of schools requiring additional on-going support. Raising Voices also reported that motivation of the Good School teacher protagonists was also an important factor influencing schools sustained implementation. This suggests that identifying and building on protagonists’ motivation may also be important for an effective program.
The Good School Toolkit intervention includes activities that foster a supportive school environment, aim to challenge negative social norms, and encourage student participation and confidence. Nevertheless, we saw that girls and students with mental health difficulties were less exposed to the Toolkit, irrespective of which school they attend. Conversely, and in line with the inclusive nature of the intervention, students reporting functional difficulties (for example, with sight or hearing), who were more absent from school, who had experienced other violence and eaten fewer meals, reported the same levels of exposure to the Toolkit as other students in their schools. These findings could indicate that there is something within all schools, reflecting broader societal and gender norms, that is preventing girls and students experiencing mental health difficulties from participating in school Toolkit activities. Also, the difference observed in Toolkit exposure for girls, compared to boys, could help explain the main study finding that the intervention was slightly more effective in reducing violence in boys overall .
Which students remain at higher risk of violence after the intervention?
Thirty-one percent of students in intervention schools still experienced physical violence from staff in the last week at endline, demonstrating that even after a highly effective intervention some children were still more at risk of violence compared to their peers. Even after accounting for level of exposure to the intervention, girls, students reporting difficulty with self-care, students who had eaten fewer meals, reported more mental health difficulties and those who experienced other violence besides physical violence from staff in the last year, remained at higher risk of physical violence from staff after the intervention.
The underlying reasons why girls might participate less in Toolkit activities, and remain at higher risk of physical violence from school staff even if they do participate in Toolkit activities, is an area requiring further investigation. Our findings may reflect the need for a social norm shift and sustained school-wide cultural change to address negative gender norms. In addition, the Toolkit may benefit from additional activities and content intentionally designed to enhance participation of these groups. This is in line with findings from other school-based intervention studies in Sub-Saharan Africa. A review of HIV prevention programs in youth in Sub-Saharan Africa that included twenty studies delivered in schools, or schools and community, concluded that “attention should go to studying implementation difficulties, sex differences in responses to interventions and determinants of exposure to interventions” . The need for behaviour change at cultural level was highlighted in a sexual reproductive health school-based intervention in Tanzania, where authors emphasised the need to train and monitor teachers to “have supportive relationships with pupils, boost pupil confidence, encourage critical thinking, challenge dominant gender norms, and not engage in physical or sexual abuse.” . While this quantitative evaluation supports the idea that the Good School Toolkit can bring about school cultural change around the use of violence, we also show that this is not universal to all schools and that harmful norms around violence use against some students remain irrespective of which school the student attends. Although gender equity is implicit in many of the processes and the design of the Toolkit, there may be value in emphasising and making it more explicit to teacher and students, in doing so highlighting Toolkit activities that specifically support gender equality and address negative gender norms.
Student risk factors for violence might reflect the circumstances that influence their likelihood of being physically punished at school. For example, students who have eaten fewer meals might be hungry and this could trigger punishment for having less attention in class. There is some evidence of behavioural and attention problems among hungry children from studies in the United States [28–30]. Students with mental health difficulties might also have difficulty concentrating and might display behaviours that could be seen as disruptive or challenging to teachers who may not have the tools or techniques to deal with children exhibiting these difficulties [31, 32]. This is particularly true in a context with limited teaching resources and large classroom sizes [12, 33] and suggests that positive discipline alternatives to physical punishment may not be well applied or may not be sufficient to prevent violence in some circumstances. Such circumstances may challenge teachers who are still transitioning to non-violent approaches to maintain discipline in their school. Teacher capacity-building around the extra support and skills required for some students in the classroom learning environment is a potential area to focus Toolkit intervention activities.
Our results draw attention to children that may have complex issues, including being poly-victimised, having difficult home environments and dealing with a variety of mental health difficulties[1, 34–36], highlighting the need for strengthening the intervention around building sustained capacity within the school system that recognises children experiencing overlapping vulnerabilities. The layered supportive environment that the Toolkit fosters might be one of the very few opportunities for marginalised children to be involved in a positive school programme, where they can be supported to form better relationships and improve communication - skills that can help build confidence and resilience that influence their choices and future life trajectories.
Can we develop indicators for programmers to monitor effectiveness?
Collecting data on program implementation, as an indicator of intervention effectiveness, can be useful for future scale-up efforts [8, 37]. The most useful indicators would be school-level, and easily captured during routine implementation monitoring. Unfortunately, none of our implementation measures collected at the school level were associated with the change in school violence over the intervention period or the prevalence measured at endline directly after intervention. These results should be interpreted with caution, as we have low power to quantitatively detect effects across only 21 intervention schools. The lack of association may also be due to the limitations of the monitoring data we collected. Data collected from schools as part of the implementation of the Toolkit had low levels of completeness, and this may have masked a real association between these measures and intervention effectiveness. We also might not have tracked important process indicators. For example, school enforcement of policies or standards promoted by administration are difficult to measure, but may be important predictors of intervention impact. Hence, refinement and reliability testing of in-school assessments, that are in line with the intervention theory of change, could be a useful addition for future program monitoring. This also underscores the importance of capturing qualitative information to understand school context as well as expert programmer’s knowledge of school specific issues relating to effective implementation of the Toolkit.
Our individual-level measures of exposure to the Toolkit were associated with intervention effect, as would be expected. School-level aggregate of teacher’s exposure was also associated with larger reductions in school violence over the implementation period. This may be a potential indicator that could be used by program implementers—although it would require surveying teachers. In addition, our results indicate that monitoring the number of school-led Toolkit activities planned each term could be a simple way to use routine programme data to identify low implementing schools. However, our results highlight that none of the process measures investigated are good indicators of overall intervention effectiveness, and should therefore not be interpreted in the same way as data on violence outcomes.
Even though the intervention is highly effective at reducing violence against children in school, we found that some schools require additional support to bring about effective and sustained change. It may be possible to increase the effectiveness of the Toolkit by increasing student and teacher exposure. The next layer of Good School programme refinement should attempt to engage with children who were less exposed—in particular, girls, those with poorer mental health and in lower school grade.
The Willington Ssekadde and the Good School implementation team, our team of interviewers and supervisors, Jane Frank Nalubega and the CHA team, our team of interviewers and supervisors, Anna Louise Barr, Heidi Grundlingh, Jennifer Horton, Professor Maria Quigley (independent statistician), our trial steering committee Professor Russell Viner (chair), Dr. Lucy Cluver, and Jo Mulligan.
This work was funded by the MRC/DfID/Wellcome Trust via the Joint Global Health Trials Scheme (to K. Devries), and the Hewlett Foundation and the Oak Foundation (to D. Naker).
Availability of data and materials
The datasets generated and/or analysed during the current study will be publicly available on LSHTM repository ten years after end of data collection in 2014.
LK managed the data collection of the endline surveys, led the process evaluation data analysis and drafted the manuscript. EA participated in the design of the study, advised on data analysis, participated in data interpretation and writing of the manuscript. AM and JN participated in data collection and writing of the manuscript. JCC managed the data collection baseline surveys and participated in the interpretation of the data and writing of the manuscript. JS managed data entry, and participated in the writing of the manuscript. SN, NK and EJW participated in data interpretation and writing of the manuscript. DE participated in the design of the study and writing of the manuscript. DN initiated the idea to do the study, participated in the design of the study, obtained funding, provided comments on data interpretation, and participated in drafting of the manuscript. KMD designed the Good Schools Study including the process evaluation, participated in data collection, data interpretation and obtained funding. All authors read and approved the final manuscript.
Ethics approval and consent to participate
The study was approved by the London School of Hygiene and Tropical Medicine Ethics Committee (6183) and the Uganda National Council for Science and Technology (SS2520). All procedures performed in studies involving human participants were in accordance with the ethical standards of the institutional and/or national research committee and with the 1964 Helsinki declaration and its later amendments or comparable ethical standards. Written voluntary informed consent was obtained from all individual participants included in the study. Headteachers provided consent for schools to participate in the study. Parents were notified and could opt children out of participation in survey data collection. Children themselves provided consent for participation. Staff provided consent for participation in survey data collection.
Dipak Naker developed the Good School Toolkit and is a Co-director of Raising Voices and Sophie Namy is the Raising Voices Learning Coordinator. Angel Mirembe and Janet Nakuti are employed in Raising Voices monitoring and evaluation division, but were managed by LSHTM staff during the study. No other author declared competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
- Devries KM, et al. School violence, mental health, and educational performance in Uganda. Pediatrics. 2014;133(1):e129–37.View ArticlePubMedGoogle Scholar
- UNICEF, Violence against Children in Tanzania: Findings from a National Survey, 2009. Summary report on the prevalence of sexual, physical and emotional violence, context of sexual violence, and health and Behavioural consequences of violence experienced in childhood. . 2011, UNICEF Tanzania, Centers for Disease Control and Prevention, and Muhimbili University of Health and Allied Sciences, Dar es Salaam, Tanzania.Google Scholar
- UNICEF. Violence against children in Kenya: findings from a 2010 National Survey. Nairobi, Kenya: Division of Violence Prevention: National Center for Injury Prevention and Control; 2012.Google Scholar
- Devries KM, et al. The good school toolkit for reducing physical violence from school staff to primary school students: a cluster-randomised controlled trial in Uganda. Lancet Glob Health. 2015;3(7):e378–86.View ArticlePubMedPubMed CentralGoogle Scholar
- Michie S, et al. Specifying and reporting complex behaviour change interventions: the need for a scientific method. Implement Sci. 2009;4:40.View ArticlePubMedPubMed CentralGoogle Scholar
- Moore GF, Audrey S, Barker M, Bond L, Bonell C, Hardeman W, et al. Process evaluation of complex interventions: Medical Research Council guidance. BMJ. 2015;350:h1258.Google Scholar
- Wierenga D, et al. What is actually measured in process evaluations for worksite health promotion programs: a systematic review. BMC Public Health. 2013;13:1190.View ArticlePubMedPubMed CentralGoogle Scholar
- Hargreaves JR, et al. Measuring implementation strength: lessons from the evaluation of public health strategies in low- and middle-income settings. Measuring implementation strength. 2016;31(7):860–7.Google Scholar
- Plummer ML, et al. A process evaluation of a school-based adolescent sexual health intervention in rural Tanzania: the MEMA kwa Vijana programme. Health Educ Res. 2007;22(4):500–12.View ArticlePubMedGoogle Scholar
- Bonell C, Jamal F, Harden A, et al. Systematic review of the effects of schools and school environment interventions on health: evidence mapping and synthesis. Southampton (UK): NIHR Journals Library; 2013. (Public Health Research, No. 1.1.) Chapter 8, Research question 3: process evaluations. Available from: https://www.ncbi.nlm.nih.gov/books/NBK262778/.
- Devries KM, et al. The good schools toolkit to prevent violence against children in Ugandan primary schools: study protocol for cluster-randomised controlled trial. Trials. 2013;14:232.View ArticlePubMedPubMed CentralGoogle Scholar
- Kyegombe N, et al. How did the good school toolkit reduce the risk of past week physical violence from teachers to students? Qualitative findings on pathways of change in schools in Luwero, Uganda. Soc Sci Med. 2017;180:10–9.View ArticlePubMedPubMed CentralGoogle Scholar
- Greco G, Knight L, Ssekadde W, et al. Economic evaluation of the Good School Toolkit: an intervention for reducing violence in primary schools in Uganda. BMJ Global Health 2018;3:e000526.View ArticlePubMedPubMed CentralGoogle Scholar
- Child JC, et al. Responding to abuse: Children's experiences of child protection in a central district, Uganda. Child Abuse & Neglect, 2014;38(10):1647-58.Google Scholar
- Prochaska JO, Velicer WF. The transtheoretical model of health behavior change. Am J Health Promot. 1997;12(1):38–48.View ArticlePubMedGoogle Scholar
- Abraham C, Michie S. A taxonomy of behavior change techniques used in interventions. Health Psychol. 2008;27(3):379–87.View ArticlePubMedGoogle Scholar
- Grant A, et al. Process evaluations for cluster-randomised trials of complex interventions: a proposed framework for design and reporting. Trials. 2013;14:15.View ArticlePubMedPubMed CentralGoogle Scholar
- Brown MB. Algorithm AS 116: the tetrachoric correlation and its asymptotic standard error. Appl Stat. 1977;26:343–55.View ArticleGoogle Scholar
- Costello AB, Osborne J. Best practices in exploratory factor analysis: four recommendations for getting the most from your analysis. Practical Assessment Res Eval. 2005;10(7). Available online: http://pareonline.net/getvn.asp?v=10&n=7.
- Zolotor AJ, et al. ISPCAN Child Abuse Screening Tool Children's Version (ICAST-C): Instrument development and multi-national pilot testing. Child Abuse Negl. 2009;33(11):833-41.Google Scholar
- Goodman R, et al. Using the strengths and difficulties questionnaire (SDQ) to screen for child psychiatric disorders in a community sample. Br J Psychiatry. 2000;177:534–9.View ArticlePubMedGoogle Scholar
- Runyan DK, Dunne MP, Zolotor AJ. Introduction to the development of the ISPCAN child abuse screening tools. Child Abuse Negl. 2009;33(11):842–5.View ArticlePubMedGoogle Scholar
- Runyan DK, et al. The development and piloting of the ISPCAN child abuse screening tool—parent version (ICAST-P). Child Abuse Negl. 2009;33(11):826–32.View ArticlePubMedGoogle Scholar
- de Vries PJ, et al. Measuring adolescent mental health around the globe: psychometric properties of the self-report Strengths and Difficulties Questionnaire in South Africa, and comparison with UK, Australian and Chinese data. Epidemiol Psychiatr Sci. 2017:1-12.Google Scholar
- Goodman A, Goodman R. Strengths and difficulties questionnaire as a dimensional measure of child mental health. J Am Acad Child Adolesc Psychiatry. 2009;48(4):400–3.View ArticlePubMedGoogle Scholar
- Michielsen K, et al. Effectiveness of HIV prevention for youth in sub-Saharan Africa: systematic review and meta-analysis of randomized and nonrandomized trials. AIDS. 2010;24(8):1193–202.View ArticlePubMedGoogle Scholar
- Wight D, Plummer M, Ross D. The need to promote behaviour change at the cultural level: one factor explaining the limited impact of the MEMA kwa Vijana adolescent sexual health intervention in rural Tanzania. A process evaluation. BMC Public Health. 2012;12:788.View ArticlePubMedPubMed CentralGoogle Scholar
- Murphy JM, et al. Relationship between hunger and psychosocial functioning in low-income American children. J Am Acad Child Adolesc Psychiatry. 1998;37(2):163–70.View ArticlePubMedGoogle Scholar
- Jyoti DF, Frongillo EA, Jones SJ. Food insecurity affects school children's academic performance, weight gain, and social skills. J Nutr. 2005;135(12):2831–9.View ArticlePubMedGoogle Scholar
- Alaimo K, Olson CM, Frongillo EA Jr. Food insufficiency and American school-aged children's cognitive, academic, and psychosocial development. Pediatrics. 2001;108(1):44–53.PubMedGoogle Scholar
- Schulte-Körne G. Mental health problems in a school setting in children and adolescents. Dtsch Arztebl Int. 2016;113(11):183–90.PubMedPubMed CentralGoogle Scholar
- Nelson JR. Designing schools to meet the needs of students who exhibit disruptive behavior. J Emot Behav Disord. 1996;4(3):147–61.View ArticleGoogle Scholar
- Elbla AIF. Is punishment (corporal or verbal) an effective means of discipline in schools?: case study of two basic schools in greater Khartoum/Sudan. Procedia - Social and Behavioral Sciences. 2012;69(Supplement C):1656–63.View ArticleGoogle Scholar
- Clarke K, et al. Patterns and predictors of violence against children in Uganda: a latent class analysis. BMJ Open. 2016;6(5):e010443.View ArticlePubMedPubMed CentralGoogle Scholar
- Nalugya-Sserunjogi J, et al. Prevalence and factors associated with depression symptoms among school-going adolescents in Central Uganda. Child Adolesc Psychiatry Ment Health. 2016;10:39.View ArticlePubMedPubMed CentralGoogle Scholar
- Devries KM, et al. Witnessing intimate partner violence and child maltreatment in Ugandan children: a cross-sectional survey. BMJ Open. 2017;7(2):e013583.Google Scholar
- Panovska-Griffiths J, et al. Optimal allocation of resources in female sex worker targeted HIV prevention interventions: model insights from Avahan in South India. PLoS One. 2014;9(10):e107066.View ArticlePubMedPubMed CentralGoogle Scholar