Skip to main content

Does refining an intervention based on participant feedback increase acceptability? An experimental approach

Abstract

Background

Participant feedback is an important consideration for increasing intervention acceptability, yet whether incorporating such feedback actually improves acceptability is rarely tested.

Purpose

The present study describes a theory-based approach to assessing whether refining an intervention based on participant feedback increases acceptability.

Methods

Three hundred and ninety-three UK adults who had previously self-harmed were exposed to the same intervention at baseline and, six months later, were randomly allocated to receive either: (a) the same version of the intervention (control group), or (b) a version of the intervention that had been refined following participant feedback (experimental group). The main outcome measure was acceptability ratings for each of the seven domains specified in the Theoretical Framework of Acceptability (TFA).

Results

Mixed ANOVAs, with control versus experimental group as the between-participants factor and time (baseline versus follow-up) as the within participants factor showed no significant changes in acceptability.

Conclusions

The null effects reported here imply that participants found both the original and modified versions of the intervention equally acceptable, and that our process of refining an intervention based on participant feedback did not impact on acceptability. Nevertheless, we have operationalised a robust approach for examining whether participant feedback impacts on the acceptability of an intervention. Further research is required to understand better how participant feedback should be incorporated into the development of healthcare interventions.

Peer Review reports

Background

Various factors affect the successful implementation of an intervention, such as its feasibility, desirability, and perceived appropriateness [1, 2]. Intervention acceptability in particular is an important consideration in the design, implementation, and evaluation of healthcare interventions [3, 4]. The UK Medical Research Council (MRC) guidelines for developing and evaluating complex interventions recommend assessing intervention acceptability (a key consideration for intervention design and refinement) by engaging potential intervention users to inform the refinements to interventions [2]. The likelihood of successful implementation and subsequent effectiveness is dependent upon perceptions of acceptability [4, 5]. For example, interventions perceived as acceptable by those delivering and/or receiving them are more likely to result in favourable outcomes including adherence to treatment programmes [6], or support for public health policy [5].

Involving people in the design and modification of interventions is recognised as an important stage in ensuring acceptability [1]. Key principles include encouraging researchers to view intervention development as iterative cycles of development to refine an intervention using feedback from people outside the research team throughout the process. Coproduction can involve researchers working together with the public or those with lived experience from the start of a research programme, sharing power and responsibility, and generating knowledge [7]. There is evidence across a broad range of healthcare settings that coproducing interventions with key stakeholders may increase perceptions of intervention acceptability [8, 9]. Additionally, coproduction, for example, may involve people with lived experience providing feedback on an existing intervention with the overarching aim of making the intervention more acceptable and feasible [1]. This could involve stakeholders and intervention developers generating ideas about content, format, style and delivery of interventions [7]. However, there are two potential limitations of this approach.

The first limitation concerns the lack of study methods explicitly testing the acceptability of a modified intervention following participant feedback. This is important because, to our knowledge, no experimental studies have been conducted that evaluate whether changes made to an intervention in response to participant feedback elicited improvements in acceptability.

The second limitation concerns the lack of use of theoretical frameworks to guide the investigation of acceptability. The Theoretical Framework of Acceptability (TFA) [4, 10] is an established guide to assess the acceptability of interventions. It defines acceptability as “a multi-faceted construct that reflects the extent to which people delivering or receiving a healthcare intervention consider it to be appropriate, based on anticipated or experienced cognitive and emotional responses to the intervention” (p.4) [4]. The TFA comprises seven domains: (1) affective attitude (how individuals feel about taking part in an intervention), (2) burden (the amount of effort required to engage with an intervention), (3) perceived effectiveness (whether individuals perceive an intervention as likely to achieve its purpose), (4) ethicality (the extent to which an intervention fits with individuals’ personal values), (5) intervention coherence (whether individuals understand an intervention and how it works), (6) opportunity costs (what is given up, such as time, to take part in an intervention), and (7) self-efficacy (how confident individuals are doing the intervention). The advantage of using the TFA, as opposed to more general approaches to evaluating acceptability, is that the TFA allows a more systematic assessment of intervention acceptability that is comparable across interventions, and enables researchers to target specific TFA domains in future iterations of interventions (e.g. addressing perceived burden of interventions) [11].

Aims

The aims of the present study were to operationalise an approach to intervention acceptability that: (a) ensures that a large representative sample of people with lived experience is involved in the process of refining an intervention to increase acceptability; and (b) adopts a theory-driven, experimental approach to evaluating whether refining an intervention based on participant feedback increases acceptability.

Methods

Overview

Ethical approval was obtained from The University of Manchester Research Ethics Committee (ref: 2020-8446-15312). All methods were performed in accordance with the relevant guidelines and regulations (Declaration of Helsinki). The study was conducted in three phases as part of a larger six-month follow-up study examining the effectiveness of an intervention for reducing self-harm (ClinicalTrials.gov Identifier: NCT04420546). In phase 1, we focused on intervention development, and phase 2 focused on developing and testing an experimental approach to evaluating whether refining an intervention based on participant feedback increases acceptability. Phase 3 is currently ongoing, and is focused on evaluating the effectiveness of the intervention.

Phase 1: intervention development

Development of the volitional help sheet for self-harm

The intervention is based on the concept of implementation intentions [12], which are “if-then” plans that help people to link a critical situation (i.e. “if”) with an appropriate response (i.e. “then”). If-then plans work by making automatic links [13] in memory between a critical situation (“If I am tempted to self-harm when I want to get some attention…”) and an appropriate response (“…then I will do something else instead of self-harming”). The volitional help sheet (VHS) is a tool designed to assist with the formation of implementation intentions to reduce self-harm, and has previously been shown to be effective in reducing self-harm in people recently admitted to the hospital for self-harm [14].

Participant suggestions for refining the intervention were coded to the specific constructs of the TFA as part of our qualitative analyses conducted during our prior intervention development work on the original intervention [15, 16]. Briefly, a directed content analysis approach was used to identify and categorise instances of the TFA domains. To increase the trustworthiness of the data, two authors were involved in the coding process (CK and CJA), with any areas of contention discussed and agreed upon accordingly. A more detailed overview of the coding process is presented elsewhere [13]. Based on the feedback provided by three-hundred and forty participants, three specific amendments were made based on open-ended comments. Firstly, participants described how it would be helpful to emphasise to participants before completing the volitional help sheet that not all situations and solutions were relevant to everyone. Therefore, changes were made to the instructions of the volitional help sheet to further increase clarity and understanding (in line with the intervention coherence domain of the TFA). Secondly, participants suggested that the VHS would be improved if they could enter their own situations and solutions if the statements were not applicable to them. Therefore, open-text fields were added to the volitional help sheet (in line with the ethicality domain of the TFA). Thirdly, participants made useful suggestions about improving the formatting and layout of the intervention webpage. Therefore, prior to distribution, the research team ensured the intervention could be viewed effectively on both a desktop computer and a mobile device (in line with burden domain of the TFA).

The original and modified versions are presented in Figs. 1 and 2 respectively, with modifications highlighted in red (see Fig. 2). The intervention was presented on a single webpage, with a list of situations alongside which participants could choose an appropriate response from a drop-down menu for each critical situation. Specific modifications were made based on the feedback provided by participants at baseline, via open ended comments in line with the seven TFA domains [15].

Fig. 1
figure 1

Volitional help sheet for self-harm (control condition)

Fig. 2
figure 2

Volitional help sheet for self-harm (experimental condition)

Phase 2: main study

Participants

A national community sample of people in the UK who had previously self-harmed were recruited via a survey panel company (YouGov), as part of a larger study (ClinicalTrials.gov Identifier: NCT04420546). Participants were incentivised in line with YouGov’s points system (respondents accumulate points for taking part in surveys, which can then be exchanged for cash or entry into a prize draw). A screening question was asked to ensure that the final sample contained people with a prior history of self-harm: “Have you ever intentionally hurt yourself/self-harmed?” Response options were “yes, I have,” “no, I have not,” or “prefer not to say.” The final sample was based on respondents answering, “yes, I have.”

Design

A mixed-measures design was employed with one between-participants factor (condition: modified intervention versus original intervention) and one within-participants factor (time: pre-randomisation versus six-month follow-up). The primary outcome measure was acceptability, divided into the seven TFA domains: Affective Attitude, Burden, Ethicality, Self-efficacy, Opportunity Costs, Intervention Coherence, and Perceived Effectiveness.

Procedure

At baseline, after participants gave informed consent, all were presented with the original intervention aimed at reducing self-harm. Participants were then randomised via online survey software to receive, at six-month follow-up, either: (a) the same version of the intervention, or (b) a version of the intervention that had been changed following participant feedback. Whilst participants were not expected to use the intervention in the period between baseline and follow-up, the volitional help sheet is intended to provide people with a means of responding to critical situations (where the urge to self-harm may be heightened), with automatic coping plans that were formed during the baseline period.

Measures

Sociodemographic variables

Demographic variables including age, gender, ethnicity, and social grade were taken using standard UK Office for National Statistics [17] measures.

Acceptability measures

Likert scale responses were developed in line with the seven Theoretical Framework of Acceptability (TFA) constructs used to assess acceptability [10]. The items were developed through consensus using the expertise within the research team, and in the absence of a suitable existing measure, psychometric properties of the measure were evaluated. Study-specific amendments were made to the items consistent with the VHS. Seven items were developed, e.g. “On a scale of 0–10, how much effort was required to use the volitional help sheet?” (burden; no effort [0]-lots of effort [10]). Item wordings were developed to closely resemble the definitions provided for each domain of the TFA [4, 10]. The items used to measure each TFA domain are presented in Supplementary File A. Participants were invited to complete the TFA measures at baseline and six-month follow-up.

Reliability was assessed using test-retest reliability (intra-class correlation coefficients). A series of two-way mixed-effects models with measures of absolute agreement were used. ICCs were determined as < 0.40 (poor), 0.40–0.75 (fair to good), and > 0.75 (excellent; [18]). Discriminant validity was assessed using inter-item correlations (Pearson’s r). Pearson’s correlation coefficient (r) was used to assess the strength of the relationship between the items. Pearson’s r is interpreted as 0.10 (small effect), 0.30 (medium effect), and 0.50 (large effect) (Cohen, 1988). A series of pairwise correlations were conducted to examine relationships between the seven TFA items. As each item is deemed to measure a different construct (Affective Attitude, Burden, Ethicality, Self-efficacy, Opportunity costs, Intervention coherence, and Perceived effectiveness), low correlation between items overall was expected (Pearson’s r < .50).

Test-retest reliability results are reported in Table 1. Data are analysed according to participants allocated to the control condition who completed each item at baseline and follow-up: Affective Attitude (n = 174), Burden (n = 170), Ethicality (n = 167), Self-efficacy (n = 170), Opportunity Costs (n = 153), Intervention Coherence (n = 167), Perceived Effectiveness (n = 168). Test–retest reliability was fair to good for all of the items (ICC 0.469-0.703).

Table 1 Reliability demonstrated by intra-class correlation coefficient (ICC) and 95% confidence intervals (CI) for TFA items (control group only)

Discriminant validity results for Time 1 are reported in Table 2. Fourteen of the correlations were small to medium (r = − .112–0.481). Three of the correlations were large: affective attitude and perceived effectiveness (r = .524), ethicality and perceived effectiveness (r = .590), and self-efficacy and intervention coherence (r = .604). Four of the correlations were non-significant. In contrast, at Time 2 (Table 2), eleven of the correlations were small to medium (r = .113–0.472). Six of the correlations were large: ethicality and affective attitude (r = .503), self-efficacy and affective attitude (r = .500), intervention coherence and self-efficacy (r = .657), perceived effectiveness and affective attitude (r = .590), perceived effectiveness and ethicality (r = .727), and perceived effectiveness and opportunity costs (r = .536). Three of the correlations were non-significant.

Table 2 Pearson’s correlations in relation to Time 1a and Time 2b data

Analysis

Descriptive statistics were used to summarise sociodemographic variables. Chi-square was used to compare our sample of people who reported a previous history of self-harm with general population data collected as part of the Adult Psychiatric Morbidity Survey [19]. The success of the randomisation procedure was checked using MANOVA, and two-way repeated measures ANOVAs were used to assess the effect of the two versions of the volitional help sheet at six-month follow-up on acceptability ratings over time. MANOVA was used to establish whether the participants who remained in the study, or dropped out before follow-up were similar at baseline in terms of their demographic characteristics and their TFA ratings. Whilst the seven TFA domains are independent constructs, it may be possible that changes in one domain may affect other domains. Therefore, all TFA domains were included in the analyses, regardless of which domains the refinements were made in line with. This approach allows the researchers to determine whether making refinements in one TFA domain, could impact acceptability ratings of other TFA domains. The between-participants factor was condition (original volitional help sheet versus modified volitional help sheet). The within-participants factor was time (baseline and six-month follow-up acceptability scores). To determine whether floor or ceiling effects were observed, we calculated the proportion of participants scoring at each point on the rating scale of each item. The recognized value of 15% of the sample was used to determine whether floor and ceiling effects were observed with the proportion of responses being at either the minimum or maximum point of the items [20]. G power software [21] was used to calculate the required sample size. Assuming statistical power of 0.95 and an error probability of 0.05, to detect a small effect size (d = 0.20), the total required sample size for a fully powered randomized controlled trial was estimated to be N = 328.

Results

Sample characteristics

The final sample (n = 393) comprised mostly women (67.7%), and 23.9% were aged 18–34 years, 22.4% were aged 35–44 years, 20.4% were aged 45–54 years, and 33.3% were aged 55 years and older. The majority of the sample was White (92.1%), and 65.6% were of higher social grade (non-manual worker) (see Table 3). Characteristics of our sample closely resembled the characteristics of people who reported a history of self-harm according to the Adult Psychiatric Morbidity Survey of the general population [19] in terms of age. However, our sample contained a lower proportion of men, a higher proportion of people from a white background, and a lower proportion of people from a minority ethnic background, compared to national data.

Table 3 Sample demographics in the two groups (final sample)

Randomisation check

The success of the randomisation procedure was checked using MANOVA (a flow diagram is presented in Fig. 3). The independent variable was condition with two levels: modified intervention versus original intervention. The dependent variables were age, gender, and baseline measures of all seven TFA domains. The multivariate test, F(9, 319) = 0.97, p = .46, ηp2 < 0.03, and all the univariate tests, Fs(1, 327) = 0.02 to 1.7, ps > 0.19, ηp2 = 0.01, were nonsignificant, indicating successful randomisation.

Fig. 3
figure 3

Flow diagram showing participant allocation to each condition

Baseline equivalence check

Equivalence between the participants who dropped out versus remained in the study with respect to baseline characteristics was checked using MANOVA. The independent variable was completed at follow-up with two levels: yes versus no. The dependent variables were age, gender, and baseline measures of all seven TFA domains. The multivariate test, F(9, 423) = 0.81, p = .61, ηp2 = 0.02, and all the univariate tests, Fs(1, 431) = 0.00 to 2.3, ps > 0.13, ηp2 s < 0.01, were nonsignificant, indicating no differences in demographic characteristics of TFA ratings.

Changes in acceptability ratings over time

The proportion of participants scoring at each point on the rating scale of each item at baseline and follow-up is presented in Table 4. There are two key findings. First, we observed ceiling effects for the intervention coherence item at baseline (16.4%) and follow-up (18.9%). Second, there was a high proportion of responses at the lower end of the opportunity costs item at baseline (33.6%) and follow-up (37.6%). Table 5 shows the results of the repeated measures ANOVA to assess changes in acceptability ratings over time. The main effects of time and condition on acceptability scores were nonsignificant for all TFA domains (see Table 5). Similarly, all time x condition interaction effects were nonsignificant (all ηp2 < 0.01).

Table 4 Proportion of responses at each point according to item at baseline and follow-up
Table 5 Two-way repeated measures ANOVAs were used to assess the effect of the two versions of the volitional help sheet on acceptability ratings over time

Discussion

This paper aimed to operationalise an approach to intervention acceptability that: (a) ensured that a large representative sample of people with lived experience is involved in the process of improving intervention acceptability; and (b) adopted a theory-driven, experimental approach to evaluating whether refining an intervention based on participant feedback increases acceptability. To our knowledge, this is the first study to apply the TFA to develop an experimental approach to assess whether incorporating participant feedback brings about improvements in intervention acceptability.

Although the number of participants exceeded the required numbers based on prior power calculations, there were no statistically significant differences in acceptability ratings between the control group (the original intervention) and the intervention group (a modified version of the volitional help sheet based on participant feedback) at six-month follow-up. Consequently, more changes may be required to the modified intervention to further increase acceptability. Future research should therefore aim to build on our findings and examine how making further adaptations may affect perceptions of acceptability, with a focus on two specific domains: opportunity costs and intervention coherence. There are two possible reasons that may explain our null findings. First, we observed a high proportion of responses at the lower end of the opportunity costs item (i.e. perceptions of what is given up, such as time, to take part in an intervention) at baseline and follow-up. This may suggest that a key area for refinement would be to explore ways of reducing the perceived costs associated with engaging in an intervention, such as providing incentives for participation [22]. Second, we observed a high proportion of responses at the higher end of the intervention coherence item at baseline and follow-up. Whilst this may suggest the instructions accompanying the original intervention as well as the changes we made to our intervention at follow-up were both perceived to be clear and well understood by participants, further acceptability research within this domain would help to ensure that the possibility of ceiling effects could be eliminated.

Implications

Incorporating participant feedback in the design of interventions is an important stage of intervention development [1], such as feedback about content, format, style and delivery of interventions [7], yet this is still an emerging field [23]. Most intervention acceptability research to-date typically relies on relatively small sub-samples of people with limited focus on actual experiences of engaging with an intervention, leaving judgements open to the “third-person effect” (i.e. perceiving an intervention as having a greater effect on others than on himself or herself) [24]. Informed by the TFA we have developed a process for researchers to systematically examine ways of improving intervention acceptability as part of iterative cycles of intervention development. However, further refinements are needed to our process to ensure our measure is sensitive enough to identify changes over the long-term. Whilst we observed no significant changes in acceptability ratings at follow-up, future intervention development research must aim to deploy appropriate measures to ensure that changes made in response to participant and public feedback actually do increase intervention acceptability.

Strengths and limitations

Previous studies applying the TFA to explore intervention development have primarily been qualitative studies [25, 26]. Few studies have deployed the TFA as a quantitative measure [10, 27]. For example, Renko et al. assessed the acceptability of a training programme to enhance teachers’ physical activity promotion. Whilst outside the scope of the present study, it would be valuable to explore the views of participants who dropped out of the study at baseline, and consequently did not provide feedback at follow-up on the views of the refined intervention. To our knowledge, the present research is the first time a study has developed an experimental methodology, based on the TFA, to evaluate whether changes made to an intervention based on participant feedback improves acceptability. Our measure is designed to assess each of the seven TFA domains: affective attitude, burden, perceived effectiveness, ethicality, intervention coherence, opportunity costs, and self-efficacy. This is important because having a systematic method of determining how to increase intervention acceptability, according to the different facets associated with acceptability, also allows researchers to examine whether processes designed to increase acceptability of interventions are effective following refinements to the intervention.

Increasing acceptability is important when developing interventions iteratively, and in line with recognised guidance for intervention development. We present an experimental approach that allows researchers to target specific TFA domains in future iterations of interventions (e.g. addressing perceived effort to engage with interventions). Another strength of our process is the involvement of stakeholders in the design and modification of interventions, allowing participants to provide specific feedback on an existing intervention with the overarching aim of increasing intervention acceptability and feasibility [1].

Conclusions

Incorporating participant feedback is an important consideration for the design, implementation and evaluation of interventions. The Theoretical Framework of Acceptability provides a framework designed to systematically examine the seven components associated with acceptability. Our tool provides a systematic method of evaluating whether refining an intervention based on participant feedback increases acceptability. With further testing across different populations and interventions we hope the tool provides researchers with a method of improving intervention acceptability.

Data availability

The datasets used and/or analysed during the current study are available from the corresponding author on reasonable request.

References

  1. O’Cathain A et al. Guidance on how to develop complex interventions to improve health and healthcare 2019. 9(8): p. e029954.

  2. Skivington K et al. A new framework for developing and evaluating complex interventions: update of Medical Research Council guidance 2021. 374: p. n2061.

  3. Moore GF, et al. Process evaluation of complex interventions: Medical Research Council guidance. BMJ: Br Med J. 2015;350:h1258.

    Article  Google Scholar 

  4. Sekhon M, Cartwright M, Francis JJ. Acceptability of healthcare interventions: an overview of reviews and development of a theoretical framework. BMC Health Serv Res. 2017;17(1):88.

    Article  PubMed  PubMed Central  Google Scholar 

  5. Diepeveen S, et al. Public acceptability of government intervention to change health-related behaviours: a systematic review and narrative synthesis. BMC Public Health. 2013;13:756.

    Article  PubMed  PubMed Central  Google Scholar 

  6. Hommel KA, et al. Telehealth behavioral treatment for medication nonadherence: a pilot and feasibility study. Eur J Gastroenterol Hepatol. 2013;25(4):469–73.

    Article  PubMed  PubMed Central  Google Scholar 

  7. INVOLVE N. Guidance on co-producing a research project. INVOLVE Southampton; 2018.

  8. Brook J, et al. Co-production of an intervention to increase retention of early career nurses: acceptability and feasibility. Nurse Educ Pract. 2020;47:102861.

    Article  PubMed  PubMed Central  Google Scholar 

  9. Madden M et al. Producing co-production: reflections on the development of a complex intervention. 2020. 23(3): p. 659–69.

  10. Sekhon M, Cartwright M, Francis JJ. Development of a theory-informed questionnaire to assess the acceptability of healthcare interventions. BMC Health Serv Res. 2022;22(1):279.

    Article  PubMed  PubMed Central  Google Scholar 

  11. Sekhon M, Cartwright M, Francis J. Application of a theoretical framework to assess intervention acceptability: a semi–structured interview study. EHPS/DHP; 2016.

  12. Gollwitzer PM. Implementation intentions: strong effects of simple plans. Am Psychol. 1999;54(7):493–503.

    Article  Google Scholar 

  13. Gollwitzer PM, Sheeran P. Implementation intentions and goal achievement: a Meta-analysis of Effects and processes, in advances in experimental social psychology. Academic Press; 2006. pp. 69–119.

  14. Armitage CJ, et al. An exploratory randomised trial of a simple, brief psychological intervention to reduce subsequent suicidal ideation and behaviour in patients admitted to hospital for self-harm. Br J Psychiatry. 2016;208(5):470–6.

    Article  PubMed  Google Scholar 

  15. Keyworth C, et al. Acceptability of a brief online theory-based intervention to prevent and reduce self-harm: a mixed methods evaluation. Journal of Medical Internet Research; 2021.

  16. Keyworth C, et al. Exploring the acceptability of a brief online theory-based intervention to prevent and reduce self-harm: a theoretically framed qualitative study. BJPsych Open. 2022;8(6):e184.

    Article  PubMed  PubMed Central  Google Scholar 

  17. Statistics. U.O.f.N. 2020.

  18. Fleiss JL. Design and analysis of clinical experiments. Volume 73. John Wiley & Sons; 2011.

  19. McManus S, et al. Prevalence of non-suicidal self-harm and service contact in England, 2000-14: repeated cross-sectional surveys of the general population. Lancet Psychiatry. 2019;6(7):573–81.

    Article  PubMed  PubMed Central  Google Scholar 

  20. McHorney CA, Tarlov AR. Individual-patient monitoring in clinical practice: are available health status surveys adequate? Qual Life Res. 1995;4(4):293–307.

    Article  CAS  PubMed  Google Scholar 

  21. Faul F, et al. G*Power 3: a flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behav Res Methods. 2007;39(2):175–91.

    Article  PubMed  Google Scholar 

  22. Martinsson C, et al. What incentives influence employers to engage in workplace health interventions? BMC Public Health. 2016;16(1):854.

    Article  PubMed  PubMed Central  Google Scholar 

  23. Wong E, Mavondo F, Fisher J. Patient feedback to improve quality of patient-centred care in public hospitals: a systematic review of the evidence. BMC Health Serv Res. 2020;20(1):530.

    Article  PubMed  PubMed Central  Google Scholar 

  24. DAVISON WP. The third-person effect in communication. Pub Opin Q. 1983;47(1):1–15.

    Article  Google Scholar 

  25. Murphy AL, Gardner DM. Pharmacists’ acceptability of a men’s mental health promotion program using the theoretical Framework of Acceptability. AIMS public health. 2019;6(2):195–208.

    Article  PubMed  PubMed Central  Google Scholar 

  26. Pavlova N, Teychenne M, Olander EK. The concurrent acceptability of a postnatal walking group: a qualitative study using the theoretical Framework of Acceptability. Int J Environ Res Public Health. 2020;17(14):5027.

    Article  PubMed  PubMed Central  Google Scholar 

  27. Renko E, et al. Acceptability, reach and implementation of a training to enhance teachers’ skills in physical activity promotion. BMC Public Health. 2020;20(1):1568.

    Article  PubMed  PubMed Central  Google Scholar 

Download references

Acknowledgements

The authors wish to thank all of the participants involved in this study.

Funding

This work was funded by the National Institute for Health and Care Research (NIHR) Greater Manchester Patient Safety Translational Research Centre (grant number: PSTRC-2016-003). C.J.A. is supported by the NIHR Manchester Biomedical Research Centre. The views expressed are those of the author(s) and not necessarily those of the NIHR or the Department of Health and Social Care.

Author information

Authors and Affiliations

Authors

Contributions

C.K. and C.J.A. conceived and designed the study. CK led the data anlysis. All authors contributed to the interpretation of the data. C.K. drafted the manuscript. All authors contributed to the manuscript edits and revisions, and approved the final version.

Corresponding author

Correspondence to Chris Keyworth.

Ethics declarations

Competing interests

The authors declare no competing interests.

Ethics approval and consent to participate

Ethical approval was obtained from The University of Manchester Research Ethics Committee (ref: 2020-8446-15312). All methods were performed in accordance with the relevant guidelines and regulations (Declaration of Helsinki). All participants gave informed consent prior to the study.

Consent for publication

Not applicable.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Electronic supplementary material

Below is the link to the electronic supplementary material.

Supplementary Material 1

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Keyworth, C., Quinlivan, L., Leather, J.Z. et al. Does refining an intervention based on participant feedback increase acceptability? An experimental approach. BMC Public Health 23, 1598 (2023). https://doi.org/10.1186/s12889-023-16344-w

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12889-023-16344-w

Keywords