Skip to main content

From intervention to interventional system: towards greater theorization in population health intervention research



Population health intervention research raises major conceptual and methodological issues. These require us to clarify what an intervention is and how best to address it.

This paper aims to clarify the concepts of intervention and context and to propose a way to consider their interactions in evaluation studies, especially by addressing the mechanisms and using the theory-driven evaluation methodology.

Main text

This article synthesizes the notions of intervention and context. It suggests that we consider an “interventional system”, defined as a set of interrelated human and non-human contextual agents within spatial and temporal boundaries generating mechanistic configurations – mechanisms – which are prerequisites for change in health. The evaluation focal point is no longer the interventional ingredients taken separately from the context, but rather mechanisms that punctuate the process of change. It encourages a move towards theorization in evaluation designs, in order to analyze the interventional system more effectively. More particularly, it promotes theory-driven evaluation, either alone or combined with experimental designs.


Considering the intervention system, hybridizing paradigms in a process of theorization within evaluation designs, including different scientific disciplines, practitioners and intervention beneficiaries, may allow researchers a better understanding of what is being investigated and enable them to design the most appropriate methods and modalities for characterizing the interventional system. Evaluation methodologies should therefore be repositioned in relation to one another with regard to a new definition of “evidence”, repositioning practitioners’ expertise, qualitative paradigms and experimental questions in order to address the intervention system more profoundly.

Peer Review reports


Population health intervention research has been defined as “the use of scientific methods to produce knowledge about policy and program interventions that operate within or outside of the health sector and have the potential to impact health at the population level” [1] (see Table 1). This research raises a number of conceptual and methodological issues concerning, among other things, the interaction between context and intervention. This paper therefore aims to synthesize these issues, to clarify the concepts of intervention and context and to propose a way of considering their interactions in evaluation studies, especially by addressing the mechanisms and using the theory-driven evaluation methodology.

Table 1 Definitions of terms used

Main text

To clarify the notions of intervention, context and system

What is an intervention?

According to the International Classification of Health Interventions (ICHI), “a health intervention is an act performed for, with or on behalf of a person or population whose purpose is to assess, improve, maintain, promote or modify health, functioning or health conditions” [2]. Behind this simple definition lurks genuine complexity, creating a number of challenges for the investigators circumscribing, evaluating and transferring these interventions. This complexity arises in particular from the strong influence of what is called the context [3], defined as a “spatial and temporal conjunction of events, individuals and social interactions generating causal mechanisms that interact with the intervention and possibly modifying its outcomes” [4]. Acknowledgement of the influence of context has led to increased interest in process evaluation, such as that described in the Medical Research Council (MRC) guideline [5]. It defines the complexity of intervention by pinpointing its constituent parts. It also stresses the need for evaluations “to consider the influence of context insofar as it affects how we understand the problem and the system, informs intervention design, shapes implementation, interacts with interventions and moderates outcomes”.

Intervention components

How should intervention and context be defined when assessing their specificities and interactions? The components of the interventions have been addressed in different ways. Some authors have introduced the concept of “intervention components” [6] and others that of “active ingredients” [7, 8] as a way to characterize interventions more effectively and distinguish them from context. For Hawe [9], certain basic elements of an intervention should be examined as a priority because they are “key” to producing an effect. She distinguishes an intervention’s theoretical processes (“key functions”) that must remain intact and transferable, from the aspects of the intervention that are structural and contingent on context. Further, she and her colleagues introduced a more systemic approach to intervention [10, 11]. Intervention could be defined as “a series of inter-related events occurring within a system where the change in outcome (attenuated or amplified) is not proportional to change in input. Interventions are thus considered as ongoing social processes rather than fixed and bounded entities” [11]. Both intervention and context are thus defined as being dynamic over time, and interact with each other.

The notion of mechanisms

To understand these interactions between context and intervention, we can use the work by Pawson and Tilley [12] on realistic evaluation. This involves analyzing the configurations between contextual parameters, mechanisms and outcomes (CMO). As such, we can consider the process of change as being marked by various intermediate states illustrated by mechanisms.

Mechanisms may be the result of a combination of factors which can be human (knowledge, attitudes, representations, psychosocial and technical skills, etc.) or material (called “non-human” by Akrich et al. [13]). The notion of mechanism has various definitions. Some authors, such as Machamer et al. [14], define them as “entities and activities organized such that they are productive of regular changes from start or set-up to finish or termination of conditions”. Others define them more as prerequisites to outcomes, as in the realistic approach: a mechanism is “an element of reasoning and reaction of an agent with regard to an intervention productive of an outcome in a given context” [15, 16]. They can be defined in health psychology as “the processes by which a behavior change technique regulates behavior” [8]. This could include, for instance, how practitioners perceive an intervention’s usefulness, or how individuals perceive their ability to change their behavior.

Due to the combinations of contextual and interventional components, the process of change therefore produces mechanisms, which in turn produce effects (final and intermediate outcomes). For instance, we could consider that a motivational interview for smoking cessation could produce different psychosocial mechanisms, such as motivation, perception of the usefulness of cessation and self-efficacy. These mechanisms influence smoking cessation. This constitutes causal chains, defined here as the way in which an ordered sequence of events in the chain causes the next event. These mechanisms may also affect their own contextual or interventional components as a system. For example, the feeling of self-efficacy could influence the choice of smoking cessation supports.

From the intervention to the interventional system

Because the mechanism is the result of the interaction between the intervention and its context, the line between intervention and context becomes blurred [17]. Thus, rather than intervention, we suggest using “interventional system”, which includes interventional and contextual components. An interventional system is produced by successive changes over a given period in a given setting.

In this case, mechanisms become key to understanding the interventional system and could generally be defined as “what characterizes and punctuates the process of change and hence, the production of outcomes”. As an illustration, they could be psychological (motivation, self-efficacy, self-control, skills, etc) in behavioral intervention or social (values shared in a community, power sharing perception, etc.) in socio-ecological intervention.

In light of the above, we propose to define the interventional system in population health intervention research as: A set of interrelated human and non-human contextual agents within spatial and temporal boundaries generating mechanistic configurations – mechanisms – which are prerequisites for change in health. In the same way, we could also consider that the intervention could in fact be an arrangement of pre-existing contextual parameters influencing their own change over time. Figure 1 illustrates this interventional system.

Fig. 1
figure 1

The interventional system

Combining methods to explore the system’s key mechanisms

Attribution versus contribution: a need for theorization

The dynamic nature of interventional systems raises the question of how best to address them in evaluation processes. Public health has historically favored research designs with strong internal validity [18], based on experimental designs. Individual randomized controlled trials are the gold standard for achieving causal attribution by counterfactual comparison in an experimental situation. Beyond the ethical, technical or legal constraints known in population health intervention research [19], trials in this field have a major drawback: they are “blind” to the contextual elements which do influence outcomes, however. Their theoretical efficacy may well be demonstrated, but their transferability is weak, which becomes an issue as intervention research is supposed to inform policy and practice [20]. Breslow [22] made the following statement: “Counterfactual causality with its paradigm, randomization, is the ultimate black box.” However, the black box has to be opened in order to understand how an intervention is effective and how it may be transferred elsewhere.

More in line with the notion of the interventional system, other models depart completely from causal attribution by counterfactual methods. They use a contributive understanding of an intervention through mechanistic interpretation, focusing on the exploration of causal chains [23]. In other words, instead of “does the intervention work? ” the question becomes “given the number of parameters influencing the result (including the intervention components), how did the intervention meaningfully contribute to the result observed?” This new paradigm promotes theory-driven evaluations (TDE) [24, 25], which could clarify intervention-contextual configurations and mechanisms. In TDEs, the configurations and mechanisms are hypothesized by combining scientific evidence and the expertise of practitioners and researchers. The hypothetical system is then tested empirically. If this is conclusive, evidence therefore exists of contribution, and causal inferences can be made. Two main categories of TDEs can be distinguished [24, 26]: realist evaluation and theories of change.

Realistic evaluation

In the first one, developed by Pawson and Tilley [12], intervention effectiveness depends on the underlying mechanisms at play within a given context. The evaluation consists in identifying context-mechanism-outcome configurations (CMOs), and their recurrences are observed in successive case studies or in mixed protocols, such a realist trials [27]. The aim is to understand how and under what circumstances an intervention works. In this approach, context is studied with and as a part of the intervention. This moves us towards the idea of an interventional system. For example, we applied this approach to the “Transfert de Connaissances en REGion” project (TC-REG project), an evaluation of a knowledge transfer scheme to improve policy making and practices in a health promotion and disease prevention setting in French regions [28]. This protocol describes the way in which we combined evidence and stakeholders’ expertise in order to define an explanatory theory. This explanatory theory (itself based on a combination of sociological and psychological classic theories) hypothesizes mechanism-context configurations for evidence-based decision-making. The three steps to build the theory in the TC-REG project [28] are: step 1/ a literature review of evidence-based strategies of knowledge transfer and mechanisms to enhance evidence-based decision making (e.g. the perceived usefulness of scientific evidence); step 2 / a seminar with decision makers and practitioners to choose the strategies to be implemented and hypothesize the mechanisms potentially activated by them, along with any contextual factors potentially influencing them (e.g. the availability of scientific data.) 3/ a seminar with the same stakeholders to elaborate the theory combining strategies, contextual factors and mechanisms to be activated. The theory is the interpretative framework for defining strategies, their implementation, the expected outcomes and all the investigation methods.

Theory of change

In theory of change [25, 29, 30], the intervention components or ingredients mentioned earlier are fleshed out and examined separately from those of context, as a way to study how they contribute to producing outcomes. As with realistic evaluation, the initial hypothesis (the theory) is based on empirical assumptions (i.e. from earlier evaluations) or theoretical assumptions (i.e. from social or psychosocial theories). What is validated (or not) is the extent to which the explanatory theory, including implementation parameters (unlike realist evaluation), corresponds to observations: expected change (i.e. 30 mins of daily physical activity); presence of individual or socio-ecological prerequisites for success (i.e. access to appropriate facilities, sufficient physical ability, knowledge about the meaning of physical activity, etc.) based on psychosocial or organizational theories (e.g. social cognitive theory, health belief model) called classic theories [31]; effectivity of actions to achieve the prerequisites for change (i.e. types of intervention or necessary environmental modifications and their effects) based on implementation theories [31] (e.g COM-B model: Capacity-Opportunity-Motvation – Behaviour Model).; effectivity of actions conducive to these prerequisites (i.e. use of the necessary intellectual, human, financial and organizational (…) resources). This can all be mapped out in a chart for checking [30]. Then, the contribution of the external factors of the intervention to the outcomes can be evaluated. For an interventional system, in both categories, the core elements to be characterized in TDE would be the mechanisms as prerequisites to outcome. The identification of these mechanisms should confirm the causal inference, rather than demonstrating causal attribution by comparison. By replicating these mechanisms, the interventions can be transferred [21, 32]. In the case of TDEs, interventional research can be developed by natural experiment [33], allowing mechanisms to be explored, in order to explain the causal inferences, in a system which is outside the control of investigators. The GoveRnance for Equity ENvironment and Health in the City (GREENH-City) project illustrates this. It aims to address the conditions in which green areas could contribute to reducing health inequality by intervening on individual, political, organizational or geographical factors [34]. The researchers combined evidence, theories, frameworks and multidisciplinary expertise to hypothesize the potential action mechanisms of green areas on health inequalities. The investigation plans to verify these mechanisms by a retrospective study via qualitative interviews. The final goal is to determine recurring mechanisms and conditions for success by cross-sectional analysis and make recommendations for towns wishing to use green areas to help reduce health inequality.

In addition, new statistical models are emerging in epidemiology. They encourage researchers to devote more attention to causal modelling. [35].

The intervention theory

For both methods, before intervention and evaluation designs are elaborated, sources of scientific, theoretical and empirical knowledge should be combined to produce the explanatory theory (with varying numbers of implementation parameters). We call this explanatory theory the “intervention theory” to distinguish it from classic generalist psychosocial, organizational or social implementation theories, determinant frameworks or action models [31], which can fuel the intervention theory. The intervention theory would link activities, mechanisms (prerequisites of outcomes), outcomes and contextual parameters in causal hypotheses.

Note that to establish the theory, the contribution of social and human sciences (e.g. sociology, psychology, history, anthropology) is necessary. For example, the psychosocial, social and organizational theories enable investigators to hypothesize and confirm many components, mechanisms and their relationships involved in behavioral or organizational interventions. In this respect, intervention research becomes subordinate to the hybridization of different disciplines.

Combination of theory-based approaches and counterfactual designs

Notwithstanding the epistemic debates [36], counterfactual designs and theory-based approaches are not opposed, but complementary. They answer different questions and can be used successively or combined during an evaluation process. More particularly, TDEs could be used in experimental design, as some authors suggest [27, 36,37,38]. This combination provides a way of comparing data across evaluations; in sites which have employed both an experimental design (true control group) and theory-based evaluation, an evaluator might, for example, look at the extent to which the success of the experimental group hinged upon the manipulation of components identified by the theory as relevant to learning.

On this basis, both intervention and evaluation could be designed better. For example, the “Évaluation de l’Efficacité de l’application Tabac Info service” (EE-TIS) project [39] combines a randomized trial with a theory-based analysis of mechanisms (motivation, self-efficacy, self-regulation, etc.) which are brought about through behavioral techniques used in an application for smoking cessation. The aim is to figure out how the application works, which techniques are used by users, which mechanisms are activated and for whom. Indeed in EE-TIS project [39], we attributed one or several behavioral change techniques [8] to each feature of the “TIS” application (messages, activities, questionnaires) and identified three mechanisms– potentially activated by them and supporting smoking cessation (i.e. motivation, self-efficacy, knowledge). This was carried out by a multidisciplinary committee in 3 steps: step 1/ two groups of researchers attributed behavior change techniques to each feature, step 2/ both groups compared their results and drew a consensus and step 3/ researchers presented their results to the committee which will in turn draw a consensus. To validate these hypotheses, a multivariate analysis embedded into the randomized control trial will make it possible to figure out which techniques influence which mechanisms and which contextual factors could moderate these links.

Other examples exist which combine a realist approach and trial designs [27, 38].

Interdisciplinarity and stakeholder involvement

A focal point in theorizing evaluation designs is the interdisciplinary dimension, especially drawing on the expertise of social and human sciences and of practitioners and intervention beneficiaries [40]. As an intervention forms part of and influences contextual elements to produce an outcome, the expertise and feedback of stakeholders, including direct beneficiaries, offers valuable insights into how the intervention may be bringing about change. In addition, this empowers stakeholders and promotes a democratic process, which is to be upheld in population health [40]. The theorization could be done through specific workshops, including researchers, practitioners and beneficiaries on an equal basis. For example, the TC-REG project [28] has held a seminar involving both prevention practitioners and researchers, the aim being to discuss literature results and different theories/frameworks in order to define the explanatory theory (with context-mechanism configurations) and intervention strategies to be planned to test it.


Population health intervention research raises major conceptual and methodological issues. These imply clarifying what an intervention is and how best to address it. This involves a paradigm shift in order to consider that in intervention research, intervention is not a separate entity from context, but rather that there is an interventional system that is different from the sum of its parts, even though each part does need to be studied in itself. This gives rise to two challenges. The first is to integrate the notion of the interventional system, which underlines the fact that the boundaries between intervention and context are blurred. The evaluation focal point is no longer the interventional ingredients taken separately from their context, but rather mechanisms punctuating the process of change, considered as key factors in the intervention system. The second challenge, resulting from the first, is to move towards a theorization within evaluation designs, in order to analyze the interventional system more effectively. This would allow researchers a better understanding of what is being investigated and enable them to design the most appropriate methods and modalities for characterizing the interventional system. Evaluation methodologies should therefore be repositioned in relation to one another with regard to a new definition of “evidence”, including the points of view of various disciplines, and repositioning the expertise of the practitioners and beneficiaries, qualitative paradigms and experimental questions in order to address the interventional system more profoundly.



Context-mechanism-outcome configurations

COM-B model:

Capacity-Opportunity-Motvation – Behaviour Model


Évaluation de l’Efficacité de l’application Tabac Info Service


GoveRnance for Equity ENvironment and Health in the City


Classification of Health Interventions (ICHI)


Medical Research Council


Transfert de Connaissances en REGion


Theory-driven evaluation.


Tabac Info service


  1. Hawe P, Potvin L. What is population health intervention research? Can J Public Health. 2009;100(Suppl 1):I8–14.

    PubMed Central  Google Scholar 

  2. WHO | International Classification of Health Interventions (ICHI) [Internet]. WHO. [cité 16 déc 2017]. Disponible sur:

  3. Shoveller J, Viehbeck S, Ruggiero ED, Greyson D, Thomson K, Knight R. A critical examination of representations of context within research on population health interventions. Crit Public Health. 2016;26(5):487–500.

    Article  Google Scholar 

  4. Poland B, Frohlich K, Cargo M. Health Promotion Evaluation Practices in the Americas. New York: Springer; 2008. p. 299–317. Disponible sur:

    Book  Google Scholar 

  5. Craig P, Dieppe P, Macintyre S, Michie S, Nazareth I, Petticrew M; Medical Research Council Guidance. Developing and evaluating complex interventions: the new Medical Research Council guidance. BMJ. 2008;337:a1655. PubMed PMID: 18824488; PubMed Central PMCID: PMC2769032.

  6. Clark AM. What are the components of complex interventions in healthcare? Theorizing approaches to parts, powers and the whole intervention. Soc Sci Med. 2013;93:185–93. Epub 2012 Apr 22. Review. PubMed PMID: 22580076.

  7. Durlak JA. Why program implementation is important. J Prev Interv Community. 1998;17(2):5–18.

    Article  Google Scholar 

  8. Michie S, Richardson M, Johnston M, Abraham C, Francis J, Hardeman W, et al. The behavior change technique taxonomy (v1) of 93 hierarchically clustered techniques: building an international consensus for the reporting of behavior change interventions. Ann Behav Med. 2013;46:81–95.

    Article  PubMed  Google Scholar 

  9. Hawe P, Shiell A, Riley T. Complex interventions: how ‘out of control’ can a randomised controlled trial be? Br Med J. 2004;328:1561–3.

    Article  Google Scholar 

  10. Shiell A, Hawe P, Gold L. Complex interventions or complex systems? Implications for health economic evaluation. BMJ. 2008;336(7656):1281–3.

    Article  PubMed  PubMed Central  Google Scholar 

  11. Hawe P, Shiell A, Riley T. Theorising interventions as events in systems. Am J Community Psychol. 2009;43:267–76.

    Article  PubMed  Google Scholar 

  12. Pawson R, Tilley N. Realistic Evaluation. London: Sage Publications Ltd; 1997.

  13. Akrich M, Callon M, Latour B. Sociologie de la traduction : Textes fondateurs. 1st éd ed. Paris: Transvalor - Presses des mines; 2006. p. 304.

    Book  Google Scholar 

  14. Machamer P, Darden L, Craver CF. Thinking about mechanisms. Philos Sci. 2000;67(1):1–25.

    Article  Google Scholar 

  15. Lacouture A, Breton E, Guichard A, Ridde V. The concept of mechanism from a realist approach: a scoping review to facilitate its operationalization in public health program evaluation. Implement Sci. 2015;10:153.

    Article  PubMed  PubMed Central  Google Scholar 

  16. Ridde V, Robert E, Guichard A, Blaise P, Olmen J. L’approche réaliste à l’épreuve du réel de l’évaluation des programmes. Can J Program Eval. 2012;26.

  17. Minary L, Kivits J, Cambon L, Alla F, Potvin L. Addressing complexity in population health intervention research: the context/intervention interface. J Epdemiology Community Health. 2017;0:1–5.

    Google Scholar 

  18. Campbell D, Stanley J. Experimental and quasi-experimental designs for research. Chicago: Rand McNally; 1966.

    Google Scholar 

  19. Alla F. Challenges for prevention research. Eur J Public Health 1 févr. 2018;28(1):1–1.

    Article  Google Scholar 

  20. Tarquinio C, Kivits J, Minary L, Coste J, Alla F. Evaluating complex interventions: perspectives and issues for health behaviour change interventions. Psychol Health. 2015;30:35–51.

    Article  PubMed  Google Scholar 

  21. Cambon L, Minary L, Ridde V, Alla F. Transferability of interventions in health education: a review. BMC Public Health. 2012;12:497.

    Article  PubMed  PubMed Central  Google Scholar 

  22. Breslow NE. Statistics. Epidem Rev. 2000;22:126–30.

    Article  CAS  Google Scholar 

  23. Mayne J. Addressing attribution through contribution analysis: using performance measures sensibly. Can J Program Eval. 2001;16(1):1–24.

    Google Scholar 

  24. Blamey A, Mackenzie M. Theories of change and realistic evaluation. Evaluation. 2007;13:439–55.

    Article  Google Scholar 

  25. Chen HT. Theory-driven evaluation. Newbury Park: SAGE; 1990. p. 326.

  26. Stame N. Theory-based evaluation and types of complexity. Evaluation. 2004;10(1):58–76.

    Article  Google Scholar 

  27. Bonell C, Fletcher A, Morton M, Lorenc T, Moore L. Realist randomised controlled trials: a new approach to evaluating complex public health interventions. Soc Sci Med 1982. 2012;75(12):2299–306.

    Google Scholar 

  28. Cambon L, Petit A, Ridde V, Dagenais C, Porcherie M, Pommier J, et al. Evaluation of a knowledge transfer scheme to improve policy making and practices in health promotion and disease prevention setting in French regions: a realist study protocol. Implement Sci. 2017;12(1):83.

    Article  PubMed  PubMed Central  Google Scholar 

  29. Weiss CH. How Can Theory-based evaluation make greater headway? Eval Rev. 1997;21(4):501–24.

    Article  Google Scholar 

  30. De Silva MJ, Breuer E, Lee L, Asher L, Chowdhary N, Lund C, et al. Theory of change: a theory-driven approach to enhance the Medical Research Council’s framework for complex interventions. Trials. 2014;15(1) [cité 5 sept 2016]. Disponible sur:

  31. Nilsen P. Making sense of implementation theories, models and frameworks. Implement Sci. 2015;21:10 [cité 18 sept 2018]. Disponible sur:

    Google Scholar 

  32. Wang S, Moss JR, Hiller JE. Applicability and transferability of interventions in evidence-based public health. Health Promot Int. 2006;21:76–83.

    Article  PubMed  Google Scholar 

  33. Petticrew M, Cummins S, Ferrell C, Findlay A, Higgins C, Hoy C, et al. Natural experiments: an underused tool for public health? Public Health. 2005;119(9):751–7.

    Article  CAS  PubMed  Google Scholar 

  34. Porcherie M, Vaillant Z, Faure E, Rican S, Simos J, Cantoreggi NL, et al. The GREENH-City interventional research protocol on health in all policies. BMC Public Health. 2017;17:820.

    Article  PubMed  PubMed Central  Google Scholar 

  35. Aalen OO, Røysland K, Gran JM, Ledergerber B. Causality, mediation and time: a dynamic viewpoint. J R Stat Soc Ser A Stat Soc. 2012;175(4):831–61.

    Article  PubMed  PubMed Central  Google Scholar 

  36. Bonell C, Moore G, Warren E, Moore L. Are randomised controlled trials positivist? Reviewing the social science and philosophy literature to assess positivist tendencies of trials of social interventions in public health and health services. Trials. 2018;19(1):238.

    Article  PubMed  PubMed Central  Google Scholar 

  37. Moore GF, Evans RE. What theory, for whom and in which context? Reflections on the application of theory in the development and evaluation of complex population health interventions. SSM Popul Health. 2017;3:132–5.

    Article  PubMed  Google Scholar 

  38. Jamal F, Fletcher A, Shackleton N, Elbourne D, Viner R, Bonell C. The three stages of building and testing mid-level theories in a realist RCT: a theoretical and methodological case-example. Trials. 2015;16(1):466.

    Article  PubMed  PubMed Central  Google Scholar 

  39. Cambon L, Bergman P, Le Faou A, Vincent I, Le Maitre B, Pasquereau A, et al. Study protocol for a pragmatic randomised controlled trial evaluating efficacy of a smoking cessation e-‘Tabac info service’: ee-TIS trial. BMJ Open. 2017;7(2):e013604.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  40. Alla F. Research on public health interventions: the need for a partnership with practitioners. Eur J Public Health. 2016;26(4):531.

    Article  PubMed  Google Scholar 

Download references


Not applicable.


Not applicable.

Availability of data and materials

Not applicable.

Author information

Authors and Affiliations



All authors read and approved the final version of the manuscript. LC and FA conceived the idea for the paper, based on their previous researches on evaluation of complex interventions, LC wrote the first draft and led the writing of the paper. LC, PT and FA helped draft the manuscript. LC acts as guarantor.

Corresponding author

Correspondence to Linda Cambon.

Ethics declarations

Authors’ information

Not applicable.

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (, which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Cambon, L., Terral, P. & Alla, F. From intervention to interventional system: towards greater theorization in population health intervention research. BMC Public Health 19, 339 (2019).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: