Skip to main content

Measuring the tail of the dog that doesn't bark in the night: the case of the national evaluation of Choose Life (the national strategy and action plan to prevent suicide in Scotland)



Learning about the impact of public health policy presents significant challenges for evaluators. These include the nebulous and organic nature of interventions ensuing from policy directives, the tension between long-term goals and short-term interventions, the appropriateness of establishing control groups, and the problems of providing an economic perspective. An example of contemporary policy that has recently been subject to evaluation is the first phase of the innovative Scottish strategy for suicide prevention (Choose Life).

Discussion and summary

This paper discusses how challenges, such as those above, were made manifest within this programme. After a brief summary of the overarching approach taken to evaluating the first phase of Choose Life, this paper then offers a set of recommendations for policymakers and evaluators on how learning from a second phase might be augmented. These recommendations are likely to have general resonance across a range of policy evaluations as they move from early planning and implementation to more mature phases.

Peer Review reports


In recent years two interrelated sets of arguments about the evaluation of public health policy have become commonplace. The first laments the lack of a robust evidence base for developing policy, and charges evaluators to do better in producing such knowledge. This, it is argued, should be brought about by the more frequent application of gold-standard approaches to complex interventions [1, 2] and through adopting 'a more integrated approach to evaluation and implementation, using the most robust and sound methodologies, and taking advantage of "natural experiments"' (p.40) [3]. This is a view endorsed by the 2004 Wanless report on improving population health [4].

The second argues that both evaluators and policymakers need to be more explicit about the kinds of knowledge that they want to generate from the study of such programmes, and the place of this evidence in subsequent policymaking [59].

In this paper, within the context of these debates, we exemplify the challenges inherent in evaluating Choose Life: The National Strategy and Action Plan to Prevent Suicide in Scotland. We then briefly summarise the overarching approach taken to evaluating the first phase of Choose Life. Finally, we set out recommendations for framing subsequent evaluation of the initiative as it enters its more mature second phase. These recommendations have implications for policymakers and evaluators within public health and in wider policy domains.

Choose Life was launched in 2002 as a major strand within the National Programme for Improving Mental Health and Well-being in Scotland [10]. As with much contemporary policy emanating from the Scottish Executive, the National Programme was charged with tackling health inequalities within a wider health improvement agenda [11]. The overall aim of Choose Life was to reduce suicide by 20% over a ten year period through the achievement of objectives such as improved early prevention and crisis response, engagement with the media and, adoption of an evidence-based approach. More detail on the guiding principles of Choose Life and its objectives and priority groups is presented in table 1 below.

Table 1 Choose Life: Principles, Objectives and Priority Groups

The Choose Life plan was to be implemented in two phases from 2003–2013. In phase one (2003–2006) £9 m was allocated to local partnerships across each local authority area in Scotland, to be used for suicide prevention work, whilst £3 m was set aside for national activities in support of the plan. Local areas were, in addition, positively encouraged to generate additional resources for Choose Life activities. A further £8.4 m was set aside for the first two years of the second phase (2006–2008). A more detailed description of the structure of the initiative is provided by Platt and colleagues [12].

An intention to evaluate phase one of Choose Life was signalled in the national strategy and action plan [10] and, following a competitive tendering process, a consortium of researchers at the Universities of Edinburgh and Glasgow, the London School of Economics and the Scottish Development Centre for Mental Health was awarded a two year contract commencing in mid-2004.


A recent comparative review of suicide prevention strategies across the globe, carried out as part of the commissioned Choose Life evaluation, highlighted that the majority of government-funded efforts to generate learning about effective approaches have focused on evaluating specific interventions rather than on national strategies per se [12]. For example, Mann and colleagues, in their review of suicide prevention strategies, look at the impact of individual components of strategies rather than at the effect of an overall programme of activity [13]. Whilst Australia commissioned a summative evaluation of its national youth suicide prevention strategy, Scotland was unique in its investment in a national level formative evaluation although other countries are now following suit. Although this provided a welcome opportunity to learn how whole systems create the context for effective suicide prevention strategies, it raised an array of challenges for would-be evaluators.

In this section we consider four specific issues that needed to be addressed in the evaluation of the first phase of Choose Life: how to learn from a complex initiative with multiple stakeholders and outcomes; understanding secular trends in reaching evaluative judgements (effectiveness); attempting to provide an economic perspective on the worth of a national programme (cost-effectiveness); and integrating data arising from a range of methodological approaches. This list, while not exhaustive, represented the most significant challenges for the evaluation team.

The nature of complexity in policy initiatives

Current approaches to making and implementing policy add to programme complexity in a number of ways. Here we briefly outline seven elements of this complexity and illustrate in Table 2 below the ways in which these were made manifest in Choose Life.

Table 2 Challenges of Complexity

First, there is a growing recognition that there has been an expansion in the number and types of roles that evaluators are expected to play [1416]. In particular, policy evaluators are now expected to provide developmental support to programmes as well as to make judgements of impact, and to provide formative as well as summative feedback. Whilst the evaluation of the first phase of Choose Life was explicitly focused on the process of implementation, there were, nonetheless, requests from funders for the evaluation team to answer questions of effectiveness. Negotiating, and remaining focused on, the purpose of an evaluation is, therefore, a key challenge; such purpose will, in addition, change as the evaluation moves from a developmental to a theory-testing stage.

Second, national programmes involving a wide range of stakeholders inevitably involve multiple outcomes [17]. Choose Life is a useful example of this. Although the national strategy set itself one long-term target (to reduce the suicide rate in Scotland by 20% between 2002 and 2013), the local areas were given seven key objectives and were expected to meet 12 milestones (progress towards these measures is detailed by Platt and colleagues [12]. Multiple outcomes do not simply muddy the waters for evaluators through increasing the range of data to be collected. A feature of complex initiatives is that there is a general lack of clarity about when change across and between outcomes is anticipated because of the unknown synergies that may occur [18]. The evaluator, therefore, can struggle to identify the most appropriate point at which to look for outcomes and the order in which they will appear. Furthermore stakeholders' outcomes can be incompatible and disagreements can occur regarding the primary focus of an impact assessment.

Third, as described by Connell and Kubisch [17], strategic approaches to tackling health and social problems exhibit 'horizontal complexity'. This means that a range of different organisations, often with mismatched agendas, have a stake in the outcomes (or funding) of an initiative. This may be particularly problematic where funding is disbursed locally to partnership bodies (Community Planning Partnerships in the case of Choose Life) whose member organisations have different approaches to mental health promotion and to how evidence is gathered and interpreted. In addition, the Choose Life programme manifested horizontal complexity at a national level, with its National Implementation Support Team charged with working in partnership with other bodies at a Scotland-wide level. Once again, experience of evaluating such initiatives has demonstrated that the evaluator has a tricky balance to maintain in satisfying the requirements of different stakeholders.

Fourth, complexity is further increased in strategies that aim to impact on organisations, specific professionals and practices, communities and individuals ('vertical complexity'). Potential outcomes can start to multiply exponentially [17]. Evaluators cannot and should not measure everything: they need to develop a purposive strategy for determining what will be assessed and maintain a focus on this throughout the course of an evaluation.

Fifth, evaluators who seek to use controlled research designs to demonstrate that observed changes can be attributed to the implementation of the programme may find that complex initiatives are ill-suited to such approaches. Where programmes are delivered in a small number of pilot areas, the thrust of policy momentum can result in programme diffusion to what might have been regarded as appropriate control sites [7, 8]. This is clearly exacerbated in strategies where no pilot work is undertaken and an initiative is launched across an entire country, as with Choose Life. Some commentators have argued that, with minor design modifications, the practical problems of carrying out randomised controlled trials can be overcome [19]. The political realities of strategy development and implementation within the UK, however, throw this into question. A randomised design is also compromised by the lack of comparability of interventions and their implementation processes between intervention areas.

A sixth ingredient of complexity is that programmes are shaped by the contexts in which they are located. The body of literature emerging from a theory-based approach to evaluation argues that it is inappropriate and simplistic to ask whether an intervention works, since an intervention is made up of a myriad of interacting components, and because some elements will work for particular groups of individuals in specific contexts and circumstances [2022]. Each of the local areas charged with implementing Choose Life has its own organisational history and was given freedom to deliver the programme in locally determined ways. Understanding context, therefore, becomes a more important evaluation task than ironing it out through a controlled design. Nevertheless, understanding the contribution made by a programme to observed change remains a concern for evaluators. We expand on this in the following section.

Finally, complex health and social programmes are launched to tackle the most intractable problems. Their long-term goals, therefore, are ambitious and, more importantly, lie beyond the timescales of the interventions set up to deliver them and the evaluations designed to measure them. This challenges evaluators to select plausible, robust and short-term intermediate or proxy measures of the more distal intended outcome(s). In the metaphor of our title, the long term outcome of interest is the absence of suicide events (the dog that doesn't bark in the night). Whilst we cannot realistically capture this within the life of an evaluation, the evaluator's goal is to ensure that process measures are appropriate proxies for this ultimate outcome (in other words, to make a case for the measurement of the tail of the dog as a predictor of its future barking behaviour). In suicide prevention evaluations the quest for appropriate proxies is made difficult by the deficiencies of the existing evidence base. There remains a significant leap in the dark, for example, from the provision of mental health awareness programmes in secondary school to the prevention of suicide attempts in the community at some future date. Likewise, there is debate about the value of proxies such as hospital-treated self-harm as a marker for later suicide [23].

Secular trends

As discussed above, demonstrating that a programme has led to a change in outcome is not in itself enough to convince observers that the intervention has played a causal role. This is true in particular when trends in the outcome of interest are on a downward slope. For example, the rates for suicides and undetermined deaths for men in Scotland rose to a peak in 1992 (34.2/100,000 population aged 15+ years), remained at (or just below) this high level until 2002 (34.1/100,000), after which there has been a decline to 27.1/100,000 in 2005 (lowest rate since 1991) [12]. The decline since 2002 may be attributable in part to Choose Life activities. However, the massive reduction in male suicide and undetermined deaths between 2002 (n = 673, rate = 34.1/100,000) and 2003 (n = 577, rate = 29.1/100,000) occurred when Choose Life had only been very partially implemented and is instead probably due, in large part, to the influence of other factors (e.g. legislation restricting paracetamol sales [2325]). In the absence of control data there is, therefore, a significant challenge in interpreting trends over time.

This is exacerbated by the relative rarity of suicide as an event. It becomes virtually impossible for evaluators to propose (dis)confirmable explanations for observed changes in trends over short time periods (in this instance, two years) and at a small area level. Teasing out the possible impact of different strategic approaches within the umbrella of Choose Life on rates of suicide is equally problematic. Once again this points to the need for plausible short-term indicators of progress but without evidence of strong indicators of suicidal behaviour this remains difficult.

Undertaking an economic analysis

As debates about scarcity of resources within the NHS grow, there is a parallel increase in policy interest in addressing the cost-effectiveness of differing approaches to health improvement, yet economic evaluations of complex public health interventions remain relatively rare [26]. Sefton [27] has argued, that bringing health economic perspectives from the field of narrowly defined technology assessment into the domain of complex policy evaluation raises both practical and epistemological difficulties. First and foremost, approaches to assessing cost-effectiveness rely on the existence of robust measures of both costs and benefits. At the level of policy and public health interventions, however, these types of data are woefully limited and the problems that beset randomised controlled trials also undermine the use of standard health economic techniques [28]. In addition, the economic notion of time preference has meant that short term outcomes, more likely to be seen in health care interventions, are usually deemed to have greater economic benefits than health promoting interventions that may take several years to achieve change. [30, 31] Second, paradigmatic conflicts between those working in the field of traditional outcome studies and those dealing with the complexities of messy policy evaluation can result in stereotypical and blinkered views of the contribution to be made by different disciplines and research approaches. For example, McDaid and colleagues argue that, within the field of suicide prevention, an economist's focus on a single outcome, such as the rate of suicide or the number of life years saved, may be viewed as inappropriately reductionist by those interested in capturing the complexity of synergistic approaches to mental health promotion [28]. Conversely, a focus on process and organisational change to the detriment of measurable outcomes may be viewed by economists as a failure to pursue an intervention's bottom line. As Sefton argues, making these 'ends meet' requires not only that evaluation teams are multidisciplinary in name but also that they practise in an integrative fashion [27].

Integrating findings

Achieving an integrated approach to evaluation is much less straightforward than it sounds. Accepting the need to address questions of process and outcome and designing an evaluation that will seek to gather data which are appropriate to both can still result in 'silo evaluations', with the final report summarising results in separate sections. In their evaluation of the Total Purchasing Pilot in primary care, Evans and colleagues illustrate this through their description of how, in the absence of an a priori theoretical framework, members of a multidisciplinary team were left to undertaken their own sub-components of a multi-method evaluation [32]. Even explicit and pre-emptive attempts to be integrative, such as the use of theory-based approaches to evaluation, can struggle to combine data from different sources [7].

Our approach

Although aware of the challenges that have beset policy evaluations in recent years when we embarked on our evaluation of the first phase of Choose Life, we nevertheless struggled to find satisfactory solutions through our evaluation approach and practice. Here we summarise the main aims of the national evaluation of Choose Life as set out within the research brief and the general evaluation framework that we used.

At its most general level the purpose of the national evaluation was to 'assess ... infrastructure and early impacts' of Choose Life and to 'set the template for the next phase' of the strategy [10]. Its aims are detailed in table 3 below:

Table 3 National Evaluation of Choose Life – Main Aims

Our overarching evaluation framework derived from Theories of Change [20, 22]. This approach, designed specifically for the evaluation of complex, multiple outcome and multi-partner programmes, aims to study cumulatively the links between the purposive activities undertaken to tackle significant social problems, their short-term outcomes and processes, and longer-term goals. It seeks to address questions about which components of an overall programme work for which individuals in which sets of circumstances. In this regard it is well suited to the evaluation of a national strategy that is expected to operate differently in different local contexts, since it aims to capture why and how local variations occur. It also explicitly recognises that different stakeholders within an intervention will have potentially competing expectations of a programme and seeks to deal with this by encouraging a collaborative approach to agreeing a theory of change and prioritising key indicators and outcomes for tracking progress.

Whilst the approach is not without its practical and conceptual difficulties [3335], it does nonetheless aim to provide an integrative framework for data generation and interpretation. For example, using the approach allows questions of effectiveness and cost-effectiveness to be embedded in a chain of events linking complex organisational processes, monitoring data and proxy measures of programme success. Its main difficulties include its time-consuming nature, the challenges of eliciting detailed and well-specified theories from stakeholders, and (of particular significance in the evaluation of Choose Life) how to bridge gaps in knowledge between unsatisfactory proxy measures and the long-term goal of a programme. This latter problem can be illustrated by comparing a suicide prevention programme with a strategy to reduce coronary heart disease: whilst a three year heart disease programme might be deemed successful if it reduces levels of risk factors such as obesity or smoking, the links between self-harm incidence and suicide incidence are less clear-cut. Nonetheless, a Theories of Change approach has provided an evaluation framework that allowed researchers from a range of disciplines to work towards an integrated set of evaluation questions and helped to direct the team towards the selection of appropriate methods, including documentary analysis, interviews, group exercises and surveys [10]. An evaluation database was designed and utilised to store and retrieve data that were analysed iteratively and thematically. The data generated were, in addition, put in the context of international approaches to suicide prevention.

Whilst no separate economic data collection was undertaken, the research instruments developed for use in the study were deliberately informed by an economic perspective. Although imperfect, these allowed the evaluation team to consider the extent to which Choose Life stimulated additional local investment from their partnership organisations, including in-kind investments and the contribution of unpaid volunteers, targeted resources at relevant priority groups, spent their funding on proven effective practice and interventions, and stimulated local innovation.

In summary our approach aimed to provide a formative assessment of whether processes and short-term outcomes at far remove from the distal outcome – reduced rates of suicide – were being encouraged at a local and national level. It is not within the remit of this paper to discuss the findings of the evaluation – these are presented elsewhere [12], however, through the approach adopted the team were able to make recommendations concerning future investments in suicide prevention, sustainability, targeting of efforts, the integration of self-harm into suicide prevention strategies and the role of local and national partnerships. To illustrate we provide some specific examples. First, in relation to targeting local action we found that the long list of priority groups set by Choose Life was unhelpful since it led to rank ordering at a local level, and, in some cases, a lack of focus on local needs assessment. Second, in terms of sustainability, our data suggest that, whilst the initiative acted as an unprecedented rallying call for cross-sectoral planning to embed suicide prevention within the mental health improvement agenda, the local and national contexts made it difficult to bring together clinical and community approaches to prevention and treatment. Finally, in terms of defining the boundaries of Choose Life there was evidence of considerable variation across the programme of how 'high risk' suicidal behaviour was conceptualised and tackled. This suggests that nationally greater consideration needs to be given to how action on self-harm is integrated within a wider suicide prevention strategy.


As discussed earlier, learning from the first phase of Choose Life was expected to feed into its second and more mature implementation phase. Many commentators who argue that public health and social care programmes, like medical interventions, should ideally be submitted to testing through randomised controlled trials acknowledge that this may be impossible in the early stages of implementation as national strategies struggle to find their local feet [36]. Instead they suggest that, once interventions have matured into stable entities, then they can be evaluated using gold standard approaches. Notwithstanding the impossibility of utilising a controlled approach to an intervention that has already been rolled out nationally, there are other difficulties with this suggestion. Above all, it is questionable whether social programmes ever become stable interventions – more likely they will strengthen or weaken and constantly evolve as part of a local and national system for tackling health improvement.

To help generate learning for the future we make the following recommendations that build on those contained in our final report [12]. First, an integrative framework such as Theories of Change should continue to shape purposive strategies for generating evidence. Second, work to determine nationally and locally acceptable proxy measures of success for phase two should be undertaken and resulting targets used to stimulate action rather than to create perverse incentives [37]. Third, continued efforts to learn about the effectiveness and cost-effectiveness of different strategies for intervening to promote mental health, prevent suicide, and alleviate its effects in different contexts should be prioritised – ongoing routine but comparable data collection on the uptake of different individual interventions and the resource contributions of unpaid volunteers to their delivery should be an important element of this process. Fourth, learning about the impact of different organisational structures on proxy indicators of progress should be undertaken through case-study evaluations of contrasting approaches. Fifth, whilst suicide prevention is a unique goal, efforts to capitalise on learning across policy domains about how to interpret/utilise evidence and mainstream good practice should be maximised. Finally, measuring trends in suicide rates will remain an important strand of monitoring progress only insofar as it is undertaken sensitively, with due cognisance of the small number of events involved and the distorting effect of 'natural' year on year fluctuations. With such a strategy in place, useful learning may emerge about how to impact positively on mental health and well-being and suicide prevention at strategic and local, as well as programme and practice, levels. Only then will we be in the position to judge whether we have appropriately measured the dog that doesn't bark in the night.


  1. Macintyre S, Chalmers I, Horton R, Smith R: Using evidence to inform health policy: case study. British Medical Journal. 2001, 322 (7280): 222-5. 10.1136/bmj.322.7280.222.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  2. Oakley A: Experimentation and social interventions: a forgotten but important history. British Medical Journal. 1998, 317 (7167): 1239-42.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  3. Judge K, Platt S, Costongs C, Jurczak K: Health Inequalities: a challenge for Europe. 2006, London: Department of Health

    Google Scholar 

  4. Wanless D: Securing good health for the whole population. 2004, Norwich: Stationery Office

    Google Scholar 

  5. Young K, Ashby D, Boaz A, Grayson L: Social science and the evidence-based policy movement. Social Policy and Society. 2002, 1 (3): 215-24. 10.1017/S1474746402003068.

    Article  Google Scholar 

  6. Sanderson I: Getting evidence into practice: perspectives on rationality. Evaluation. 2004, 10 (3): 366-79. 10.1177/1356389004048283.

    Article  Google Scholar 

  7. Bauld L, Judge K, Barnes M, Benzeval M, Mackenzie M, Sullivan H: Promoting social change: the experience of Health Action Zones in England. Journal of Social Policy. 2005, 34 (3): 427-45. 10.1017/S0047279405008858.

    Article  Google Scholar 

  8. Mackenzie M, Blamey A, Hanlon P: Using and generating evidence: policy makers' reflections on commissioning and learning from the Scottish Health Demonstration Projects. Evidence & Policy. 2006, 2 (2): 211-226.

    Article  Google Scholar 

  9. Judge K, Bauld L: Learning from Policy Failure. European Journal of Public Health. 2006, 16: 341-343. 10.1093/eurpub/ckl068.

    Article  PubMed  Google Scholar 

  10. Scottish Executive: Choose Life: A National Strategy and Action Plan to Prevent Suicide in Scotland. 2002, Edinburgh: The Stationery Office

    Google Scholar 

  11. Scottish Executive: Improving Health in Scotland – The Challenge. 2003, Edinburgh: The Stationery Office

    Google Scholar 

  12. Platt S, McLean J, McCollam A, Blamey A, Mackenzie M, McDaid D, Maxwell M, Halliday E, Woodhouse A: Evaluation of the first phase of Choose Life: the National Strategy and Action Plan to Prevent Suicide in Scotland. 2006, Edinburgh: Scottish Executive Social Research, []

    Google Scholar 

  13. Mann J, Apter A, Bertolote J, Beautrais A, Currier D, Haas A, Hegerl U, Lonnqvist J, Malone K, Marusic A, Mehlum L, Patton G, Phillips M, Rutz W, Rihmer Z, Schmidtke A, Shaffer D, Silverman M, Takahashi Y, Varnik A, Wasserman D, Yip P, Hendin H: Suicide prevention strategies: a systematic review. Journal of the American Medical Association. 2005, 294 (16): 2064-2074. 10.1001/jama.294.16.2064.

    Article  CAS  PubMed  Google Scholar 

  14. Martin S, Sanderson I: Evaluating public policy experiments: measuring outcomes, monitoring processes or managing pilots?. Evaluation. 1999, 5: 245-258. 10.1177/13563899922208977.

    Article  Google Scholar 

  15. Maynard R: Whether a sociologist, economist, psychologist or simply a skilled evaluator: lessons from evaluation practice in the United States. Evaluation. 1999, 6: 471-480. 10.1177/13563890022209433.

    Article  Google Scholar 

  16. Barnes M, Bauld L, Benzeval M, Judge K, Mackenzie M, Sullivan H: Health Action Zones: partnerships for health equity. 2005, London: Routledge

    Google Scholar 

  17. Connell J, Kubisch A: Applying a theory of change approach to the evaluation of comprehensive community initiatives: progress, prospects and problems. New approaches to evaluating community initiatives. Volume 2: theory, measurement, and analysis. Edited by: Fulbright-Anderson K, Kubisch A, Connell J. 1998, Washington DC: The Aspen Institute

    Google Scholar 

  18. Sridharan S, Campbell B, Zinzow H: Developing a stakeholder driven anticipated timeline of impact for evaluation of social programs. Amercian Journal of Evaluation. 2006, 27 (2): 148-162. 10.1177/1098214006287990.

    Article  Google Scholar 

  19. Rosen L, Manor O, Engelhard D, Zucker D: In defense of the randomised controlled trial for health promotion research. American Journal of Public Health. 2006, 96 (7): 1181-1186. 10.2105/AJPH.2004.061713.

    Article  PubMed  PubMed Central  Google Scholar 

  20. Connell JP, Kubisch A, Schorr LB, Weiss CH: New approaches to evaluating community initiatives. Volume 1: concepts, methods and contexts. 1995, Washington DC: The Aspen Institute

    Google Scholar 

  21. Pawson R, Tilley N: Realistic evaluation. 1997, London: Sage Publications

    Google Scholar 

  22. Fulbright-Anderson K, Kubisch A, Connell J, eds: New approaches to evaluating community initiatives. Volume 2: theory, measurement, and analysis. 1998, Washington DC: The Aspen Institute

    Google Scholar 

  23. Maris R: The relation of nonfatal suicide attempts to completed suicides. The Assessment and Prediction of Suicide. Edited by: Maris R, Berman A, Maltsberger J, Yufit R. 1992, New York: Guildford Press, 362-380.

    Google Scholar 

  24. Hawton K, Townsend E, Deeks J, Appleby L, Gunnell D, Bennewith O, Cooper J: Effects of change in legislation on pack size of paracetamol and salicylate on self-poisoning in the United Kingdom: before and after study. BMJ. 2001, 322: 1203-7. 10.1136/bmj.322.7296.1203.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  25. Bateman DN, Gorman DR, Bain M, Inglis JHC, House FR, Murphy D: Legislation restricting paracetamol sales and patterns of self-harm and death resulting from paracetamol-containing preparations in Scotland. British Journal of Clinical Pharmacology. (online early: doi:10.1111/j.1365-2125.2006.02668.x)

  26. McDaid D, Needle J: Economic Evaluation and Public Health. 2006, Cardiff: Welsh Assembly Government Health Promotion Division

    Google Scholar 

  27. Sefton T: Economic Evaluation in the Social Welfare Field: Making Ends Meet. Evaluation. 2003, 9 (1): 73-91. 10.1177/1356389003009001005.

    Article  Google Scholar 

  28. McDaid D, Halliday E, Mackenzie M, MacLean J, Maxwell M, McCollam A, Platt S, Woodhouse A: Issues in the economic evaluation of suicide prevention strategies: practical and methodological challenges. 2007, London: Personal Social Services Research Unit, London School of Economics and Political Science, Discussion Paper 2407

    Google Scholar 

  29. Kelly M, McDaid D, Ludbrooke A, Powell J: Economic appraisal pf public health interventions. Briefing Paper. 2005, London: Health Development Agency

    Google Scholar 

  30. Ganiats TG: Prevention, policy, and paradox: what is the value of future health?. American Journal of Preventive Medicine. 1997, 13 (1): 12-7.

    CAS  PubMed  Google Scholar 

  31. Sheldon TA: Discounting in health care decision-making: time for a change?. Journal of Public Health Medicine. 1992, 14 (3): 250-6.

    CAS  PubMed  Google Scholar 

  32. Evans D, Mays N, Wyke S: Evaluating complex policies: shat have we learned from total purchasing?. The Purchasing of Health Care from Primary Care Organisations. Edited by: Mays N, Wyke S, Malbon G, Goodwin N. 2001, Buckingham: Open University Press

    Google Scholar 

  33. Sullivan H, Barnes M, Matka E: Building collaborative capacity through "theories of change": early lessons from the evaluation of Health Action Zones in England. Evaluation. 2002, 8: 205-226. 10.1177/1358902002008002514.

    Article  Google Scholar 

  34. Barnes M, Sullivan H, Matka E: Evidence, understanding and complexity: evaluation in non-linear systems. Evaluation. 2003, 9: 263-282. 10.1177/13563890030093003.

    Article  Google Scholar 

  35. Mackenzie M, Blamey A: The practice and the theory: lessons from the application of a Theories of Change approach. Evaluation. 2005, 11 (2): 151-168. 10.1177/1356389005055538.

    Article  Google Scholar 

  36. Campbell M, Fitzpatrick R, Haines A, Kinmonth A, Sandercock P, Speigelhalter D, Tyrer P: Framework for design and evaluation of complex interventions to improve health. BMJ. 2000, 321: 694-696. 10.1136/bmj.321.7262.694.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  37. Bevan G, Hood C: Have targets improved performance in the English NHS?. BMJ. 1998, 332: 419-422. 10.1136/bmj.332.7538.419.

    Article  Google Scholar 

Pre-publication history

Download references


The Research Unit in Health, Behaviour and Change (RUHBC) (Stephen Platt) is funded by the Chief Scientist Office of the Scottish Executive Health Department (SEHD). SEHD also funded the evaluation of the first phase of Choose Life. The opinions expressed in this paper are those of the authors, not of SEHD.

Author information

Authors and Affiliations


Corresponding author

Correspondence to Mhairi Mackenzie.

Additional information

Competing interests

The author(s) declare that they have no competing interests.

Authors' contributions

All authors were involved in conception and design of the Choose Life evaluation or in data acquisition and interpretation. MM was the primary author of the paper with revisions provided by other authors, in particular, SP. All authors read and approved the final manuscript.

Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Mackenzie, M., Blamey, A., Halliday, E. et al. Measuring the tail of the dog that doesn't bark in the night: the case of the national evaluation of Choose Life (the national strategy and action plan to prevent suicide in Scotland). BMC Public Health 7, 146 (2007).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: