Skip to main content

A scoping review of evaluation frameworks and their applicability to real-world physical activity and dietary change programme evaluation

Abstract

Background

Physical activity and dietary change programmes play a central role in addressing public health priorities. Programme evaluation contributes to the evidence-base about these programmes; and helps justify and inform policy, programme and funding decisions. A range of evaluation frameworks have been published, but there is uncertainty about their usability and applicability to different programmes and evaluation objectives, and the extent to which they are appropriate for practitioner-led or researcher-led evaluation. This review appraises the frameworks that may be applicable to evaluation of physical activity and/or dietary change programmes, and develops a typology of the frameworks to help guide decision making by practitioners, commissioners and evaluators.

Methods

A scoping review approach was used. This included a systematic search and consultation with evaluation experts to identify evaluation frameworks and to develop a set of evaluation components to appraise them. Data related to each framework’s general characteristics and components were extracted. This was used to construct a typology of the frameworks based on their intended programme type, evaluation objective and format. Each framework was then mapped against the evaluation components to generate an overview of the guidance included within each framework.

Results

The review identified 71 frameworks. These were described variously in terms of purpose, content, or applicability to different programme contexts. The mapping of frameworks highlighted areas of overlap and strengths and limitations in the available guidance. Gaps within the frameworks which may warrant further development included guidance on participatory approaches, non-health and unanticipated outcomes, wider contextual and implementation factors, and sustainability.

Conclusions

Our typology and mapping signpost to frameworks where guidance on specific components can be found, where there is overlap, and where there are gaps in the guidance. Practitioners and evaluators can use these to identify, agree upon and apply appropriate frameworks. Researchers can use them to identify evaluation components where there is already guidance available and where further development may be useful. This should help focus research efforts where it is most needed and promote the uptake and use of evaluation frameworks in practice to improve the quality of evaluation and reporting.

Peer Review reports

Background

Programmes that aim to increase physical activity and improve dietary behaviours in individuals, groups and populations play a central role in addressing local, national and global public health priorities [1, 2]. Recent strategies have advocated approaches that are multi-sectorial, community-centred and evidence-based [1, 3,4,5]. Understanding if, when, and how these programmes are effective is important to justify policy, programme and funding decisions, and to inform and improve future decisions and practice. In order to achieve this, there is a need for appropriate and comprehensive programme evaluation [6, 7].

Practice-based evidence is generated from formal evaluation of programmes in real-world settings and is a fundamental part of evidence-based public health [8,9,10]. Those involved in the design, delivery and commissioning of physical activity and dietary change programmes are expected to evaluate programmes and contribute to the evidence base. However, real-world behaviour change programmes are complex and difficult to evaluate [11, 12]. The challenges of programme evaluation may relate to contextual factors that influence the complexity of the programme itself, e.g. its setting, target population, intervention function(s), or intended outcome(s) [12], or to factors that influence the evaluation priorities and objectives, e.g. differing stakeholder evaluation needs and organisational, political or resourcing factors [13]. Some of the practical challenges in conducting evaluation include the use of appropriate evaluation methods and tools, understanding what counts as evidence and how that is applied, and the roles of practitioners and researchers in evaluating real-world programmes [7, 9, 11, 14, 15].

Evaluation frameworks facilitate a systematic approach to evaluation and can help mitigate against some of the above challenges. Frameworks can enable multiple stakeholders to gain a shared understanding of the programme and evaluation process, and help to identify and agree upon appropriate objectives and methods. In this way, they can facilitate a more comprehensive evaluation, and may improve the fit between researcher-led and practitioner-led evaluation approaches [14]. A range of evaluation frameworks have been published. These include those developed specifically for use in programmes targeting specific health behaviours, conditions or populations (e.g. physical activity programmes [16,17,18]), those developed for health promotion and public health programmes more broadly (e.g. RE-AIM [19]), and generic frameworks intended to be applicable across a range of contexts, settings and sectors (e.g. Realist Evaluation [20]).

It is noteworthy that there is wide variation in the use of terminology used to describe frameworks, in the format of different frameworks, and in the context and ways in which they are intended to be used. Differentiating between frameworks, guidance, models or tools can be a challenge [21]. In this review the term ‘evaluation framework’ is used to include any structured guidance which facilitates a systematic evaluation of the implementation or outcomes of a programme. A ‘generic’ framework is used to refer to one that is intended for use across a range of contexts, settings and sectors, as opposed to one that has been developed for use in a specific context or field. Several frameworks have been developed for evaluation of programme implementation (process evaluation), whilst others focus on programme effectiveness (outcome evaluation) or are intended to facilitate an overall or comprehensive evaluation. In order to understand the content and focus of the frameworks and the contexts in which they may be applied, we have referred to the individual elements encompassed within evaluation as an “evaluation component”.

Many frameworks and developments in evaluation come from the research community, yet their intended audience and purpose is often unclear. For example, questions remain about the extent to which these frameworks are intended for use in practitioner-led or researcher-led evaluation, and their applicability to different evaluation objectives, programmes, and contexts.

Previous reviews of evaluation frameworks have been limited to frameworks which evaluate specific aspects of a programme, for example health inequalities [22], or methods used in health programme evaluations [23, 24]. Within the field of implementation science, reviews have focused on frameworks for translation of research to practice [25, 26]. The review by Denford et al. [27] made a valuable contribution by providing an overview of guidance available to support evaluation of public health programmes. However, it was limited to a subset of 48 documents created or sourced by national and international organisations and published since 2000. As a result some key evaluation frameworks published before 2000 or within the academic literature were not included, such as RE-AIM [19] and Realist Evaluation [20]. Denford et al. included various guidance documents intended for use in evaluating programmes targeting a broad range of health behaviours and health problems (e.g., smoking, asthma), as well as generic ones. Whilst they suggested that the wealth and breadth of available evaluation guidance may be a limiting factor in the ability of practitioners to access and apply appropriate guidance, the resulting review [27] and associated online catalogue [28] may still overwhelm practitioners seeking guidance on how to evaluate their specific programme.

To resolve some of this complexity we sought to develop a typology of frameworks, to help guide decision making by those involved in programme evaluation. The purpose was to appraise the frameworks that may be applicable for the evaluation of physical activity or dietary change programmes. By mapping the frameworks against a range of evaluation components (such as elements of process or outcome evaluation), we aimed to develop an overview of guidance included in each framework, enabling practitioners, commissioners and evaluators to identify and agree which frameworks may best meet their needs.

Objectives

1. To identify published frameworks that can be used for evaluation of physical activity and/or dietary change programmes.

2. To identify each framework’s stated scope in order to assess their applicability to different evaluation objectives, programmes and contexts.

3. To identify and map which evaluation components are encompassed within each framework.

4. To use the findings to develop a typology of frameworks.

Method

A scoping review approach was used, as this allowed the extent and nature of the literature on evaluation guidance to be identified and an overview of the available frameworks to be developed [29,30,31]. In line with the stages of a scoping review [29, 30], the process involved identification of the research question, a systematic search, consultation with experts, and mapping of the frameworks against different components of evaluation. We followed the PRISMA–ScR statement for the reporting of scoping reviews [32].

Search strategy

To identify any frameworks that could be applied to physical activity and/or dietary change programmes, we used a broad search strategy to find those intended for use in public health, health promotion and generic programmes as well as those developed specifically for use in evaluating physical activity and dietary change programmes. Firstly, a search was conducted in Scopus. As a meta-database, including records from MEDLINE and EMBASE as well as other sources, Scopus is the world’s largest abstract and citation database of peer-reviewed literature. It contains sources across a range of fields including medicine, sciences, humanities and social sciences. The following search strategy was used: (TITLE ((framework OR model OR guid* OR tool)) AND TITLE-ABS-KEY ((“physical activity” OR exercise OR diet OR obes* OR overweight OR “public health” OR “health promotion”)) AND TITLE-ABS-KEY (communit*) AND TITLE-ABS-KEY (evaluat*)). No date restriction was applied. The search was undertaken in March 2018. All sources identified from the search were downloaded into the Endnote reference manager, and any duplicates were removed.

Secondly, between April and September 2018, we searched for grey literature on the websites of key organisations interested in evaluation of physical activity and/or dietary change programmes, using “evaluation framework” as a search term. This included the World Health Organization (WHO), Public Health England (PHE), Sport England, and the Centers for Disease Control and Prevention (CDC). Additional sources were identified from the authors’ existing files. We consulted evaluation experts and stakeholders including academics, those involved in public health policy development and evaluation, and evaluation consultants within the domains of physical activity or dietary change, to augment the search results. These experts were contacted and asked to provide feedback on the list of frameworks we had identified by the search strategy and to identify any omissions. Reference lists were examined for additional relevant sources.

Sources were screened by title and abstract, and then by full text (JF). Full text screening was independently validated (KM) and disagreements resolved through discussion. Consensus could not be reached for six sources, which were checked by a third reviewer (AJ) and agreed through further discussion.

Inclusion and exclusion criteria

Inclusion and exclusion criteria were defined a priori and applied to all sources (JF). Table 1 provides details of the full inclusion and exclusion criteria. Sources were included from both the academic and grey literature that described a framework to support systematic evaluation of a physical activity and/or dietary change programme, including generic, public health or health promotion frameworks applicable to physical activity or dietary change programmes. Academic literature included journal articles and books. Grey literature was defined as all other printed and electronic documents published by organisations and agencies. Web-based sources were included if they provided systematic guidance on how to conduct an evaluation but excluded if they were an organisation’s general website without guidance. Only sources in English were included.

Table 1 Inclusion and Exclusion Criteria

Data extraction and synthesis

To address the first and second objective, a data extraction template was used to collate information about each framework. The name of each framework was identified. Where no framework name was provided in the source, a short name was given based on the authors’ description in the title or abstract. To assess each framework’s scope and applicability to the evaluation of physical activity and/or dietary change programmes, data extraction fields included the stated evaluation objective, the types of programme it was intended for, and additional data related to general characteristics of each framework, e.g. its intended audience, format and development process.

To address the third objective we developed a set of data extraction fields to enable us to appraise whether each framework provided any guidance on a range of evaluation components, and what that guidance comprised. We have used the term ‘evaluation component’ to refer to individual elements encompassed within evaluation; for example elements of process or outcome evaluation. The list of evaluation components included in the data extraction template was identified a priori, and developed through a process of consensus building. We initially identified a list of evaluation components that were informed by recommendations for good practice in the evaluation literature, for example implementation, reach and unanticipated outcomes [12, 33,34,35]. This was further developed through consultation with evaluation experts, who were contacted and asked to comment on the appropriateness of the evaluation components we had identified and to identify any gaps or additional components based on their personal experience and knowledge of programme evaluation. Table 2 shows the full list of evaluation components grouped into those related to: (1) process evaluation, (2) outcome evaluation and (3) study design. Grouping programme context, theory of change and logic models within process evaluation components aligns with its inclusion in the UK Medical Research Council (MRC) Process Evaluation guidance [35], and recognises the crucial role of logic models in the early stages of developing an evaluation plan, in reporting causal assumptions about how a programme works, and informing process and outcome questions and methods. Where possible, pre-defined categorical responses were developed to facilitate the data extraction, coding and synthesis.

Table 2 Evaluation Components Agreed for Data Extraction and Mapping of Frameworks

Where authors had described the scope of a framework variably, and where terms were not mutually exclusive, multiple terms were noted in the data extraction table. For example, terms such as community or practice based were used interchangeably to describe a study, intervention, setting or population. Where frameworks gave more detailed guidance on specific evaluation components, we also extracted a summary of what the guidance comprised. For each evaluation component we assessed whether the framework simply mentioned or provided more detailed guidance on how to evaluate or break down the relevant component.

Data extraction was completed by JF. To verify the data extraction, a random sample of 20 sources was checked independently by AJ and WH. Differences were resolved through discussion and used to establish agreed definitions that were then applied to further data extraction.

Framework format, programme type and evaluation objectives are typically used to describe frameworks. We therefore used these aspects to develop our typology for the frameworks. For the purposes of categorising the frameworks within the typology we used the dominant term presented in the description and content of the source as the basis for identifying each framework’s most defining characteristic. The extracted data was also used to map each framework against the evaluation components in order to provide an overview of the guidance encompassed within the frameworks. A narrative synthesis of the findings is presented.

Results

Study selection

The initial search in Scopus yielded 1604 sources once duplicates were removed. An additional 24 sources were identified from the grey literature search and consultation process, and a further 60 sources were identified from reference lists. Many articles were identified as ineligible from their title alone, mostly because they related to conceptual models, treatment models, or conditions not relevant to physical activity or diet. If there was any uncertainty regarding the potential eligibility of a paper, it was included in the next stage of the screening process. After screening of titles and abstracts 168 full-text sources were assessed for eligibility (PRISMA diagram, Fig. 1).

Fig. 1
figure 1

PRISMA diagram of screening process

At full-text screening 83 sources were included and 85 were excluded. Of those excluded, 37 were reported evaluation studies that used one or more framework(s) and three were sources that critically appraised framework(s) [36,37,38]. The reference lists of these sources were searched to identify the index papers that described the frameworks mentioned.

Sources which described programme and evaluation practices in general terms, e.g. [39], and those which described a specific measurement tool, e.g. photovoice [40] and memorable messages [41] were excluded. Other sources were also excluded if they reported a framework linked to a specific intervention and in such a way that it was not generalisable (e.g. Framework for Washington State’s Healthy Communities Projects [42]). Planning frameworks that were solely for guidance on the design and development of an intervention were also excluded (e.g. [43,44,45], but a number were retained where they included guidance related to evaluation [46,47,48,49,50].

For frameworks which were described in more than one publication, for example in full and summary articles, we included both sources to facilitate data extraction and analysis, e.g. PRECEED-PROCEDE [46, 51], the CDC Framework [15, 52], UK MRC Guidance [12, 53,54,55], and Impact Pathway Analysis [56, 57]. Data were extracted from 83 sources, describing 71 evaluation frameworks.

Identification of the evaluation frameworks available

A brief description of each framework is provided in Additional File 1 and an overview of their general characteristics is provided in Additional File 2. Table 3 lists the frameworks included in the review, grouped by decade of publication and source (academic/grey literature). All included frameworks were published during the last three decades (1990 onwards). Forty-two were described in academic publications and 23 in the grey literature. Six frameworks were reported in both the grey and academic literature [35, 52,53,54,55,56,57, 114,115,116].

Table 3 Included frameworks grouped by decade of publication and source

Table 3 also indicates the format of each framework. This ranged from highly structured to more flexible guidance. Thirty of the frameworks were presented as a set of steps; typically, these steps align with the stages of programme development and implementation. Twenty-four frameworks were presented as a set of indicators or questions, ranging from those that included a small number of key indicators [19, 79, 81, 93] to those that encompassed a longer checklist of evaluation criteria or questions [16,17,18, 112, 117]. The remaining 17 provided flexible evaluation guidance.

Sources generally described the framework development as being based on (i) some combination of literature review, consultation and testing, (ii) experiences of conducting evaluation(s), or (iii) prior frameworks or theory. Many of the more recently published frameworks referred to earlier ones as informing their development, such as realist evaluation [20], utilization-focused evaluation [61], PRECEDE-PROCEED [46] and intervention mapping [63]. Several frameworks formatted as a set of steps mentioned the CDC framework [52] and other step-based frameworks [73, 75] as informing their development. Several frameworks formatted as a checklist referred to RE-AIM [19] as informing the indicators.

Seventeen frameworks provided guidance or links to sources for additional support or training in using the framework. Those that gave more detailed guidance of training and support, including links to additional resources, were predominantly published within the grey literature and had an online presence [95, 101, 103, 104, 107].

Scope of the evaluation frameworks and development of a typology

There was considerable heterogeneity in the terminology used to describe the scope of the frameworks. Authors described them variously in terms of purpose, content, or applicability to different programme and/or evaluation contexts. Additional File 2 shows the range of the descriptors used by authors. For example, thirty-one sources mentioned the frameworks were intended for use in real world or practice-based settings, and 22 were intended for use in community-based programmes, with these terms often used interchangeably. Others were described as applicable to specific intervention functions (e.g. health education [117] or policy [77, 94, 105]), or specific intervention or study types (e.g. complex interventions [44, 54, 84], natural experiments [114] or cluster randomised trials [91]). These terms were not mutually exclusive so were not used to categorise the frameworks and develop the typology but are indicated within Additional file 2.

Programme type

Despite this variability in descriptors used by authors, we used the intended programme type as the primary categorisation to develop the typology, followed by the evaluation objective and the framework format. These characteristics enabled us to group the frameworks by applying the dominant description provided by the authors as an indication of a framework’s most defining characteristics. Figures 2a-c show the typology which signposts to each framework within the categories.

Fig. 2
figure 2

a Typology of evaluation frameworks intended for use in physical activity, dietary change or behaviour change programmes. b Typology of frameworks intended for use in health promotion or public health programmes. c Typology of frameworks intended for use in generic programmes

Twelve frameworks were stated as intended for use in physical activity and/or dietary change programme evaluation, and one as for use in behaviour change interventions [96] (Fig. 2a). Forty-eight were described as for use in public health or health promotion programmes. Some of these clearly stated how their components related to health promotion principles. However, several used the terms health promotion and public health interchangeably, and these were therefore grouped together (Fig. 2b). A further ten frameworks were described as applicable to a range of programme types and we have grouped these as intended for generic programme evaluation (Fig. 2c).

Evaluation objective

Frameworks were also described variously in terms of their evaluation focus or objective, and we used this to further develop the typology shown in Figs. 2a-c. Fifty-two were stated as providing guidance on overall programme evaluation, nine as specific to process evaluation and one as specific to outcome evaluation. Several of the frameworks provided guidance on evaluating specific programme elements such as empowerment [83], partnerships and participation [68, 78, 80, 81, 87, 92], contextual factors [50], or legacy [76]. Four frameworks were described as ‘planning frameworks’ but incorporated guidance on evaluation [46,47,48,49]; these are grouped separately within the typology (Figs. 2a-c). Other frameworks that included guidance to facilitate both evaluation and planning, but were not specifically described as ‘planning frameworks’, e.g. [50] are not grouped separately.

Mapping frameworks against evaluation components

Frameworks were mapped against seven process and four outcome evaluation components (i.e. describing programme context, using theory of change, logic models, reach, implementation, maintenance, any other process measures, behaviour, health, non-health and unanticipated outcomes), as well as against the eight components of study design and reporting (see Table 2). Tables 4a-c and 5a-c provide an overview of the mapping. Describing programme context, theory of change, and logic models are crucial to informing process and outcome evaluation, we therefore included these alongside process evaluation components in Table 4a-c. The mapping enabled us to develop an overview of the guidance included in each of the frameworks and appraise their applicability to different evaluation objectives and to physical activity and/or dietary change programmes.

Table 4 a Frameworks intended for use in physical activity, dietary change or behaviour change programmes mapped against process and outcome evaluation components. Light grey shading indicates the component is mentioned, dark grey shading indicates more detailed guidance on how to break down or evaluate the component. b. Frameworks intended for use in evaluating public health and health promotion programmes mapped against process and outcome evaluation components. Light grey shading indicates the component is mentioned, dark grey shading indicates more detailed guidance on how to break down or evaluate the component. c Frameworks intended for use in evaluating generic programmes mapped against process and outcome evaluation components. Light grey shading indicates the component is mentioned, dark grey shading indicates more detailed guidance on how to break down or evaluate the component.
Table 5 a Frameworks intended for use in evaluating physical activity, dietary change or behaviour change programmes mapped against study design, evaluation approach and reporting components. Light grey shading indicates the component is mentioned, dark grey shading indicates more detailed guidance on how to break down or evaluate the component. b Frameworks intended for use in evaluating public health and health promotion programmes mapped against study design, evaluation approach and reporting components. Light grey shading indicates the component is mentioned, dark grey shading indicates more detailed guidance on how to break down or evaluate the component. c Frameworks intended for use in evaluating generic programmes mapped against study design, evaluation approach and reporting components. Light grey shading indicates the component is mentioned, dark grey shading indicates more detailed guidance on how to break down or evaluate the component.

Many frameworks mentioned components without any further details (shaded in light grey in the tables), whilst others provided detailed descriptions of how the components may be broken down or evaluated (shaded in dark grey in the tables). For ease of navigation, the frameworks in Tables 4a-c and 5a-c are grouped and listed in the same order as in the typology (Figs. 2a-c). Most frameworks included guidance on a range of both process and outcome evaluation components. Eleven frameworks did not provide any guidance on outcome evaluation and were specific to process evaluation e.g. [73,74,75, 81]. Frameworks intended to facilitate evaluation of specific programme elements focused on a narrower range of components that aligned with their stated purpose [50, 76, 78, 80, 93].

Process evaluation components

Guidance on the key components of process evaluation were included in most frameworks, e.g. describing contextual factors of programmes, identifying and describing causal mechanisms or theories of change, reach and implementation. The frameworks providing the most comprehensive and detailed guidance on these components include the MRC guidance on process evaluation of complex interventions [12], Center of Excellence for Training and Research Translation (Center TRT) Framework [105], Victoria Government Department of Health (DoH) Evaluation Framework [111], the Physical Activity Evaluation Handbook [102] and the Standard Evaluation Frameworks (SEFs) [16,17,18]. Other process evaluation components were included within fewer frameworks. For example, guidance on evaluation of sustainability was limited, with only thirteen frameworks providing more details of how to evaluate it, e.g. [76, 93]. A small number of frameworks mentioned other process components such as adaptation, exposure, capacities, training, partnerships, satisfaction, and community changes; however, details of how to evaluate these components were limited. Over half the frameworks identified logic models as a useful tool in programme planning and evaluation. Several of these provide more detailed information, examples and/or templates to support the development of logic models [12, 48, 95, 108].

Outcome evaluation components

Guidance on outcome evaluation components was more variable than for process evaluation components. Frameworks designed for use in physical activity and/or dietary change related programmes provided more detailed information on evaluation of behavioural and health outcomes than the more generic evaluation frameworks. Evaluation of non-health outcomes was typically only mentioned briefly in the frameworks, with only seven providing any level of detail [67, 68, 76, 80, 102, 109, 111]. Only about one third of the frameworks mentioned evaluation of unanticipated outcomes, and none provided further information on how to evaluate them.

Study design components

Tables 5a-c shows the frameworks mapped against components related to study design, including evaluation at different time points, stakeholder involvement, participatory approaches, data collection and analysis, and reporting of findings. Most frameworks identified the importance of stakeholder involvement and/or participatory evaluation approaches. Few provided information on how to incorporate this, with a few exceptions that did provide detailed guidance on participatory evaluation methods [56, 57, 68, 69, 79].

Most frameworks mentioned the importance of conducting evaluation that is appropriate to a programme’s stage of development, and many were presented as a set of steps aligned to stages of programme development and implementation. Most also mentioned evaluation at different time points (i.e. baseline and follow-up), mainly in relation to outcome measures only. Several frameworks used the terms formative and summative evaluation but gave limited information on how they were defining them, or how to do these types of evaluation. Exceptions to this were frameworks that gave a more detailed explanation of the role of formative and pilot studies in developing an intervention [33, 53].

Guidance on data collection and data analysis was highly variable. Several frameworks provided explanations of appropriate use of experimental designs and quantitative and qualitative methods [20, 46, 54, 75]. Others provided more detailed guidance on specific data collection methods and measures [16,17,18, 33, 72, 86, 100, 110]. Only thirteen frameworks provided information to guide data analysis. There was more consistency in the inclusion of guidance on data collection and analysis within the frameworks described as specific to physical activity and/or dietary change programmes than in the other categories of frameworks.

Finally, guidance on dissemination and reporting also varied. Many frameworks mentioned the importance of this aspect within the cycle of evidence-based practice, but few provided information about where and how to report findings to different target audiences.

Discussion

Our scoping review identified 71 evaluation frameworks, considerably more than previous reviews of evaluation frameworks within the field of public health [25,26,27]. The broad search strategy we applied enabled us to identify frameworks developed within a range of domains that we could add to those included in these earlier reviews. The focused set of inclusion and exclusion criteria we then applied meant that we only included frameworks specific to or generalisable to physical activity and/or dietary change programmes. In addition to the 12 frameworks specifically intended for physical activity and/or dietary change programme evaluation, we identified a further 59 intended for public health, health promotion, behaviour change or generic programmes that were applicable to physical activity and/or dietary change programmes.

Our review has highlighted the plethora of frameworks available; previous reviews [27] reported this as a potential challenge to practitioners and evaluators navigating and making use of the available guidance. Our review also highlighted the variability in terms used by authors to describe the purpose and scope of the frameworks. Although we identified a growing number of frameworks developed by and for practitioners, e.g. [102, 103, 106, 107, 111], in many frameworks the intended audience was unclear. Terms used to describe programme types were poorly defined and were often used interchangeably. Some phrases such as ‘natural experiment’ and ‘real-world’ were used to refer to the evaluation approach and the intervention itself, whilst others (e.g. behaviour change and sustainability) were used to refer to both intervention processes and outcomes. Several frameworks which stated they were intended to support both programme planning and evaluation provided insufficient details about how these facilitated evaluation. The lack of clarity in the extent to which frameworks are intended to be used by researcher-led or practitioner-led evaluation, and in their applicability to different programmes and evaluation objectives, has implications for those using the available guidance. There needs to be a greater consensus of how terms are defined within public health evaluation. An agreed common language would enable those involved in programme evaluation to understand more clearly the applicability of the different frameworks and would help this research area to move forward.

Our typology and mapping resolves some of that complexity in purpose and scope of frameworks by signposting to relevant frameworks and by developing an overview of what guidance is encompassed within each. Our appraisal of frameworks has highlighted areas of overlap, strengths and limitations in the guidance available to support programme evaluation. For example, the inclusion of key process evaluation components (e.g. describing programme contexts and causal mechanisms, reach, and use of logic models) in most frameworks reflects the growing understanding of the importance of these aspects of evaluation to facilitate a more detailed understanding of whether and how a programme works [7, 33,34,35, 118]. These components represent strengths within the existing guidance, and areas where there is already an abundance of guidance.

The mapping process and appraisal also identified components where more guidance would be beneficial. We found limited guidance on participatory approaches, non-health and unanticipated outcomes, and wider programme components (e.g. resources, training, delivery, adaptation, partnerships, organisational structures), and sustainability. These components represent aspects of evaluation that require further development of guidance. Stakeholder involvement or participatory evaluation was mentioned in all but nine of the frameworks, reflecting the growing recognition of the importance of stakeholder engagement in evaluation decisions and processes [34, 84]. However, detailed guidance on how to incorporate participatory evaluation methods was only provided by seven frameworks [34, 56, 64, 68, 73, 80, 81], and represents another area where further development of guidance would be beneficial. Compared to other categories within the typology, frameworks specific to physical activity programmes more consistently provided guidance on evaluation of health and behavioural outcomes, including the use of appropriate data collection and analysis methods. By their nature these components are specific and therefore may be difficult to define within more generic frameworks. Frameworks developed to facilitate evaluation of specific programme elements, such as sustainability [76, 93], and those intended to facilitate evaluation of partnerships [78, 80, 92] or community [68, 69, 80] also addressed some of the gaps within the more generic frameworks.

Our mapping and typology signpost to frameworks where guidance on specific components can be found. Although availability does not necessarily equate to accessibility or usability of information, the mapping of frameworks can be used to help understand some of the strengths and limitations within the guidance provided. Further investigation of whether and how frameworks have been used may provide insight into how fit for purpose they are, and the benefits and challenges of applying them within physical activity or dietary change programme evaluation. Furthermore, the typology and mapping can be used by practitioners, commissioners and evaluators of physical activity and/or dietary change programmes to identify frameworks relevant to their evaluation needs. They can also be used by researchers and those interested in developing evaluation guidance to identify evaluation components where it would be most useful to focus their efforts, rather than developing more guidance for components where there is already an abundance of guidance. Our categorisation could also be used by researchers publishing frameworks to more clearly report how these are intended to be used, and for those reporting evaluation studies to more clearly state how they have been used.

Strengths and limitations

Our broad search strategy enabled a comprehensive review which identified 71 frameworks within the academic and grey literature. By drawing on frameworks developed within different domains, we have added to previous reviews [25, 27] to map a wide range of evaluation frameworks applicable to physical activity and/or dietary change programmes.

Our scoping review methods, which included consultation with experts, helped to maximise the chances of identifying relevant frameworks, and of applying relevant components which were based on consensus to appraise the frameworks. It was not our intention to apply a formal consensus building method, however we recognise that the use of a more formalised process would be an alternative approach. By consulting both practice and research-based experts we are confident that the results will be of interest and value to both practitioners and researchers concerned with evaluation of physical activity and/or dietary change programmes.

There are limitations of the review. The review only included sources published in the English language. The heterogeneity and ambiguity in use of terminology was a methodological challenge during screening, data extraction and synthesis. Frameworks intended to support specialist evaluation aspects such as health economic evaluation and evaluation of programmes using digital technologies (e.g., mobile health) are critical to practice and policy decisions, however we excluded these frameworks due to their specificity and also due to the large number available. A separate review of the available guidance to support these specialist evaluation aspects would be beneficial.

Conclusion

We have added to previous reviews of evaluation frameworks, and identified 71 frameworks applicable to physical activity and/or dietary change programme evaluation. There is an abundance of frameworks available to support programme evaluation. Our typology and mapping signpost to frameworks where guidance on specific components can be found, where there is overlap in their scope and content, and where there are gaps in the guidance. Practitioners and evaluators can use the typology and mapping to identify, agree upon and apply appropriate frameworks. Researchers who develop evaluation guidance can use them to identify evaluation components for which there are gaps in available guidance. This should help focus research efforts where it is most needed and promote uptake and use of appropriate evaluation frameworks in practice to improve the quality of evaluation and reporting.

Availability of data and materials

All data generated or analysed during this study are included in this published article [and its supplementary information files]. All data used in this manuscript is in the public domain.

Abbreviations

CDC:

Center for Disease Control and Prevention

Center TRT:

Center of Excellence for Training and Research Translation

DoH:

Department of Health

GPAT:

Good Practice Appraisal Tool

HEBS:

Health Education Board Scotland

HP:

Health Promotion

LEAP:

Learning, Evaluation and Planning

MRC:

Medical Research Council

MMIPP:

Model for Management of Intervention Programme Planning

NICE:

National Institute for Health and Care Excellence

OPEN:

Obesity Prevention through the European Network

PIP:

Programme Impact Pathway

PIPA:

Participatory Impact Pathway Analysis

PA:

Physical Activity

PH:

Public Health

PHE:

Public Health England

RE-AIM:

Reach, Efficacy, Adoption, Implementation, Maintenance

SEF:

Standard Evaluation Framework

WHO:

World Health Organization

References

  1. Public Health England. Everybody Active, Every Day: An evidence-based approach to physical activity. London: Public Health England; 2014.

  2. World Health Organization. Global action plan on physical activity 2018–2030: more active people for a healthier world. Geneva: World Health Organization; 2018.

    Google Scholar 

  3. National Institute for Health and Care Excellence (NICE). Obesity: Working with local communities. London: NICE; 2012.

  4. England S. Towards an active nation. London: Sport England; 2016.

    Google Scholar 

  5. HM Government. Sporting future: a new strategy for an active nation. London: Cabinet Office; 2015.

    Google Scholar 

  6. Public Health England. A Guide to Community-Centred Approaches to Health and Well-being Full Report. London; 2015. Contract No.: 2014711.

  7. Lobo R, Petrich M, Burns SK. Supporting health promotion practitioners to undertake evaluation for program development. BMC Public Health. 2014;14:1315.

    Article  PubMed  PubMed Central  Google Scholar 

  8. Honeycutt S, Hermstad A, Carvalho ML, Arriola KRJ, Ballard D, Escoffery C, et al. Practice to evidence: using Evaluability assessment to generate practice-based evidence in rural South Georgia. Health Educ Behav. 2017;44(3):454–62.

    Article  PubMed  Google Scholar 

  9. Li V, Carter SM, Rychetnik L. Evidence valued and used by health promotion practitioners. Health Educ Res. 2015;2:193.

    Article  Google Scholar 

  10. Lee RG, Garvin T. Moving from information transfer to information exchange in health and health care. Soc Sci Med. 2003;56:449–64.

    Article  PubMed  Google Scholar 

  11. Rychetnik L, Bauman A, Laws R, King L, Rissel C, Nutbeam D, et al. Translating research for evidence-based public health: key concepts and future directions. J Epidemiol Community Health. 2012;66(12):1187–92.

    Article  PubMed  Google Scholar 

  12. Moore G, Audrey S, Barker M, Bond L, Bonell C, Cooper C, et al. Process evaluation in complex public health intervention studies: the need for guidance. J Epidemiol Community Health. 2015;68(2):101–2.

    Article  Google Scholar 

  13. Habicht JP, Victora CG, Vaughan JP. Evaluation designs for adequacy, plausibility and probability of public health programme performance and impact. Int J Epidemiol. 1999;28(1):10–8.

    Article  CAS  PubMed  Google Scholar 

  14. Wimbush E, Watson J. An Evaluation framework for health promotion: theory. Quality and Effectiveness Evaluation. 2000;6(3):301–21.

    Google Scholar 

  15. Milstein B, Wetterhall S. A framework featuring steps and standards for program Evaluation. Health Promot Pract. 2000;1(3):221–8.

    Article  Google Scholar 

  16. Cavill N, Roberts K, Rutter H. Standard Evaluation framework for physical activity interventions. Oxford: National Obesity Observatory; 2012.

    Google Scholar 

  17. Cavill N, Roberts K, Rutter H. Standard Evaluation framework for dietary interventions. Oxford: National Obesity Observatory; 2012.

    Google Scholar 

  18. Roberts K, Cavill N, Rutter H. Standard Evaluation framework for weight management interventions. London: National Obesity Observatory; 2009.

    Google Scholar 

  19. Glasgow RE, Vogt TM, Boles SM. Evaluating the public health impact of health promotion interventions: the RE-AIM framework. Am J Public Health. 1999;89(9):1322–7.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  20. Pawson R, Tilley N. Realistic Evaluation. London: Sage Publications Ltd; 1997.

    Google Scholar 

  21. Nilsen P. Making sense of implementation theories, models and frameworks. Implement Sci. 2015;10.

  22. Davies JK, Sherriff N. The gradient in health inequalities among families and children: a review of evaluation frameworks. Health Policy. 2011;101(1):1–10.

    Article  PubMed  Google Scholar 

  23. Chambers AH, Kolbe A, Murphy K. Designs and methods used in published Australian health promotion evaluations. Aust N Z J Public Health. 1992-2011;2015:222–6.

    Google Scholar 

  24. Lim MSC, Wright CJC, Carrotte ER, Pedrana AE. Reach, engagement, and effectiveness: a systematic review of evaluation methodologies used in health promotion via social networking sites. Health Promotion Journal of Australia. 2016;27(3):187–97.

    Article  PubMed  Google Scholar 

  25. Milat A, Li B. Narrative review of frameworks for translating research evidence into policy and practice. Public Health Research & Practice. 2017;27(1):e2711704.

    Article  Google Scholar 

  26. Escoffery C, Lebow-Skelley E, Udelson H, Böing EA, Wood R, Fernandez ME, et al. A scoping study of frameworks for adapting public health evidence-based interventions. Transl Behav Med. 2018;9(1):1–10.

    Article  PubMed Central  Google Scholar 

  27. Denford S, Abraham C, Callaghan M, Aighton P, De Vocht F, Arris S. A review of Grey and academic literature of evaluation guiance relevant to public health interventions. BMC Health Serv Res. 2017;17:643.

    Article  PubMed  PubMed Central  Google Scholar 

  28. Evaluation in Health and Wellbeing: Guidance Summaries [Internet]. Public Health England. 2018 [cited 13 February 2019]. Available from: https://www.gov.uk/government/publications/evaluation-in-health-and-wellbeing-guidance-summaries/evaluation-in-health-and-well-being-guidance-summaries.

  29. Arksey H, O'Malley L. Scoping studies: towards a methodological framework. Int J Soc Res Methodol. 2005;8(1):19–32.

    Article  Google Scholar 

  30. Levac D, Colquhoun H, O'Brien KK. Scoping studies: advancing the methodology. Implement Sci. 2010;5:69.

    Article  PubMed  PubMed Central  Google Scholar 

  31. Grant MJ, Booth A. A typology of reviews: an analysis of 14 review types and associated methodologies. Health Information & Libraries Journal. 2009;26(2):91–108.

    Article  Google Scholar 

  32. Tricco AC, Lillie E, Zarin W, K.K. OB, Colquhoun H, Levac D. PRISMA Extension for Scoping Reviews (PRISMAScR): Checklist and Explanation. Ann Intern Med. 2018;169:467–73.

  33. Bauman A, Nutbeam D. Evaluation in a nutshell: McGraw-hill education Australia; 2013.

    Google Scholar 

  34. Judd J, Frankish CJ, Moulton G. Setting standards in the evaluation of community-based health promotion programmes— a unifying approach. Health Promot Int. 2001;16(4):367–80.

    Article  CAS  PubMed  Google Scholar 

  35. Moore G, Audrey S, Barker M, Bond L, Bonell C. Hardeman W, et al. UK Medical Research Council (MRC) Gudiance: Process Evaluation of Complex Interventions; 2015.

    Google Scholar 

  36. Vatcharavongvan P, Hepworth J, Marley J. The application of the parallel track model in community health promotion: a literature review. Health and Social Care in the Community. 2013;21(4):352–63.

    Article  PubMed  Google Scholar 

  37. Dauphinee WD. The role of theory-based outcome frameworks in program evaluation: considering the case of contribution analysis. Medical Teacher. 2015;37(11):979–82.

    Article  PubMed  Google Scholar 

  38. Ory MG, Altpeter M, Belza B, Helduser J, Zhang C, Smith ML. Perceived utility of the RE-AIM framework for health promotion/disease prevention initiatives for older adults: A case study from the U.S. evidence-based disease prevention initiative. Frontiers in Public Health. 2015;2(143).

  39. McLeroy KR, Bibeau D, Steckler A. Glanz KJHeq. An ecological perspective on health promotion programs. 1988;15(4):351–77.

    CAS  Google Scholar 

  40. Kramer L, Schwartz P, Cheadle A, Rauzon S. Using Photovoice as a participatory Evaluation tool in Kaiser Permanente's community health initiative. Health Promot Pract. 2013;14(5):686–94.

    Article  PubMed  Google Scholar 

  41. Davis LA, Morgan SE, Mobley AR. The utility of the memorable messages framework as an intermediary Evaluation tool for fruit and vegetable consumption in a nutrition education program. Health Education and Behavior. 2015;43(3):321–7.

    Article  PubMed  Google Scholar 

  42. Smith LT, Johnson DB, Lamson E, Sitaker M. A framework for developing evaluation tools used in Washington State's Healthy Communities projects. Preventing Chronic Disease. 2006;3(2).

  43. Lubans DR, Lonsdale C, Cohen K, Eather N, Beauchamp MR, Morgan PJ, et al. Framework for the design and delivery of organized physical activity sessions for children and adolescents: Rationale and description of the 'SAAFE' teaching principles. International Journal of Behavioral Nutrition and Physical Activity. 2017;14(24).

  44. Angeles RN, Dolovich L, Kaczorowski J, Thabane L. Developing a theoretical framework for complex community-based interventions. Health Promot Pract. 2014;15(1):100–8.

    Article  PubMed  Google Scholar 

  45. Laverack G, Labonte R. A planning framework for community empowerment goals within health promotion. Health Policy Plan. 2000;15(3):255–62.

    Article  CAS  PubMed  Google Scholar 

  46. Green LW, Kreuter MW. Health Promotion Planning : An Educational and Environmental Approach 2nd ed. Mountain View, Calif.: Mayfield Publishing Co.; 1991.

  47. Bartholomew LK, G.S. P, Kok G, Gottlieb NH. Planning Health Promotion Programs : An Intervention Mapping Approach: John Wiley & Sons, Incorporated; 2006.

  48. W K Kellogg Foundation. Logic Model Development Guide. Battle Creek Michigan: WK Kellogg Foundation; 2004.

  49. Poland B, Krupa G, McCall D. Settings for health promotion: an analytic framework to guide intervention design and implementation. Health Promot Pract. 2009;10(4):505–16.

    Article  PubMed  Google Scholar 

  50. Vanderkruik R, McPherson ME. A contextual factors framework to inform implementation and Evaluation of public health initiatives. Am J Eval. 2017;38(3):348–59.

    Article  Google Scholar 

  51. Crosby R, Noar SM. What is a planning model? An introduction to PRECEDE-PROCEED. J Public Health Dent. 2011;71(SUPPL. 1):S7–S15.

    Article  PubMed  Google Scholar 

  52. Centers for Disease Control and Prevention (CDC). Framework for program evaluation in public health. 1999;48(RR-11).

  53. Craig P, Dieppe P, Macintyre S, Michie S, Nazareth I, Petticrew M. Developing and evaluating complex interventions: new guidance. Medical Research Council; 2008.

    Google Scholar 

  54. Craig P, Dieppe P, Macintyre S, Michie S, Nazareth I, Petticrew M. Developing and evaluating complex interventions: The new Medical Research Council guidance. British Medical Journal. 2008;337( e1655).

  55. Moore G, Audrey S, Barker M, Bond L, Bonell C, Cooper C, et al. Process evaluation in complex public health intervention studies. British Medical Journal. 2015;2015, 350:h1258(2):101–2.

  56. Douthwaite B, Alvarez S, Thiele G, Mackay R. Participatory Impact Pathway Analysis: A practical method for project planning and evaluation. The Institutional Learning and Change (ILAC); 2008. Contract No.: Brief 17 May 2008.

  57. Douthwaite B. Alvarez S. Participatory Impact Pathways Analysis Manual. Available from: http://pipamethodology.pbworks.com/w/page/70345685/PIPA%20Manual.

  58. Israel BA, Cummings KM, Dignan MB, Heaney CA, Perales DP, Simons-Morton BG, et al. Evaluation of health education programs: current assessment and future directions. Health Educ Q. 1995;22(3):364–89.

    Article  CAS  PubMed  Google Scholar 

  59. Poland BD. Knowledge development and evaluation in, of and for healthy community initiatives. Part I: guiding principles. Health Promot Int. 1996;11(3):237–47.

    Article  Google Scholar 

  60. Hawe P, Dageling D, Hall J. Evaluating health promotion: a health workers guide. Maclennan & Petty Pty Ltd: Sydney; 1990.

    Google Scholar 

  61. Patton M. Utilization-focused evaluation: the new century text. London SAGE. 1997.

  62. Brännström IA, Persson LÅM, Wall SI. Towards a framework for outcome assessment of health intervention: conceptual and methodological considerations. Eur J Pub Health. 1994;4(2):125–30.

    Article  Google Scholar 

  63. Bartholomew LK, Parcel GS, Kok G. Intervention mapping: a process for developing theory- and evidence-based health education programs. Health Educ Behav. 1998;25.

  64. Sanderson C, Haglund BJA, Tillgren P, Svanström L, Östenson CG, Holm LE, et al. Effect and stage models in community intervention programmes; and the development of the model for management of intervention programme preparation (MMIPP). Health Promot Int. 1996;11(2):143–56.

    Article  Google Scholar 

  65. Nutbeam D. Evaluating health promotion—Progress. Problems and solutions Health Promotion International. 1998;13(1):27–44.

    Article  Google Scholar 

  66. Goodman RM. Principles and tools for evaluating community-based prevention and health promotion programs. Journal of Public Health Management and Practice. 1998;4(2):37–47.

    Article  CAS  PubMed  Google Scholar 

  67. Kegler MC, Twiss JM, Look V. Assessing community change at multiple levels: the genesis of an evaluation framework for the California healthy cities project. Health Education and Behavior. 2000;27(6):760–79.

    Article  CAS  PubMed  Google Scholar 

  68. Wallerstein N, Polascek M, Maltrud K. Participatory evaluation model for coalitions: the development of system indicators. Health Promot Pract. 2002;3(3):361–73.

    Article  Google Scholar 

  69. Wallerstein N. A participatory evaluation model for healthier communities: developing indicators for New Mexico. Public Health Rep. 2000;115(2–3):199–204.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  70. Rootman I, Goodstadt M, Hyndman B, McQueen DV, Potvin L, Springett J, et al. Evaluation in health promotion: principles and perspectives Geneva: WHO regional publications European series; 2001.

    Google Scholar 

  71. Harris R, Hardman E. A formative model of service evaluation. Journal of Clinical Excellence. 2001;3(2):69–73.

    Google Scholar 

  72. O'Connor-Fleming ML, Parker E, Higgins H, Gould T. A framework for evaluating health promotion programs. Health Promotion Journal of Australia. 2006;17(1):61–6.

    Article  PubMed  Google Scholar 

  73. Steckler AB, Linnan L. Process evaluation for public health interventions and research. San Francisco, Calif.: Jossey-Bass; 2002.

  74. Saunders RP, Evans MH, Joshi P. Developing a process-Evaluation plan for assessing health promotion program implementation: a how-to guide. Health Promot Pract. 2005;6(2):134–47.

    Article  PubMed  Google Scholar 

  75. Baranowski T, Stables G. Process evaluations of the 5-a-day projects. Health Educ Behav. 2000;27(2):157–66.

    Article  CAS  PubMed  Google Scholar 

  76. Beery WL, Senter S, Cheadle A, Greenwald HP, Pearson D, Brousseau R, et al. Evaluating the legacy of community health initiatives: a conceptual framework and example from the California Wellness Foundation's health improvement initiative. Am J Eval. 2005;26(2):150–65.

    Article  Google Scholar 

  77. Chinman M, Imm P, Wandersman A, Kaftarian S, Neal J, Pendleton K, et al. Using the getting to outcomes (GTO) model in a statewide prevention initiative. Health Promot Pract. 2001;2(4):302–9.

    Article  Google Scholar 

  78. Butterfoss FD, Francisco VT. Evaluating community partnerships and coalitions with practitioners in mind. Health Promot Pract. 2004;5(2):108–14.

    Article  PubMed  Google Scholar 

  79. Themessl-Huber M, Lazenbatt Anne A, Taylor J. Overcoming health inequalities: a participative evaluation framework fit for the task. J R Soc Promot Health. 2008;128(3):117–22.

    Article  PubMed  Google Scholar 

  80. Jolley G, Lawless A, Hurley C. Framework and tools for planning and evaluating community participation, collaborative partnerships and equity in health promotion. Health Promotion Journal of Australia. 2008;19(2):152–7.

    Article  PubMed  Google Scholar 

  81. South J, Fairfax P, Green E. Developing an assessment tool for evaluating community involvement. Health Expect. 2005;8(1):64–73.

    Article  PubMed  PubMed Central  Google Scholar 

  82. Kruger DJ, Morrel-Samuels S, Davis-Satterla L, Harris-Ellis BJ, Slonim A. Developing a cross-site Evaluation tool for diverse health interventions. Health Promot Pract. 2010;11(4):555–61.

    Article  PubMed  Google Scholar 

  83. Brandstetter S, Curbach J, Lindacher V, Rueter J, Warrelmann B, Loss J. Empowerment for healthy nutrition in German communities: a study framework. Health Promot Int. 2017;32(3):500–10.

    PubMed  Google Scholar 

  84. Jolley G. Evaluating complex community-based health promotion: addressing the challenges. Evaluation and Program Planning. 2014;45:71–81.

    Article  PubMed  Google Scholar 

  85. Dunne A, Scriven A, Howe A. Research partnership and knowledge transfer in the development of a generic evaluation toolkit for health promotion interventions. Translational Research for Primary Healthcare2012. p. 147–160.

  86. Aarestrup AK, Jørgensen TS, Due P, Krølner R. A six-step protocol to systematic process evaluation of multicomponent cluster-randomised health promoting interventions illustrated by the boost study. Evaluation and Program Planning. 2014;46:58–71.

    Article  PubMed  Google Scholar 

  87. Wagemakers A, Koelen MA, Lezwijn J, Vaandrager L. Coordinated action checklist: a tool for partnerships to facilitate and evaluate community health promotion. Glob Health Promot. 2010;17(3):17–28.

    Article  PubMed  Google Scholar 

  88. Masso M, Quinsey K, Fildes D. Evolution of a multilevel framework for health program evaluation. Aust Health Rev. 2017;41(3):239–45.

    Article  PubMed  Google Scholar 

  89. Mantziki K, Renders CM, Westerman MJ, Mayer J, Borys JM, Seidell JC. Tools for a systematic appraisal of integrated community-based approaches to prevent childhood obesity. BMC Public Health. 2018;18(1).

  90. Hoddinott P, Allan K, Avenell A, Britten J. Group interventions to improve health outcomes: a framework for their design and delivery. BMC Public Health. 2010;10.

  91. Grant A, Treweek S, Dreischulte T, Foy R, Guthrie B. Process evaluations for cluster-randomised trials of complex interventions: a proposed framework for design and reporting. Trials. 2013;14:15.

    Article  PubMed  PubMed Central  Google Scholar 

  92. Wagemakers A, Vaandrager L, Koelen MA, Saan H, Leeuwis C. Community health promotion: a framework to facilitate and evaluate supportive social environments for health. Evaluation and Program Planning. 2010;33(4):428–35.

    Article  PubMed  Google Scholar 

  93. Goldsmith R, Harris S. Thinking inside the box: the health cube paradigm for health and wellness program Evaluation and design. Population Health Management. 2013;16(5):291–5.

    Article  PubMed  PubMed Central  Google Scholar 

  94. World Health Organization. Health promotion Evaluation: recommendations to policy-makers. Geneva: World Health Organization; 1998.

    Google Scholar 

  95. W.K. Kellogg Foundation. The Step-by-Step Guide to Evaluation USA: W.K. Kellogg Foundation; 2017.

  96. National Institute for Health and Care Excellence (NICE). Behaviour Change: individual approaches Public Health Guideline [PH49]. London: NICE; 2014. Available from: https://www.nice.org.uk/guidance/ph49/chapter/1-recommendations#recommendation-15-monitor-behaviour-change-interventions.

  97. Taylor M, Purdue D, Wilson M, Wilde P. Evaluating community projects: a practical guide. Joseph Rowntree Foundation: York; 2005. Available from: https://www.jrf.org.uk/report/evaluating-community-projects-practical-guide.

    Google Scholar 

  98. The Centre for the Advancement of Community Based Public Health. An Evaluation Framework for Community Health Programs. Durham, NC.; 2000.

  99. Dugdill L, Stratton G. Evaluating sport and physical activity interventions: a guide for practitioners: University of Salford; 2007.

    Google Scholar 

  100. Health System Intelligence Project (HSIP). The Health Planner's Toolkit Module 8: Evaluation 2008. Available from: http://www.ontla.on.ca/library/repository/mon/22000/283845.pdf.

  101. National Health Service (NHS) Health Scotland. LEAP for health: learning, Evaluation and planning. Edinburgh: Health Scotland; 2003. Available from: http://www.healthscotland.com/uploads/documents/308-LEAP_for_health.pdf.

    Google Scholar 

  102. US Department of Health and Human Services. Physical Activity Evaluation Handbook. Atlanta GA; 2002.

  103. Sport England. Evaluation framework: measurement and Evaluation guidance tools and resources. London: Sport England; n.d. Available from: https://evaluationframework.sportengland.org/.

  104. Better Evaluation. Better Evaluation rainbow framework and planning tool 2014. Available from: http://www.betterevaluation.org.

    Google Scholar 

  105. Leeman J, Sommers J, Vu M, Jernigan J, Payne G, Thompson D, et al. An evaluation framework for obesity prevention policy interventions. Prev Chronic Dis. 2012;9.

  106. Community Toolbox. Community Toolbox: evaluating community programs and initiative. Kansas: Centre for community health and development of University of Kansas; 2018. Available from: https://ctb.ku.edu/en/evaluating-initiative.

  107. West of England Academic Health Science Network, BNSSG Research and Evidence Team, NIHR CLAHRC West. Evaluation Works: A toolkit to support commissioning of health and care services 2016. Available from: http://www.nhsevaluationtoolkit.net/.

  108. Cavill N, Roberts K, Ells L. Evaluation of weight management, physical activity and dietary interventions: an introductory guide. Oxford: Public Health England; 2015.

    Google Scholar 

  109. HM Treasury. The Magenta book: guidance for Evaluation. London: HM Treasury; 2011.

  110. Ontario Agency for Health Protection and Promotion (Public Health Ontario), Snelling S, Meserve A. Evaluating health promotion programs: introductory workbook. Toronto, ON: Queen's Printer for Ontario; 2016. Available from: https://www.publichealthontario.ca/en/erepository/Evaluating_health_promotion_programs_workbook_2016.pdf.

  111. Prevention and Population Health Branch. Evaluation framework for health promotion and disease prevention programs. Melbourne: Victorian Government Department of Health and Human Services; 2011. Available from: https://www2.health.vic.gov.au.

  112. World Health Organization Regional Office of Europe. Good practice appraisal tool for obesity prevention Programmes, projects, initiatives and interventions. Copenhagen: World Health Organization; 2011. Available from: http://www.euro.who.int/__data/assets/pdf_file/0007/149740/e95686.pdf?ua=1.

    Google Scholar 

  113. Pérez-Escamilla R, Segura-Pérez S, Damio G. Applying the program impact pathways (PIP) evaluation framework to school-based healthy lifestyles programs: workshop Evaluation manual. Food Nutr Bull. 2014;35:S97–S107.

    Article  PubMed  Google Scholar 

  114. Craig P, Cooper C, Gunnell D, Haw S, Lawson K, Macintyre S, et al. Using natural experiments to evaluate population health interventions: guidance for producers and users of evidence. London: Medical Research Council; 2012. Available from: https://mrc.ukri.org/documents/pdf/natural-experiments-guidance/.

    Google Scholar 

  115. Craig P, Cooper C, Gunnell D, Haw S, Lawson K, Macintyre S, et al. Using natural experiments to evaluate population health interventions: new Medical Research Council guidance. J Epidemiol Community Health. 2012;66(12):1182–6.

    Article  PubMed  Google Scholar 

  116. Hand RK, Abram JK, Brown K, Ziegler PJ, Parrott JS, Steiber AL. Development and validation of the guide for effective nutrition interventions and education (GENIE): a tool for assessing the quality of proposed nutrition education programs. J Nutr Educ Behav. 2015;47(4):308–16.

    Article  PubMed  Google Scholar 

  117. Abram JK, Hand RK, Parrott JS, Brown K, Ziegler PJ, Steiber AL. From the academy: what is your nutrition program missing? Finding answers with the guide for effective nutrition interventions and education (GENIE). J Acad Nutr Diet. 2015;115:122–30.

    Article  PubMed  Google Scholar 

  118. Smith BJ, Rissel C, Shilton T, Bauman A. Advancing evaluation practice in health promotion. Health Promotion Journal of Australia. 2016;27(3):184–6.

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

The authors would like to thank Nick Cavill, Darcy Hare and Harry Rutter who contributed as part of the expert consultation.

Funding

The work was undertaken by the Centre for Diet and Activity Research (CEDAR), a UKCRC Public Health Research Centre of Excellence. Funding from the British Heart Foundation, Cancer Research UK, Economic and Social Research Council, Medical Research Council, the National Institute for Health Research, and the Wellcome Trust, under the auspices of the UK Clinical Research Collaboration, is gratefully acknowledged. The funders had no role in any element of this research.

Author information

Authors and Affiliations

Authors

Contributions

JF, AJ & WH conceptualised the research questions, developed and agreed the search strategy, data extraction and synthesis. JF conducted the search, screening, data extraction and analysis. KM checked and validated the screening of sources. AJ & WH checked and validated the data extraction. All authors contributed to the manuscript, critically reviewed and approved the final manuscript.

Corresponding author

Correspondence to Judith F. Fynn.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary information

Additional File 1.

Brief descriptions of the frameworks.xlsx. A short summary of each framework based on the stated scope and aim provided by the original authors of each framework.

Additional File 2.

General characteristics of the frameworks.xlsx. Summary of data extracted to identify the scope and general characteristics of each framework.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Fynn, J.F., Hardeman, W., Milton, K. et al. A scoping review of evaluation frameworks and their applicability to real-world physical activity and dietary change programme evaluation. BMC Public Health 20, 1000 (2020). https://0-doi-org.brum.beds.ac.uk/10.1186/s12889-020-09062-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://0-doi-org.brum.beds.ac.uk/10.1186/s12889-020-09062-0

Keywords