The approach described in this paper to engage stakeholders to interpret and synthesize community data reflects the increasing realization among public health researchers and practitioners that evaluation of complex community programs requires a shift from narrowly defined effectiveness studies that maximize internal validity to a focus on generalizability and external validity, while recognizing the heterogeneity of local contexts. This requires a balance between identifying features of the program that are transferrable across locations and local adaptations that matter [18]. At this point, because of the separation that exists between community practitioners or program implementers and researchers or evaluators, this dual goal can only be accomplished by partnering across multiple stakeholders via the kind of process described in this paper. This brings together those able to interpret local data with those who are able to guide the identification of generalizable themes that transcend context.
These collaborations are difficult for several reasons. First, because of funding streams and training they are usually researcher or evaluator led, and without careful planning and communication, could easily create the impression that local participation is only solicited to serve an external research agenda. Second, even if the objective of the collaboration is not for research but for joint learning about what works, the emphasis by researchers or evaluators on rigorous validation of the data and a process of critical inquiry can be experienced as judgmental by some stakeholders, resulting in a barrier to the open communication needed for joint meaning making.
In the SCALE evaluation, the evaluation team had the benefit of having worked with the communities and the implementation team for four years, initially in a formative evaluation role that supported communities within real time through their process of implementation [19]. Evaluation team members had personal relationships with the community stakeholders, and, therefore, a degree of trust had been established over time. Additionally, a lot of careful planning and preparation went into both the design and the execution of the process to ensure that the locus of the evaluation was centered around the communities and that their ownership and knowledge of their data was recognized and appreciated.
In order to forestall any perception that the data of some communities were favored over others, our approach explicitly designed the first two cycles of synthesis to create subgroups that focused on the common components of the theory of change across communities. Community specific interpretations were not undertaken until the third cycle by which time most stakeholders were comfortable with the synthesis process, and thematic consensus on the subgroups data had been achieved. This design also made it possible to begin conversations about generalizability early, since the second cycle focused on cross community themes.
On the execution side, each cycle was preceded by communication sessions to explain the goals for the cycle, and community members were encouraged to provide suggestions about the methodology. Each synthesis meeting began with an update of progress to date to enable members who had not attended previous meetings to be caught up. All templates used for synthesis were created using collaborative group software, were extensively tested with community and implementation stakeholders and were editable by everyone irrespective of whether they participated in the synthesis meetings. Recognizing the wide variation in evaluation expertise across the stakeholder groups, roles for community members were clearly established and agreed upon so that each team member could bring their own area of expertise into the synthesis process. Stakeholders were informed of the time commitment needed in advance, but competing priorities made it difficult for several of them to continue in the group meetings, necessitating follow-up conversations that consumed additional time and resources. Overall, the evaluation team significantly underestimated the time and labor required to manage the synthesis cycles to assure adequate data quality, community input and overall engagement.
Some of the extensive effort needed was due to the intrinsic heterogeneity across the SCALE communities that required piecing together disparate data to elicit both community specific and generalizable insights. The varying quality and heterogeneity of the data and the need to shore up gaps in data through supplementary collection was another major factor that contributed to additional time and effort. Even though each SCALE community was tackling a different health topic, they followed a common theory of change and were all provided with common tools for data collection. For a variety of reasons beyond the scope of this paper, some of the data collection tools were felt by the communities to be burdensome and not valuable, and, therefore, overall community compliance with the data collection recommendations was poor.
To address these considerations, future attempts at participatory evaluation of complex community or public health interventions would benefit from promoting the salience of these approaches during implementation and not just after. As mentioned previously, complex interventions require an iterative testing process; building the capacity of program implementers or community coalitions to develop systems for routine cross-sector data collection, and to conduct synthesis cycles within their communities or program sites in partnership with community members, could encourage improved data collection and use during implementation. Further, this approach could also prepare the data required to answer broader evaluation questions about external validity and generalizability at the end of project, without the need to start from scratch. Moreover, since a key goal of the participatory synthesis is to encourage meaning making by those who are closest to the data, building evaluation capability within community coalitions enables them to engage their community members in evaluating the effects of transformation efforts such as SCALE. Adopting this “empowerment evaluation” approach ensures both that the power of interpretation is situated within communities and that high-quality data is available for both local and higher order decision making [20].