- Open Access
Comparing output from two methods of participatory design for developing implementation strategies: traditional contextual inquiry vs. rapid crowd sourcing
Implementation Science volume 17, Article number: 46 (2022)
Participatory design methods are a key component of designing tailored implementation strategies. These methods vary in the resources required to execute and analyze their outputs. No work to date has examined the extent to which the output obtained from different approaches to participatory design varies.
We concurrently used two separate participatory design methods: (1) field observations and qualitative interviews (i.e., traditional contextual inquiry) and (2) rapid crowd sourcing (an innovation tournament). Our goal was to generate and compare information to tailor implementation strategies to increase the use of evidence-based data collection practices among one-to-one aides working with children with autism. Each method was executed and analyzed by study team members blinded to the output of the other method. We estimated the personnel time and monetary costs associated with each method to further facilitate comparison.
Observations and interviews generated nearly double the number of implementation strategies (n = 26) than did the innovation tournament (n = 14). When strategies were classified into implementation strategies from the Expert Recommendations for Implementing Change (ERIC) taxonomy, there was considerable overlap in the content of identified strategies. However, strategies derived from observations and interviews were more specific than those from the innovation tournament. Five strategies (13%) reflected content unique to observations and interviews and 3 (8%) strategies were unique to the innovation tournament. Only observations and interviews identified implementation strategies related to adapting and tailoring to context; only the innovation tournament identified implementation strategies that used incentives. Observations and interviews required more than three times the personnel hours than the innovation tournament, but the innovation tournament was more costly overall due to the technological platform used.
There was substantial overlap in content derived from observations and interviews and the innovation tournament, although there was greater specificity in the findings from observations and interviews. However, the innovation tournament yielded unique information. To select the best participatory design approach to inform implementation strategy design for a particular context, researchers should carefully consider unique advantages of each method and weigh the resources available to invest in the process.
Comparing output from two methods of participatory design for developing implementation strategies: traditional contextual inquiry vs. rapid crowd sourcing
Implementation strategies, the methods used to improve the use and sustainment of evidence-based practices (EBPs), are the backbone of implementation science . Carefully selecting implementation strategies that target theorized implementation mechanisms , guided by contextual inquiry to identify implementation determinants , is de rigueur for implementation efforts. There are several published methods for selecting and tailoring implementation strategies , all of which emphasize the importance both of collaborating with stakeholders, or “end-users,” of the intervention being implemented, and of drawing on methods from participatory design to do so. Because implementation is context dependent, implementation strategies will be more successful if those leading implementation efforts learn from practitioners who will use the intervention of interest. Outputs from contextual inquiry ideally identify implementation determinants, which then drive implementation strategy selection [4, 5].
Implementation science has borrowed from other disciplines to add participatory tools to inform the design of implementation strategies. Participatory design approaches have a long history in many fields and aim to involve core stakeholders directly in co-designing tools, products, and other processes. Participatory design originated as a collaborative method of product design that engages end-users in designing new products using a variety of different methodologies [6, 7]. Participatory design methods can include observation of product use in the field, interviews, and focus groups, as well as traditional user-centered design practices such as obtaining reflective feedback from stakeholders on various mock-ups or prototypes throughout the design phase [6, 8]. Primarily originating in the information technology space , the adoption of participatory design methods for designing implementation strategies is more recent.
Most approaches to participatory design in implementation science (e.g., concept mapping, implementation mapping) rely heavily on inputs obtained through rigorous contextual inquiry, using methods such as field observations and qualitative interviews [9, 10]. However, these approaches to participatory design are time intensive and costly. It is also not clear which participatory design approach an implementation scientist should pick when engaging end-users in implementation strategy development for a target EBP, nor is it known whether different approaches will lead to varying results. To our knowledge, no one has compared the results of using different approaches of participatory design on the resultant implementation strategy output . In particular, it would be useful for the field to identify if alternative rapid participatory design methods can produce comparable outcomes to the more traditional methods that can take months to execute and analyze.
To this end, this study compared output from two participatory design approaches:  a contextual inquiry approach (i.e., field observations and qualitative interviews with stakeholders) versus  a more recent addition to the panoply of participatory design tools, an innovation tournament, which is intended as a more rapid approach to gaining end-user feedback to inform implementation strategy design. Innovation tournaments involve a kind of crowd sourcing—a competition in which members of a group (in this case, relevant stakeholders to an implementation problem) can submit ideas about how to address a challenge. These ideas compete with one another through several rounds of vetting. Other members of the group can vote for and comment on participants’ different ideas (thus engaging in the reflective practice from stakeholders that is integral to all participatory design methods; 6). In the last round, experts work with the originators of the most popular and promising ideas to refine them for a last round of voting. Innovation tournaments have become particularly popular in health care and other industries [12,13,14]. Stewart and colleagues recently used one to identify potential implementation strategies to increase the use of evidence-based practices in outpatient mental healthcare, demonstrating their feasibility and potential utility in generating tailored implementation strategies in behavioral health . As innovation tournaments are designed to quicker, cheaper, and less burdensome on stakeholder partners, identifying whether their output produces comparable results to those of more traditional contextual inquiry approaches has important implications for how we use participatory design in implementation science.
Our secondary aim was to categorize the resources involved with executing and analyzing the data obtained from each participatory design approach. Given the breadth of potential implementation strategies that emerged from Stewart and colleagues’ innovation tournament, we hypothesized the innovation tournament would produce output comparable to that obtained from observations and qualitative interviews, with less burden on stakeholders and research staff.
This work was conducted as an exploratory project as part of the Penn ALACRITY Center which aims to integrate behavioral economics and implementation science principles to increase evidence-based practice use . In this exploratory project, we aimed to employ principles of participatory and user-centered design to design a mobile application to support data collection among behavioral health technicians (BHTs) who work one-on-one with children with an autism spectrum disorder in schools . Quantitative data collection on child response to intervention and progress monitoring is a core EBP component of applied behavioral analysis (ABA) for children with autism yet is poorly implemented in practice . As formative work in the design and development of this app, and in keeping with the goals of the Penn ALACRITY Center to test new methods of participatory design, our team concurrently  collected observations and interviews with BHTs and interviews with their supervisors and  conducted an innovation tournament inviting BHTs and their supervisors, to participate. We synthesized data obtained from each method using a systematic, iterative processes to identify implementation strategies.
To avoid cross-contamination of output from each method, the research team divided into two separate groups for the duration of data collection and synthesis so that each team was blinded to outputs or information obtained from the other method of participatory design. We briefly describe the participatory design strategy and data analysis below. The City of Philadelphia (Approval #2019-32) and University Pennsylvania (Approval #849995) Institutional Review Boards approved the study.
Observations and interviews
We recruited 9 BHTs and 7 BHT supervisors from four different agencies in Philadelphia (63% female; 31% White; 44% Black; 6% Asian; 19% Hispanic/Latinx). Recruited BHTs had to be working with a school-aged child with autism. BHTs allowed us to observe their data collection practices in the field and participated in a subsequent qualitative interview over the phone. Two BHTs completed a qualitative interview only due to COVID-19 social distancing restrictions put into place in the final month of planned data collection. Recruited BHT supervisors participated in phone interviews only. To protect BHT participant confidentiality, BHT supervisors were recruited separately from participating agencies, rather than as matched BHT/BHT supervisor pairs. We obtained informed consent from each participant prior to any research activities.
Prior to observations, author HJN trained research team members in a structured observation coding system adapted to document barriers and facilitators to BHT data collection [19, 20]. Coders were trained to identify opportunities for data collection on the part of BHTs across 3-min intervals, whether or not the BHTs engaged in data collection on available opportunities and take field notes on observed barriers and facilitators to data collection.Footnote 1 After obtaining BHT consent and formal approval from the school in which the BHT worked, research staff traveled to the BHT’s work site for a 1-h field observation. Two research team members (EBH and BR) attended the first two observations to ensure reliable use of the structured field observation form; one team member completed the remaining observations and debriefed with the second team member immediately after. During each observation, research staff observed BHT data collection practices from the back of the classroom using the structured form. Research staff noted, in real time, perceived barriers to data collection they observed. At the conclusion of the 1-h observation, staff spent approximately 5 min asking the BHT to expand on observed barriers (e.g., “I noticed you didn’t take data in X interval. What, if anything, got in the way of you doing so?”). After the observation, BHTs scheduled a qualitative interview.
Qualitative interview guides for BHTs and BHT supervisors were designed to assess data practices BHTs used, assess barriers to engaging in quantitative data collection, and elicit recommendations on how quantitative data collection could be made easier for BHTs. BHTs additionally answered standard belief elicitation questions and questions adapted from a theory-guided interview created by Potthoff and colleagues [21, 22]. Copies of the interview guides are available upon request. Team members met and reviewed emerging themes following every 2–3 interviews; when two successive interviews did not yield any new substantive information across both stakeholders, we determined that saturation was reached. Thematic saturation was reached after the 9th BHT interview and the 7th BHT supervisor interview.
The interviewer audio-recorded the interview and a member of the research team transcribed the recording. The researchers who conducted the field observation (EBH or BR) analyzed the interviews, so that information learned via observations could guide interpretation of qualitative responses. The research team developed a structured codebook via iterative review of three transcripts and applied it to the remaining transcripts. The a priori code of “barriers and facilitators to quantitative data collection” was the primary code of interest for this study. Coders met regularly to review codes and resolve discrepancies.
From the coded transcripts, we amassed a list of 10 identified major barriers across BHTs and BHT supervisors. The same two staff members systematically reviewed each identified barrier and generated one or more discrete implementation strategies that would address each barrier drawing on their review of all qualitative data and field observation notes collaboratively and systematically, guided by leading frameworks to guide implementation strategy design: the Expert Recommendations for Implementing Change (ERIC) taxonomy and the Behavior Change Wheel [23, 24]. To ensure comprehensiveness, a third team member who had administered qualitative interviews but was not involved in the coding process (KAZ) then reviewed the list of barriers and independently generated corresponding implementation strategies. This third team member remained masked to the original generated implementation strategies prior to a consensus meeting and did not identify any additional implementation strategies. This process resulted in a final set of 26 discrete implementation strategies, with two to five strategies generated per barrier (some generated strategies also addressed multiple barriers).
Our innovation tournament procedure was guided by Stewart and colleagues, which was developed through the Penn Medicine Center for Health Care Innovation [13, 15]. We invited behavioral health professionals working with children with autism in Philadelphia schools (n = 21), including BHTs and their supervisors, who had provided us their email addresses at previous in-person site visits. We sent them email invitations to participate in an innovation tournament titled “Your Idea Matters! The Data Collection Improvement Challenge.” We also asked behavioral health agency leaders (n = 27) and team meeting leaders at in-person site visits (n = 5) to forward recruitment invitations to their staff. Some BHTs may have received the invitation more than once (i.e., from a direct email from our team, from their agency leader, and from their team meeting leader). We sent agency leaders a priming email approximately 2 weeks prior to the start of the tournament, as advance contact aligns with recommendations in survey design . We then sent the first invitation email to agency leaders, team meeting leaders, and BHTs, which included the link to participate, and 5 reminder emails over the course of 5 weeks. During the second week of reminders, we contacted agency leaders to participate and invited their staff to participate by phone.
We made the web link to the innovation tournament available for 5 weeks in October 2019. Participants opened the link and provided consent before participating. We hosted the online tournament on the “Your Big Idea Platform” that facilitates crowdsourcing for ideas and solutions. Innovation challenges are posted to the platform, and participants can submit an idea and comment and rate ideas submitted by other participants. As such, the electronic platform supports both the submission of ideas as well as hosting the full list of suggestions made for participants to review, rate, and vote for the suggestions they think represent the best solutions for the challenge problem.
Developing the right tournament prompt (i.e., the question for participants) is critical to the success of an innovation tournament [13, 15]. We iteratively designed the prompt for this study. The research team generated the question language via email and in weekly meetings and then sought feedback from 2 BHTs on the wording and design of the prompt. The final prompt asked, “What would make data collection about children with autism easier, more useful, and more motivating for you?” Participants could submit and rate as many ideas as they liked.
Eleven individuals submitted 14 ideas. Procedures for reviewing, rating, identifying “winning ideas,” and categorizing output followed established procedures . Briefly, the research team formed a Challenge Committee consisting of an expert panel of stakeholders to vote on the submissions. The committee included two city administrators, two agency stakeholders, one autism content expert, and one behavioral science expert. The committee rated the 14 ideas on their potential impact as an implementation strategy. Participants with the top three ideas were considered “winning participants;” these participants were later celebrated at a community event intended to foster community engagement in the implementation of evidence-based practices for autism.
To categorize information shared through the innovation tournament platform into clear implementation strategies, three members of the research team (JEB, RES, DSM) met and applied aspects of the NUDGE (Narrow, Understand, Discover, Generate, Evaluate) framework to extrapolate the discrete implementation strategies from the information submitted by innovation tournament participants. See  for a detailed description of how the NUDGE framework is applied to analysis of innovation tournament data. Briefly, this is an iterative process of:  generating hypotheses about behavioral barriers that interfere with EBP delivery through structured brainstorming that links the reported barriers to core principles from behavioral science (e.g., cognitive biases and heuristic thinking),  de-duplicating hypotheses, and  rapidly validating hypothesized barriers through expert consultation, review of the literature, and confirmation by core stakeholders. This output then informs generation of tailored implementation strategies to directly address the behavioral barriers identified. We selected the NUDGE framework to guide generation of implementation strategies here as it aligned with methods of the only other innovation tournament used in mental health care to date [15, 26].
Resources and cost of each method
We captured the time and costs associated with each method, guided by principles of time-driven activity-based costing, a method increasingly employed in implementation science . Researchers who executed each participatory design method estimated the time and money spent throughout the process of data collection and synthesis.
After output from each method was used to generate implementation strategies, guided by each method’s respective analytic framework (i.e., the NUDGE framework for the innovation tournament and the ERIC and the Behavior Change Wheel for the traditional contextual inquiry), we then categorized each implementation strategy using the ERIC taxonomy to facilitate comparison of output content. Table 1 shows the implementation strategies that the two participatory design approaches generated. The first author (EBH) classified all implementation strategies across the two approaches using the ERIC taxonomy, using the clusters identified by Waltz and colleagues [24, 28]. An independent reviewer (HJN) was trained in the ERIC taxonomy and then double coded 50% of all strategies, with 100% concordance achieved without the need for a consensus procedure. To further facilitate comparison, we also classified how, within each ERIC cluster, the implementation strategy generated from each method compared when examined at sub-clusters identified by Powell and colleagues 2015 within the ERIC taxonomy .
As seen in Table 1, observations and interviews resulted in nearly double the number of implementation strategies (n = 26) than did the innovation tournament (n = 14). Together, both methods identified implementation strategies that spanned eight of the nine clusters of implementation strategies identified by Waltz and colleagues . Only the observations and interviews generated implementation strategies in the cluster of strategies related to “adapting and tailoring to context” and only the innovation tournament generated implementation strategies in the cluster of strategies related to “utilize incentive strategies.” No implementation strategies from either method were generated in the cluster of engaging consumers (e.g., involving children and their families).
At the more granular level, resultant strategies across both methods spanned 10 discrete implementation strategy categories from the ERIC taxonomy. Although the observations and interviews generated more strategies than the innovation tournament did, a substantial portion of the output from observation and interviews overlapped conceptually with that from the innovation tournament data, albeit with a greater level of specificity emerging from observations and interviews. Specifically, when examined by implementation strategy category, 32 of the 40 total identified implementation strategies (80%) reflected overlapping content obtained from each method. Areas of overlap at this level of implementation strategy category were observed for changing the physical structure, equipment and regulatory processes (n = 2 strategies from innovation tournament, n = 3 strategies from observations/interviews), changing record systems (n = 4 strategies from innovation tournament, n = 6 strategies from observations/interviews), providing clinical supervision (n = 2 strategies from innovation tournament, n = 5 strategies from observations/interviews), facilitating relay of clinical data to providers (n = 1 strategy from innovation tournament, n = 1 strategy from observations/interviews), and conducting educational meetings (n = 2 strategies from innovation tournament, n = 6 strategies from observations/interviews). With respect to areas of conceptual divergence between the two methods, only the observations and interviews yielded implementation strategies in the categories of promoting adaptability (n = 2 strategies), conducting local needs assessment (n = 1 strategy), and reminding clinicians (n = 2 strategies). Only the innovation tournament yielded implementation strategies in the categories of: audit and feedback (n = 1 strategy), and altering incentive/allowance structures (n = 2 strategies).
As noted above, within each of these areas, findings from the observations and interviews tended to yield more specific and somewhat more actionable implementation recommendations than the innovation tournament did. For example, the innovation tournament identified the importance of “improving training” and “providing skills training in functional analysis.” In contrast, observation and interview data yielded 6 specific recommendations to improve training in response to specific implementation barriers to data collection that BHTs reported. For example, specific strategies included training on best data collection practices, training for active and outdoor times, and training on best practices for data collection when clients display problem behaviors. Similarly, for “provide clinical supervision,” the innovation tournament indicated the importance of offering in vivo coaching and providing immediate feedback on data collection. In contrast, the observation and interview data yielded 5 more specific recommendations about how supervisory processes could better address specific barriers to evidence-based data collection for BHTs.
Table 2 shows the estimated costs of each participatory design method, separated by preparation costs, data collection costs, and data synthesis costs. Overall, personnel time was substantially higher for the observations and interviews (~ 263 h) than for the innovation tournament (~ 68 h). However, the overall cost of the innovation tournament, which relied on an electronic platform to power idea submission and allow participants to comment on and rate ideas submitted by other participants, was much higher than that for observations and interviews.
To our knowledge, this is the first effort to directly compare the output of two distinct methods of participatory design and their resultant implications for implementation strategy design. We compared the output of two different participatory methodological approaches—an innovation tournament and observations and interviews. Data suggested a high degree of convergence in output across the two strategies when examining concordance at the broadest conceptual levels (i.e., within the general theoretical clusters of implementation strategies). However, the information yielded by the traditional contextual inquiry was much richer and detailed than that obtained from the innovation tournament. Overall, the observations and interviews generated a greater number of recommendations and more specific recommendations than did the innovation tournament. Given that a growing body of literature highlights the importance of tailored implementation strategies to lead to optimal implementation outcomes [2, 4], we might hypothesize that the more specific implementation strategies generated by the traditional contextual inquiry may be more effective than the more general strategies generated by the innovation tournament. However, the contextual inquiry took almost four times as long. That said, each method has strengths and limitations that might inform a decision to use one method or another. Table 3 presents a high-level overview of the strengths and limitations of each method studied to guide researchers’ decisions about how and when to use one method over another.
While our results do suggest that observations and interviews overall resulted in more specific implementation strategies than did the innovation tournament, there was still unique information produced by each. As such, results from this study do not suggest a clear recommendation that one participatory design method is clearly superior to the other. The time and resource element should be given significant consideration. Overly long contextual inquiry efforts to design implementation strategies may further delay the integration of evidence into practice or providing stakeholders with critical feedback during the implementation. It remains an empirical question how to trade off efficiency in contextual inquiry and detail in implementation strategy output to optimize acceleration of research into practice. Researchers should also carefully weigh the potential information gaps that may occur from using one participatory design method and consider stakeholder preferences for implementation strategies. For example, contextual inquiry was better at generating strategies relating to adapting and tailoring interventions to context, whereas the innovation tournament generated more strategies related to incentives, the latter being considered “most useful” by clinicians, supervisors, and payers . Surprisingly, no implementation strategies from either method were generated in the cluster of engaging consumers, and perhaps direct prompts are needed to generate strategies that include this important stakeholder group. Future work could explore how integrating methods, such as conducting an initial innovation tournament followed by rapid stakeholder feedback (e.g., “member checking”), could perhaps lead to greater efficiency in the contextual inquiry process than more traditional methods. For example, innovation tournament output could be reviewed briefly with local champions to delineate the specific suggestions that emerged from the formal contextual inquiry. This would be an important area for future research to determine if comparable specificity of information could be obtained in a more expedient fashion through an innovation tournament compared to interviews and observations.
It is also important to note that while we might hypothesize that the more specific strategies generated by observations and interviews might be more effective than the more general ones generated by the innovation tournament, our data are not able to determine whether the output of one participatory design method is more effective, acceptable, or feasible than another. An important next step in this line of research will be to share the resultant strategies from each method with stakeholders and evaluate their perception of each strategy’s acceptability and feasibility. Whether stakeholder perceptions vary as a function of the chosen participatory design approach it was derived from is an important question to explore in future research.
Ours findings suggested clear limitations of innovation tournaments as compared to more traditional contextual inquiry. That said, the innovation tournament did yield both somewhat comparable and unique information (related to incentives) to that obtained from the observations and interviews and required substantially less time investment from both the research team and stakeholder participants to complete. The primary driver of the cost of the innovation tournament was the electronic platform used in this study. Researchers hoping to conduct an innovation tournament to gain more rapid contextual inquiry could use cheaper HIPAA-compliant survey platforms to elicit ideas from stakeholders in response to an innovation prompt. However, to include rounds of coding and stakeholder involvement in rating of ideas, along with various incentives for the “winners’ of the innovation tournament, would be more costly. Should costs for such technology reduce in the future, innovation tournaments may hold potential to be a more cost-efficient way to gather rapid information from stakeholders and should be continued to be studied as an alternative method to engage with stakeholders to address implementation challenges.
There are several reasons why output may differ between the two methods we studied. Observations and interviews allowed for the “up close and personal” examination of the BHT workflow and environment, as well as opportunities for follow-up questions; this may have facilitated obtaining information about adapting and tailoring to context that did not emerge from the innovation tournament. In contrast, the innovation tournament may have been more likely to identify incentive-related strategies because the prompt allowed for emotional “distancing” on the topic from respondents. Specifically, innovation tournament participants were asked to report on what would make data collection for BHTs easier and more motivating broadly (in contrast to speaking specifically to the barriers and facilitators they faced in data collection in the qualitative interview). Participants may have been more likely to think that their colleagues would be motivated by rewards or other incentives to engage in data collection, even if they perceived that they had their own intrinsic motivation to do so, and that the innovation tournament was more likely to elicit this type of information. Participants also may have felt more comfortable sharing ideas about monetary rewards in a more anonymous setting like the innovation tournament, rather than face-to-face during interviews.
Finally, it is possible that the fundamentally different approach to the framing of the inquiry in each approach led to the divergence in findings. During the qualitative interviews, research staff asked open-ended questions barriers and facilitators to data collection and then deduced potential solutions to the information shared in the generation of implementation strategies. In contrast, the innovation tournament asked respondents to report directly on suggested solutions to the challenge of how to make data collection easier and more motivating and then the research team refined the offered strategies. Therefore, in many ways, the implementation strategies generated from the observations and interviews incorporate the researchers’ experience and expertise in a way that the innovation tournament does not.
There are also likely common disadvantages to both methods that are important to note. In general, people tend to be inaccurate reporters on what would change their own behavior . Outputs from both methods may suffer from biased sampling, as all data comes from participants amenable to collaborating with researchers. This may have contributed to the lack of implementation strategies from both methods related to engaging consumers; had we included youth and their families in each participatory design process, results may have differed.
We note several study limitations. First, for both participatory design methods, we had relatively small samples sizes. While our observations and interviews reached thematic saturation, our sample for the innovation tournament was lower than others’ . BHTs comprise a largely independent contractor workforce, and they often work independently at schools, which may have contributed to our difficulty engaging them. In addition, the innovation tournament is not necessarily designed for concluding when saturation is reached, as there is limited control over sample size. Second, we did not evaluate the acceptability, feasibility, or effectiveness of any of the identified implementation strategies, so we are unable to evaluate the fit or effectiveness of each participatory design approach, only to highlight points of convergence and divergence. Third, we employed a different framework to generate implementation strategies in the innovation tournament (i.e., the NUDGE framework) than within the traditional contextual inquiry (i.e., ERIC/Behavior Change Wheel for). While our rationale for using the NUDGE framework was to replicate methods of the only other published innovation tournament for mental healthcare, it is possible that this decision may have led to additional divergence in outcomes between the innovation tournament and traditional contextual inquiry. Fourth, as noted above, we did not interview youth and families in this study, nor did we invite them to participate in the innovation tournament, which may have led to a gap in identified implementation strategies. This underscores the potential for exclusion of select stakeholder groups to greatly impact the output of a participatory design approach. Finally, it remains an open question whether one might find the same points of overlap and divergence in output between observations and interviews with an innovation tournament when conducted in another setting (e.g., within a large healthcare system or with providers outside of the mental health space). Our hope is that proposed strengths and limitations of each method delineated in Table 3 can help support researchers working in other healthcare spaces to determine which method will best address their implementation questions; however, this remains a question for future research.
This study also has some notable strengths. Our use of two separate and blinded research teams to execute each participatory design method concurrently allowed us to be confident in the independence of the findings from each method. Importantly, this study represents the first effort to compare the results that emerge from two methods of participatory design, setting the stage for future research to refine selection and execution of participatory methods to optimize the design of effective implementation strategies.
This study indicates that two methods of participatory design—observations/interviews and an innovation tournament—yield substantial overlap in the information obtained that can be used for implementation strategy development. However, in this study, the observations and interviews resulted in more specific and tailored implementation strategies than did the innovation tournament. That said, there was also unique output generated by each. Given the time and resources required to engage in comprehensive contextual inquiry, selecting the best participatory design approach to inform implementation strategies necessitates that researchers carefully consider what each method offers (strategies relating to adapting and tailoring to context vs. incentives) and weigh the resources available (e.g., personnel time vs. money) to invest in the process. This study advances our understanding of the strengths and limitations of different approaches to participatory design, which is critical for helping implementation researchers and policy makers select the approach best suited to address their implementation question.
Availability of data and materials
Data will be made available upon request. Requests for access to the data can be sent to the corresponding author.
Additional information about the coding system and training procedures are available from the first author upon request.
Behavioral health technician
Applied behavior analysis
Sales AE, Wilson PM, Wensing M, Aarons GA, Armstrong R, Flottorp S, et al. Implementation science and implementation science communications: our aims, scope, and reporting expectations. Implement Sci. 2019;14(1):77. https://doi.org/10.1186/s13012-019-0922-2.
Powell BJ, Beidas RS, Lewis CC, Aarons GA, McMillen JC, Proctor EK, et al. Methods to improve the selection and tailoring of implementation strategies. J Behav Health Serv Res. 2017;44(2):177–94. https://doi.org/10.1007/s11414-015-9475-6.
Lane-Fall MB, Curran GM, Beidas RS. Scoping implementation science for the beginner: locating yourself on the “subway line” of translational research. BMC Med Res Methodol. 2019;19(1):133. https://doi.org/10.1186/s12874-019-0783-z.
Waltz TJ, Powell BJ, Fernández ME, Abadie B, Damschroder LJ. Choosing implementation strategies to address contextual barriers: diversity in recommendations and future directions. Implement Sci. 2019;14(1):42. https://doi.org/10.1186/s13012-019-0892-4.
Kok G, Gottlieb NH, Peters GJY, Mullen PD, Parcel GS, Ruiter RAC, et al. A taxonomy of behaviour change methods: an intervention mapping approach. Health Psychol Rev. 2016;10(3):297–312. https://doi.org/10.1080/17437199.2015.1077155.
Robertson T, Simonsen J. Challenges and opportunities in contemporary participatory design. Des Issues. 2012;28(3):3–9. https://doi.org/10.1162/DESI_a_00157.
Simonsen J, Robertson T. Routledge international handbook of participatory design. London: Routledge; 2013.
Dopp AR, Parisi KE, Munson SA, Lyon AR. A glossary of user-centered design strategies for implementation experts. Transl Behav Med. 2019;9(6):1057–64. https://doi.org/10.1093/tbm/iby119.
Green AE, Fettes DL, Aarons GA. A concept mapping approach to guide and understand dissemination and implementation. J Behav Health Serv Res. 2012;39(4):362–73. https://doi.org/10.1007/s11414-012-9291-1.
Fernandez ME, ten Hoor GA, van Lieshout S, Rodriguez SA, Beidas RS, Parcel G, et al. Implementation mapping: using intervention mapping to develop implementation strategies. Front Public Health. 2019;7(1):158. https://doi.org/10.3389/fpubh.2019.00158.
Davis M, Beidas RS. Refining contextual inquiry to maximize generalizability and accelerate the implementation process. Implement Res Pract. 2021;2(1):1–4. https://doi.org/10.1177/2F2633489521994941.
Ranard BL, Ha YP, Meisel ZF, Asch DA, Hill SS, Becker LB, et al. Crowdsourcing--harnessing the masses to advance health and medicine, a systematic review. J Gen Intern Med. 2014;29(1):187–203. https://doi.org/10.1007/s11606-013-2536-8.
Terwiesch C, Mehta SJ, Volpp KG. Innovating in health delivery: the Penn medicine innovation tournament. Healthcare. 2013;1(1–2):37–41. https://doi.org/10.1016/j.hjdsi.2013.05.003.
Asch DA, Terwiesch C, Mahoney KB, Rosin R. Insourcing health care innovation. N Engl J Med. 2014;370(19):1775–7. https://doi.org/10.1056/NEJMp1401135.
Stewart RE, Williams N, Byeon YV, Buttenheim A, Sridharan S, Zentgraf K, et al. The clinician crowdsourcing challenge: using participatory design to seed implementation strategies. Implement Sci. 2019;14(1):63. https://doi.org/10.1186/s13012-019-0914-2.
Beidas RS, Volpp KG, Buttenheim AN, Marcus SC, Olfson M, Pellecchia M, et al. Transforming mental health delivery through behavioral economics and implementation science: protocol for three exploratory projects. JMIR Res Protoc. 2019;8(2):e12121. https://doi.org/10.2196/12121.
Nuske HJ, Buck JE, Ramesh B, Becker-Haimes EM, Zentgraf K, Mandell DS. Making progress monitoring easier and more motivating: developing a client data collection app incorporating user-centered design and behavioral economics insights. Soc Sci. 2022;11(3):106. https://doi.org/10.3390/socsci11030106.
Azad GF, Locke J, Downey MM, Xie M, Mandell DS. One-to-one assistant engagement in autism support classrooms. Teach Educ Spec Educ. 2015;38(4):337–46. https://doi.org/10.1177/2F0888406415603208.
Lopetegui M, Yen PY, Lai A, Jeffries J, Embi P, Payne P. Time motion studies in healthcare: what are we talking about? J Biomed Inform. 2014;49(1):292–9. https://doi.org/10.1016/j.jbi.2014.02.017.
Nuske HJ, McGhee Hassrick E, Bronstein B, Hauptman L, Aponte C, Levato L, et al. Broken bridges—new school transitions for students with autism spectrum disorder: a systematic review on difficulties and strategies for success. Autism. 2019;23(2):306–25. https://doi.org/10.1177/1362361318754529.
Sutton S, French DP, Hennings SJ, Mitchell JO, Wareham NJ, Griffin S, et al. Eliciting salient beliefs in research on the theory of planned behaviour: the effect of question wording. Curr Psychol. 2003;22(3):234–51. https://doi.org/10.1007/s12144-003-1019-1.
Potthoff S, Presseau J, Sniehotta FF, Breckons M, Rylance A, Avery L. Exploring the role of competing demands and routines during the implementation of a self-management tool for type 2 diabetes: a theory-based qualitative interview study. BMC Med Inform Decis Mak. 2019;19(1):23. https://doi.org/10.1186/s12911-019-0744-9.
Michie S, Atkins L, West R. The behaviour change wheel: a guide to designing interventions. London: Silverback Publishing; 2014.
Waltz TJ, Powell BJ, Chinman MJ, Smith JL, Matthieu MM, Proctor EK, et al. Expert recommendations for implementing change (ERIC): protocol for a mixed methods study. Implement Sci. 2014;9(1):39. https://doi.org/10.1186/1748-5908-9-39.
Dillman DA, Phelps G, Tortora R, Swift K, Kohrell J, Berck J, et al. Response rate and measurement differences in mixed-mode surveys using mail, telephone, interactive voice response (IVR) and the Internet. Soc Sci Res. 2009;38(1):1–18. https://doi.org/10.1016/j.ssresearch.2008.03.007.
Stewart RE, Beidas RS, Last BS, Hoskins K, Byeon YV, Williams NJ, et al. Applying NUDGE to inform design of EBP implementation strategies in community mental health settings. Adm Policy Ment Health. 2021;48(1):131–42. https://doi.org/10.1007/s10488-020-01052-z.
Cidav Z, Mandell D, Pyne J, Beidas R, Curran G, Marcus S. A pragmatic method for costing implementation strategies using time-driven activity-based costing. Implement Sci. 2020;15(1):28. https://doi.org/10.1186/s13012-020-00993-1.
Waltz TJ, Powell BJ, Matthieu MM, Damschroder LJ, Chinman MJ, Smith JL, et al. Use of concept mapping to characterize relationships among implementation strategies and assess their feasibility and importance: results from the expert recommendations for implementing change (ERIC) study. Implement Sci. 2015;10(1):109. https://doi.org/10.1186/s13012-015-0295-0.
Powell BJ, Waltz TJ, Chinman MJ, Damschroder LJ, Smith JL, Matthieu MM, et al. A refined compilation of implementation strategies: results from the expert recommendations for implementing change (ERIC) project. Implement Sci. 2015;10(1):1–14. https://doi.org/10.1186/s13012-015-0209-1.
Candon M, Williams N, Zentgraf K, Buttenheim A, Bewtra M, Beidas RS, et al. Variation in stakeholder preferences for implementing evidence-based practices in behavioral health care. Psychiatr Serv. 2022.
Asch DA, Rosin R. Innovation as discipline, not fad. N Engl J Med. 2015;373(7):592–4. https://doi.org/10.1056/nejmp1506311.
The authors gratefully acknowledge that this study would not be possible without our collaborating agencies in the City of Philadelphia. We would also like to thank all of the individuals who participated in the study and all who helped recruit those clinicians, making it possible.
Funding for this research project is supported by the following grants from NIH: NIMH P50113840 (MPIs: Beidas, Buttenheim, Mandell) and NIDA K23 DA 048167 (PI: Rebecca Stewart).
Ethics approval and consent to participate
All procedures have been approved by the City of Philadelphia and University of Pennsylvania IRB.
Consent for publication
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Becker-Haimes, E.M., Ramesh, B., Buck, J.E. et al. Comparing output from two methods of participatory design for developing implementation strategies: traditional contextual inquiry vs. rapid crowd sourcing. Implementation Sci 17, 46 (2022). https://doi.org/10.1186/s13012-022-01220-9
- Participatory design
- Evidence-based practice
- Implementation science
- Implementation strategy design