Hostname: page-component-78c5997874-dh8gc Total loading time: 0 Render date: 2024-11-10T20:49:43.188Z Has data issue: false hasContentIssue false

Using a process improvement approach to identifying barriers to research in a CTSA hub environment

Published online by Cambridge University Press:  19 August 2020

Douglas Grabenstetter*
Affiliation:
Mechanical Engineering Department, Industrial Engineering Program, Milwaukee School of Engineering University, Milwaukee, WI, USA Clinical Translational Science Institute, Medical College of Wisconsin, Milwaukee, WI, USA
Ramez Rashid
Affiliation:
Mechanical Engineering Department, Industrial Engineering Program, Milwaukee School of Engineering University, Milwaukee, WI, USA
Jeff Whittle
Affiliation:
Mechanical Engineering Department, Industrial Engineering Program, Milwaukee School of Engineering University, Milwaukee, WI, USA Department of Medicine, Division of General Internal Medicine, Medical College of Wisconsin, Milwaukee, WI, USA Primary Care Division, Clement J. Zablocki VA Medical Center, Milwaukee, WI, USA
*
Address for correspondence: D. Grabenstetter, PhD, CMBB, Mechanical Engineering Department, Milwaukee School of Engineering University, 1025 N Broadway, Milwaukee, WI53202, USA. Email: grabenstetter@msoe.edu
Rights & Permissions [Opens in a new window]

Abstract

For the past 4 years, as part of the National Institutes of Health (NIH) Clinical and Translational Science Award (CTSA) grant award number UL1TR001436, the Clinical Translational Science Institute of Southeast Wisconsin (CTSI) has used process engineering approaches to identify and understand barriers that local researchers and other stakeholders face when engaging in clinical and translational science. We describe these approaches and present preliminary results. We identified barriers from published and unpublished work at other CTSA hubs, supplemented by surveys and semi-structured interviews of CTSI faculty. We then used a multifaceted approach to organize, visualize, and analyze the barriers. We have identified 27 barriers to date. We ranked their priority for CTSI to address based on the barrier’s impact, the feasibility of intervention, and whether addressing the barrier aligned with CTSI’s institutional role. This approach provides a systematic framework to scope and address the “barriers to research problem” at CTSI institutions.

Type
Special Communications
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
© The Association for Clinical and Translational Science 2020

Introduction

While the past two decades have experienced significant advancements in science and medicine, various sources suggest that the average time for scientific findings to be translated into improvements in public health exceeds 10 years [Reference Heller and de Melo-Martin1Reference Morris, Wooding and Grant3]. In part to address this delay, the National Institutes of Health’s National Center for Advancing Translational Sciences (NCATS) has funded over 55 Clinical and Translational Science Award (CTSA) program hubs. CTSA hubs have been tasked to work individually and as a group to speed the translation of research discoveries into improved health and healthcare. This task requires that hub leaders identify key barriers, prioritize where efforts are most likely to yield benefits and appropriately allocate resources. This complex mission is made more difficult by the need to collaborate across organizational entities within a hub and within the CTSA network. Further, within each academic medical center, there are many processes that impact research projects at a macro level, but do not belong to, and often cannot be addressed by, anyone in the research team.

To address similar challenges, industries from automobile [Reference Liker and Meier4] and pharmaceutical [Reference Carleysmith, Dufton and Alria5Reference Sirbu and Baisan8] manufacturing to healthcare provision [Reference Martinez, Chavez-Valdez and Holt9Reference Improta, Cesarelli and Montuori11] have used Lean and Six Sigma process (quality) improvement methods. Further, research has pointed out that CTSA processes lend themselves to Lean Six Sigma (LSS) approaches [Reference Schweikhart and Dembe12]. While a limited number of published works have examined the use of these approaches in the CTSA domain [Reference Schweikhart and Dembe12Reference Schmidt, Shay and Saygin14], we found no examples of the use of LSS tools as part of an integrated methodology to visualize, identify, and prioritize opportunities to address barriers to performing research, one of the key tasks for CTSA hubs. In our paper, we present the approach we have used over the last 3 years at the Clinical Translational Science Institute of Southeast Wisconsin (CTSI), a partnership among eight academic and clinical organizations. Our efficiency team included a process engineering expert, a physician-scientist, a quantitative analyst, and an experienced CTSA evaluator.

Methods and Results

First, in order to inform our research, we searched PubMed, ProQuest Central, and Google with a variety of keywords, including LSS, barriers, translational research, and continuous improvement. We reviewed over 30 papers. Next, within a LSS framework, we organized our efforts using the Six Sigma define, measure, analyze, improve, and control (DMAIC) approach [Reference Montgomery15] and used a variety of tools drawn from both academic and LSS traditions. The present report focuses on the first three phases of DMAIC — defining, measuring, and analyzing the problem, but we also discuss the use of LSS tools to initiate the Improve phase, when we seek to mitigate specific research barriers.

Define and measure

Based on the prior literature (e.g., [Reference Heller and de Melo-Martin1]), we defined our problem as: Researchers working in CTSI institutions experience many barriers that seriously hamper their ability to make medical discoveries and translate them to the bedside and community. Further, they do not have an organized approach to identify and prioritize them.

In our Measure phase, in order to inform our LSS barrier identification efforts, as previously mentioned, we first identified barriers through a literature review, including work by the CTSA hub located at the University of Florida. In addition to a published qualitative analysis,[Reference Behar-Horenstein and Feng16] the University of Florida CTSA investigators shared summary results from a series of surveys conducted from 2011 to 2015. These annual satisfaction surveys included a section on perceived barriers to research. Further, to acquire a deeper understanding of our local barriers, we conducted a series of semi-structured interviews with CTSI stakeholders in the research process [Reference DiCicco-Bloom and Crabtree17]. Stakeholders included individuals managing research relevant processes (e.g., institutional review board administrators, department business managers) and a mix of junior, mid-, and senior-level researchers at institutions affiliated with our CTSI. Twenty-three interviews were conducted over an 11-month period, using an interview guide (Appendix 1 in the Supplementary Material) aimed at: (1) eliciting barriers to research that the stakeholder had recognized, (2) understanding their knowledge of the services offered by CTSI, and (3) identifying services or support they believe would better support their research and research at CTSI institutions more broadly. The confidential face-to-face interviews were conducted by members of the quality and efficiency (Q&E) team, who summarized the interview, then returned the summary to the interviewee to obtain clarifications, corrections, and additional insights post-interview. These semi-structured interviews exposed multiple barriers we had not previously recognized; these were subsequently added to a graphical LSS tool that we describe next.

Once the initial list of barriers was compiled, we organized and displayed them using a graphical LSS tool: The cause-and-effect diagram [Reference Montgomery15]. This process engineering tool is also known as a “Fishbone Diagram,” since it displays the problem (research inefficiency) as the fish head and the causes (in this case, barriers to efficient research) as bones (Fig. 1). The tool presents causes (in our case, barriers) in a hierarchical format.

Fig. 1. Cause-and-effect (fishbone) diagram approach.

For example, referring to Fig. 2, the top-level barrier is described as data issues. The lower level barrier (child), which is a type of data issue, is represented by a horizontal bone and is described as CTSI collaborative database. The various barriers (grandchildren) connected to the collaborative database appear as angular bones.

Fig. 2. A portion of fishbone diagram.

The full list of 27 barriers from the fishbone diagram is displayed in Appendix 2 in the Supplementary Material. This use of the fishbone diagram simplified presentations to key stakeholders (e.g., CTSI leaders and advisory board members), allowing them to see the scope of the problem and its hierarchical nature. It also provided the CTSI team with a graphical method to present projects aimed at mitigating specific barriers. In our hands, the fishbone diagram has been a “living” document; as new barriers are identified, they are categorized and displayed on the drawing and the filename version is incremented. As far as we are aware, this is the first published use of this tool to visualize research barriers.

Analyze

Our next step was to identify the more important barriers. To do this, we added barrier questions to an extant annual survey (CTSI user satisfaction survey), which is used to inform CTSI leadership decision-making. We asked respondents to rank the importance of the aforementioned 27 barriers using a 10-point ordinal scale ranging from 0 (not a barrier) to 10 (extreme barrier — top 1 or 2 barriers to my research). To improve the user experience and increase the accuracy of response, a slider was used (Appendix 3 in the Supplementary Material). During 2019, this survey was sent to 417 individuals who are affiliated with CTSI service, support, and programs, 75 (18%) of whom responded., Of the respondents, 56% were male, 59% were engaged in clinical/patient-oriented research, 16% in basic/lab research, and 8% worked on social science research. We categorized research experience as 7 or more years (60% of respondents), 3–7 years (28%), or less than 3 years (12%). To ensure responses were informed, we limited the barriers portion of the survey to individuals with 3 or more years of research experience and who had conducted research within the previous year.

We present the top 10 barriers in Table 1 as measured by the mean score for each barrier. Lack of Time to Conduct Research was considered the most important barrier by the respondents.

Table 1. Barriers survey results

We next used a multicriteria ranking tool (Fig. 3) to incorporate pragmatic factors into our decision-making process This tool used four factors to assess which barriers were the most appropriate targets for CTSI efforts. First, Q&E, assessed the barrier’s impact on the overall efficacy and speed of the research process, as measured by the mean score from the barriers survey (for the remaining factors, project team members assigned a score from 1 to 10 for each barrier). The second factor, CTSI influence, reflects the perceived ability of the CTSI to influence the barrier, either directly or through relationships. The more influence CTSI has a barrier, the higher the score for this factor. Third, cost to CTSI, assesses the likely cost to CTSI of mitigating the barrier; the lower the cost to mitigate the barrier, the higher its score on this factor. Finally, importance to the CTSA grant reflects whether addressing the impediment had been an implicit or explicit “deliverable” in our funded proposal. If it was judged to be an explicit deliverable, it was ranked highly.

Fig. 3. Multicriteria ranking tool and guidelines.

Next, in consultation with the local CTSA management team, we weighted the importance of each of the four factors to decision makers. Not surprisingly, given our dependence on grant funding, importance to the CTSA grant was weighted highest (weight = 4), while cost to the CTSI was weighted lowest (weight = 1).

Comparing the multicriteria rank of “Lack of Time to Conduct Research” and “Unawareness of Shared Resources and Equipment” illustrates the importance of this step. Although “Lack of Time to Conduct Research” was identified as the most important barrier to research in our CTSI user survey, our multicriteria ranking tool suggested that it was nowhere near the best target for intervention. It scored low on CTSI as the CTSI does not have an influence on the time faculty members have protected for research – and “Importance to the CTSA Grant” increases in faculty protected time was not a key NIH expectation in either current or expected requests for CTSA awardees. Instead, our tool scored Unawareness of Shared Resources and Equipment as the best target for intervention. To illustrate how the calculation was executed, we note that the impact of Unawareness of Shared Resources and Equipment on Q&E was ranked as moderate (4.7). However, given that the CTSI is able to directly address this barrier, it was ranked as a high impact (9). Likewise, since the projected cost to address awareness is low and awareness of CTSA support capabilities is a grant imperative, both were ranked as having a high impact. Thus, the multicriteria ranking tool score is (4.7 * 1.5) + (9 * 2) + (8 * 1) +(9 * 4) = 69.0. In contrast, note that Lack of Time to Conduct Research receives a score of only (5.9 * 1.5) + (1 * 2) + (1 * 1) + (1 * 4) = 15.9.

Thus, our multicriteria ranking tool is complementary to the information we gain from our survey of CTSI researchers, described above. The tool allows us to incorporate input from subject-matter experts and important stakeholders to calculate multicriteria rankings of the expected value of focusing CTSI efforts on mitigating specific barriers, enabling rationale selection of intervention targets. By using a formal, defined, process, we are more able to explain to leaders why a specific barrier is (or is not) an appropriate target for CTSI resources.

Improve Phase

While occasionally a barrier might be easily mitigated, in our experience, most of the barriers that have come to our attention are of sufficient complexity that we have used the DMAIC approach for projects to address them. To facilitate the define (problem definition) step in the process of addressing a barrier, we use a modified Six Sigma project definition worksheet (Fig. 4). Working alone, or with a project team member, the project sponsor completes the worksheet to begin the project scoping process; the worksheet then guides our team’s initial conversations with the project sponsors and provide an initial point of engagement with self-guided project teams. This is similar to the method described by others [Reference Improta, Cesarelli and Montuori11]. An example of a resulting project was one which helped to address the barrier “Unawareness of shared resources and equipment”. The project was conducted with one of the CTSA collaborating institutions such that a team of industrial engineering students worked with the Medical College of Wisconsin Department of Medicine. The DMAIC approach was used by the students, the current process was mapped, and a new improved process was successfully developed and documented along with an equipment tracking tool.

Fig. 4. Scoping document.

Discussion

We have presented a novel stepwise approach to the problem of identifying and addressing barriers to efficiently conducting high-quality translational research, based upon the DMAIC problem-solving process. This approach allowed us to objectively determine a comprehensive set of these barriers and to funnel down to key targets. Furthermore, our approach’s novel use of hierarchical visualization (i.e., fishbone diagram) and a multicriteria ranking tool, provided effective means of communication with CTSI leadership.

While these techniques have been widely used in various industries, we believe they have not been widely used by CTSA awardees. We believe that future users in CTSA environments should describe their implementation of these and similar techniques, since their usefulness may vary across institutions. We also encourage descriptions of successful resolution of specific problems [Reference Schmidt, Goros and Parsons13, Reference Schmidt, Shay and Saygin14]. Broader reporting of positive and negative experiences at individual CTSA awardees may reveal factors that influence whether specific solutions work at some institutions but not others.

This project has implications for the increasing emphasis on dissemination and implementation (D&I) activities and D&I science among CTSA hubs. Originally conceived as a method to enhance the adoption and integration of medical innovations into everyday clinical practice [Reference Bauer, Damschroder and Hagedorn18, Reference Mittman, Weiner, Proctor and Handley19], several CTSA hubs have recognized that similar processes can enhance the translational research process. Instructive examples [Reference Leppin and Mahoney20, Reference Quanbeck and Mitchell21] and a theoretical model for integrating D&I into a translational research framework have been published [Reference Leppin and Mahoney20]. Since these studies have demonstrated that D&I efforts can enhance the impact of a wide range of CTSA activities, the techniques we describe can help CTSA leaders prioritize areas where D&I resources, which are often in short supply, are best allocated. Moreover, the use of the prioritization process we suggest, which allows for comparison of disparate areas of research, may suggest that D&I requires increased attention by hub leaders.

We do note that while several authors have emphasized D&I as a method to enhance translation across the T1–T5 spectrum, we have focused on increasing the efficiency of research in general – most, but far from all, of the barriers we discussed are significant within a translational stage, though some (e.g., barriers to collaborating across partners) clearly have implications for translation across stages. When one hub identifies a method to increase efficiency, D&I has a crucial role in making sure that that method moves rapidly within the CTSA consortium and beyond. Indeed, given that the value of the processes we describe herein has been recognized for decades in other industries, D&I may have an important role to play in ensuring that they are adopted across the CTSA consortium.

We note several caveats regarding our description and identification of the barriers. First, we identified barriers at just two CTSA hubs; other CTSA hubs may have other barriers we did not identify. Within our hubs, both our surveys had limited response rates, so that non-response bias could be present. However, these response rates are not lower than other surveys of translational researchers [Reference Estapé-Garrastazu, Noboa-Ramos and De Jesús-Ojeda22Reference Dao, Kota and James24]. Similarly, while we conducted just 23 semi-structured interviews at one of our hubs, we note that many reports in the literature have been based on fewer interviews [Reference Boldrup-Tingleff, Hounsgard and Bradley25Reference Mallon, Gooberman-Hill and Blom28]. In addition, we note that the barriers we identified are broadly similar to those described in other works.

We also note that while all CTSA quality improvement teams likely have access to experienced medical researchers, not all have access to robust process engineering expertise (the Milwaukee School of Engineering is a CTSI institution that does have such expertise). For example, it is not trivial to translate barriers from a fishbone diagram into a set of meaningful survey questions or develop a qualitative ranking tool that captures leadership priorities. At the same time, we realize that, in many cases, such expertise may indeed be present within an affiliated medical center’s quality improvement department [Reference Martinez, Chavez-Valdez and Holt9Reference Improta, Cesarelli and Montuori11].

Finally, we note that others may use different LSS tools or implement them differently. For example, one alternative to using the fishbone diagram for data visualization is the affinity diagram [Reference Munro, Govindarajan and Zyrmiak29]. Similarly, others could question several choices we made for our qualitative ranking tool. For example, one could use a metric other than the mean score of responses as the ranking for a barrier’s impact on research Q&E or consider a larger range for ranking within the tool. Others could also use different weights or even different considerations to inform a pragmatic decision about what barriers should be targeted for intervention and score them differently. For instance, we note that more quantitative approach, such as the analytical hierarchy process (AHP) [Reference Saaty30] may likewise be utilized. Further, it is noted that the weights and considerations themselves may be modified over time if the decision-making culture changes. We believe our approach, which reflected the input of multiple stakeholders, was useful, but welcome efforts to improve it.

In summary, we note that use of process engineering tools provides a structured approach that may assist CTSA hubs in identifying the processes where their efforts can most favorably impact the Q&E of translational research at their institution. These techniques also encourage hub leaders to explicitly consider the factors likely to influence successful intervention on these processes. Lastly, the use of explicit techniques such as the one we outline facilitates clear communication about why specific interventions are undertaken by CTSA hubs.

Supplementary material

To view supplementary material for this article, please visit https://doi.org/10.1017/cts.2020.522.

Acknowledgments

The authors wish to express their gratitude to the reviewers for their comments and recommendations which helped to improve the quality of our paper. We also acknowledge the seminal input of Emily Connors, MS, who initiated this project. Finally, we thank the faculty members whose participation in surveys and qualitative interviews made the project possible.

The project described was supported by the National Center for Advancing Translational Sciences, National Institutes of Health, award number UL1TR001436. The content is solely the responsibility of the author(s) and does not necessarily represent the official views of the NIH.

Disclosures

The authors have no conflicts of interest to declare.

References

Heller, C, de Melo-Martin, I. Clinical Translational Science Awards: can they increase the efficiency and speed of Clinical and Translational Research? Academic Medicine 2009; 4: 424432.CrossRefGoogle Scholar
Lenfant, C. Clinical research to clinical practice — lost in translation? New England Journal of Medicine 2003; 349: 868874.CrossRefGoogle ScholarPubMed
Morris, ZS, Wooding, S, Grant, J. The answer is 17 years, what is the question? Understanding time lags in translational research. Journal of the Royal Society of Medicine 2011; 104: 510520.CrossRefGoogle ScholarPubMed
Liker, J, Meier, D. The Toyota Way Fieldbook – A Practical Guide for Implementing Toyota’s 4P’s. McGraw Hill, 2005.Google Scholar
Carleysmith, S, Dufton, A, Alria, K. Implementing Lean Sigma in pharmaceutical research and development: a review by practitioners. R&D Management 2009; 1: 95106.Google Scholar
Selko, A. Pharmaceutical’s Lean Prescription. Industry Week [Internet], 2011 [cited July 11, 2011]. (https://www.industryweek.com/leadership/companies-executives/article/21941921/pharmaceuticals-lean-prescription)Google Scholar
Kuiper, K, van Raalte, M, Does, RJ. Quality quandaries: improving the overall equipment effectiveness at a pharmaceutical company. Quality Engineering 2014; 6: 478483.CrossRefGoogle Scholar
Sirbu, D, Baisan, CE. Lean Manufacturing Benefits in a Pharmaceutical Production Plant Applied Mechanics and Materials 2014; 657: 991995.CrossRefGoogle Scholar
Martinez, EA, Chavez-Valdez, R, Holt, NF, et al. Successful implementation of a perioperative glycemic control protocol in cardiac surgery: barrier analysis and intervention, using Lean Six Sigma. Anesthesiology Research and Practice 2011; 2011: 110.CrossRefGoogle ScholarPubMed
Martens, L, Gooder, G, Wold, JFH, et al. Structured syncope care pathways based on Lean Six Sigma methodology optimizes resource use with shorter time to diagnosis and increased diagnostic yield. PLoS One [Internet], 2014 [cited June 13, 2014]. (https://doi.org/10.1371/journal.pone.0100208)CrossRefGoogle Scholar
Improta, G, Cesarelli, M, Montuori, P, et al. Reducing the risk of healthcare-associated infections through Lean Six Sigma: the case of the medicine areas at the Federico II University Hospital in Naples (Italy). Journal of Evaluation in Clinical Practice 2018; 2: 338346.CrossRefGoogle Scholar
Schweikhart, SA, Dembe, AG. The Applicability of Lean and Six Sigma Techniques to Clinical and Translational Research. Journal of Investigative Medicine 2009; 57: 748755.CrossRefGoogle ScholarPubMed
Schmidt, S, Goros, M, Parsons, HM, et al. Improving initiation and tracking of research projects at an academic health center: A Case Study. Evaluation & the Health Professions 2017; 3: 372379.CrossRefGoogle Scholar
Schmidt, S, Shay, LA, Saygin, C, et al. Improving pilot project application and review processes: A novel application of lean six sigma in translational science. Journal of Clinical and Translational Science 2018; 2: 135138.CrossRefGoogle ScholarPubMed
Montgomery, D. Statistical Quality Control, 7th ed. Hoboken, NJ: John Wiley & Sons, Inc., 2013.Google Scholar
Behar-Horenstein, LS., Feng, X. What open-ended comments reveal: an analysis of a Clinical Translational Science Institute’s Annual Surveys. The Qualitative Report, 2018; 23(8), 20032018. (https://nsuworks.nova.edu/tqr/vol23/iss8/15)Google Scholar
DiCicco-Bloom, B, Crabtree, BF. The qualitative research interview. Medical Education 2006; 40: 314–21.CrossRefGoogle ScholarPubMed
Bauer, MS, Damschroder, L, Hagedorn, H, et al. An introduction to implementation science for the non-specialist. BMC Psychology 2015; 3:112.CrossRefGoogle ScholarPubMed
Mittman, BS., Weiner, BJ., Proctor, EK., & Handley, MA. Expanding D&I science capacity and activity within NIH Clinical and Translational Science Award (CTSA) Programs: guidance and successful models from national leaders. In Implementation Science 2015; 10(S1): A38.CrossRefGoogle Scholar
Leppin, AL, Mahoney, JE, et al. Situating dissemination and implementation sciences within and across the translational science spectrum. Journal of Clinical and Translational Science 2019; 4(3): 152158.CrossRefGoogle Scholar
Quanbeck, A, Mitchell, J et al. Building capacity for dissemination and implementation to maximize impact in a CTSA: The University of Wisconsin story. Journal of Clinical and Translational Science 2019; 4(3): 209215.CrossRefGoogle Scholar
Estapé-Garrastazu, ES, Noboa-Ramos, C, De Jesús-Ojeda, L, et al. Clinical and Translational Research Capacity Building Needs in Minority Medical and Health Science Hispanic Institutions. Clinical Translational Science 2014; 7: 406412.CrossRefGoogle ScholarPubMed
Nokes, KM, Nelson, DA, McDonald, MA, et al. Faculty perceptions of how community-engaged research is valued in tenure, promotion, and retention decisions. Clinical Translational Science Journal 2013; 4: 259266.CrossRefGoogle Scholar
Dao, HD, Kota, P, James, JA, et al. Assessment of Translational and Interdisciplinary Clinical Research at an Oklahoma Health Sciences Center. Journal of the Oklahoma State Medical Association 2015; 3: 93101.Google Scholar
Boldrup-Tingleff, E, Hounsgard, L, Bradley, SK, et al. A matter of trust and distrust: a qualitative investigation of parents’ perceptions about the use of mechanical restraint on their adult children in a forensic psychiatric setting. Journal of Forensic Nursing 2019; 2: 120130.CrossRefGoogle Scholar
Brys, NA, Whittle, J, Safdar, N. Development of a veteran engagement toolkit for researchers. Journal of Comparative Effectiveness Research 2018; 6: 595602.CrossRefGoogle Scholar
Long, JC, Winata, T, Debono, D, et al. Process evaluation of a behavior change approach to improving clinical practice for detecting hereditary cancer. BMC Health Services Research 2019; 1: 180.CrossRefGoogle Scholar
Mallon, C, Gooberman-Hill, R, Blom, A, et al. Surgeons are deeply affected when patients are diagnosed with prosthetic joint infection. PLoS One [Internet], 2018 [cited November 28, 2018]. (10.1371/journal.pone.0207260)CrossRefGoogle Scholar
Munro, RA, Govindarajan, R, Zyrmiak, DJ. Affinity Diagram. Chapter 7 In: The Certified Six Sigma Greenbelt Handbook. 2nd ed. Milwaukee, WI: ASQ Quality Press; 2015. 103104.Google Scholar
Saaty, TL. Fundamentals of Decision Making and Priority Theory. Pittsburgh, PA: RWS Publications, 2001.Google Scholar
Figure 0

Fig. 1. Cause-and-effect (fishbone) diagram approach.

Figure 1

Fig. 2. A portion of fishbone diagram.

Figure 2

Table 1. Barriers survey results

Figure 3

Fig. 3. Multicriteria ranking tool and guidelines.

Figure 4

Fig. 4. Scoping document.

Supplementary material: File

Grabenstetter et al. supplementary material

Appendices 1-3

Download Grabenstetter et al. supplementary material(File)
File 262.5 KB