key: cord-0752791-lo8jyuib authors: Musshafen, Leslie A.; Poger, Jennifer M.; Simmons, William R.; Hoke, Alicia M.; Hanson, Laura N.; Bondurant, Whitney W.; McCullough, Jody R.; Kraschnewski, Jennifer L. title: Strengthening the clinical research workforce through a competency-based orientation program: Process outcomes and lessons learned across three academic health institutions date: 2021-09-13 journal: Journal of Clinical and Translational Science DOI: 10.1017/cts.2021.852 sha: 17e2b020c0b996d383b0d9018f77093bc41eaea6 doc_id: 752791 cord_uid: lo8jyuib Clinical research coordinators are increasingly tasked with a multitude of complex study activities critical to scientific rigor and participant safety, though more than half report not receiving appropriate training. To determine the reproducibility of an established clinical research workforce orientation program, collaborative partners across Clinical and Translational Science Award institutions seeded core principles and structure from Mayo Clinic’s Clinical Research Orientation program within Penn State University and the University of Mississippi Medical Center from 2019 to 2021. Training concepts were established and tied to those domains deemed critical by the Joint Task Force for Clinical Trial Competency for the conduct of clinical research at the highest levels of safety and quality possible. Significant knowledge and confidence gains and high overall program satisfaction were reported across participants and partner sites, despite programs being required to pivot from traditional, in-person formats to entirely virtual platforms as a result of the COVID-19 pandemic. The successful standardization and translation of foundational clinical research training has important efficiency and efficacy implications for research enterprises across the USA. Well-trained research support staff are a key pillar of research infrastructure in promoting the highest quality translational and clinical science. Though overall responsibility for the conduct of a study lies with the principal investigator, clinical research coordinators (CRCs) are increasingly delegated investigator responsibilities and complex study activities [1] [2] [3] . While core activities traditionally related to the recruitment, consenting, and care coordination of study participants, the role of CRCs has evolved to routinely include submission and maintenance of regulatory documents, study budget preparation and management, collection and processing of specimens, and liaison for healthcare providers, other study personnel, study participants, regulatory bodies, and sponsors alike [1] . Given the vital contributions of CRCs, the Clinical and Translational Science Award (CTSA) Research Coordinator Taskforce was created with a focus on enhancing CRC support and training [1] . The Taskforce surveyed all active CTSAs in 2008 and found that less than half (45%) of responding CRCs reported receiving appropriate training on all of the tasks their position required [1] . Further, Inspectional Observation Summaries by the US Food & Drug Administration in 2017 found the most frequently cited audit deficiency was the failure to establish, maintain, and follow standard operating procedures [4] . Lack of proper training and standardized processes to carry out research best practices coupled with expanded responsibilities of CRCs creates barriers to conducting efficacious and ethically sound research [5] . As such, the Taskforce recommended institutions conduct gap analyses of their training programs to determine areas of weakness in CRC training, including core competencies and career development [1] . Follow-up surveys continue to expose gaps in training despite evidence that minimal training and experience are correlated with lower self-reported competency among CRCs to conduct the myriad research responsibilities they are tasked [6] . This further increases the risk of research staff turnover and burnout [7] . Identifying and learning from innovators of successful and sustainable CRC training programs is an initial step toward best practice implementation at other institutions. Mayo Clinic was one of the first 12 institutions to receive the CTSA from the National Institutes of Health. The resulting Mayo Clinic Center for Clinical and Translational Science (CCaTS) has created, sustained, and adapted educational and operational resources to train and support the coordination of clinical research studies since 2015. By leveraging institutional resources as well as extramural funding, CCaTS has created a centralized research infrastructure that avoids costly duplication and inefficient silos. CCaTS established the Mayo Clinic Clinical Research Orientation program (MCCRO) to provide new research staff with the foundational knowledge needed to safely and efficiently conduct clinical trials of the highest quality, as well as to provide continuing professional development opportunities for faculty, trainees, and other research team members throughout their careers [8] . The MCCRO program serves as the standard onboarding program for all clinical research support staff at the Mayo Clinic, focusing on key concepts and processes throughout all stages of the clinical research lifecycle including study development (investigator-initiated studies) or assessment (externally initiated studies), startup, conduct, closure, and results dissemination [8] . Subsets of tasks at each stage are detailed according to research regulations and internal processes at the Mayo Clinic (Fig. 1) . Informed by the Morrison, Ross, and Kemp model of instructional design [9] , the MCCRO program relies on a blended learning/ flipped classroom approach to deliver material and promote knowledge retention through completion of 20 h of online training modules, 20 h of in-person classroom instruction led by subjectmatter experts (SMEs), and 40 h of hands-on, mentor-guided work in the participant's assigned unit (Supplemental Fig. 1 ). Upon completion of the program, participants are expected to exhibit competency to: recognize the basic principles of human subjects protection and the significance of their role in Good Clinical Practice; identify key activities involved in coordinating a research protocol; and build a list of professional contacts and resources to support the coordination of a clinical study. Participant satisfaction surveys of program offerings, an exam required of all CRCs at the end of their first year of employment, and input from program mentors are used as program evaluation checkpoints and feedback loops. Nearly 1300 professionals have completed the MCCRO program to date with overwhelmingly positive feedback and improved operational efficiencies, making it a well-suited training program model for other CTSAs across the country. Penn State University College of Medicine (PSU) and the University of Mississippi Medical Center (UMMC) were both separately grappling with developing sustainable clinical research training programs in 2018. PSU conducted a needs assessment to gain a better understanding of the education, experience, research interests, and training needs of those on their campus. Roughly 80% of the 130 responses received from those who selfidentified as PSU research support personnel expressed interest in training related to study development/management. The majority (75%) reported not receiving a formal research orientation upon hire, highlighting a stark gap between needs and resources (email communication, June 2021). At UMMC, the institution's clinical research portfolio was rapidly expanding with the construction of a dedicated Clinical Research and Trials Unit within the main hospital. Historically, individual departments had been charged with training clinical research support professionals in their area, resulting in a disjointed and incongruous training schema across the institution. Gaps in training were magnified in areas with less robust research infrastructure. The need for a standardized, institution-wide training program as a critical component to the institution's clinical research infrastructure and continued success became apparent to institutional research leaders. Representatives from each site recognized an opportunity for collaboration and set out in 2019 to determine the reproducibility of the MCCRO program through implementation at PSU and UMMC. Each institution offered an opportunity to explore programmatic adaptations based on size, research portfolio, and unique training needs. Implementation of the MCCRO at partner sites, PSU and UMMC, and the subsequent evaluations of each are reported. With funding through a CTSA administrative supplement (3UL1TR002014-03S2) and support from senior leadership across sites, a core team was defined to include experienced personnel from central research offices at each organization (Table 1) . A kickoff visit to Mayo Clinic and participation in subsequent virtual orientation sessions allowed core team members from PSU and UMMC the opportunity to engage in the MCCRO program first-hand and meet with program staff. Existing program structures, training content, schedules, assessments, and related templates were shared among partners. PSU and UMMC then focused efforts on identifying both commonalities between the sites, as well as unique program adaptations that would be needed at their respective institutions. Subsequent core team visits to PSU and UMMC were organized to gain a better understanding of each site's organizational culture, structure, and workforce. Monthly virtual meetings were held with all partner sites to review implementation efforts and troubleshoot challenges. Common program components implemented at partner sites The Core Competency Framework (CCF) established by the Joint Task Force for Clinical Trial Competency asserts that clinical research professionals should exhibit competency in up to eight domains to safely and ethically carry out their roles [10] . All partner sites agreed at the outset of the collaboration that the CCF would serve as the foundation by which training objectives and participant knowledge assessments would be established and tied. Orientation sessions at both partner sites were developed to address topics related to study startup (i.e., activation, setup), study conduct (i.e., recruitment, fiscal management, regulatory oversight, reporting) and study closure (i.e., data management and 2 Leslie A. Musshafen et al. dissemination), as guided by the CCF [10] . In doing so, training was harmonized across locations. Pre-/posttest (PPT) assessments were developed at partner sites based on their respective orientation curriculum to evaluate participants' knowledge of key competencies within clinical research both before and after orientation participation. CCF competencies across nearly all domains were assessed, though specific CCF sublevels tied to assessment questions varied by site. A total of six CCF competency sublevels across three domains − Ethical and Participant Safety Considerations, Clinical Study Operations (Good Clinical Practices), and Data Management and Informatics − were addressed across all sites to allow for comparisons of knowledge change across institutions. PSU revised their PPT assessment over the duration of this collaboration as a function of curriculum changes and participant feedback; however, crosssite competencies remained the same. Composite variables were created for questions that addressed the same competency. McNemar's test was employed to compare individual knowledge scores from pretest to posttest. The Wilcoxon signed-rank test was used to compare percentage correct medians at pretest to those at posttest and also to evaluate differences between sites. Both prior to and immediately following orientation, participants were asked to use a four-point Likert scale to report their level of confidence in performing tasks related to study development, participant consent, study management, and results dissemination. An ordinal logistic regression with a generalized estimating equation model was employed to evaluate repeated measures across time points. Odds ratios were calculated to compare confidence levels from preorientation to postorientation. Comparisons between median preorientation and postorientation confidence levels were drawn using the Wilcoxon signed-rank test. Sites also assessed participant satisfaction of content presented and presenter effectiveness as a measure of program efficacy using a Likert scale. Space for qualitative feedback was included within evaluations to allow participants an opportunity to report experiences or provide suggestions not otherwise captured in quantitative survey items. Participants at PSU completed daily session evaluations, while those at Mayo Clinic and UMMC completed evaluations at the conclusion of the program. Finally, with support from research leadership at partner sites, site-specific program eligibility and waiver requirements were outlined. Partner sites collaborated with their respective Human Resources office and central research service areas to identify eligible program participants. Faculty and staff instructors were identified based on subject-matter expertise. PSU and UMMC agreed that offering quarterly orientation sessions would allow adequate accommodation for all eligible staff. To develop a plan for orientation content and structure, PSU utilized results from a research support staff assessment, two employee focus groups, and exploration meetings with over 20 content experts. The PSU program was developed to foundationally support investigator-initiated research initiatives and as a complement to industry-sponsored research topics offered in the MCCRO curriculum. Translational research topics in behavioral and community-based spheres were added focus areas. The The resulting orientation program at UMMC was coined CREW (Clinical Research Education for the Workforce) and housed within the Office of Clinical Trials. CREW originally included 20 h of foundational online instruction, along with 5 half-day, in-person sessions. After the first cohort expressed challenges stepping away from assigned job areas for several consecutive days; however, the schedule was altered to limit in-person sessions to two full days (Supplemental Figs. 3 and 5) . The 20 h of online instruction remained unchanged. As the COVID-19 pandemic spread across the USA in 2020, all sites pivoted to move sessions and participant interactions entirely online with minimal programing disruptions. Mayo Clinic transitioned from a blended in-person/virtual orientation to a completely virtual offering using Blackboard Collaborate and Zoom. PSU delivered their program synchronously via the Zoom platform, maximizing interactivity through built-in chat, breakout room, and annotation features. UMMC transitioned the in-person component of its training to the Cisco Webex platform and utilized games, polls, and quizzes to maximize engagement. PSU received support to launch the orientation from the Vice Dean for Research and Graduate Studies. As a new institutional offering without demonstrated efficacy, managers were permitted to opt in /opt out their employee(s). Invitations were sent to managers and employees whose job title fell into the defined PSU Research job family. Program coordinators specifically aimed to identify eligible employees within 3 to 4 months of their start date; participants with varying lengths of service regularly requested and attended orientation, however. In all, 71 participants completed the PSU program over five sessions offered from October 2019 to January 2021. At the request of the program developers, the UMMC Associate Vice Chancellor for Research messaged all clinical department Chairs to introduce the program and affirm its alignment with strategic priorities. Posts to an internal listserv and email messages to identified eligible staff were then used to broadly announce the program and register participants. Any interested student, staff, or faculty member on campus was invited to attend, though all new employees functioning as a CRC or those new to a CRC role at the institution were required to complete the program within the first 3 months of their hire date. Existing employees functioning as a CRC, regardless of title, were required to complete the program within 1 year. Existing employees could request an exemption with (1) their supervisor's approval, and (2) either Certified Clinical Research Professional or Certified Clinical Research Coordinator credentials, or a passing score (≥80%) on the CREW competency exam. From October 2019 to January 2021, a total of 115 participants completed the UMMC program over five orientation sessions. Only 14 exemptions were requested and approved during that time. Total knowledge scores are reported in Table 2 . PSU cohorts 1−3 and 4−5 are reported separately due to the changes made to the assessment between these cohorts. Sites found significant increases in knowledge scores from pretest to posttest (PSU cohorts 4−5 only). When combining data across sites for the six shared competencies (n = 605), overall knowledge scores from pretest to posttest also significantly increased. No significant differences in crosscompetency knowledge scores were identified between sites. PSU implemented the participant confidence assessment as part of their PPT assessment from the outset of the program's implementation; UMMC adopted the confidence assessment beginning with cohort 3. A significant increase in participant confidence to perform all assessed clinical research activities was observed at both PSU and UMMC, with participants having significantly higher odds of increased confidence postorientation as compared to preorientation (Table 3) . A significant increase in the median confidence level of participants from pretest to posttest was also observed at both sites. Attendees at all sites reported high overall program satisfaction scores and presenter effectiveness. Mayo Clinic and UMMC utilized a five-point scale to gauge satisfaction, to include "poor" (1), "fair" (2), "satisfactory" (3), "good" (4), and "excellent" (5). Respectively, mean program satisfaction were 4.09 and 4.31, with average presenter effectiveness scores of 4.34 and 4.36. PSU used a four-point scale, including "not at all satisfied" (1), "not very satisfied" (2), "satisfied" (3), and "very satisfied" (4). The mean program satisfaction score at PSU was 3.77, with a mean presenter effectiveness score of 3.55. Leslie A. Musshafen et al. Qualitative feedback from participants was overwhelmingly positive. Several participants expressed an appreciation for learning how their duties fit within the research lifecycle, and for the opportunity to share experiences and solutions. Even experienced research team members at PSU commented that the orientation "tied everything together" with regard to the research process, and provided "tips and strategies previously unknown." A UMMC participant expressed appreciation for learning best practices and being able to immediately implement changes based on the program. A Mayo Clinic participant called the program "an eye-opening experience to learn about the key components to a clinical trial." Participants across sites found the networking and career modeling aspects of the program particularly beneficial. This effort successfully confirmed the reproducibility of a clinical research workforce training program across CTSA organizations and reinforced evidence that collaboration supports best practices in standardizing research training and development. An institutional approach to onboarding study staff provides new hires with dedicated support, alleviates burden on study teams, and helps ensure research of the highest safety and ethical standards. Lack of a centralized training program may further exacerbate staffing challenges and turnover often experienced in these positions. The COVID-19 pandemic accentuated the value of an effective and efficient clinical research training program. Academic institutions have experienced turnover in research support positions related to individuals leaving the workforce, temporary interruptions in hiring practices as a result of financial stabilization responses, and competing industry recruitment efforts. As universities rebuild their workforce, the benefit of an established orientation program for new professionals is evident. Results from our programs indicate participants experienced knowledge and confidence gains associated with performing the varied responsibilities of clinical research professionals. Participants were highly satisfied with their respective programs and no significant knowledge differences were noted between sites. We believe the homogeneity of results achieved across programs points to both the benefit and the critical nature of successful collaborations. The value of our synergistic partnership extended beyond positive participant outcomes to also include valuable byproducts from sharing best practices and lessons learned throughout. Each site contributed and benefited from the partnership in different, yet tangible ways. As other institutions consider opportunities to implement new or improve existing clinical research workforce offerings, we propose seven key steps: 1. Assess your landscape (needs of research staff, existing training, SMEs, resources) 2. Identify intended program audience 3. Secure institutional buy-in and support (including funding, as needed) 4. Establish the core framework for orientation (integration within a learning management system, as needed) 5. Establish assessment metrics 6. Develop supplemental/ongoing trainings 7. Market the program Dedicated resources to support both the initial development and the maintenance of programs cannot be understated. Personnel effort is necessary to develop and maintain content, infrastructure to deliver program content, and evaluative methods to measure program results. A mix of talents is suggested to include individuals with clinical research subject-matter expertise, as well as those with experience in curriculum-building, public speaking, and the selected content delivery platform(s). Inclusion of skilled marketing professionals from the outset is also advised to assist with establishing a program "brand" identity and consistent program communications. The described standardized program was implemented at unique research organizations with aspects tailored to account for these distinctions. Results reported are therefore not generalizable to all research organizations or study staff. Further, most program participants continued to perform their position duties while attending orientation sessions. Knowledge and confidence gains may therefore not be directly attributable to the program itself but may have resulted in part due to experience gained over the assessment period. This is limited, however, in that many program participants were seasoned research staff. Self-reported confidence also comes with its own set of limitations, including under-or over-reporting, and is therefore a less reliable indicator of program efficacy [11] . Expansion of the program to other CTSA sites and evaluation of the efficiency and effectiveness of doing so could provide a step towards standardized training and competencies for clinical research study staff. Utilization of a collaborative learning space like the Development, Implementation, and Assessment of Novel Training in Domain-based Competencies, or "DIAMOND" portal to share onboarding processes/documents, competency assessments, standardized job descriptions, and training materials could also serve as a resource towards the goals of improved training and increased efficiencies [12] [13] [14] . Given each institutions' investment in such programs and strong collaborative relationships, Mayo Clinic, PSU, and UMMC are well positioned to support these future endeavors. The critical need for academic health centers to assess the training, support, and career development requirements of clinical research coordinators: Recommendations from the Clinical and Translational Science Award Research Coordinator Taskforce Variability in protocol design complexity by phase and therapeutic area Appendix D: discussion paper. The clinical trials enterprise in the United States: a call for disruptive innovation Food & Drug Administration Education and training of clinical and translational study investigators and research coordinators: a competency-based approach Perceived professional competence of clinical research coordinators Burnout in clinical research coordinators in the United States Operationalizing and Maintaining a Comprehensive Assessment-Driven Clinical Research Orientation. Webinar for the Association of Designing Effective Instruction Moving from compliance to competency: a harmonized core competency framework for the clinical research professional Flawed self-assessment: implications for health, education, and the workplace Training and Assessment Digital Network Competency index for clinical research professionals (CIRCP) -assessment with scoring guide Competency indices to assess the knowledge, skills and abilities of clinical research professionals