Abstract
Background
Measurement of neonatal team resuscitation performance is critical to identify opportunities for improvement and to target education. An effective tool to measure team performance during infant resuscitations is lacking.
Methods
We developed an in-hospital infant resuscitation performance tool (Infa-RePT) using the modified Delphi method. We employed a QI framework and targeted interventions, including the use of role responsibility checklists, mock codes, and an educational video. We tracked Infa-RePT scores, mock code team attendance, and confidence surveys. Our specific aim was to improve Infa-RePT score from a baseline of 7.4 to <5 (lower is better) over a 26-month period.
Results
Twenty-five elements reached >80% consensus as essential components to include on the Infa-RePT. Independent observation showed 86% concordance on checklist items. Simulation (n = 26) and unit-based code (n = 10) Infa-RePT scores showed significant improvement after project start from 7.4 to 4.2 (p < 0.01) with special cause variation noted on control chart analysis. No significant difference was observed between simulations and in-unit codes. Staff confidence self-reports improved over the study period.
Conclusions
Use of a novel scoring tool can help monitor team progress over time and identify areas for improvement. Focused interventions can improve resuscitation team performance.
Impact
-
We developed and used a novel, comprehensive measurement tool for team infant resuscitation performance in both simulation and in-unit settings.
-
Using QI methodology, team performance improved after the enhancement of a mock code simulation program.
-
Review of team performance scores can highlight key areas to target interventions and monitor progress over time.
Similar content being viewed by others
Introduction
Infant resuscitation events are rare emergencies that require critical thinking, teamwork, clinical skills, and effective communication. High-quality team performance during resuscitations is key for successful clinical care of infants.1 Professional organizations have developed clear algorithms to direct teams in the proper steps of resuscitation with guides, including cardiopulmonary resuscitation (CPR), Neonatal Resuscitation Program (NRP),2 and Pediatric Acute Life Support (PALS).3 While resuscitation algorithms give specific recommendations related to practical steps during resuscitation, they do not fully address the complexities of executing those steps in the context of a team with individuals from different training backgrounds and with varying skill levels. Team training focused on crisis resource management and team communication are critical components of successful team performance during resuscitation.4,5,6 Team training in high reliability principles and culture of safety can improve consistency of practice and collaboration across disciplines to improve quality of care.7 Conceptual frameworks, including the Systems Engineering Initiative for Patient Safety model, can help teams understand the complex interactions between human factors including interactions between systems, processes, and outcomes.8,9 Decision support tools have been shown to improve compliance with NRP algorithms.10,11 Use of Read-Do checklists have also been shown to be effective tools for improving team performance.12,13 Other tools have been developed focused on adult emergency response with inclusion of behavioral components and practical steps of resuscitation,14,15,16,17 but to our knowledge, none currently exist that are efficient, facile, infant-focused, and applicable to both in-unit and simulation settings. Tools to reliably track overall infant code team performance are lacking, thus limiting the ability to evaluate metrics critical for optimal outcomes.
Measurement is a key component of effective quality improvement as it allows a team to assess the current state of practice, set goals, and monitor effects of improvement interventions.18 Tracking baseline team performance and identifying recurring themes can allow for focused educational efforts targeted at individualized, specific factors. Monitoring progress over time allows teams to assess effectiveness of interventions. In conjunction with tools that help track performance, simulation practice for rare events can be used to both evaluate and improve performance. Simulation practice and tool use have been shown to improve team resuscitation performance.19,20 Mock code programs allow for team practice in a controlled, consequence-free setting. Multidisciplinary team simulation practice allows for targeted educational efforts to be created and practiced.
Our project goals were to (1) gain consensus on the essential elements of infant team resuscitation and create a simple tool that would include both clinical resuscitation steps and crisis resource management factors, (2) use the tool to facilitate feedback during debriefs for immediate impact, and (3) use the tool to document performance over time, follow trends, and identify specific data-driven learning opportunities and test targeted quality improvement interventions.
Methods
Context
Boston Children’s Hospital (BCH) neonatal intensive care unit (NICU) is an urban 24-bed tertiary/quaternary care unit in Boston, MA, USA with approximately 650 annual admissions, of whom 80% are admitted to the medical service and 20% to the surgical service. All patients are out-born. PALS is used as the default resuscitation algorithm. Resuscitation events are typically attended by multidisciplinary staff, including attending neonatologists, neonatology fellows, neonatal nurse practitioners, respiratory therapists, clinical nurse specialists, nurses, and occasionally subspecialist physicians, pharmacists, and families. Prior to this study, the NICU had a well-established Crisis Resource Management (CRM) program that occurred off-site in our simulation center. The NICU CRM program offers monthly, 4-h courses, focused on team communication and understanding the human and systems factors that influence patient safety. We also completed debriefs after each simulation and in-unit resuscitation event.
Tool development
Our study used a group consensus method to identify key, observable elements critical for effective resuscitation. The modified Delphi method provides a framework to use data, literature review, and expert opinions to reach consensus agreement.21,22 The insight and opinion from a group of experts provides improved results compared to any single opinion. Given the complexity of infant resuscitation and limited evidence of key elements indicating successful team performance, expert opinion was chosen as the most accurate way to gain collective expertise and reduce bias to develop a comprehensive tool. We performed a literature search about key elements of code team performance and a review of 2 years of post-resuscitation debriefing documentation to generate ideas of key elements for inclusion. The multidisciplinary expert team included two institutional leaders (NICU Medical Director and Nurse Leadership), two neonatal simulation-trained expert facilitators (MD and RN), two neonatal fellows, two nurse practitioners, two neonatal nurses, and one respiratory therapist. For each cycle, respondents were asked if they agreed or disagreed with inclusion of each element on the checklist. An element was included if >80% of experts agreed it should be included.
After definition of essential elements, the modified Delphi process was used again to define an importance score and assign a weighted value to each element. The multidisciplinary expert panel was asked to define the importance of each checklist element, with 1 point for “major” and 0.5 points for “moderate.” Mean score and standard deviation were calculated. The importance score for the element was finalized after >80% of experts agreed.
Quality improvement interventions and study of the interventions
After consensus was obtained and the infant resuscitation team performance tool (Infa-RePT) was finalized (Fig. 1), we developed a multidisciplinary study team to use quality improvement methodology to improve team performance and created a key driver diagram (Fig. 2). In the baseline period (January 2019–December 2019), study team members completed the Infa-RePT score based on their individual observations during neonatal resuscitations in the simulation or in-unit environment. Study team members were observers and not part of the resuscitation efforts. The first ten resuscitation events were scored independently by a minimum of two expert reviewers and these scores were assessed for inter-rater agreement. We also tabulated the frequency of omission of key elements to identify unit trends in the baseline period and created a priority matrix using this data to inform Plan–Do–Study–Act cycles and opportunities for improvements through targeted interventions.
In the intervention period (January 2020–March 2021), the team used quality improvement methodology to perform tests of change and monitor team performance over time using the Infa-RePT tool. Key interventions included (1) creation of role responsibility read-do checklists for leaders and recorders, (2) enhancement of a mock code program, and (3) creation and dissemination of a mock code video highlighting key learning points. Other interventions included formalization of a code review process and staff tool training via in-person staff education, email education, and in real time prior to simulations.
The team reviewed pre-intervention resuscitation debriefs and identified “clarification of role responsibilities” as an opportunity for focused improvement. In response to this, the first intervention included creation of two role-responsibility checklists that included reminders for the “leader” and “recorder” roles. The tools were designed as “read-do” checklists that could be used in real time with key reminders. After creation, the tools were piloted, staff were educated, and then the checklists were placed on all code carts in the NICU. The second intervention was the creation of a monthly multidisciplinary, high-fidelity simulation mock code program, which was 30 min in length, held in proximity to the NICU, and included participants working that day in the NICU. Three expert facilitators led each simulation and debrief. Lessons learned from the sessions were disseminated broadly in a unit-wide educational email to broaden impact. Infa-RePT scores were completed at the end of each mock code with team input. Staff were also educated on the use of Infa-RePT tool to ensure appropriate use during in-unit resuscitation events. After review of all Infa-RePT scores, as a third intervention our multidisciplinary QI team created an infant resuscitation video with educational pause points to highlight key considerations during the code event. All staff were invited to view the video and reminders were circulated prior to mock code participation. This new mock code program enhancement was in addition to the ongoing crisis resource management simulation sessions. An additional project intervention included formalization of the code review process during multidisciplinary leadership meetings to ensure thorough review, systematically track issues, and fast-track system changes.
All staff members are required to maintain PALS certification and some may have participated in PALS training during the study period. No specific additional resuscitation training was performed during the study period, but lessons learned from debriefs were shared across the unit through safety briefs, email educational updates, and practice changes as per routine unit practice.
Measures
Infa-RePT performance scores were collected over the baseline and intervention periods as an outcome measure. Given the important benefit of having a full team present to participate in the mock code practicing in their trained role, we also tracked the multi-disciplinary attendance as a process measure. Ten roles were selected as essential for each mock code and the percentage of roles filled was tracked. Roles included attending, fellow, neonatal nurse practitioner, nurse leader/charge nurse, bedside nurse, two code cart nurses, recorder, nurse assistant, and respiratory therapist. To study tool generalizability, we also compared scores obtained from simulation events to in-unit events.
Analysis
To determine whether special cause variation was found after interventions, we developed statistical process charts (SPC) using the Chartrunner software to make X bar S charts. Excel was used to make run charts for analysis. Statistics are presented as percentage or mean ± standard deviation and t test was used to compare groups.
Results
Modified Delphi process for tool development
The initial search and review process yielded 27 candidate elements that were grouped into key categories for the modified Delphi process. Twelve multidisciplinary experts participated in the modified Delphi process. Two cycles were performed. After the second cycle, >80% agreement among the team determined the final 25 elements to include in the checklist. The group also collectively decided whether each element should be evaluated with options of “yes/no” or “yes/somewhat/no.” The team placed elements under main categories crucial for effective team resuscitation to be CPR, PALS, Role Responsibilities, Communication and Other Elements (Fig. 1).
The next step included ranking the elements based on importance as “moderate” or “major.” After 4 modified Delphi cycles, a >80% consensus was reached on each element. The elements on the tool were assigned a point value, with a lower score indicating a better performance. Elements of moderate and major importance were given a 0.5 and 1 point score, respectively, if not observed (Fig. 1). The highest (worst) possible score is 42.
Tool testing and baseline data collection
To assess inter-rater variability, a subset of the expert team consisting of at least 2 (range 2–4) observers scored 10 simulation resuscitations with an average agreement of 86% concordance (range 76–96%, SD 7.3) for checklist elements. Fifteen elements had ≥90% concordance and only 4 elements had ≤70% concordance (Table 1). The element with the lowest concordant response was “Appropriate switch off of compressors,” which only was agreed upon 50% of the time (Table 1). Overall, the scoring agreement was very high indicating low inter-rater variability and high reliability of the tool.
During the pre-intervention period, 20 resuscitations were observed and scored. Infa-RePT scores were monitored over time with an average total score of 7.4 (range 2–12, SD 2.0). This included 16 crisis resource management simulations and 8 in-unit resuscitations. For 16 simulation scenarios, the average performance score was 7.6 (range 3.5–12, SD 2.8) and for 8 in-unit events, the average performance score was 5.7 (range 2–10.5, SD 3). Seventeen elements were observed >75% of the time. The most frequently observed elements were “RN code cart role clearly identified,” “other team member role clearly identified,” “proper specialists were called in a timely way if needed,” “family notified and updated promptly”, and “debrief occurred,” all of which each occurred >95% of the time. The least frequently observed element was “Reversible causes reviewed (Hs and Ts),” which were observed in 21% of resuscitations. Other elements that were not frequently observed were “appropriate pulse checks” 37% of the time, “code team used clear communication” 46% of the time, and “appropriate switch of the compressors” only 50% of the time (Table 2). Team compliance with key elements was lacking most frequently in the CPR category, followed by the PALS category.
Outcome and process measures
The Infa-RePT performance score significantly improved (decreased) and demonstrated special cause variation on SPC chart analysis with a decrease from 7.4 baseline (or “pre”) to 4.2 during the project (Fig. 3). Center line shift occurred after the second main intervention, the enhancement of the mock code program. Variability also decreased and improved as shown in S analysis on SPC chart. The multidisciplinary team attendance at the mock code sessions remained high throughout the project period with a median of 95% on run chart analysis (Fig. 4). No difference was seen between Infa-RePT score based on setting of resuscitation; in-unit (n = 10; average score = 5) and simulation (n = 26; average score = 6.2) were not significantly different (p = 0.3). (Supplemental Fig. 1). The educational video had 93 individual views (as of 11 March 2021) with 110 mock code participants attending simulations after the video was made available.
Team confidence scores
Staff confidence in performing their role during resuscitation scenarios improved after the project started. We conducted a pre-project survey (n = 72) and a post-mock code survey (n = 103) of multidisciplinary staff in which they reported their confidence in different code scenarios. For resuscitations requiring intubation and those requiring chest compressions and medication delivery respectively, 68 and 63% of pre-survey respondents “strongly agreed” or “agreed” that they felt confident that they could perform their role well, compared to 91 and 96% in the post-survey. Pre-survey results also identified additional opportunities for focused education; only 41 and 22% “strongly agreed” or “agreed” that they felt confident to perform their role well in resuscitations requiring delivery of electrical shocks and emergent pericardiocentesis, respectively. These procedures will be used for future mock code scenarios to improve confidence and provide opportunity for practice.
Multidisciplinary staff reported that expansion of simulation program to include monthly mock codes in addition to the monthly CRM simulations was beneficial. Pre-project survey respondents reported that 68% felt that practice during simulation could help them perform their role better during in-unit patient resuscitations. Post-mock code survey results showed that 96% of respondents agreed that mock code increased their confidence to perform their role, 92% agreed that the mock code increased their medical knowledge, and 96% agreed that the mock code increased their communication skills.
Discussion
Collaborative team performance is critical to provide the best care possible to critically ill infant patients. When multidisciplinary teams work together effectively, patient outcomes can be optimized, medical errors can be minimized, and efficiency can be improved. While some tools have been successfully used in pediatric settings such as simulation-only, pediatric intensive care unit, or emergency rooms, no current measurement tool was available to capture the critical code performance tailored to the in-patient NICU setting, including such items as knowledge and communication. Given no clear evidence of the key factors to measure team performance, expert consensus was used to develop a novel tool to fill this gap in ability to measure performance in a standardized manner.
Use of this tool to follow team performance over time showed areas for growth and improvement. To target these areas, simulation scenarios through an enhanced mock code program allowed for the multi-disciplinary team to gain more practice with rare in-unit events. The implementation of this program led to a significant improvement in overall performance as measured by the Infa-RePT. Checklists have been shown to improve team performance. The tool itself is a checklist that the team can use during the debrief to assess performance of providers at varying skill levels simultaneously. It allows the ability for the team to reflect on opportunities for improvement immediately after an event. Over time, the tool can be used to collect composite data and monitor for trends or key vulnerabilities that can direct education, system changes, or focused interventions using quality improvement methodology. Sustainability of the tool usage is critical to ensure continued improvements. Embedding usage into workflow and regularly scheduled unit feedback are two methods currently employed.
Our study had several limitations. The modified Delphi method is a consensus generating methodology and focuses on expert agreement and not on direct evidence to show validity. This tool was studied at one site with no delivery room and patients up to 6 months corrected age, where we use PALS for resuscitation, not NRP. Therefore, we could not directly test delivery room resuscitations. Because this tool is focused on PALs resuscitations for in-unit code events, it would likely need to be modified for use in delivery room resuscitations, when teams may be smaller and role responsibilities different. To gain additional reliability, validity, and generalizability data, the tool could be tested in delivery room settings, and in NICUs more broadly. During the study, different team members participated in each resuscitation. Therefore, we were unable to assess serial individual performance but instead focused on group data over time. Additionally, there were several potentials for observer bias affecting the resuscitation scores because the team was not blinded, observations were performed in real-time, and observers had external knowledge of the participants. Also, time-based improvements may have contributed to score differences. Analysis also did not account for possible clustering effects caused by individual team members participating in variable numbers of resuscitations. Event complexity was not recorded and may affect team performance. Despite this, in general, simulation events were more complex than in-unit resuscitations and no difference in scores was found between these two environments (Supplemental Fig. 1).
Evaluating resuscitation performance and providing meaningful team feedback is essential to improve skills. To our knowledge, no current tool for use both in simulation and in-unit resuscitations incorporates both the resuscitation algorithm steps and key communication and team training crisis resource management elements that are essential for effective team performance. This tool can be used for immediate feedback to teams after resuscitation events in both the simulation and in-unit environment. It is facile and efficient and can be integrated into a debriefing session. The tool results are valuable not only to the individual team members but also to NICU leadership and educators who can use the specifics of the resuscitation performance data over time to target education and track effectiveness of the educational interventions. We showed how targeted interventions can improve team performance over time.
Conclusions
Through the modified Delphi method, we developed an in-hospital infant team resuscitation performance scoring tool (Infa-RePT) based on expert consensus. Healthcare teams can utilized the Infa-RePT to capture scores across varied settings (simulation or clinical) to target educational efforts to improve team performance and infant resuscitation skills. Interventions including enhancement of a mock code program, role responsibility tool development, and addition of an educational video improved team performance.
Data availability
The data generated during this study are available from the corresponding author on reasonable request.
References
Yamada, N. K., Kamlin, C. O. F. & Halamek, L. P. Optimal human and system performance during neonatal resuscitation. Semin. Fetal Neonatal Med. 23, 306–311 (2018).
Weiner, G. M. & Zaichkin, J. Textbook of Neonatal Resuscitation 7th edn (AAP, 2016).
National Health Care Provider Solutions. Pediatric Advanced Life Support: Provider Handbook (Bookboon, 2016).
Gilfoyle, E. et al. Improved clinical performance and teamwork of pediatric interprofessional resuscitation teams with a simulation-based educational intervention. Pediatr. Crit. Care Med. 18, e62–e69 (2017).
Norris, E. M. & Lockey, A. S. Human factors in resuscitation teaching. Resuscitation 83, 423–427 (2012).
Knight, L. J. et al. Improving code team performance and survival outcomes: Implementation of pediatric resuscitation team training. Crit. Care Med. 42, 243–251 (2014).
Veazie, S. et al. Implementing high-reliability organization principles into practice: a rapid evidence review. J. Patient Saf. 18, e320–e328 (2022).
Carayon, P. et al. Work system design for patient safety: the SEIPS model. Qual. Saf. Health Care 15, i50–i58. (2006).
Holden, R. J. et al. SEIPS 2.0: a human factors framework for studying and improving the work of healthcare professionals and patients. Ergonomiocs 56, 1669–1686 (2013).
Roitsch, C. M., Patricia, K. E., Hagan, J. L., Arnold, J. L. & Sundgren, N. C. Tablet-based decision support tool improves performance of neonatal resuscitation: a randomized trial in simulation. Simul. Healthc. 15, 243–250 (2020).
Fuerch, J. H., Yamada, N. K., Coelho, P. R., Lee, H. C. & Halamek, L. P. Impact of a novel decision support tool on adherence to Neonatal Resuscitation Program algorigthm. Resuscitation 88, 52–56 (2015).
Stentz, M. J., Wiepking, M. D., Hodge, K. A., Ramonell, R. P. & Jabaley, C. S. Checklists improve team performance during simulated extracorporeal membrane oxygenation emergencies: a randomized trial. Crit. Care Explor. 3, e0404 (2021).
De Bie Dekker, A. J. R. et al. Testing the effects of checklists on team behaviour during emergencies on general wards: an observational study using high-fidelity simulation. Resuscitation 157, 3–12 (2020).
Reid, J. et al. The Simulation Team Assessment Tool (STAT): development, reliability and validation. Resuscitation 83, 879–886 (2012).
McKay, A., Walker, S. T., Brett, S. J., Vincent, C. & Sevdalis, N. Team performance in resuscitation teams: comparison and critique of two recently developed scoring tools. Resuscitation 83, 1478–1483 (2012).
Perkins, G. D., Hulme, J. & Tweed, M. J. Variability in the assessment of advanced life support skills. Resuscitation 50, 281–286 (2001).
Guise, J. M. et al. Validation of a tool to measure and promote clinical teamwork. Simul. Healthc. 3, 217–223 (2008).
Institute for Healthcare Improvement. Institute for Healthcare Improvement: Resources: How to Improve (Ihi.Org, 2015).
Garvey, A. A. & Dempsey, E. M. Simulation in neonatal resuscitation. Front. Pediatr. 8, 59 (2020).
Katheria, A., Rich, W. & Finer, N. Development of a strategic process using checklists to facilitate team preparation and improve communication during neonatal resuscitation. Resuscitation 84, 1552–1557 (2013).
Humphrey-Murto, S. et al. The use of the Delphi and other consensus group methods in medical education research: a review. Acad. Med. 92, 1491–1498 (2017).
Diamond, I. R. et al. Defining consensus: a systematic review recommends methodologic criteria for reporting of Delphi studies. J. Clin. Epidemiol. 67, 401–409 (2014).
Ogrinc, G. et al. SQUIRE 2.0 (Standards for Quality Improvement Reporting Excellence): revised publication guidelines from a detailed consensus process. J. Nurs. Care Qual. 31, 1–8 (2016).
Funding
K.T.L. is supported by BCH Quality Scholar Grant; R.M.R. is supported by T32HD098061.
Author information
Authors and Affiliations
Contributions
K.T.L. conceptualized and designed the study, collected data, carried out the analyses, drafted the initial manuscript, and reviewed and revised the manuscript. R.M.R., C.O., and D.C. contributed to the design of the study, data collection, and reviewed and revised the manuscript. A.H. contributed to project study design and data analysis and reviewed and revised the manuscript. All authors approved the final manuscript as submitted and agree to be accountable for all aspects of the work.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Ethics approval and consent to participate
The Institutional Review Board determined that the study was exempt from needing approval. This manuscript was written using the SQUIRE guidelines23 (see Supplemental Table 1). No consent was required.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
About this article
Cite this article
Leeman, K.T., Raju, R.M., O’Brien, C. et al. Development and use of an infant resuscitation performance tool (Infa-RePT) to improve team performance. Pediatr Res 93, 56–62 (2023). https://doi.org/10.1038/s41390-022-02097-6
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1038/s41390-022-02097-6