Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Performance Assessment of Emergency Teams and Communication in Trauma Care (PERFECT checklist)—Explorative analysis, development and validation of the PERFECT checklist: Part of the prospective longitudinal mixed-methods EPPTC trial

  • David Häske,

    Roles Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Validation, Writing – original draft, Writing – review & editing

    Affiliations Faculty of Medicine, Eberhard Karls University Tübingen, Tübingen, Germany, DRK Rettungsdienst Reutlingen, Reutlingen, Germany

  • Stefan K. Beckers,

    Roles Conceptualization, Formal analysis, Investigation, Methodology, Writing – review & editing

    Affiliations Department of Anaesthesiology, Faculty of Medicine, University Hospital RWTH, Aachen, Aachen, Germany, Emergency Medical Service, Fire Department, City of Aachen, Aachen, Germany

  • Marzellus Hofmann,

    Roles Conceptualization, Formal analysis, Investigation, Methodology, Writing – review & editing

    Affiliation Faculty of Health, University of Witten/Herdecke, Witten, Germany

  • Rolf Lefering,

    Roles Methodology, Validation, Writing – review & editing

    Affiliation Institute for Research in Operative Medicine, University of Witten/Herdecke, Cologne, Germany

  • Christine Preiser,

    Roles Methodology, Validation, Writing – review & editing

    Affiliations Coordination Centre for Health Services Research, University Hospital Tübingen, Tübingen, Germany, Institute of Occupational and Social Medicine and Health Services Research, University Hospital Tübingen, Tübingen, Germany

  • Bernhard Gliwitzky,

    Roles Conceptualization, Funding acquisition, Writing – review & editing

    Affiliation Megamed Emergency Management GbR, Maikammer, Germany

  • Paul Alfred Grützner,

    Roles Supervision, Writing – review & editing

    Affiliation Department of Trauma and Orthopaedic Surgery, BG Hospital Ludwigshafen, Ludwigshafen, Germany

  • Ulrich Stöckle,

    Roles Project administration, Supervision, Writing – review & editing

    Affiliation Department of Traumatology and Reconstructive Surgery, BG Hospital Tübingen, Tübingen, Germany

  • Matthias Münzberg

    Roles Conceptualization, Investigation, Project administration, Supervision, Writing – review & editing

    Matthias.Muenzberg@bgu-ludwigshafen.de

    Affiliations Department of Trauma and Orthopaedic Surgery, BG Hospital Ludwigshafen, Ludwigshafen, Germany, Centre of interdisciplinary Rescue- and Emergency Medicine, BG Hospital Ludwigshafen, Ludwigshafen, Germany

Abstract

Background

Trainings in emergency medicine are well structured, but examinations are rarely validated. We are evaluating the impact of pre-hospital emergency trainings on participants and patient care and developed and validated a checklist to assess emergency trainings.

Methods

We used videos recorded at the time points directly before (t0), directly after (t1), and one year after (t2) training to develop the PERFECT checklist (Performance Assessment of Emergency Teams and Communication in Trauma Care). The videos were assessed using semi-qualitative/linguistic analysis as well as expert panel appraisal and recommendations using the Delphi method. The checklist was tested for validity and reliability.

Results

The inter-rater reliability (ICC = 0.99) and internal consistency (α = 0.99) were high. Concurrent validity was moderate to high (r = 0.65 –r = 0.93 (p<0.001)). We included scales for procedures, non-technical skills, technical skills and global performance. The procedures were done faster in the mean over the timeline (t0: 2:29, 95%CI 1:54–3:03 min., t1: 1:11, 95%C 0:53–1:30 min, t2: 1:14, 95%CI 0:56–1:31 min.). All experts rated the recorded scenarios at t0 with the lowest sum score (mean 31±8), with a significantly better performance of the teams at t1 (mean 69±7). The performance at t2 (mean 66 ± 13) was slightly lower than at t1, but still better than at t0. At t1 and t2, linguistic analysis showed a change in the team leaders communication behaviour, which can be interpreted as a surrogate parameter for reduced stress.

Conclusion

The PERFECT checklist has a good validity and high reliability for assessing trauma procedures and teamwork

Introduction

The care of emergency patients with their countless facets is a great challenge. Non-traumatological patients require a wide range of approaches due to their numerous comorbidities and interactions, whereas due to physiological peculiarities pediatric emergency patients need completely different priorities and strategies in their care. The care for seriously injured patients in turn still deserves considerable attention because a large number of these patients are young, of a working age and the injuries generally have major physical, emotional and socio-economic consequences [1,2]. Today, there is a wide variety of offers for training in acute and emergency care, for example for trauma, resuscitation, pre-hospital or in-hospital emergency situations.

Although these trainings are well structured, in accordance with guidelines, and usually conclude with a written (multiple choice) and practical success evaluation [36], the examinations are rarely and not consistently validated [7]. In the study of medicine for the objective assessment of skills and abilities, OSCEs (objective structured clinical examination) are used and these are a valid and reliable tool [810].

There are numerous tests for non-technical skills [11] and technical skills [12,13] as well as for trainings or real patient care [14]. Although early experience in OSCE in emergency medicine has been published [15], there is just little literature compared to other medical subjects [16].

Rationale

In our study, we investigated how an emergency medical service (EMS) system is influenced by systematic training. The reason for a new approach in this EMS was a decreased employee satisfaction and loss of quality in patient care [17]. We knew that the chosen training models have no significant impact on mortality [18], so we chose a prospective longitudinal mixed-methods design, including video analysis of training sequences, to view the impact in its entirety [19]. However, existing assessments or OSCEs were not suitable for verifying changes other than mortality from training, because they do not have the required technical, cognitive and communicative characteristics.

Specific aims

This article describes the explorative, semi-qualitative development of a checklist for assessment and verification of video analyses of emergency medical trainings. The checklist should enable an objective evaluation and comparability of scenarios or even real trauma patient care by means of point scores.

Methods

Context

This analysis was part of the mixed-methods longitudinal EPPTC (Effect of Paramedic Training on Pre-hospital Trauma Care) study evaluating the subjective and objective changes after emergency trainings in participants and real patient care. The complete study is described in the previously published study protocol as well as partial results [1921].

Intervention

To improve the pre-hospital care of patients, the medical director of EMS decreed that all of the approximately 300 paramedics had to be trained in PHTLS.

The two-day PHTLS courses teach paramedics and emergency physicians how to improve pre-hospital care for trauma patients [22]. These courses can be regarded as a worldwide standard in the pre-hospital care of seriously injured patients. They use different teaching methods (e.g. lectures, practical case studies, skill training), with a close instructor-participant ratio (1:4), many practical activities and continuous interaction. The courses are conducted by certified instructors (physicians, paramedics, etc.). The priority-based structure ABCDE (Airway, Breathing, Circulation, Disability and Exposure) is taught intensively and practiced in scenario-based training sessions, as well as various skills. The PHTLS statements are similar to the key recommendations of the "German Guideline on Treatment of Patients with Severe and Multiple Injuries" [6].

Study of the intervention

To assess the impact of the interventions, we chose a mix methods approach. For the longitudinal analysis, we used three measuring points. The first measuring point was just prior to the course (t0), the second measuring point was directly after the course (t1) and the third measuring point was one year after the course (t2). The current publication is concerned with the analysis of videos of the trainings, using qualitative, quantitative and linguistic approaches and with the development of a checklist to assess and to compare the performance of the teams in the training videos in an objective way similar to an OSCE.

Measurements

We used detailed video analysis for the measurements. For this purpose, three videos were selected at random for each measuring point. To create the scenario-checklist, the videos were analysed, and the results were reviewed, adapted and refined by an expert panel.

The videos were recorded during the trainings for paramedics in context of the overall project (EPPTC-study) [19]. Recording times were at each measuring point. A Panasonic HD Camcorder HC-V100 on a tripod was used for recording and data was stored on SD Memory Cards. The scenarios simulated the pre-hospital care of severely injured patients. Amateur actors represented patients with a leading severe thoracic injury and dislocated ankle fracture, however, always with different causes and stories. Injuries should correspond to an Injury Severity Score (ISS) of approx. 38 (abbreviated injury scale (AIS): AIS 0-0-5-2-3-1). For the present analysis, three videos from each of the three measuring points were randomly selected from the records.

Explorative analysis.

To gain an impression of the data, we performed an explorative analysis before validation. First, we analysed the timing of the measurements, differentiated according to the three measuring points. Second, we ran a qualitative analysis to develop items for the checklist. Data coding and analyses were performed with the qualitative software program MAXQDA 12 (Berlin, Germany) and followed the methodological concept of a directed qualitative content analysis [23]. Team performance, medical measures, communication characteristics and behaviour were encoded directly into the program. Communication from the team leader to the patient and to the team was transcribed and coded for linguistic analysis.

For linguistic quantitative analysis of the communication between the team leader and team, we used the program "Linguistic Inquiry and Word Count (LIWC)" (Lawrence Erlbaum Associates Inc, Texas). This method categorises word count, sentence punctuation, negation (no, never, not), proportion of words with more than 6 letters (big words), approvals (yes, OK, mmhmm) and first person plural (us, our, us) as well as psychological classifications like positive emotions (happy, handsome, good), anger (hatred, annoying) and cognitive processes (cause, knowledge, effect, perhaps) and fillers [24,25]. The program counts the words in the transcribed text and calculates the percentage of total words that match the specific categories. This analysis did not include other aspects of phonetic language with para-verbal and non-verbal events.

Expert panel

An expert panel consisting of emergency physicians, medical didactics, sociologists, and human scientists (DH, SB, MH, CP, BG, MM) assessed the results of the video analysis. The Delphi method was used to discuss the codes for the checklist and to discuss the applicability for the practical assessment of scenarios and to reach a final consensus for the determination of the content validity.

Subsequently the inter-rater reliability was tested. Videos of all three time points (t0, t1, t2) were blinded to the time points and then assessed by six experienced instructors (physicians, paramedics with ATLS / PHTLS / ALS instructor level) independently.

Statistical analysis

To assess the construct validity, we performed a principal component factor analysis (PCA) with varimax rotation. Eigenvalues greater than 0.5 were chosen for the component factors, and factor charges of at least 0.5 were sought for the interpretation of the factor structure.

To calculate the internal consistency of the scales, we used Cronbach’s Alpha. For concurrent validity we correlated the global performance scale, non-technical scale, primary assessment scale and procedure scale.

An inter-rater reliability > 0.8 is recommended [26]. We used the intra-class correlation (ICC), because of continuous variables and more than two raters. Because the evaluation is usually done by a single rater, a two-way random model with single measure ICC (3.1) was used [27].

Pearson coefficient was used to describe interval-scaled correlation. A two-tailed p-value < 0.05 was usually considered as statistically significant. For continuous variables, data is shown as mean ± standard or 95%-confidence interval. For categorical variables, percentages are presented. All data were analysed using the statistical software SPSS (Version 24.0, IBM Inc., Armonk, NY, USA).

Ethical considerations

The Ethics Committee of the Medical Faculty of the Eberhard Karls University of Tübingen and the University Hospital approved the study proposal, number 197/ 2013BO2. The study is registered in the German Clinical Trials Register with the ID DRKS00004713.

Data collection and analysis were aligned with the data protection officer at the University of Tübingen and the University Hospital of Tübingen. The video recordings were voluntary for the participants and were made after their written consent.

Results

Explorative analysis

As part of the qualitative analysis, we ultimately generated 84 codes. The extensive codes gave a differentiated picture of the training but were too complex to use during training or possibly during patient care. Examples of differentiations are the differences between, for example, oxygen applied, oxygen administration ordered, and oxygen administration controlled. During the Delphi process of the Expert Panel, the codes were reduced to items suitable for the checklist.

The chronology of primary assessment showed that most of the procedures were performed earlier at t1 and (t2) as t0 and the confidence intervals became mostly narrower. On average, the procedures were carried out at minute 2:29, 95%CI 1:54–3:03 at t0, at minute 1:11, 95%C 0:53–1:30 at t1 and at minute 1:14, 95%CI 0:56–1:31 at t2. Fig 1 shows the corresponding results with ABCDE approach.

thumbnail
Fig 1. The figure shows the mean time of the measures performed in the primary assessment, grouped according to three different measuring points.

The graph t1 fits best the (linear) ABCDE-approach, followed by the graph t2. T0 has the most divergence from the t1.

https://doi.org/10.1371/journal.pone.0202795.g001

The linguistic analysis of the communication of the team leader with the team showed a changed communication behaviour from measuring point to measuring point. Fig 2 shows the changes in the different categories. The obvious change is the increase of big words and articles, while the cognitive and social words, as well as emotions, decrease.

thumbnail
Fig 2. Linguistic analysis of the communication from the team leader to the team, over the three times t0, t1, t2.

The increase of big words and articles is obvious, while the cognitive and social words, as well as emotions, decrease.

https://doi.org/10.1371/journal.pone.0202795.g002

Expert panel

Based on the analysis and the expert experience, the checklist Performance Assessment of Emergency Teams and Communication in Trauma Care” (PERFECT Checklist) was created (S1 File). It includes seven scales with a minimum of 6 points and a maximum of 100 points (Fig 3).

thumbnail
Fig 3. Proportions of the scales in the checklist, based on their maximum points.

https://doi.org/10.1371/journal.pone.0202795.g003

The first scale "primary assessment" includes 25 items with the options "application executed" and "timely", which means that each item allows two, overall up to 50 points. The scale "secondary assessment" includes four items with the options "application executed" and "timely". The value of each option was set at 0.5 points, so four points are the possible maximum. Additionally, the time of interventions can be documented.

The expert panel defined a scale "procedures" with five items, in which the most important characteristics for trauma care were defined, which cannot be mapped in other scales. Each item has the value of two points, which means overall twelve points.

The scale "technical skills" includes five skill items and an additional overall item. Each item has a checkbox for "executed" and "indication correct?" as well as a 4-point-performance scale. But only the "skills overall" rating is included in the calculation of the checklist points, with a maximum of six points.

The scale "trauma communication" includes eight items (simply rated) with specific communication points or signal words which were recognised in the qualitative analysis. The maximum is eight points.

The qualitative analysis showed very heterogeneous non-technical skills of the teams. The expert panel chose four items regarding situation awareness and decision-making, leadership and teamwork, workload management and communication. A 4-point performance scale was added, which makes a maximum of 16 points.

Finally, a 6-point "global performance scale" was added to incorporate the experience and judgment of the raters.

Validity and reliability

For validation, 36 videos were reviewed by experts as described previously. All experts rated the recorded scenarios at t0 with the lowest sum score (mean 31 ± 8), with a significantly better performance of the teams at t1 (mean 69 ± 7). At t2 the performance was still better (mean 66 ± 13) than at t0, but slightly lower than at t1. This inter-rater agreement is visualised in Fig 4.

thumbnail
Fig 4. This figure shows the high inter-rater agreement (ICC = 0.993) of the six raters as well as the different appraisal at different measuring points.

The evaluation of the reviewers shows a significantly better performance of the teams after the course than before, but with a slight deterioration from right after the course to one year later.

https://doi.org/10.1371/journal.pone.0202795.g004

Construct validity

The Kaiser-Meyer-Olkin Measure of sampling adequacy was 0.848, Bartlett's Test of Sphericity was χ2 = 207.456 (p<0.001), therefore data showed a good eligibility for PCA.

Three factors with eigenvalues ≥0.5 were extracted by PCA and explained 89.7% of the overall variance. The first factor explained 66.6% of the total variance and included 5 items with factor loadings between 0.83 and 0.92. The second factor explained 14.1% of the variance and comprised just one item with a factor loading of 0.93. The third factor also comprised one item and explained 9.0% of the variance with a factor loading of 0.98. The first factor included the primary assessment, procedures, trauma communication, non-technical skills and the global performance scale and, in accordance with the rating of the experts, explained the most important aspects regarding the quality of simulated trauma care. The second factor included the secondary assessment, which was rarely completed in contrast to the other scales. The third factor was well explained by the technical skills.

Concurrent validity

The sum score and the sum of the primary assessment showed an excellent correlation (r = 0.916, p<0.001), as well as the sum score with the non-technical skills (r = 0.912, p<0.001) and the sum score with the global rating scale (r = 0.912, p<0.001), shown in Table 1.

Internal consistency of the scales

Scales with several items have a good consistency: Cronbach’s alpha was 0.93 for the 27-item-scale "primary assessment", 0.87 for the 4-item-scale "secondary assessment", 0.95 for the 6-item-scale "procedures", 0.93 for the 8-item-scale "communication", 0.96 for the 4-item-scale "non-technical skills" and overall 0.99.

Inter-rater reliability

Inter-rater reliability showed very good values overall and for specific scales. ICC was for the "primary assessment" 0.93, for "secondary assessment" 0.85, for "procedures" 0.93, for the 8-item-scale "communication" 0.93, for "non-technical skills" 0.96 and overall 0.99.

Discussion

The aim of emergency trainings is to gain assurance manual skills and to act more effective in life-threatening situations by improving the knowledge as well as the structured care of emergency patients. Therefore, cognitive knowledge, technical skills, and procedures with clinical judgment are necessary and elementary for subjective safety in trauma care [28].

A variety of courses are available, but the existing assessments did not match our needs. The PERFECT checklist was developed using qualitative and quantitative analysis and the expertise of experienced academics, clinicians and emergency medicine trainers. The result combines the assessment of (technical) skills, non-technical skills and procedural performance, which are essential for clinical competence in trauma care.

The qualitative analysis gathered data in detail. In the end we separated items in, for example, "indicated", "executed", "executed correctly", "executed without indication". In a second step, we had to decide between analytical details whose clinical relevance may be marginal and practicality of the checklist during scenarios. We chose to focus on clinical aspects and practicality and therefore combined several items in groups with the same clinical relevance, for example we established the item “looked for possible respiratory failure” instead of the detailed items cyanosis, thoracic excursions, breathing work, and diminished tidal volume.

Linguistic analysis was developed in the 1990s to analyse cockpit communication regarding language errors and workload [29]. It resulted in recommendations to keep communication simple by short and clear words, because under increasing stress the brain’s memory capacity decreases to a few seconds [30]. The transferability of the findings to acute medicine is widespread and acknowledged [31,32]. Our linguistic analysis of the communication from the team leader to the team at measuring point t0 to t2 showed an increasing number of words in the categories “big words” and "articles" (Fig 2). The relationships may be multifactorial, but with the knowledge that communication, and respectively speech, becomes shorter and monosyllable under stress [30] these categories can be indicated as surrogate parameters for stress. Stress has a relevant influence on the technical performance, but can be compensated by non-technical skills [33]. In our analysis, we interpreted the increasing number of words in the categories "big words" and "articles" as a surrogate marker, which indicates declined stress levels over the measurement times. This result is also reflected in the assessment of the course participants, as their subjective safety in the care of severely injured patients improved over the measuring points and after the course [28]. Additionally, our "non-technical skills scale" and "primary assessment scale" showed a good correlation with r = 0.745.

In the qualitative analysis of the scenarios, the difference in the non-technical skills of the teams was remarkable. Although the PHTLS system does not provide any NTS teaching content, they have been added to the checklist. For the NTS scale we used a 4-point performance scale, which was labelled and matched with the relevant literature [31,32]. Wallin et al. have also evaluated the impact of training on medical students, with improved clinical skills, but there was no improvement in teamwork [34]. In contrast, in our exploration we saw a significant improvement in teams with mixed experience, but these findings must be confirmed with an adequate sample size.

The checklist developed shows a very good reliability and validity. Our ICC (0.99) is notably higher than the ICC of similar assessment tools (0.6–0.8) [3537]. The appraisal of the expert panel was in our opinion more important than a calculation of a content validity index, especially because the content validity is always subject to a certain subjectivity and it is strictly speaking not a test quality criterion [26]. Concurrent validity shows high correlation. Along with this, we have established the equally well correlated global performance scale, with which Dankbaar et al. had already had positive experiences [38]. Therefore, it is important to have well-trained users of the checklist.

The different coding of the scales was also discussed in the expert panel. For classical OSCE, binary codes are often used but are at least equal to multi-level rating [39]. Frequently, technical skills tend to use binary rating scales rather than communicative ones [40], and their number of points or range can vary widely [35,37,41]. We decided to use binary coding in the assessments and procedures and used 4-point rating scales for NTS and the skill performance. The global rating scale has a 6-point rating scale.

Qualitative video analysis also showed that some teams met all the points on the scale, but they did not perform in a timely manner, in any structure or at speed, or in terms of rapid identification and management of life-threatening conditions. We also knew that a structured patient treatment has a significant influence on subjective safety in trauma care [28], which is why the expert panel had implemented the scale "procedures" and even weighed their points twice. Thus, the factor can be evaluated once with the two items in the procedure scale, on the other hand, the individual measures have the option of "timely" to note. In addition, the time can be recorded independently of the checklist by the real time measurement.

As described, the PERFECT-checklist was developed to evaluate trainings from PHTLS-courses. In principle, however, it is possible to use this validated checklist for each prehospital trauma training according to a comparable standard. The standard in trauma care in Germany is the "Level 3 guideline on the treatment of patients with severe/multiple injuries" [42]. For PHTLS it is proven that the key recommendations fit the Level 3 guideline [6]. It is to be assumed, but not proven, that other established training programs, such as ITLS, trauma management come to similar statements. Thus, the PERFECT-checklist fits all comparable standards in the acute care of trauma patients.

It should also be possible to use the checklist for real patient care in prehospital setting or in the trauma room as well as for simulation. However, this has not been investigated yet. Even the individual scales of the checklist can be used, whose validation index can be taken from the result part.

Limitations

The biggest weakness of the checklist is the impossibility to make a differentiated assessment of the skills. This concerns the individual skills (execution, indication, performance, etc.) as well as the embedding into the points system of the checklist to ensure the comparability of scenarios. For example, a scenario with three skills could theoretically receive a triple skill score, while a scenario with only one skill could be very well performed but would have automatically scored less on the one skill. So we decided to include only one overall skill item in the calculation, but to include differentiated skills as a memo for instructors.

Our checklist was developed on scenarios of emergency trainings and not during real patient care and so no statement can be made about how it works in real patient care.

Conclusion

The importance of a systematic approach instead of a personal approach to team training in high-risk emergency care is crucial [43] and vitally important for improving public health and potentially reducing the mortality of patients. This requires appropriate training, as well as validated opportunities to review long-term training success. With the support of the PERFECT checklist, a validated tool is now available to evaluate trauma trainings with comparable standards.

Supporting information

Acknowledgments

We would like to thank the following for their contribution: participating paramedics and the respective EMS agencies, HSK Hospital Wiesbaden and special thanks to Götz Brodermann, Marc Dieroff, and Nobert Hagner, who were the initiators of the project and organised the study project at that site. We thank Günter Bildstein MSc MBA, Niklas Heinemann, Timo Schädler and Deniz Uzun for the test runs of the checklists.

References

  1. 1. Debus F, Lefering R, Frink M, Kuhne CA, Mand C, Bucking B, et al. Numbers of Severely Injured Patients in Germany. A Retrospective Analysis From the DGU (German Society for Trauma Surgery) Trauma Registry. Dtsch Arztebl Int. 2015; 112: 823–829. pmid:26754119
  2. 2. Statistisches Bundesamt. Wirtschaftsbereiche—Verkehrsunfälle—Beteiligte und Verunglückte; 06.07.2017. https://www.destatis.de/DE/ZahlenFakten/Wirtschaftsbereiche/TransportVerkehr/Verkehrsunfaelle/Tabellen/VerletzteAlter.html. Accessed 18 July 2017.
  3. 3. Münzberg M, Mutschler M, Paffrath T, Trentzsch H, Wafaisade A, Walcher F, et al. Level of Evidence Analysis for the Latest German National Guideline on Treatment of Patients with Severe and Multiple Injuries and ATLS. World J Surg. 2015; 39: 2061–2067. pmid:25894400
  4. 4. Häske D, Gliwitzky B, Semmel T, Schädler T, Casu S, Grusnick H-M, et al. Advanced Medical Life Support (AMLS). Notfall Rettungsmed. 2013; 16: 611–616.
  5. 5. Greif R, Lockey AS, Conaghan P, Lippert A, de Vries W, Monsieurs KG. Ausbildung und Implementierung der Reanimation. Notfall Rettungsmed. 2015; 18: 1016–1034.
  6. 6. Häske D, Stuke L, Bernhard M, Heller AR, Schweigkofler U, Gliwitzky B, et al. Comparison of the Prehospital Trauma Life Support recommendations and the German national guideline on treatment of patients with severe and multiple injuries. J Trauma Acute Care Surg. 2016; 81: 388–393. pmid:27257703
  7. 7. Ringsted C, Lippert F, Hesselfeldt R, Rasmussen MB, Mogensen SS, Frost T, et al. Assessment of Advanced Life Support competence when combining different test methods—reliability and validity. Resuscitation. 2007; 75: 153–160. pmid:17467869
  8. 8. Crossley J, Humphris G, Jolly B. Assessing health professionals. Medical Education. 2002; 36: 800–804. pmid:12354241
  9. 9. Hodges B, Regehr G, Hanson M, McNaughton N. Validation of an objective structured clinical examination in psychiatry. Acad Med. 1998; 73: 910–912. pmid:9736854
  10. 10. Müller S, Dahmen U, Settmacher U. Objective Structured Clinical Examination (OSCE) an Medizinischen Fakultäten in Deutschland—eine Bestandsaufnahme. Gesundheitswesen. 2016.
  11. 11. Passauer-Baierl S, Baschnegger H, Bruns C, Weigl M. Interdisziplinäre Teamarbeit im OP: Identifikation und Erfassung von Teamarbeit im Operationssaal. Zeitschrift für Evidenz, Fortbildung und Qualität im Gesundheitswesen. 2013.
  12. 12. Pape-Koehler C, Immenroth M, Sauerland S, Lefering R, Lindlohr C, Toaspern J, et al. Multimedia-based training on Internet platforms improves surgical performance: a randomized controlled trial. Surg Endosc. 2013; 27: 1737–1747. pmid:23475016
  13. 13. Martin JA, Regehr G, Reznick R, MacRae H, Murnaghan J, Hutchison C, et al. Objective structured assessment of technical skill (OSATS) for surgical residents. Br J Surg. 1997; 84: 273–278. pmid:9052454
  14. 14. Carter EA, Waterhouse LJ, Kovler ML, Fritzeen J, Burd RS. Adherence to ATLS primary and secondary surveys during pediatric trauma resuscitation. Resuscitation. 2013; 84: 66–71. pmid:22781213
  15. 15. Lunenfeld E, Weinreb B, Lavi Y, Amiel GE, Friedman M. Assessment of emergency medicine. A comparison of an experimental objective structured clinical examintion with a practical examination. Medical Education. 1991; 25: 38–44. pmid:1997827
  16. 16. Ruesseler M, Weinlich M, Byhahn C, Muller MP, Junger J, Marzi I, et al. Increased authenticity in practical assessment using emergency case OSCE stations. Adv Health Sci Educ Theory Pract. 2010; 15: 81–95. pmid:19609700
  17. 17. Häske D, Kreinest M, Wölfl CG, Frank C, Brodermann G, Horter J, et al. Bericht aus der Praxis: Strukturierte Fortbildung zur Verbesserung der Versorgungsqualität im Rettungsdienst. Einsatz-Supervision als neuer Ansatz im Bereich der Rettungsdienst-Fortbildung in Wiesbaden und im Rheingau-Taunus-Kreis? Bericht aus der Praxis. Zeitschrift für Evidenz, Fortbildung und Qualität im Gesundheitswesen. 2013; 107: 484–489.
  18. 18. Johansson J, Blomberg H, Svennblad B, Wernroth L, Melhus H, Byberg L, et al. Prehospital Trauma Life Support (PHTLS) training of ambulance caregivers and impact on survival of trauma victims. Resuscitation. 2012; 83: 1259–1264. pmid:22366502
  19. 19. Häske D, Beckers SK, Hofmann M, Wölfl CG, Gliwitzky B, Grützner PA, et al. The effect of paramedic training on pre-hospital trauma care (EPPTC-study): a study protocol for a prospective semi-qualitative observational trial // The effect of paramedic training on pre-hospital trauma care (EPPTC-study). A study protocol for a prospective semi-qualitative observational trial. BMC medical education. 2014; 14: 32. pmid:24528532
  20. 20. Häske D, Beckers SK, Hofmann M, Lefering R, Gliwitzky B, Wölfl CG, et al. Quality of Documentation as a Surrogate Marker for Awareness and Training Effectiveness of PHTLS-Courses. Part of the Prospective Longitudinal Mixed-Methods EPPTC-Trial. Part of the prospective longitudinal mixed-methods EPPTC-trial. PLoS One. 2017; 12: e0170004. pmid:28107394
  21. 21. Häske D, Beckers SK, Hofmann M, Lefering R, Grützner PA, Stöckle U, et al. Subjective safety and self-confidence in prehospital trauma care and learning progress after trauma-courses. Part of the prospective longitudinal mixed-methods EPPTC-trial. Scand J Trauma Resusc Emerg Med. 2017; 25: 1–9.
  22. 22. NAEMT. Präklinisches Traumamanagement. Prehospital trauma life support (PHTLS). 2nd ed. München: Elsevier, Urban & Fischer; 2012.
  23. 23. Hsieh H-F, Shannon SE. Three approaches to qualitative content analysis. Qual Health Res. 2005; 15: 1277–1288. pmid:16204405
  24. 24. Francis ME, Pennebaker JW. LIWC: Linguistic Inquiry and Word Count. Technical Report. Dallas, TX: Southern Methodist University; 1993.
  25. 25. Wolf M, Horn AB, Mehl MR, Haug S, Pennebaker JW, Kordy H. Computergestützte quantitative Textanalyse. Diagnostica. 2008; 54: 85–98.
  26. 26. Bortz J, Döring N. Forschungsmethoden und Evaluation für Human- und Sozialwissenschaftler. 4th ed. Heidelberg: Springer Medizin Verl.; 2009.
  27. 27. Shrout PE, Fleiss JL. Intraclass correlations. Uses in assessing rater reliability. Psychological Bulletin. 1979; 86: 420–428. pmid:18839484
  28. 28. Häske D, Beckers SK, Hofmann M, Lefering R, Grützner PA, Stöckle U, et al. Impact of PHTLS-Courses on Team Performance and Communication in trauma care. Part of the prospective longitudinal mixed-methods EPPTC-trial.
  29. 29. Sexton JB, Helmreich RL. Analyzing Cockpit Communications. The Links Between Language, Performance, Error, and Workload. Journal of Human Performance in Extreme Environments. 2000; 5.
  30. 30. Vetter B, Gasch B, Padosch SA. Medizinisches Handeln in komplexen Notfallsituationen. Kompetent und erfolgreich entscheiden, kommunizieren und führen. Anaesthesist. 2015; 64: 298–303.
  31. 31. Doepfer A-K, Seemann R, Merschin D, Stange R, Egerth M, Münzberg M, et al. Sicherheitskultur in Orthopädie und Unfallchirurgie. Kurskonzept: Interpersonal Competence by Deutsche Gesellschaft für Orthopädie und Unfallchirurgie (DGOU) and Lufthansa Aviation Training. Ophthalmologe. 2017.
  32. 32. Rall M, Gaba DM. Human performance and patient safety. In: Miller RD, editor. Miller’s Anesthesia. Philadelphia: Elsevier Churchhill Livingstone; 2009. pp. 93–150.
  33. 33. Krage R, Zwaan L, Tjon Soei Len L, Kolenbrander MW, van Groeningen D, Loer SA, et al. Relationship between non-technical skills and technical performance during cardiopulmonary resuscitation. Does stress have an influence. Emerg Med J. 2017; 34: 728–733. pmid:28844039
  34. 34. Wallin C-J, Meurling L, Hedman L, Hedegård J, Felländer-Tsai L. Target-focused medical emergency team training using a human patient simulator: effects on behaviour and attitude. Med Educ. 2007; 41: 173–180. pmid:17269951
  35. 35. Cooper S, Cant R, Porter J, Sellick K, Somers G, Kinsman L, et al. Rating medical emergency teamwork performance: Development of the Team Emergency Assessment Measure (TEAM). Resuscitation. 2010; 81: 446–452. pmid:20117874
  36. 36. DeMoor S, Rehim SA, Olmsted R, Myers JG, Parker-Raley J. Evaluating Trauma Team Performance in a Level I Trauma Center: Validation of the Trauma Team Communication Assessment (TTCA-24). J Trauma Acute Care Surg. 2017. pmid:28452896
  37. 37. Steinemann S, Berg B, DiTullio A, Skinner A, Terada K, Anzelon K, et al. Assessing teamwork in the trauma bay: introduction of a modified “NOTECHS” scale for trauma. The American Journal of Surgery. 2012; 203: 69–75. pmid:22172484
  38. 38. Dankbaar MEW, Stegers-Jager KM, Baarveld F, van Merrienboer Jeroen J G, Norman GR, Rutten FL, et al. Assessing the assessment in emergency care training. PLoS One. 2014; 9: e114663. pmid:25521702
  39. 39. Nikendei C, Jünger J. OSCE—praktische Tipps zur Implementierung einer klinisch-praktischen Prüfung. GMS Z Med Ausbild. 2006; 23: 47.
  40. 40. Newble D. Techniques for measuring clinical competence. Objective structured clinical examinations. Med Educ. 2004; 38: 199–203. pmid:14871390
  41. 41. Parker-Raley J, Mottet TP, Lawson KA, Duzinski SV, Cerroni A, Mercado M. Investigating pediatric trauma team communication effectiveness phase one. The development of the assessment of pediatric resuscitation communication. Journal of Communication in Healthcare. 2013; 5: 102–115.
  42. 42. Level 3 guideline on the treatment of patients with severe/multiple injuries. AWMF Register-Nr. 012/019. Eur J Trauma Emerg Surg. 2018; 44: 3–271. pmid:29654333
  43. 43. Patterson MD, Geis GL, Falcone RA, LeMaster T, Wears RL. In situ simulation: detection of safety threats and teamwork training in a high risk emergency department. BMJ Quality & Safety. 2013; 22: 468–477. pmid:23258390