Methods: A pilot study was designed to precede a national experimental study. Basic orientation and advanced evaluator training modules were developed. These modules included orientation documents, StrengthsFinder Inventory instructions, a training video for the Creighton Competency Evaluation Instrument (CCEI), a training webinar, and a coaching webinar. Study instruments included the CCEI, student performance videos created for the National League for Nursing (NLN) Project to Explore the Use of Simulation for High Stakes, a demographic survey, and the StrengthsFinder Inventory Survey. With NLN approval, the student performance videos and the performance assessment tool produced for and used in the NLN feasibility study were used in the pilot study. A training intervention for faculty evaluators was developed. Five simulation experts completed the training intervention and the performance evaluation procedure. Reliability and correlational analysis was performed to evaluate the impact of training and faculty personality characteristics on inter/intra rater reliability. Feedback was collected from the participants to guide modifications to the content and process of the intervention in preparation for a regional, multi-site, experimental study, which began in the fall of 2016.
Results: All pilot participants were female. Three participants were ages 51 to 60; one was age 61-70; and one was age 31-40. Four participants held a master's degree and one held a doctoral degree as the highest academic credential. The participants taught in associate, baccalaureate, and entry-level master’s programs. The five participants taught in three different states. Only one participant taught in a program currently conducting high-stakes assessment in simulation. Quantitative analysis was conducted on the CCEI video evaluations. When the analysis of the six videos was compared with the three training videos, a large increase in inter-rater reliability was noted for two subscales: Assessment and Clinical Judgment. Two subscales: Communication and Patient Safety showed little difference. The two overall measures, Yes/No Competency, and Overall score showed little difference. The results were reported in an aggregated format, which obscured the differences between the separate evaluations of the training videos and the experimental videos. Even though the statistics were reported in aggregate, it still appears that the training intervention helped the participants to develop a more shared mental model of evaluation. These statistics will be analyzed and reported both individually and in aggregate for the full study.
Conclusion: It was evident that conducting a pilot study was invaluable. When data collection instruments, study procedures, and data analysis are complex, one can expect difficulties that require problem solving. This study has raised some critical questions relative to high stakes assessment, including: 1) what is the “right” amount and format of training? and 2) how do you help teams of faculty develop a shared mental model? This pilot study provided the opportunity to implement study procedures and make changes where issues and problems were discovered.
See more of: Research Sessions: Oral Paper & Posters