Improvement in Scoring Consistency for the Creighton Simulation Evaluation Instrument ©

Mary E. Parsons, Kimberly S. Hawkins, Maribeth Hercinger, Martha J. Todd, Julie A. Manz, Xiang Fang

Research output: Contribution to journalArticle

9 Citations (Scopus)

Abstract

Background: The Creighton Simulation Evaluation Instrument © is a valid and reliable instrument for evaluation of students in the simulated clinical environment. Increased use of simulation and requests for widespread use of the instrument prompted the development of an educational program to maintain reliability. Method: The educational program included review of the instrument and dialogue among faculty to determine expectations of student performance. Percentage agreement and kappa coefficient scores were calculated pre- and posttraining. Results: Improved percentage agreement scores were noted. Seventy-five percent of the kappa scores after implementation of the educational program demonstrate moderate, substantial, or almost perfect agreement. Conclusions: Education and faculty dialogue improve consistency of student assessment and add richness to the learning experience.

Original languageEnglish
JournalClinical Simulation in Nursing
Volume8
Issue number6
DOIs
StatePublished - Jul 2012

Fingerprint

Scoring
educational program
Students
simulation
Evaluation
evaluation
Percentage
Simulation
Program Development
dialogue
student
Learning
Percent
Education
Valid
Coefficient
Demonstrate
learning
performance
education

All Science Journal Classification (ASJC) codes

  • Nursing (miscellaneous)
  • Education
  • Modeling and Simulation

Cite this

Improvement in Scoring Consistency for the Creighton Simulation Evaluation Instrument © . / Parsons, Mary E.; Hawkins, Kimberly S.; Hercinger, Maribeth; Todd, Martha J.; Manz, Julie A.; Fang, Xiang.

In: Clinical Simulation in Nursing, Vol. 8, No. 6, 07.2012.

Research output: Contribution to journalArticle

@article{9fb779f0e1c942faa185d74de1fe899f,
title = "Improvement in Scoring Consistency for the Creighton Simulation Evaluation Instrument {\circledC}",
abstract = "Background: The Creighton Simulation Evaluation Instrument {\circledC} is a valid and reliable instrument for evaluation of students in the simulated clinical environment. Increased use of simulation and requests for widespread use of the instrument prompted the development of an educational program to maintain reliability. Method: The educational program included review of the instrument and dialogue among faculty to determine expectations of student performance. Percentage agreement and kappa coefficient scores were calculated pre- and posttraining. Results: Improved percentage agreement scores were noted. Seventy-five percent of the kappa scores after implementation of the educational program demonstrate moderate, substantial, or almost perfect agreement. Conclusions: Education and faculty dialogue improve consistency of student assessment and add richness to the learning experience.",
author = "Parsons, {Mary E.} and Hawkins, {Kimberly S.} and Maribeth Hercinger and Todd, {Martha J.} and Manz, {Julie A.} and Xiang Fang",
year = "2012",
month = "7",
doi = "10.1016/j.ecns.2012.02.008",
language = "English",
volume = "8",
journal = "Clinical Simulation in Nursing",
issn = "1876-1399",
publisher = "Elsevier Inc.",
number = "6",

}

TY - JOUR

T1 - Improvement in Scoring Consistency for the Creighton Simulation Evaluation Instrument ©

AU - Parsons, Mary E.

AU - Hawkins, Kimberly S.

AU - Hercinger, Maribeth

AU - Todd, Martha J.

AU - Manz, Julie A.

AU - Fang, Xiang

PY - 2012/7

Y1 - 2012/7

N2 - Background: The Creighton Simulation Evaluation Instrument © is a valid and reliable instrument for evaluation of students in the simulated clinical environment. Increased use of simulation and requests for widespread use of the instrument prompted the development of an educational program to maintain reliability. Method: The educational program included review of the instrument and dialogue among faculty to determine expectations of student performance. Percentage agreement and kappa coefficient scores were calculated pre- and posttraining. Results: Improved percentage agreement scores were noted. Seventy-five percent of the kappa scores after implementation of the educational program demonstrate moderate, substantial, or almost perfect agreement. Conclusions: Education and faculty dialogue improve consistency of student assessment and add richness to the learning experience.

AB - Background: The Creighton Simulation Evaluation Instrument © is a valid and reliable instrument for evaluation of students in the simulated clinical environment. Increased use of simulation and requests for widespread use of the instrument prompted the development of an educational program to maintain reliability. Method: The educational program included review of the instrument and dialogue among faculty to determine expectations of student performance. Percentage agreement and kappa coefficient scores were calculated pre- and posttraining. Results: Improved percentage agreement scores were noted. Seventy-five percent of the kappa scores after implementation of the educational program demonstrate moderate, substantial, or almost perfect agreement. Conclusions: Education and faculty dialogue improve consistency of student assessment and add richness to the learning experience.

UR - http://www.scopus.com/inward/record.url?scp=84863732030&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=84863732030&partnerID=8YFLogxK

U2 - 10.1016/j.ecns.2012.02.008

DO - 10.1016/j.ecns.2012.02.008

M3 - Article

VL - 8

JO - Clinical Simulation in Nursing

JF - Clinical Simulation in Nursing

SN - 1876-1399

IS - 6

ER -