Abstract
Background: The Creighton Simulation Evaluation Instrument © is a valid and reliable instrument for evaluation of students in the simulated clinical environment. Increased use of simulation and requests for widespread use of the instrument prompted the development of an educational program to maintain reliability. Method: The educational program included review of the instrument and dialogue among faculty to determine expectations of student performance. Percentage agreement and kappa coefficient scores were calculated pre- and posttraining. Results: Improved percentage agreement scores were noted. Seventy-five percent of the kappa scores after implementation of the educational program demonstrate moderate, substantial, or almost perfect agreement. Conclusions: Education and faculty dialogue improve consistency of student assessment and add richness to the learning experience.
Original language | English |
---|---|
Journal | Clinical Simulation in Nursing |
Volume | 8 |
Issue number | 6 |
DOIs | |
State | Published - Jul 2012 |
Fingerprint
All Science Journal Classification (ASJC) codes
- Nursing (miscellaneous)
- Education
- Modeling and Simulation
Cite this
Improvement in Scoring Consistency for the Creighton Simulation Evaluation Instrument © . / Parsons, Mary E.; Hawkins, Kimberly S.; Hercinger, Maribeth; Todd, Martha J.; Manz, Julie A.; Fang, Xiang.
In: Clinical Simulation in Nursing, Vol. 8, No. 6, 07.2012.Research output: Contribution to journal › Article
}
TY - JOUR
T1 - Improvement in Scoring Consistency for the Creighton Simulation Evaluation Instrument ©
AU - Parsons, Mary E.
AU - Hawkins, Kimberly S.
AU - Hercinger, Maribeth
AU - Todd, Martha J.
AU - Manz, Julie A.
AU - Fang, Xiang
PY - 2012/7
Y1 - 2012/7
N2 - Background: The Creighton Simulation Evaluation Instrument © is a valid and reliable instrument for evaluation of students in the simulated clinical environment. Increased use of simulation and requests for widespread use of the instrument prompted the development of an educational program to maintain reliability. Method: The educational program included review of the instrument and dialogue among faculty to determine expectations of student performance. Percentage agreement and kappa coefficient scores were calculated pre- and posttraining. Results: Improved percentage agreement scores were noted. Seventy-five percent of the kappa scores after implementation of the educational program demonstrate moderate, substantial, or almost perfect agreement. Conclusions: Education and faculty dialogue improve consistency of student assessment and add richness to the learning experience.
AB - Background: The Creighton Simulation Evaluation Instrument © is a valid and reliable instrument for evaluation of students in the simulated clinical environment. Increased use of simulation and requests for widespread use of the instrument prompted the development of an educational program to maintain reliability. Method: The educational program included review of the instrument and dialogue among faculty to determine expectations of student performance. Percentage agreement and kappa coefficient scores were calculated pre- and posttraining. Results: Improved percentage agreement scores were noted. Seventy-five percent of the kappa scores after implementation of the educational program demonstrate moderate, substantial, or almost perfect agreement. Conclusions: Education and faculty dialogue improve consistency of student assessment and add richness to the learning experience.
UR - http://www.scopus.com/inward/record.url?scp=84863732030&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84863732030&partnerID=8YFLogxK
U2 - 10.1016/j.ecns.2012.02.008
DO - 10.1016/j.ecns.2012.02.008
M3 - Article
AN - SCOPUS:84863732030
VL - 8
JO - Clinical Simulation in Nursing
JF - Clinical Simulation in Nursing
SN - 1876-1399
IS - 6
ER -