Article Text

Download PDFPDF
Inter-rater reliability in the Paediatric Observation Priority Score (POPS)

Abstract

Objective The primary objective of this study was to determine the level of inter-rater reliability between nursing staff for the Paediatric Observation Priority Score (POPS).

Design Retrospective observational study.

Setting Single-centre paediatric emergency department.

Participants 12 participants from a convenience sample of 21 nursing staff.

Interventions Participants were shown video footage of three pre-recorded paediatric assessments and asked to record their own POPS for each child. The participants were blinded to the original, in-person POPS. Further data were gathered in the form of a questionnaire to determine the level of training and experience the candidate had using the POPS score prior to undertaking this study.

Main outcome measures Inter-rater reliability among participants scoring of the POPS.

Results Overall kappa value for case 1 was 0.74 (95% CI 0.605 to 0.865), case 2 was 1 (perfect agreement) and case 3 was 0.66 (95% CI 0.58 to 0.744).

Conclusion This study suggests there is good inter-rater reliability between different nurses’ use of POPS in assessing sick children in the emergency department.

  • nursing
  • inter-rater reliability
  • early warning score
  • emergency severity index

Statistics from Altmetric.com

Request Permissions

If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.