Article Text

PDF

The Quality of Practice Committee of the RCPCH
  1. N McIntosh1,
  2. J H Baumer2
  1. 1University of Edinburgh, UK
  2. 2Derriford Hospital, Plymouth, UK
  1. Correspondence to:
    Prof. N McIntosh
    Department of Child Life and Health, University of Edinburgh, 20 Sylvan Place, Edinburgh EH9 1UW, UK; neil.mcintoshed.ac.uk

Statistics from Altmetric.com

Update on the clinical effectiveness programme

Failure by doctors to incorporate strong research evidence into their clinical practice delays improvement in mortality and morbidity. The logo of the Cochrane collaboration1 (fig 1) shows the clear evidence of benefit from the randomised controlled trials of antenatal steroids in preterm labour available in 1982 if studies had been subject to meta-analysis.2 Antenatal steroids took over a decade after this point to be widely incorporated into obstetric practice despite the clear evidence that neonatal mortality and subsequent neurodevelopmental morbidity were reduced.3 Similarly suboptimal management was shown in a proportion of children with Kawasaki disease in the UK in 1990.4,5 Only 60% received intravenous gammaglobulin, some in an inadequate dose, despite clear evidence of benefit from randomised controlled trials. A further example might be professional advice on the sleeping position for babies and the risk of cot death.6

Figure 1

 Logo of the Cochrane collaboration.

There are a number of reasons for this. First, unless doctors are practising in a very narrow field there are just too many peer reviewed journals to read. In addition it is only by careful critical appraisal that the research evidence can be set in context of what is already known. This is a time consuming task, and requires a skill that not all doctors yet possess. There is also a genuine lack of knowledge in many medical areas, and one of the important functions of a guideline is to highlight research gaps. Advances in medical knowledge from additional research are also important in improving clinical management.

Clinical practice guidelines are “systematically developed statements to assist both practitioner and patient in decisions about appropriate health care for specific clinical circumstances”.7 It is only through the systematic search for and appraisal of research evidence that truly “evidence based” recommendations can be produced. Clinical guidelines provide assistance for a number of decisions that together comprise a clinical pathway. Many evidence based clinical guidelines are supported by systematic reviews of research evidence. A systematic review is “any summary that attempts to address a focused clinical question using methods designed to reduce the likelihood of bias”,8 Health technology assessments critically appraise all evidence on a specific intervention and give guidance for its use. Both clinical guidelines and systematic reviews can help to identify priorities for primary clinical research.

In 1996 the Royal College of Paediatrics and Child Health (RCPCH) mandated the Quality of Practice Committee (QPC) to:

“act to improve clinical practice by various means—including the production of guidelines. These guidelines will be evidence based, rigorously scrutinised by peer groups and should be eminently subject to audit.”

The QPC oversees the College’s clinical effectiveness programme, including clinical guidelines, the dissemination of systematic reviews, and clinical audit. Any statement about clinical practice made by or on behalf of the College is appraised by the QPC, which then makes recommendations to its Council. Effectively the QPC is the “guardian” of the College with regard to paediatric practice. The College’s Health Services Committee undertakes a similar role for statements pertaining to the organisation or configuration of services. However, any evidence based guidance on clinical practice may have implications for service delivery.

An evidence based clinical practice guideline9,10 will contain a mixture of recommendations where there is supporting evidence and others where no research evidence exists. The College formally recognises three categories of statement about clinical practice (see table 1), and has different mechanisms for managing them.

Table 1

 College categories of clinical practice statements

The resources required to create a well produced evidence based clinical guideline are prodigious. The QPC has developed a procedure for appraising clinical guidelines11 to ensure that only those produced according to best practice are given college endorsement. This involves checking the guideline methodology using the AGREE instrument12 and then reviewing all the recommendations supported by strong evidence alongside the original publications, to confirm that they accurately reflect the evidence.

The scope and clinical questions to be included are crucial if a guideline is to be useful to professionals and patients. It must be underpinned by a fully documented and rigorous literature review. Each relevant research study should be critically appraised and the level of evidence (1–5) documented using a system that defines a hierarchy of evidence relevant to the type of question. The grading system used for recommendations (A–D) likewise creates a hierarchy dependent on both the strength and applicability of the supporting evidence. It is therefore straightforward for readers to judge the strength of supporting evidence for each recommendation. Other important features of a clinical guideline include the involvement of all relevant professionals, and of children and their carers, the use of a formal consensus process (for example, Delphi)13 where evidence is lacking, piloting of the guideline, and the inclusion of audit criteria, written patient information, and educational material. A well produced evidence based guideline will also identify priorities for further research.

At present the QPC, for resource reasons, only appraises A and B recommendations. This is being addressed as sometimes C and D recommendations have even more clinical relevance and importance despite a weaker evidence base. The College Council gives final approval to endorse a clinical guideline. The college standards represent best practice but the approval process gives a measure of discretion so that while only well produced clinical guidelines are endorsed, the standards have not been set so high that they are unattainable.

The production of clinical guidelines in itself does not change practice.14 In order to do so, the barriers to change need to be understood. A clear message should be promoted through multiple mechanisms (a “multi-faceted” approach), including championing by leaders, educational packages, and audit. Audit alone appears to be a weak driver for change, whereas championing by leaders in the field appears more effective (using “eminence based medicine” to promote “evidence based medicine”). Those developing a clinical guideline should therefore consider at the outset what are the potential barriers to implementation, and agree a supporting dissemination and implementation strategy that includes relevant supporting educational material for the profession and parents and children.

In the last five years the QPC has identified, appraised, and disseminated the appraisals of over 20 evidence based clinical guidelines on a wide variety of topics (table 2). Further guidelines are currently being appraised. The clinical guideline on chronic fatigue syndrome (CFS/ME)15 is the first one to be initiated and carried through entirely by the RCPCH. The Quality of Practice Committee remained independent of the guideline development group, and was therefore able to undertake the quality assurance without conflict of interest.

Table 2

 Clinical guidelines appraised and in receipt of college approval

Not all the identified clinical guidelines have been endorsed in the appraisal process, however: one produced through the very reputable auspices of the Scottish Intercollegiate Guidelines Network (SIGN) contained significant discrepancies between the evidence cited and the recommendations,16 despite apparently appropriate guideline development methodology and despite consultation including the RCPCH. A revised version has been published by SIGN,17 but not appraised as the literature search was by then over five years old. Another guideline commissioned by the National Institute for Clinical Excellence (NICE) on eating disorders18 gave insufficient guidance on a number of important clinical issues for paediatric practice, including differential diagnosis and the management of complications. A third guideline on the management of urinary tract infections (from the American Academy of Pediatrics19) showed the importance of ensuring that the scope includes the appropriate outcomes. Although the RCPCH did endorse this guideline, the QPC appraisal commented that it did not address the link between childhood urinary tract infections, renal abnormalities, and long term outcome with or without treatment.20 Without this link, the recommendations for initial management of children following a urinary tract infection could not be considered valid. These examples illustrate some of the potential pitfalls of clinical guideline development, and justify the college’s careful approach.

The Quality of Practice Committee, through strong links with SIGN, has for several years submitted suggestions for future guideline topics. NICE is now developing more evidence based clinical guidelines for a number of child health topics, and the QPC submits suggestions to its Advisory Committee for Topic Selection. The final decision about topic selection rests with the Secretary of State for Health and the Welsh Assembly. The College has also submitted priority topics for research identified through clinical guideline development to the National Coordinating Centre for Health Technology Appraisals.

The number of new paediatric clinical guidelines from NICE and SIGN will increase and other reputable sources (for example, the British Thoracic Society) are coming on-stream. Together with the existing evidence based guidelines, a formidable array of college endorsed guidelines should soon be a reality.

The QPC approach has meant that the RCPCH has been in a position to endorse more clinical guidelines, from a wide variety of sources, than would be the case if only college initiated guidelines were endorsed. So, in what direction should the clinical guidelines programme be developing further? It has been suggested that the college might start producing “best practice guidelines or statements”, without following a rigorous evidence based approach, so as to cover an even wider range of topics. This would run counter to the College’s current policy, and it could not claim to be following best practice in relation to evidence based medicine. It would also undermine the College’s position relating to poorly produced clinical guidelines submitted to it, and weaken the authority of well produced evidence based guidelines.

The other approach is to strengthen the college’s mechanisms for promoting the messages from the clinical guidelines already being developed. NICE has until recently considered the implementation of its guidelines as outside its remit. The promotional material and mechanisms for disseminating its messages to clinicians are therefore usually weak, and the college could assist with this.

The guideline appraisals circulated to College members with the College newsletters now take the form of a summary of the clinical guidelines including the scope and all the recommendations. Only the A and B recommendations were previously included, but this has now changed. In future the College “standards for clinical guidelines” will include a stronger section on local implementation; this should include how to develop a local protocol/integrated care pathway and how to undertake an audit of compliance.

In addition, the College should take the lead in planning the launch of new guidelines, perhaps in collaboration with guideline developers including NICE. This could include the preparation of educational material, which would be likely to help their local implementation. The college needs to state more clearly to paediatricians what is expected of them when the college endorses a clinical guideline: unless they are adopted into practice the efforts of guideline developers are wasted.

In summary, the College’s clinical effectiveness programme, overseen by the Quality of Practice Committee, has already endorsed a number of evidence based clinical guidelines, including its first in-house guideline on chronic fatigue syndrome (CFS/ME).15 It now needs to focus more directly on the local implementation and audit of well produced guidelines, and so help paediatricians achieve demonstrable improvements in the quality of delivered care.

Update on the clinical effectiveness programme

REFERENCES

View Abstract

Footnotes

  • Competing interests: none declared

Request permissions

If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.