Guidance notes on National Reporting and Learning System official statistics publications

Similar documents
NRLS national patient safety incident reports: commentary

NRLS organisation patient safety incident reports: commentary

Data handling notes for Organisation Patient Safety Incident Reports (NRLS Official Statistics) for April to September 2016

Data Quality Notes. Dimension 1: Relevance. Dimension 2: Accuracy. Alison Pryce (Senior Statistician).

Pressure ulcers: revised definition and measurement. Summary and recommendations

Reference costs 2016/17: highlights, analysis and introduction to the data

Patients registered at a GP Practice

Venous thromboembolism risk assessment data collection Quarter /18 (October to December 2017)

Venous thromboembolism risk assessment data collection Quarter /18 (January to March 2018)

Mental health development PLICS cost collection guidance 2017/18

2020 Objectives July 2016

Improvement and assessment framework for children and young people s health services

Learning from Deaths Framework Policy

Data on Written Complaints in the NHS Q4 Provisional Experimental statistics

NATIONAL INSTITUTE FOR HEALTH AND CARE EXCELLENCE. Health and Social Care Directorate Quality standards Process guide

Learning from adverse events. Learning and improvement summary

GUIDANCE ON SUPPORTING INFORMATION FOR REVALIDATION FOR SURGERY

Supporting information for appraisal and revalidation: guidance for Supporting information for appraisal and revalidation: guidance for ophthalmology

Consolidated pathology network Clinical governance guide

LEARNING FROM DEATHS (Mortality Policy)

A fresh start for registration. Improving how we register providers of all health and adult social care services

Activity planning: NHS planning refresh 2018/19 acute and ambulance provider activity plan template

Seven day hospital services: case study. South Warwickshire NHS Foundation Trust

Annual Complaints Report 2014/15

The non-executive director s guide to NHS data Part one: Hospital activity, data sets and performance

Our next phase of regulation A more targeted, responsive and collaborative approach

A&E Attendances and Emergency Admissions

Venous thromboembolism risk assessment data collection Quarter /18 (July to September 2017)

Addressing ambulance handover delays: actions for local accident and emergency delivery boards

Mandating patient-level costing in the ambulance sector: an impact assessment

An improvement resource for the district nursing service: Appendices

A&E Attendances and Emergency Admissions

SAFE STAFFING GUIDELINE

Improving compliance with oral methotrexate guidelines. Action for the NHS

SUPPORTING DATA QUALITY NJR STRATEGY 2014/16

Welsh Government Response to the Report of the National Assembly for Wales Public Accounts Committee Report on Unscheduled Care: Committee Report

A. Commissioning for Quality and Innovation (CQUIN)

Changes in practice and organisation surrounding blood transfusion in NHS trusts in England

Indicator 5c Mortality Survey

NHS Performance Statistics

Policy on Learning from Deaths

Scottish Hospital Standardised Mortality Ratio (HSMR)

Supporting information for appraisal and revalidation: guidance for Occupational Medicine, June 2014

service users greater clarity on what to expect from services

Blood Transfusion Policy. Version Number: 6.1 Controlled Document Sponsor: Controlled Document Lead: On: December 2014.

Monthly and Quarterly Activity Returns Statistics Consultation

Supporting information for appraisal and revalidation: guidance for psychiatry

NHS Sickness Absence Rates. January 2016 to March 2016 and Annual Summary to

Mental Health Crisis Pathway Analysis

European Reference Networks. Guidance on the recognition of Healthcare Providers and UK Oversight of Applications

Recommendations from National Patient Safety Agency alerts that remain relevant to the Never Events list 2018

Guidance notes to accompany VTE risk assessment data collection

National Reporting and Learning Service (NRLS) Data Quality Standards. Guidance for organisations reporting to the Reporting and Learning System (RLS)

CRITICAL CAPACITY A SHORT RESEARCH SURVEY ON CRITICAL CARE BED CAPACITY. March Intensive Care Medicine. The Faculty of

Reducing Medication Errors

Prevention and control of healthcare-associated infections

Linking quality and outcome measures to payment for mental health

Sponsoring director: Purpose: Decision Assurance For information Disclosable X Non-disclosable

Utilisation Management

2017/18 and 2018/19 National Tariff Payment System Annex E: Guidance on currencies without national prices. NHS England and NHS Improvement

Supporting information for appraisal and revalidation: guidance for pharmaceutical medicine

Safer use of anticoagulants: the NPSA patient safety alert Steve Chaplin MSc, MRPharmS

2017 National NHS staff survey. Results from Royal Cornwall Hospitals NHS Trust

NHS performance statistics

NHS and independent ambulance services

Direct Commissioning Assurance Framework. England

#NeuroDis

NHS performance statistics

The new CQC approach to hospital inspection. Ann Ford Head of Hospital Inspection (North West) June 2014

Guideline scope Intermediate care - including reablement

Directorate Medical Operations Patients and Information Nursing Policy Commissioning Development

Prescribing & Medicines: Reimbursement and remuneration paid to dispensing contractors

The Welsh NHS Confederation s response to the inquiry into cross-border health arrangements between England and Wales.

Improving ethnic data collection for equality and diversity monitoring NHSScotland

Supporting information for appraisal and revalidation: guidance for Occupational Medicine, April 2013

Learning from Patient Deaths: Update on Implementation and Reporting of Data: 5 th January 2018

NHS WALES INFORMATICS SERVICE DATA QUALITY STATUS REPORT ADMITTED PATIENT CARE DATA SET

NHS Sickness Absence Rates

General Practice Extended Access: March 2018

Update on the reporting and monitoring arrangements and post-infection review process for MRSA bloodstream infections

Learning from Deaths Policy. This policy applies Trust wide

Hospital Maternity Activity

ANTI-COAGULATION MONITORING

NHS TAYSIDE MORTALITY REVIEW PROGRAMME

Monitoring the Mental Health Act 2015/16 SUMMARY

OFFICIAL. Integrated Urgent Care Key Performance Indicators and Quality Standards Page 1 of 20

NHS inpatient admission and outpatient referrals and attendances

Public Health Skills and Career Framework Multidisciplinary/multi-agency/multi-professional. April 2008 (updated March 2009)

SCHEDULE 2 THE SERVICES Service Specifications

Percentage of provider spells with an invalid primary diagnosis code

2017 National NHS staff survey. Results from Dorset County Hospital NHS Foundation Trust

Mortality Report Learning from Deaths. Quarter

Same day emergency care: clinical definition, patient selection and metrics

National Standards for the Conduct of Reviews of Patient Safety Incidents

2016 National NHS staff survey. Results from Wirral University Teaching Hospital NHS Foundation Trust

Demand and capacity models High complexity model user guidance

Charlotte Banks Staff Involvement Lead. Stage 1 only (no negative impacts identified) Stage 2 recommended (negative impacts identified)

Status: Information Discussion Assurance Approval. Claire Gorzanski, Head of Clinical Effectiveness

Evaluation of NHS111 pilot sites. Second Interim Report

Trust Policy and Procedure Document Ref. No: PP (17) 283. Central Alerting System (CAS) Policy and Procedure. For use in: For use by: For use for:

Transcription:

Guidance notes on National Reporting and Learning System official statistics publications September 2017

We support providers to give patients safe, high quality, compassionate care, within local health systems that are financially sustainable.

Contents 1. Introduction... 2 2. NRLS data collection... 4 3. NRLS data publications... 4 4. Interpreting the data... 5 5. Contact us for help... 14 1 > Contents

1. Introduction This is a guide to using and interpreting our official statistics outputs: the national patient safety incident reports (NaPSIR; previously called the quarterly data summaries, QDS); and organisational patient safety incident reports (OPSIR). It gives background to the National Reporting and Learning System (NRLS), a summary of our data publications, and the caveats and other considerations when using the data. The document should be read alongside the NaPSIR and OPSIR data tables. Patient safety incident definition We use the definition of a patient safety incident as any unintended or unexpected incident that could have or did lead to harm for one or more patients receiving NHS-funded healthcare. This definition was developed by the National Patient Safety Agency (NPSA) in the Seven steps to patient safety: full reference guide (2004). Examples include: a patient has a severe allergic reaction to a medication a patient s breathing is suppressed after a syringe driver s flow rate is set inappropriately high an incision is made on the wrong limb of a patient scheduled for a joint replacement. The National Reporting and Learning System The NRLS was established by the NPSA in late 2003 as a voluntary scheme for reporting patient safety incidents; its primary function is to enable learning from these incidents and to reduce their recurrence. All NHS organisations in England and Wales have been able to report to the system since 2005. NHS Improvement now manages the NRLS and has statutory functions in relation to patient safety: to operate the NRLS and to use information from the NRLS and elsewhere to develop advice and guidance for the NHS to reduce risks to patients. 2 > Guidance notes on NRLS official statistics publications

In April 2010, it became mandatory for NHS organisations to report all patient safety incidents resulting in severe harm or death to the Care Quality Commission (CQC). They can fulfil this requirement by reporting incidents via the NRLS. To foster openness and encourage continual increases in reporting, reporting to the NRLS remains voluntary with the exception of certain severe incidents (see below). This means NRLS data does not provide the definitive number of patient safety incidents occurring in the NHS. We use information from incident reports to identify trends and develop patient safety resources such as Patient Safety Alerts. All patient safety incident reports submitted to the NRLS coded as resulting in severe harm or death are individually reviewed by NHS Improvement clinicians to make sure we learn as much as we can from them and take action at a national level where appropriate. NRLS incident reports and data are shared with a range of national bodies to support the identification of hazards and the development of patient safety guidance and solutions. These organisations include: NHS England, Public Health England (PHE), the Medicines and Healthcare products Regulatory Agency (MHRA), CQC and the royal colleges. Other data sources Other systems and organisations collect data that relates to patient safety incidents; however, the NRLS is the only national database that includes all types of patient safety incident. Many of these other data collections have been created specifically for a particular type of incident, have a broader remit, or cover a wider geography; as such their comparability with the NRLS cannot be assumed. Other systems include the: strategic executive information system (StEIS); CQC notification database; (MHRA) yellow card scheme and serious adverse blood reactions & events (SABRE); NHS safety thermometer; PHE notifications database; and serious hazards of transfusion (SHOT) scheme. Databases are not the only source of patient safety information. The patient perspective, the voice of the patient, has been shown to be fundamental to understanding patient safety issues; for example, a key failing identified in the Mid Staffordshire NHS Foundation Trust Public Inquiry was that the trust board did not listen sufficiently to its patients and staff. The Department of Health has instructed trusts to use the summary hospital-level mortality indicator (SHMI) in their quality accounts. Quality accounts are reports on the quality of the services provided, measured by looking at patient safety, the effectiveness of the treatments that patients receive and patient feedback about the care provided. 3 > Guidance notes on NRLS official statistics publications

2. NRLS data collection The NRLS collects data on patient safety incidents in England and Wales. Organisations are encouraged to report patient safety incidents to the NRLS regularly and consistently across the year. Frequent and regular reporting can be viewed as an indication of a positive reporting culture. Most patient safety incident reports are submitted electronically to the NRLS from an NHS organisation s own local risk management system. These local systems are often used for a number of purposes, not just recording patient safety incidents. As a result, local definitions and classification systems often differ and must be aligned to the NRLS for accurate reporting (see Section 5 below). A small number of incidents are reported directly using our online eforms by individuals and organisations that do not have local risk management systems. Organisations vary in how their local systems are set up, how many incidents are reported locally and how frequently they upload data to the NRLS. To reduce the burden on local organisations of having to manually enter data both locally and nationally, the NRLS enables the national submission of digital reports stored on local systems. There can be a time lag between an incident occurring and being reported to the NRLS for a variety of reasons (see Section 5 below). Regular reporting is encouraged to minimise this lag. 3. NRLS data publications Purpose NHS Improvement publishes three routine data outputs using the NRLS data: the national patient safety incident reports (NaPSIR), organisation patient safety incident reports (OPSIR) and monthly data summaries. The NaPSIR and OPSIR are classified as official statistics and are considered here in more detail. The NaPSIR and OPSIR are designed to make NRLS data more accessible and they contribute to the pan-government initiative for transparency in data. Specifically, increased transparency together with more thorough reporting and analysis of patient safety-related incidents provides a real opportunity for the NHS at a local and national level to share experiences and learn from these incidents. 4 > Guidance notes on NRLS official statistics publications

Increased transparency is key to: improving the safety of services for patients improving outcomes and productivity in NHS services promoting higher quality and more efficient services, choice and accountability facilitating enhanced commissioning driving economic growth by enabling the development of tools to support users, commissioners and providers of NHS services. As mentioned above, the NRLS and patient safety incidents are not the only source of information relating to patient safety and our published data must be considered alongside other local and national intelligence. Frequency and format NaPSIR and OPSIR data is published every six months in March and September on the NHS Improvement website. Publication dates are announced in advance for the forthcoming year and data are published at 9.30 am on these dates as workbooks in Excel and CSV formats. Official statistics By definition official statistics, such as NaPSIR and OPSIR, are statistical outputs produced by the UK Statistics Authority s executive office (the Office for National Statistics), by central government departments and agencies or the devolved administrations in Northern Ireland, Scotland and Wales, and by other Crown bodies (over 200 in total). The UK Statistics Authority Code of Practice for Official Statistics guides their production, management and dissemination to ensure high standards are met, and their content is well explained in a subjective and impartial manner. We continually look to improve the way we produce and publish our official statistics in line with requirements of the Code of Practice. 4. Interpreting the data Many factors have to be considered when interpreting the NRLS data to reduce the chance of misinterpretation. These factors and the steps we take to reduce their impact are detailed below; some of these factors have already been alluded to above. 5 > Guidance notes on NRLS official statistics publications

Data quality assurance Assurance happens at both the local (validation at point of data collection and submission) and national (validation at the point of data collection and submission) levels. As most patient safety incident data is collected on local systems (see Section 3 above) and then uploaded to the NRLS, we need to ensure local information is correctly mapped to the NRLS fields. Thus all systems used to supply data to the NRLS must be approved as NRLS compliant by NHS Improvement before routine uploading can start. Submitted data uploads are only accepted onto the NRLS if essential fields have been entered, to ensure a core minimum dataset is captured. During production of the NaPSIR and OPSIR a range of quality assurance processes are undertaken, including regular feedback of reporting data for local verification by data suppliers, verification of denominator information and complete quality/cross-checking of the final data tables and associated documentation. This ensures the published outputs reflect the information reported to us. Sources of error and bias Patient safety incident data in general is prone to reporting error and bias and NRLS data is no exception to this. Error and bias will affect the number, type and temporality of reported incidents and how the data is interpreted. Users must also remember that as the number of incidents reported reflects reporting culture rather than the definitive number of patient safety incidents occurring. Known sources of error and bias are summarised below. Source of reports In England 98% of incidents are reported by NHS acute/general hospitals, mental health service, and community nursing, medical and therapy service care settings. The NRLS data is likely to under-represent patient safety incidents happening elsewhere in the healthcare sector, for example in general practice, dental surgeries, community pharmacies and independent providers. Organisational restructuring within the NHS may impact on who reports incidents. For example, in 2013 primary care trusts were dissolved and replaced by clinical commissioning groups (CCGs). CCG reporting requirements had not been established at this point and, historically, the reporting of incidents by the non-acute setting is likely to have been affected by this. Further examples are given in Figure 1. We are working with non-acute/non-nhs providers of NHS-commissioned care to improve their reporting. 6 > Guidance notes on NRLS official statistics publications

Figure 1: Changes with a one-off impact on reporting 7 > Guidance notes on NRLS official statistics publications

Type of incidents reported National and local initiatives on patient safety have often focused on certain staff groups and attitudes to reporting incidents may also differ between staff groups. This can impact on the types of incidents that are reported, such as falls being overrepresented in reported incidents. The level of harm reported for an incident may consciously or unconsciously be modified depending on the local or national profiles of specific incident types. Again the direction and size of any resulting bias is impossible to estimate. Changes in policy Changes in policy may affect the type of incidents reported, with an increase in the reporting of particular incidents if their profile is raised (see Figure 1 above). For example, from April 2010 it became mandatory to report serious incidents to the CQC. StEIS is the main mechanism for reporting such incidents, but as NHS acute and mental health organisations can fulfil this requirement by reporting via NRLS reporting (such incidents are then shared with CQC), this policy change may have affected serious incident reporting to the NRLS. Another example is when the requirements for reporting suicides were widened to encompass actual or apparent suicides of people with an open episode of care in specialist mental health services. Previously only suicides relating to patient safety incidents were required to be reported. Seasonality The NRLS dataset is subject to two types of seasonality: Incident seasonality reflects real differences across the year in when certain types of incident occur. For example, postsurgical morbidity and mortality appear to vary over the year. 1 Reporting (or administrative ) seasonality is an artefact of when reports are received by the NRLS: notable peaks every six months coincide with the cut-off deadlines for data submission for the NaPSIR and OPSIR publications. Users should be aware of incident seasonality when examining data across time, but this should not be controlled for as it reflects real fluctuations in incidents. However, reporting seasonality reflects patterns of reporting rather than when incidents occur. We are working to achieve a more even distribution in reporting 1 Englesbe MJ, Pelletier SJ, Magee JC et al (2007) Seasonal variation in surgical outcomes as measured by the American College of Surgeons-National Surgical Quality Improvement Program (ACS-NSQIP). Ann Surg 246(3): 456 462. 8 > Guidance notes on NRLS official statistics publications

over time by publishing a monthly data table of incidents reported by organisations in each of the previous 12 months. Reporting delay Users are encouraged to report their incidents to the NRLS regularly, at least once a month. However, there are known reporting delays with batches of incidents being sent close to the cut-offs for the NaPSIR and OPSIR publications. Users may also only report or update incidents once a full investigation has been completed, often a considerable time after the incident occurred. The reported and occurring datasets To present NRLS patient safety incident data as accurately as possible, two different datasets are used. This helps to mitigate the issues discussed above around reporting lags and seasonality in the data. The reported dataset is used to look at patterns in reporting, such as frequency and timeliness. It contains incidents reported to the NRLS within a specified time period. It may include incidents that occurred a long time before they were reported. The occurring dataset is used to look at patient safety incident characteristics. It contains incidents reported as actually taking place in a specific time period. This dataset reflects seasonality in when incidents occur; analysis based on this dataset may be biased by fluctuation in numbers over time due to reporting delays. Comparability over time All of the factors discussed above in relation to bias affect how NRLS data is interpreted over time. In general, reporting is increasing over time. This is a positive finding as it reflects a maturing of the reporting culture and maximises the potential for learning. However, it can be difficult to disentangle patterns in specific incidents from the general trend of increased reporting. Therefore, when reviewing changes over time, we recommend users: use proportions or percentages rather than actual numbers (to allow for the differences in the underlying numbers of incidents) either use the same time period for the previous year or a full year s worth of data (to take seasonality into account) 9 > Guidance notes on NRLS official statistics publications

check that any change/difference is not due to new/amended national mandatory reporting requirements or local organisational restructuring. Understanding the degree of harm of patient safety incidents The degree of harm reported to the NRLS should record the actual degree of harm suffered by the patient as a direct result of the patient safety incident. However, this is not always the case. Sometimes reporters give the potential degree of harm of an incident instead. For example, the resulting degree of harm is occasionally coded as severe for near misses where no harm resulted as the impact was prevented. Reporters may also code the degree of harm as severe when the patient is expected to suffer severe but temporary harm (eg severe bruising), which conflicts with the NRLS definition of significant and permanent harm (see below). For some incidents, particularly those affecting patients with multiple co-morbidities or those near the end of their life, it may be difficult to determine at the time of reporting if the level of harm recorded was a direct result of the patient safety incident. There are five NRLS codes for the degree of harm: no harm a situation where no harm occurred: either a prevented patient safety incident or a no harm incident low harm any unexpected or unintended incident that required extra observation or minor treatment and caused minimal harm to one or more persons moderate harm any unexpected or unintended incident that resulted in further treatment, possible surgical intervention, cancelling of treatment, or transfer to another area, and which caused short-term harm to one or more persons severe harm any unexpected or unintended incident that caused permanent or long-term harm to one or more persons death any unexpected or unintended event that caused the death of one or more persons. 10 > Guidance notes on NRLS official statistics publications

Differences between the NaPSIR and OPSIR The NaPSIR and OPSIR fulfil different purposes and differences between them must be appreciated if they are compared (Table 1). By definition the NaPSIR provides a national perspective. Summary data is provided by the date the incident was reported to the NRLS, with more detailed analysis by date of occurrence and care setting. In contrast the OPSIR provides organisation-level breakdowns that show the number and rate of incidents, based on the date incidents occurred. Data quality indicators based on patterns of reporting are included. Due to the impact of organisation type on the anticipated level of patient safety incidents, the OPSIR groups organisations by type. Table 1: Main features of the NaPSIR and OPSIR workbooks Feature NaPSIR OPSIR Purpose To provide a national picture of the reporting of patient safety incidents and of the characteristics of incidents (type, care setting, degree of harm) This dataset forms the basis of the indicator Improving the culture of safety reporting within Domain 5 of the NHS operating framework (Treating and caring for people in a safe environment and protecting them from avoidable harm) To provide data on individual organisation s reporting and patient safety characteristics Different NHS organisations provide different services and serve different populations. Therefore, to make comparisons as meaningful as possible, the NRLS groups NHS organisations into clusters of similar organisations* Dataset type Dynamic Fixed/static Dataset used Reported and occurring datasets Reported and occurring datasets Time period covered Reported dataset: Rolling quarters from October to December 2003 to the most recent quarter available Occurring dataset: Rolling quarters covering the last four available quarters The most recent six months only Updated Every six months Every six months Geography/ breakdown All geographical locations, by care setting England, by individual NHS organisation (organised by cluster) Inclusions The following care settings: The following organisation types: 11 > Guidance notes on NRLS official statistics publications

acute/general mental health service community nursing, medical and therapy service learning disabilities service ambulance service general practice community pharmacy community and general dental service community optometry/ optician service acute/general hospital mental health service community trusts ambulance service *Information on clusters is available in or accompanies the relevant publication. Figures for previous quarters may change slightly (figures for four consecutive quarters are given in each workbook for incidents occurring, from tab 5 onwards in the workbook) as the NRLS is a dynamic system (and incidents can be reported, or updated, at any time after the event). The reported dataset refers to incidents reported by, or within, a certain time period. The occurring dataset refers to incidents occurring by, or within, a certain time period. See above for more information. Suppression of statistics based on small numbers The number of incidents an organisation reports varies both over time and compared to other organisations. The reasons for this are varied and may, for example, reflect real differences in the number of incidents or problems with data upload functionality. Statistics such as percentages and rates are unreliable measures when the underlying numbers are small as differentiating random fluctuation (the role of chance) from true changes in the statistic can be impossible. Where an organisation reports 10 or fewer incidents as occurring during the specific time period (OPSIR), reporting rates and medians are not calculated or published and comparisons are not made. For both the NaPSIR and OPSIR publications, percentages are not calculated where fewer than 30 incidents in total (over the most recently available preceding four quarters [NaPSIR] and six months [OPSIR]) are reported as occurring. 12 > Guidance notes on NRLS official statistics publications

Understanding high and low reporting rates Incidents are reported to the NRLS voluntarily and for the purposes of learning. The number reported by each organisation therefore reflects reporting culture, and is not necessarily the actual number of incidents occurring. A low reporting rate for an organisation should not necessarily be interpreted as a safe environment; it may represent under-reporting. Conversely a high reporting rate should not be interpreted as unsafe ; it may represent a more open culture. Some NHS organisations report to the NRLS daily, others quarterly. In many cases, incidents are grouped and submitted to the NRLS in large batches. It should never be assumed that the total numbers of patient safety incidents are representative of totals across the NHS. The reporting culture varies between organisation types: reporting in secondary care is far more common than in primary care, while ambulance and mental health organisations have the most varied reporting patterns. NHS Improvement s guidance is that organisations should be reporting incidents to the NRLS on a regular basis and at least once a month. Calculation and interpretation of NRLS reporting rates A rate is the frequency of occurrence of a phenomenon in the population under study that is, how often an event happens in a given period of time in the population at risk of the event. For example, if the event is a patient falling in ward A in a given hospital, the population at risk could be all patients admitted to ward A in that hospital. Rates are important in comparing different patient populations and have always been given for the NRLS reporting data. We calculate a reporting rate as: Reported number of patient safety incidents (numerator) Potential opportunities for these incidents to occur (denominator) Using the example above, this would be calculated as: Number of falls in ward A Number of patients admitted to ward A For this rate to be valid, reliable and therefore meaningful, both the number of patient safety incidents (numerator) and the potential opportunities for those incidents to occur (denominator) need to be as accurate as possible. The falls example above is simplistic for the purposes of illustration; as discussed above known factors affect the number of incidents reported and this will affect the accuracy of the numerator. Additionally there are issues with identifying the denominator that best captures the population at risk; for example, the type of 13 > Guidance notes on NRLS official statistics publications

service(s) provided by an organisation and the patient safety incident being examined both affect the population at risk, thus the choice of denominator. NRLS statistics use quarterly KH03 data on average daily overnight bed occupancy (published by NHS England) to estimate total number of bed days for the specified time period of the analysis. Due to the nature of KH03 data this denominator is only appropriate for acute and mental health trusts and a rate currently cannot be calculated for NHS ambulance and community trusts. 5. Contact us for help If you have any questions about the NRLS data collection, the published data or your organisation s data please contact the NRLS team at: nhsi.nrls.datarequests@nhs.net 14 > Guidance notes on NRLS official statistics publications

Contact us: NHS Improvement Wellington House 133-155 Waterloo Road London SE1 8UG 0300 123 2257 enquiries@improvement.nhs.uk improvement.nhs.uk Follow us on Twitter @NHSImprovement This publication can be made available in a number of other formats on request. NHS Improvement 2017 Publication code: CG 33/17