Hospital Adverse Event Reporting Program: an Initial Evaluation

Similar documents
UNIVERSITY OF MISSISSIPPI MEDICAL CENTER PATIENT SAFETY PLAN

WPSC Teleconference Avoiding Never Events. Linda Furkay, PhD, RN Patient Safety Adverse Event Officer

Development and assessment of a Patient Safety Culture Dr Alice Oborne

A Client Satisfaction Survey at a Large Rural Medical Facility

Quality Improvement and Patient Safety (QPS) Ratchada Prakongsai Senior Manager

Version: 3.0. Effective from: 29/08/2012

10 Steps to Maximizing Nurses Full Scope of Practice Utilization in Primary Care Settings

Creating a Change Team

Midwest Alliance for Patient Safety Patient Safety Organization Getting Started with a PSO. An Illinois Hospital Association Company

HOUSE OF REPRESENTATIVES STAFF ANALYSIS REFERENCE ACTION ANALYST STAFF DIRECTOR SUMMARY ANALYSIS

2017 Good Catch Program: Blueprint Companion Guide

Executive Summary DIRECTORS MANAGERS CNO/CNE. Respondent Profile 32% 26% 17%

Mandatory Public Reporting of Hospital Acquired Infections

Risk Management in the ASC

Quality Improvement Plans (QIP): Progress Report for Q3

Enhancing Patient Quality and Safety with Compliance

Hospital Survey on Patient Safety Culture: Debrief and Action Planning

2015 Lasting Change. Organizational Effectiveness Program. Outcomes and impact of organizational effectiveness grants one year after completion

SHARED SPACE and SERVICES

Company/Project Name. Table of Contents

READMISSION ROOT CAUSE ANALYSIS REPORT

Improving the Pre-Empted Medication Error Reporting System at St. Charles Hospital, Port Jefferson, NY

Bedside Teaching Creating Competent Physicians

REVIEW OF SEQOHS STANDARDS

Performance Improvement Projects (PIP) Clinic May 13, 2016

Bureau of Milwaukee Child Welfare Period 7 Corrective Action Plan: Jeanine B. v. Doyle Settlement Agreement May 24, 2010

Recommendations for Adoption

Root Cause Analysis. Why things happen

MedPAC June 2013 Report to Congress: Medicare and the Health Care Delivery System

Collaborating to Make Communities Healthier: A Case Study in Community Coalitions

Collaborating to Make Communities Healthier: A Case Study in Community Coalitions

Tools, Resources and Modules

Expanding Improvement Science Competencies: Successes & Challenges Terry L. Jones RN, PhD. utexas.edu/nursing

Success Story Winner 2010

On the CUSP: Stop BSI

Diagnostic Waiting Times

September 25, Via Regulations.gov

EXECUTIVE SUMMARY THE LOS ANGELES FAMILY AIDS NETWORK (LAFAN) 2003 HIV/AIDS CARE NEEDS ASSESSMENT 1

Request for Grant Proposals. September 2, 2009

Positive and Safe Management of Post incident Support and Debrief. Ron Weddle Deputy Director, Positive and Safe Care

More staff in country/remote areas had attended one training session only compared to their metropolitan counterparts (58% versus 45%).

3/30/2015. Objectives. Cooking Up a QAPI: Recipe for Success Under the new COPs Part 1

Indicator Definition

University of Michigan Health System. Current State Analysis of the Main Adult Emergency Department

Using MEDMARX for Reporting and Benchmarking. Anne Skinner, RHIA Katherine Jones, PhD, PT

Cross-Site Data Reporting and. Evaluation. Phillip W. Graham, DrPH. Pamela Roddy, PhD. PFS Cross-site Evaluation Project Director CSAP, PEPC COR

UPMC POLICY AND PROCEDURE MANUAL

RACGP Standards for general practices (5 th edition)

Root Cause Analysis LITE (RCA Lite)

North Carolina Department of Commerce Small Cities Community Development Block Grant Program (CDBG)

Leveraging Nursing Expertise with USP<800>

University of Michigan Health System Part IV Maintenance of Certification Program [Form 12/1/14]

Quality Assessment and Assurance. Guidance Training (F520) (o)

Busines Professi. Regulation. Annual Report Fiscal Year Florida Departmentof. Florida State Boxing Commission

STRENGTHENING THE REGIONAL CONSERVATION PARTNERSHIP PROGRAM FOR THE CHESAPEAKE BAY REGION

Disclosures. Relevant Financial Relationship(s): Nothing to Disclose. Off Label Usage: Nothing to Disclose 6/1/2017. Quality Indicators

UPMC POLICY AND PROCEDURE MANUAL

2014 Edition FUNDRAISING WITH ARTEZ INTERACTIVE WHITE PAPER FACEBOOK ARTEZ.COM FACEBOOK.COM/ARTEZINTERACTIVE

PERSON CENTERED CARE PLANNING HONORING CHOICE WHILE MITIGATING RISK

IDAHO NONPROFIT SECTOR IN BRIEF

Laura Hempstead, DO, FACOFP AODME April 22/2015

CAREER SERVICES USE OF SOCIAL MEDIA TECHNOLOGIES

DRAFT SUPPLEMENTARY GUIDANCE & DRAFT SIMPLIFIED PLANNING ZONE SCHEME CENTRAL BORDERS BUSINESS PARK, TWEEDBANK

Annual provider survey results 94%

Diagnostic Waiting Times

NATIONAL LOTTERY CHARITIES BOARD England. Mapping grants to deprived communities

NetworkNotes. U.S. Behavioral Health Plan, California (USBHPC) News for Clinicians and Facilities Fall 2009

Handling Organisational Complaints

SPSP Medicines. Prepared by: NHS Ayrshire and Arran

A HOSPITAL SELF-ASSESSMENT INVENTORY

Diagnostic Waiting Times

Quality Impact Assessment Policy

Caregivers and Digital Health: A Survey of Trends and Attitudes of Massachusetts Family Caregivers

2.b.iii ED Care Triage for At-Risk Populations

HUDSON CORRECTIONAL FACILITY REENTRY UNIT

Service Agreements. Mike Davies, MD FACP

Certification Body Customer Satisfaction Survey 2017 Summary Report

Guidance for Setting up and Engaging Patients and Family Members on Patient Councils

TE18 Review Process and Responsibilities

RISK MANAGEMENT AND PATIENT SAFETY

California HIPAA Privacy Implementation Survey: Appendix A. Stakeholder Interviews

Non-Time Limited Supportive Housing Program for Youth Request for Proposals for Supportive Housing Providers (RFP)

Planning for Improved Access and Orientation Anjali Joseph Jain EDRA 35

DESIGN COMPETITIONS: Why? And what it takes.

A CME Activity Developed by National Jewish Health and Medscape Education

Nursing Certification Research Perceived Value of Certification Tool (PVCT)

Lincoln Lancaster County Health Department, NE. Accreditation Preparation and Quality Improvement Demonstration Sites Project.

4/26/2017. I ll Do It My Way, Thank You Performance Improvement Strategies for Home Care. Session Objectives. Session Agenda

WHITE PAPER. Taking Meaningful Use to the Next Level: What You Need to Know about the MACRA Advancing Care Information Component

ERN Assessment Manual for Applicants 2. Technical Toolbox for Applicants

SAFEGUARDING CHILDEN POLICY. Policy Reference: Version: 1 Status: Approved

August 15, Dear Mr. Slavitt:

Data on Written Complaints in the NHS Q4 Provisional Experimental statistics

MSCRF Discovery Program

National Survey on Consumers Experiences With Patient Safety and Quality Information

PATIENTS PERSPECTIVES ON HEALTH CARE IN THE UNITED STATES: NEW JERSEY

Root Cause Analysis: The NSW Health Incident Management System

CACFP : Conducting Five-Day Reconciliation in the Child and Adult Care Food Program, with Questions and Answers

PERSPECTIVES. Under Pressure: Front-Line Experiences of Medi-Cal Eligibility Workers. Overview. Current Environment

Document Details Clinical Audit Policy

Transcription:

Hospital Adverse Event Reporting Program: an Initial Evaluation When the Patient Safety Commission began the Adverse Event Reporting Program in May, 2006, we promised hospitals we would complete an evaluation within six months. To honor our pledge, we fielded a survey in November. Our goal was to measure progress in creating a reporting program that supports quality improvement activity. We were specifically interested in comments about the reporting template, the impact of the program, and our ability to put the reported data to good use. To provide a useful evaluation we developed two versions of our survey one for hospitals that had submitted adverse event reports (24 hospitals as of November, 2006) and one for those that had not (28 hospitals). Our response rates for completed surveys: Overall: 32 of 52 hospitals completed surveys ( 60%) Among those that had filed an adverse event report: 19 of 24 (79%) Among those that had NOT submitted an adverse event report: 13 of 28 (46%) Following is a brief summary of findings and an overview of our next steps. Complete survey results are available on the Commission s website at www.oregonpatientsafety.org. Independent of this evaluation we will publish a report of 2006 adverse event data later in January. Overall Assessment In general it seems that the Adverse Event Reporting Program is off to a good start, though we need to continue to raise the visibility of the program. Senior management and boards of directors appear to be more aware of the Reporting Program than hospital clinical staff. Overall there was a clearly articulated need for streamlining the reporting process and for providing hospitals with more results (especially quarterly reports), action items, feedback and training. Based on our evaluation, we are now ready to make improvements to the reporting program. Planned changes include: Shortening the reporting form. Providing written guidance on what to include in the form. Clarifying definitions of events and harm levels, especially as related to disclosure requirement. Initiating discussions with hospitals around how to handle overlap with peer review. Identifying Commission announcements more clearly when sending out via E-mail; including all reports/announcements on the website. Increasing contact with hospitals to assist in identifying reportable events and providing hospitalspecific trainings and support. Developing an Advisory Committee for analysis of adverse events and assistance with improving RCAs. In the coming weeks we will work with our partners to simplify and improve the adverse event reporting program. For further information or comments, please contact: Jim Dameron, Administrator jim.dameron@oregonpatientsafety.org or Leslie Ray, Field Coordinator leslie.ray@oregonpatientsafety.org

Initial Evaluation of Hospital Reporting Program, Jan. 2007 Page 2 of 5 The Reporting Tool and Process In general, respondents indicated that the Commission s definitions, reporting tool and process were relatively easy to use and consistent with already-in-place procedures. However, respondents suggested that the Commission: clarify definitions of serious, less serious adverse events, harm levels, and reporting near-misses ; recommend how to handle adverse events when the cause is difficult to identify; provide understandable boundaries around peer review; reduce overlap in categories of contributing factors; reconsider action plan categories; eliminate need for back and forth communication to clarify reports. We specifically asked hospitals to help us identify barriers to participation in the reporting program. For hospitals that have submitted reports, 7 of 19 (37%) indicated they faced challenges to full participation. Among the barriers mentioned: differing criteria for initiating an RCA, especially for temporary events; physician lack of engagement and concern over liability; time and resource constraints; internal inconsistencies and gaps in policy, procedures. Among the hospitals that did not submit reports, identifying an appropriate event to report was the most frequently cited barrier, followed by the type/amount of information requested (see below for frequency of each barrier noted and please keep in mind the small number of responses): Barriers Identified by Respondents from Hospitals that Have Yet to Submit a Report n = 11 Notifying the patient of an adverse event 4 Completing the investigation/root cause analysis 4 Barriers Identifying appropriate events to report 7 Yes No Concerns about confidentiality 3 Type/amount of information requested 5 1 2 3 4 5 6 7 8 9 10 11 P rogram Visibility to To be successful, the Patient Safety Commission s reporting program needs to be an active and visible part of a hospital s information structure. To judge our progress, we asked participants rate our visibility using a five point scale (1 = high visibility; 5 = low visibility). The overall

Initial Evaluation of Hospital Reporting Program, Jan. 2007 Page 3 of 5 results suggest we have made progress on the administrative side, but need to continue to reach out to the clinical teams. Among all hospitals, senior management (2.3) and Boards (2.8) were more likely to know about the Patient Safety Commission and its reporting program than medical or nursing staff (3.2 and 3.4 respectively). Clearly, we have not yet saturated the entire hospital s culture. For example, among hospitals submitting reports, visibility within the reporting department was fairly high (1.9), but in the hospital overall it was lower, averaging 3.6. Comparison of visibility between those hospitals submitting reports and those who have yet to submit a report is below. Visibility of the Patient Safety Reporting Program within Hospitals that Have Submited Reports and Those Who Have Not 1 High Visibility - 5 Low Visibility Sr Mgt Hospital Area Board Med Staff Nsg Staff 1.00 2.00 3.00 4.00 5.00 Visibility Program Impact To determine if the Commission is beginning to make a real difference we asked participants (who had filed a report), Has the Patient Safety Commission had an impact on your hospital s % Yes Internal Incident Reporting 37% Investigation of Adverse Events 42% Ability to learn from Adverse Events 53% Culture of Safety 58% Almost 6 in 10 respondents said that the Commission has had a positive impact on their organization s culture of safety; this is a good start. But clearly we can do more. We also asked the hospitals that had not yet submitted reports to evaluate their organization s overall attitude toward the reporting program (from very positive to very negative). The results

Initial Evaluation of Hospital Reporting Program, Jan. 2007 Page 4 of 5 within Risk Management are encouraging, while the results among clinical staff reinforce the need to work more directly with physicians and nurses: Your Organization s attitude toward the reporting program (Hospitals that have not yet submitted adverse event reports)* Very or somewhat Neutral Somewhat or Department positive very negative Within the risk/legal departments 72.7% 18.2% 9.1% (1/11) With senior management 63.6% 27.3% 9.1% (1/11) With nursing staff 27.3% 72.7% 0.0% With medical staff 18.2% 54.5% 27.3% (3/11) With the Board of Directors 50.0% 50.0% 0.0% * Please keep in mind that only 11 respondents answered this question. Program Activities As the Commission receives information from participating hospitals it is critical that we put that information to good use. To see how we are doing we asked about our recent program activity. For example, in 2006 we published several types of announcements (Bulletins, Safety Tips), issued a report on Colored Wristband Use, and held two teleconferences for small and rural hospitals. As the following graph suggests, almost all of the respondents saw the two bulletins, and a large percent also saw the report on Colored Wristband Use. However, a much smaller percent remembered seeing either of the two patient safety tips. Percentage of Respondents Who Had Seen Safety Announcements n = 30 Ticket to Ride Show Me the Mark Type of Announcement Colored Wristbands Kinked Tubing MRI's 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% Respondents Among respondents who have seen the various announcements a large majority found them helpful. However the perceived usefulness of safety tips varied by whether a hospital had submitted reports or not. Those who had were much more positive:

Initial Evaluation of Hospital Reporting Program, Jan. 2007 Page 5 of 5 Reported Usefulness of Safety Announcements n = 29 Ticket to Ride Show Me the Mark Type of Announcement Colored Wristbands Kinked Tubing MRI's 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% Respondents Summary This informal evaluation of the first six months of the Hospital Reporting Program suggests we have made some real progress in raising awareness and adding value to on-going patient safety efforts. However in 2007 we need to redouble our efforts to reach out to clinical teams, to offer root cause analysis training, to make thoughtful recommendations based on aggregate data.