Safer Care - Measuring to Manage and Improve
Fast-forward to today and patient safety is on the agenda worldwide. In Canada the first nationwide study of adverse events in hospitals was published in 2004 (Baker et al. 2004). Healthcare providers, the new Canadian Patient Safety Institute, provincial institutes and task forces, and many others are working to respond to the results of the study.
While medical practice has changed since the days of Ernest Codman, what has not changed is the focus on having good information to guide quality improvement efforts. Measures are required at a variety of levels (see Figure 1). For instance, broad-based global metrics provide information about the prevalence of adverse events and their impact on patients. Healthcare organizations often seek to track patient safety outcomes for their patients, as well as related processes of care. Individual quality improvement teams also require detailed information to monitor their progress in specific areas. This information may be collected as part of rapid cycle improvement or other change processes and will evolve over time depending on the focus of quality improvement efforts.
From the Global PerspectiveIn 2004 approximately one in four (23%) Canadian adults 15 years and older said that they or a member of their family had experienced an adverse event related to their medical care (Canadian Institute for Health Information 2004). That translates to about 5.2 million people across the country. Almost a third (30%) said that the most recent event happened within the last year.
In hospitals, the largest study in Canada of adverse events found that between 9,250 and 23,750 medical and surgical adult patients with overnight hospital stays in 2000-2001 experienced a preventable adverse event and later died (Baker et al. 2004). Interestingly, the public tends to estimate much lower numbers of deaths. In 2003-2004 the Canadian Institute for Health Information (CIHI) commissioned a telephone survey of just over 4,200 adults across the country. Only 7% of respondents thought that 10,000 or more Canadians die in hospital each year from preventable adverse events (Canadian Institute for Health Information 2005a). Another 21% said that they did not know how many people died annually. Similarly, most respondents to a 2000 survey in the U.S. believed that fewer in-hospital deaths due to preventable errors occurred than estimated by authors of a landmark study by the Institute of Medicine (Blendon et al. 2002).
While many studies provide overall estimates of adverse event rates, more detailed research demonstrates that the frequency of specific types of adverse events varies widely. For example, adverse events related to medications are much more common than those related to infected blood transfusions (see Table 1). Emerging data also suggest that rates may vary significantly from one part of the country to another. Regional in-hospital hip fracture rates, for instance, ranged from 0.5 to 3.4 per 1,000 seniors admitted to Canadian acute care hospitals in 2001-2002 to 2003-2004 (excludes Quebec and Manitoba) (Canadian Institute for Health Information 2005b).
A View From The Organizational PerspectiveKnowing the extent to which adverse events occur within a population provides a baseline from which to start to measure change. However, individual health care organizations may also use more detailed qualitative and quantitative measures to track their progress towards safer care and to identify opportunities for quality improvement.
One option is to compare outcomes with other similar healthcare providers. The majority of hospital executives in all five countries surveyed by the Commonwealth Fund in 2003 felt that this approach would be somewhat or very effective in improving quality of care (see Figure 2) (Blendon et al. 2004). Compared with other countries, Canadian hospital executives were among the most supportive of disclosing quality information, such as the rates of nosocomial infections and medical errors, to the public.
Already, a number of initiatives are underway that build on these premises. For example, a number of health regions and hospitals across the country participate in the CIHI/Hay Group Benchmarking Comparison of Canadian Hospitals. In recent years, this project has included a range of patient safety indicators. At a regional level, comparable data on selected outcomes of care (e.g., 30-day in-hospital acute myocardial infarction mortality) and patient safety (e.g., in-hospital hip fracture rates) are produced through the CIHI/Statistics Canada Health Indicators Project. There are also several provincial initiatives and efforts to provide comparable information for different specialty areas.
Additional initiatives are also emerging. For example "Safer Healthcare Now!", a grassroots patient safety campaign aimed at reducing preventable complications and deaths, is testing the use of intervention-specific process and outcome measures, as well as broad-based safety indicators. Originally developed in the United Kingdom, Hospital Standardized Mortality Ratios (HSMRs) compare observed versus expected deaths on a hospital-specific basis, adjusted for the age, sex, diagnoses, and admission status of its patients (Jarman et al. 2005). The Institute for Healthcare Improvement in the United States is now using this measure to track the success of its 100,000 Lives patient safety campaign, and it will be a core measure for the Canadian Safer Healthcare Now! campaign.
HSMRs provide a baseline from which hospitals can track and compare their results over time. In 2000, for example, the Walsall Hospitals NHS Trust in England had 1,080 deaths compared with the 830 that would be expected based on the patient mix that they cared for (Jarman et al. 2005). This translates into an HSMR or 130, the highest level of any hospital in the country at the time. Through a series of concerted improvements, over a four-year period they reduced their HSMR to 93. That represents a reduction of 295 observed compared with expected deaths per year.
From the Team PerspectiveMeasurement is at the heart of many quality improvement efforts. For example, Plan, Do, Study, Act (PDSA) cycles are being used by healthcare teams across Canada and around the world. This approach uses pragmatic data collection and measurement activities to inform and support incremental changes in the process of care. For the local teams leading these initiatives, measurement is not the goal; rather it is a tool that facilitates progress towards the goal. Unlike measurement for research, data used by quality improvement teams often involves smaller samples and less complex collection methods (Institute for Healthcare Improvement 2005).
In some cases, teams may be able to build on shared approaches to data collection and analysis. For example, The Guidelines Applied in Practice (GAP), endorsed by the American Heart Association and widely accepted internationally, outlines five specific practices at time of discharge that have been shown to reduce mortality in patients with heart disease (Parsons et al. 2002). Application of GAP-related improvements is one of six strategies in Safer Healthcare Now! The campaign intends to provide tools that can be used by individual teams to track their progress over time.
ConclusionsAccording to the World Health Organization (WHO), adverse events represent "a challenge to quality of care, a significant avoidable cause of human suffering, and a high toll in financial loss and opportunity cost to health services" (WHO 2002). To address this challenge, WHO, in conjunction with its partners, launched the World Alliance for Patient Safety in October 2004 to reduce the number of preventable illnesses, injuries, and deaths patients experience during their care.
In Canada and elsewhere, in order to know whether progress is being made and where further opportunities for improvement might exist, high-quality information is required at multiple levels. At a macro level, we need to know how many Canadians experience preventable adverse events, as well as how the situation is changing over time. As Ernst Codman pointed out almost a century ago, health regions and healthcare providers also need more detailed information to understand the progress of their quality improvement initiatives and patient outcomes following care. And finally, healthcare teams can test rapid improvement strategies by collecting and rapidly responding to data that tracks the results of their efforts.
About the Author
Kira Leeb, MA, Manager, Health Services Research, has managed the production of CIHI's annual report on the health of Canada's healthcare system and has led the production of more focused reports or research on Canada's healthcare system.
Jennifer Zelmer, PhD, is Vice-President, Research and Analysis. In this role, Dr. Zelmer leads an integrated program of health services and population health-related analytical and research initiatives.
Greg Webster, MSc, Director, Research and Indicator Development. In this role, Mr. Webster is responsible for providing vision, leadership and direction in fostering CIHI's relationships with the health services research community, in developing comparative health indicators and in enhancing CIHI's analytical capacity.
Indra Pulcins, PhD, Manager, Health Indicators. Dr. Pulcins conducts health services research and is responsible for the development of health indicators at CIHI, with a particular focus on outcomes and patient safety.
Corresponding Author: Greg Webster, Director, Research and Indicator Development, Canadian Institute for Health Information (CIHI), 90 Eglinton Avenue East, Suite 300, Toronto, Ontario M4P 2Y3, Phone: (416) 481-1616 Ext. 3508, Fax: (416) 481-2950, e-mail: email@example.com, web-site: www.cihi.ca
Baker G. R., P. G. Norton, V. Flintoft, R. Blais, A. Brown, J. Cox, E. Etchells, W. A. Ghali, P. Hébert, S. R. Majumdar, M. O'Beirne, L. Palacios-Derflingher, R. J. Reid, S. Sheps and R. Tamblyn. 2004. "The Canadian Adverse Events Study: the incidence of adverse events among hospital patients in Canada." Canadian Medical Association Journal 170(11): 1678-86.
Blendon, R. J., C. M. DesRoches, M. Brodie, J. M. Benson, A. B. Rosen, E. Schneider, D. E. Altman, K. Zapert, M. J. Herrmann, and A. E. Steffenson. 2002. "Views of Practicing Physicians and the Public on Medical Errors." New England Journal of Medicine 347(24): 1933-40.
Blendon R.J., C. Schoen, C. M. DesRoches, R. Osborn, K. Zapert, and E. Raleigh. 2004. "Confronting Competing Demands To Improve Quality: A Five-Country Hospital Survey." Health Affairs 23(3): 119-35.
Canadian Institute for Health Information (CIHI). 2004. Health Care in Canada, 2004. Ottawa.
Canadian Institute for Health Information (CIHI). 2005a. "The Berger Population Health Monitor." Ottawa. (Unpublished, 2004.)
Canadian Institute for Health Information (CIHI). 2005b. Health Indicators, 2005. Ottawa.
Institute for Healthcare Improvement. 2005. Measures. http://www.ihi.org/IHI/Topics/Improvement/
Jarman B., A. Bottle, P. Aylin, and M. Browne. 2005. "Monitoring changes in hospital standardized mortality rations." British Medical Journal 330: 329.
Parsons L.S., C.V. Pollack, K. Newby, and K.A. Littrell. 2002. "Death rates lower in hospitals that follow heart attack guidelines." American Heart Association's Scientific Sessions 2002: Abstract 3557.
World Health Organization. 2002. "Quality of Care: Patient Safety." Report by the Secretariat to the Fifty-Fifth World Health Assembly. Geneva.
Be the first to comment on this!
Personal Subscriber? Sign In
Note: Please enter a display name. Your email address will not be publically displayed