- Research
- Open access
- Published:
A retrospective analysis using comorbidity detecting algorithmic software to determine the incidence of International Classification of Diseases (ICD) code omissions and appropriateness of Diagnosis-Related Group (DRG) code modifiers
BMC Medical Informatics and Decision Making volume 24, Article number: 309 (2024)
Abstract
Background
The mechanism for recording International Classification of Diseases (ICD) and diagnosis related groups (DRG) codes in a patient’s chart is through a certified medical coder who manually reviews the medical record at the completion of an admission. High-acuity ICD codes justify DRG modifiers, indicating the need for escalated hospital resources. In this manuscript, we demonstrate that value of rules-based computer algorithms that audit for omission of administrative codes and quantifying the downstream effects with regard to financial impacts and demographic findings did not indicate significant disparities.
Methods
All study data were acquired via the UCLA Department of Anesthesiology and Perioperative Medicine’s Perioperative Data Warehouse. The DataMart is a structured reporting schema that contains all the relevant clinical data entered into the EPIC (EPIC Systems, Verona, WI) electronic health record. Computer algorithms were created for eighteen disease states that met criteria for DRG modifiers. Each algorithm was run against all hospital admissions with completed billing from 2019. The algorithms scanned for the existence of disease, appropriate ICD coding, and DRG modifier appropriateness. Secondarily, the potential financial impact of ICD omissions was estimated by payor class and an analysis of ICD miscoding was done by ethnicity, sex, age, and financial class.
Results
Data from 34,104 hospital admissions were analyzed from January 1, 2019, to December 31, 2019. 11,520 (32.9%) hospital admissions were algorithm positive for a disease state with no corresponding ICD code. 1,990 (5.8%) admissions were potentially eligible for DRG modification/upgrade with an estimated lost revenue of $22,680,584.50. ICD code omission rates compared against reference groups (private payors, Caucasians, middle-aged patients) demonstrated significant p-values < 0.05; similarly significant p-value where demonstrated when comparing patients of opposite sexes.
Conclusions
We successfully used rules-based algorithms and raw structured EHR data to identify omitted ICD codes from inpatient medical record claims. These missing ICD codes often had downstream effects such as inaccurate DRG modifiers and missed reimbursement. Embedding augmented intelligence into this problematic workflow has the potential for improvements in administrative data, but more importantly, improvements in administrative data accuracy and financial outcomes.
Contributions to Literature |
---|
• Computer algorithms are able to automatically audit ICD coding omission by targeting disease states using raw EHR data |
• By auditing ICD codes, DRG codes can be evaluated for proper application of severity modifiers automatically |
• Insight into the differing rates of ICD code omission among groups by age, sex, payor, and ethnicity |
Introduction
The International Classification of Diseases (ICD) (the most widely used disease catalogue today) originated in 1900 with 180 diseases and has steadily developed into over 120,000 codes [1]. These codes offer incredible specificity allowing healthcare coders to assign precise disease states to patients’ medical records. Beyond good record keeping, ICD codes have become the foundation of disease data for subsequent layers of codes such as the Diagnosis Related Groups (DRG) codes. Developed in 1976, DRG codes are a means to structure the overlapping risk and complexity of patient diagnoses, in-patient medical care and surgical procedures performed in order to arrive at an estimated hospital resource utiliziation [2]. As of 1983, Medicare’s Inpatient Prospective Payment System made ICD and DRG codes the basis for inpatient hospital payments in the United States [3].
The recording or ICD and DRG codes from a patient chart is performed by a certified medical coder. At the completion of a hospitalization, coders review the medical record to assign primary and secondary ICD diagnosis codes along with any applicable procedure codes. ICD codes are then used to assign DRG codes via coding algorithms; with high-acuity ICD codes justifying comorbidity modifiers for DRG codes [4]. The Center for Medicare and Medicaid Services (CMS) supports up to three tiers of DRG comorbidity modifiers: base level DRG indicating that patients do not have any complicating diagnoses outside of the reason for hospital admission, Comorbid Condition diagnoses known as “CC,” and Major Comorbid Condition known as “MCC.” [5] The inclusion of CCs or MCCs in the patient’s coded chart shifts the assigned DRG within the group to one that on average justifies increased resource utilization and therefore increased hospital payment to reflect the more resource intensive healthcare needs.
Because of healthcare coding and billing regulation, modern electronic health records (EHR) have adapted tools that allow physicians to document ICD codes within treatment workflows and often require ICD justification when ordering procedures, medications, or laboratory specimens. This is typically done using systematized nomenclature of medicine clinical terms (SNOMED-CT) which allows clinicians to search a disease reference catalogue while storing the associated ICD codes in the EHR [6]. Provider entries are stored in staging areas within the EHR facilitating the process for certified coders to migrate the ICD codes to the subsequent billing schemas [7].
Given their structure, ICD codes are used for more than hospital billing. ICD codes are the cornerstone of epidemiologic disease identification in administrative databases and are often used for both Continuous Quality Improvement (CQI) and Health Services Research (HSR). At a national level, payor databases (Medicare or private insurers) and disease specific registries are large repositories with longitudinal patient level data that are frequently indexed by ICD codes. These administrative databases have been heavily used in health economic & outcomes research, precision medicine, quality assessment, hospital & physician report cards, and benchmarking [8,9,10]. Groups like the Agency for Healthcare Research and Quality (AHRQ) routinely use ICD coding to facilitate the development of clinical practice guidelines [11]. Lastly, these databases are used by payors, such as CMS, to facilitate future reimbursement level for procedures and DRGs [12].
Despite the need for accurate ICD coding, many manuscripts have been written about the possible sources of ICD code inaccuracies and have identified sources of error [1, 9, 13,14,15]. As far back as 1984, the Office of the Inspector General found that 61.7% of inpatient medical records within their study sample had ICD coding errors that favored hospital billing [14]. A subsequent study by the Inspector General found that 14.7% of ICD errors were related to diagnosis codes which would have resulted in payment above the base level DRG [16]. However, both of these studies and many of the other studies that followed have relied on manual clinician chart review to find these coding errors and associated data gaps [17, 18]. While informative about the problems and deficiencies inherent in using administrative data, these studies do little to provide any insight or direction toward a scalable solution.
As an alternative to relying on administrative data, researchers have commonly preferred to use raw clinical EHR data [10]. Using data repositories and intricate disease identifying algorithms, previous manuscripts have been able to demonstrate high fidelity data to assign disease states [19, 20]. Using similar algorithm-based disease identification methods, the gap between clinical and administrative data can be evaluated more closely.
The recent rise of AI, especially neural networks with natural language processing, offers a solution to inaccuracies in ICD coding by automating the documentation of disease states with high accuracy. Neural networks can detect complex patterns in EHRs, reducing human error. However, this comes with a tradeoff: while AI models offer speed and efficiency, they lack the flexibility to fine-tune definitions like rule-based algorithms. As AI systems are often opaque, they may sacrifice coding precision in favor of broader pattern recognition, posing a challenge for ensuring accuracy in clinical settings.
In this manuscript, we hypothesize that rules-based algorithms relying on raw EHR data could be used to audit codes (ICD and DRG) automatically and at scale without the need for manual chart review. As a primary outcome, we evaluate the incidence of missing ICD codes for co-morbid diseases in a cohort of hospital admissions using automated rules-based algorithms based on structured EHR data. Secondarily, we evaluate the extent to which missing ICD codes for the true clinical picture led to an incorrect DRG assignment for the admission, as well as the potential financial impact of these omissions. Lastly, we examine the missing documentation for clinically important demographic associations, especially as it relates to historically underserved groups.
Materials and methods
Data extraction
This study (IRB# 15–000518) qualified for IRB exemption status through the UCLA Human Research Protection Program by virtue of having no direct patient contact and using a de-identified dataset. All study data were acquired via a previously published Department of Anesthesiology and Perioperative Medicine at UCLA’s Perioperative Data Warehouse (PDW) [21]. The PDW is a structured reporting data schema that contains all the relevant clinical data entered into the EPIC (EPIC Systems, Verona, WI) electronic health record (EHR). Data are acquired via Clarity, the relational database created by EPIC for data analytics and reporting. While Clarity contains raw clinical data, the PDW was designed to organize, filter, and improve data so that it can be used reliably for creating robust disease phenotypes using clinical logic. Other published manuscripts deriving data from the PDW can be found in the reference Sects. [21,22,23,24,25,26,27,28].
Despite the name including the word “perioperative” the DataMart contains data on all UCLA patients regardless of their operative status. It includes data from both the Ronald Reagan UCLA main hospital (500 + beds), Santa Monica UCLA hospital (275 + beds), and all the affiliated outpatient locations.
Inclusion criteria
Data were extracted for patients who had an inpatient admission to any UCLA health system hospital in calendar year 2019. The rationale for choosing 2019 was to avoid any incomplete billing and only evaluate inpatients that have since been discharged with established DRG designations. Additionally, 2019 was chosen as opposed to 2020 to avoid confounding effects due to the COVID-19 pandemic. If an admission did not have a DRG assigned, or there was no single primary DRG code, then it was excluded from the analysis. Each of these analyses were done at the hospitalization-encounter level so that if a patient was admitted to the hospital multiple times, each admission was treated independently.
Creating algorithms to identify comorbid diseases
Selected diseases were curated from the CMS list of diseases that met criteria for DRG modifiers. The authors chose diseases that had clear objective criteria that could be mined from EHR data, but it is entirely feasible to create more algorithms that phenotype other disease states. Two of the authors (EG, IH), with experience in designing disease-based algorithms, reviewed the CMS list then came to a consensus with the disease list shown in Table 1 along with the modifier level, ICD codes and structured definitions, and the data sources [19, 22, 29]. The data used for each disease algorithm used mostly clinical data in the form of labs results, echocardiography reads, physician and nursing documentation, medication administrations, operative procedures, and orders. The specific for each individual algorithm are detailed in Table 1. Using this data, each admission was then flagged as having or not having each of the diseases.
Identifying cases with missing administrative data
For each of the eighteen disease algorithms, a global list of ICD was generated by looking at historic hospital admission that satisfied the algorithm logic. Then the list was manually reviewed to only include the ICD codes that were clinically relevant to the specific algorithm and qualify for DRG modifiers. In an example, the algorithm evaluating myocardial injury was satisfied by ICD codes for myocardial infarction, acute ischemic disease, or unspecified cardiac injury. The final list of ICD codes is found in Table 1 with accompanying algorithm definitions. Admissions that had clinical evidence of the disease in the algorithm but did not have any of the selected ICD codes were flagged as missing applicable ICD data.
Categorizing the diagnosis related groups (DRGs)
A complete list of DRG codes was downloaded from the Center for Medicare and Medicaid Services (CMS) website: www.CMS.gov. Within the downloaded file, data for each DRG included the DRG value and a text description.
After importing the DRG data into the PDW database, each unique DRG was designated a comorbidity level; “MCC”, “CC”, “Base”, or “Singlet.” While some DRG codes are “Doublets” indicating that there are only 2 possible states, they typically have a “Base/CC” and “CC/MCC” designation and modification is still partly allowable. MCC indicated that the DRG included a major comorbid condition (MCC) which was triggered from the text description in the CMS data containing the phrase “WITH MCC” at the end of the description such as “PLEURAL EFFUSION WITH MCC.” The same applied for DRG’s with the comorbid condition (CC) designation. These code descriptions ended with the phrase “WITH CC’’ such as “PLEURAL EFFUSION WITH CC.” For codes where the description included verbiage that excluded MCC or CC levels, such as “PLEURAL EFFUSION WITHOUT CC/MCC’’, we assigned the “Base” level. This implied that the same diagnosis DRG had the ability to scale upwards. Lastly, the “Singlet” designation was given to codes that had no indication of the presence of CC/MCC modifiers and had no codes with the same descriptors to indicate the ability to increase or decrease the level of complexity.
Once the codes were classified using discrete column variables to indicate comorbidity level, all the text modifiers were stripped out of the text descriptions. For example, “PLEURAL EFFUSION WITH MCC”, “PLEURAL EFFUSION WITH CC,” and “PLEURAL EFFUSION WITHOUT CC/MCC” where all grouped to “PLEURAL EFFUSION’’ and classified as a single disease state. This approach allowed us to identify diagnoses where there exists the ability to change the DRG based on the presence or absence of relevant comorbidities.
Evaluating DRGs for modifiers
When an algorithm detected the presence of a disease state, and the billed DRG comorbidity level was below the comorbidity level of the algorithm disease comorbidity, a flag was assigned to the admission to indicate an inaccurate comorbidity level was coded since the missing ICD code would have justified a higher DRG modifier. For example, admissions that were flagged by the acute pancreatitis algorithm (MCC worthy), but only had existing DRG codes with base or CC level modifiers, were flagged as appropriate for MCC designation since pancreatitis was appropriate based on conservative limits of clinical laboratory data. There are three possible upgrade types: base to CC, CC to MCC, and base to MCC. In the event that a different disease algorithm flagged an admission for comorbidity level modification, the highest DRG modifier level was used; essentially eliminating redundancy. If an admission DRG was given the “Singlet” designation, then there was no consideration for possible modification. This process was applied to each hospitalization encounter.
Categorizing hospital payors
Payors were classified into three categories: Medicare, Medi-Cal, and Private. The payor grouping was done manually by evaluating the 2019 payors in the EMR by financial class. Medi-Cal and Medicaid were consolidated to Medi-Cal, while Medicare and Medicare Advantage were consolidated to Medicare. All the other payor types were grouped into the private category.
Incorporating weighting factors
Using the CMS website, we downloaded data on the 2019 DRG Relative Weighting Factors (RWF) [24]. RWF are used in hospital billing to assign hospital reimbursement to DRG values, essentially putting a monetary value to each DRG based on resource intensity. In practice different payors negotiate and contract assigning dollar amounts to RWF (in fact this may also differ by provider and insurance line of business for a given payor). For analysis purposes, we estimated that a single RWF point was worth $20,000 for private payors, $10,000 for Medicare, and $7,500 for Medicaid (Medi-Cal). These values were informed by our previous experience using RWF [30].
When evaluating hospital admissions that were flagged for upgrade by the disease algorithms, we calculated the delta RWF between the originally billed DRG, and the algorithm proposed upgraded DRG with modifiers. When multiplying the delta RWF by the estimated payor-based RWF value, we were able to derive the lost revenue for not coding the higher level DRG.
Statistical methods
Incidence rates were computed for each condition for both primary and secondary outcomes (ICD/DRG) and 95% confidence intervals were constructed using the binomial distribution. Analyses were conducted using R v4.1.0.
In looking for findings did not indicate significant disparities between age, sex, and payor, these data were analyzed using two methods, Standardized Mean Difference (SMD) and direct comparison of each group to a reference group within each category.
SMD: Given the large sample size, the analysis is overpowered in determining significant differences. In other words, a 1% difference would be statistically significant with little clinical value. Published guidelines for interpreting the magnitude of the SMD in the social sciences include: small, SMD 0-0.4; medium, SMD 0.4–0.6; and large, SMD 0.6-1.0 [25].
Reference Groups: Reference groups were chosen from the race, payor, and age categories to compare the incidence of ICD miscoding within each category. The index groups were Caucasians (non-Hispanic), private payors, and middle-aged patients. Each comparison was then represented by a p-value with values < 0.05 being statistically significant.
De-identification
While the algorithm scripts were written using the PDW which houses identified data, all research extraction was done using de-identification methods that meet the standards of the UCLA health system. Any unnecessary identifier was removed from the data and any necessary linkages were subjected to a hash function using an random salt that is unique to each patient. Lastly, all dates were modified to be relative values starting at the first encounter date that each patient had in the EMR data.
Results
Data were analyzed from January 1, 2019, to December 31, 2019. In total, 34,982 hospitalizations met inclusion criteria for having DRG assignment data with 34,104 (97.5%) of the admissions having complete data and a primary DRG designated. 5,388 (15.4%) of the DRG codes were at the base comorbidity level, 13,227 (37.8%) had a CC modifier, 12,591 (36%) had a MCC modifier, and 3,776 (10.8%) had no applicable modifier and were considered “Singlet.” These data are summarized in a supplemental table.
Application of algorithms
The results of how many admissions (individual hospitalization encounters) were flagged by each disease algorithm and the count of applicable ICD codes can be found in Table 2. In total, 11,520 (32.9%) hospital admissions were flagged by our algorithms as having a disease state with no corresponding ICD code in the EHR. Among these flagged cases, some of the most notable were Acidemia with 63.2% of cases missing the appropriate ICD code and Delirium, which had an even higher omission rate of 80.3%. Hyponatremia also had a significant percentage of omissions, with 62.3% of cases missing ICD codes.
In cases where the originally assigned DRG comorbidity modifier level was below the algorithm-derived comorbidity level, the hospital admission was flagged as not having adequate ICD codes specifically assigned to the index hospitalization. The results of the DRG upgrades per disease state are also demonstrated in Table 2. For instance, Pancreatitis had the highest percentage of cases flagged for potential DRG upgrade, with 30.7% of cases identified as needing a change in the comorbidity modifier. The GCS ≤ 8 category was also notable, with 18.3% of admissions requiring a DRG upgrade.
The DRG upgrades per disease state, also demonstrated in Table 2, showed that 1,127 (3.2%) admissions were flagged for upgrade from the base level to CC level, 185 (0.53%) admissions were flagged for upgrade from the base level to MCC level, and 678 (1.9%) admissions were flagged for upgrade from CC to MCC level. The breakdown of how many upgrades were from base to MCC level vs. CC to MCC level is demonstrated in Fig. 1. Hyponatremia was flagged for the highest number of potential upgrades, with 431 cases (4.4%) needing a higher DRG modifier. Lastly, Table 2 demonstrates the specificity of the algorithms by indicating the number of occurrences where the ICD codes were available, but the disease states did not meet the algorithms’ criteria for inclusion. For example, Acidemia had 2,055 cases (45.6%) where ICD codes were present but did not meet the algorithm’s criteria, showing the importance of this approach in highlighting potential misclassifications.
Distribution of DRG Upgrade Eligibility by Algorithm Type Fig. 1 illustrates the incidence of admissions with DRG modifiers that are at a lower acuity than the disease states associated with the applicable algorithms. Only cases that can justify upgrading to Comorbid Condition (CC) or Major Comorbid Condition (MCC) are shown
Accounting for potential revenue optimization
Accounting for the difference in RWF points between the DRG codes originally assigned to admission and the DRG codes assigned by the algorithm, we were able to create the delta RWF values. Multiplying the delta RWF values by the payor-dependent dollar estimated per point, we calculated the total lost revenue by admission due to miscoding. In total, we found that the value of miscoding DRGs amounted to a loss of $22,680,584.50. $4,971,278 was attributed to upgrades from the base to CC comorbidity levels, $2,480,785 was attributed to upgrades from the base to MCC comorbidity levels, and $15,228,522 was attributed to upgrades from the CC to MCC comorbidity levels. A financial summary is available in Table 3 where the data is also broken down by payor.
Demographic association
Demographic information for the hospitalizations flagged by the algorithms for not having a proper ICD code applied to the chart is summarized in Table 4. When using the standardized mean difference (SMD) calculated for each table row, as described in the statistical method section, they all fall into the small association category (SMD < = 0.4). This indicates that there is minimal association and little clinical relevance. The data in the table did not restrict the number of ICD codes for payor maximums and used all the codes entered in the EMR for the admission.
Alternatively, the p-value generated by comparison of each patient group to the reference groups (Caucasians non-Hispanic, private payors, and middle aged) all demonstrated significant p-value < 0.001 with the exception of Asian patients, patients > 79 years old, and patients with ethnicity listed as “other – non Hispanic.”
ICD limits
In some cases, payors will limit the number of ICD codes that the hospital is able to submit. This requires billers to choose the most optimal ICD codes to maximize revenue while also ensuring accurate account of the hospitalization. In our data, 6,493 Medicare cases met the maximal 25 ICD codes limit. Of those, 156 where eligible for un upgrade for one of the three upgrade types. It’s unclear which of the ICD were ultimately submitted as the filtration can occurs downstream of our system.
Discussion
In this manuscript, we successfully used rules-based algorithms and raw structured EHR data to identify ICD codes that were not coded in patients’ medical record claims for completed hospitalizations. These missing ICD codes often had downstream effects such as incorrect DRG modifiers and ultimately incorrect reimbursement. Overall, 34% of hospital admissions were deemed as missing an ICD code and 5.8% of admissions had miscoded DRG modifiers. 0.6% of the miscoded DRG codes were eligible for MCC modifiers despite not even having a CC modifier. This resulted in an estimated $22,680,584.50 USD of unbilled revenue over the course of a single year for this study. These results indicate there may be value in using these kinds of algorithms to identify the 19 disease states that were the focus to improve medical coding and enhance ICD/DRG accuracy.
In this study, our goal was not to fully assess the accuracy or inaccuracy of all ICD codes, but rather to test the hypothesis that rules-based algorithms can be used to identify hospital admissions that might be eligible for upward ICD adjustments and improved billing and / or payment. For an ICD code to be eligible for billing, the disease must not only be present but also addressed by a treating provider. The algorithms here only identify that the clinical criteria were present, not that the necessary chart documentation of management and / or treatment was also present. For example, a patient that had hyponatremia by laboratory results, but no documentation by the provider, was identified by our algorithms as having hyponatremia but is not eligible for increased billing due to a lack of documentation/treatment. This emphasizes that there is still a need for human review followed by clarification of documentation queries to treating providers with the hope to use augmented intelligence to aid in the process. Thus, this manuscript sheds light on discrepancies between ICD codes and the true clinical picture, but the study was not designed to inform the accuracy of the ICD coding as it relates to the necessary clinical documentation.
Considering the differences between clinical presence of a disease and the regulatory requirements for coding, we believe that this study is most reflective of the fact that modern EHRs increasingly seem to have large volumes of relevant data but are often hard to navigate by practicing providers (and administrative staff such as billers & coders). A single hospital admission may generate a multitude of laboratory and radiographic results, and dozens of provider notes. It is not possible for a human to review this magnitude of data on their own. We believe that automated computer-based algorithms, such as those used here, can be a potential part of the solution.
At the outset of our study, we hypothesized that the missing diagnoses might be based on patient demographics. Issues of systemic bias in healthcare are now well documented [31,32,33]. It is certainly plausible that these systemic issues resulted in providers being more likely to miss key conditions in some patients as opposed to others. Looking at the standard mean difference (SMD), all the variation was in the minimal difference range. However, when comparing the frequencies of groups with missing ICD coding against reference groups (private payors, Caucasians, middle-aged patients) we were able to derive statistically significant p-values. Similarly significant p-value where demonstrated between patients of opposite sexes. The only comparisons that were not significant included Caucasian vs. Asian patients, Caucasian vs. “other non-Hispanic” patients, and middle-aged vs. elder patients. Given the discrepancy between the significant p-values using reference groups and a small deference by SMD, this raises the concern of having statistical significance without clinical significance and far more research is necessary to understand these implications.
Inaccurate or incomplete ICD coding have a significant impact on a health system beyond direct financial reimbursement. Given their widespread use, and assumptions of accuracy, hospital leadership, regional groups, payers and even the federal government routinely use ICD codes for risk adjustment purposes and researchers often use ICD codes in retrospective studies examining healthcare utilization, patient outcomes, and similar metrics [34,35,36]. As healthcare transitions towards value-based care paradigms these secondary uses are likely to take on increased importance.
There are limitations to this study. First, this is a single center retrospective analysis carried out at a large academic medical center. The results here may not be directly applicable to all other hospitals and certainly might be different in centers with different levels of acuity. It is also important to note that many assumptions were made to provide a clear analysis – these assumptions may not always be perfect for a given patient. Thus, the exact numbers cited in this manuscript may not be 100% accurate; however, we believe the overall conclusions are correct. There are also exceptions where certain ICD codes do not qualify for modification of a DRG like in the example of morbid obesity as it pertains to a DRG for bariatric surgery since the obesity is implied – these exclusions were not accounted for in our analysis and may affect the final rates that were reported. Nonetheless, given the magnitude of missing ICD codes and potential for changes in DRGs we believe that the results certainly demonstrate room for improvement. Secondly, the financial analysis that was used to estimate the lost annual revenue from the selected disease algorithms was only an estimate. Reimbursement for the relative weight factors (RWF) were estimated based on insurance categorization. To derive a more accurate number, work would need to be done in conjunction with billing data to arrive at a more specific insurer/policy reimbursement rate. Also, over the course of the study, factors such as inflation and other reimbursement changes occur naturally over time which were not accounted for in our financial modeling. Lastly, despite the algorithms indicating existence of a disease, it’s possible that the disease did not receive any treatment during the given hospitalization and thus does not warrant billing consideration. This suggests that algorithms have the potential to “up code” for disease states that do not justify ICD coding. We believe that the algorithms are specific enough to indicate active disease states that are being addressed.
In trying to create a more accurate picture of ICD / DRG coding and actual patient disease burden, these algorithms fail in identifying patients that are assigned codes that are at a higher acutely then appropriate. While technically possible, the algorithms were designed to identify diseases in a “rule-in” approach rather than “rule-out.” This is something we plan to address in future works.
Despite these limitations, we believe this manuscript demonstrates that algorithms based on raw EHR data can be used to identify disease states. Similarly, this manuscript adds to the existing body of work demonstrating issues with relying on ICD codes as a complete picture of a given patient’s clinical risk. The implementation of systems, like these algorithms, can likely have benefit for clinician staff in the form of clinical decision support and for the coding teams as admission summaries. Embedding augmented intelligence into this problematic workflow has the potential for improvements in administrative data, but more importantly, improvements in administrative data accuracy and financial outcomes.
Conclusion
We successfully used rules-based algorithms and raw structured EHR data to identify omitted ICD codes from inpatient medical record claims. Using a short list of eighteen diseases, 34% of the algorithm positive hospital admissions were identified as having an ICD code omission and 5.8% of the admissions had miscoded DRG modifiers. Using these automatable algorithms can help augment the coding process to improve the accuracy of ICD coding and potentially minimize findings did not indicate significant disparities.
Data availability
Given that the data used for this study is property of UCLA, it is not available for sharing publicly. Furthermore, the code is written based on a proprietary system and cannot be shared readily in the current form without possible copyright issues from the EHR vendor.
Abbreviations
- ICD:
-
International Classification of Diseases
- DRG:
-
Diagnosis Related Groups
- CMS:
-
The Center for Medicare and Medicaid Services
- HER:
-
Electronic Health Records
- PDW:
-
UCLA’s Perioperative Data Warehouse
- MCC:
-
Major Comorbid Condition
- CC:
-
Comorbid Condition
- RWF:
-
Relative Weighting Factors
- SMD:
-
Standardized Mean Difference
References
O’Malley KJ, Cook KF, Price MD, Wildes KR, Hurdle JF, Ashton CM. Measuring diagnoses: ICD code accuracy. Health Serv Res. 2005;40:1620–39.
Zafirah SA, Nur AM, Puteh SEW, Aljunid SM. Potential loss of revenue due to errors in clinical coding during the implementation of the Malaysia diagnosis related group (MY-DRG®) Casemix system in a teaching hospital in Malaysia. BMC Health Serv Res 2018;18.
PROVISIONS AFFECTING THE FINANCING OF THE SOCIAL SECURITY SYSTEM. 42 U.S.C. § 1305. 1983. https://www.govinfo.gov/content/pkg/STATUTE-97/pdf/STATUTE-97-Pg65.pdf
Falen TJ. Learning to Code With ICD-9-CM 2011. In: LWW, 2010:41–68. http://books.google.lu/books?id=7lGOQgAACAAJ
Calore KA, Iezzoni L. Disease staging and PMCs: can they improve DRGs? Med Care. 1987;25:724–37.
Lee D, de Keizer N, Lau F, Cornet R. Literature review of SNOMED CT use. J Am Med Inf Assoc. 2014;21:e11–9.
Bowman S, Coordinating SNOMED-CT. ICD-10: getting the most out of electronic health record systems. J Am Heal Inf Manag Assoc. 2005;76:60–1.
Mitchell JB, Bubolz T, Paul JE, Pashos CL, Escarce JJ, Muhlbaier LH, Wiesman JM, Young WW, Eptein RS, Javitt JC. Using medicare claims for outcomes research. Med Care. 1994;32:JS38–51.
Iezzoni LI. Assessing quality using administrative data. Ann Intern Med. 1997;127:666–73.
Pendergrass SA, Crawford DC. Using Electronic Health Records To Generate Phenotypes for Research. Curr Protoc Hum Genet 2019;100.
Fonteyn M. The Agency for Health Care Policy and Research guidelines: implications for home health care providers. AACN Clin Issues 1998;9.
Services C, for M& M. FY 2023 IPPS Final Rule. 2023. http://www.cms.gov/Medicare/Medicare-Fee-for-Shttps://www.cms.gov/medicare/acute-inpatient-pps/fy-2023-ipps-proposed-rule-home-page
Stausberg J, Lehmann N, Kaczmarek D, Stein M. Reliability of diagnoses coding with ICD-10. Int J Med Inf. 2008;77:50–7.
Hsia DC, Krushat WM, Fagan AB, Tebbutt JA, Kusserow RP. Accuracy of Diagnostic Coding for Medicare Patients under the prospective-payment system. N Engl J Med. 1988;318:352–5.
Campbell PG, Malone J, Yadla S, Chitale R, Nasser R, Maltenfort MG, Vaccaro A, Ratliff JK. Comparison of ICD-9-based, retrospective, and prospective assessments of perioperative complications: Assessment of accuracy in reporting - clinical article. J Neurosurg Spine. 2011;14:16–22.
SERVICES USDOHAH. The Supplementary Appendices for the Medicare 2011 Improper Payment Rate Report. 2011:1–55.
Nashed A, Zhang S, Chiang CW, Zitu M, Otterson GA, Presley CJ, Kendra K, Patel SH, Johns A, Li M, Grogan M, Lopez G, Owen DH, Li L. Comparative assessment of manual chart review and ICD claims data in evaluating immunotherapy-related adverse events. Cancer Immunol Immunother. 2021;70:2761–9.
Schaefer JW, Riley JM, Li M, Cheney-Peters DR, Venkataraman CM, Li CJ, Smaltz CM, Bradley CG, Lee CY, Fitzpatrick DM, Ney DB, Zaret DS, Chalikonda DM, Mairose JD, Chauhan K, Szot MV, Jones RB, Bashir-Hamidu R, Mitsuhashi S, Kubey AA. Comparing reliability of ICD-10-based COVID-19 comorbidity data to manual chart review, a retrospective cross-sectional study. J Med Virol. 2022;94:1550–7.
Hofer IS, Cheng D, Grogan T. A retrospective analysis demonstrates that a failure to Document Key Comorbid diseases in the anesthesia preoperative evaluation associates with increased length of Stay and Mortality. Anesth Analg; 2021.
Wei W-Q, Denny JC. Extracting research-quality phenotypes from electronic health records to support precision medicine. Genome Med. 2015;7:1–14. http://genomemedicine.com/content/7/1/41
Hofer IS, Gabel E, Pfeffer M, Mahbouba M, Mahajan A. A systematic Approach to Creation of a Perioperative Data Warehouse. Anesth Analg. 2016;122:1880–4.
Gabel E, Hofer IS, Satou N, Grogan T, Shemin R, Mahajan A, Cannesson M. Creation and validation of an automated algorithm to Determine Postoperative Ventilator requirements after cardiac surgery. Anesth Analg. 2017;124:1423–30.
Childers CP, Hofer IS, Cheng DS, Maggard-Gibbons M. Evaluating surgeons on Intraoperative Disposable Supply costs: details Matter. J Gastrointest Surg; 2018.
Gabel E, Shin J, Hofer I, Grogan T, Ziv K, Hong J, Dhillon A, Moore J, Mahajan A, Cannesson M. Digital Quality Improvement Approach reduces the need for rescue antiemetics in high-risk patients. Anesth Analg 2018:1.
Epstein RH, Hofer IS, Salari V, Gabel E. Successful implementation of a Perioperative Data Warehouse using another hospital’s published specification from Epic’s Electronic Health Record System. Anesth Analg. 2021;132:465–74.
Mišić V V, Gabel E, Hofer I, Rajaram K, Mahajan A. Machine Learning Prediction of Postoperative Emergency Department Hospital Readmission. Anesthesiology. 2020;132:968–80.
Mišić VV, Rajaram K, Gabel E. A simulation-based evaluation of machine learning models for clinical decision support: application and analysis using hospital readmission. NPJ Digit Med. 2021;4:98. http://www.ncbi.nlm.nih.gov/pubmed/34127786
Ershoff BD, Grogan T, Hong JC, Chia PA, Gabel E, Cannesson M. Hydromorphone Unit Dose affects intraoperative dosing: an observational study. Anesthesiology 2020:981–91.
Hofer IS, Cheng D, Grogan T, Fujimoto Y, Yamada T, Beck L, Cannesson M, Mahajan A. Automated Assessment of existing patient’s revised Cardiac Risk Index using Algorithmic Software. Anesth Analg 2018:1.
Services C, for M& M. Hospital Inpatient Prospective Payment System (IPPS). Frequently Asked Questions (FAQs) – Market-Based MS-DRG Relative Weight Data Collection and Change in Methodology for Calculating MS-DRG Relative Weights. https://www.cms.gov/files/document/frequently-asked-questions-faqs-market-based-ms-drg-relative-weight-data-collection-and-change.pdf
Cohen J. Statistical Power Analysis for the behavioral sciences. Routledge; 2013.
Fitzgerald C, Hurst S. Implicit bias in healthcare professionals: a systematic review. BMC Med Ethics 2017;18.
Marcelin JR, Siraj DS, Victor R, Kotadia S, Maldonado YA. The impact of unconscious Bias in Healthcare: how to recognize and mitigate it. J Infect Dis. 2019;220:S62–73.
Truong HP, Luke AA, Hammond G, Wadhera RK, Reidhead M, Joynt Maddox KE. Utilization of Social Determinants of Health ICD-10 Z-Codes among hospitalized patients in the United States, 2016–2017. Med Care. 2020;58:1037–43.
Wei MY, Luster JE, Chan CL, Min L. Comprehensive review of ICD-9 code accuracies to measure multimorbidity in administrative data. BMC Health Serv Res 2020;20.
Sharabiani MTA, Aylin P, Bottle A. Systematic review of comorbidity indices for administrative data. Med Care. 2012;50:1109–18.
Acknowledgements
Dr. David Battleman, MD for offering perspective an ideas.
Funding
This work was an original manuscript with no compensation given to the authors.
Author information
Authors and Affiliations
Contributions
EG: Created and coding the disease algorithms. Developed research idea and wrote/edited manuscript. JG: Developed research idea and wrote/edited manuscript. TG: Developed statistical models. IH: Created and coding the disease algorithms. Developed research idea and wrote/edited manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
This study conformed to the ethical guidelines of the 1975 Declarations of Helsinki and was reviewed by the UCLA Human Research Protection Program, Medical IRB 1 Subgroup (https://ohrpp.research.ucla.edu/). Waiver of Informed Consent: The UCLA IRB waived the requirement for informed consent or ethics approval under 45 CFR 46.116(d) for the entire study. The UCLA IRB waived the requirement for HIPAA Research Authorization for the research.
Consent for publication
Not applicable.
Competing interests
Drs. Gabel and Hofer are co-founders of Extrico Health Inc., a healthcare analytics company. Dr. Gal and Mr. Grogan have no competing interests.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Electronic supplementary material
Below is the link to the electronic supplementary material.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Gabel, E., Gal, J., Grogan, T. et al. A retrospective analysis using comorbidity detecting algorithmic software to determine the incidence of International Classification of Diseases (ICD) code omissions and appropriateness of Diagnosis-Related Group (DRG) code modifiers. BMC Med Inform Decis Mak 24, 309 (2024). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12911-024-02724-8
Received:
Accepted:
Published:
DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12911-024-02724-8