Loading [MathJax]/extensions/MathEvents.js
Clinical Errors From Acronym Use in Electronic Health Record: A Review of NLP-Based Disambiguation Techniques | IEEE Journals & Magazine | IEEE Xplore

Clinical Errors From Acronym Use in Electronic Health Record: A Review of NLP-Based Disambiguation Techniques


Illustration of the different components of E-Health Infrastructure that comprises of various healthcare stakeholders, including physicians involved in clinical interacti...

Abstract:

The adoption of Electronic Health Record (EHR) and other e-health infrastructures over the years has been characterized by an increase in medical errors. This is primaril...Show More

Abstract:

The adoption of Electronic Health Record (EHR) and other e-health infrastructures over the years has been characterized by an increase in medical errors. This is primarily a result of the widespread usage of medical acronyms and abbreviations with multiple possible senses (i.e., ambiguous acronyms). The advent of Artificial Intelligence (AI) technology, specifically Natural Language Processing (NLP), has presented a promising avenue for tackling the intricate issue of automatic sense resolution of acronyms. Notably, the application of Machine Learning (ML) techniques has proven to be highly effective in the development of systems aimed at this objective, garnering significant attention and interest within the research and industry domains in recent years. The significance of automating the resolution of medical acronym senses cannot be overstated, especially in the context of modern healthcare delivery with the widespread use of EHR. However, it is disheartening to note that comprehensive studies examining the global adoption of EHR, assessing the impact of acronym usage on medical errors within EHR systems, and reporting on the latest trends and advancements in ML-based NLP solutions for disambiguating medical acronyms remain severely limited. In this current study, we present a detailed overview on medical error, its origins, unintended effects, and EHR-related errors as a subclass of clinical error. Furthermore, this paper investigates the adoption of EHR systems in developed and developing nations, as well as the review concludes with an examination of various artificial intelligence techniques, particularly machine learning algorithms for medical acronym and abbreviation disambiguation in EHRs.
Illustration of the different components of E-Health Infrastructure that comprises of various healthcare stakeholders, including physicians involved in clinical interacti...
Published in: IEEE Access ( Volume: 11)
Page(s): 59297 - 59316
Date of Publication: 09 June 2023
Electronic ISSN: 2169-3536

Funding Agency:


CCBY - IEEE is not the copyright holder of this material. Please follow the instructions via https://creativecommons.org/licenses/by/4.0/ to obtain full-text articles and stipulations in the API documentation.
SECTION I.

Introduction

Acronyms and other types of word abbreviation are compelling in usage and gaining traction in the clinical domain. One of the motivating factors behind the widespread usage of acronyms and abbreviations among physicians is the rapid adoption of the Electronic Health Record (EHR) systems in clinical settings. This electronic version of patient health history often contains all important administrative clinical data relevant to a patient’s care and is maintained by healthcare providers over time. EHR provides a digital format for a patient’s paper medical chart. Unlike traditional paper records, EHRs offer real-time, patient-centric records that can be securely accessed by authorized users. In actual fact, most health reports are now written at the time of service by doctors who type, dictate (using speech recognition software), enter notes into a semi-structured or templated document entry system, or utilize a combination of these approaches [1]. However, these shorter word forms frequently have multiple various meanings (i.e., ambiguous), and thereby making automated or semi-automated extraction from notes difficult. Moreover, they also pose several communication challenges, which are capable of jeopardizing the patient’s safety [2], [3], [4].

It is of utmost importance to accurately decipher and provide the intended sense or full form of acronyms and other abbreviated word forms commonly found in clinical reports. This task is just as vital as comprehending the overall content of the document. With the proliferation of acronyms and their multiple meanings, as highlighted in numerous studies [2], [4], [5], [6], [7], the issue of acronym ambiguity is widely acknowledged as a significant challenge that impacts effective communication in the medical field. Traditionally, making sense of acronyms and other confusing language has been a matter of human perception and interpretation. However, the increasing usage of clinical acronyms due to extensive adoption of EHR systems presented a new challenge for researchers: developing and improving an automated sense resolution technique for medical acronyms and abbreviations in clinical notes. While it is much easier for human (i.e., expert) readers to derive the intended meaning of an acronym or abbreviation based on the context of usage in a sentence, it remains very challenging to automate the process, due to several challenging factors.

The development and implementation of automatic algorithms for medical acronyms and abbreviations sense resolution in clinical notes such as EHRs is an important topic in medical NLP, and it is regarded as a subset of Word Sense Disambiguation (WSD). The task of automating the resolution of intended sense of acronym and abbreviation in clinical note primarily involves teaching a computer program to accurately provide the intended complete meaning of the acronym based on the context of its usage in the document. To assist knowledge-based applications like error detection, decision support, as well as surveillance, a variety of Natural Language Processing (NLP) algorithms have been presented to obtain patient health information from narrative health records.

However, existing automated medical NLP approaches still suffer several challenges, which could be attributed to a paucity of complete inventories of acronyms and their various senses (sense inventory), as well as the dearth of effective methods for disambiguating abbreviations with numerous senses [8]. With the advancements of modern machine learning (ML) technologies, several domain-specific medical problems, such as disease detection [9], [10], mortality prediction [11], [12], [13], [14], and patient monitoring [14], [15], [16] have been successfully addressed. As a result, numerous authors have been encouraged to explore the enormous potential of various ML algorithms for the automation of clinical text extraction and medical acronym disambiguation. While several authors have contributed various comprehensive reviews of existing research studies in the field of medical NLP over the years (as evidenced in Table 1), there is scarcity of studies that provide a comprehensive and structured overview of medical errors stemming from acronym use in EHR. Specifically, studies that encompass the extent of EHR adoption, the challenges associated with it, and provide a critical assessment of current trends and advances in machine learning-based NLP solutions for automated sense resolution of acronyms and abbreviations in clinical notes.

TABLE 1 Existing Survey Articles on Clinical Acronyms Disambiguation
Table 1- 
Existing Survey Articles on Clinical Acronyms Disambiguation

In this study, we offer a comprehensive overview of medical errors, including their unintended consequences, with a specific focus on EHR-related errors as a distinct subclass. Additionally, we delve into the adoption of EHR systems in both developed and developing countries, examining implementation challenges and issues in each region in detail. In addition, we have performed a literature review focusing on NLP algorithms that can automatically disambiguate clinical acronyms and abbreviations within EHRs. Our review encompasses articles published between 2012 and 2023. To sum up, we highlight recommendations and future research directions. The unique contributions of this study can be summarized in fourfold:

  1. We contribute a comprehensive and structured overview of medical errors, their origins, unintended effects, and EHR-related errors as a subclass of clinical error.

  2. We investigate the adoption and implementation profundity of EHR systems in developed and developing nations, as well as identify various implementation and usages challenges.

  3. This article examines various machine learning algorithms for medical acronyms and abbreviation disambiguation in EHRs.

  4. Various issues and challenges for the development of accurate NLP tools for automatic acronyms disambiguation were discussed, future research trajectories were also identified.

The remainder of this paper is organized into sections. Section II presents a detailed overview of errors in medical practice. Section III describes EHR, its adoption in developing countries, and the issues and challenges encountered. In Section IV, we review previous work applying AI for acronyms disambiguation in clinical notes. In Sections V and VI we discussed our findings and concluded the paper.

SECTION II.

Overview of Medical Error

Medical error is a serious issue in public health because it is difficult and time-consuming to identify all consistent causes of errors. It is widely regarded as one of the most critical threats to patients’ health [24]. According to the Institute of Medicine’s report on healthcare safety, To Err is Human, approximately 98,000 Americans die each year from avoidable medical errors [25], [26]. Correspondingly, a 2006 report by the Institute of Medicine of the National Academies further revealed that medication errors affect at least 1.5 million individuals annually, making them one of the most frequent types of medical errors. Report states, the cost of treating drug-related injuries in hospitals alone amounts to \$ 3.5 billion per year, exclusive of lost wages and productivity or additional healthcare costs [27].

In its executive summary released in the year 2000, the National Academy of Medicine, which was then called the Institute of Medicine, defined error in clinical and medical practice as “the failure of the planned action to be completed as intended (omission) or the use of the wrong plan to achieve an aim (commission)”. Despite efforts to reduce medical errors, they remain the primary cause of harmful patient incidents and a significant financial liability for healthcare systems worldwide. Patient injury resulting from unsafe medical care is globally recognized as a leading cause of morbidity and mortality [28], [29], [30]. The World Health Organization defines patient injury as “an incident that results in harm to a patient, such as impairment of structure or function of the body and/or any adverse effect arising therefrom or associated with plans or actions taken during the provision of healthcare” [31].

Patient injuries caused by medical errors can manifest at both the individual and system levels [32], [33]. These injuries may result from an unintended act (omission or commission) or an action that does not achieve its intended outcome [34], [35]. They can also be caused by the failure of a planned action to be executed as intended (an error of execution), the utilization of an incorrect plan to achieve an aim (an error of planning), or an aberration from the process of care [32].

Over the years, the error taxonomy in the clinical domain has continued to evolve, becoming more comprehensive and precise in categorizing factors and events that can be prevented. This evolution has been driven by the need to account for variations in measurement techniques, variables, and detection methods, and study populations. Moreover, the absence of an internationally standardized taxonomy that describes what constitutes an error, potential error, or error cause is also a reason for the growing taxonomy [36]. In Figure 1, we present the taxonomy of error in medical practice proposed by [37]. In this figure, two major error causes are identified, and medical errors at the individual level are further classified into four main categories: medication error, diagnostic errors, treatment errors, and clinical errors. Each of these categories is further subdivided into subcategories that describe the types of errors that can occur. This taxonomy helps healthcare professionals recognize and prevent different types of errors that can occur at different levels of medical healthcare delivery.

FIGURE 1. - Taxonomy of errors in medical practice, which classifies error causes into two main categories: individual errors and system errors. Within the individual errors category, errors related to medication, diagnosis, treatment procedures, and clinical procedures are included. Technical and organizational errors, on the other hand, fall under the system errors category.
FIGURE 1.

Taxonomy of errors in medical practice, which classifies error causes into two main categories: individual errors and system errors. Within the individual errors category, errors related to medication, diagnosis, treatment procedures, and clinical procedures are included. Technical and organizational errors, on the other hand, fall under the system errors category.

A. Common Causes of Error in Medicine

The most common and preventable cause of patient injury is medical error, with potential causes including incorrect medication, incorrect administration route, and incorrect medication time [24], [38]. Some errors are unintentional, such as those caused by misunderstanding or miscommunication [39]. Furthermore, medical errors can be complicated; some are insignificant, while others can take the life of a patient with a long-life expectancy [32]. Thus, it is not surprising when some research findings [32], [40] revealed that, after cancer and heart disease, medical error is the third leading cause of death in the United States. Several other studies on medical errors show that the causes of medical errors include medication errors, diagnoses errors, surgical errors, therapeutic errors, procedural errors, never-happen events, facility accidents, hospital acquired infections, referral errors, error of uncoordinated care, missed warning signs, and untimely discharge from health facility [24], [38], [41]. In Table 2, we provide a comprehensive list of common types of medical error and their potential causes as discussed in previous literature.

TABLE 2 Common Causes of Error in Medical Practices
Table 2- 
Common Causes of Error in Medical Practices

While medical errors continue to be a major hazard to patients’ well-being, their victims far exceed the patient [40]. In the occurrence of errors in healthcare, four groups are impacted in a domino effect: the patient and family (first victim), healthcare personnel (second victim), hospital reputation (third victim), and patients who are affected as a result (fourth victims). Safe, dependable, and patient-focused care is crucial and paramount as the main and ultimate goal of medicine. To reduce the occurrence of medical errors, the use of electronic health records (EHRs) and other e-Health systems have been promoted as possible methods.

B. EHR-Related Medical Errors and Safety Concerns

The adoption of EHR systems has significantly reduced medication and a few communication-related errors, but its performance in reducing diagnostic and technology-related errors has not been particularly encouraging or promising [64]. It, however, produces a new type of communication-related error as healthcare information technology, which may occur during interaction with the system, information retrieval, and decision making. Although the use of clinical acronyms and abbreviations is not new in medicine, it becomes more pervasive with the adoption of EHR. While abbreviations were once limited to prescription writing, they have already become quite popular in many parts of medical reporting. During surgery, in the emergency room, and at the time of discharge, medical abbreviations are utilized across all medical and surgical departments [65].

A common application zone in which medical abbreviations are frequently employed, and cause concern, is when drafting prescription orders. However, the over utilization of abbreviation in a flaw-prone system like EHR, with insufficient backup to detect errors, increases the risk of conceivable mistakes at some stages of prescription order interpretation and/or drugs dispensing by practitioners. Among the most significant implications of EHR-related errors is patient safety. Errors such as incorrect medication dosages, missed diagnoses, or delayed treatment plans can have harmful effects on patients, including adverse reactions, complications, and even death. The complex nature of EHR systems and the potential for human error can contribute to these errors, making it crucial for healthcare providers to remain vigilant and attentive to their use of these systems.

Moreover, EHR-related errors can also have a broader impact on healthcare outcomes. They can lead to unnecessary healthcare costs, wasted resources, and inefficiencies in care delivery. For example, incorrect information in EHRs can lead to duplicate tests, procedures, and prescriptions, increasing healthcare costs and reducing the quality of care. EHR-related errors can also contribute to the fragmentation of care, making it more challenging for providers to collaborate effectively and provide coordinated care.

Finally, EHR-related errors can also affect the trust patients have in their healthcare providers and the healthcare system as a whole. Patients rely on their providers to accurately record and interpret their medical information, and any errors can erode this trust and negatively impact patient satisfaction and engagement. In conclusion, EHR-related errors can have significant implications for patient safety, healthcare outcomes, and the healthcare system. Understanding the information flow and reduced similarity of these errors can aid in the development of effective strategies and tools to prevent and address these errors, enhancing the quality of healthcare and patient outcomes.

C. Unintended Consequences of EHR-Related Medical Errors

Until recently, the usage of abbreviations was still unregulated, and there is no general law dictating which abbreviations can and cannot be used. Concerns relating to patient safety from use of ambiguous medical abbreviation can broadly be classified as adverse events that affected the patient, near misses that did not affect the patient, or unsafe events that bring up possibility of a safety incident [8]. Moreover, several other unintended consequences and medical errors also emerge as a result of EHR adoption-related technological and socio-technical difficulties, such as usability challenges, disruptions in clinical processes, and dangerous workarounds to circumvent technology-related constraints [66], [67], [68]. EHR-related safety issues are difficult to detect and address because they are frequently multidimensional, involving not only potentially harmful EHR technological elements but also EHR user behaviors, organizational characteristics, as well as policies and guidelines to follow for EHR-related activities. To address the challenges of EHR-related patient safety, comprehensive and more recent “socio-technical” methods that account for these components are required [69], [70], [71].

Using an eight-dimension socio-technical conceptual model, [69] analyzed technical and non-technical dimensions of safety in order to identify and evaluate both new and recurring EHR safety issues found in multiple reports. The study’s findings revealed that EHR safety concerns were caused by both unsafe technology and unsafe technology use. Also, in a study that aims to find out what types of EHR-related incidents have happened so far and how they are categorized, the authors in [70] found that the error categories may also entail non-technical issues and may not just concern the technology features of the EHRs.

In a similar vein, according to a national survey conducted by [72] on safety, quality enhancement, and healthcare administration officials, health information technology — more specifically, EHR systems — safety and performance were identified as the most worrying dangers in 2013. The number of deaths caused by mistakes that could have been avoided is still going up, and the recommended information technology (IT) fix has revealed a new mistake. This prompts Rajasekar in [64] to ask, “Is healthcare-IT a knight, a knave, or a pawn?”. Because of how quickly EHRs are being adopted and how ubiquitous acronyms are becoming in such systems, there are concerns about how information will be extracted in the future, which could put patient safety at risk. In summary, medical error remains one of the greatest concerns in clinical settings, as evidently shown in various reports considered in this study as well as their expanding taxonomy over the years.

This error has numerous possible causes and often occurs as a result of omission or commission during healthcare delivery. Among other causes of medical error, miscommunication error is the most frequently occurring one. Since the primary focus of medicine is patients’ rights to safe, trustworthy, and patient-centered care, several measures have been adopted in the practice to reduce or possibly eliminate the occurrence of error in medical practice. Among these measures is the adoption of EHRs and other e-Health infrastructure. While the adoption of EHR makes the use of medical abbreviation more ubiquitous, it however, compounded the long-standing concern about its rampant usage and the associated danger to patients’ safety.

EHR-related errors are subset of the wider category of medical errors, that specifically refer to errors or adverse events that result from the utilization of EHRs, such as errors in data entry, documentation, ordering, and retrieval of patient information. EHR-related errors can occur due to a variety of factors, such as system design flaws, user errors, and workflow issues. While several prior studies suggested addressing this ravaging issue in healthcare as part of a comprehensive patient safety and quality improvement program, recent investigations are also exploring the use of technological innovations such as artificial intelligence as a possible solution. In the section that follows, we present an overview of the EHR, its adoption and implementation level in both the developed and developing countries, issues, challenges and opportunities, as well as EHR-adoption and clinical acronyms use.

SECTION III.

Electronic Health Record

Several healthcare organizations have recently shifted away from traditional paper-based medical report records and toward Electronic Health Record (EHR) systems, in which patients’ longitudinal medical information is stored in an electronic format on a data repository. EHR, like other types of electronic health (e-health) infrastructure, uses information and communication technology to digitize and automate healthcare delivery operations and tasks [3]. This quick transition has led to the availability of enormous amounts of clinical records in EHR systems, much of which contains helpful patient data [8]. In addition to the physical space savings brought about by the transition from paper to digital records, EHR has played a key role in improving healthcare effectiveness, fostering decision-making, and improving management [73]. The medical Decision Support Systems (DSS), which support all sorts of healthcare service providers, including doctors, employees, and management, in making decisions, are connected to the EHR system. It facilitates, among other things, quick and precise judgments on billing, diagnosis, and data analysis. It makes it easier to make prompt, precise judgments about things like diagnosis, laboratory tests, invoicing and payment processing, and analysis of data [74], [75]. In Figure 2, we provide an illustration of an e-health infrastructure for healthcare delivery. Various elements of e-health infrastructure were provided that include physician support system, knowledge base system, health management information system, EHR etc.

FIGURE 2. - Illustration of the different components of E-Health Infrastructure that comprises of various healthcare stakeholders, including physicians involved in clinical interaction and diagnosis, support systems like physiotherapy and clinical therapy, the health management information system, the health logistics system, the electronic health record system, etc.
FIGURE 2.

Illustration of the different components of E-Health Infrastructure that comprises of various healthcare stakeholders, including physicians involved in clinical interaction and diagnosis, support systems like physiotherapy and clinical therapy, the health management information system, the health logistics system, the electronic health record system, etc.

EHR systems provide multiple user access, sharing, and updating from local and remote locations via appropriate user interfaces and network connections, eliminating the need to retrieve and move paper files from a crowded file room. This technology is used to create and store patient data in an electronic manner. The system collects patient data such as complaints, test orders, prescriptions, diagnosis, and procedures. This system allows approved health providers to access data gathered, and data can be analyzed, updated, and electronically annotated even if other authorized health providers are utilizing the same patient data [76]. In the medical industry, several NLP algorithms have been employed to find important information in clinical notes in order to improve patient care and facilitate clinical trials [77].

Despite the obvious benefits of implementing these systems, there are significant challenges to overcome when implementing a truly interoperable EHR between primary and secondary care [78], [79]. These concerns are typically associated with the installation process rather than the product of the EHR vendor [80]. As a result, the implementation strategy is critical, and it should be viewed as a continuous procedure that begins with procurement and continues through all phases of design, implementation, testing, installation, and optimization [81]. In the subsequent subsections, we will discuss the challenges and opportunities for EHR adoption and implementation in both affluent as well as low- and middle-income countries. We will also investigate the causes and recommend strategic plans to avoid problems and challenges associated with EHR adoption.

A. EHR Adoption and Implementation Level

Since the mid-1990s, IT capabilities in healthcare delivery have garnered significant attention and financing, particularly in industrialized countries such as the United States, Australia, and the United Kingdom. The shift from traditional paper-based records to EHRs has been uneven and has not followed overall IT trends in several parts of the world. In certain regions, such as Scandinavia and the United Kingdom, primary health care was the first to implement EHR systems, while university clinics at large hospitals pioneered the development in other parts of the world [82]. Despite numerous studies showing the advantages of electronic health records, adoption is still low in developing countries [74], [83], [84], [85]. In fact, many hospitals in developing nations still record patient information on paper [85], [86]. Table 3 shows a few EHR systems that have been implemented and adopted in developing countries.

TABLE 3 Some EHR Systems Adopted in Developing Countries
Table 3- 
Some EHR Systems Adopted in Developing Countries

Due to the belief that the adoption of EHRs can help to improve the quality of healthcare, the use of EHR systems has increased in Europe, the United States, and other developed countries [92]. Nevertheless, there are still issues in the majority of these regions because of healthcare practices and laws, the diversity of regional languages, the emergence of numerous non-interoperable EHR systems, and a lack of or inconsistent clinical documentation in EHR [93]. EHR system creation remains a difficult task that involves a careful match of local needs to available technologies and resources. There is a lack of experience building EHR systems for the developing world; criteria, goals, and local restrictions are likely to be more different [94].

In [95], the authors pointed out that it is difficult to create scalable and functional EHR systems in developing countries and that online access control is necessary for EHRs. Relying solely on an online access control system is not advised, especially in developing nations, as access to the server may be hampered by a variety of unfavorable circumstances, such as frequent power outages that render the server unavailable. Because of the inability to select an access control method, EHRs are no longer accessible. While other contexts are resource-constrained, it is difficult to develop a single EHR architecture and implementation that will meet the needs of all environments. Adaptability, complexity, cost, incentives and external policy, implementation climate, technological constraints, standards boundaries, attitudinal constraints - individual and organizational behaviors - are other hurdles to EHR deployment () [96], [97].

While the developing world still witnesses a poor uptake of EHRs, recent years have been characterized by numerous studies aimed at identifying the critical factors affecting the adoption of EHR in the healthcare system, particularly in developing countries. A study by [74] revealed that patients are generally more optimistic about using electronic health (e-health) services than doctors in developing countries. Several other studies [98], [99], [100], [101], [102] also indicated that doctors in these countries tend to be reluctant to adopt new technologies. Further noting that, doctors’ negative attitudes toward the EHR system are a result of technology anxiety and inefficient doctor-patient communication. In [103], the authors argue that physicians are also concerned that the EHR system will disrupt workflow and change current work practices.

Moreover, healthcare facilities in low- and middle-income nations, like those in affluent ones, also generate terabytes of multimedia data on a monthly basis in the course of documenting patients’ health status and the care delivery process [104]. However, lack of standard criteria for health information data, technological problems, system interoperability, a lack of readily available, well-trained clinician informatics teams to oversee the process, privacy and confidentiality concerns, a small pool of potential vendors in the market, and vendor churn are other obstacles to the adoption of EHR systems in developing nations [105]. Also, the attitude of the government and key policymakers contribute massively to the low adoption rate in this region.

While the issue of EHR adoption is minimal in the industrialized world, issues such as EHR interoperability with clinical processes persist. In an effort to better understand safety dangers associated with EHRs and potential interoperability issues with other health IT during the care process, [105] examined reports of patient safety events (PSEs) stored in a database comprising different provider organizations. In this study, the authors discovered that, of the 1.735 million PSE reports, 2625 were identified as being connected to health IT, and further research revealed that radiology, laboratory, and pharmacy systems interfaces accounted for the majority of EHR interoperability PSE reports (i.e., medication-related). The majority of interoperability concerns in these clinical domains were connected to obtaining information from other health IT systems rather than giving information to other systems.

B. Issues, Challenges and Recommendations

The main problems with the design and use of EHRs, according to the findings of this study, are the lack of a common and unifying architecture, the lack of concept extraction standards (like standard sense inventories), and the limited robust capacity of analytical engines. There is a big need for a unified architecture that will support standardized interfaces for connecting different analytical engines made by different groups, allowing them to work together both in terms of meaning and process. After the passage of the Health Information Technology for Economic and Clinical Health Act in 2009, the adoption of EHRs in the United States increased rapidly, with approximately 86% of office-based physicians and 96% of non-federal hospitals using an EHR system as of 2017 [106]. Other studies also shown that developed countries such as Canada, the United Kingdom, Australia, and Denmark have made strides in the design and implementation of EHR as part of their national e-health infrastructure, though they have also struggled in their EHR mandate due to user resistance to EHR support for clinical processes, provider burnout, decreased satisfaction [106], [107], [108], [109], and EHR interoperability with other e-health systems [105].

Based on the study’s findings, the following recommendations are offered: Governments and other key policymakers in developing countries could increase EHR system adoption by making legislation and implementing social strategies to encourage physicians to use the EHR system and by providing technical competence and training to make EHR system use easier. Data security and privacy remain critical issues in e-health. In developed countries, like Europe, e-health is based on policies to protect the data of citizens who are patients. So that patients can trust the system, developing countries must also pass laws to protect personal information.

Several studies show that clinicians in developed countries are trained using innovative technologies like 3D simulations, virtual reality, and robotics. Information and communication technology (ICT) is also part of the curriculum for medical courses in these countries. Medical informatics, bioinformatics, computational biology, and health informatics are among the new courses that have begun. The availability of ICT skills among clinicians is likely to result in the acceptance and use of e-health in primary care. This is because clinicians who know how to use ICT can see how it can help them carry out and improve the different processes in which they are involved. So, to have a positive view of electronic medical records, a good knowledge of information communication technology is required.

Numerous concepts have been suggested to address clinician burnout and patient dissatisfaction. Epic Systems Inc. in Verona, Wisconsin, USA, for example, created the Inpatient Provider Efficiency Profile (IP PEP) tool to determine where inpatient EHR users spend the majority of their time. Other regions continue to have a high demand for solutions of this type, which can only be met through impactful research aimed at understanding the root cause and scope of the problem. It is worthwhile to examine the system’s evolution trend when analyzing the prospects of EHR. At the moment, the primary function of EHR systems is to store patient data for use by healthcare providers while maintaining data integrity and confidentiality. To provide a benchmark against which current and future EHR systems can be measured, Navigli in [110] defined significantly improved health care through computer technological competences as the vision of future patient records and patient record systems.

C. EHR and Clinical Abbreviation Use

The longstanding practice of using acronyms and abbreviations in medical documentation is deeply ingrained in medical practice, with even the most junior medical and nursing school graduates being very comfortable with their usage [65]. The number of healthcare professionals who use these medical acronyms, as well as their frequency of use, is unknown, but it is certain that the number is substantial. In any medical document or drug prescription at a healthcare facility, at least one abbreviation can be found on each page of the patient’s medical record. However, the question of whether the usage of medical abbreviations is dangerous and how many patients have been harmed as a result remains a topic for further research.

In an attempt to unveil the extent of inappropriate use of medical abbreviation in medical records and their comprehension issues, Hamiel et al. in [111] utilized as a model, ophthalmology consults in a tertiary hospital. In this study, the authors first mapped out the frequency of general English abbreviation in the department’s EHR. The study included most frequent English abbreviations in the design of their cross-sectional survey to assess the attitude of non-ophthalmologist physicians toward abbreviation use and comprehension. The authors reported that of 437 records screened and 235 responses gotten, only 42.5% got at least 10% of the abbreviations, and none of the respondents was able to get all abbreviation correctly as intended.

As reported in [65], misinterpretation of ambiguous medical abbreviations has resulted in consistent increase in reported errors, some of which have caused adverse events, received by the US Institute of Safe Medication Practices (ISMP). The Joint Commission has provided updated regulations and a brief list of harmful medical abbreviations and dose expressions that should never be used to ensure safe usage of medical abbreviations. However, despite numerous condemnations from prominent organizations, medical acronyms are still heavily utilized, especially with the adoption of EHRs. Medical abbreviations are almost always present in medical records or prescriptions, putting junior healthcare staff in a difficult position to decipher the drug orders without the presence of the healthcare provider who wrote the abbreviations. The other option is to refuse to carry out any order with confusing abbreviations, but this may risk the patient’s health.

Researchers are now having to deal with this lingering problem for the electronic form of medical records created in EHR. Since automated sense resolution of medical acronyms in such systems is core to avoiding improper communication, different medical NLP approaches have been suggested to deal with this problem. However, due to the nature of the task, most of these approaches still have trouble understanding and interpreting ambiguous short word form. Moreover, EHRs make it hard to figure out what an acronym means because it lacks a comprehensive list of acronyms full meaning and their methods for resolving acronyms’ intended full sense remain not particularly good. In summary, medical errors can come from a variety of sources, but since the adoption of EHR has gained momentum, EHR-related errors from clinical acronyms use have been on the rise.

SECTION IV.

NLP in Clinical Narratives

With the help of machine learning (ML) and deep learning (DL) methods, artificial intelligence (AI) in healthcare is better than other technologies at gathering information, processing it, and giving a clear result. When AI is used in medicine, the main goal is to find out how prevention or treatment affects how well a patient does. Even though AI has been used to handle medical data, the implementation of Electronic Health Record (EHR) systems has increased the utilization of AI technologies in healthcare. Due to the wide use of EHRs, the healthcare community now has access to a lot of discharge summaries, laboratory reports, progress notes, and other types of information. These clinical notes, particularly progress notes, include the patient’s most recent pertinent information. Even though the information is particularly important for describing a patient’s medical situation, it is mostly written down as unstructured text using highly specialized clinical language [112] and sometimes, ambiguous acronyms. In Fig. 3, we provide a good example of the use of acronym MR in two clinical notes. In Note 1, the acronym stands for “mental retardation”. Based on the context of this note, it is quite simple to interpret it incorrectly as “mitral regurgitation,” which may prompt additional imaging tests and the postponement or avoidance of non-urgent interventions due to a higher risk of complications. In Note 2, MR is defined as “mitral regurgitation”.

FIGURE 3. - Illustration of ambiguous acronym occurrence in clinical narratives, where the acronym “MR” is used in Note 1 and 2, but with different intended full forms.
FIGURE 3.

Illustration of ambiguous acronym occurrence in clinical narratives, where the acronym “MR” is used in Note 1 and 2, but with different intended full forms.

However, it is widely established that unstructured data contain pertinent, extensive, and nuanced information regarding the symptoms trajectory and treatment processes done by and upon patients [65], [113], [114], making the problem of automatically extracting accurate and precise information from narrative notes worthwhile to tackle [115]. A significant amount of structured content is affected by high-flown manipulations or interpretation errors, such as upcoding and misclassification, but narrative content is primarily used to facilitate practitioners’ recollection and as a means of doctor-doctor communication across different work shifts [65], [113]. A special kind of AI approach called Natural language processing (NLP) or text mining techniques, are the foundation for the methods used to process the unstructured data found in EHRs. These techniques are also instrumental for extracting relevant information from unstructured EHR data, such as clinical narratives, and transforming it into a structured format that can be easily analyzed.

Studies have shown that doctors and nurses often use acronyms and abbreviations in their clinical notes, which can hide important information like diseases or procedures. This section will go over previous research on the disambiguation of acronyms and abbreviations in clinical notes with NLP. We conducted a search primarily on Google Scholar and the Scopus database using combination keywords such as Natural language processing, NLP, electronic health record, EHR, text mining, acronyms, abbreviations, word sense disambiguation, WSD, and clinical note in order to examine the application of NLP in medicine for acronyms and abbreviation disambiguation. Even though some earlier publications on the topic of NLP are used, we focused on articles published in the previous five years because NLP applied in medicine is a relatively new field. We concentrated on publications published in the most prestigious journals dealing with clinical informatics, such as JAMIA, Computer Methods and Programs in Biomedicine, Journal of Medical Internet Research, Journal of Biomedical Informatics, and International Journal of Medical Informatics. Our search also considered papers that appear in the proceedings of prestigious international conferences and workshops like ACL, EMNLP, BioNLP, NAACL etc.

A. Medical Acronyms Sense Disambiguation

Clinical narratives in patient Electronic Health Records (EHRs) are riddled with abbreviations [116]. Disambiguating these abbreviations is a critical first step in developing clinical decision support technologies that rely on narratives [117]. Disambiguation of clinical acronyms and abbreviations is a subset of Word Sense Disambiguation (WSD), which is the process of determining the sense of a word to be activated based on the context of usage [118]. Several studies in general English and other languages have proposed numerous distinct approaches to WSD. Among these approaches are supervised machine learning [119], [120], semi-supervised machine learning [121], unsupervised machine learning [122], [123], and knowledge-based [19]. Because of the inherent linguistic fundamentals, researchers have been able to apply similar methods to biomedical literature and clinical text [17].

For medical acronyms and abbreviations, a variety of WSD methods have been proposed, including traditional supervised machine learning-based approaches with optimized features [2], [124], vector space model-based approaches [125], [126], hyper-dimensional computing-based algorithms [25], semi-supervised approaches [127], and unsupervised approaches grounded on topic-modeling method [107]. While disambiguation approaches based on supervised learning methods have been proven to be more effective compared to those based on un- or semi-supervised and knowledge-based methods, they do have a limitation. They often require upfront human intervention to appropriately label the input and output data. However, labeling huge datasets is cost prohibitive and labor-intensive. Additionally, majority of early NLP techniques to clinical abbreviations disambiguation require training a separate classifier for each ambiguous term, which can be challenging, and particularly impractical [128]. To overcome various challenges in this task, numerous approaches have been suggested including those that leverage statistical power across all instances and scale efficiently with the size of the vocabulary. In this subsection, we summarize the most recent ML-based and neural-based NLP techniques for acronym disambiguation in EHR. We examined earlier works based on the techniques used to carry out a variety of automated sense resolution tasks, including the identification of acronyms, the development of sense inventories, and other significant topics.

1) Machine Learning-Based Methods

Abbreviation disambiguation is a difficult issue for computers but has been tackled by a number of different approaches over the past two decades. Most of these methods [117], [126], [129] are based on supervised algorithms, such as Support Vector Machines (SVMs), Naive Bayes, and Random Forest classifiers which are trained on the number of times different senses appear together in clinical abstracts with automatically tagged medical concepts. During this time, semi-supervised, unsupervised, word-embedding, and knowledge-based algorithms also gained traction. In general, these methods are called “artificial intelligence-based,” and the most successful ones in this group are those that use machine learning.

Machine-learning (ML) techniques concentrate on the use of data and algorithms to imitate how humans learn, and progressively enhance its accuracy and performance. Unlike algorithmic programing, this branch of AI enables computer systems to learn from data and improve its performance on a given task, without such knowledge being expressly programmed in its memory. Using statistical methods, ML algorithms are trained to make classifications or predictions, and to learn from crucial patterns and insight in data. Different from traditional statistical and knowledge-based approaches, where human intervention is required at every stage including when variables of interest are being selected as well as during the creation of model with capability to predict outcome [130], ML-based approaches are fully automated and requires no human intervention. Moreover, the output of statistical models is by and large falsely influenced by the variables selected to be included by the user, and it does not also enable performance optimization. The output of ML-based models, however, is free from all these restrictions.

To disambiguate abbreviations and determine their intended meaning, classical ML-based NLP approaches utilize the local context of the abbreviation with respect to the statement they appear. Over the last couple of years, several ML-based NLP approaches have been explored in various studies as a viable innovation for resolving ambiguous acronyms in medical reports. For instance, [127] in their attempt to show the suitability of semi-supervised approaches as a viable alternative to fully supervised machine learning methods employed the Reverse substitution (RS) method, which automatically generates training data by swapping out expansions for the corresponding abbreviations. In this study, a large hand-annotated medical record with instances of seventy-four (74) ambiguous abbreviations was used to test a number of semi-supervised classification algorithms. The authors pointed out that despite large discrepancies between training and test corpora, classifiers nevertheless manage to attain up to 90% accuracy. However, as shown by the fact that the dispersion of terms in their full and shortened forms are usually not the same, RS results in unbalanced training sets [131].

Using three techniques from their prior research —word sense disambiguation techniques for clinical abbreviations, clustering-based semi-automated methods to generate potential abbreviation senses, and machine learning-based algorithms to recognize abbreviations from a clinical corpus — [8] also created a system for medical acronyms detection and disambiguation (CARD). Using medical text data from the Vanderbilt University Medical Center (VUMC), the suggested methodology was utilized to build two detailed sense inventories for abbreviations in clinic visit notes and discharge summaries. By achieving an F1 score of 0.755 for detecting and resolving sense of all abbreviations in a corpus from the VUMC discharge summaries using the sense inventories generated from discharge summaries (cTAKES) [132], CARD outperformed the existing clinical NLP system (MetaMap) [133], [134].

In [135], the authors applied unsupervised approaches for expanding and disambiguating medical acronyms and abbreviations. To disambiguate multi-sense clinical acronyms, the authors combine statistical machine translation with document-context neural language models in this research. In addition, the study looks into the usage of mismatched training data and self-training. These algorithms are tested using nursing progress notes and achieve 71.6% disambiguation accuracy without the use of any manual annotation. In their research, [136] proposed investigating the impact of word-embedding on acronym disambiguation by offering two word-embedding models based on [137]. The model proposed in this paper was tested using ScienceWISE and MSH datasets. The paper’s experimental results show that word-embedding has a considerable impact on the accuracy of acronym disambiguation without knowledge bases, and the results are also remarkably robust across datasets from diverse domains.

Nakayama et al., in [138] proposed using abbreviation normalization, lemmatization, and stop word removal to extract structured information from nursing notes. The authors created a pipeline that conducted straightforward abbreviation disambiguation, using popular preprocessing methods used in natural language processing, and then used sentiment analysis and dimensionality reduction based on topic modeling to create meaningful features from clinical notes. The study discovered that using abbreviation disambiguation in nursing notes for subsequent topic modeling and sentiment analysis enhanced prediction of in-hospital and 30-day mortality while controlling for comorbidities. In [139], Jaber and Martinez studied four approaches to trained two supervised machining learning methods with pre-trained word embeddings integrated within these pipelines as features. Their training features includes the context information of the target abbreviation and dataset used for training is small and obtained from a university-linked health care center. The two supervised ML algorithms are SVM and Naïve Bayes, with SVM showing the best performance in all the four studied approaches.

While ML algorithms have been shown in multiple studies to be successful in clinical acronym and abbreviation disambiguation, their practical deployment in clinical systems is currently limited. Furthermore, the lack of sufficient training data prevented further development and deployment of algorithms for automatic abbreviation disambiguation. Besides, these methods only utilized the local contextual information contained in the data for earning. However, recently evolving methods are now leveraging the contextualized embeddings obtained from deep learning models (such as BERT and ELMo model variants) to fine-tune abbreviation disambiguation models. In the subsection that follows, we provide a detailed review of some recent works that adopt this approach.

2) Deep Learning-Based Methods

While machine learning is widely used in today’s NLP, classical ML-based methods primarily involve optimizing weights for pre-designed representations and features created by humans. Deep learning (DL) is a subclass of ML whose objective is to explore how computers can leverage data to develop features and representations that are suitable for intricate interpretation tasks. Recent research has demonstrated that DL approaches are at the forefront of many NLP tasks [140], [141]. Despite several research demonstrating the appropriateness of DL-based techniques for biological and clinical data and possible applications, the use of deep learning approaches in medical text simplification has been gradual. There could be various explanations for this. While deep architectures can be significantly more efficient than other traditional algorithms at capturing fine intricacies in data structure, Deep neural networks (DNNs), particularly convolutional neural networks, are extremely complicated machines with hundreds of millions of weights, making them data hungry, and demanding huge, supervised data for training and regularization.

In recent years, several papers on the disambiguation of clinical acronyms have utilized the technique. The authors in [117] for instance, trained a Convolutional Neural Network (CNN) to differentiate between various acronyms senses. In this study, the authors utilized three datasets, the first two of which were 1,001 longitudinal patient records obtained from the Cleveland Clinic in Ohio (USA) and totaling 117,526 clinical notes. The first dataset was produced automatically using reversed substitution, the second dataset was manually annotated from set-aside notes, and the third dataset was produced by a team at the University of Minnesota [142] and made publicly available. It contains 37,500 occurrences of 75 abbreviations, with approximately 500 occurrences per abbreviation. Ultimately, they determined that the CNN model performed the best (with an accuracy of 1–4 points on all three datasets) compared to more conventional methods such as SVM.

Also, in [131], the authors proposed a method to enhance a model’s generalizability via cutting-edge data augmentation strategies. The suggested techniques make use of relevant data from biomedical ontologies, including global context information and related medical concepts, within the medical note. The author trained the model on a public dataset (MIMIC III) and evaluated its performance on automatically generated and manually annotated datasets from various sources (MIMIC III, CASI, i2b2). Together, these techniques improve abbreviation disambiguation accuracy by up to 17% on hand-labeled data without sacrificing performance on a MIMIC III held-out test set.

While differentiating between abbreviation disambiguation and abbreviation expansion, Kim et al., [143] noted that expansion is much easier as compared to disambiguation of abbreviation. In this paper, the authors examined two approaches, namely non-sense-based and sense-based approaches to abbreviation expansion. The former utilizes state-of-the-art language models alongside unstructured information contained in clinical notes, exclusively on a lexical level. On the other hand, the latter considered the two tasks of abbreviation expansion and disambiguation, while integrating sense information alongside the unstructured data for language model training. Particularly, the language models considered in this work are the masked BERT [144] language model and the second is novel adaptation of the XLNet language model [145] called length-agnostic permutation XLNet. The findings of this research support the notion that expanding abbreviations might be less challenging than disambiguating them, since the non-sense-based methods perform better than the sense-based methods.

To improve the accuracy of abbreviation disambiguation in biomedicine, the authors in [146] proposed a disambiguation pipeline based on graph attention neural network. The semantic class, part of speech, as well as the words in context of the ambiguous acronyms are taking as the input to the proposed model, while constructing the graph with the sentence and features of disambiguation as nodes. To dynamically modify the edge weight between two adjacent nodes, the author applied a multi-head graph attention mechanism. In this study, the author considered 28 ambiguous biomedical vocabularies from MSH dataset. However, the approach followed by authors is that of the early work in clinical abbreviation disambiguation where separate classifiers are trained for each ambiguous word.

Jaber and Martínez in [147] explore the deep contextualized representations from large language model like BERT for disambiguating clinical abbreviation. This work investigates a series of fine-turning approaches on clinical abbreviation disambiguation by performing various experiments with different pretrained clinical BERT. The utilization of deep contextualized representations from Bioclinical, BlueBERT, and MS_BERT pretrained models in one-fits-all classifiers led to enhanced accuracy on the University of Minnesota (UMN) dataset. Specifically, the Bioclinical, BlueBERT, and MS_BERT models attained accuracy rates of 98.99%, 98.75%, and 99.13%, respectively.

The authors in reference [7] utilize deep learning models trained on public web data to decipher abbreviations and shorthand in clinical text by substituting abbreviations with their corresponding meanings. The type of models utilized in this study are Text-to-Text Transfer Transformers (T5), specifically, designed as encoder-decoder architectures. The result reported in this work uses the T5 80B variation. Furthermore, the authors experimented with three other variants of T5 models, with all been pretrained on web corpus using the MLM (Masked Language Modeling) loss function. These models include T5 11B, T5 large, with 770 million parameters, and T5 small with 60 million parameters. The study introduces a novel translation model capable of detecting and expanding numerous abbreviations in real clinical notes with exceptional accuracy. The model achieves accuracies ranging from 92.1% to 97.1% on multiple external test datasets, outperforming board-certified physicians with a total accuracy of 97.6% compared to 88.7%.

The authors in [148] suggested a few-shot learning strategy to make the most out of the limited annotated data currently available for clinical abbreviation disambiguation. To acquire better contextualized sentence representations for clinical abbreviation disambiguation, a neural network based on topic-attention was specifically used. Their model is tested on a manually built balanced dataset after being trained on a training set that includes 30 abbreviation words taken from a publicly available dataset of clinical domain acronyms and abbreviations. The author demonstrates that adding subject information to the sentence representation significantly improves effectiveness on small-scale uneven training datasets as compared to many baseline models.

The base model’s average AUC is 0.7189, whereas the topic-attention model (ELMo+Topic) suggested in this study obtains an average AUC of 0.8196. The fact that it is expensive and time-consuming to create hand-labeled medical abbreviation datasets for training and testing DNN models, coupled with the fact that there are only a few Clinical Abbreviation datasets containing training data and labels available, research employing deep learning-based method are limited. Moreover, only limited research attention has been given to the issue of data sparsity over the years. To address this issue, which has been one of the most significant barriers preventing the use of many current NLP methods in clinical settings, [149] present MeDAL, a sizable medical text dataset annotated for abbreviation disambiguation, developed for natural language understanding pre-training in the medical domain. The authors utilized this dataset to pre-train several deep learning-based models of common architectures, including LSTM, LSTM+attention and ELECTRA, while conclusively demonstrating that such pre-training improves performance and computational efficiency when fine-tuning on subsequent medical tasks.

Overall, recent advances in deep learning techniques have been greatly outperforming classical statistical and knowledge-based algorithms, particularly in healthcare application, due to their ability to process and analyse vast amounts of complex medical data more accurately and efficiently. These improvements can be summarised as follows:

  1. Large and Complex Data Handling capability: Deep learning algorithms are capable of processing larger and more complex datasets compared to traditional machine learning methods. This is because deep learning models use multiple layers of artificial neural networks to extract features and learn complex patterns in the data.

  2. Data Integration: Deep learning techniques can be used to integrate data from multiple sources, such as EHRs, genomics, and social determinants of health, to provide a more comprehensive view of patient health. This can help healthcare providers make more informed decisions about patient care.

  3. Automated Feature Engineering: Deep learning methods can automatically extract relevant features from data, eliminating the need for manual feature engineering in traditional machine learning. This is especially useful when dealing with large and complex datasets where manual feature engineering is time-consuming and error prone.

  4. Non-Linear Relationships: Deep learning algorithms can model non-linear relationships between inputs and outputs, which traditional machine learning methods cannot easily capture. This is useful in tasks like image and speech recognition, where the relationships between inputs and outputs can be highly complex.

  5. Transfer Learning: Deep learning models can be trained on large datasets and then fine-tuned on smaller datasets for specific tasks. This transfer learning approach can save time and resources, as it avoids the need to train deep learning models from scratch.

  6. Scalability: Deep learning algorithms can be easily scaled to handle large datasets and compute-intensive tasks by leveraging distributed computing platforms and graphics processing units (GPUs). This makes it possible to process massive amounts of data in a reasonable amount of time.

  7. Enhanced Accuracy: One of the main advantages of deep learning techniques is their ability to improve accuracy. Neural networks can learn from vast amounts of data and recognize complex patterns in the data that may be difficult for classical statistical or knowledge-based algorithms to identify.

In summary, the use of deep learning techniques in healthcare has shown great promise in improving patient outcomes and reducing healthcare costs. As these technologies continue to develop and become more widely adopted, we can expect to see even more significant advancements in healthcare.

SECTION V.

Discussion

Patient safety is a critical healthcare problem due to the consequences of iatrogenic injuries. Medication errors in critical care are frequent, dangerous, and preventable. Although the primary objective for adopting the EHR system was to resolve some problems bedeviling medical healthcare delivery, its adoption has led to an increase in medical errors. The use of acronyms increased as a result of the transition from traditional paper-based records to electronic health records, many of which are notorious for being ambiguous. However, human factor research carried out in non-medical situations suggests that requiring higher vigilance from medical practitioners may not lead to a meaningful improvement in safety. Finding issues and redesigning defective systems appears to be a more effective way to reduce human error.

As a remedy, NLP has been adopted as a viable solution to disambiguate medical acronyms and abbreviations in clinical notes. Models built using sophistical NLP algorithms are usually integrated into an EHR system, and often function both as an auto-completion tool during medical report writing as well as intended full meaning suggester during information retrieval. The model, however, performs these functions while taking cognizant of the context of the acronym’s usage in the report statement. Over the years, several approaches have been proposed. Notably, artificial intelligence-based solutions have been generally recognized as the best solution for seeking a complete form to acronyms in medical note. Existing AI methods applied over the years include knowledge-based methods, statistical approach, and machine learning based techniques. Recent approaches to Natural Language Processing (NLP) based on neural networks and deep learning, on the other hand, have achieved substantial advances, surpassing conventional statistical as well as knowledge-based methods on a number of tasks. A noteworthy limitation observed in most of the existing literature is that datasets of only hundreds or thousands of documents are used to train machine learning models.

Another issue that marred the use of this technology in healthcare is ethical concerns. This is related to the potential for bias in the algorithms used in NLP and deep learning techniques. If the algorithms are not designed and trained appropriately, they may produce biased results that could negatively impact certain patient populations, such as those with specific health conditions or demographic characteristics. This could result in discriminatory treatment or exacerbate existing health disparities. Additionally, there are concerns about the accountability and transparency of NLP and deep learning algorithms. Healthcare providers must be able to explain how these algorithms make decisions and ensure that the decisions are fair and justifiable. Patients and other stakeholders should also be able to understand and question the decisions made by these algorithms.

Furthermore, the use of NLP and deep learning techniques in healthcare could also raise issues related to informed consent. Patients may not fully understand the implications of allowing their data to be used in research and may not be aware of the potential risks and benefits of these techniques. Healthcare providers must ensure that patients are fully informed and provide clear and transparent consent processes. In conclusion, the use of NLP and deep learning techniques in healthcare has significant potential to improve patient care and health outcomes. However, ethical considerations related to patient privacy, bias, accountability, and informed consent must be addressed to ensure that these techniques are used in an ethical and responsible manner. More details on other constraints and suggested recommendations are provided in the subsection that follows.

A. Difficulties and Challenges

  1. Scarcity of labeled training data. Modern machine learning (ML) techniques, particularly deep learning approaches, are highly promising for automating a range of clinical and research frameworks, but they require a significant amount of annotated data for training. Due to the cognitive difficulty of abbreviation disambiguation task and the considerable variation in the quality of the data, annotating EHR data can be costly and difficult [131], [150]. Unfortunately, there are frequently insufficient amounts of useful EHR data, and since neural networks need a lot of text to train on, this is a problem. Additionally, for certain tasks, only certified annotators may be qualified to perform annotations, further adding to the cost and difficulty of the process. Ensuring annotation quality can also be challenging as annotators may have disagreements, especially when dealing with domain-specific long forms or concepts of abbreviations. Furthermore, a significant portion of publicly available annotated data in this field is in English, which presents additional challenges for research in other languages. The sparse and imbalanced nature of datasets in this particular domain has several consequences. One of the main consequences is that it restricts the development and deployment of methods for automated abbreviation disambiguation. Due to the limited availability of data, methods built solely on these datasets are susceptible to overfitting and may not be applicable to abbreviations that are not present in the training data.

  2. Imbalance data. A common challenge faced in real-world datasets, such as clinical notes, is the presence of imbalanced class distributions. This means that some classes are more prevalent than others, resulting in insufficient training input for traditional classifiers. The uneven distribution of data presents challenges in creating precise automated abbreviation disambiguation systems. It becomes particularly difficult to train a model that can accurately predict the full form of an abbreviation when there is limited data available for learning. This issue arises from the imbalance in class distribution within publicly available datasets, which can result in overfitting and render existing methods ineffective for abbreviations not included in the training data.

  3. Privacy. Because of the highly sensitive nature of the information contained in EHRs, as well as the existence of regulatory laws such as the (US) Health Insurance Portability and Accountability Act (HIPAA), maintaining privacy within analytic pipelines is critical [151], [152]. As a result, additional privacy-protecting steps are frequently required before any downstream tasks or data can be shared with others. It is an expensive process to remove personal information from a large corpus of EHRs [150]. It is difficult to automate and requires domain experts as annotators.

  4. Explainability and interpretability. The issue of deep learning explainability and interpretability is not peculiar to application in healthcare alone, it is a common problem that is still being instigated in general deep learning research. When compared to other techniques, deep neural networks can produce better results. However, they are frequently regarded as “black boxes” in many disciplines [153], [154]. A neural network model typically has a large number of trainable parameters, which makes it extremely difficult to interpret the model. Moreover, neural networks have complex architectures and non-linear layers making them more difficult to interpret than linear models, which are typically simpler and easier to understand. To increase model transparency, a few efforts have recently been made to create explainable deep neural networks [155], [156], [157], [158].

B. Future Research Direction

The fundamental cause of poor data use for training of supervised ML/DL models for disambiguation is annotation difficulties observed by supervised machine learning algorithms. However, the use of more contemporary deep learning methods has been hampered by the necessity to generate a large, labeled corpus. Among the potential prospects for this area of research is the use of generative models to generate medical acronyms and corresponding possible senses that may be used to train a machine learning model. Also, while methods such as pattern matching, language modeling, and machine learning have produced promising results in the disambiguation of acronyms, there is need to explore the potentials of deep learning-based methods, especially those based on the most recent powerful natural language models such as GPT-3, BERT, diffusion models etc.

In addition, to encourage the development of transparent DL models for healthcare applications, future research should focus on the development and training of explainable and interpretable deep learning models. Such models should scale beyond the current models with only static knowledge distillation, while incorporating dynamism. Also unsupervised learning is a promising research direction in deep learning, as it can help to overcome some of the limitations of supervised learning, such as the need for labeled data, and can enable more autonomous and adaptive learning systems.

SECTION VI.

Conclusion

Electronic health records (EHRs), which are computerized compilations of patient healthcare events and observations, are widely utilized in the medical field and are indispensable for healthcare delivery, operations, and research. Clinical narratives in patient Electronic Health Records (EHRs) are notoriously difficult to automate due to the prevalence of abbreviations [116]. The majority of the information stored in EHRs consists of unstructured text (e.g., doctor notes, surgical summaries), which is typically underutilized for secondary purposes. Recent advances in NLP neural networks and deep learning techniques have surpassed classical statistical and knowledge-based algorithms in a number of tasks. This article provides a comprehensive overview on medical error, EHR and its adoption in both developed and developing nations. We also discussed EHR-related error from acronyms and abbreviation usage, while summarizing current ML- and DL-based NLP approaches for abbreviation disambiguation in clinical setting. This paper adopted a comprehensive approach to carefully review EHR-related medical errors caused by clinicians’ interactions with the EHR systems. These errors could appear during the data entry process and/or during subsequent information retrieval, and it is possible that automated WSD could be used as an effective remedy. A detailed overview was presented on medical error, its causes, and unintended repercussions. We also discussed EHR-related errors as a subset in the general medical error. We examined EHR adoption issues and challenges in both developed and developing countries, while suggesting appropriate solutions. To sum up, we critically analyzed the research trend and future direction in AI and NLP for medical acronyms sense resolution in clinical note.

This survey found that despite the widespread success of deep learning techniques in the general NLP domain, it remains challenging to apply them in the healthcare system owing to the scarcity as well as complexity of obtaining domain-specific text data and the ongoing evolution of research seeking the interpretability and explainability of deep learning techniques. In terms of decision-making and interpretability, better knowledge and information extraction from unstructured data and practical merging of both structured and unstructured data are prospective directions for this field’s future development. To atone for the lack of labeled textual data, another approach for EHR tasks could be unsupervised learning or adoption of pre-trained models, which can be fine-tuned with the limited data available via transfer learning. This study is intended to inspire readers and advance future NLP developments or clinical domain acronym disambiguation.

Despite the widespread success of DL-based techniques in the general NLP domain, this survey found that applying them in the healthcare system remains difficult due to the scarcity and complexity of obtaining domain-specific textual data, as well as the ongoing evolution of research seeking the interpretability and explainability of deep learning techniques. Better knowledge and information extraction from unstructured data, as well as practical merging of both structured and unstructured data, are prospective directions for this field’s future development in terms of decision-making and interpretability.

ACKNOWLEDGMENT

The authors would like to thank Universiti Teknologi PETRONAS and the Institute of Health and Analytics for their support toward this study.

References

References is not available for this document.