Abstract
Accurate recognition and linking of oncologic entities in clinical notes is essential for extracting insights across cancer research, patient care, clinical decision-making, and treatment optimization. We present the Neuro-Symbolic System for Cancer (NSSC), a hybrid AI framework that integrates neurosymbolic methods with named entity recognition (NER) and entity linking (EL) to transform unstructured clinical notes into structured terms using medical vocabularies, with the Unified Medical Language System (UMLS) as a case study. NSSC was evaluated on a dataset of clinical notes from breast cancer patients, demonstrating significant improvements in the accuracy of both entity recognition and linking compared to state-of-the-art models. Specifically, NSSC achieved a 33% improvement over BioFalcon and a 58% improvement over scispaCy. By combining large language models (LLMs) with symbolic reasoning, NSSC improves the recognition and interoperability of oncologic entities, enabling seamless integration with existing biomedical knowledge. This approach marks a significant advancement in extracting meaningful information from clinical narratives, offering promising applications in cancer research and personalized patient care.
Graphical abstract

Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Cancer remains a major public health problem, ranking as the second leading cause of death worldwide. In particular, breast cancer is the most common cancer in women worldwide and the second most common cancer in general [1]. According to the World Health Organization,Footnote 1 2.26 million women worldwide will be diagnosed with breast cancer in 2022. Furthermore, cancer treatment is an expensive process with significant social and economic consequences for patients and healthcare systems [2].
Electronic health records (EHRs) encode relevant information about a patient, representing a rich source for supporting clinical research [3]. The oncology care process generates a large amount of information that describes the evolution of cancer in patients [4]. Physicians capture this information in the EHR using narrative clinical notes [5]. Extracting and mining this information is critical to supporting oncology research and improving patient outcomes. However, extracting this information is challenging due to the complexity of natural language. The use of natural language processing (NLP) in the biomedical domain has increased the possibility of automatically extracting information from clinical oncology narratives [6,7,8]. Recently, deep learning-based approaches have shown their feasibility in obtaining accurate information in the processing of clinical narratives on cancer [9,10,11].
In the Spanish language, several studies have also aimed to extract information in the cancer domain [12,13,14]. However, these approaches have focused only on recognizing medical entity recognition or on the detection of negations and uncertainties [15, 16] using separate processes. They also do not use methods for linking entities or structuring the information after it has been extracted. Another interesting reference in Spanish is the Cantemist Challenge (CANcer TExt Mining Shared Task) [17], presented during IberLEF 2020. This effort represents the first collective attempt to evaluate and improve the development of tools for named entity recognition, concept normalization, and clinical coding, specifically targeting cancer-related information in Spanish. However, this challenge focused only on a single entity, tumor morphology, and did not define a complete pipeline. Instead, it was divided into three completely independent phases: NER, NORM, and CODING.
The state-of-the-art NLP tools scispaCy [18] and BioFalcon [19] are designed to recognize medical entities and associate them with terms in controlled vocabularies (e.g., UMLS)—as illustrated in Fig. 1. However, these tools often overlook contextual information and rely solely on string similarity. Furthermore, while they can be applied to Spanish notes, there is no specific Spanish version of these tools.
Motivating example: a clinical note snippet from a breast cancer patient, with correct UMLS terms as the gold standard (translated into English for clarity). Both scispaCy and BioFalcon fail to disambiguate and identify the correct terms, illustrating the need for contextual and semantic information to ensure accurate entity recognition in medical texts
In conclusion, the complexity of extracting valuable information from EHRs, especially in oncology, poses significant challenges due to the nature of clinical language. Existing approaches, especially in Spanish, either focus on isolated tasks such as entity recognition or do not implement entity linking or information structuring. In addition, state-of-the-art tools such as scispaCy and BioFalcon, while effective in English, lack tailored versions for Spanish clinical notes and rely primarily on string similarity.
Problem statement We address the problem of oncologic entity linking from medical notes, which involves accurately recognizing oncologic entities within the text and linking them to their corresponding terms in medical vocabularies.
Proposed solution We present a novel approach called the NeuroSymbolic System for Cancer (NSSC), a hybrid AI framework specifically designed to extract and organize oncology terms from clinical narratives. NSSC processes free-text clinical documents written in a narrative format, automatically identifies all relevant medical entities, and links them to the appropriate terms in the Unified Medical Language System (UMLS). To the best of our knowledge, NSSC is the first system capable of automatically structuring information from oncology clinical narratives, specifically for breast cancer, in Spanish.
Our approach integrates symbolic reasoning with language models to effectively recognize oncologic entities and link them to UMLSFootnote 2 based on real breast cancer clinical notes. First, language models such as BERT [20] are used to exploit the contextual knowledge encoded in clinical notes during the NER step. This first step is already addressed in [21]. Next, a symbolic system is able to reason over a background knowledge that encodes relationships between medical concepts and their labels and semantic types in UMLS. Finally, large language models (LLMs) such as GPT-4 are used for their advanced natural language understanding capabilities. These models assist in disambiguating the most appropriate UMLS term(s) by evaluating and interpreting the context around each recognized medical entity, thereby selecting the most relevant term from a number of potential options.
We have conducted extensive validation experiments to assess the NSSC effectiveness. We have developed an in-house benchmark to compare NSSC with existing approaches. The validation process includes rigorous evaluation metrics to demonstrate the robustness and reliability of NSSC in handling different clinical scenarios.
Our work makes several significant contributions to the field of clinical natural language processing, particularly in the area of oncology:
-
First integrated system for Spanish-specific clinical narratives: We present the first fully integrated system for extracting, linking, and structuring oncologic entities from Spanish-language clinical narratives. In doing so, we address the significant gap in the availability of NLP tools for Spanish medical texts.
-
Generalizable and cost-effective system: Our approach is designed to be adaptable beyond breast cancer, making it versatile for other types of cancer and potentially other medical conditions. It is also cost-effective because it does not rely on resource-intensive pre-labeling of data.
-
Novel disambiguation method based on LLMs: We propose a novel disambiguation method that exploits the capabilities of LLMs, such as GPT-4. This approach goes beyond traditional string similarity methods by using the contextual understanding power of LLMs to accurately identify and link the most appropriate UMLS terms, even in cases of ambiguous or complex medical terminology.
The rest of the paper is organized as follows: Section 2 summarizes the state of the art, and Section 3 presents the addressed problem and our proposed solution based on a hybrid AI system. Next, the NSSC algorithms are described in Section 4, and the results of the empirical evaluation are reported in Section 5. We discuss the observed results in Section 6, then the limitations are described in Section 7, and finally, we conclude with the conclusions in Section 8.
2 Related work
Recent advances in hybrid AI systems have demonstrated significant improvements in various domains. The Neuro-Symbolic System for Cancer (NSSC), CapMatch [22], and Perceptive Capsule Network (PCapN) [23] exemplify this trend. NSSC integrates symbolic reasoning, LLMs, and deep learning for oncologic entity recognition, while CapMatch combines capsule networks, contrastive learning, and knowledge distillation for human activity recognition (HAR) from wearable sensors. Similarly, PCapN employs capsule networks and distillation methods for multivariate time series classification. Despite their different applications—NSSC in medical text, CapMatch in sensor data, and PCapN in time series—each system emphasizes advanced representation learning, knowledge transfer, and context-aware processing, demonstrating the versatility of hybrid AI approaches.
Building on these recent advances, this section reviews the state of the art in normalization of free text in biomedical data, focusing on three core areas: named entity recognition-NER (Section 2.1), entity linking-EL (Section 2.2), and named entity disambiguation-NED (Section 2.3). We analyze different approaches, including those based on deep learning, knowledge bases (KB), symbolic methods, and hybrid AI systems. We also highlight gaps in current approaches, setting the stage for further innovation (Section 2.4).
2.1 Named entity recognition
Named entity recognition (NER) is a technique used to classify words in free text into predefined categories such as person names, locations, or medical terms. State-of-the-art NER techniques predominantly rely mainly on deep learning methods [24], incorporating input representations such as GloVe with BERT document-level embeddings [25], Cloze-style language model embeddings [26], or GloVe with aggregated contextual embeddings [27]. These inputs are processed by context encoders such as LSTM [25, 26] or RNN [27], and the outputs are decoded by tag decoders such as CRF [26, 27]. These approaches achieve high performance, with accuracy rates of up to 93.5% on the CoNLL-2003 dataset [26] and 92.07% on the OntoNotes5.0 dataset [25].
Although NER has traditionally focused on identifying entities without explicit linking, recent advances have integrated knowledge bases (KBs) to improve entity recognition accuracy by leveraging pre-existing entries. However, this approach is more naturally aligned with the entity linking process, as discussed in Section 2.2. NER methods based on KBs often have limitations compared to deep learning models, primarily due to their dependence on predefined databases, limited adaptability, and lack of contextual understanding. Conversely, deep learning models excel are characterized by scalability, contextual understanding, and automatic learning, resulting in more accurate and adaptive entity recognition. However, they require large annotated corpora to perform effectively, which can be a significant barrier in specialized domains, such as breast cancer.
There are specific cases where corpus annotation is unnecessary, especially when the text is very similar to entries in the KB. In such scenarios, KB methodologies that incorporate rule-based systems can be advantageous. For instance, [28] proposes a novel approach to train NER models without labeled data, achieving a seven-point improvement in F1 scores on the CoNLL 2003 and Reuters/Bloomberg datasets.
The use of NLP approaches to extract information from clinical texts has increased in recent years [29,30,31]. Specifically, within cancer-related clinical texts, models such as RoBERTa have demonstrated good performance, as shown by [32]. While deep learning methods achieve high accuracy and adaptability, they still struggle with domain-specific challenges, particularly when context is complex or the data is sparse. In contrast, NER methods using KBs lack the contextual understanding and adaptability of DL models but can be beneficial in specific scenarios where text closely matches KB entries. This highlights a gap in the development of more robust approaches that can handle unannotated data and specialized domains, such as clinical text, where models such as RoBERTa have shown promise. Future research should focus on bridging the gap between the scalability of deep learning methods and the accuracy of KB-based approaches while improving domain-specific NER capabilities.
2.2 Entity linking
Entity linking (EL), also known as entity normalization, is the process of matching entities mentioned in a text with their corresponding entries in a KB. This step typically follows the extraction of entities from the text, where (entity, label) tuples are obtained. Given the complexity of natural language—including misspelling and inherent ambiguity— this information must then be normalized.
State-of-the-art EL models use neural architectures that have proven superior to classical ML methods [33]. These models can be categorized into four main types: joint mention detection and linking, global models, domain-independent approaches (including zero-shot methods), and cross-lingual techniques. Recent advances have shifted towards self-attention architectures such as BERT for mention encoding, with zero-shot methods becoming increasingly prevalent. Cross-encoder architectures, such as E-BERT [34] and BART-based models [35], have also demonstrated strong performance for joint tasks.
The EL task has seen applications using KBs, from graph traversal methods [36] to neural network-based approaches [37]. De Cao et al. [38] introduced a generative model that eliminates the need for hard negative sampling during training. These methods improve EL performance by integrating additional knowledge such as entity definitions, entity types, and knowledge graph (KG) triples to support training [39]. Despite these advances, models still rely heavily on a fixed number of candidates, approach the problem as a classification task, and require extensive training data. To mitigate the dependence on labeled datasets, some techniques [40] have begun to use unlabeled corpora. Recent efforts have focused on developing zero-shot models capable of generalizing EL to previously unseen entities [41, 42].
Symbolic approaches in artificial intelligence have also gained popularity [43]. Sakor et al. [19, 44] propose rule-based methods for linking entities in short texts using knowledge graphs such as DBpedia [45] and Wikidata [46]. These approaches highlight the importance of linguistic rules in understanding sentence structure, which helps in identifying the context identification of entities. Similarly, [47] presents a lightweight method for linking entities in Wikidata based on heuristic rules.
Hybrid AI approaches to EL, which combine symbolic reasoning with neural networks, are an active area of research aimed at improving the explainability of algorithms. Jiang et al. [48] introduced LNN-EL, a neuro-symbolic EL approach that combines human-defined rule templates with neural learning, providing interpretability and transferability without extensive labeled data. However, this method is primarily designed for short texts, such as questions, which limits its generalizability to more complex domains. Plu et al. [49] propose a hybrid method that combines a linguistic-based approach combined with annotation coverage facilitated by a KB.
In [50], they perform EL as an information retrieval (IR) task. However, in [51], they propose BertMCN, which uses BERT as a multi-classification task to normalize for mapping health-related entity mentions. In addition, [52] improves the model of Pattisapu et al. [53] for medical concept normalization by jointly learning the representations of the input and target concepts using RoBERTa as the embedding generator. This approach improves accuracy by up to 2.31% on three standard datasets. The improvement is significant because previous approaches only train the transformation of concepts into embeddings separately from the mentions that appear in the text. Recent work, such as Gallego et al. [54] (2024), proposed ClinLinker, which implements a two-stage pipeline using a SapBERT bi-encoder for candidate retrieval and a SapBERT cross-encoder for re-ranking, specifically tailored to handle Spanish medical texts using UMLS as KB.
The EL literature highlights gaps in generalizability across domains and scalability with large KBs. While zero-shot and cross-lingual methods, they still struggle with unseen entities and low-resource languages. Furthermore, the integration of symbolic and neural approaches is limited, with current models often failing to handle complex and ambiguous entities effectively. Future research should focus on developing more domain-agnostic models, improving efficiency with large KBs, improving zero-shot learning, and better integrating symbolic reasoning with neural methods to address these challenges.
2.3 Named entity disambiguation
Named entity disambiguation (NED) is an NLP task that aims at resolving ambiguities arising from named entities in text. It can be considered as a subtask of EL, where NED is responsible for selecting the most appropriate candidate from a list generated for a mention in the text, based on context [55].
The NED task has been a subject of investigation for some time. For example, [56] proposed a method using a disambiguation SVM kernel trained on an online encyclopedia to detect and disambiguate named entities in open domain text. Novel approaches have been explored using large language models (LLMs) for disambiguating homonyms in academic KGs [57].
Symbolic approaches utilizing KBs have been a significant area of interest for NED tasks. For instance, [58] addresses spatial NED by identifying and assigning precise coordinates to ambiguous place names in text. Additionally, [59] focuses on applying NED to short text fragments in KGs, proposing an approach that includes context expansion with WordNet, coherence exploitation between entities, relation-based similarity calculation, and the use of syntactic features to improve the accuracy of query answering systems.
In [60], a technique for integrating cross-domain data by transferring structural knowledge from a general text KB to the medical field significantly improved disambiguation performance on benchmark datasets such as MedMentions and BC5CDR. Supervised classification approaches that identify informative keywords in the context of named entities have improved transferability [61]. Graph Neural Networks (GNNs) such as GraphSAGE, R-GCN, and MAGNN achieve state-of-the-art results by representing entities as query graphs and employing effective negative sampling strategies [62].
Recently, BELHD [63] introduced two major improvements: (i) the extension of homonyms in the KB with unique disambiguating strings, ensuring unambiguous linking decisions, and (ii) the implementation of a new strategy in contrastive learning that selects candidates more effectively, improving the training process. Further research [64] has also adopted NED, focusing on a retrieve-and-rerank approach.
Despite progress, the state of the art in NED reveals several research gaps and opportunities for further exploration. Traditional approaches using SVM and symbolic methods with KBs have made progress in disambiguating entities, especially in specific domains such as geography and short text fragments. However, these methods often struggle with complex or cross-domain scenarios where large LLMs and advanced neural methods, such as GNNs, show promise. There is still a need for more effective context handling, especially in domain-specific applications such as medical texts. Recent techniques, like contrastive learning and improved homonym handling, represent steps forward, but further research is needed to improve the integration of domain-specific knowledge with general-purpose models, to increase the portability of these approaches across different domains, and to develop more robust methods for handling ambiguous entities in different contexts.
2.4 Summary and gaps in existing research
Several research gaps remain in NER, EL, and NED. Current deep learning models, while powerful, require large amounts of annotated data, which are often lacking in specialized domains such as oncology. Meanwhile, KB-based approaches lack the contextual adaptability of DL models. Future research should focus on developing more robust hybrid models that combine the contextual understanding of deep learning with the specificity and interpretability of KB-based approaches. Additionally, there is a need for models that can better handle low-resource languages and domain-specific nuances, particularly in the biomedical domain. To solve the problem of oncologic entity linking, three approaches can be considered:
-
Symbolic AI systems: These are rule-based systems that use syntactic or lexicographic rules from a KB. They provide clear and understandable solution paths but struggle with variations in text mentions that differ from vocabulary labels, since concepts can be expressed in multiple ways, and the same word can have different meanings depending on the context.
-
Sub-symbolic AI systems: These systems, often based on deep learning, require extensive annotated data. They typically outperform symbolic systems in most NLP tasks but lack interpretability and rely heavily on extensive training data.
-
Hybrid systems: These systems combine symbolic and sub-symbolic approaches to exploit the strengths of both, aiming for high performance while maintaining some degree of interpretability.
Table 1 provides a selection of related articles. Most approaches focus only on individual steps such as NER or EL, with few proposing a complete pipeline. To the best of the author’s knowledge, no approach is specialized in the cancer domain or specifically tailored to Spanish texts. NSSC addresses the need for cost-effective solutions to extract insights from oncologic free-text records by introducing a hybrid AI system. This system operates without the need for extensive data annotation and utilizes an explainable decision-making process.
3 NSSC—problem statement and proposed solution
The Neuro-Symbolic System for Cancer (NSSC) is designed to provide structure and semantics to unstructured short medical notes, addressing the critical need for normalization of clinical notes within specific medical domains, such as oncology. The primary motivation for developing NSSC is the lack of robust tools and methodos capable of effectively transforming these unstructured texts into structured data, which is crucial for advancing clinical research, decision-making, and patient care.
Our approach formulates the problem as an optimization task, with the goal of maximizing the accuracy of medical entity recognition and linkage while minimizing the associated costs. NSSC utilizes a hybrid AI model that integrates three distinct paradigms: deep learning, symbolic reasoning, and generative AI, each of which contributes uniquely to NLP tasks. By orchestrating components based on these three paradigms, NSSC effectively recognizes medical entities and links them to standardized medical terminologies within controlled vocabularies, such as the UMLS. Furthermore, the modular design of NSSC ensures its generalizability across different medical domains beyond oncology.
A formal definition of the problem addressed by NSSC is provided in the following subsection, as well as a design pattern outlining the core components of the proposed framework. Table 2 summarizes the notation used in the specification of NSSC.
3.1 Formal problem statement
Given a set \(\mathcal {I}\) of medical entities, a set \(\mathcal {J}\) of medical terms from a controlled medical vocabulary \(\mathcal {V}\), and a short note T. The problem of oncologic entity linking is to recognize from T, a set \(h^*\) of the correct medical entities in \(\mathcal {I}\) and their links to terms in \(\mathcal {J}\), while minimizing the cost of \(\delta (h^*,T)\).
-
\(h^*\) is the set of recognized medical entities and their links to medical terms in \(\mathcal {V}\) that optimizes \(\delta (.)\), i.e.,
$$\begin{aligned} h^*=\{(e,t) \mid e \in \mathcal {I} \wedge ~ t \in \mathcal {J} \wedge ~ e \textit{ appears in T } \wedge \textit{ t is the medical term for e } \} \end{aligned}$$ -
\(\delta (h,T)\) is a utility function that quantifies the cost of correctly recognizing the medical entities in T and the terms in \(\mathcal {J}\) that express the meaning of the linked entity in the controlled vocabulary \(\mathcal {V}\).
$$\begin{aligned} \delta (h,T)= \phi (h,T) + \rho (h,T) + \alpha (h,T) \end{aligned}$$-
Cost of correctly recognizing medical entities in T: \(\phi (h,T)=\sum _{(e,t) \in h} \phi '((e,t),T)\):
$$ \phi '((e,t),T)={\left\{ \begin{array}{ll} 0, & \text {if { e} is a correct medical entity in { T}}\\ 1, & \text {otherwise}; \end{array}\right. } $$ -
Cost of correctly linking medical entities in T: \(\rho (h,T)=\sum _{(e,t) \in h} \rho '((e,t),T)\):
$$\begin{aligned} \rho '((e,t),T)={\left\{ \begin{array}{ll} 0, & \text {if { t} is the correct term for the medical entity { e} in { T}}\\ 1, & \text {otherwise}; \end{array}\right. } \end{aligned}$$ -
Cost of executing a model for performing named entity recognition and linking: \(\alpha (h,T)=\sum _{(e,t) \in h}\) \(\alpha '((e,t),T)\), where \( \alpha '((e,t),T)\) quantifies the cost of processing the tasks of recognizing e and linking e to t.
-
3.2 Hybrid AI system: a boxology design pattern
NSSC is designed as a hybrid approach to recognize medical entities and to associate the recognized entities with medical terms in a controlled vocabulary. This hybrid approach combines a language model trained with annotated medical entities, a rule-based system guided by symbolic background knowledge, and a large language model to solve the optimization problem in Eq. 1.
The modular design pattern in Fig. 2 illustrates the usage of a hybrid AI system that combines these three AI subsystems. This compositional design pattern is based on the basic design patterns proposed by van Bekkum et al. [70] to describe hybrid AI systems. Thus, NSSC comprises the following components:
-
Named entity recognition (NER) component: This component aims to identify oncologic entities within clinical text. Medical entities are recognized from medical notes based on a BERT-based language model pre-trained with oncologic terms from clinical notes. The NER component receives a clinical note T and produces a set of medical entities \(\mathcal {I}\); for each recognized entity e in \(\mathcal {I}\), it assigns a label \(l_e\) representing its context in the clinical note. This annotation is captured in the function \(\gamma (.)\) in a way that \(\gamma (T)\)=\(\mathcal {I}\). Using a BERT-based NER procedure, the NER component systematically analyzes unstructured clinical notes in T to recognize relevant entities such as tumor types, treatment modalities, and relevant clinical concepts. This task of entity recognition task includes a contextual analysis mechanism to capture the broader context surrounding the identified oncologic entities. Contextual information is crucial for the disambiguation of entities. The NER component trains a model on a diverse dataset of clinical notes, ensuring the accurate extraction of entities along with their contextual information. The process to train this model is described in detail by García-Barragán et al. [21].
-
Medical entity linking (MEL) component: A rule-based model is utilized to link medical entities to a list of candidate medical terms; a symbolic background knowledge serves as an extensional database to deduce candidate links. The MEL component receives the set \(\mathcal {I}\) of medical entities recognized by the NER component and generates a set of pairs (e, \(t_e\)) where e belongs to \(\mathcal {I}\) and \(t_e\) is a set of candidate medical terms in UMLS. The MEL component resorts to a symbolic background knowledge (BK) to identify the potential terms from UMLS associated with e. The BK is built on top of the UMLS and represents medical terms using factual statements that associate the term with labels and semantic types. Since for the same medical term, different organizations may have assigned different labels to the term, BK keeps track of the number of duplicate labels and ranks the labels per term based on this number; we call this number the linking score of the labels of a term. Given an entity e, the MEL component searches on BK, the potential terms from UMLS that will be part of \(t_e\). To increase confidence in the correctness of the terms in \(t_e\), the MEL component follows a heuristic-based approach, which assumes that the labels with the highest values of the linking score are the most appropriate labels for e. Following techniques proposed by Sakor et al. [44], the MEL component indexes e and extends it with their synonyms. A BM25\(^\dagger \) algorithm [71] is executed on the BK for querying and ranking entities in \(\mathcal {I}\) based on linking score, including label, semantic type, score, and definition in each index. As a result, for each entity e in \(\mathcal {I}\), a set \(t_e\) of the top k candidates for links in UMLS is generated.
-
Medical entity disambiguation (MED) component: Symbolic background knowledge is used for prompt engineering, and a set of prompts is generated to query a large language model (LLM) whose responses allow the disambiguation of the medical terms associated with each recognized entity. The generated prompts aim to maximize the efficiency of the large language model while minimizing the cost of using the model. The MED component receives triples \((e,l_e, t_e)\) and generates the set h corresponding to a solution of the optimization specified in Eq. 1. This component uses the predictive capabilities of LLMs and the reasoning process performed over BK to identify for each e and a singleton set \(t_e\). The model captures the broader context surrounding the identified oncologic entities (provided as input), ensuring a holistic understanding of the entities and their interaction within the clinical notes, as well as the contextual information of the entity encoded in the BK. The medical entity linking process facilitates the extraction of CUIs based on the context and semantics of the identified entities, improving interoperability, consistency, and semantic alignment. This contextual information—collected from the background knowledge—is exploited for performing a Chain of Thought (CoT) [72] for engineering the prompts posed over the LLM.
4 The NSSC algorithms
This section introduces the algorithms that implement NSSC, focusing on how they integrate symbolic reasoning with advanced language models to address the complexities of oncologic entity linking. By using a background knowledge base and a modular design, NSSC effectively handles the variations in medical terminology and context that are prevalent in clinical narratives. This ensures that terms such as “mama derecha” from the vocabulary are correctly recognized even when expressed differently, such as “mama parte derecha,” in clinical notes. The initial search generates a set of potential candidate matches. The algorithm then uses heuristics to determine whether these candidate terms should be passed on to an LLM for a more fine-grained disambiguation process. Actually, this mimics the process a human would use to associate a Spanish term with a medical vocabulary:
-
1.
Recognize an entity in the text.
-
2.
Search for the entity without any modifications.
-
3.
Search for similar terms in the list of candidates.
-
4.
If not found, translate the entity into English and perform the search again.
NSSC has three configurable parameters: the query type \(q_{\text {type}}\), a threshold of the symbolic module \(\theta _{\text {sim}}\), and a prompt \(p_{LLM}\) to disambiguate concepts. Each parameter is described in Table 3.
The NSSC background knowledge (BK) has four specialized modules categorized by article [73]:
-
Domain knowledge and ontology
-
Indexer stores and indexes vocabulary medical terms for retrieval.
-
Semantic information is used to store domain external information to make decisions about the disambiguation process.
-
-
Trained models
-
Clinical NER corresponds to the trained model in the breast cancer domain for clinical entity recognition.
-
LLM: a generic model which is responsible for disambiguating candidates.
-
NSSC is presented in Section 4.1, which describes how a free text flow through the whole system. Each of the three components of our NSSC system is explained separately, entity recognition in Section 4.2, entity linking in Section 4.3, and entity disambiguation in Section 4.4. Finally, the algorithm complexity is discussed in Section 4.5.
4.1 NSSC flowchart
After training the models and configuring all parameters, NSCC normalizes new Spanish free text and extracts the corresponding CUIs. This online process is illustrated in Fig. 3. NSSC receives T, an oncologic clinical note which is initially passed to the NER module to obtain a set of entities \(\mathcal {I}\), each represented as a tuple \((e, l_e)\). Subsequently, for each entity in \(\mathcal {I}\), the medical entity linking module consults the BK using rule-based heuristics to make a decision. This decision determines whether we have an acceptable term t for an entity e, do not have it, or do not need to disambiguate. If disambiguation is required, the medical entity disambiguation module creates a prompt \(p_{LLM}\) with the medical entity e, each label \(l_e\), the candidate terms \(t_e\), and their corresponding UMLS definitions. Ultimately, NSSC either successfully selects the best terms, finds them inadequate, or provides a translation of the term. If the term is translated, it is searched again, and the process is repeated.
NSSC flowchart: clinical note T with the set \(h^*\), which is a set of normalized terms aligned to UMLS. NSSC takes as input a short medical note in unstructured format and transforms this text into a set of medical entities. For each recognized entity, a background knowledge (BK) is searched to identify candidates for links in UMLS. In case disambiguation is required, semantic information from the candidate terms is used for Chain of Thoughts (CoT) prompting over an LLM
For instance, given an input T, “Paciente diagnosticado con carcinoma ductal in situ,” the NSSC system will extract the set of entities \(\mathcal {I} = \) {(“carcinoma,” “cancer concept”), (“ductal,” “cancer type”), (“in situ,” “cancer expansion”)}. Then, for each tuple, it assigns the most appropriate CUIs for each concept. For the concept “carcinoma,” the result might be (“carcinoma,” “Cancer Concept,” “{‘C0007097’}”), where “C0007097” is the CUI associated with the concept of “carcinoma.”
4.2 Named entity recognition
NSSC uses a BERT-based model to perform NER. We choose for this technique because of the nature of natural language text, where the quality of written notes tends to be suboptimal due to the pressures faced by clinicians. Consequently, these models are essential because they adapt to language nuances and are able to identify words even with grammatical errors. These models require an annotated corpus to learn.
Algorithm 1 outlines the steps of the named entity recognition process. It receives a clinical note T and extracts a set of entities \(\mathcal {I}\). This set contains tuples \((e, l_e)\), where \(l_e\) is the medical label and e represents the text that appears in the prediction for a label \(l_e\) by the NER model. This model is used after a preprocessing step. This step aims to prepare the clinical text before performing the information extraction task by splitting the clinical narrative into sentences and tokenizing each sentence. Acronyms are then transformed into full descriptions, such as by converting “ca.” to “cáncer.” This transformation is achieved using a combination of regular expressions and a dictionary of the most common acronyms in the cancer domain.
In a previous work [21], a model was trained to extract medical concepts using multilingual BERT [20]. NSSC utilizes this model, this approach that takes advantage of a transfer learning technique to perform clinical NER in the field of breast cancer. Transfer learning is achieved by fine-tuning the BERT model with a classification layer on top, as described in [16]. However, in [32], it is shown that RoBERTa generally outperforms BERT in medical text. In any case, the entity identification is a required step of the process that can be performed with the state-of-the-art methods that the researcher decides.
4.3 Medical entity linking
The NSSC medical entity linking uses a BK to generate the candidates; it is configured based on the following parameters:
-
Background knowledge base: We have built a background knowledge [73] based on UMLS. For each medical entity, the background knowledge includes all its labels, semantic types, definitions, and synonyms in other knowledge bases, such as DBpedia or Wikidata. These descriptions provide contextual information collected by the communities contributing to UMLS and the knowledge bases.
-
Search engine: The medical entity linking system resorts to OpenSearchFootnote 3 to search for background knowledge. OpenSearch relies on indexes and search methods and is able to scale to large knowledge bases.
-
Query to be posted to the search engine. As we will see, experiments have been conducted to find the most appropriate configuration of the search. The query may have the following configurable parameters:
-
k: It controls the number of candidates that the BK should return for a given medical entity. It is represented by a natural number.
-
fuzziness: The degree of fuzziness or tolerance in matching concepts.
-
*
Exact: This method is strictly an exact match. Only results that exactly match the query concepts in the BK.
-
*
Basic fuzzy: This method uses basic fuzzy matching, allowing for slight variations in the query terms.
-
*
Multi-match: This query method represents the highest level of fuzziness.
-
*
-
lang: The language of the terms is saved.
-
*
SPA: Spanish language.
-
*
ENG: English language.
-
*
-
boost words: Keywords that receive a boost in relevance.
-
-
Similarity function in the optimization module: It determines the relationship between the vocabulary term and a medical entity.
Algorithm 2 is a sketch of the algorithm implemented for the medical entity linking system. When the results are retrieved from the search engine, they are ranked according to their relevance to the query. Each result is assigned a relevance score based on how well it matches the search query. With all of the results, we have a set of candidates that could be used directly as input to query the LLM. The result of the search will be a set of candidates (\(t_e\)) along with the ranking value.
The choice of query type is another customizable parameter within our system. Queries can be Exact, Basic_Fuzzy, Multi_match_SPA, Multi_match, and Multi_match_boosted. Details in Section 5.1.
4.3.1 Filtering
The results of searching in the knowledge base can be large. As we know, optimizing the number of tokens sent to an LLM is crucial for several reasons related to efficiency, computational cost, and model limitations. In particular,
-
1.
Response time: Smaller inputs typically result in faster response times, which is critical for applications that require real-time or near real-time interactions.
-
2.
Model output constraints: Some models have constraints on the maximum length of generated responses, and sending fewer tokens ensures that the response fits within those constraints.
-
3.
Noise reduction: Sending extraneous or irrelevant tokens to the model can introduce noise and potentially affect the quality of the generated output.
-
4.
Mitigate model bias: Expanded inputs can weaken the influence of essential details, potentially leading to biased or less accurate results.
As a consequence, our approach includes an optimization step in which the search results are grouped by cui and ranked with a matching function. We propose to use a similarity function s(.,.) that, when given a term of the candidate set (result of a search query in the search engine, \(t_e\)), returns a numeric value in the range of 0 to 1. Then, a threshold is defined so that only those candidates above the threshold are included in the final set of filtered candidates. This matching function is implemented in the Python documentation.Footnote 4
NSSC uses a threshold \(\theta _{\text {sim}}\) that varies with context and requires configuration. The choice of this threshold for our case study is described in detail in Table 8.
4.3.2 Heuristic decision
At this stage, three different scenarios can unfold: (i) there are no candidates are in the set, indicating that the concept does not appear in background knowledge; (ii) the list contains only a single term; or (iii) the list encompasses multiple terms. In the second case, a direct mapping is established. In contrast, for the last scenario in our methodology, we utilize an LLM to facilitate the determination of the optimal solution, taking into account contextual nuances, the range of ranked candidates, and the definition of the concept in the specific vocabulary.
4.4 Medical entity disambiguation
An LLM is used by NSSC to produce a result, considering the contextual information provided by the previous phase. This model offers insights into likely meanings based on the context in which the term is used. Creating a robust context is key to improving the performance of language models in disambiguation tasks. A detailed explanation of this process can be found in Algorithm 3. When using LLMs, several techniques can be used to control the behavior of the model and achieve the desired results. One important method is prompt engineering, which involves creating specific prompts to guide the model’s behavior. These prompts may include explicit instructions, contextual setting, or special formatting to guide the model towards the expected output. Two important prompt engineering techniques are as follows:
-
Few-shot learning [74] involves giving the model a limited set of examples or demonstrations that illustrate the target behavior. This method helps the model better understand the intended task or concept more effectively. The model then uses these examples to generalize and formulate responses to novel prompts or questions, demonstrating understanding beyond the specific instances provided.
-
Chain of Thought (CoT) [72] involves crafting prompts that encourage the model to reveal its reasoning process in a step-by-step manner, similar to how a human might think aloud while solving a problem. This approach is particularly effective for complex tasks, because it seeks not only the correct answer but also the logical path to that answer. By breaking down the thought process, CoT enhances the model’s ability to handle complex questions and provides users with a clear understanding of how the model reached its conclusions, thus improving transparency and trustworthiness.
Although NSSC is LLM-agnostic, as a proof of concept, we implemented this step using the OpenAI API.Footnote 5 This API involves determining the appropriate model and includes configuration parameters such as the maximum number of tokens and the temperature of the language model. The optimal configuration of these parameters is essential to achieve the desired behavior.
In Section 5, the effects of combining different possibilities to assess their impact on the LLM’s outcome. Although the full text of the note in the prompt could offer the potential for richer interactions with the LLM, several challenges must be effectively addressed to make this approach successful.
-
Computational costs: LLMs, especially those hosted in the cloud, can have computational costs associated with processing large amounts of text. The more tokens you send, the more resources are required for analysis.
-
Response time: Sending lengthy prompts may result in longer response times from the LLM. Although this is not a real-time issue, it is something to consider.
-
Privacy and sensitive information: Sending complete notes may involve sharing sensitive or private information with external language model services, which raises privacy concerns.
-
Model capacity limits: Some language models have input limits and may have a maximum token limit per request, and sending very long requests may result in truncation or incomplete processing.
The best prompt for our approach is illustrated in Table 9, which is a CoT prompt. An example of CoT is illustrated in Fig. 4. In this example, we show how, by using the CoT prompt technique, the LLM can intelligently select the most appropriate CUI based on the contextual information provided. The language model is required to justify its decision by explaining whether a suitable CUI exists; if not, it should provide a translation of the term. In addition, each candidate is given the CUI, the label, the semantic group, and the definition of the CUI is provided. Because UMLS often provides multiple definitions for each CUI, we give preference to the definitions provided by the NCI (National Cancer Institute)Footnote 6 because it is a recognized and authoritative source of cancer-related information.
4.5 Algorithmic complexity
The time complexity of NSSC is determined by the sum of the time complexities of the three algorithms of which it is composed. Since NSSC takes a text input that is divided into an arbitrary number of tokens, the key parameters to consider for each module are as follows:
-
NER: The BERT-based model must predict an entity for each token for a number of tokens. The preprocessing function time is minimal compared to the NER model.
-
\(s\): Number of tokens is by divided the input text.
-
\(b\): Number of neurons in the NER model.
-
-
MEL: The BM25 algorithm has to compare all the terms in UMLS with the entities extracted from the previous step. Then, it has to filter the subset of candidates and decide whether to disambiguate or not.
-
\(r\): Number of entities extracted from the text.
-
\(u\): Number of terms in the UMLS.
-
\(k\): Number of filtered candidates.
-
-
MED: A prompt must be created for each entity that needs to be disambiguated. This prompt is then processed by the LLM, which generates a response token by token. For each new token, all the neurons in the model must be computed.
-
\(l\): Number of neurons in the LLM.
-
\(p\): Length of prompt, expressed in number of tokens.
-
\(o\): Answer length generated by the LLM.
-
Given all these parameters, the time complexity of NSSC can be expressed as
As is known, the disambiguation process is the most computationally expensive since the number of neurons in an LLM is about 100 billion. However, the NSSC algorithm is optimized to minimize the use of the LLM whenever possible. In the best case, the complexity of NSSC only involves the NER and MEL modules.
Table 4 summarizes the space complexity for the various components involved in the deployment, focusing on the required disk space.
5 Validation experiments
This section reports on the results of the experimental study of the NSSC performance. The aim is to assess NSSC and compare it with existing approaches. We use a case study in breast cancer with real clinical notes. The experimental study is guided by the following research questions (RQs):
RQ1: How does the integration of background knowledge through different query types affects the overall performance and effectiveness of NSSC?
RQ2: How does the cost-effectiveness of utilizing NSSC models for natural language processing tasks vary when employing different prompt techniques, taking into account the influence of token length on accuracy?
RQ3: How can the combination of LLMs and the background knowledge contribute to better disambiguation of entity linking, particularly in cases where multiple entities share similar or ambiguous surface forms within clinical notes?
The experimental setup is as follows:
Dataset. We have generated an in-house corpus containing clinical notes from nearly 600 breast cancer patients. All the notes correspond to the category of clinical judgement. The corpus has been annotated following the process described in [21] resulting in a total of 17,986 entities. Table 5 reports on the labels that have been used in the annotation process, their support in the corpus, and the number of tokens. The former is calculated using the spaCyFootnote 7 tokenizer.
Figure 5 illustrates an annotated clinical note from our NER corpus. The phrase “Diagnosticada con carcinoma ductal infiltrante de mama derecha” contains four entities: cancer concept, cancer type, cancer expansion, and cancer location. It is important to note that identifying these named entities is only the first step. Subsequently, these medical entities are normalized to the UMLS vocabulary.
In order to validate the NSSC, we introduce an in-house validation benchmark, called the Breast-Norm-Benchmark, which is constructed from the NER corpus described above. This benchmark includes each term from the NER corpus only once, meaning that duplicate terms are not repeated. This benchmark is described in Table 6 and consists of 438 unique terms. These terms, along with their associated entities, will be the context provided to the LLM.
Below are two examples from the Breast-Norm-Benchmark:
-
Text: “cuadrante superior externo mama izquierda” Context: “Cancer Location” CUIs: “{‘C1266903’}”
-
Text: “mastectomía ahorradora” Context: “Surgery” CUIs: “{‘C0024881’, ‘C1997268’}”
During medical entity linking, we encounter the challenge that many UMLS terms are available only in English and not in Spanish. In our background knowledge—created using UMLS 2023AB–\(-\)25.8% of the terms are in English, while only 10.8% are in Spanish. To overcome this issue, we first search for the term in Spanish. If it is not found, we use the LLM to translate the term into English and then search for its corresponding English term in UMLS.
Implementation NSSC is implemented in Python 3.10. The source code is located in https://github.com/SDM-TIB/NSSC. The experiments are executed in an Intel Core i9-7900X CPU with a clock speed of 3.30 GHz, equipped with 20 CPU cores, organized into 10 cores per socket, each supporting 2 threads per core. The CPU architecture was identified as x86_64. The RAM consisted of 4 x 16GiB DDR4 synchronous modules, each running at 2133 MHz. The O.S. is Linux with kernel version 5.15.0-83-generic, based on Ubuntu 20.04.1 LTS.
Metrics The performance of NSSC is measured in terms of accuracy (3). We use this metric because an entity can have more than one valid CUI in UMLS (e.g., the entity “chemotherapy” can be valid with the following CUIs: “C0013216,” “C3665472,” “C1571591”). This metric focuses on the proportion of correct (gold) values that the system also identified.
where:
-
G = set of gold (correct) values
-
R = set of outputs by NSSC.
Furthermore, we use Accuracy@N (Eq. 4) to allow comparison with other systems, as classical methods cannot retrieve the best CUIs and are limited to providing a top-N set of results.
where:
-
\( R_{\text {{top-N}}} \) includes the top N responses in the system’s output.
Execution time corresponds to the elapsed time spent to execute an NLP tool to solve the problem of oncologic entity linking in the entire benchmark. The execution time is measured using the Python library time. It is acknowledged that the timing of a general-purpose computer is influenced by the operating system scheduler. As a result, the reported times represent an average of five different runs conducted at various times. Finally, the metric tokens(.) quantifies the number of tokens generated by an LLM. In our experiments, tokens(.) is calculated using the OpenAI API.
Baselines The following NLP tools are considered as baselines:
-
scispaCy [18] is a specialized approach designed for processing biomedical and scientific texts, based on the spaCy library’s robust framework. It specifically addresses the complex requirements of biomedical information, providing an extensive array of tools and models designed for functions like named entity recognition, disambiguation, among others.
-
BioFalcon [19] is an entity recognition and linking engine. It extends the background knowledge and the target KG of the Falcon approach from DBpedia and Wikidata to UMLS to support entity linking to UMLS. BioFalcon effectively maps entities and relations within a short text to its mentions of a background knowledge graph. It overcomes the challenges of short text using a light-weight linguistic approach based on a background knowledge graph.
NSSC has three configurable parameters (see Table 3) that must to be set appropriately. To achieve this, we present ablation studies described in Section 5.1, where we tune each module separately in a sequential search space. Since the term must be among the candidates selected by the LLM, the linking part must be optimized to present the correct candidates. Then, we compared NSSC with the parameters selected in the ablation study with the baselines in Section 5.2.
5.1 Ablation studies
This section details the process of configuring various components of NSSC. The initial results related to the NER model are described in [21]. Subsequently, our focus shifts exclusively to the entity linking aspect. Ablation studies are performed to distinguish the collaborative and independent functionalities of medical entity linkage and medical disambiguation.
In systems with a constrained number of modules, each equipped with its unique configuration parameters, the parameter search process becomes crucial. In our specific context, where modules operate autonomously, we advocate a strategic approach to parameter search. The primary purpose of each module is to optimize the accuracy of the candidates it presents. In simpler terms, each module should return the most accurate and relevant candidates, prioritizing their ranking within the top selections. We address RQ1 by presenting a comparative analysis in Table 7. To configure the entity linking module, we have tested five different querying methods, each with an increasingly wide range of search results:
-
1.
Exact: This method is strictly exact match. Only results that exactly match the query terms in the knowledge base are returned.
-
2.
Basic_fuzzy: This method uses basic fuzzy matching, allowing for slight variations in the query terms.
-
3.
Multi_match_SPA: This query method represents the highest level of fuzziness and targets Spanish terms.
-
4.
Multi_match: The same level of fuzziness is as Multi_match_SPA, but can also return terms from any language.
-
5.
Multi_match_boosted: This query method can return results in any language and boosted terms related to cancer which are prioritized in the search results, that is, if the term cancer appears in the term candidate, it has a higher score.
The metrics used for comparison are the following:
-
Accuracy@1 (%): This represents the accuracy in retrieving the best-fit term from the BK. For instance, the Multi_match_boosted method has an accuracy of 54.02%, indicating that in 54.02% of the cases, the top term retrieved was the correct one.
-
Accuracy@15 (%): This measures the accuracy when retrieving a broader set of terms, up to the 15 best matches. The Multi_match_boosted method demonstrates an 77.64% accuracy rate for the top 15 terms, suggesting that while the single best term may not always be correct, expanding the search to the top 15 yields a much higher likelihood of retrieving the correct terms.
-
Execution time (s): Indicates the time spent in this initial phase of term retrieval. The exact method is the fastest, taking only 1.72 s, while the Multi_match_boosted method is the slowest, taking 137.46 s. The time is calculated as the sum of the duration required to search all the terms in the benchmark.
The entity linking component, which is an integral part of the model’s output, includes a heuristic decision step. The effect of changing the threshold in the optimized module, as shown in Table 8, when combined with the BK response, reveals changes in performance.
In our benchmark, Breast-Norm-Benchmark, the optimal value for \(\theta _{\text {{opt}}}\) is 0.96. This means that the LLM will only be used if all candidates have less than 96% similarity, or if there are candidates with more than 96% similarity and more than one CUI. Table 9 reports on the results observed by combining medical entity linking and disambiguation when playing prompt engineering over LLMs with symbolic reasoning against the background knowledge. The reported results aim to address RQ2 and RQ3.
The results show that a well-designed prompt and a competent LLM can significantly improve the performance of the medical entity linking component. However, a poor application of LLMs together with inappropriate prompts can drastically reduce the performance of the approach and increase the execution time. This is because the nature of the text in clinical notes differs significantly from the terms in UMLS. Forcing disambiguation on an LLM with a high tendency to hallucinate is worse than just getting the CUIs provided by the medical entity linking component. The best parameters for our case study are \( q_{\text {type}} \): Multi_match_boosted, \( \theta _{\text {sim}} \): 0.96, \( p_{\text {LLM}} \): CoT (GPT4). These parameters were used for the rest of the validation study.
5.2 Main results
We present a comparative analysis of NSSC against scispaCy and BioFalcon, focusing specifically on its performance in terms of accuracy and execution time. The objective is to identify the most effective model for processing oncology clinical notes. The results are presented in Table 10. Accuracy@5 is chosen because no term in the benchmark has more than five possible CUIs.
The NSSC model clearly outperforms the baselines with an exemplary accuracy of 88.34%. This represents a considerable advance over the scispaCy models (en_core_sci_sm and en_core_sci_lg), which both hover around an accuracy of 28.84%, and the en_core_sci_scibert variant at 30.21%. BioFalcon, while superior to scispaCy, still falls short with an accuracy of 55.02%.
The execution time of the NSSC model, at 2502.34 s, is significantly higher than that of its competitors. In stark contrast, the scispaCy models demonstrate exceptional efficiency, with en_core_sci_sm and en_core_sci_lg models completing tasks in approximately 3.65 and 3.72 s, respectively, and the en_core_sci_scibert variant in 12.98 s. BioFalcon also reports an execution time of 154.33 s for the whole dataset.
6 Discussion
The architecture of NSSC, as compared to other approaches in Table 1, is a cancer domain-specific model designed to link terms to a biomedical vocabulary such as UMLS. Since UMLS comprises various vocabularies, NSSC is agnostic to the specific vocabulary being used. Other approaches focus on specific tasks such as NER or EL, while NSSC presents a complete pipeline that transforms clinical note free text into various vocabulary concepts. Other methods, such as those proposed in [54, 68], use BERT embeddings to calculate the similarity between terms and entities. However, NSSC uses the BM25 algorithm as other approaches [19, 50], which is faster, more efficient, and does not require any training data. However, this approach to disambiguate candidates requires the use of LLMs, which involves crafting prompts and results in increased memory and processing time.
NSSC solves the oncology-medicine linking problem with the highest accuracy. However, the execution time increases by almost three orders of magnitude compared to the smallest version of scispaCy. Despite the higher computational cost, one of the NSSC strengths lies in its potential to generalize across a wide range of diseases, not just oncology. The adaptability of BERT-based NER with symbolic reasoning and LLM enables the framework to be extended to different medical domains. Training the model on diverse medical datasets allows it to fine-tune entity identification and linking for different diseases, enhancing its versatility in clinical text processing. This feature makes NSSC a valuable asset for advancing information extraction across the healthcare landscape.
The time complexity of the algorithm is higher than other approaches that do not use LLMs, because models like GPT-4 are significantly more expensive than smaller domain-specific models due to the large number of parameters, which are in the order of billions. However, the approach is optimized to avoid using the LLM in every case, making it more efficient. In addition, the use of the LLM allows the approach to be generalized to other domains.
NSSC uses contextual information as a critical input for understanding disease entities in clinical notes. Different diseases require different levels of context sensitivity; for example, oncology entities often require a nuanced understanding of dependencies, while simpler diseases may benefit from less detailed contextual analysis. Future enhancements should include disease-specific contextual knowledge, allowing the model to adjust its context sensitivity based on the unique characteristics of each disease. This extension will ensure that the approach remains tailored to the nuances of different medical conditions, optimizing performance across a range of diseases.
The use of advanced AI models in healthcare requires careful consideration of ethical implications, particularly patient privacy, consent, and the handling of sensitive medical data. As these models operate within clinical records, strict adherence to privacy standards and ethical guidelines is essential. Transparency in decision-making and ongoing collaboration with healthcare professionals are also critical to building trust in the technology and ensuring its ethical use in healthcare.
7 Limitations
This paper points out some limitations, mainly in the results.
-
The execution time of the NSSC model averages between 5 and 15 s. This timeframe is limited by the response time of the OpenAI API, which can vary. This problem can be solved if the LLMs are executed locally and do not depend on network overhead.
-
The experiments were performed only on a benchmark of more than 400 terms, could escalate significantly if we were to extract all the terms from a real hospital database. Therefore, financial considerations should be taken into account.
-
The use of an NER model before the transition to the linking phase may propagate errors from the initial phase to the subsequent phase. However, state-of-the-art models have remarkably high accuracy metrics that make potential errors manageable and adapt to account for them. In the end, the LLM is responsible for the final decision.
It is important to note that this execution time may make the model unsuitable for real-time applications. However, once a term is linked, there is no need for repetitive mapping, paving the way for the creation of a knowledge base that links terms with their corresponding context. This can be implemented with a cache database that stores the entity, label, and linked terms.
8 Conclusions and future work
The proposed hybrid AI system, NSSC, outperforms existing approaches and thus represents a promising framework to advance the field of information extraction from clinical narratives. Generalization to other diseases, coupled with considerations of training costs and disease-specific contextual information, highlights the potential impact of NSSC on diverse healthcare applications.
Although the benefits of NSSC are significant, it is critical to recognize the computational costs associated with training such advanced models. Training BERT-based NER, reasoning over symbolic systems, and running LLMs typically requires significant computational resources and large annotated datasets. The cost of acquiring and preprocessing such datasets, coupled with the computational demands during training, can be challenging, especially in resource-constrained environments. Future research should focus on reducing training costs and optimizing model architectures for resource-constrained environments. These efforts will improve the accessibility and adoption of our framework. In addition, future research should prioritize improving disease-specific contextual adaptability. In this way, NSSC contributes to ongoing efforts to harness the power of AI to transform medical information extraction for improved patient care and medical research.
Data Availability
The corpus used to validate the entity linking approach, Breast-Norm-Benchmark, is available at https://doi.org/10.5281/zenodo.11185980.
Notes
References
Amaral LAN, Scala A, Barthelemy M, Stanley HE (2000) Classes of small-world networks. Proc Natl Acad Sci 97(21):11149–11152
Spasić I, Livsey J, Keane JA, Nenadić G (2014) Text mining of cancer-related information: review of current status and future directions. Int J Med Inform 83(9):605–623. https://doi.org/10.1016/j.ijmedinf.2014.06.009
Kehl KL, Xu W, Lepisto E, Elmarakeby H, Hassett MJ, Van Allen EM, Johnson BE, Schrag D (2020) Natural language processing to ascertain cancer outcomes from medical oncologist notes. JCO Clin Cancer Inform 4:680–690
Bose P, Srinivasan S, Sleeman WC, Palta J, Kapoor R, Ghosh P (2021) A survey on recent named entity recognition and relationship extraction techniques on clinical texts. Appl Sci (Switzerland) 11(18). https://doi.org/10.3390/app11188319
Zeng Z, Deng Y, Li X, Naumann T, Luo Y (2019) Natural language processing for EHR-based computational phenotyping. IEEE/ACM Trans Comput Biol Bioinform 16(1):139–153. https://doi.org/10.1109/TCBB.2018.2849968arXiv:1806.04820
Zhou Y, Ju C, Caufield JH, Shih K, Chen C, Sun Y, Chang K-W, Ping P, Wang W (2021) Clinical named entity recognition using contextualized token representations. arXiv:2106.12608
Yang X, Mu D, Peng H, Li H, Wang Y, Wang P, Wang Y, Han S et al (2022) Research and application of artificial intelligence based on electronic health records of patients with cancer: systematic review. JMIR Med Inform 10(4):33799
Zhang X, Zhang Y, Zhang Q, Ren Y, Qiu T, Ma J, Sun Q (2019) Extracting comprehensive clinical information for breast cancer using deep learning methods. Int J Med Inform 132(September):103985. https://doi.org/10.1016/j.ijmedinf.2019.103985
Hernandez-Boussard T, Kourdis PD, Seto T, Ferrari M, Blayney DW, Rubin D, Brooks JD (2017) Mining electronic health records to extract patient-centered outcomes following prostate cancer treatment. AMIA. Annual Symposium proceedings. AMIA Symposium 2017:876–882
Solarte-Pabón O, Blazquez-Herranz A, Torrente M, Rodríguez-Gonzalez A, Provencio M, Menasalvas E (2021) Extracting cancer treatments from clinical text written in Spanish: a deep learning approach. In: 2021 IEEE 8th International conference on Data Science and Advanced Analytics (DSAA), pp 1–6. IEEE
Najafabadipour M, Zanin M, Rodríguez-González A, Gonzalo-Martín C, Nuñez García B, Calvo V, Luis Cruz Bermudez J, Provencio M, Menasalvas E (2019) Recognition of time expressions in Spanish electronic health records. In: 2019 IEEE 32nd International symposium on Computer-Based Medical Systems (CBMS), pp 69–74. https://doi.org/10.1109/CBMS.2019.00025
Solarte-Pabón O, Blazquez-Herranz A, Torrente M, Rodríguez-Gonzalez A, Provencio M, Menasalvas E (2021) Extracting cancer treatments from clinical text written in Spanish: a deep learning approach. In: 2021 IEEE 8th International conference on Data Science and Advanced Analytics (DSAA), pp 1–6. https://doi.org/10.1109/DSAA53316.2021.9564137
Solarte-Pabón O, Torrente M, Garcia-Barragán A, Provencio M, Menasalvas E, Robles V (2022) Deep learning to extract breast cancer diagnosis concepts. In: 2022 IEEE 35th International symposium on Computer-Based Medical Systems (CBMS), pp 13–18. https://doi.org/10.1109/CBMS55023.2022.00010
Santiso S, Pérez A, Casillas A, Oronoz M (2020) Neural negated entity recognition in Spanish electronic health records. J Biomed Inform 105 (December 2019):103419. https://doi.org/10.1016/j.jbi.2020.103419
Pabón OS, Montenegro O, Torrente M, González AR, Provencio M, Menasalvas E (2022) Negation and uncertainty detection in clinical texts written in Spanish: a deep learning-based approach. PeerJ Comput Sci 8:913
Miranda-Escalada A, Farré E, Krallinger M (2020) Named entity recognition, concept normalization and clinical coding: overview of the cantemist track for cancer text mining in Spanish, corpus, guidelines, methods and results. IberLEF@ SEPLN, 303–323
Neumann M, King D, Beltagy I, Ammar W (2019) ScispaCy: fast and robust models for biomedical natural language processing. In: Demner-Fushman D, Cohen KB, Ananiadou S, Tsujii J (eds) Proceedings of the 18th BioNLP workshop and shared task, pp 319–327. Association for Computational Linguistics, Florence, Italy. https://doi.org/10.18653/v1/W19-5034https://aclanthology.org/W19-5034
Sakor A, Singh K, Patel A, Vidal M-E (2020) Falcon 2.0: an entity and relation linking tool over wikidata. In: Proceedings of the 29th ACM international conference on information; knowledge management. CIKM ’20. ACM, Online. https://doi.org/10.1145/3340531.3412777
Devlin J, Chang MW, Lee K, Toutanova K (2019) BERT: pre-training of deep bidirectional transformers for language understanding. NAACL HLT 2019 - 2019 conference of the north american chapter of the association for computational linguistics: human language technologies - proceedings of the conference. 1(Mlm):4171–4186. arXiv:1810.04805
García-Barragán A, Solarte-Pabón O, Nedostup G, Provencio M, Menasalvas E, Robles V (2023) Structuring breast cancer Spanish electronic health records using deep learning. In: 2023 IEEE 36th International symposium on Computer-Based Medical Systems (CBMS), pp 404–409. IEEE
Xiao Z, Tong H, Qu R, Xing H, Luo S, Zhu Z, Song F, Feng L (2023) Capmatch: semi-supervised contrastive transformer capsule with feature-based knowledge distillation for human activity recognition. IEEE Trans Neural Netw Learn Syst 1–15. https://doi.org/10.1109/TNNLS.2023.3344294
Xiao Z, Xu X, Xing H, Zhao B, Wang X, Song F, Qu R, Feng L (2024) DTCM: deep transformer capsule mutual distillation for multivariate time series classification. IEEE Trans Cognit Dev Syst 16(4):1445–1461. https://doi.org/10.1109/TCDS.2024.3370219
Li J, Sun A, Han J, Li C (2020) A survey on deep learning for named entity recognition. IEEE Trans Knowl Data Eng 34(1):50–70
Luo Y, Xiao F, Zhao H (2020) Hierarchical contextualized representation for named entity recognition. In: Proceedings of the AAAI conference on artificial intelligence 34:8441–8448
Baevski A, Edunov S, Liu Y, Zettlemoyer L, Auli M (2019) Cloze-driven pretraining of self-attention networks. In: Inui K, Jiang J, Ng V, Wan X (eds) Proceedings of the 2019 conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp 5360–5369. Association for Computational Linguistics, Hong Kong, China. https://doi.org/10.18653/v1/D19-1539https://aclanthology.org/D19-1539
Jiang Y, Hu C, Xiao T, Zhang C, Zhu J (2019) Improved differentiable architecture search for language modeling and named entity recognition. In: Inui K, Jiang J, Ng V, Wan X (eds) Proceedings of the 2019 conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp 3585–3590. Asso-ciation for Computational Linguistics, Hong Kong, China. https://doi.org/10.18653/v1/D19-1367https://aclanthology.org/D19-1367
Lison P, Barnes J, Hubin A, Touileb S (2020) Named entity recognition without labelled data: a weak supervision approach. In: Jurafsky D, Chai J, Schluter N, Tetreault J (eds) Proceedings of the 58th annual meeting of the association for computational linguistics, pp 1518–1533. Association for Computational Linguistics, Online. https://doi.org/10.18653/v1/2020.acl-main.139https://aclanthology.org/2020.acl-main.139
Kim Y, Lee JH, Choi S, Lee JM, Kim J-H, Seok J, Joo HJ (2020) Validation of deep learning natural language processing algorithm for keyword extraction from pathology reports in electronic health records. Sci Rep 10(1):1–9
Martina S, Ventura L, Frasconi P (2020) Classification of cancer pathology reports: a large-scale comparative study. IEEE J Biomed Health Inform 24(11):3085–3094. https://doi.org/10.1109/JBHI.2020.3005016arXiv:2006.16370
Bitterman D, Chen Lin H, Finan S, Warner J, Mak R, Savova G (2020) Extracting radiotherapy treatment details using neural network-based natural language processing. In: Annual meeting of the american society for radiation oncology, Cham
Solarte-Pabón O, Montenegro O, García-Barragán A, Torrente M, Provencio M, Menasalvas E, Robles V (2023) Transformers for extracting breast cancer information from Spanish clinical narratives. Artif Intell Med 143:102625
Sevgili Ö, Shelmanov A, Arkhipov M, Panchenko A, Biemann C (2022) Neural entity linking: a survey of models based on deep learning. Semantic Web 13(3):527–570
Poerner N, Waltinger U, Schütze H (2020) E-BERT: efficient-yet-effective entity embeddings for BERT. In: Cohn T, He Y, Liu Y (eds) Findings of the association for computational linguistics: EMNLP 2020, pp 803–818. Association for Computational Linguistics, Online. https://doi.org/10.18653/v1/2020.findings-emnlp.71https://aclanthology.org/2020.findings-emnlp.71
Cao ND, Izacard G, Riedel S, Petroni F (2021) Autoregressive entity retrieval. In: International conference on learning representations. https://openreview.net/forum?id=5k8F6UU39V
Usbeck R, Ngonga Ngomo A-C, Auer S, Gerber D, Both A, Coelho S (2014). AGDISTIS - graph-based disambiguation of named entities using linked data. https://doi.org/10.1007/978-3-319-11964-9_29
Onando M, Singh K, Vyas A, Shekarpour S, Vidal M-E, Auer S (2020) Encoding knowledge graph entity aliases in attentive neural network for wikidata entity linking, pp 328–342. https://doi.org/10.1007/978-3-030-62005-9_24
Cao ND, Izacard G, Riedel S, Petroni F (2021) Autoregressive entity retrieval
Ayoola T, Tyagi S, Fisher J, Christodoulopoulos C, Pierleoni A (2022) ReFinED: an efficient zero-shot-capable approach to end-to-end entity linking
Le P, Titov I (2019) Distant learning for entity linking with automatic noise detection
Logeswaran L, Chang M-W, Lee K, Toutanova K, Devlin J, Lee H (2019) Zero-shot entity linking by reading entity descriptions
Wu L, Petroni F, Josifoski M, Riedel S, Zettlemoyer L (2020) Scalable zero-shot entity linking with dense entity retrieval
Hitzler P, Eberhart A, Ebrahimi M, Sarker MK, Zhou L (2022) Neuro-symbolic approaches in artificial intelligence. Nat Sci Rev 9(6):035. https://doi.org/10.1093/nsr/nwac035https://academic.oup.com/nsr/article-pdf/9/6/nwac035/43952953/nwac035.pdf
Sakor A, Mulang IO, Singh K, Shekarpour S, Vidal ME, Lehmann J, Auer S (2019) Old is gold: linguistic driven approach for entity and relation linking of short text. In: Proceedings of the 2019 conference of the north american chapter of the association for computational linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp 2336–2346
Auer S, Bizer C, Kobilarov G, Lehmann J, Cyganiak R, Ives Z (2007) DBpedia: a nucleus for a web of open data. In: Aberer K, Choi K-S, Noy N, Allemang D, Lee K-I, Nixon L, Golbeck J, Mika P, Maynard D, Mizoguchi R, Schreiber G, Cudré-Mauroux P (eds) The Semantic Web. Springer, Berlin, Heidelberg, pp 722–735
Vrandečić D (2012) Wikidata: a new platform for collaborative data collection. WWW ’12 Companion, pp 1063–1064. Association for Computing Machinery, New York, NY, USA. https://doi.org/10.1145/2187980.2188242
Delpeuch A (2020) OpenTapioca: lightweight entity linking for wikidata
Jiang H, Gurajada S, Lu Q, Neelam S, Popa L, Sen P, Li Y, Gray A (2021) LNN-EL: a neuro-symbolic approach to short-text entity linking. In: Zong C, Xia F, Li W, Navigli R (eds) Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (Volume 1: Long Papers), pp 775–787. Association for Computational Linguistics, Online. https://doi.org/10.18653/v1/2021.acl-long.64https://aclanthology.org/2021.acl-long.64
Plu J, Rizzo G, Troncy R (2015) A hybrid approach for entity recognition and linking. In: Semantic web evaluation challenges: second semwebeval challenge at ESWC 2015, Portorož, Slovenia, May 31-June 4, 2015, Revised Selected Papers, pp 28–39. Springer
Ji Z, Wei Q, Xu H (2020) BERT-based ranking for biomedical entity normalization. AMIA summits on translational science proceedings 269
Kalyan KS, Sangeetha S (2021) BertMCN: mapping colloquial phrases to standard medical concepts using BERT and highway network. Artif Intell Med 102008
Kalyan KS, Sangeetha S (2020) Medical concept normalization in user-generated texts by learning target concept embeddings. In: Holderness E, Jimeno Yepes A, Lavelli A, Minard A-L, Pustejovsky J, Rinaldi F (eds) Proceedings of the 11th international workshop on health text mining and information analysis, pp 18–23. Association for Computational Linguistics, Online. https://doi.org/10.18653/v1/2020.louhi-1.3https://aclanthology.org/2020.louhi-1.3
Pattisapu N, Patil S, Palshikar G, Varma V (2020) Medical concept normalization by encoding target knowledge. In: Machine learning for health workshop, pp 246–259. PMLR
Gallego F, López-García G, Gasco-Sánchez L, Krallinger M, Veredas FJ (2024) ClinLinker: medical entity linking of clinical concept mentions in Spanish. In: Franco L, Mulatier C, Paszynski M, Krzhizhanovskaya VV, Dongarra JJ, Sloot PMA (eds) Computational Science – ICCS 2024. Springer, Cham, pp 266–280
Duan S, Guang Y, Bu W, Yang J (2023) A survey of named entity disambiguation in entity linking. In: 2023 3rd International conference on Intelligent Communications and Computing (ICC), pp 296–303. https://doi.org/10.1109/ICC59986.2023.10421092
Bunescu R, Pasca M (2006) Using encyclopedic knowledge for named entity disambiguation. In: 11th Conference of the european chapter of the association for computational linguistics, pp 9–16
Liu S, Fang Y (2023) Use large language models for named entity disambiguation in academic knowledge graphs. In: 2023 3rd International conference on Education, Information Management and Service Science (EIMSS 2023), pp 681–691. Atlantis Press
Kafando R, Decoupes R, Roche M, Teisseire M (2023) SNEToolkit: spatial named entities disambiguation toolkit. SoftwareX. 23:101480
Bouarroudj W, Boufaida Z, Bellatreche L (2022) Named entity disambiguation in short texts over knowledge graphs. Knowl Inf Syst 64(2):325–351
Varma M, Orr L, Wu S, Leszczynski M, Ling X, Ré C (2021) Cross-domain data integration for named entity disambiguation in biomedical text. In: Moens M-F, Huang X, Specia L, Yih SW-t (eds) Findings of the Association for Computational Linguistics: EMNLP 2021, pp 4566–4575. Association for Computational Linguistics, Punta Cana, Dominican Republic. https://doi.org/10.18653/v1/2021.findings-emnlp.388https://aclanthology.org/2021.findings-emnlp.388
Wang X, Tsujii J, Ananiadou S (2009) Classifying relations for biomedical named entity disambiguation. In: Proceedings of the 2009 conference on empirical methods in natural language processing, pp 1513–1522
Vretinaris A, Lei C, Efthymiou V, Qin X, Özcan F (2021) Medical entity disambiguation using graph neural networks. Proceedings of the 2021 international conference on management of data. https://doi.org/10.1145/3448016.3457328
Garda S, Leser U (2024) BELHD: improving biomedical entity linking with homonoym disambiguation
Angell R, Monath N, Mohan S, Yadav N, McCallum A (2021) Clustering-based inference for biomedical entity linking
Sung M, Jeon H, Lee J, Kang J (2020) Biomedical entity representations with synonym marginalization. In: Jurafsky D, Chai J, Schluter N, Tetreault J (eds) Proceedings of the 58th annual meeting of the association for computational linguistics, pp 3641–3650. Association for Computational Linguistics, Online. https://doi.org/10.18653/v1/2020.acl-main.335.https://aclanthology.org/2020.acl-main.335
Logeswaran L, Chang M-W, Lee K, Toutanova K, Devlin J, Lee H (2019) Zero-shot entity linking by reading entity descriptions. In: Proceedings of the 57th annual meeting of the association for computational linguistics
Chen L, Varoquaux G, Suchanek FM (2021) A lightweight neural model for biomedical entity linking. Proc AAAI Conf Artif Intell 35(14):12657–12665. https://doi.org/10.1609/aaai.v35i14.17499
Chen S, Wang J, Jiang F, Lin C-Y (2020) Improving entity linking by modeling latent entity type information. Proceedings of the AAAI conference on artificial intelligence 34:7529–7537
García-Barragán A (2024). Breast-Norm-Benchmark. https://doi.org/10.5281/zenodo.12703934
Bekkum M, Boer M, Harmelen F, Meyer-Vitali A, Teije A (2021) Modular design patterns for hybrid learning and reasoning systems. Appl Intell 51(9)
Robertson S, Zaragoza H et al (2009) The probabilistic relevance framework: BM25 and beyond. Foundations and Trends® in Information Retrieval 3(4):333–389
Wei J, Wang X, Schuurmans D, Bosma M, Ichter B, Xia F, Chi E, Le Q, Zhou D (2023) Chain-of-thought prompting elicits reasoning in large language models
Doan S, Conway M, Phuong TM, Ohno-Machado L (2014) Natural language processing in biomedicine: a unified system architecture overview, pp 275–294. Springer, New York, NY. https://doi.org/10.1007/978-1-4939-0847-9_16
Labrak Y, Rouvier M, Dufour R (2023) A zero-shot and few-shot study of instruction-finetuned large language models applied to clinical and biomedical tasks
Funding
This article is supported by the European Union’s Horizon 2020 research and innovation program under grant agreement no. 875160, project CLARIFY. Maria-Esther Vidal is partially funded by Leibniz Association, program “Leibniz Best Minds: Program for Women Professors,” project TrustKG-Transforming Data in Trustable Insights; Grant P99/2020.
Author information
Authors and Affiliations
Corresponding authors
Ethics declarations
Conflict of interest
The authors declare no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix A. Prompt costs
Appendix A. Prompt costs
We use models gpt-3.5-turbo and gpt-4 though Microsoft Azure. To establish a connection with the OpenAI API, we utilize the Azure endpoint. The price of this API depends on the model, and there are different prices for the input tokens and the output tokens, as detailed in https://azure.microsoft.com/en-us/pricing/details/cognitive-services/openai-service/.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
García-Barragán, Á., Sakor, A., Vidal, ME. et al. NSSC: a neuro-symbolic AI system for enhancing accuracy of named entity recognition and linking from oncologic clinical notes. Med Biol Eng Comput 63, 749–772 (2025). https://doi.org/10.1007/s11517-024-03227-4
Received:
Accepted:
Published:
Version of record:
Issue date:
DOI: https://doi.org/10.1007/s11517-024-03227-4









