Named Entity Recognition (NER) is a fundamental task in natural language processing that involves identifying and classifying entities such as names, locations, and organizations within text. Although public NER resources such as annotated datasets and annotation services exist across various domains, no single resource typically supports all entity types required for specific downstream applications. Additionally, the availability of training data to effectively develop NER systems for different domain classification schemas is often limited due to constraints on time, quality, and annotation costs. In this paper, we propose to address these issues by a transfer learning approach, validating the hypothesis that with limited resources, the target domain labels and their distributions can be learned by exploiting features obtained from the source domain. The proposed approach achieves comparable performance with respect to fine-tuning state-of-the-art transformer-based models, using a limited amount of resources in terms of GPU, CPU and RAM.
Balducci, G., Fersini, E., Messina, E. (2026). Cross-Domain Named Entity Recognition: A Resource-Efficient Transfer Learning Approach. In Natural Language Processing and Information Systems 30th International Conference on Applications of Natural Language to Information Systems, NLDB 2025, Kanazawa, Japan, July 4–6, 2025, Proceedings, Part II (pp.105-115). Springer Science and Business Media Deutschland GmbH [10.1007/978-3-031-97144-0_10].
Cross-Domain Named Entity Recognition: A Resource-Efficient Transfer Learning Approach
Balducci, G
;Fersini, E;Messina, E
2026
Abstract
Named Entity Recognition (NER) is a fundamental task in natural language processing that involves identifying and classifying entities such as names, locations, and organizations within text. Although public NER resources such as annotated datasets and annotation services exist across various domains, no single resource typically supports all entity types required for specific downstream applications. Additionally, the availability of training data to effectively develop NER systems for different domain classification schemas is often limited due to constraints on time, quality, and annotation costs. In this paper, we propose to address these issues by a transfer learning approach, validating the hypothesis that with limited resources, the target domain labels and their distributions can be learned by exploiting features obtained from the source domain. The proposed approach achieves comparable performance with respect to fine-tuning state-of-the-art transformer-based models, using a limited amount of resources in terms of GPU, CPU and RAM.| File | Dimensione | Formato | |
|---|---|---|---|
|
Balducci-2026-NLDB 2025-AAM.pdf
embargo fino al 01/07/2026
Tipologia di allegato:
Author’s Accepted Manuscript, AAM (Post-print)
Licenza:
Licenza open access specifica dell’editore
Dimensione
393.66 kB
Formato
Adobe PDF
|
393.66 kB | Adobe PDF | Visualizza/Apri Richiedi una copia |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.


