Large Language Models (LLMs) are increasingly deployed in real-world applications, raising urgent concerns around their safety, reliability, and ethical behavior. While existing safety evaluations have primarily focused on English, low- and mid-resource languages such as Italian remain critically underexplored. In this paper, we present the first comprehensive and multidimensional evaluation of LLM safety in the Italian language. We assess seven state-of-the-art LLMs across key safety dimensions using several automatic moderators tailored to cover the Italian settings. Furthermore, we analyze the challenges of adapting English-centric safety benchmarks to Italian via machine translation, highlighting limitations and proposing best practices for developing culturally and linguistically grounded evaluation frameworks.

Rizzi, G., Magazzù, G., Sormani, A., Pulerà, F., Scalena, D., Fersini, E. (2025). Uncovering Unsafety Traits in Italian Language Models. In Proceedings of the Eleventh Italian Conference on Computational Linguistics (CLiC-it 2025) (pp.974-982). CEUR Workshop Proceedings.

Uncovering Unsafety Traits in Italian Language Models

Rizzi, G;Magazzù, G;Pulerà, F;Scalena, D;Fersini, E
2025

Abstract

Large Language Models (LLMs) are increasingly deployed in real-world applications, raising urgent concerns around their safety, reliability, and ethical behavior. While existing safety evaluations have primarily focused on English, low- and mid-resource languages such as Italian remain critically underexplored. In this paper, we present the first comprehensive and multidimensional evaluation of LLM safety in the Italian language. We assess seven state-of-the-art LLMs across key safety dimensions using several automatic moderators tailored to cover the Italian settings. Furthermore, we analyze the challenges of adapting English-centric safety benchmarks to Italian via machine translation, highlighting limitations and proposing best practices for developing culturally and linguistically grounded evaluation frameworks.
paper
Italian Language; Large Language Models (LLMs); Safety Evaluation;
English
Eleventh Italian Conference on Computational Linguistics (CLiC-it 2025) - September 24-26, 2025
2025
Bosco, C; Jezek, E; Polignano, M; Sanguinetti, E
Proceedings of the Eleventh Italian Conference on Computational Linguistics (CLiC-it 2025)
9791224305873
2025
4112
974
982
https://ceur-ws.org/Vol-4112/
open
Rizzi, G., Magazzù, G., Sormani, A., Pulerà, F., Scalena, D., Fersini, E. (2025). Uncovering Unsafety Traits in Italian Language Models. In Proceedings of the Eleventh Italian Conference on Computational Linguistics (CLiC-it 2025) (pp.974-982). CEUR Workshop Proceedings.
File in questo prodotto:
File Dimensione Formato  
2025.clicit-1.91.pdf

accesso aperto

Descrizione: CEUR WORKSHOP PROCEEDINGS
Tipologia di allegato: Publisher’s Version (Version of Record, VoR)
Licenza: Creative Commons
Dimensione 1.2 MB
Formato Adobe PDF
1.2 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/10281/592501
Citazioni
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
Social impact