Resilience of Named Entity Recognition models against adversarial attacks
Salvato in:
| Pubblicato in: | International Journal of Electronics and Telecommunications vol. 71, no. 3 (2025), p. 1-7 |
|---|---|
| Autore principale: | |
| Pubblicazione: |
Polish Academy of Sciences
|
| Soggetti: | |
| Accesso online: | Citation/Abstract Full Text - PDF |
| Tags: |
Nessun Tag, puoi essere il primo ad aggiungerne!!
|
| Abstract: | Adversarial Attacks are actions that aims to mislead models by introducing subtle and often imperceptible changes in model’s input. Providing resilience for such kind of risk is key for all Natural Language Processing (NLP) task specific models. Current state of the art solution for one of NLP task Named Entity Recognition (NER) is usage of transformer based solutions. Previous solution where based on Conditional Random Fields (CRF).This research aims to investigate and compare the robustness of both transformer-based and CRF-based NER models against adversarial attacks. By subjecting these models to carefully crafted perturbations, we seek to understand how well they can withstand attempts to manipulate their input and compromise their performance. This comparative analysis will provide valuable insights into the strengths and weaknesses of each architecture, shedding light on the most effective strategies for enhancing the security and reliability of NER systems. |
|---|---|
| ISSN: | 2081-8491 2300-1933 0035-9386 0867-6747 |
| DOI: | 10.24425/ijet.2025.153623 |
| Fonte: | Advanced Technologies & Aerospace Database |