A Study of Errors in the Output of Large Language Models for Domain-Specific Few-Shot Named Entity Recognition
DOI:
https://doi.org/10.21248/jlcl.38.2025.281Keywords:
large language models, few-shot named entity recognition, error analysisAbstract
This paper proposes an error classification framework for a comprehensive analysis of the output that large language models (LLMs) generate in a few-shot named entity recognition (NER) task in a specialised domain. The framework should be seen as an exploratory analysis complementary to established performance metrics for NER classifiers, such as F1 score, as it accounts for outcomes possible in a few-shot, LLMbased NER task. By categorising and assessing incorrect named entity predictions quantitatively, the paper shows how the proposed error classification could support a deeper cross-model and cross-prompt performance comparison, alongside a roadmap for a guided qualitative error analysis.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2025 Elena Volkanovska

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.