Robust fine-tuning for low-resource NLP: Combining adversarial and metric-based learning to mitigate overfitting

Kyuri Choi, Youngjoong Ko

Research output: Contribution to journalArticlepeer-review

Abstract

Learning from low-resource training samples is challenging since the model can memorize features that are irrelevant to the given task, commonly known as overfitting. As neural networks grow in size due to their enhanced effectiveness, they increasingly face the risk of overfitting, especially when trained on limited data. This situation where larger models are both more capable but more prone to overfitting necessitates novel strategies to maintain generalization. Existing solutions often rely on preserving pre-trained model weights to prevent overfitting while harnessing rich information from the pre-training phase. However, these approaches encounter performance trade-offs between in-distribution and out-of-distribution and lack analysis of how pre-trained language models (PLMs) overfit training data. In this work, we analyze the tendency of PLMs to overfit salient features within the constrained data distribution, especially domain-specific features. Motivated by this observation, we propose a training method that reduces the influence of domain information in the embedding space to prevent overfitting on specific feature when working with low-resource samples. Our approach demonstrates promising improvements in diverse out-of-distribution settings while maintaining comparable performance on in-distribution test sets.

Original languageEnglish
Article number127737
JournalExpert Systems with Applications
Volume288
DOIs
StatePublished - 1 Sep 2025

Keywords

  • Adversarial learning
  • Low-resource classification
  • Metric-based learning
  • Overfitting
  • Pre-trained language models

Fingerprint

Dive into the research topics of 'Robust fine-tuning for low-resource NLP: Combining adversarial and metric-based learning to mitigate overfitting'. Together they form a unique fingerprint.

Cite this