Comparing and Improving Active Learning Uncertainty Measures for Transformer Models by Discarding Outliers

Research output: Contribution to journalResearch articleContributedpeer-review

Contributors

Abstract

Despite achieving state-of-the-art results in nearly all Natural Language Processing applications, fine-tuning Transformer-encoder based language models still requires a significant amount of labeled data to achieve satisfying work. A well known technique to reduce the amount of human effort in acquiring a labeled dataset is Active Learning (AL): an iterative process in which only the minimal amount of samples is labeled. AL strategies require access to a quantified confidence measure of the model predictions. A common choice is the softmax activation function for the final Neural Network layer. In this paper, we compare eight alternatives on seven datasets and show that the softmax function provides misleading probabilities. Our finding is that most of the methods primarily identify hard-to-learn-from samples (commonly called outliers), resulting in worse than random performance, instead of samples, which actually reduce the uncertainty of the learned language model. As a solution, this paper proposes Uncertainty-Clipping, a heuristic to systematically exclude samples, which results in improvements for most methods compared to the softmax function.

Details

Original languageEnglish
JournalInformation systems frontiers
Publication statusE-pub ahead of print - 26 Jun 2024
Peer-reviewedYes

External IDs

ORCID /0000-0001-8107-2775/work/174431839
ORCID /0000-0002-5985-4348/work/174432432

Keywords

Keywords

  • Active Learning, Calibration, Deep Neural Networks, Softmax, Transformer, Uncertainty