Active Learning on Pre-trained Language Model with Task-Independent Triplet Loss

Seungmin Seo, Donghyun Kim, Youbin Ahn, Kyong Ho Lee

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Active learning attempts to maximize a task model's performance gain by obtaining a set of informative samples from an unlabeled data pool. Previous active learning methods usually rely on specific network architectures or task-dependent sample acquisition algorithms. Moreover, when selecting a batch sample, previous works suffer from insufficient diversity of batch samples because they only consider the informativeness of each sample. This paper proposes a task-independent batch acquisition method using triplet loss to distinguish hard samples in an unlabeled data pool with similar features but difficult to identify labels. To assess the effectiveness of the proposed method, we compare the proposed method with state-of-the-art active learning methods on two tasks, relation extraction and sentence classification. Experimental results show that our method outperforms baselines on the benchmark datasets.

Original languageEnglish
Title of host publicationAAAI-22 Technical Tracks 10
PublisherAssociation for the Advancement of Artificial Intelligence
Pages11276-11284
Number of pages9
ISBN (Electronic)1577358767, 9781577358763
Publication statusPublished - 2022 Jun 30
Event36th AAAI Conference on Artificial Intelligence, AAAI 2022 - Virtual, Online
Duration: 2022 Feb 222022 Mar 1

Publication series

NameProceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022
Volume36

Conference

Conference36th AAAI Conference on Artificial Intelligence, AAAI 2022
CityVirtual, Online
Period22/2/2222/3/1

Bibliographical note

Funding Information:
This research was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIP; Ministry of Science, ICT & Future Planning) (No. NRF-2019R1A2B5B01070555). Kyong-Ho Lee is the corresponding author.

Publisher Copyright:
Copyright © 2022, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.

All Science Journal Classification (ASJC) codes

  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Active Learning on Pre-trained Language Model with Task-Independent Triplet Loss'. Together they form a unique fingerprint.

Cite this