Repository logo
 

Prompt-based Few-Shot Text Classification With Multi-Granularity Label Augmentation and Adaptive Verbalizer

aut.relation.articlenumber58
aut.relation.endpage58
aut.relation.issue1
aut.relation.journalInformation
aut.relation.startpage58
aut.relation.volume17
dc.contributor.authorHuang, Deling
dc.contributor.authorLi, Zanxiong
dc.contributor.authorYu, Jian
dc.contributor.authorZhou, Yulong
dc.date.accessioned2026-01-14T21:13:58Z
dc.date.available2026-01-14T21:13:58Z
dc.date.issued2026-01-08
dc.description.abstract<jats:p>Few-Shot Text Classification (FSTC) aims to classify text accurately into predefined categories using minimal training samples. Recently, prompt-tuning-based methods have achieved promising results by constructing verbalizers that map input data to the label space, thereby maximizing the utilization of pre-trained model features. However, existing verbalizer construction methods often rely on external knowledge bases, which require complex noise filtering and manual refinement, making the process time-consuming and labor-intensive, while approaches based on pre-trained language models (PLMs) frequently overlook inherent prediction biases. Furthermore, conventional data augmentation methods focus on modifying input instances while overlooking the integral role of label semantics in prompt tuning. This disconnection often leads to a trade-off where increased sample diversity comes at the cost of semantic consistency, resulting in marginal improvements. To address these limitations, this paper first proposes a novel Bayesian Mutual Information-based method that optimizes label mapping to retain general PLM features while reducing reliance on irrelevant or unfair attributes to mitigate latent biases. Based on this method, we propose two synergistic generators that synthesize semantically consistent samples by integrating label word information from the verbalizer to effectively enrich data distribution and alleviate sparsity. To guarantee the reliability of the augmented set, we propose a Low-Entropy Selector that serves as a semantic filter, retaining only high-confidence samples to safeguard the model against ambiguous supervision signals. Furthermore, we propose a Difficulty-Aware Adversarial Training framework that fosters generalized feature learning, enabling the model to withstand subtle input perturbations. Extensive experiments demonstrate that our approach outperforms state-of-the-art methods on most few-shot and full-data splits, with F1 score improvements of up to +2.8% on the standard AG’s News benchmark and +1.0% on the challenging DBPedia benchmark.</jats:p>
dc.identifier.citationInformation, ISSN: 2078-2489 (Print); 2078-2489 (Online), MDPI AG, 17(1), 58-58. doi: 10.3390/info17010058
dc.identifier.doi10.3390/info17010058
dc.identifier.issn2078-2489
dc.identifier.issn2078-2489
dc.identifier.urihttp://hdl.handle.net/10292/20495
dc.languageen
dc.publisherMDPI AG
dc.relation.urihttps://www.mdpi.com/2078-2489/17/1/58
dc.rights© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
dc.rights.accessrightsOpenAccess
dc.subject08 Information and Computing Sciences
dc.subject46 Information and computing sciences
dc.subjectfew-shot learning
dc.subjecttext classification
dc.subjectprompt tuning
dc.subjectdata augmentation
dc.subjectadversarial training
dc.titlePrompt-based Few-Shot Text Classification With Multi-Granularity Label Augmentation and Adaptive Verbalizer
dc.typeJournal Article
pubs.elements-id750755

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Huang et al_2026_Prompt based few shot text classification.pdf
Size:
1.05 MB
Format:
Adobe Portable Document Format
Description:
Journal article

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
1.37 KB
Format:
Plain Text
Description: