A novel prompt-tuning method: Incorporating scenario-specific concepts into a verbalizer

The verbalizer, which serves to map label words to class labels, is an essential component of prompt-tuning. In this paper, we present a novel approach to constructing verbalizers. While existing methods for verbalizer construction mainly rely on augmenting and refining sets of synonyms or related w...

Full description

Saved in:
Bibliographic Details
Published inExpert systems with applications Vol. 247; p. 123204
Main Authors Ma, Yong, Luo, Senlin, Shang, Yu-Ming, Li, Zhengjun, Liu, Yong
Format Journal Article
LanguageEnglish
Published Elsevier Ltd 01.08.2024
Subjects
Online AccessGet full text

Cover

Loading…
Abstract The verbalizer, which serves to map label words to class labels, is an essential component of prompt-tuning. In this paper, we present a novel approach to constructing verbalizers. While existing methods for verbalizer construction mainly rely on augmenting and refining sets of synonyms or related words based on class names, this paradigm suffers from a narrow perspective and lack of abstraction, resulting in limited coverage and high bias in the label-word space. To address this issue, we propose a label-word construction process that incorporates scenario-specific concepts. Specifically, we extract rich concepts from task-specific scenarios as label-word candidates and then develop a novel cascade calibration module to refine the candidates into a set of label words for each class. We evaluate the effectiveness of our proposed approach through extensive experiments on five widely used datasets for zero-shot text classification. The results demonstrate that our method outperforms existing methods and achieves state-of-the-art results. •Retrieving scenario-specific concepts as label word candidates.•Proposing a novel class-name-free cascade approach for label-word refining.•Constructing a verbalizer that has wide-ranging coverage and minimal bias.•Reporting fresh state-of-the-art results.
AbstractList The verbalizer, which serves to map label words to class labels, is an essential component of prompt-tuning. In this paper, we present a novel approach to constructing verbalizers. While existing methods for verbalizer construction mainly rely on augmenting and refining sets of synonyms or related words based on class names, this paradigm suffers from a narrow perspective and lack of abstraction, resulting in limited coverage and high bias in the label-word space. To address this issue, we propose a label-word construction process that incorporates scenario-specific concepts. Specifically, we extract rich concepts from task-specific scenarios as label-word candidates and then develop a novel cascade calibration module to refine the candidates into a set of label words for each class. We evaluate the effectiveness of our proposed approach through extensive experiments on five widely used datasets for zero-shot text classification. The results demonstrate that our method outperforms existing methods and achieves state-of-the-art results. •Retrieving scenario-specific concepts as label word candidates.•Proposing a novel class-name-free cascade approach for label-word refining.•Constructing a verbalizer that has wide-ranging coverage and minimal bias.•Reporting fresh state-of-the-art results.
ArticleNumber 123204
Author Shang, Yu-Ming
Liu, Yong
Li, Zhengjun
Luo, Senlin
Ma, Yong
Author_xml – sequence: 1
  givenname: Yong
  orcidid: 0000-0003-1400-534X
  surname: Ma
  fullname: Ma, Yong
  email: znsoft@163.com
  organization: School of Information and Electronics, Beijing Institute of Technology, Haidian District, Beijing, China
– sequence: 2
  givenname: Senlin
  surname: Luo
  fullname: Luo, Senlin
  email: luosenlin2019@126.com
  organization: School of Information and Electronics, Beijing Institute of Technology, Haidian District, Beijing, China
– sequence: 3
  givenname: Yu-Ming
  surname: Shang
  fullname: Shang, Yu-Ming
  email: shangym@bupt.edu.cn
  organization: School of Cyberspace Security, Beijing University of Posts and Telecommunications, Haidian District, Beijing, China
– sequence: 4
  givenname: Zhengjun
  orcidid: 0000-0001-9518-5391
  surname: Li
  fullname: Li, Zhengjun
  email: zjli@bit.edu.cn
  organization: School of Information and Electronics, Beijing Institute of Technology, Haidian District, Beijing, China
– sequence: 5
  givenname: Yong
  surname: Liu
  fullname: Liu, Yong
  email: liuyong03@qianxin.com
  organization: Qi-AnXin Technology Group, QAX Security Center, Xicheng District, Beijing, China
BookMark eNp9kM1KAzEUhYNUsK2-gKu8wIz5m2Yibkrxp1Bwo-AuZJI7mtImQxIr-vR2qGtXFw58h3O_GZqEGACha0pqSujiZltD_jI1I0zUlHFGxBma0lbyaiEVn6ApUY2sBJXiAs1y3hJCJSFyit6WOMQD7PCQ4n4oVfkMPrzjPZSP6G7xOtiYhphMGdNsIZjkY5UHsL73FtsYLAwlYx9KxAYfIHVm538gXaLz3uwyXP3dOXp9uH9ZPVWb58f1armpLGtoqXrn-oYLIxrVUaOsc7SRCwoCOioZACGiNxyEpaxtXdMTAMV516pGWaWU4XPETr02xZwT9HpIfm_St6ZEj270Vo9u9OhGn9wcobsTBMdlBw9JZ-vh-IrzCWzRLvr_8F-Mg3Dn
Cites_doi 10.1162/dint_a_00013
10.1162/tacl_a_00324
10.3233/SW-140134
ContentType Journal Article
Copyright 2024 Elsevier Ltd
Copyright_xml – notice: 2024 Elsevier Ltd
DBID AAYXX
CITATION
DOI 10.1016/j.eswa.2024.123204
DatabaseName CrossRef
DatabaseTitle CrossRef
DatabaseTitleList
DeliveryMethod fulltext_linktorsrc
Discipline Computer Science
EISSN 1873-6793
ExternalDocumentID 10_1016_j_eswa_2024_123204
S0957417424000691
GroupedDBID --K
--M
.DC
.~1
0R~
13V
1B1
1RT
1~.
1~5
4.4
457
4G.
5GY
5VS
7-5
71M
8P~
9JN
9JO
AAAKF
AABNK
AACTN
AAEDT
AAEDW
AAIAV
AAIKJ
AAKOC
AALRI
AAOAW
AAQFI
AARIN
AAXUO
AAYFN
ABBOA
ABFNM
ABMAC
ABMVD
ABUCO
ABYKQ
ACDAQ
ACGFS
ACHRH
ACNTT
ACRLP
ACZNC
ADBBV
ADEZE
ADTZH
AEBSH
AECPX
AEKER
AENEX
AFKWA
AFTJW
AGHFR
AGJBL
AGUBO
AGUMN
AGYEJ
AHHHB
AHJVU
AHZHX
AIALX
AIEXJ
AIKHN
AITUG
AJOXV
AKRWK
ALEQD
ALMA_UNASSIGNED_HOLDINGS
AMFUW
AMRAJ
AOUOD
APLSM
AXJTR
BJAXD
BKOJK
BLXMC
BNSAS
CS3
DU5
EBS
EFJIC
EO8
EO9
EP2
EP3
F5P
FDB
FIRID
FNPLU
FYGXN
G-Q
GBLVA
GBOLZ
HAMUX
IHE
J1W
JJJVA
KOM
MO0
N9A
O-L
O9-
OAUVE
OZT
P-8
P-9
P2P
PC.
PQQKQ
Q38
RIG
ROL
RPZ
SDF
SDG
SDP
SDS
SES
SEW
SPC
SPCBC
SSB
SSD
SSL
SST
SSV
SSZ
T5K
TN5
~G-
29G
AAAKG
AAQXK
AAXKI
AAYXX
ABKBG
ABXDB
ACNNM
ADJOM
ADMUD
ASPBG
AVWKF
AZFZN
CITATION
EJD
FEDTE
FGOYB
G-2
HLZ
HVGLF
HZ~
LG9
LY1
LY7
M41
R2-
SBC
SET
WUQ
XPP
ZMT
ID FETCH-LOGICAL-c251t-fddf534a459b1a9cdd15761e4eb172ee004fa3e4c1288d5f0ee933b8959c999a3
IEDL.DBID .~1
ISSN 0957-4174
IngestDate Thu Sep 12 16:54:03 EDT 2024
Sat Apr 13 16:37:26 EDT 2024
IsPeerReviewed true
IsScholarly true
Keywords Zero-shot
Prompt learning
Verbalizer construction
Text classification
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-c251t-fddf534a459b1a9cdd15761e4eb172ee004fa3e4c1288d5f0ee933b8959c999a3
ORCID 0000-0003-1400-534X
0000-0001-9518-5391
ParticipantIDs crossref_primary_10_1016_j_eswa_2024_123204
elsevier_sciencedirect_doi_10_1016_j_eswa_2024_123204
PublicationCentury 2000
PublicationDate 2024-08-01
2024-08-00
PublicationDateYYYYMMDD 2024-08-01
PublicationDate_xml – month: 08
  year: 2024
  text: 2024-08-01
  day: 01
PublicationDecade 2020
PublicationTitle Expert systems with applications
PublicationYear 2024
Publisher Elsevier Ltd
Publisher_xml – name: Elsevier Ltd
References Zhao, Wallace, Feng, Klein, Singh (b38) 2021
Brown, Mann, Ryder, Subbiah, Kaplan, Dhariwal, Neelakantan, Shyam, Sastry, Askell, Agarwal, Herbert-Voss, Krueger, Henighan, Child, Ramesh, Ziegler, Wu, Winter, Amodei (b1) 2020; volume 33
Meng, Zhang, Huang, Xiong, Ji, Zhang, Han (b23) 2020
Wu, Li, Wang, Zhu (b32) 2012
Lehmann, Isele, Jakob, Jentzsch, Kontokostas, Mendes, Hellmann, Morsey, van Kleef, Auer, Bizer (b16) 2015
Schick, Schütze (b28) 2021
Lewis, Liu, Goyal, Ghazvininejad, Mohamed, Levy, Stoyanov, Zettlemoyer (b17) 2020
Eysenck, Brysbaert (b6) 2018
Li, Liang (b18) 2021
Hambardzumyan, Khachatrian, May (b8) 2021
Raffel, Shazeer, Roberts, Lee, Narang, Matena, Zhou, Li, Liu (b26) 2020; 21
Zhang, Li, Chen, Deng, Bi, Tan, Huang, Chen (b36) 2021
Xu, Wang, Qiu, Luo, Xu, Huang, Huang (b33) 2023
Yuan, Neubig, Liu (b35) 2021; volume 34
Toshniwal, Shi, Shi, Gao, Livescu, Gimpel (b31) 2020
Hu, Ding, Wang, Liu, Wang, Li, Wu, Sun (b11) 2022
Kong, Jiang, Zhuang, Lyu, Zhao, Zhang (b15) 2020
Maas, Daly, Pham, Huang, Ng, Potts (b21) 2011
Liu, Ott, Goyal, Du, Joshi, Chen, Levy, Lewis, Zettlemoyer, Stoyanov (b19) 2019
Liu, Yuan, Fu, Jiang, Hayashi, Neubig (b20) 2021
Jiang, Xu, Araki, Neubig (b14) 2020; 8
Devlin, Chang, Lee, Toutanova (b3) 2019
McAuley, Leskovec (b22) 2013
Cui, Hu, Ding, Huang, Liu (b2) 2022
Ji, Wang, Shi, Zhang, Wang, Yan (b13) 2019; 1
Reynolds, McDonell (b27) 2021
Schick, Schütze (b29) 2021
Dunning (b5) 1993; 19
Huang, Meng, Han (b12) 2022
Shin, Razeghi, Logan IV, Wallace, Singh (b30) 2020
Zhang, Zhao, LeCun (b37) 2015; volume 28
Radford, Wu, Child, Luan, Amodei, Sutskever (b25) 2019
Howard, Ruder (b10) 2018
Petroni, Rocktäschel, Riedel, Lewis, Bakhtin, Wu, Miller (b24) 2019
Yin, Hay, Roth (b34) 2019
Ding, Hu, Zhao, Chen, Liu, Zheng, Sun (b4) 2021
Gao, Fisch, Chen (b7) 2021
Han, Zhang, Ding, Gu, Liu, Huo, Qiu, Yao, Zhang, Zhang, Han, Huang, Jin, Lan, Liu, Liu, Lu, Qiu, Song, Zhu (b9) 2021
Lewis (10.1016/j.eswa.2024.123204_b17) 2020
Meng (10.1016/j.eswa.2024.123204_b23) 2020
Zhang (10.1016/j.eswa.2024.123204_b36) 2021
Hambardzumyan (10.1016/j.eswa.2024.123204_b8) 2021
Zhang (10.1016/j.eswa.2024.123204_b37) 2015; volume 28
Raffel (10.1016/j.eswa.2024.123204_b26) 2020; 21
Lehmann (10.1016/j.eswa.2024.123204_b16) 2015
Zhao (10.1016/j.eswa.2024.123204_b38) 2021
Kong (10.1016/j.eswa.2024.123204_b15) 2020
Li (10.1016/j.eswa.2024.123204_b18) 2021
Toshniwal (10.1016/j.eswa.2024.123204_b31) 2020
Liu (10.1016/j.eswa.2024.123204_b19) 2019
Schick (10.1016/j.eswa.2024.123204_b28) 2021
Eysenck (10.1016/j.eswa.2024.123204_b6) 2018
McAuley (10.1016/j.eswa.2024.123204_b22) 2013
Gao (10.1016/j.eswa.2024.123204_b7) 2021
Ji (10.1016/j.eswa.2024.123204_b13) 2019; 1
Reynolds (10.1016/j.eswa.2024.123204_b27) 2021
Wu (10.1016/j.eswa.2024.123204_b32) 2012
Xu (10.1016/j.eswa.2024.123204_b33) 2023
Dunning (10.1016/j.eswa.2024.123204_b5) 1993; 19
Devlin (10.1016/j.eswa.2024.123204_b3) 2019
Han (10.1016/j.eswa.2024.123204_b9) 2021
Howard (10.1016/j.eswa.2024.123204_b10) 2018
Brown (10.1016/j.eswa.2024.123204_b1) 2020; volume 33
Radford (10.1016/j.eswa.2024.123204_b25) 2019
Ding (10.1016/j.eswa.2024.123204_b4) 2021
Maas (10.1016/j.eswa.2024.123204_b21) 2011
Liu (10.1016/j.eswa.2024.123204_b20) 2021
Cui (10.1016/j.eswa.2024.123204_b2) 2022
Hu (10.1016/j.eswa.2024.123204_b11) 2022
Yin (10.1016/j.eswa.2024.123204_b34) 2019
Yuan (10.1016/j.eswa.2024.123204_b35) 2021; volume 34
Jiang (10.1016/j.eswa.2024.123204_b14) 2020; 8
Schick (10.1016/j.eswa.2024.123204_b29) 2021
Huang (10.1016/j.eswa.2024.123204_b12) 2022
Petroni (10.1016/j.eswa.2024.123204_b24) 2019
Shin (10.1016/j.eswa.2024.123204_b30) 2020
References_xml – volume: 8
  start-page: 423
  year: 2020
  end-page: 438
  ident: b14
  article-title: How can we know what language models know?
  publication-title: Transactions of the Association for Computational Linguistics
  contributor:
    fullname: Neubig
– year: 2021
  ident: b20
  article-title: Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing
  contributor:
    fullname: Neubig
– start-page: 3816
  year: 2021
  end-page: 3830
  ident: b7
  article-title: Making pre-trained language models better few-shot learners
  publication-title: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (volume 1: Long papers)
  contributor:
    fullname: Chen
– start-page: 2463
  year: 2019
  end-page: 2473
  ident: b24
  article-title: Language models as knowledge bases?
  publication-title: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP)
  contributor:
    fullname: Miller
– start-page: 7871
  year: 2020
  end-page: 7880
  ident: b17
  article-title: BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension
  publication-title: Proceedings of the 58th annual meeting of the association for computational linguistics
  contributor:
    fullname: Zettlemoyer
– year: 2021
  ident: b4
  article-title: OpenPrompt: An open-source framework for prompt-learning
  contributor:
    fullname: Sun
– start-page: 438
  year: 2023
  end-page: 446
  ident: b33
  article-title: Making pre-trained language models end-to-end few-shot learners with contrastive prompt tuning
  publication-title: Proceedings of the sixteenth ACM international conference on web search and data mining
  contributor:
    fullname: Huang
– start-page: 4582
  year: 2021
  end-page: 4597
  ident: b18
  article-title: Prefix-tuning: optimizing continuous prompts for generation
  publication-title: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (volume 1: Long papers)
  contributor:
    fullname: Liang
– volume: 1
  start-page: 238
  year: 2019
  end-page: 270
  ident: b13
  article-title: Microsoft concept graph: Mining semantic concepts for short text understanding
  publication-title: Data Intelligence
  contributor:
    fullname: Yan
– start-page: 3914
  year: 2019
  end-page: 3923
  ident: b34
  article-title: Benchmarking zero-shot text classification: Datasets, evaluation and entailment approach
  publication-title: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP)
  contributor:
    fullname: Roth
– year: 2019
  ident: b25
  article-title: Language models are unsupervised multitask learners
  contributor:
    fullname: Sutskever
– volume: volume 34
  start-page: 27263
  year: 2021
  end-page: 27277
  ident: b35
  article-title: BARTScore: Evaluating generated text as text generation
  publication-title: Advances in neural information processing systems
  contributor:
    fullname: Liu
– start-page: 605
  year: 2022
  end-page: 614
  ident: b12
  article-title: Few-shot fine-grained entity typing with automatic label interpretation and instance generation
  publication-title: Proceedings of the 28th ACM SIGKDD conference on knowledge discovery and data mining
  contributor:
    fullname: Han
– start-page: 142
  year: 2011
  end-page: 150
  ident: b21
  article-title: Learning word vectors for sentiment analysis
  publication-title: Proceedings of the 49th annual meeting of the association for computational linguistics: Human language technologies
  contributor:
    fullname: Potts
– start-page: 2225
  year: 2022
  end-page: 2240
  ident: b11
  article-title: Knowledgeable prompt-tuning: Incorporating knowledge into prompt verbalizer for text classification
  publication-title: Proceedings of the 60th annual meeting of the association for computational linguistics (volume 1: Long papers)
  contributor:
    fullname: Sun
– volume: volume 28
  year: 2015
  ident: b37
  article-title: Character-level convolutional networks for text classification
  publication-title: Advances in neural information processing systems
  contributor:
    fullname: LeCun
– start-page: 328
  year: 2018
  end-page: 339
  ident: b10
  article-title: Universal language model fine-tuning for text classification
  publication-title: Proceedings of the 56th annual meeting of the association for computational linguistics (volume 1: Long papers)
  contributor:
    fullname: Ruder
– year: 2021
  ident: b9
  article-title: Pre-trained models: Past, present and future
  contributor:
    fullname: Zhu
– volume: 19
  start-page: 61
  year: 1993
  end-page: 74
  ident: b5
  article-title: Accurate methods for the statistics of surprise and coincidence
  publication-title: Computational Linguistics
  contributor:
    fullname: Dunning
– year: 2021
  ident: b36
  article-title: Differentiable prompt makes pre-trained language models better few-shot learners
  publication-title: International conference on learning representations
  contributor:
    fullname: Chen
– start-page: 4222
  year: 2020
  end-page: 4235
  ident: b30
  article-title: AutoPrompt: Eliciting knowledge from language models with automatically generated prompts
  publication-title: Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP)
  contributor:
    fullname: Singh
– year: 2015
  ident: b16
  article-title: DBpedia - A large-scale, multilingual knowledge base extracted from Wikipedia
  publication-title: Semantic Web
  contributor:
    fullname: Bizer
– start-page: 165
  year: 2013
  end-page: 172
  ident: b22
  article-title: Hidden factors and hidden topics: Understanding rating dimensions with review text
  publication-title: Proceedings of the 7th ACM conference on recommender systems
  contributor:
    fullname: Leskovec
– start-page: 4171
  year: 2019
  end-page: 4186
  ident: b3
  article-title: BERT: Pre-training of deep bidirectional transformers for language understanding
  publication-title: Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: Human language technologies, volume 1 (long and short papers)
  contributor:
    fullname: Toutanova
– year: 2018
  ident: b6
  article-title: Fundamentals of coginition
  contributor:
    fullname: Brysbaert
– start-page: 2339
  year: 2021
  end-page: 2352
  ident: b29
  article-title: It’s not just size that matters: Small language models are also few-shot learners
  publication-title: Proceedings of the 2021 conference of the North American chapter of the association for computational linguistics: Human language technologies
  contributor:
    fullname: Schütze
– start-page: 4921
  year: 2021
  end-page: 4933
  ident: b8
  article-title: WARP: word-level adversarial reprogramming
  publication-title: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (volume 1: Long papers)
  contributor:
    fullname: May
– start-page: 1326
  year: 2020
  end-page: 1340
  ident: b15
  article-title: Calibrated language model fine-tuning for in- and out-of-distribution data
  publication-title: Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP)
  contributor:
    fullname: Zhang
– start-page: 166
  year: 2020
  end-page: 176
  ident: b31
  article-title: A cross-task analysis of text span representations
  publication-title: Proceedings of the 5th workshop on representation learning for NLP
  contributor:
    fullname: Gimpel
– start-page: 255
  year: 2021
  end-page: 269
  ident: b28
  article-title: Exploiting cloze-questions for few-shot text classification and natural language inference
  publication-title: Proceedings of the 16th conference of the european chapter of the association for computational linguistics: Main volume
  contributor:
    fullname: Schütze
– start-page: 12697
  year: 2021
  end-page: 12706
  ident: b38
  article-title: Calibrate before use: Improving few-shot performance of language models
  publication-title: Proceedings of the 38th international conference on machine learning
  contributor:
    fullname: Singh
– start-page: 9006
  year: 2020
  end-page: 9017
  ident: b23
  article-title: Text classification using label names only: A language model self-training approach
  publication-title: Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP)
  contributor:
    fullname: Han
– year: 2022
  ident: b2
  article-title: Prototypical verbalizer for prompt-based few-shot tuning
  contributor:
    fullname: Liu
– year: 2019
  ident: b19
  article-title: RoBERTa: A robustly optimized BERT pretraining approach
  contributor:
    fullname: Stoyanov
– start-page: 481
  year: 2012
  end-page: 492
  ident: b32
  article-title: Probase: A probabilistic taxonomy for text understanding
  publication-title: Proceedings of the 2012 ACM SIGMOD international conference on management of data
  contributor:
    fullname: Zhu
– volume: 21
  start-page: 1
  year: 2020
  end-page: 67
  ident: b26
  article-title: Exploring the limits of transfer learning with a unified text-to-text transformer
  publication-title: Journal of Machine Learning Research
  contributor:
    fullname: Liu
– start-page: 1
  year: 2021
  end-page: 7
  ident: b27
  article-title: Prompt programming for large language models: Beyond the few-shot paradigm
  publication-title: Extended abstracts of the 2021 CHI conference on human factors in computing systems
  contributor:
    fullname: McDonell
– volume: volume 33
  start-page: 1877
  year: 2020
  end-page: 1901
  ident: b1
  article-title: Language models are few-shot learners
  publication-title: Advances in neural information processing systems
  contributor:
    fullname: Amodei
– start-page: 328
  year: 2018
  ident: 10.1016/j.eswa.2024.123204_b10
  article-title: Universal language model fine-tuning for text classification
  contributor:
    fullname: Howard
– start-page: 3816
  year: 2021
  ident: 10.1016/j.eswa.2024.123204_b7
  article-title: Making pre-trained language models better few-shot learners
  contributor:
    fullname: Gao
– year: 2021
  ident: 10.1016/j.eswa.2024.123204_b9
  contributor:
    fullname: Han
– start-page: 1
  year: 2021
  ident: 10.1016/j.eswa.2024.123204_b27
  article-title: Prompt programming for large language models: Beyond the few-shot paradigm
  contributor:
    fullname: Reynolds
– start-page: 3914
  year: 2019
  ident: 10.1016/j.eswa.2024.123204_b34
  article-title: Benchmarking zero-shot text classification: Datasets, evaluation and entailment approach
  contributor:
    fullname: Yin
– start-page: 2225
  year: 2022
  ident: 10.1016/j.eswa.2024.123204_b11
  article-title: Knowledgeable prompt-tuning: Incorporating knowledge into prompt verbalizer for text classification
  contributor:
    fullname: Hu
– start-page: 4171
  year: 2019
  ident: 10.1016/j.eswa.2024.123204_b3
  article-title: BERT: Pre-training of deep bidirectional transformers for language understanding
  contributor:
    fullname: Devlin
– start-page: 2463
  year: 2019
  ident: 10.1016/j.eswa.2024.123204_b24
  article-title: Language models as knowledge bases?
  contributor:
    fullname: Petroni
– volume: 19
  start-page: 61
  issue: 1
  year: 1993
  ident: 10.1016/j.eswa.2024.123204_b5
  article-title: Accurate methods for the statistics of surprise and coincidence
  publication-title: Computational Linguistics
  contributor:
    fullname: Dunning
– start-page: 1326
  year: 2020
  ident: 10.1016/j.eswa.2024.123204_b15
  article-title: Calibrated language model fine-tuning for in- and out-of-distribution data
  contributor:
    fullname: Kong
– volume: volume 28
  year: 2015
  ident: 10.1016/j.eswa.2024.123204_b37
  article-title: Character-level convolutional networks for text classification
  contributor:
    fullname: Zhang
– year: 2022
  ident: 10.1016/j.eswa.2024.123204_b2
  contributor:
    fullname: Cui
– volume: 1
  start-page: 238
  issue: 3
  year: 2019
  ident: 10.1016/j.eswa.2024.123204_b13
  article-title: Microsoft concept graph: Mining semantic concepts for short text understanding
  publication-title: Data Intelligence
  doi: 10.1162/dint_a_00013
  contributor:
    fullname: Ji
– start-page: 9006
  year: 2020
  ident: 10.1016/j.eswa.2024.123204_b23
  article-title: Text classification using label names only: A language model self-training approach
  contributor:
    fullname: Meng
– start-page: 255
  year: 2021
  ident: 10.1016/j.eswa.2024.123204_b28
  article-title: Exploiting cloze-questions for few-shot text classification and natural language inference
  contributor:
    fullname: Schick
– volume: volume 33
  start-page: 1877
  year: 2020
  ident: 10.1016/j.eswa.2024.123204_b1
  article-title: Language models are few-shot learners
  contributor:
    fullname: Brown
– year: 2021
  ident: 10.1016/j.eswa.2024.123204_b4
  contributor:
    fullname: Ding
– volume: 8
  start-page: 423
  year: 2020
  ident: 10.1016/j.eswa.2024.123204_b14
  article-title: How can we know what language models know?
  publication-title: Transactions of the Association for Computational Linguistics
  doi: 10.1162/tacl_a_00324
  contributor:
    fullname: Jiang
– year: 2021
  ident: 10.1016/j.eswa.2024.123204_b20
  contributor:
    fullname: Liu
– year: 2015
  ident: 10.1016/j.eswa.2024.123204_b16
  article-title: DBpedia - A large-scale, multilingual knowledge base extracted from Wikipedia
  publication-title: Semantic Web
  doi: 10.3233/SW-140134
  contributor:
    fullname: Lehmann
– start-page: 438
  year: 2023
  ident: 10.1016/j.eswa.2024.123204_b33
  article-title: Making pre-trained language models end-to-end few-shot learners with contrastive prompt tuning
  contributor:
    fullname: Xu
– start-page: 481
  year: 2012
  ident: 10.1016/j.eswa.2024.123204_b32
  article-title: Probase: A probabilistic taxonomy for text understanding
  contributor:
    fullname: Wu
– start-page: 2339
  year: 2021
  ident: 10.1016/j.eswa.2024.123204_b29
  article-title: It’s not just size that matters: Small language models are also few-shot learners
  contributor:
    fullname: Schick
– start-page: 142
  year: 2011
  ident: 10.1016/j.eswa.2024.123204_b21
  article-title: Learning word vectors for sentiment analysis
  contributor:
    fullname: Maas
– year: 2018
  ident: 10.1016/j.eswa.2024.123204_b6
  contributor:
    fullname: Eysenck
– year: 2019
  ident: 10.1016/j.eswa.2024.123204_b25
  contributor:
    fullname: Radford
– start-page: 4222
  year: 2020
  ident: 10.1016/j.eswa.2024.123204_b30
  article-title: AutoPrompt: Eliciting knowledge from language models with automatically generated prompts
  contributor:
    fullname: Shin
– start-page: 7871
  year: 2020
  ident: 10.1016/j.eswa.2024.123204_b17
  article-title: BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension
  contributor:
    fullname: Lewis
– start-page: 4582
  year: 2021
  ident: 10.1016/j.eswa.2024.123204_b18
  article-title: Prefix-tuning: optimizing continuous prompts for generation
  contributor:
    fullname: Li
– start-page: 165
  year: 2013
  ident: 10.1016/j.eswa.2024.123204_b22
  article-title: Hidden factors and hidden topics: Understanding rating dimensions with review text
  contributor:
    fullname: McAuley
– start-page: 605
  year: 2022
  ident: 10.1016/j.eswa.2024.123204_b12
  article-title: Few-shot fine-grained entity typing with automatic label interpretation and instance generation
  contributor:
    fullname: Huang
– year: 2019
  ident: 10.1016/j.eswa.2024.123204_b19
  contributor:
    fullname: Liu
– volume: volume 34
  start-page: 27263
  year: 2021
  ident: 10.1016/j.eswa.2024.123204_b35
  article-title: BARTScore: Evaluating generated text as text generation
  contributor:
    fullname: Yuan
– start-page: 4921
  year: 2021
  ident: 10.1016/j.eswa.2024.123204_b8
  article-title: WARP: word-level adversarial reprogramming
  contributor:
    fullname: Hambardzumyan
– year: 2021
  ident: 10.1016/j.eswa.2024.123204_b36
  article-title: Differentiable prompt makes pre-trained language models better few-shot learners
  contributor:
    fullname: Zhang
– volume: 21
  start-page: 1
  issue: 140
  year: 2020
  ident: 10.1016/j.eswa.2024.123204_b26
  article-title: Exploring the limits of transfer learning with a unified text-to-text transformer
  publication-title: Journal of Machine Learning Research
  contributor:
    fullname: Raffel
– start-page: 166
  year: 2020
  ident: 10.1016/j.eswa.2024.123204_b31
  article-title: A cross-task analysis of text span representations
  contributor:
    fullname: Toshniwal
– start-page: 12697
  year: 2021
  ident: 10.1016/j.eswa.2024.123204_b38
  article-title: Calibrate before use: Improving few-shot performance of language models
  contributor:
    fullname: Zhao
SSID ssj0017007
Score 2.4773178
Snippet The verbalizer, which serves to map label words to class labels, is an essential component of prompt-tuning. In this paper, we present a novel approach to...
SourceID crossref
elsevier
SourceType Aggregation Database
Publisher
StartPage 123204
SubjectTerms Prompt learning
Text classification
Verbalizer construction
Zero-shot
Title A novel prompt-tuning method: Incorporating scenario-specific concepts into a verbalizer
URI https://dx.doi.org/10.1016/j.eswa.2024.123204
Volume 247
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwnV1LSwMxEA6lXrz4Fuuj5OBN0nabZB_eSrG0ir1oobclyWahotvSbhU8-NudabKiCB48btjA8mXyzcwy8w0hlyKxcQiOguksEgzyL8FUaGPGtYyNjEyoQux3vh-Hw4m4ncppjfSrXhgsq_Tc7zh9w9Z-pe3RbC9ms_YDBAfgDiG1E8i5roMdnBHYdOvjq8wD5ecip7cXMXzbN864Gi-7ekPtoa5oYWThh7X9ck7fHM5gj-z4SJH23Mfsk5otDshuNYWB-kt5SKY9Wsxf7TMFLnxZlKxc468O6kZDX9MRClVuxIpxFbWbIDueM-ywxCohalzf4orOinJOFQXT1hCbv9vlEZkMbh77Q-bnJTADUUrJ8izLJRdKyEQHKjFZFkA2EVgBfBx1rYX7kCtuhQGfFGcy71ibcK7jRCYG4kTFj0m9mBf2hFAewbaww3MJ4AQBVwLOA7CMjYi6WugGuaqAShdOFiOt6sWeUoQ1RVhTB2uDyArL9MfhpsDbf-w7_ee-M7KNT65O75zUy-XaXkDsUOrmxjiaZKs3uhuOPwGbYcN7
link.rule.ids 315,786,790,4521,24144,27957,27958,45620,45714
linkProvider Elsevier
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwnV3JTsMwELWq9gAXdkRZfeCGTJvazsKtqkApXS60Um-WnThSESRVm4LE1zOuHQRC4sDVyUjRi_3mTTILQtcs0qEPjoKoNGAE4i9GpK9DQhUPEx4kvvRNvfNo7MdT9jjjsxrqVbUwJq3Scb_l9A1bu5WWQ7O1mM9bTyAOwB1CaMcM55oK9gbjgcfqqNHtD-Lx18-EoG2rpuF-Ygxc7YxN89Krd9N-qMNujbhw89p--advPudhD-04sYi79nn2UU3nB2i3GsSA3bk8RLMuzos3_YKBDl8XJSnX5msHttOh73Df9Krc9Cs2q6Z9EwTIBTFFliZRCCe2dHGF53lZYIlhdyuQ5x96eYSmD_eTXkzcyASSgFApSZamGadMMh4pT0ZJmnoQUHiaASUHHa3hSGSSapaAWwpTnrW1jihVYcSjBKSipMeonhe5PkGYBmDmt2nGARzPo5LBKwEsw4QFHcVUE91UQImF7YwhqpSxZ2FgFQZWYWFtIl5hKX68XwHU_Yfd6T_trtBWPBkNxbA_HpyhbXPFpu2do3q5XOsLkBKlunRb5RNvG8Yx
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=A+novel+prompt-tuning+method%3A+Incorporating+scenario-specific+concepts+into+a+verbalizer&rft.jtitle=Expert+systems+with+applications&rft.au=Ma%2C+Yong&rft.au=Luo%2C+Senlin&rft.au=Shang%2C+Yu-Ming&rft.au=Li%2C+Zhengjun&rft.date=2024-08-01&rft.issn=0957-4174&rft.volume=247&rft.spage=123204&rft_id=info:doi/10.1016%2Fj.eswa.2024.123204&rft.externalDBID=n%2Fa&rft.externalDocID=10_1016_j_eswa_2024_123204
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=0957-4174&client=summon
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=0957-4174&client=summon
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=0957-4174&client=summon