PDF (7.8 MB)
Collect
Submit Manuscript
Show Outline
Outline
Abstract
Keywords
References
Show full outline
Hide outline

PCP-tuning: Personalized Continuous Prompt Tuning for Few-Shot Learning

Ting LIUShaotian CAIXiaojun CHEN()Qin ZHANG
School of Computer Science and Technology, Shenzhen University, Shenzhen Guangdong 518071, China
Show Author Information

Abstract

Pre-trained language models have achieved remarkable performance in few-shot learning with the rise of “prompt learning”, where the key problem is how to construct a suitable prompt for each example. Sample and prompt will be combined as a new input to language model (LM). A series of prompt construction methods have been proposed recently, some of these methods are for discrete prompt construction, and some focus on continuous prompt construction, both of them normally apply a unified prompt to all examples. However, the results show that it is hard to find a perfect unified prompt that works for all examples in a task, one prompt can only help LM assign the correct class to some samples in the downstream classification task and give the wrong result to others. To this end, we propose a novel personalized continuous prompt tuning (PCP-tuning) method to learn personalized prompts that are tailored to each sample's semantic for few-shot learning. Two calibration techniques are proposed to control the distribution of generated prompts for better prompts. Extensive experimental results on ten benchmark tasks demonstrate the superior performance of our method.

Article ID: 2096-7675(2024)01-0059-010

References

[1]
AISHAN W, WEI W L, ZAOKERE K. Sentiment analysis based on BiLSTM+Attention in sports field[J]. Journal of Xinjiang University (Natural Science Edition in Chinese and English), 2020, 37(2): 142-149. (in Chinese)
[2]
ZENG R, HUANG D Q, WEI X, et al. Short-term traffic flow forecast based on modified WOA optimized LSTM neural network[J]. Journal of Xinjiang University (Natural Science Edition in Chinese and English), 2022, 39(2): 242-248. (in Chinese)
[3]
TAN X, TUERGEN Y, AISHAN W, et al. Uygur words clustering based on the similarity calculation[J]. Journal of Xinjiang University (Natural Science Edition), 2012, 29(1): 104-107. (in Chinese)
[4]
YALIQING A, HALIDAN A, CHEN Y. Uygur text filtering based on vector space model[J]. Journal of Xinjiang University (Natural Science Edition), 2015, 32(2): 221-226. (in Chinese)
[5]
RAFFEL C, SHAZEER N, ROBERTS A, et al. Exploring the limits of transfer learning with a unified text-to-text transformer[EB/OL]. 2019: arXiv: 1910.10683. http://arxiv.org/abs/1910.10683.pdf.
[6]
SUI D B, CHEN Y B, MAO B J, et al. Knowledge guided metric learning for few-shot text classification[C]//Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Online. Stroudsburg, PA, USA: Association for Computational Linguistics, 2021: 3266-3271.
[7]
LIU P F, YUAN W Z, FU J L, et al. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing[J]. ACM Computing Surveys, 2023, 55(9): 195.
[8]
BROWN T B, MANN B, RYDER N, et al. Language models are few-shot learners[C]//Proceedings of the 34th International Conference on Neural Information Processing Systems. December 6-12, 2020, Vancouver, BC, Canada. ACM, 2020: 1877-1901.
[9]
SCHICK T, SCHÜTZE H. Exploiting cloze-questions for few-shot text classification and natural language inference[C]//Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume. Online. Stroudsburg, PA, USA: Association for Computational Linguistics, 2021: 255-269.
[10]
SCHICK T, SCHÜTZE H. It's not just size that matters: Small language models are also few-shot learners[C]//Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Online. Stroudsburg, PA, USA: Association for Computational Linguistics, 2021: 2339-2352.
[11]
SHIN T, RAZEGHI Y, LOGAN R L, et al. AutoPrompt: Eliciting knowledge from language models with automatically generated prompts[C]//Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). Online. Stroudsburg, PA, USA: Association for Computational Linguistics, 2020: 4222-4235.
[12]
GAO T Y, FISCH A, CHEN D Q. Making pre-trained language models better few-shot learners[C]//Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). Online. Stroudsburg, PA, USA: Association for Computational Linguistics, 2021: 3816-3830.
[13]
HAN X, ZHAO W L, DING N, et al. PTR: Prompt tuning with rules for text classification[J]. AI Open, 2022, 3: 182-192.
[14]
HU S D, DING N, WANG H D, et al. Knowledgeable prompt-tuning: Incorporating knowledge into prompt verbalizer for text classification[C]//Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Dublin, Ireland. Stroudsburg, PA, USA: Association for Computational Linguistics, 2022: 2225-2240.
[15]
ZHONG Z X, FRIEDMAN D, CHEN D Q. Factual probing is[MASK]: Learning vs. learning to recall[C]//Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Online. Stroudsburg, PA, USA: Association for Computational Linguistics, 2021: 5017-5033.
[16]
LIU X, ZHENG Y N, DU Z X, et al. GPT understands, too[EB/OL]. (2023-08-26)[2023-09-03]. .
[17]
LI X L, LIANG P. Prefix-tuning: Optimizing continuous prompts for generation[C]//Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). Online. Stroudsburg, PA, USA: Association for Computational Linguistics, 2021: 4582-4597.
[18]
LESTER B, AL-RFOU R, CONSTANT N. The power of scale for parameter-efficient prompt tuning[C]//Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. Online and Punta Cana, Dominican Republic. Stroudsburg, PA, USA: Association for Computational Linguistics, 2021: 3045-3059.
[19]
GU Y X, HAN X, LIU Z Y, et al. PPT: Pre-trained prompt tuning for few-shot learning[C]//Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Dublin, Ireland. Stroudsburg, PA, USA: Association for Computational Linguistics, 2022: 8410-8423.
[20]
CHEN J A, YANG Z C, YANG D Y. MixText: Linguistically-informed interpolation of hidden space for semi-supervised text classification[C]//Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Online. Stroudsburg, PA, USA: Association for Computational Linguistics, 2020: 2147-2157.
[21]
YU M, GUO X X, YI J F, et al. Diverse few-shot text classification with multiple metrics[C]//Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers). New Orleans, Louisiana. Stroudsburg, PA, USA: Association for Computational Linguistics, 2018: 1206-1215.
[22]
HAN X, ZHU H, YU P F, et al. FewRel: A large-scale supervised few-shot relation classification dataset with state-of-the-art evaluation[C]//Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Brussels, Belgium. Stroudsburg, PA, USA: Association for Computational Linguistics, 2018: 4803-4809.
[23]
BANSAL T, JHA R, MCCALLUM A. Learning to few-shot learn across diverse natural language classification tasks[C]//Proceedings of the 28th International Conference on Computational Linguistics. Barcelona, Spain (Online). Stroudsburg, PA, USA: International Committee on Computational Linguistics, 2020: 5108-5123.
[24]
CHEN T, KORNBLITH S, NOROUZI M, et al. A simple framework for contrastive learning of visual representations[C]//Proceedings of the 37th International Conference on Machine Learning. ACM, 2020: 1597-1607.
[25]
HE K M, FAN H Q, WU Y X, et al. Momentum contrast for unsupervised visual representation learning[C]//2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Seattle, WA, USA. IEEE, 2020: 9726-9735.
[26]
JIANG T, JIAO J, HUANG S H, et al. PromptBERT: Improving BERT sentence embeddings with prompts[C]//Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing. Abu Dhabi, United Arab Emirates. Stroudsburg, PA, USA: Association for Computational Linguistics, 2022: 8826-8837.
[27]
WANG A, SINGH A, MICHAEL J, et al. GLUE: A multi-task benchmark and analysis platform for natural language understanding[C]//Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP. Brussels, Belgium. Stroudsburg, PA, USA: Association for Computational Linguistics, 2018: 353-355.
[28]
HU M Q, LIU B. Mining and summarizing customer reviews[C]//Proceedings of the Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. August 22-25, 2004, Seattle, WA, USA. ACM, 2004: 168-177.
[29]
VOORHEES E M, TICE D M. Building a question answering test collection[C]//Proceedings of the 23rd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval. July 24-28, 2000, Athens, Greece. ACM, 2000: 200-207.
[30]
WIEBE J, WILSON T, CARDIE C. Annotating expressions of opinions and emotions in language[J]. Language Resources and Evaluation, 2005, 39(2): 165-210.
[31]
PASZKE A, GROSS S, MASSA F, et al. PyTorch: An imperative style, high-performance deep learning library[EB/OL]. 2019: arXiv: 1912.01703. http://arxiv.org/abs/1912.01703.pdf.
Journal of Xinjiang University(Natural Science Edition in Chinese and English)
Pages 59-68
Cite this article:
LIU T, CAI S, CHEN X, et al. PCP-tuning: Personalized Continuous Prompt Tuning for Few-Shot Learning. Journal of Xinjiang University(Natural Science Edition in Chinese and English), 2024, 41(1): 59-68. https://doi.org/10.13568/j.cnki.651094.651316.2023.09.17.0001
Metrics & Citations  
Article History
Copyright
Return