Discover the SciOpen Platform and Achieve Your Research Goals with Ease.
Search articles, authors, keywords, DOl and etc.
The springing up of large language models (LLMs) has shifted the community from single-task-orientated natural language processing (NLP) research to a holistic end-to-end multi-task learning paradigm. Along this line of research endeavors in the area, LLM-based prompting methods have attracted much attention, partially due to the technological advantages brought by prompt engineering (PE) as well as the underlying NLP principles disclosed by various prompting methods. Traditional supervised learning usually requires training a model based on labeled data and then making predictions. In contrast, PE methods directly use the powerful capabilities of existing LLMs (e.g., GPT-3 and GPT-4) via composing appropriate prompts, especially under few-shot or zero-shot scenarios. Facing the abundance of studies related to the prompting and the ever-evolving nature of this field, this article aims to 1) illustrate a novel perspective to review existing PE methods within the well-established communication theory framework, 2) facilitate a better/deeper understanding of developing trends of existing PE methods used in three typical tasks, and 3) shed light on promising research directions for future PE methods.
Cheng K M, Li Z Y, Li C, Xie R J, Guo Q, He Y B, Wu H Y. The potential of GPT-4 as an AI-powered virtual assistant for surgeons specialized in joint arthroplasty. Annals of Biomedical Engineering , 2023, 51(7): 1366–1370. DOI: 10.1007/s10439-023-03207-z.
Cascella M, Montomoli J, Bellini V, Bignami E. Evaluating the feasibility of ChatGPT in healthcare: An analysis of multiple clinical and research scenarios. Journal of Medical Systems , 2023, 47(1): Article No. 33. DOI: 10.1007/s10916-023-01925-4.
George A S, George A S H. A review of ChatGPT AI’s impact on several business sectors. Partners Universal International Innovation Journal , 2023, 1(1): 9–23. DOI: 10.5281/zenodo.7644359.
Liu P F, Yuan W Z, Fu J L, Jiang Z B, Hayashi H, Neubig G. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Computing Surveys , 2023, 55(9): 195. DOI: 10.1145/3560815.
Radford A, Wu J, Child R, Luan D, Amodei D, Sutskever I. Language models are unsupervised multitask learners. OpenAI blog , 2019, 1(8): Article No. 9.
Jiang Z B, Xu F F, Araki J, Neubig G. How can we know what language models know? Transactions of the Association for Computational Linguistics , 2020, 8: 423–438. DOI: 10.1162/tacl_a_00324.
Han X, Zhao W L, Ding N, Liu Z Y, Sun M S. PTR: Prompt tuning with rules for text classification. AI Open , 2022, 3: 182–192. DOI: 10.1016/j.aiopen.2022.11.003.
Latané B. Dynamic social impact: The creation of culture by communication. Journal of Communication , 1996, 46(4): 13–25. DOI: 10.1111/j.1460-2466.1996.tb01501.x.
Orbe M P. From the standpoint(s) of traditionally muted groups: Explicating a co-cultural communication theoretical model. Communication Theory , 1998, 8(1): 1–26. DOI: 10.1111/j.1468-2885.1998.tb00209.x.
Segrin C, Abramson L Y. Negative reactions to depressive behaviors: A communication theories analysis. Journal of Abnormal Psychology , 1994, 103(4): 655–668. DOI: 10.1037/0021-843X.103.4.655.
Shannon C E. A mathematical theory of communication. The Bell System Technical Journal , 1948, 27(3): 379–423. DOI: 10.1002/j.1538-7305.1948.tb01338.x.
Raffel C, Shazeer N, Roberts A, Lee K, Narang S, Matena M, Zhou Y Q, Li W, Liu P J. Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research , 2020, 21(1): 140.
Ben-David E, Oved N, Reichart R. PADA: Example-based prompt learning for on-the-fly adaptation to unseen domains. Transactions of the Association for Computational Linguistics , 2022, 10: 414–433. DOI: 10.1162/ tacl_a_00468.
Li B H, Hou Y T, Che W X. Data augmentation approaches in natural language processing: A survey. AI Open , 2022, 3: 71–90. DOI: 10.1016/j.aiopen.2022.03.001.
Zhou Z H, Wu J X, Tang W. Ensembling neural networks: Many could be better than all. Artificial Intelligence , 2002, 137(1/2): 239–263. DOI: 10.1016/S0004-3702(02)00190-X.
Gu Z H, Fan J, Tang N, Cao L, Jia B W, Madden S, Du X Y. Few-shot text-to-SQL translation using structure and content prompt learning. Proceedings of the ACM on Management of Data , 2023, 1(2): 147. DOI: 10.1145/3589292.
Yang Q, Liu Y, Chen T J, Tong Y X. Federated machine learning: Concept and applications. ACM Trans. Intelligent Systems and Technology , 2019, 10(2): 12. DOI: 10.1145/3298981.
Schick T, Udupa S, Schütze H. Self-diagnosis and self-debiasing: A proposal for reducing corpus-based bias in NLP. Transactions of the Association for Computational Linguistics , 2021, 9: 1408–1424. DOI: 10.1162/tacl_a_ 00434.