Context Learning
In-context learning (ICL) is a paradigm shift in machine learning, focusing on enabling models to adapt to new tasks using only a few examples provided within the input, without requiring parameter updates. Current research emphasizes understanding ICL's mechanisms, particularly within transformer-based large language models, and improving its effectiveness through techniques like enhanced example selection, chain-of-thought prompting, and addressing issues such as spurious correlations and copy bias. This research is significant because ICL offers a more efficient and adaptable approach to many machine learning problems, impacting fields ranging from natural language processing and computer vision to scientific computing and beyond.
Papers
Measuring and Mitigating Constraint Violations of In-Context Learning for Utterance-to-API Semantic Parsing
Shufan Wang, Sebastien Jean, Sailik Sengupta, James Gung, Nikolaos Pappas, Yi Zhang
Boosting Cross-lingual Transferability in Multilingual Models via In-Context Learning
Sunkyoung Kim, Dayeon Ki, Yireun Kim, Jinsik Lee
Adversarial Demonstration Attacks on Large Language Models
Jiongxiao Wang, Zichen Liu, Keun Hee Park, Zhuojun Jiang, Zhaoheng Zheng, Zhuofeng Wu, Muhao Chen, Chaowei Xiao
Coverage-based Example Selection for In-Context Learning
Shivanshu Gupta, Matt Gardner, Sameer Singh
Estimating Large Language Model Capabilities without Labeled Test Data
Harvey Yiyun Fu, Qinyuan Ye, Albert Xu, Xiang Ren, Robin Jia
EXnet: Efficient In-context Learning for Data-less Text classification
Debaditya Shome, Kuldeep Yadav
RetICL: Sequential Retrieval of In-Context Examples with Reinforcement Learning
Alexander Scarlatos, Andrew Lan
Active Learning Principles for In-Context Learning with Large Language Models
Katerina Margatina, Timo Schick, Nikolaos Aletras, Jane Dwivedi-Yu
Skill-Based Few-Shot Selection for In-Context Learning
Shengnan An, Bo Zhou, Zeqi Lin, Qiang Fu, Bei Chen, Nanning Zheng, Weizhu Chen, Jian-Guang Lou
In-Context Probing: Toward Building Robust Classifiers via Probing Large Language Models
Afra Amini, Massimiliano Ciaramita
Label Words are Anchors: An Information Flow Perspective for Understanding In-Context Learning
Lean Wang, Lei Li, Damai Dai, Deli Chen, Hao Zhou, Fandong Meng, Jie Zhou, Xu Sun
Dr.ICL: Demonstration-Retrieved In-context Learning
Man Luo, Xin Xu, Zhuyun Dai, Panupong Pasupat, Mehran Kazemi, Chitta Baral, Vaiva Imbrasaite, Vincent Y Zhao
Learning Relevant Contextual Variables Within Bayesian Optimization
Julien Martinelli, Ayush Bharti, Armi Tiihonen, S. T. John, Louis Filstroff, Sabina J. Sloman, Patrick Rinke, Samuel Kaski
CTQScorer: Combining Multiple Features for In-context Example Selection for Machine Translation
Aswanth Kumar, Ratish Puduppully, Raj Dabre, Anoop Kunchukuttan
Make a Choice! Knowledge Base Question Answering with In-Context Learning
Chuanyuan Tan, Yuehe Chen, Wenbiao Shao, Wenliang Chen
Concept-aware Training Improves In-context Learning Ability of Language Models
Michal Štefánik, Marek Kadlčík
Small Language Models Improve Giants by Rewriting Their Outputs
Giorgos Vernikos, Arthur Bražinskas, Jakub Adamek, Jonathan Mallinson, Aliaksei Severyn, Eric Malmi
Measuring Inductive Biases of In-Context Learning with Underspecified Demonstrations
Chenglei Si, Dan Friedman, Nitish Joshi, Shi Feng, Danqi Chen, He He
Friendly Neighbors: Contextualized Sequence-to-Sequence Link Prediction
Adrian Kochsiek, Apoorv Saxena, Inderjeet Nair, Rainer Gemulla