
Large language models (LLMs) are advanced AI systems that use deep learning techniques and vast datasets to understand, generate, and predict human language5. Their primary function is to process and generate text with coherent communication, enabling them to perform a wide range of tasks such as translation, summarization, information retrieval, and conversational interactions.

The main applications of LLMs include natural language processing, text generation, question answering, summarization, and translation2. They are used in various industries, such as healthcare, education, and social interactions, to enhance language understanding and automate tasks.

In-context learning (ICL) in Large Language Models (LLMs) faces limitations such as uneven performance, significant computational overhead, and context length restrictions. Traditional ICL methods often struggle with adaptability and efficiency, requiring extensive context windows. These issues highlight the need for more scalable and effective solutions to improve the effectiveness and control of ICL in LLMs.