Fine-tuning vs. In-Context Learning: New Research Reveals Best Practices for LLM Customization

A study by Google DeepMind and Stanford sheds light on how fine-tuning and in-context learning each impact large language model generalization, guiding developers seeking maximum value from enterprise data with Artificial Intelligence.

Researchers from Google DeepMind and Stanford University have conducted a comprehensive study comparing two prominent methods for adapting large language models (LLMs) to downstream tasks: fine-tuning and in-context learning. Fine-tuning involves retraining a pre-trained LLM on a specialized data subset to alter its internal parameters, while in-context learning (ICL) uses tailored examples within the prompt to guide the model’s outputs without modifying its underlying structure. To ensure rigorous testing, the researchers used synthetic datasets with complex relationships and replaced familiar terms with nonsense words, ruling out the influence of prior knowledge learned during pre-training.

The study subjected LLMs to a series of logical and deductive challenges involving tasks such as relationship reversals and syllogisms, using both fine-tuning and ICL strategies. Experimental results revealed that models relying on ICL displayed superior generalization to novel tasks compared to their fine-tuned counterparts. However, ICL is more computationally expensive at inference time, as it requires feeding large context prompts for every model use. In contrast, standard fine-tuning is less flexible on unfamiliar data, but it does not incur repeated inference-time costs.

To bridge the gap between generalization and efficiency, the research team introduced an innovative hybrid approach dubbed ´augmented fine-tuning.´ This method enriches the fine-tuning dataset by integrating new, inferred examples generated using the LLM’s own ICL capabilities, employing both local strategies (individual fact manipulation) and global strategies (holistic data linkage). When these augmented datasets were used for fine-tuning, the resulting models outperformed those trained with standard fine-tuning or ICL alone. The findings suggest that this hybrid technique delivers both broader generalization and greater cost-effectiveness for enterprise deployment. The researchers caution that augmented fine-tuning introduces its own upfront computational overhead, but recommend its consideration wherever standard fine-tuning falls short. Overall, these insights provide actionable guidance for enterprises seeking to reliably adapt LLMs to domain-specific requirements using Artificial Intelligence.

69

Impact Score

Most UK firms see Artificial Intelligence training gap as shadow tool use grows

New research finds that 6 in 10 UK businesses say employees lack comprehensive Artificial Intelligence training, even as shadow use of unapproved tools becomes widespread and investment surges. Executives warn that without stronger skills, governance and strategy, many organisations risk missing out on expected Artificial Intelligence returns.

COSO issues internal control roadmap for governing generative artificial intelligence

COSO has released governance guidance that applies its Internal Control-Integrated Framework to generative artificial intelligence, offering audit-ready control structures and implementation tools for organizations. The publication details capability-based risk mapping, aligned controls, and practical templates to help institutions manage emerging technology risks.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.