Name: Towards AI Legal Name: Towards AI, Inc. Description: Towards AI is the world's leading artificial intelligence (AI) and technology publication. Read by thought-leaders and decision-makers around the world. Phone Number: +1-650-246-9381 Email: pub@towardsai.net
228 Park Avenue South New York, NY 10003 United States
Website: Publisher: https://towardsai.net/#publisher Diversity Policy: https://towardsai.net/about Ethics Policy: https://towardsai.net/about Masthead: https://towardsai.net/about
Name: Towards AI Legal Name: Towards AI, Inc. Description: Towards AI is the world's leading artificial intelligence (AI) and technology publication. Founders: Roberto Iriondo, , Job Title: Co-founder and Advisor Works for: Towards AI, Inc. Follow Roberto: X, LinkedIn, GitHub, Google Scholar, Towards AI Profile, Medium, ML@CMU, FreeCodeCamp, Crunchbase, Bloomberg, Roberto Iriondo, Generative AI Lab, Generative AI Lab VeloxTrend Ultrarix Capital Partners Denis Piffaretti, Job Title: Co-founder Works for: Towards AI, Inc. Louie Peters, Job Title: Co-founder Works for: Towards AI, Inc. Louis-François Bouchard, Job Title: Co-founder Works for: Towards AI, Inc. Cover:
Towards AI Cover
Logo:
Towards AI Logo
Areas Served: Worldwide Alternate Name: Towards AI, Inc. Alternate Name: Towards AI Co. Alternate Name: towards ai Alternate Name: towardsai Alternate Name: towards.ai Alternate Name: tai Alternate Name: toward ai Alternate Name: toward.ai Alternate Name: Towards AI, Inc. Alternate Name: towardsai.net Alternate Name: pub.towardsai.net
5 stars – based on 497 reviews

Frequently Used, Contextual References

TODO: Remember to copy unique IDs whenever it needs used. i.e., URL: 304b2e42315e

Resources

Our 15 AI experts built the most comprehensive, practical, 90+ lesson courses to master AI Engineering - we have pathways for any experience at Towards AI Academy. Cohorts still open - use COHORT10 for 10% off.

Publication

Fine-Tuning LLMs in 2025: Techniques, Trade-offs, and Use Cases
Data Science   Latest   Machine Learning

Fine-Tuning LLMs in 2025: Techniques, Trade-offs, and Use Cases

Author(s): Yuval Mehta

Originally published on Towards AI.

Fine-Tuning LLMs in 2025: Techniques, Trade-offs, and Use Cases
Photo by Growtika on Unsplash

Fine-tuning large language models (LLMs) has become a fundamental component of AI development in recent years, moving from a specialized technical endeavor. Enhancing model performance is no longer the only goal; efficiency, accessibility, alignment, and safety are now also important considerations.

1. Full Fine-Tuning: The Classic Route

During the GPT-2 era, complete fine-tuning was standard. The entire model was retrained by developers using a dataset unique to the domain. Even though this method still provides the greatest level of control and specialization, particularly for activities like medical imaging or legal document parsing, it is very expensive:

Pros:

  • Complete control over model behavior
  • High performance on domain-specific tasks

Cons:

  • High compute and memory demands
  • Requires large datasets
  • Time-consuming

2. Parameter-Efficient Fine-Tuning (PEFT): Power Without the Price Tag

Low-Rank Adaptation (LoRA)

By introducing trainable low-rank matrices into pre-existing layers, LoRA avoids the necessity of updating each parameter. This maintains performance while significantly lowering the need for training.

Quantized LoRA (QLoRA)

Introduced in 2023, QLoRA significantly reduces memory consumption by combining quantization techniques with LoRA, enabling the fine-tuning of billion-parameter models on consumer GPUs.

“QLoRA democratizes LLM fine-tuning by reducing the hardware barrier.”

Adapter Layers

In adapter-based fine-tuning, tiny plug-in layers are trained while the main model is frozen. This modularity makes it possible for:

  • Easy task-switching
  • Simplified deployment
  • Reduced compute cost

Pros:

  • Minimal computational overhead
  • Modular and reusable
  • Ideal for limited-resource settings

Cons:

  • May not match full fine-tuning performance
  • Requires architecture-specific integration

3. Instruction Tuning: Teaching LLMs to Follow Orders

Instruction tuning provides a model a variety of tasks that are labeled with instructions instead of training it on a single job. This method improves zero-shot capability and generalization.

Notable Projects

  • T0 (Sanh et al.)
  • FLAN (Google Research)

Pros:

  • Better generalization
  • Improved zero-shot performance

Cons:

  • Requires diverse labeled instruction datasets
  • Doesn’t enforce output alignment by default

4. RLHF: Aligning Models with Human Values

A key component of aligned models such as ChatGPT is Reinforcement Learning from Human Feedback (RLHF). It operates in three phases:

  1. Collect human-labeled comparisons
  2. Train a reward model
  3. Optimize the LLM via reinforcement learning

Pros:

  • Better alignment with human preferences
  • Produces more helpful and safe responses

Cons:

  • Data collection is expensive and slow
  • Hard to scale and replicate

5. System-2 Fine-Tuning: Toward Reflective Reasoning

System-2 Fine-Tuning, which draws inspiration from cognitive science, enables models to reason systematically and purposefully. It highlights:

  • Structured thought
  • Planning and reflection
  • Integrative multi-hop reasoning

Though still emerging, work by Park et al. (2025) indicates promising applications in legal reasoning and scientific research.

Pros:

  • Improved reasoning and planning
  • Enhanced robustness and interpretability

Cons:

  • Still under development
  • Requires careful supervision

6. Prompt Tuning & Soft Prompting: Lightweight Steering

Prompt tweaking provides a lightweight substitute for contexts with limitations or access that is limited to APIs. Rather than changing the model, it learns the best prompts to guide behavior.

In particular, soft prompting is subtle and powerful because it works at the embedding level.

Pros:

  • Fast and low-cost
  • Requires no access to model internals

Cons:

  • Limited control over deep model behavior
  • Less effective for complex reasoning tasks

At a Glance: When to Use What

  • High specialization with ample resources → Use Full Fine-Tuning
  • Resource-constrained environments → Opt for Parameter-Efficient Fine-Tuning (like LoRA or QLoRA)
  • Enhancing instruction-following capabilities → Go with Instruction Tuning
  • Aligning outputs with human preferences → Choose Reinforcement Learning from Human Feedback (RLHF)
  • Rapid prototyping without model access → Use Prompt Tuning
  • Integrating new knowledge and reasoning → Apply System-2 Fine-Tuning
AI-generated image by Napkin AI

The Big Picture: Fine-Tuning as a Craft

In 2025, fine-tuning is about knowing your tools and when to use them, not just about choosing one approach. For added robustness, you may add some RLHF and combine LoRA with instruction tweaking.

It’s simultaneously a strategic choice, a performance trade-off, and a design choice.

“Fine-tuning today is less about code and more about craftsmanship.”

References

  1. Hu et al., 2021 — LoRA: Low-Rank Adaptation of Large Language Models
  2. Dettmers et al., 2023 — QLoRA: Efficient Finetuning of Quantized LLMs
  3. Pfeiffer et al., 2020 — AdapterFusion: Non-Destructive Task Composition for Transfer Learning
  4. Sanh et al., 2021 — Multitask Prompted Training of Language Models (T0)
  5. Ouyang et al., 2022 — Training language models to follow instructions with human feedback
  6. Lester et al., 2021 — The Power of Scale for Parameter-Efficient Prompt Tuning
  7. Park et al., 2025 — System-2 Fine-Tuning for Robust Integration of New Knowledge

Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming a sponsor.

Published via Towards AI


Take our 90+ lesson From Beginner to Advanced LLM Developer Certification: From choosing a project to deploying a working product this is the most comprehensive and practical LLM course out there!

Towards AI has published Building LLMs for Production—our 470+ page guide to mastering LLMs with practical projects and expert insights!


Discover Your Dream AI Career at Towards AI Jobs

Towards AI has built a jobs board tailored specifically to Machine Learning and Data Science Jobs and Skills. Our software searches for live AI jobs each hour, labels and categorises them and makes them easily searchable. Explore over 40,000 live jobs today with Towards AI Jobs!

Note: Content contains the views of the contributing authors and not Towards AI.