We Benchmarked 12 Small Language Models Across 8 Tasks to Find the Best Base Model for Fine-Tuning

7 min read Original article ↗

TL;DR

We benchmarked 12 small language models across 8 tasks to find the best base model for fine-tuning. Here’s what we found:

Fine tuned SLMs can beat much larger models: fine tuned Qwen3-4B matches or exceeds GPT-OSS-120B (a 30x larger teacher) on 7 of 8 benchmarks, with the remaining one within 3 percentage points. On SQuAD 2.0, the fine-tuned student beats the teacher by 19 points. This means you can get frontier-level accuracy at a fraction of the cost, running entirely on your own hardware.

Best fine-tuned performance: The Qwen3 family dominates, claiming 4 of the top 6 spots. Qwen3-4B-Instruct-2507 ranks #1, followed by Qwen3-8B at #2. The Llama family takes the remaining spots (#3, #5, #7), while Gemma and SmolLM2 trail behind.

Most tunable (biggest gains from fine-tuning): The Llama family leads, with Llama-3.2-1B-Instruct and Llama-3.2-3B-Instruct ranking #1 and #2. Smaller models generally show larger improvements as they start weaker but benefit most from fine-tuning.

Introduction

If you’re building AI applications that need to run on-device, on-premises, or at the edge, you’ve probably asked yourself: which small language model should I fine-tune? The SLM landscape is crowded (Qwen, Llama, Gemma, Granite, SmolLM) and each family offers multiple size variants. Picking the wrong base model can mean weeks of wasted compute or a model that never quite hits production quality.

We ran a systematic benchmark to answer this question with data. Using the distil labs platform, we fine-tuned 12 models across 8 diverse tasks (classification, information extraction, open-book QA, and closed-book QA), then compared their performance against each other and against the teacher LLM used to generate synthetic training data.

This post answers four practical questions:

  1. Which model produces the best results after fine-tuning?
  2. Which model is the most tunable? (i.e., gains the most from fine-tuning)
  3. Which model has the best base performance? (before any fine-tuning)
  4. Can our best student actually match the teacher?

Methodology

We evaluated the following models:

  • Qwen3 family: Qwen3-8B, Qwen3-4B-Instruct-2507, Qwen3-1.7B, Qwen3-0.6B. Note that we turned off thinking for this family to ensure an even playing field.
  • Llama family: Llama-3.1-8B-Instruct, Llama-3.2-3B-Instruct, Llama-3.2-1B-Instruct
  • SmolLM2 family: SmolLM2-1.7B-Instruct, SmolLM2-135M-Instruct
  • Gemma family: gemma-3-1b-it, gemma-3-270m-it
  • Granite: granite-3.3-8b-instruct

For each model, we measured:

  • Base score: Few-shot performance with prompting alone
  • Finetuned score: Performance after training on synthetic data generated by our teacher (GPT-OSS 120B)

Our 8 benchmarks span classification (TREC, Banking77, Ecommerce, Mental Health), document understanding (docs), and question answering (HotpotQA, Roman Empire QA, SQuAD 2.0). For details on the evaluation setup and the benchmarks themselves, see our platform benchmarking post.

Results aggregation: To create fair measurement, we ranked models on each benchmark individually, then computed average ranks across all tasks and plotted the 95% confidence interval as error bars. Lower average rank = better overall performance.

Question 1: Which Model Produces the Best Results After Fine-Tuning?

Winner: Qwen3-4B-Instruct-2507 (average rank: 2.25)

ModelAverage Rank95% CI
Qwen3-4B-Instruct-25072.25±1.03
Qwen3-8B2.75±1.37
Llama-3.1-8B-Instruct4.00±1.42
Qwen3-1.7B4.44±1.60
Llama-3.2-3B-Instruct4.56±1.73
Qwen3-0.6B5.11±1.86

The Qwen3 family dominates the top spots, with Qwen3-4B-Instruct-2507 taking first place. Notably, this 4B model outranks the larger Qwen3-8B, suggesting that for distillation tasks, the more recent version of Qwen3 (July 25 update) yields the best result than the previous 8B SLM.

Key takeaway: If you want the best possible fine-tuned model and have GPU memory for ~4B parameters, Qwen3-4B-Instruct-2507 is your top choice.

Question 2: Which Model is the Most Tunable?

Winner: Llama-3.2-1B-Instruct (average rank: 3.44)

Here we measure tunability — the improvement from base to fine-tuned performance (finetuned_score - base_score). A highly tunable model starts weaker but gains dramatically from fine-tuning.

ModelAverage Rank95% CI
Llama-3.2-1B-Instruct3.44±1.31
Llama-3.2-3B-Instruct4.67±1.93
Qwen3-0.6B4.78±1.78
SmolLM2-1.7B-Instruct5.00±1.46
gemma-3-270m-it5.00±2.77

Interestingly, the tunability ranking inverts the size hierarchy. Smaller models like Llama-3.2-1B and Qwen3-0.6B show the largest gains from fine-tuning. The largest models (Qwen3-8B, granite-3.3-8b) rank near the bottom for tunability - not because they’re worse, but because they start strong and have less room to improve.

Key takeaway: If you’re constrained to very small models (<2B parameters), don’t despair. These models benefit most from fine-tuning and can close much of the gap to larger models.

Question 3: Which Model Has the Best Base Performance?

Winner: Qwen3-8B (average rank: 1.75)

Before any fine-tuning, which model performs best out of the box?

ModelAverage Rank95% CI
Qwen3-8B1.75±0.72
granite-3.3-8b-instruct2.57±0.84
Qwen3-4B-Instruct-25073.75±1.27
Llama-3.1-8B-Instruct4.14±2.11
Qwen3-1.7B4.78±1.02

As expected, base performance correlates with model size. The 8B models claim the top spots, with Qwen3-8B showing remarkably consistent performance across all benchmarks (lowest standard deviation).

Key takeaway: If you need strong zero-shot/few-shot performance without fine-tuning, larger models are still your best bet. But remember - this advantage shrinks after fine-tuning.

Question 4: Can Our Best Student Match the Teacher?

Yes. Qwen3-4B-Instruct-2507 matches or exceeds the teacher on 7 out of 8 benchmarks.

BenchmarkTeacherQwen3-4B FinetunedQwen3-4B BaseΔ (Student vs Teacher)
TREC0.890.930.51+0.03
Banking770.920.890.87-0.03
Docs0.820.840.64+0.02
Ecommerce0.880.900.75+0.03
HotpotQA0.930.930.88+0.00
Mental Health0.810.820.78+0.01
Roman Empire QA0.750.800.65+0.05
SQuAD 2.00.520.710.26+0.19

The fine-tuned 4B student beats the 120B+ teacher on 6 benchmarks, ties on 1 (HotpotQA), and falls slightly short on 1 (Banking77, within margin of error). The most dramatic improvement is on SQuAD 2.0 closed-book QA, where the student exceeds the teacher by 19 percentage points — a testament to how fine-tuning can embed domain knowledge more effectively than prompting.

Key takeaway: A 4B parameter model, properly fine-tuned, can match or exceed a model 30x its size. This translates to ~30x lower inference costs and the ability to run entirely on-premises.

Practical Recommendations

Based on our benchmarks, here’s how to choose your base model:

Your ConstraintRecommended ModelWhy
Maximum accuracyQwen3-4B-Instruct-2507Best fine-tuned performance across tasks
Very limited compute (<2B params)Llama-3.2-1B-Instruct or Qwen3-0.6BHighest tunability — gains most from fine-tuning
No fine-tuning possibleQwen3-8BBest zero-shot/few-shot performance
Edge deployment (mobile, IoT)Qwen3-0.6BGood tunability at minimal size

Next Steps

This benchmark is just a starting point, and we are actively working to make these results even more robust:

  • Evaluate more models: The SLM landscape evolves rapidly. We plan to add newer releases like Qwen3.5, Phi-4, and Mistral variants as they become available.
  • Run more iterations: Currently, our results are averaged over a limited number of runs. We’re expanding to more iterations per benchmark to tighten confidence intervals and ensure the rankings are statistically reliable.
  • Expand benchmark coverage: We want to include additional task types like summarization, code generation, and multi-turn dialogue to give a fuller picture of model capabilities.

We’ll update this post as new data comes in. If there’s a specific model or task you’d like us to benchmark, let us know!

Training Details

Each model was fine-tuned on synthetic data generated using our distillation pipeline (see Small Expert Agents from 10 Examples for details on the data generation process). For each benchmark, we generated 10,000 training examples using the teacher model (GPTOss-120B).

Fine-tuning was performed using the default distil labs configuration: 4 epochs with learning rate of 5e-5, linear learning rate scheduler and LoRA with rank 64.

All models were trained with identical hyperparameters. Evaluation was performed on held-out test sets that were never seen during training or synthetic data generation.

Conclusion

Not all small models are created equal but the differences narrow dramatically after fine-tuning. Our benchmarks show that Qwen3-4B-Instruct-2507 delivers the best overall fine-tuned performance, matching a 120B+ teacher while being deployable on a single consumer GPU. For extremely constrained environments, smaller models like Llama-3.2-1B show exceptional tunability and can close much of the gap.

The bottom line: fine-tuning matters more than base model choice. A well-tuned 1B model can outperform a prompted 8B model. If you have a task in mind, start with a short description and a handful of examples - we’ll take it from there.

Want to try this yourself? Sign up for distil labs and train your first expert agent in under 24 hours.

References