The Difference Between a Base LLM and an Instruction-Tuned LLM
Introduction
Large language models (LLMs) can behave very differently depending on how they were trained.
BaseLLMs are trained purely on next-token prediction over a large corpus of text. Instruction-tuned LLMs, by contrast, are further trained to follow prompts in a more helpful and structured way.
To explore how these two types of language models behave, we will take a look at two models from Hugging Face's SmolLM family:
- SmolLM on Hugging Face: a base model trained solely with next-token prediction
- SmolLM-Instruct on Hugging Face: the same model, but further fine-tuned on prompt–response pairs