Brandon-Tiny 10M: A 3-Phase Training Pipeline for Ultra-Small Instruction-Following Language Models

Samuel Cortes · February 2026
Code: github.com/xaskasdf/brandon-tiny · Model: HuggingFace · PDF

Abstract

We present Brandon-Tiny 10M, a 10.7M parameter instruction-following language model that achieves competitive performance with models 3x its size through a novel 3-phase training pipeline combining foundation pre-training, knowledge distillation, and instruction fine-tuning. Built on a deep-narrow Llama 2 architecture enhanced with DenseFormer weighted averaging, Value Residual connections, and register tokens, our model demonstrates that careful training orchestration can compensate for limited parameter budgets. Trained entirely on a single RTX 3090 GPU, Brandon-Tiny 10M achieves a fine-tuning validation loss of 2.40 (surpassing our own 30M parameter models at 2.61), generates coherent text, follows instructions at 80% accuracy, and exhibits zero repetition loops across 20 free-generation tests. We describe our complete experimental journey across 8 model variants and analyze the key factors contributing to performance at this extreme scale.

1. Introduction

The dominant trend in language modeling has been scaling up: more parameters, more data, more compute. Yet practical deployment often demands the opposite — models that run on edge devices, embedded systems, or resource-constrained environments. While research on efficient LLMs has flourished (MobileLLM, SmolLM2, Phi), most work targets the 125M–7B range. Models under 50M parameters remain largely unexplored for instruction-following tasks.

Our motivation was concrete: we wanted a language model capable of running natively on a PlayStation 2's Emotion Engine, which has 32 MB of RAM and 4 MB of VRAM. Inspired by TinyStories (Eldan & Li, 2023), we searched for existing models small enough for this constraint and found none with instruction-following capabilities. The project's name — Brandon Tiny — originated from a Cloudflare tunnel URL (sugar-alaska-brandon-tiny.trycloudflare.com) generated while serving a custom agentic chat, which sounded so much like a language model name that it stuck.

We address this gap by systematically exploring what is achievable at 10.7M parameters. Our key contributions:

  1. A 3-phase training pipeline (Pretrain → Distill → Finetune) specifically designed for ultra-small models, where each phase compensates for limitations of the previous one.
  2. Extensive ablation across 8 model variants, revealing that training methodology matters more than raw parameter count at this scale.
  3. Anti-repetition training techniques combining label smoothing, unlikelihood training, and entropy regularization that completely eliminate generation loops.
  4. Practical guidelines for training sub-50M parameter instruction models on consumer hardware.

3. Architecture

3.1 Base Architecture

Brandon-Tiny 10M follows the Llama 2 architecture with modifications from MobileLLM (deep-narrow design with block sharing):

ComponentSpecification
Dimensionsdim=256, hidden=720
Layers24 (12 unique, block sharing)
Attention8 heads, 2 KV heads (GQA 4:1)
Vocabulary8,192 (SentencePiece BPE)
Max sequence512 tokens
ActivationSwiGLU
NormalizationRMSNorm
Position encodingRoPE (theta=10000)
Total parameters10,706,776

3.2 Architectural Enhancements

Three enhancements from recent literature, validated on smaller experiments before applying to the final model:

DenseFormer (Pagliardini et al., 2024): Depth-Weighted Averaging (DWA) adds learnable weighted connections from all previous layer outputs to each subsequent layer, improving gradient flow in deep-narrow architectures.

Value Residual Learning (Wang et al., 2024): The value projection from layer 0 is added to all subsequent layers' value computations, preserving early representations that tend to be lost in deep networks.

Register Tokens (Darcet et al., 2024): 4 learnable tokens prepended to the input sequence act as "information sinks" for attention, reducing attention entropy collapse observed in small models.

3.3 MobileLLM Block Sharing

Adjacent pairs of layers share weights, effectively giving 12 unique parameter blocks applied twice. This halves the parameter count while maintaining depth, following the finding from MobileLLM that depth matters more than width for small models.

3.4 Design Rationale

The deep-narrow + block sharing design was chosen over standard proportional scaling based on our ablation studies:

Model VariantArchitecturePretrain LossFinetune Loss
10M v2 baselinedim=192, 16 layers, sharing1.923.92
10M Enhanced v2dim=256, 24 layers, sharing + DWA + VR + Reg3.732.92

The enhanced architecture trades pretrain perplexity for dramatically better downstream performance, suggesting that the enhancements improve the model's ability to compress knowledge into limited parameters.

4. Training Pipeline

4.1 Overview

Our 3-phase pipeline addresses the core challenge of ultra-small models: they lack the capacity to learn everything from scratch. Each phase adds a different type of knowledge:

Phase 1: Foundation Pretrain    → Language modeling basics
Phase 2: Knowledge Distillation → Compressed knowledge from larger teacher
Phase 3: Instruction Finetune   → Task-specific behavior + anti-repetition

4.2 Phase 1: Foundation Pre-training

Data: Mixed corpus of 600M tokens — 40% Wikipedia English, 30% SmolLM Corpus, 30% Synthetic data (generated via GPT-4o-mini across diverse topics).

Configuration: Learning rate 8e-4 with WSD schedule (MiniCPM), 15,000 steps, effective batch 65,536 tokens/step. Warmup 500 steps, Stable 70%, Decay 20%. Weight decay 0.1, gradient clipping 1.0, bfloat16.

Result: val_loss = 4.39, perplexity = 80.8. The WSD schedule's decay phase consistently delivered the largest quality improvements (often 0.3–0.5 loss reduction).

4.3 Phase 2: Knowledge Distillation

Distillation from a pre-trained 30M parameter teacher (Brandon-Tiny 30M v2, val_loss 3.26) into our 10M student — a 2.8x compression ratio.

Configuration: Reverse KL Divergence (mode-seeking), temperature 2.0, alpha 0.5, learning rate 4e-4 with WSD, 7,500 steps.

Why Reverse KLD? Following MiniPLM (ICLR 2025), reverse KLD is mode-seeking rather than mean-seeking, which better suits small students that cannot cover the full distribution of the teacher. The student learns to focus on the teacher's most confident predictions.

Result: val_loss = 4.84 (measured against hard targets; soft-target loss was much lower).

4.4 Phase 3: Instruction Fine-tuning

Data: Merged instruction dataset (75,502 examples) — 57K curated chat instructions, 19,944 reasoning/CoT examples, ~200 pretrain replay examples (to prevent catastrophic forgetting).

Chat Format: ChatML with <|im_start|> / <|im_end|> tokens, loss computed only on assistant responses.

Anti-repetition Training: Label smoothing (0.1), Unlikelihood training (alpha=0.5), Entropy regularization (beta=0.01). This triple-technique approach eliminates severe repetition observed in early versions.

Configuration: Learning rate 2e-5, cosine schedule, 12,000 steps, weight decay 0.01.

Result: val_loss = 2.3995 (new record across all models).

5. Experiments and Ablations

5.1 Model Variants

ModelParamsArchitecturePretrain LossFinetune Loss
10M v2 baseline10.7MDeep-narrow, sharing1.923.92
10M MTP10.7M+ Multi-Token Prediction*2.103.45
10M Enhanced10.7M+ DWA + VR + Registers3.732.92
10M Dream10.7MTernary + Looped3.812.98
10M Synthetic-only10.7MEnhanced, synthetic pretrain1.963.62
10M Optimal10.7MEnhanced + 3-phase4.392.40
30M v2 (reference)30.0MDeep-narrow, sharing3.262.61

*MTP: Multi-Token Prediction (Gloeckle et al., 2024), where the model predicts the next N tokens simultaneously via N auxiliary heads. Improved pretrain loss regularity but did not translate to downstream gains at this scale.

5.2 Key Findings

Finding 1: Low pretrain loss ≠ good downstream performance. The synthetic-only model achieved the best pretrain loss (1.96) but one of the worst finetune losses (3.62). Conversely, the Optimal model had the highest pretrain loss (4.39) but the best finetune loss (2.40).

Finding 2: The 3-phase pipeline enables 10M to beat 30M. Our 10M Optimal (val_loss 2.40) significantly outperforms our best 30M model (val_loss 2.61), despite having 3x fewer parameters. Caveat: validation sets differ between model variants, so this comparison reflects training pipeline effectiveness rather than a controlled head-to-head.

Finding 3: Anti-repetition training is essential at this scale. Without the triple anti-repetition techniques, all models exhibited some degree of repetitive generation. With it, zero sequence loops across 20 free-generation tests.

Finding 4: Architectural enhancements compound. DenseFormer + Value Residual + Registers together improve finetune loss by ~1.0 point compared to baseline.

Finding 5: Ternary quantization doesn't work at this scale. The Dream architecture resulted in the worst instruction-following rate (28%) with incoherent generation.

5.3 Benchmark Results

Standard academic benchmarks (0-shot, 1000 examples per task):

BenchmarkScoreRandomDelta
BLiMP (Grammar)73.3%50.0%+23.3
HellaSwag (Commonsense)32.4%25.0%+7.4
ARC-Easy (Science)30.6%25.0%+5.6
PIQA (Physical Intuition)54.7%50.0%+4.7
Winogrande (Coreference)50.3%50.0%+0.3
LAMBADA (Last Word)8.8%0.0%+8.8
Wikitext-2 PPL224.2

BLiMP is the standout result, with several subtasks scoring above 90%: sentential negation NPI licensor (100%), existential there quantifiers (95.2%), wh-vs-that no gap long distance (95.1%).

For comparison, the Super Tiny Language Models paper (2405.14159) reported their 50M model scoring 25.6% on HellaSwag and 21% on ARC-Easy. Our 10.7M model outperforms both. Note: this comparison should be interpreted cautiously — differences in tokenizer vocabulary (8K vs 32K), evaluation subset size, and data composition may contribute to the gap.

6. Qualitative Analysis

6.1 Generation Examples

Identity awareness:

User: What is your name?
Brandon-Tiny: I'm Brandon Tiny, a small language model designed
to help in limited environments.

Creative writing:

User: Tell me a short story about a cat.
Brandon-Tiny: Once upon a time, there was an old but young cat
named Charlie who lived in the small town of Pagima. Every day,
he set out on adventures...

Factual (with hallucination):

User: What is the capital of France?
Brandon-Tiny: The capital of the French Empire was Goé. [HALLUCINATION]

6.2 Failure Modes

  • Factual hallucination: Confidently generates plausible but incorrect facts
  • Mathematical inability: Cannot reliably perform arithmetic
  • Vocabulary artifacts: Occasionally produces garbled words ("cocooked", "tated")
  • Template pollution: Some responses include training artifacts

7. Infrastructure

7.1 Hardware

All training on a single NVIDIA RTX 3090 (24GB VRAM):

  • Phase 1: ~3 hours (15K steps)
  • Phase 2: ~1.5 hours (7.5K steps)
  • Phase 3: ~2 hours (12K steps)
  • Total: ~6.5 hours wall clock time

7.2 Inference

  • Speed: 21 tokens/second on RTX 3090
  • VRAM: 51.5 MB allocated
  • Model size: 42.8 MB (fp32), 21.4 MB (bf16)

7.3 Tokenizer

Custom SentencePiece BPE tokenizer with 8,192 vocabulary, trained on TinyStories + SmolLM text. Includes ChatML special tokens (<|im_start|>, <|im_end|>, <|bos|>, <|eos|>, <|pad|>).

8. Limitations

  • Not suitable for factual tasks: The model hallucinates freely.
  • English only: Trained exclusively on English data.
  • Short context: 512 token maximum sequence length.
  • No safety alignment: No RLHF/DPO training.
  • Evaluation challenges: Standard LLM benchmarks (MMLU, GSM8K) are meaningless at this scale.

9. Conclusion

Brandon-Tiny 10M demonstrates that with careful architecture design, multi-phase training, and knowledge distillation, a 10.7M parameter model can achieve instruction-following capabilities previously associated with much larger models. Our 3-phase pipeline (Pretrain → Distill → Finetune) is the key innovation, enabling the model to outperform a 30M parameter counterpart on fine-tuning loss. The complete training pipeline runs in under 7 hours on a single consumer GPU.

Since this paper's completion, we have extended Brandon-Tiny into a Vision-Language-Action (VLA) controller for robotics, adding a MobileNetV3-Small vision encoder and 25 discrete action tokens for motor control (see companion paper: Brandon-VLA). The VLA pipeline adds three additional training phases: VLA adaptation, vision conditioning, and GRPO online reinforcement learning in a MuJoCo simulation environment.

Additional future work includes: (1) scaling the 3-phase pipeline to our 110M model, (2) quantization for true edge deployment, and (3) real-world sim-to-real transfer of the VLA controller.

10. Reproducibility

All code, configurations, and training scripts are available at github.com/xaskasdf/brandon-tiny. The complete training pipeline can be run with:

python scripts/train_10m_optimal.py

Model checkpoints and the tokenizer are available on HuggingFace.

References

  • Liu, Z., et al. (2024). MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases. ICML 2024.
  • Allal, L.B., et al. (2025). SmolLM2: When Smol Goes Big — Data-Centric Training of a Small Language Model. arXiv:2502.02737.
  • Hillier, D., Guertler, L., Tan, C., Agrawal, P., Chen, R., Cheng, B. (2024). Super Tiny Language Models. arXiv:2405.14159.
  • Pagliardini, M., Mohtashami, A., Fleuret, F., Jaggi, M. (2024). DenseFormer: Enhancing Information Flow in Transformers via Depth Weighted Averaging. arXiv:2402.02622.
  • Wang, Z., et al. (2024). Value Residual Learning For Alleviating Attention Concentration In Transformers. arXiv:2410.17897. ACL 2025.
  • Darcet, T., Oquab, M., Mairal, J., Bojanowski, P. (2024). Vision Transformers Need Registers. ICLR 2024.
  • Hu, S., et al. (2024). MiniCPM: Unveiling the Potential of Small Language Models with Scalable Training Strategies. arXiv:2404.06395.
  • Gu, Y., Zhou, H., Meng, F., Zhou, J., Huang, M. (2025). MiniPLM: Knowledge Distillation for Pre-Training Language Models. ICLR 2025. arXiv:2410.17215.
  • Welleck, S., et al. (2020). Neural Text Generation with Unlikelihood Training. ICLR 2020.
  • Touvron, H., et al. (2023). Llama 2: Open Foundation and Fine-Tuned Chat Models. arXiv:2307.09288.
  • Eldan, R. & Li, Y. (2023). TinyStories: How Small Can Language Models Be and Still Speak Coherent English? arXiv:2305.07759.
  • Gloeckle, F., et al. (2024). Better & Faster Large Language Models via Multi-Token Prediction. ICML 2024.
  • Su, J., et al. (2024). RoFormer: Enhanced Transformer with Rotary Position Embedding. Neurocomputing.
  • Shazeer, N. (2020). GLU Variants Improve Transformer. arXiv:2002.05202.
  • Ainslie, J., et al. (2023). GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints. EMNLP 2023.