DiffutronLM-0.3B-Instruct

Diffutron is a parameter-efficient, Masked Diffusion Language Model (MDLM) specifically designed for the Turkish language. Unlike standard autoregressive models that generate text one token at a time, Diffutron generates text by iteratively refining sequences in parallel, allowing for simultaneous consideration of the entire sentence context.

Despite its compact size of 307 million parameters, DiffutronLM-0.3B-Instruct achieves highly competitive performance against much larger, multi-billion-parameter autoregressive baselines on Turkish NLP benchmarks.

πŸ“Œ Model Details

  • Model Type: Masked Diffusion Language Model (MDLM)
  • Base Architecture: jhu-clsp/mmBERT-base (Multilingual Encoder)
  • Language: Turkish
  • Parameter Count: 307M (0.3B)
  • Context Length: 256 tokens (Instruct), 512 tokens (Base)
  • Training Libraries: dllm, PyTorch

πŸš€ Architecture & Approach

Diffutron departs from traditional next-token prediction. It treats text generation as a discrete diffusion process:

  1. Forward Process: Clean text is gradually corrupted into a sequence of <mask> tokens.
  2. Reverse Process: The model learns to denoise the sequence globally, attending to visible context bi-directionally to predict the original tokens.

This non-autoregressive paradigm compresses linguistic knowledge efficiently, allowing this 0.3B model to punch significantly above its weight class.

πŸ“š Training Pipeline

The model was developed through a resource-efficient, multi-stage training pipeline:

1. Continual Pre-training (CPT)

To adapt the multilingual backbone to Turkish without catastrophic forgetting, we employed a high-rank LoRA strategy (r=256, Ξ±=256) targeting all linear modules (Attention and MLP).

  • Data: ~2 million sequences sourced from Havadis (news), Temiz-OSCAR (web), and Turkish Wikipedia.
  • Result: Perplexity on the Bilkent Turkish Writings Dataset dropped significantly from 3.42 (base) to 2.75.

2. Progressive Instruction-Tuning (SFT)

To unlock generative instruction-following capabilities, we utilized a two-stage supervised fine-tuning approach:

  • Stage 1 (General Adaptation): Trained on metunlp/LlamaTurk-Instruction-Set for 20 epochs to establish fundamental instruction-following behaviors.
  • Stage 2 (Complex Specialization): Trained on the nuanced turkish-nlp-suite/InstrucTurca dataset for 8 epochs with an increased batch size, enhancing the model's ability to handle intricate, domain-specific Turkish commands.

πŸ“Š Evaluation Results

The model was evaluated on a representative subset of the CETVEL Benchmark Suite. DiffutronLM-0.3B (2nd Stage) demonstrates remarkable parameter efficiency, outperforming models up to 7x its size (e.g., Kumru-2B and TURNA-1.1B) on average scores.

Benchmark Diffutron-1st-Stage (0.3B) Diffutron-2nd-Stage (0.3B) TURNA (1.1B) Kumru (2B) Kanarya (2B) Llama-3.2 (3B) Trendyol (7B) Aya-101 (13B)
Belebele_TR 22.22 27.00 22.56 29.00 28.11 55.78 36.22 22.89
EXAMS_TR 25.95 27.74 23.66 30.03 30.03 26.21 28.50 22.90
IronyTR 50.67 52.00 48.33 51.00 50.00 50.17 50.00 52.17
News_Cat 23.20 32.40 32.80 26.40 66.80 64.00 81.20 20.00
MNLI_TR 33.29 32.81 34.94 36.42 33.40 34.76 35.19 27.90
STS_TR 17.77 18.78 14.21 11.75 12.91 12.91 15.52 16.97
XCOPA_TR 53.80 52.00 55.80 54.00 64.20 54.60 61.00 59.60
Average 32.41 34.68 33.19 34.09 40.78 42.63 43.95 31.78

πŸ’» Usage

Because Diffutron is a Masked Diffusion Language Model, it requires inference strategies distinct from standard causal generation. We recommend using the dllm library or custom generation loops tailored for discrete diffusion.

Generation Parameters Used in Paper:

  • Longer Context: Steps: 128, Temp: 0.1, Block Length: 32, Repetition Penalty: 1.2
  • Shorter Context: Steps: 64, Remask: low_conf, Stochastic: False, CFG: 0.0

⚠️ Limitations

  • Multilingual Backbone: Built upon a multilingual encoder rather than a native Turkish foundation model.
  • Context Window: Restricted to a 256-token context window for generation, limiting its use in long-form summarization or document-level generation.
  • Data Nuances: Instruction datasets rely heavily on translations or synthetic data, which may occasionally miss subtle cultural contexts.

πŸ“ Citation

If you use Diffutron in your research, please cite our preprint:

@misc{diffutron2026,
  author = {Kocabay, Şuayp Talha and Akkuş, Talha Rüzgar},
  title = {Diffutron: A Masked Diffusion Language Model for Turkish Language},
  year = {2026},
  publisher = {Hugging Face},
  howpublished = {\url{https://huggingface.co/collections/diffutron/diffutronlm}}
}
Downloads last month
12
Safetensors
Model size
0.3B params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for diffutron/DiffutronLM-0.3B-Instruct

Finetuned
(1)
this model

Dataset used to train diffutron/DiffutronLM-0.3B-Instruct

Collection including diffutron/DiffutronLM-0.3B-Instruct