YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Templar-I: Permissionless Distributed Training

A 1.2B-parameter causal language model trained with Gauntlet, an incentive system that rewards permissionless contributors for useful pseudo-gradients on the Bittensor network. [Paper]


Overview

  • Setting: Fully open, permissionless, internet-scale training; no control over who registers or their hardware.
  • Mechanism: Two-stage peer filtering (uptime/reliability/sync) + scoring per-peer gradient quality.
  • Run: 20K communication rounds; FineWebEdu data; top 15 peers aggregated per round with up to 250 registered peers.
  • Result: On a per-iteration basis, convergence outpaced a centralized AdamW baseline; downstream metrics are competitive.

Gauntlet

  • Stage 1: Filters peers by uptime, reliability, and synchronization.
  • Stage 2: Estimates loss before/after applying each peer’s pseudo-gradients to evaluate its contribution.
  • Ratings: Uses OpenSkill to track competitiveness across time.
  • Aggregation: In each round, aggregate updates from the top-scoring G=15 peers.

Training setup

  • Data: FineWeb-edu [11].
  • Rounds: 20,000 communication rounds (evaluation windows matched rounds).
  • Tokens: 100-200B
  • Baseline for comparison: Centralized AdamW trained for 120B tokens.

Quickstart

from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

model_id = "tplr/TEMPLAR-I"
tok = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16, device_map="auto")

Results

Downstream Benchmarks (zero-shot)

Model Dataset Tokens HellaSwag (acc_norm) PIQA (acc_norm) ARC-E (acc)
TEMPLAR-1B FineWebEdu 100B–200B 51.0 71.4 59.2
DeMo 1B [12] Dolmo 100B 48.0 70.0 55.0
AdamW DDP 1B FineWebEdu 120B 51.0 71.9 58.9

Per-Iteration Loss

Training loss


Citation

If you use this model or Gauntlet, please cite it as follows:

@article{lidin2025incentivizing,
  title={Incentivizing Permissionless Distributed Learning of LLMs},
  author={Lidin, Joel and Sarfi, Amir and Pappas, Evangelos and Dare, Samuel and Belilovsky, Eugene and Steeves, Jacob},
  journal={arXiv preprint arXiv:2505.21684},
  year={2025}
}
Downloads last month
2
Safetensors
Model size
1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for tplr/TEMPLAR-I

Adapters
4 models