SPES-7B

SPES-7B is a pretrained language model released as part of paper:

Pretraining A Large Language Model using Distributed GPUs: A Memory-Efficient Decentralized Paradigm

Model Details

  • Model name: SPES-7B
  • Model type: Causal language model
  • Parameters: 7B
  • Framework: SPES
  • License: Apache-2.0

Project Links

Intended Use

This model is intended for:

  • research on decentralized LLM pretraining
  • research on MoE training and synchronization
  • experimentation and evaluation of pretrained language models

Citation

If you use this model, please cite the SPES paper.

@article{zhang2026spes,
  title={Pretraining A Large Language Model using Distributed GPUs: A Memory-Efficient Decentralized Paradigm},
  author={Zhang, Jinrui and Xiao, Chaodong and Wu, Aoqi and Zhang, Xindong and Zhang, Lei},
  year={2026}
}
Downloads last month
13
Safetensors
Model size
7B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including zjr2000/SPES-7B

Paper for zjr2000/SPES-7B