An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
Cross-referenced across 55 tracked directories
#638
Popularity Rank
1 / 55
Listed In
Emerging
Adoption Stage
12/22/2020
Created
7,400
GitHub Stars
Score: 100/100
0 dependency vulnerabilities found
Run an AI-powered security scan to analyze this package's source code for vulnerabilities, prompt injection vectors, data exfiltration risks, and behavior mismatches.
Scans fetch actual source code from the GitHub repository, not just the README.
Open-source framework for fine-tuning and evaluating LLMs. It simplifies the process of experimenting with different training configurations and makes it easy to reproduce and share results, supporting features like LoRA, QLoRA, DeepSpeed, PEFT, and multi-GPU setups.
...moreveRL is a flexible and efficient RL framework for LLMs.
A simple, performant and scalable Jax LLM!
Leandro von Werra <leandro.vonwerra@gmail.com>
Train transformer language models with reinforcement learning.
1,100
Forks
93
Open Issues
2/3/2026
Last Commit
Recently added to the ecosystem