Releases: AI-Hypercomputer/maxtext
Releases · AI-Hypercomputer/maxtext
maxtext-v0.2.0
Changes
- Qwen3-Next is now supported.
- New
tpu-post-traintarget in PyPI. Please also use this installation option for running vllm_decode. See the MaxText installation instructions for more info. - New MaxText structure! MaxText has been restructured according to RESTRUCTURE.md. Please feel free to share your thoughts and feedback.
- Muon optimizer is now supported.
- DeepSeek V3.1 is now supported. Use existing configs for DeepSeek V3 671B and load in V3.1 checkpoint to use model.
- New RL and SFT Notebook tutorials are available.
- The ReadTheDocs documentation site has been reorganized.
- Multi-host support for GSPO and GRPO is now available via new RL tutorials.
- A new guide, What is Post Training in MaxText?, is now available.
- Ironwood TPU co-designed AI stack announced. Read the blog post on its co-design with MaxText.
- Optimized models tiering documentation has been refreshed.
- Added Versioning. Check out our first set of release notes!
- Post-Training (SFT, RL) via Tunix is now available.
- Vocabulary tiling (PR) is now supported in MaxText! Adjust config
num_vocab_tilingto unlock more efficient memory usage. - The GPT-OSS family of models (20B, 120B) is now supported.
Deprecations
- Many MaxText modules have changed locations. Core commands like train, decode, sft, etc. will still work as expected temporarily. Please update your commands to the latest file locations
- install_maxtext_github_deps installation script replaced with install_maxtext_tpu_github_deps
tools/setup/setup_post_training_requirements.shfor post training dependency installation is deprecated in favor of pip installation
maxtext-tutorial-v1.5.0
Merge pull request #2898 from AI-Hypercomputer:tests_docker_image PiperOrigin-RevId: 850456883
maxtext-tutorial-v1.4.0
maxtext-tutorial-v1.4.0
maxtext-tutorial-v1.3.0
Merge pull request #2706 from AI-Hypercomputer:mohit/tokamax_quant_gmm PiperOrigin-RevId: 834605168
maxtext-tutorial-v1.2.0: Merge pull request #2676 from AI-Hypercomputer:pypi_release
PiperOrigin-RevId: 832378885
Recipe Branch for TPU performance results
Merge pull request #2539 from AI-Hypercomputer:qinwen/latest-tokamax PiperOrigin-RevId: 823749360
maxtext-tutorial-v1.0.0
Merge pull request #2538 from AI-Hypercomputer:mohit/fix_docker PiperOrigin-RevId: 822796389
tpu-recipes-v0.1.5
Use this release for tpu-recipes that require version tpu-recipes-v0.1.5
maxtext-v0.1.0
Our first MaxText PyPI package is here! MaxText is a high performance, highly scalable, open-source LLM library and reference implementation written in pure Python/JAX and targeting Google Cloud TPUs and GPUs for training. We are excited to make it easier than ever to get started.
Users can now install MaxText through pip, both for local development and through stable PyPI builds. Please see our MaxText Installation Guide for more setup details.
Going forward, this page will document notable changes as we release new versions of MaxText.
tpu-recipes-v0.1.4
Use this release for tpu-recipes that require version tpu-recipes-v0.1.4