Skip to content

Releases: AI-Hypercomputer/maxtext

pre-nnx-v0.1.0

16 Jun 22:46
Compare
Choose a tag to compare

Use this release for the latest MaxText version that fully depends on Flax Linen (no NNX).

tpu-recipes-v0.1.3

03 Jun 18:04
Compare
Choose a tag to compare

Use this release for tpu-recipes that require version tpu-recipes-v0.1.3

tpu-recipes-v0.1.2

16 Apr 19:33
Compare
Choose a tag to compare

Use this release for tpu-recipes that require version tpu-recipes-v0.1.2

tpu-recipes-v0.1.1

11 Apr 02:29
Compare
Choose a tag to compare

Use this release for tpu-recipes that require version tpu-recipes-v0.1.1

pre-module-v0.1.0

08 Apr 21:14
Compare
Choose a tag to compare
pre-module-v0.1.0 Pre-release
Pre-release

Release prior to module refactor for older train API.

With this release or prior: python3 MaxText/train.py MaxText/configs/base.yml run_name=...

Soon after this release: python3 -m MaxText.train MaxText/configs/base.yml run_name=...

tpu-recipes-v0.1.0

02 Apr 04:50
Compare
Choose a tag to compare

Use this release for tpu-recipes that require version tpu-recipes-v0.1.0

MoE v1.0.0

10 Sep 06:36
Compare
Choose a tag to compare

MoE v1.0.0 supports:

  • Megablox with Fully Sharded Data Parallelism (FSDP) and Token Parallelism (TP)
  • Dropping strategies with FSDP, TP, and Expert Parallelism (EP)