You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It installs MaxText and then for post-training, it installs primarily the following:
65
-
66
-
a. [Tunix](https://github.com/google/tunix) as the LLM Post-Training Library, and
67
-
68
-
b. `vllm-tpu` which is
69
-
[vllm](https://github.com/vllm-project/vllm) and
70
-
[tpu-inference](https://github.com/vllm-project/tpu-inference) and thereby
71
-
providing TPU inference for vLLM, with unified JAX and PyTorch support.
72
-
73
-
### Option 2: From Github
74
-
75
-
For using a version newer than the latest PyPI release, you could also install the latest vetted versions of the dependencies from MaxText in the following way:
For instructions on installing MaxText with post-training dependencies on your VM, please refer to the [official documentation](https://maxtext.readthedocs.io/en/latest/install_maxtext.html) and use the `maxtext[tpu-post-train]` installation path to include all necessary post-training dependencies.
For instructions on installing MaxText with post-training dependencies on your VM, please refer to the [official documentation](https://maxtext.readthedocs.io/en/latest/install_maxtext.html) and use the `maxtext[tpu-post-train]` installation path to include all necessary post-training dependencies.
43
30
44
31
## Setup environment variables
45
32
@@ -53,7 +40,7 @@ Set the following environment variables before running SFT.
export BASE_OUTPUT_DIRECTORY=<output directory to store run logs># e.g., gs://my-bucket/my-output-directory
@@ -76,15 +63,15 @@ This section explains how to prepare your model checkpoint for use with MaxText.
76
63
If you already have a MaxText-compatible model checkpoint, simply set the following environment variable and move on to the next section.
77
64
78
65
```sh
79
-
exportPRE_TRAINED_MODEL_CKPT_PATH=<gcs path for MaxText checkpoint># e.g., gs://my-bucket/my-model-checkpoint/0/items
66
+
exportMAXTEXT_CKPT_PATH=<gcs path for MaxText checkpoint># e.g., gs://my-bucket/my-model-checkpoint/0/items
80
67
```
81
68
82
69
### Option 2: Converting a Hugging Face checkpoint
83
70
84
71
Refer the steps in [Hugging Face to MaxText](../../guides/checkpointing_solutions/convert_checkpoint.md#hugging-face-to-maxtext) to convert a hugging face checkpoint to MaxText. Make sure you have correct checkpoint files converted and saved. Similar as Option 1, you can set the following environment and move on.
85
72
86
73
```sh
87
-
exportPRE_TRAINED_MODEL_CKPT_PATH=<gcs path for MaxText checkpoint># e.g., gs://my-bucket/my-model-checkpoint/0/items
74
+
exportMAXTEXT_CKPT_PATH=<gcs path for MaxText checkpoint># e.g., gs://my-bucket/my-model-checkpoint/0/items
88
75
```
89
76
90
77
## Run SFT on Hugging Face Dataset
@@ -95,8 +82,8 @@ Now you are ready to run SFT using the following command:
Refer the steps in [Hugging Face to MaxText](../../guides/checkpointing_solutions/convert_checkpoint.md#hugging-face-to-maxtext) to convert a hugging face checkpoint to MaxText. Make sure you have correct checkpoint files converted and saved. Similar as Option 1, you can set the following environment and move on.
96
96
97
97
```bash
98
-
exportMODEL_CHECKPOINT_PATH=<gcs path for MaxText checkpoint># gs://my-bucket/my-checkpoint-directory/0/items
98
+
exportMAXTEXT_CKPT_PATH=<gcs path for MaxText checkpoint># gs://my-bucket/my-checkpoint-directory/0/items
0 commit comments