Skip to content

Commit e3dbd54

Browse files
Merge pull request #3370 from AI-Hypercomputer:490518973_fix
PiperOrigin-RevId: 882042990
2 parents 05c5083 + 22fc9e6 commit e3dbd54

2 files changed

Lines changed: 21 additions & 50 deletions

File tree

.github/workflows/run_jupyter_notebooks.yml

Lines changed: 0 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -96,20 +96,6 @@ jobs:
9696
9797
papermill "$notebook" "$output_name" -k maxtext_venv
9898
done
99-
- name: Record Commit IDs
100-
shell: bash
101-
run: |
102-
echo "--- MaxText and Post-Training Repositories Commit IDs ---"
103-
echo "maxtext: ${GITHUB_SHA:0:7}"
104-
105-
declare -a repos=("tunix" "vllm" "tpu-inference")
106-
for repo_dir in "${repos[@]}"; do
107-
if [ -d "$repo_dir" ]; then
108-
echo "$repo_dir: $(git -C "$repo_dir" rev-parse --short HEAD)"
109-
else
110-
echo "Warning: $repo_dir directory not found."
111-
fi
112-
done
11399
- name: Upload Outputs
114100
if: always()
115101
uses: actions/upload-artifact@v4

docs/tutorials/posttraining/rl_on_multi_host.md

Lines changed: 21 additions & 36 deletions
Original file line numberDiff line numberDiff line change
@@ -122,20 +122,24 @@ gcloud auth configure-docker
122122
docker run hello-world
123123
```
124124

125-
### Option 1: Install stable releases of post-training dependencies
125+
### Option 1: From PyPI releases (Recommended)
126126

127-
> **Caution:** RL in MaxText is currently broken with stable releases of
128-
> post-training dependencies. We are working on fixing this and recommend
129-
> following
130-
> [Option 2: Install from Git repositories of post-training dependencies](#option-2-install-from-git-repositories-of-post-training-dependencies)
131-
> in the meantime.
127+
Get the latest stable release of MaxText from PyPI. This will automatically pull
128+
compatible versions of post-training dependencies, such as [Tunix](https://github.com/google/tunix),
129+
[vLLM](https://github.com/vllm-project/vllm), and
130+
[tpu-inference](https://github.com/vllm-project/tpu-inference).
131+
132+
```bash
133+
git clone https://github.com/AI-Hypercomputer/maxtext.git
134+
cd maxtext
135+
136+
# checkout the latest stable release here: https://pypi.org/project/maxtext/
137+
export MAXTEXT_VERSION=0.2.0
138+
git checkout maxtext-v${MAXTEXT_VERSION?}
139+
```
132140

133141
Run the following script to create a Docker image with stable releases of
134-
MaxText, [Tunix](https://github.com/google/tunix),
135-
[vLLM](https://github.com/vllm-project/vllm), and
136-
[tpu-inference](https://github.com/vllm-project/tpu-inference) dependencies.
137-
This installs `vllm-tpu` which provides TPU inference for vLLM with unified JAX
138-
and PyTorch support. The build process takes approximately 10-15 minutes.
142+
MaxText, and its post-training dependencies. The build process takes approximately 10-15 minutes.
139143

140144
```bash
141145
bash dependencies/scripts/docker_build_dependency_image.sh WORKFLOW=post-training
@@ -147,34 +151,15 @@ For experimental features (such as improved pathwaysutils resharding API), use:
147151
bash dependencies/scripts/docker_build_dependency_image.sh WORKFLOW=post-training-experimental
148152
```
149153

150-
### Option 2: Install from Git repositories of post-training dependencies
151-
152-
You can also locally clone the [tunix](https://github.com/google/tunix),
153-
[tpu-inference](https://github.com/vllm-project/tpu-inference), and
154-
[vllm](https://github.com/vllm-project/vllm.git) repositories and then build the
155-
docker image with these local sources. To get a set of compatible commit IDs for
156-
`maxtext`, `tunix`, `tpu-inference`, and `vllm`, follow these steps:
154+
### Option 2: From Github
157155

158-
1. Navigate to the
159-
[MaxText Package Tests](https://github.com/AI-Hypercomputer/maxtext/actions/workflows/build_and_test_maxtext.yml?query=event%3Aschedule)
160-
GitHub Actions workflow.
161-
162-
2. Select the latest successful run.
163-
164-
3. Within the workflow run, find and click on the `maxtext_jupyter_notebooks (py312)` job, then expand the `run` job.
165-
166-
4. Locate the `Record Commit IDs` step. The commit SHAs for `maxtext`, `tunix`,
167-
`tpu-inference`, and `vllm` that were used in that successful run are listed
168-
in the logs of this step.
169-
170-
5. Prior to installation, ensure that the `maxtext`, `tunix`, `vllm`, and `tpu-inference` repositories are synchronized to the specific commits recorded from the CI logs. For each repository, use the following command to switch to the correct commit: `git checkout <commit_id>`.
171-
172-
**Note:** Clone these repositories as siblings of the `maxtext` directory (e.g.,
173-
in the same parent directory). After cloning, run the build from inside the
174-
`maxtext` repository so it picks up the local sources:
156+
For using a version newer than the latest PyPI release, you could also build the Docker image with the latest vetted versions of post-training dependencies and MaxText in the following way:
175157

176158
```bash
177-
bash dependencies/scripts/docker_build_dependency_image.sh WORKFLOW=post-training POST_TRAINING_SOURCE=local
159+
git clone https://github.com/AI-Hypercomputer/maxtext.git
160+
cd maxtext
161+
162+
bash dependencies/scripts/docker_build_dependency_image.sh WORKFLOW=post-training
178163
```
179164

180165
### Upload the Docker Image

0 commit comments

Comments
 (0)