Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
68 commits
Select commit Hold shift + click to select a range
391a1fc
one step off async training recipe
imh966 Jun 27, 2025
338c2a9
simplify trainer
imh966 Jun 30, 2025
071ddc2
fix resource pool config and simplify the trainer yaml file
imh966 Jul 1, 2025
71569f5
separate actor and rollout class
imh966 Jul 3, 2025
78ef6f2
update name of recipe and add license
imh966 Jul 3, 2025
e274747
Merge branch 'volcengine:main' into recipe/async_training
ArronHZG Jul 7, 2025
b9a9618
one_step_off_policy megatron
ArronHZG Jul 7, 2025
8dc0034
use fsdp2 and clear useless code
lalala-2 Jul 7, 2025
5ea1c00
fix config
lalala-2 Jul 7, 2025
69d58c4
fix
lalala-2 Jul 7, 2025
6cdaf2e
one_step_off_policy dapo_7b 2 node
ArronHZG Jul 7, 2025
36ed4f6
recipe/one_step_off_policy
ArronHZG Jul 8, 2025
a1966ef
opt gen_next_batch
lalala-2 Jul 8, 2025
40df88f
Merge branch 'recipe/async_training_megatron' of https://github.com/i…
lalala-2 Jul 8, 2025
5d52efa
4_12_megatron
ArronHZG Jul 8, 2025
59f6be9
4_12_megatron
ArronHZG Jul 8, 2025
dfabe15
megatron config
ArronHZG Jul 8, 2025
40e8816
megatron config
ArronHZG Jul 8, 2025
fc76d4f
fix megatron
lalala-2 Jul 8, 2025
dedc436
Merge branch 'recipe/async_training_megatron' of https://github.com/i…
lalala-2 Jul 8, 2025
344581f
megatron config
ArronHZG Jul 8, 2025
0091f52
megatron config
ArronHZG Jul 9, 2025
283f7fd
megatron config
ArronHZG Jul 9, 2025
6871a29
cross epoch
ArronHZG Jul 9, 2025
1b96322
ruff format
ArronHZG Jul 9, 2025
652f91f
# Copyright 2025 Meituan Ltd. and/or its affiliates
ArronHZG Jul 9, 2025
b36918c
add Copyright
ArronHZG Jul 9, 2025
84b712d
optim sh
ArronHZG Jul 9, 2025
4685463
python3
ArronHZG Jul 9, 2025
7f3d1db
update recipe
ArronHZG Jul 10, 2025
592f393
add doc
ArronHZG Jul 10, 2025
2fb1cd9
Merge branch 'volcengine:main' into recipe/async_training
ArronHZG Jul 10, 2025
dff8f56
update date
ArronHZG Jul 11, 2025
648cb44
update date
ArronHZG Jul 11, 2025
c2395f7
config
ArronHZG Jul 11, 2025
165c1b2
Revert "fix config"
lalala-2 Jul 11, 2025
aaa356e
fix error
lalala-2 Jul 11, 2025
03f1dec
update is_last_step
ArronHZG Jul 11, 2025
e2007ef
one_step_off_policy
ArronHZG Jul 11, 2025
204d624
update readme
ArronHZG Jul 14, 2025
19fac39
e2e_one_step_off_policy
ArronHZG Jul 14, 2025
c1b86ec
add e2e test for one_step_off_policy
ArronHZG Jul 14, 2025
492ff98
add e2e test for one_step_off_policy
ArronHZG Jul 14, 2025
1e7aa47
add e2e test for one_step_off_policy
ArronHZG Jul 14, 2025
8ab0834
add e2e test for one_step_off_policy
ArronHZG Jul 14, 2025
22dc212
format
ArronHZG Jul 14, 2025
dcbfb0c
ruff check
ArronHZG Jul 14, 2025
1e8cee3
add megatron test
ArronHZG Jul 14, 2025
27c9816
Merge pull request #2 from imh966/recipe/async_training_e2e_test
ArronHZG Jul 14, 2025
727320b
Merge branch 'volcengine:main' into recipe/async_training
ArronHZG Jul 14, 2025
8727916
rm spmd
ArronHZG Jul 14, 2025
42ddeed
CI check fix some error
ArronHZG Jul 14, 2025
5ffd8b4
merge main
ArronHZG Jul 15, 2025
1c9b6eb
change author
ArronHZG Jul 15, 2025
8772b14
update e2e_one_step_off_policy CI rule
ArronHZG Jul 15, 2025
c8468e6
update comments
ArronHZG Jul 15, 2025
d8dd8b0
Merge branch 'volcengine:main' into recipe/async_training
ArronHZG Jul 15, 2025
659b108
update ruff
ArronHZG Jul 15, 2025
9b5646a
Fix pre-commit error: sort imports in async_main_ppo.py
openhands-agent Jul 15, 2025
1ed49c7
rollout.nnodes
ArronHZG Jul 16, 2025
754cfae
update code and doc by comments
ArronHZG Jul 16, 2025
8df1c1b
ruff
ArronHZG Jul 16, 2025
1837fc7
update code and doc by comments
ArronHZG Jul 16, 2025
c56467f
update docs
ArronHZG Jul 16, 2025
174d94a
Merge branch 'recipe/async_training' of https://github.com/imh966/ver…
ArronHZG Jul 16, 2025
e3db358
Merge branch 'recipe/async_training' into recipe/async_training_rollo…
ArronHZG Jul 16, 2025
8e5b714
Merge pull request #3 from imh966/recipe/async_training_rollout_nodes
ArronHZG Jul 16, 2025
40b2ebe
Merge branch 'volcengine:main' into recipe/async_training
ArronHZG Jul 16, 2025
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
144 changes: 144 additions & 0 deletions .github/workflows/e2e_one_step_off_policy.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,144 @@
# # Tests layout

# Each folder under tests/ corresponds to a test category for a sub-namespace in verl. For instance:
# - `tests/trainer` for testing functionality related to `verl/trainer`
# - `tests/models` for testing functionality related to `verl/models`
# - ...

# There are a few folders with `special_` prefix, created for special purposes:
# - `special_distributed`: unit tests that must run with multiple GPUs
# - `special_e2e`: end-to-end tests with training/generation scripts
# - `special_npu`: tests for NPUs
# - `special_sanity`: a suite of quick sanity tests
# - `special_standalone`: a set of test that are designed to run in dedicated environments

# Accelerators for tests
# - By default tests are run with GPU available, except for the ones under `special_npu`, and any test script whose name ends with `on_cpu.py`.
# - For test scripts with `on_cpu.py` name suffix would be tested on CPU resources in linux environment.

# # Workflow layout

# All CI tests are configured by yaml files in `.github/workflows/`. Here's an overview of all test configs:
# 1. A list of always triggered CPU sanity tests: `check-pr-title.yml`, `secrets_scan.yml`, `check-pr-title,yml`, `pre-commit.yml`, `doc.yml`
# 2. Some heavy multi-GPU unit tests, such as `model.yml`, `vllm.yml`, `sgl.yml`
# 3. End-to-end tests: `e2e_*.yml`
# 4. Unit tests
# - `cpu_unit_tests.yml`, run pytest on all scripts with file name pattern `tests/**/test_*_on_cpu.py`
# - `gpu_unit_tests.yml`, run pytest on all scripts with file without the `on_cpu.py` suffix.
# - Since cpu/gpu unit tests by default runs all tests under `tests`, please make sure tests are manually excluded in them when
# - new workflow yaml is added to `.github/workflows`
# - new tests are added to workflow mentioned in 2.


name: e2e_one_step_off_policy

on:
# Trigger the workflow on push or pull request,
# but only for the main branch
# For push, for now only anti-patterns are specified so it is more conservative
# and achieves higher coverage.
push:
branches:
- main
- v0.*
paths:
- "**/*.py"
- "!**/*.md"
- "!**/*.sh"
# Other entrypoints
- "!examples/*trainer*"
- "!tests/**"
- "!verl/trainer/main_*.py"
- "!verl/trainer/fsdp_sft_trainer.py"
- "!recipe/**"
- "recipe/one_step_off_policy"
pull_request:
branches:
- main
- v0.*
paths:
- "**/*.py"
- "!**/*.md"
- "!**/*.sh"
# Other entrypoints
- "!examples/**"
- "!tests/**"
- "!verl/trainer/main_*.py"
- "!verl/trainer/fsdp_sft_trainer.py"
# Other recipes
- "!recipe/**"
# Home
- "recipe/one_step_off_policy"
# Entrypoints
- ".github/workflows/e2e_one_step_off_policy.yml"
- "examples/data_preprocess/gsm8k.py"
- "tests/special_e2e/run_one_step_off_policy.sh"

# Cancel jobs on the same ref if a new one is triggered
concurrency:
group: ${{ github.workflow }}-${{ github.ref }}
cancel-in-progress: ${{ github.ref != 'refs/heads/main' }}

# Declare permissions just read content.
permissions:
contents: read

jobs:
# Test FSDP2 strategy
e2e_one_step_off_policy_fsdp2:
runs-on: [L20x8]
timeout-minutes: 50 # Increase timeout for async training
env:
HTTP_PROXY: ${{ secrets.PROXY_HTTP }}
HTTPS_PROXY: ${{ secrets.PROXY_HTTPS }}
NO_PROXY: "localhost,127.0.0.1,hf-mirror.com"
HF_ENDPOINT: "https://hf-mirror.com"
HF_HUB_ENABLE_HF_TRANSFER: "0" # This is more stable
ACTOR_STRATEGY: "fsdp2"
container:
image: verlai/verl:app-verl0.4-vllm0.8.5-mcore0.12.1
options: --gpus all --shm-size=10g
steps:
- uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2
with:
fetch-depth: 0
- name: Install the current repository
run: |
pip3 install --no-deps -e .[test,gpu]
- name: Prepare GSM8K dataset
run: |
python3 examples/data_preprocess/gsm8k.py
- name: Running the E2E test with one_step_off_policy algorithm (FSDP2)
run: |
ray stop --force
bash tests/special_e2e/run_one_step_off_policy.sh

# Test Megatron strategy
e2e_one_step_off_policy_megatron:
runs-on: [L20x8]
timeout-minutes: 50 # Increase timeout for async training
env:
HTTP_PROXY: ${{ secrets.PROXY_HTTP }}
HTTPS_PROXY: ${{ secrets.PROXY_HTTPS }}
NO_PROXY: "localhost,127.0.0.1,hf-mirror.com"
HF_ENDPOINT: "https://hf-mirror.com"
HF_HUB_ENABLE_HF_TRANSFER: "0" # This is more stable
ACTOR_STRATEGY: "megatron"
container:
image: verlai/verl:app-verl0.4-vllm0.8.5-mcore0.12.1
options: --gpus all --shm-size=10g
steps:
- uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2
with:
fetch-depth: 0
- name: Install the current repository
run: |
pip3 install --no-deps -e .[test,gpu]
- name: Prepare GSM8K dataset
run: |
python3 examples/data_preprocess/gsm8k.py
- name: Running the E2E test with one_step_off_policy algorithm (Megatron)
run: |
ray stop --force
bash tests/special_e2e/run_one_step_off_policy.sh

Loading