-
Notifications
You must be signed in to change notification settings - Fork 963
verl v0.2.1 & v0.3 release checklist #354
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
What can I help about the 'tool calling examples' part? |
Will megatron context parallelism be supported in the future? |
Yes. We will use mcore that supports cp by default. |
@BearBiscuit05 See #344, I outlined the main challenge. I think it should be relatively straightforward if veRL can start using I imagine we can have GRPO-trained reasoners in the future that learns when to use tools as part of their |
I talked to vllm maintainer yesterday. It seems that there should be no blocking if we switch from generate to chat. Do you mind give it a try to call chat using SPMD style offline inference? |
Not very familiar with inference, but I think I’m starting to get the hang of it. Does this mean I need to build a new |
You should be able to replace Case 1: Detokenize the tokenized
8000
inputs we use for
|
The second choice would incur significant overhead when tokenizing on-the-fly (typically 2x slowdown in generation, which is basically unacceptable). I guess we will need to seek solution for case 1 |
Got it. I'll give it a try. |
It seems that the context parallelism in the model part has not been implemented yet. Is this function currently available? |
Not right now, but if you check this roadmap, once verl upgrades MCore, cp will be support. |
Is it possible to optimize startup time? I noticed when using veRL, it is significantly slower to launch a job than when using Huggingface TRL |
Disabling torch.compile is useful, as it can also hang PPO training when enabling use_remove_padding. #387 |
@maksimstw thanks for the feedback! Would you like to provide a PR with this option? |
when will you release the "sglang integration" part? |
how to install v0.3 |
## Summary Providing an option in the config to turn off the `torch.compile` used in `dp_actor.py` ## Usage Adding the following line to the driver or cli scripts to turn off `torch.compile`. ```python +actor_rollout_ref.actor.use_torch_compile=False ``` Otherwise, `torch.compile` will be used by default ## Related Issue #354 #245 --------- Signed-off-by: Hongpeng Guo <hpguo@anyscale.com>
Item solved in #554 |
hi @JarvisFei , v0.3 is not fully released but you are welcome to try verl main branch with |
As we are already making quite some progress in the main branch, I suggest we freeze code week for v0.3 and push the rest of the features to v0.4 |
Moving discussions to #710 |
…lcengine#554) ## Summary Providing an option in the config to turn off the `torch.compile` used in `dp_actor.py` ## Usage Adding the following line to the driver or cli scripts to turn off `torch.compile`. ```python +actor_rollout_ref.actor.use_torch_compile=False ``` Otherwise, `torch.compile` will be used by default ## Related Issue volcengine#354 volcengine#245 --------- Signed-off-by: Hongpeng Guo <hpguo@anyscale.com>
…lcengine#554) ## Summary Providing an option in the config to turn off the `torch.compile` used in `dp_actor.py` ## Usage Adding the following line to the driver or cli scripts to turn off `torch.compile`. ```python +actor_rollout_ref.actor.use_torch_compile=False ``` Otherwise, `torch.compile` will be used by default ## Related Issue volcengine#354 volcengine#245 --------- Signed-off-by: Hongpeng Guo <hpguo@anyscale.com>
…lcengine#554) ## Summary Providing an option in the config to turn off the `torch.compile` used in `dp_actor.py` ## Usage Adding the following line to the driver or cli scripts to turn off `torch.compile`. ```python +actor_rollout_ref.actor.use_torch_compile=False ``` Otherwise, `torch.compile` will be used by default ## Related Issue volcengine#354 volcengine#245 --------- Signed-off-by: Hongpeng Guo <hpguo@anyscale.com>
v0.2.1
log_prob_micro_batch_size
is smaller than world_size, and fix the case when "the evaluation dataset size is not divisible by the world_size" Hangs during vllm rollout, no error message #12 (comment)ray start
Add instructions on how to run verl on multi-node #278doc: add multinode training and debug tutorial #585
v0.3
The text was updated successfully, but these errors were encountered: