8000 docs: update readme.md features by snowmanwwg · Pull Request #348 · NVIDIA-NeMo/RL · GitHub
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

docs: update readme.md features #348

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
May 9, 2025
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 3 additions & 4 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -43,15 +43,14 @@ What you can expect:
- ✅ **Learning Algorithms** - GRPO (Group Relative Policy Optimization), SFT (Supervised Fine-Tuning), and DPO (Direct Preference Optimization).
- ✅ **Multi-Turn RL** - multi-turn generation and training for RL with tool use, games, etc.
- ✅ **Large Model Support** - Native PyTorch support for models up to 32B parameters.
- ✅ **Advanced Parallelism** - FSDP2, TP, and SP for efficient training.
- ✅ **Advanced Parallelism** - PyTorch native FSDP2, TP, and SP for efficient training.
- ✅ **Worker Isolation** - Process isolation between RL Actors (no worries about global state).
- ✅ **Environment Isolation** - Dependency isolation between components.

- 🔜 **(Even) Larger Model Support** - Native PyTorch & Megatron.
- 🔜 **Improved Native Performance** - Improve training time for Native Pytorch Models.
- 🔜 **Megatron Policy** - Support advanced parallelism in training with Megatron Core.
- 🔜 **Megatron Inference** - Support Megatron Inference for day-0 support for new megatron models.
- 🔜 **(even) Larger Model Support with Long(er) Sequence** - Support advanced parallelism in training with Megatron Core.
- 🔜 **MoE Models** - Support DeepseekV3 and Llama4.
- 🔜 **Megatron Inference** - Support Megatron Inference for day-0 support for new megatron models.

## Prerequisites

Expand Down
Loading
0