-
Notifications
You must be signed in to change notification settings - Fork 2.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error with latest setup.py trl.extras.vllm_client - Server is not up yet #543
Comments
What code do you use for training? Dyou set up a vllm server with |
Thanks for your response! I am using grpo training cmd for DeepSeek R1 and Simple RL:
do I need to set up a vllm server first with |
Yes, check the latest doc: https://huggingface.co/docs/trl/en/grpo_trainer#speed-up-training-with-vllm-powered-generation. |
Thanks! Now following solution for training example in cmd works for me: use GPU 0 for vllm-server, others for training if you have 8 GPUs total then in another session:
|
This comment has been minimized.
This comment has been minimized.
I got the same problem, I solved it by "unset http_proxy" and "unset https_proxy" |
Hi team,
The latest main commit with updated trl has error, I cannot reproduce with 8 GPUs grpo.py
my previous build based on Commit 8782fa6
is working
error log:
Can anyone help me with that?
The text was updated successfully, but these errors were encountered: