-
Notifications
You must be signed in to change notification settings - Fork 445
Removes the generate thread #1054
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good to me generally, but I worry about error handling now.
total_processed = 0 | ||
iteration_count = 0 | ||
|
||
while not self._should_exit(): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
similar question to above, what happens if some other process dies and the vllm worker should stop?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Then it gets killed when Ray shuts down:
open-instruct/open_instruct/grpo_fast.py
Line 2708 in c3f79a3
ray.shutdown() |
output.request_id, output, complete_output, current_time | ||
) | ||
|
||
if self.verbose and iteration_count % 100 == 0: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
i never really used this logging so fine with removing but was there any other impetus for doing so?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Truthfully, I asked Claude to remove all the debug logging (I had a bunch of debug logging in an earlier version of this PR) and it also removed this. I thought it was fine so I kept it. Open to changing this if you prefer.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
no, all good! Was just wondering why the change.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Tested and seems good to me!
The only reason we had the generate thread previously was to fix a locking issue that we had when updating the weights on the actor. Turns out that, with vLLM v1, the locking issue goes away!
Step time decreases significantly on the runs with
inflight_updates=True
(wandb):Runs with
inflight_updates=True
:Runs with
inflight_updates=False
: