-
Notifications
You must be signed in to change notification settings - Fork 473
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Peteish13 #739
base: main
Are you sure you want to change the base?
Peteish13 #739
Conversation
This reverts commit 6cc6f62.
…to peteish13-augusta
…to peteish13-augusta
@@ -1036,6 +1036,10 @@ def eval(self) -> Dict[str, Any]: | |||
|
|||
del eval_batches | |||
|
|||
# Eval compiles a bunch more versions, and the result is terrible. This way we get back to zero. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
What do you that the result is terrible?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
so this prompted me to look into this a bit more and I think I've found a better solution: just mark the model input sizes as dynamic. I tested this out in OLMo-core and it appears to work well.
allenai/OLMo-core#105
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think it compiles a bunch of versions for different batch sizes, because that's how we call it during eval, and then they stick around. In all of my early runs I had high tps until the first eval, and then low tps afterwards. This is what fixed it.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I tried dynamic
and it was bad. I don't remember the way in which it was bad, but it didn't work. That's why I added that version in the first place.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ok, oh well. I tested with nightly so maybe it's just better now with recent compiler advances.
torch.compile()
compile()
to one block at a time.