-
Notifications
You must be signed in to change notification settings - Fork 160
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Improved stability of litellm models for reasoning models. #538
base: main
Are you sure you want to change the base?
Conversation
Hey @JoelNiklaus, since you have made most commits around LiteLLM.
|
Hi @satpalsr,
|
@JoelNiklaus Thanks. I'll just drop it as separate issue for anyone to pick. |
kwargs["caching"] = False | ||
logger.info("Response is empty, retrying without caching") | ||
response = litellm.completion(**kwargs) | ||
|
||
if content and "<think>" in content: | ||
logger.debug(f"Removing <think> tags from response: {content}") |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Why are we removing think tags from the answer here ? I think it should be done in the metric function no ?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
If we are evaluating a reasoning model the grader will look at the thinking tokens unless we remove them. We would need to remove them in every metric function otherwise.
Co-authored-by: Nathan Habib <[email protected]>
No description provided.