-
Notifications
You must be signed in to change notification settings - Fork 2.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: add support for copilot models #3273
base: main
Are you sure you want to change the base?
Conversation
I think for the claude models, we should keep haiku as the weak models. Having sonnet-3.5 as the weak model for sonnet-3.7 is expensive and overkill. |
@paul-gauthier Since the changes in litellm side were merged at BerriAI/litellm#8577, could you have a look at this to see if I'm missing anythings? |
Please change the weak models to haiku. sonnet-3.5 costs the same as sonnet-3.7 - there's no point in making 3.5 the weaker model. |
Sorry to intrude here, are the sonnet costs relevant when used via GH Copilot? Isn't it a flat fee regardless of the model used? |
Fair point. One could even argue that the weak model could be the same as the strong model in that case - why have them be different? My concern is twofold. The first is simple: Using such a heavy model is overkill for the weak model. The second is that (some) people who have used proxies to reroute Aider to copilot have gotten warning emails from Microsoft. It's not clear what triggers it - perhaps too much |
use_repo_map: true | ||
examples_as_sys_msg: true | ||
extra_params: | ||
max_tokens: 8192 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
why is the number of tokens so small? As far as i see both 3.7 and 3.5 accept 200k context: https://docs.anthropic.com/en/docs/about-claude/models/all-models.
do I misunderstand how max_tokens
works?
solve #2227
Wait for BerriAI/litellm#8577 to be merged