Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG] Azure GPT 4o is limited to maximum 4096 output tokens even though model supports 16,384 #677

Open
1 task
tsoernes opened this issue Nov 12, 2024 · 1 comment
Labels
type: bug Something isn't working

Comments

@tsoernes
Copy link

Description

Cannot set it higher than 4096 even though model supports it.

Device and browser

Linux

Screenshots and more

No response

Willingness to Contribute

  • πŸ™‹β€β™‚οΈ Yes, I would like to contribute a fix.
@tsoernes tsoernes added the type: bug Something isn't working label Nov 12, 2024
@enricoros
Copy link
Owner

@tsoernes confirmed. Since the APIs don't tell us the token window and completion sizes, we have to have mappings in the code inside Big-AGI, and the model ID for azure is probably not available either (you have probably the deployment name).

Probably in the future we'd need to make the value fully user editable, or finally OpenAI/Azure will have descriptive APIs where they return the token sizes.

To fix this now, you need to go in the code, models.ts, and look for OpenAI / Azure.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
type: bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants