-
Notifications
You must be signed in to change notification settings - Fork 250
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feature: Support new OpenAI o1 reasoning models #575
Comments
I assume they'll open it up in the near future but currently to use the reasoning models via the API you need to hit tier 5 https://platform.openai.com/docs/guides/rate-limits/usage-tiers?context=tier-five meaning you have to have bought $1000 in tokens |
@Alextibtab if you need to test anything I have this level of API access. |
lmao if you have tier 5 then feel free to use it. For now we have to wait till GA for API usage. btw the API costs would be pretty high for o1 from my testing. |
@abatilo I have access and just tried to implement your diff-- experiencing the same freeze at |
I mean they haven't even publish o1 API yet, so there is nothing we can do for sure. I suspect it is still streaming, just that we need to figure out how to display CoT reasoning |
@aarnphm There might be a misunderstanding. There is an API but it's closed to certain tiers. It's mostly the same as the current chat completion endpoint but doesn't support streaming. According to OAI, they don't plan on returning the full CoT tokens. Unless maybe they've changed their mind. I think we would mostly just need to add code paths for handling non-streaming results to make this work |
Yes the API is open for tier 5 and up. What I'm referring to is on their API reference they have yet to update the example for o1, or specify specific data_type for SSE during CoT. I don't think they will ever publish CoT tokens (that is their moat apparently). But this is irrelevant in this case. The chat after CoT are still streaming afaict |
I've added a pull request that adds O1 support, without interfering with other models, and also solving the response hanging issue people had experienced in this thread. Have an awesome day guys <3 |
Well done @LessComplexity ! Fantastic work |
@aarnphm |
thanks |
Feature request
I would like if Avante could prompt against the OAI reasoning models
Motivation
These reasoning models are allegedly more capable at coding.
Other
There are a few differences in the API now:
temperature
must be set to 1 if you're using eithero1-preview
oro1-mini
models.max_tokens
is not used. Insteadmax_completion_tokens
must be used in the configsystem
role/message. I think we need to remove this lineEven with these changes, I'm not getting a successful end to end flow. I've never contributed to the avante.nvim codebase and I'm not entirely sure what else to try at the moment to get things working.
So far, my total diffs look like so:
I've gotten this far by trying to use the openai provider and seeing it fail and return an error message. This time, it's not returning anything. I see
Generating response ...
and it never changes.The text was updated successfully, but these errors were encountered: