Skip to content

Conversation

@openingnow
Copy link
Contributor

This PR introduces minimally working openAI-compatible /v1/responses API by converting /v1/responses request into /v1/chat/completions request.

Only text generation is supported and several fields such as IDs (of response and messages) are omitted.

If this appears too unfinished for a merge at this stage, please let me know and I'll convert it to a draft.

Copy link
Collaborator

@ngxson ngxson left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

please also add a test with openai python library to prove that this is openai-compat.

see tools/server/tests/unit/test_chat_completion.py

@wqerrewetw
Copy link

#14702

@pwilkin
Copy link
Collaborator

pwilkin commented Dec 20, 2025

@openingnow

If this appears too unfinished for a merge at this stage, please let me know and I'll convert it to a draft.

Actually, I think this is a very good approach - let's start with a minimal implementation that will permit Codex to work properly with Llama.cpp and we can add more complexities later.

@openingnow openingnow requested a review from CISC as a code owner December 21, 2025 04:22
@github-actions github-actions bot added the python python script changes label Dec 21, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

examples python python script changes server

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants