Skip to content

OpenAI compatible API server for YandexGPT and YandexART API calls

Notifications You must be signed in to change notification settings

sazonovanton/YandexGPT_to_OpenAI

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

35 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

YandexGPT API to OpenAI API translator

Server that translates OpenAI API requests to YandexGPT/YandexART API requests.
It is useful for those who have a tool that expects OpenAI (like) API usage, but want to use YandexGPT API.

How it works

The API server listens for OpenAI API requests and sends them to Yandex GPT and returns translated response to the client.
Supports text generation, text embeddings and image generation (learn more).

How to use

At first you will need API key and catalog ID from Yandex Cloud. Roles needed: ai.languageModels.user and ai.imageGeneration.user.
Then clone this repository:

git clone https://github.com/sazonovanton/YandexGPT_to_OpenAI
cd YandexGPT_to_OpenAI

After that follow the instructions below and then use the API by setting openai.base_url to http://<your_host>:<your_port>/v1 in your program.

Authentication

Authentication is done by providing a token in the header. You can use generated tokens from the utils/tokens.py script or use your own Yandex cloud API keys as token for the API.

Generate tokens

Create tokens by running the following command from project root directory:

python utils/tokens.py

Tokens are stored in the data/tokens.json file.

User own keys

You can set API so that users can use their own Yandex cloud API keys as token (as <CatalogID>:<SecretKey>) for the API by setting Y2O_BringYourOwnKey to True in the environment variables.

Setup with Docker

  1. Install Docker if you haven't already.
curl -fsSL https://get.docker.com -o get-docker.sh
sh get-docker.sh
  1. Make changes in the docker-compose.yml file setting environment variables.
  2. Run the API
docker-compose up -d

Setup without Docker

  1. Install the requirements
pip install -r requirements.txt
  1. Create environment variables for the OpenAI API key and the YandexGPT API key. You can do this by creating a .env file in the root of the project with the following content:
Y2O_SecretKey=None
Y2O_CatalogID=None
Y2O_BringYourOwnKey=False
Y2O_Host=127.0.0.1
Y2O_Port=8520
Y2O_ServerURL=http://127.0.0.1:8520
Y2O_LogFile=logs/y2o.log
Y2O_LogLevel=INFO

Here are default values.
If you set Y2O_BringYourOwnKey to True, then users are able to use their own Yandex cloud API keys as token for the API (token in that case <CatalogID>:<SecretKey>).
If you set Y2O_SecretKey and Y2O_CatalogID, then users will be able to use the API without providing their own keys but using given token instead.
Y2O_ServerURL needed to send an URL of the generated image in the response, default is http://localhost:8520.
3. Run the API

python app.py

SSL

If you want to use SSL, you can set paths to the SSL key and certificate files via environment variables in the docker-compose.yml file or in the .env file.

Y2O_SSL_Key=ssl/private.key
Y2O_SSL_Cert=ssl/cert.pem

What is implemented

/chat/completions

Chat completions translates to TextGeneration.completion

  • ✅ Streaming
  • ⬜ Vision (not supported)
  • ⬜ Tools (not supported)

/embeddings

Embeddings translates to Embeddings.textEmbedding (not fully tested)

  • ✅ encoding_format: float, base64

/images/generations

Images generations translates to ImageGenerationAsync.generate.
Generates one JPEG image, does not return revised prompt. When URL is requested, it saves generated image to the data/images directory and deletes it after one hour. Directory is cleaned on server start. Recommended to use response_format: b64_json to get image as base64 encoded string as it provided by YandexART and not additionaly processed.

  • ✅ response_format: b64_json, url
  • ❕ size - sets aspect ratio, not width and height (weight of width and height in image)
  • ❌ n - number of images to generate (always 1)

/models

Models translates to list of Yandex Foundation models that is stored in data/model_list.json file (can be incomplete)

Usage

Here is an example of how to use the API.

Chat completions

cURL example

curl http://<your_host>:<your_port>/v1/chat/completions \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer $TOKEN" \
  -d '{
    "model": "yandexgpt/latest",
    "messages": [
      {
        "role": "system",
        "content": "You are a helpful assistant."
      },
      {
        "role": "user",
        "content": "Hello!"
      }
    ]
  }'

Python example

import os
from openai import OpenAI

client = OpenAI(
    api_key=os.getenv("TOKEN"),
    base_url="http://<your_host>:<your_port>/v1",
)
chat_completion = client.chat.completions.create(
    messages=[
        {
            "role": "user",
            "content": "Say this is a test",
        }
    ],
    model="yandexgpt/latest",
)

Image generations

cURL example

Generate an image and get the URL:

curl http://<your_host>:<your_port>/v1/images/generations \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer $TOKEN" \
  -d '{
    "model": "yandex-art/latest",
    "prompt": "A painting of a cat",
    "response_format": "url"
  }'

Retrive the image by the URL in the response:

curl -H "Authorization: Bearer $TOKEN" -H "Content-Type: application/json" -O http://<your_host>:<your_port>/images/<id>.jpg

Python example

Returing image as base64 encoded string:

from openai import OpenAI

client = OpenAI(
    api_key=os.getenv("TOKEN"),
    base_url="http://<your_host>:<your_port>/v1",
)
image_generation = client.images.generate(
    model="yandex-art/latest",
    prompt="A painting of a cat",
    response_format="b64_json"
)