diff --git a/output/openapi/elasticsearch-openapi.json b/output/openapi/elasticsearch-openapi.json index 4e19513388..1f26606b87 100644 --- a/output/openapi/elasticsearch-openapi.json +++ b/output/openapi/elasticsearch-openapi.json @@ -22546,15 +22546,15 @@ "examples": { "PutLlamaRequestExample1": { "description": "Run `PUT _inference/text_embedding/llama-text-embedding` to create a Llama inference endpoint that performs a `text_embedding` task.", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/inference/embeddings\"\n \"dimensions\": 384,\n \"model_id\": \"all-MiniLM-L6-v2\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/inference/embeddings\",\n \"dimensions\": 384,\n \"model_id\": \"all-MiniLM-L6-v2\" \n }\n}" }, "PutLlamaRequestExample2": { "description": "Run `PUT _inference/completion/llama-completion` to create a Llama inference endpoint that performs a `completion` task.", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\"\n \"model_id\": \"llama3.2:3b\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\",\n \"model_id\": \"llama3.2:3b\" \n }\n}" }, "PutLlamaRequestExample3": { "description": "Run `PUT _inference/chat-completion/llama-chat-completion` to create a Llama inference endpoint that performs a `chat_completion` task.", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\"\n \"model_id\": \"llama3.2:3b\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\",\n \"model_id\": \"llama3.2:3b\" \n }\n}" } } } diff --git a/output/openapi/elasticsearch-serverless-openapi.json b/output/openapi/elasticsearch-serverless-openapi.json index 2c63931259..82911e2ebe 100644 --- a/output/openapi/elasticsearch-serverless-openapi.json +++ b/output/openapi/elasticsearch-serverless-openapi.json @@ -13341,15 +13341,15 @@ "examples": { "PutLlamaRequestExample1": { "description": "Run `PUT _inference/text_embedding/llama-text-embedding` to create a Llama inference endpoint that performs a `text_embedding` task.", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/inference/embeddings\"\n \"dimensions\": 384,\n \"model_id\": \"all-MiniLM-L6-v2\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/inference/embeddings\",\n \"dimensions\": 384,\n \"model_id\": \"all-MiniLM-L6-v2\" \n }\n}" }, "PutLlamaRequestExample2": { "description": "Run `PUT _inference/completion/llama-completion` to create a Llama inference endpoint that performs a `completion` task.", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\"\n \"model_id\": \"llama3.2:3b\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\",\n \"model_id\": \"llama3.2:3b\" \n }\n}" }, "PutLlamaRequestExample3": { "description": "Run `PUT _inference/chat-completion/llama-chat-completion` to create a Llama inference endpoint that performs a `chat_completion` task.", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\"\n \"model_id\": \"llama3.2:3b\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\",\n \"model_id\": \"llama3.2:3b\" \n }\n}" } } } diff --git a/output/schema/schema.json b/output/schema/schema.json index 8da975684a..a28d21882c 100644 --- a/output/schema/schema.json +++ b/output/schema/schema.json @@ -178062,17 +178062,17 @@ "PutLlamaRequestExample1": { "description": "Run `PUT _inference/text_embedding/llama-text-embedding` to create a Llama inference endpoint that performs a `text_embedding` task.", "method_request": "PUT _inference/text_embedding/llama-text-embedding", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/inference/embeddings\"\n \"dimensions\": 384,\n \"model_id\": \"all-MiniLM-L6-v2\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/inference/embeddings\",\n \"dimensions\": 384,\n \"model_id\": \"all-MiniLM-L6-v2\" \n }\n}" }, "PutLlamaRequestExample2": { "description": "Run `PUT _inference/completion/llama-completion` to create a Llama inference endpoint that performs a `completion` task.", "method_request": "PUT _inference/completion/llama-completion", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\"\n \"model_id\": \"llama3.2:3b\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\",\n \"model_id\": \"llama3.2:3b\" \n }\n}" }, "PutLlamaRequestExample3": { "description": "Run `PUT _inference/chat-completion/llama-chat-completion` to create a Llama inference endpoint that performs a `chat_completion` task.", "method_request": "PUT _inference/chat-completion/llama-chat-completion", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\"\n \"model_id\": \"llama3.2:3b\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\",\n \"model_id\": \"llama3.2:3b\" \n }\n}" } }, "inherits": { diff --git a/specification/inference/put_llama/examples/request/PutLlamaRequestExample1.yaml b/specification/inference/put_llama/examples/request/PutLlamaRequestExample1.yaml index 709663248e..7f49bd5343 100644 --- a/specification/inference/put_llama/examples/request/PutLlamaRequestExample1.yaml +++ b/specification/inference/put_llama/examples/request/PutLlamaRequestExample1.yaml @@ -1,12 +1,14 @@ # summary: -description: Run `PUT _inference/text_embedding/llama-text-embedding` to create a Llama inference endpoint that performs a `text_embedding` task. +description: + Run `PUT _inference/text_embedding/llama-text-embedding` to create a Llama inference endpoint that performs a + `text_embedding` task. method_request: 'PUT _inference/text_embedding/llama-text-embedding' # type: "request" value: |- { "service": "llama", "service_settings": { - "url": "http://localhost:8321/v1/inference/embeddings" + "url": "http://localhost:8321/v1/inference/embeddings", "dimensions": 384, "model_id": "all-MiniLM-L6-v2" } diff --git a/specification/inference/put_llama/examples/request/PutLlamaRequestExample2.yaml b/specification/inference/put_llama/examples/request/PutLlamaRequestExample2.yaml index 6ec845bf59..8a323a7b58 100644 --- a/specification/inference/put_llama/examples/request/PutLlamaRequestExample2.yaml +++ b/specification/inference/put_llama/examples/request/PutLlamaRequestExample2.yaml @@ -6,7 +6,7 @@ value: |- { "service": "llama", "service_settings": { - "url": "http://localhost:8321/v1/openai/v1/chat/completions" + "url": "http://localhost:8321/v1/openai/v1/chat/completions", "model_id": "llama3.2:3b" } } diff --git a/specification/inference/put_llama/examples/request/PutLlamaRequestExample3.yaml b/specification/inference/put_llama/examples/request/PutLlamaRequestExample3.yaml index 40124abab7..cf73fb2b61 100644 --- a/specification/inference/put_llama/examples/request/PutLlamaRequestExample3.yaml +++ b/specification/inference/put_llama/examples/request/PutLlamaRequestExample3.yaml @@ -1,12 +1,14 @@ # summary: -description: Run `PUT _inference/chat-completion/llama-chat-completion` to create a Llama inference endpoint that performs a `chat_completion` task. +description: + Run `PUT _inference/chat-completion/llama-chat-completion` to create a Llama inference endpoint that performs a + `chat_completion` task. method_request: 'PUT _inference/chat-completion/llama-chat-completion' # type: "request" value: |- { "service": "llama", "service_settings": { - "url": "http://localhost:8321/v1/openai/v1/chat/completions" + "url": "http://localhost:8321/v1/openai/v1/chat/completions", "model_id": "llama3.2:3b" } }