From bd46e69693be0c6ba334494dbd0e0b6bd34121df Mon Sep 17 00:00:00 2001 From: lcawl Date: Tue, 19 Aug 2025 11:55:00 -0700 Subject: [PATCH 1/2] Fix typos in Llama request examples --- .../put_llama/examples/request/PutLlamaRequestExample1.yaml | 2 +- .../put_llama/examples/request/PutLlamaRequestExample2.yaml | 2 +- .../put_llama/examples/request/PutLlamaRequestExample3.yaml | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/specification/inference/put_llama/examples/request/PutLlamaRequestExample1.yaml b/specification/inference/put_llama/examples/request/PutLlamaRequestExample1.yaml index 709663248e..b6c592f3e4 100644 --- a/specification/inference/put_llama/examples/request/PutLlamaRequestExample1.yaml +++ b/specification/inference/put_llama/examples/request/PutLlamaRequestExample1.yaml @@ -6,7 +6,7 @@ value: |- { "service": "llama", "service_settings": { - "url": "http://localhost:8321/v1/inference/embeddings" + "url": "http://localhost:8321/v1/inference/embeddings", "dimensions": 384, "model_id": "all-MiniLM-L6-v2" } diff --git a/specification/inference/put_llama/examples/request/PutLlamaRequestExample2.yaml b/specification/inference/put_llama/examples/request/PutLlamaRequestExample2.yaml index 6ec845bf59..8a323a7b58 100644 --- a/specification/inference/put_llama/examples/request/PutLlamaRequestExample2.yaml +++ b/specification/inference/put_llama/examples/request/PutLlamaRequestExample2.yaml @@ -6,7 +6,7 @@ value: |- { "service": "llama", "service_settings": { - "url": "http://localhost:8321/v1/openai/v1/chat/completions" + "url": "http://localhost:8321/v1/openai/v1/chat/completions", "model_id": "llama3.2:3b" } } diff --git a/specification/inference/put_llama/examples/request/PutLlamaRequestExample3.yaml b/specification/inference/put_llama/examples/request/PutLlamaRequestExample3.yaml index 40124abab7..ab6a6f9606 100644 --- a/specification/inference/put_llama/examples/request/PutLlamaRequestExample3.yaml +++ b/specification/inference/put_llama/examples/request/PutLlamaRequestExample3.yaml @@ -6,7 +6,7 @@ value: |- { "service": "llama", "service_settings": { - "url": "http://localhost:8321/v1/openai/v1/chat/completions" + "url": "http://localhost:8321/v1/openai/v1/chat/completions", "model_id": "llama3.2:3b" } } From e0c6083e1f1aa621938518901e004acc1516ff92 Mon Sep 17 00:00:00 2001 From: lcawl Date: Tue, 19 Aug 2025 11:59:36 -0700 Subject: [PATCH 2/2] Generate output --- output/openapi/elasticsearch-openapi.json | 6 +++--- output/openapi/elasticsearch-serverless-openapi.json | 6 +++--- output/schema/schema.json | 6 +++--- .../put_llama/examples/request/PutLlamaRequestExample1.yaml | 4 +++- .../put_llama/examples/request/PutLlamaRequestExample3.yaml | 4 +++- 5 files changed, 15 insertions(+), 11 deletions(-) diff --git a/output/openapi/elasticsearch-openapi.json b/output/openapi/elasticsearch-openapi.json index 4e19513388..1f26606b87 100644 --- a/output/openapi/elasticsearch-openapi.json +++ b/output/openapi/elasticsearch-openapi.json @@ -22546,15 +22546,15 @@ "examples": { "PutLlamaRequestExample1": { "description": "Run `PUT _inference/text_embedding/llama-text-embedding` to create a Llama inference endpoint that performs a `text_embedding` task.", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/inference/embeddings\"\n \"dimensions\": 384,\n \"model_id\": \"all-MiniLM-L6-v2\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/inference/embeddings\",\n \"dimensions\": 384,\n \"model_id\": \"all-MiniLM-L6-v2\" \n }\n}" }, "PutLlamaRequestExample2": { "description": "Run `PUT _inference/completion/llama-completion` to create a Llama inference endpoint that performs a `completion` task.", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\"\n \"model_id\": \"llama3.2:3b\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\",\n \"model_id\": \"llama3.2:3b\" \n }\n}" }, "PutLlamaRequestExample3": { "description": "Run `PUT _inference/chat-completion/llama-chat-completion` to create a Llama inference endpoint that performs a `chat_completion` task.", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\"\n \"model_id\": \"llama3.2:3b\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\",\n \"model_id\": \"llama3.2:3b\" \n }\n}" } } } diff --git a/output/openapi/elasticsearch-serverless-openapi.json b/output/openapi/elasticsearch-serverless-openapi.json index 2c63931259..82911e2ebe 100644 --- a/output/openapi/elasticsearch-serverless-openapi.json +++ b/output/openapi/elasticsearch-serverless-openapi.json @@ -13341,15 +13341,15 @@ "examples": { "PutLlamaRequestExample1": { "description": "Run `PUT _inference/text_embedding/llama-text-embedding` to create a Llama inference endpoint that performs a `text_embedding` task.", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/inference/embeddings\"\n \"dimensions\": 384,\n \"model_id\": \"all-MiniLM-L6-v2\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/inference/embeddings\",\n \"dimensions\": 384,\n \"model_id\": \"all-MiniLM-L6-v2\" \n }\n}" }, "PutLlamaRequestExample2": { "description": "Run `PUT _inference/completion/llama-completion` to create a Llama inference endpoint that performs a `completion` task.", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\"\n \"model_id\": \"llama3.2:3b\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\",\n \"model_id\": \"llama3.2:3b\" \n }\n}" }, "PutLlamaRequestExample3": { "description": "Run `PUT _inference/chat-completion/llama-chat-completion` to create a Llama inference endpoint that performs a `chat_completion` task.", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\"\n \"model_id\": \"llama3.2:3b\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\",\n \"model_id\": \"llama3.2:3b\" \n }\n}" } } } diff --git a/output/schema/schema.json b/output/schema/schema.json index 8da975684a..a28d21882c 100644 --- a/output/schema/schema.json +++ b/output/schema/schema.json @@ -178062,17 +178062,17 @@ "PutLlamaRequestExample1": { "description": "Run `PUT _inference/text_embedding/llama-text-embedding` to create a Llama inference endpoint that performs a `text_embedding` task.", "method_request": "PUT _inference/text_embedding/llama-text-embedding", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/inference/embeddings\"\n \"dimensions\": 384,\n \"model_id\": \"all-MiniLM-L6-v2\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/inference/embeddings\",\n \"dimensions\": 384,\n \"model_id\": \"all-MiniLM-L6-v2\" \n }\n}" }, "PutLlamaRequestExample2": { "description": "Run `PUT _inference/completion/llama-completion` to create a Llama inference endpoint that performs a `completion` task.", "method_request": "PUT _inference/completion/llama-completion", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\"\n \"model_id\": \"llama3.2:3b\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\",\n \"model_id\": \"llama3.2:3b\" \n }\n}" }, "PutLlamaRequestExample3": { "description": "Run `PUT _inference/chat-completion/llama-chat-completion` to create a Llama inference endpoint that performs a `chat_completion` task.", "method_request": "PUT _inference/chat-completion/llama-chat-completion", - "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\"\n \"model_id\": \"llama3.2:3b\" \n }\n}" + "value": "{\n \"service\": \"llama\",\n \"service_settings\": {\n \"url\": \"http://localhost:8321/v1/openai/v1/chat/completions\",\n \"model_id\": \"llama3.2:3b\" \n }\n}" } }, "inherits": { diff --git a/specification/inference/put_llama/examples/request/PutLlamaRequestExample1.yaml b/specification/inference/put_llama/examples/request/PutLlamaRequestExample1.yaml index b6c592f3e4..7f49bd5343 100644 --- a/specification/inference/put_llama/examples/request/PutLlamaRequestExample1.yaml +++ b/specification/inference/put_llama/examples/request/PutLlamaRequestExample1.yaml @@ -1,5 +1,7 @@ # summary: -description: Run `PUT _inference/text_embedding/llama-text-embedding` to create a Llama inference endpoint that performs a `text_embedding` task. +description: + Run `PUT _inference/text_embedding/llama-text-embedding` to create a Llama inference endpoint that performs a + `text_embedding` task. method_request: 'PUT _inference/text_embedding/llama-text-embedding' # type: "request" value: |- diff --git a/specification/inference/put_llama/examples/request/PutLlamaRequestExample3.yaml b/specification/inference/put_llama/examples/request/PutLlamaRequestExample3.yaml index ab6a6f9606..cf73fb2b61 100644 --- a/specification/inference/put_llama/examples/request/PutLlamaRequestExample3.yaml +++ b/specification/inference/put_llama/examples/request/PutLlamaRequestExample3.yaml @@ -1,5 +1,7 @@ # summary: -description: Run `PUT _inference/chat-completion/llama-chat-completion` to create a Llama inference endpoint that performs a `chat_completion` task. +description: + Run `PUT _inference/chat-completion/llama-chat-completion` to create a Llama inference endpoint that performs a + `chat_completion` task. method_request: 'PUT _inference/chat-completion/llama-chat-completion' # type: "request" value: |-