diff --git a/output/openapi/elasticsearch-openapi.json b/output/openapi/elasticsearch-openapi.json index 26c52f9e09..69e3596591 100644 --- a/output/openapi/elasticsearch-openapi.json +++ b/output/openapi/elasticsearch-openapi.json @@ -20481,13 +20481,13 @@ "inference" ], "summary": "Create a Mistral inference endpoint", - "description": "Creates an inference endpoint to perform an inference task with the `mistral` service.\n ##Required authorization\n* Cluster privileges: `manage_inference`", + "description": "Create an inference endpoint to perform an inference task with the `mistral` service.\n ##Required authorization\n* Cluster privileges: `manage_inference`", "operationId": "inference-put-mistral", "parameters": [ { "in": "path", "name": "task_type", - "description": "The task type.\nThe only valid task type for the model to perform is `text_embedding`.", + "description": "The type of the inference task that the model will perform.", "required": true, "deprecated": false, "schema": { @@ -84672,7 +84672,9 @@ "inference._types.MistralTaskType": { "type": "string", "enum": [ - "text_embedding" + "text_embedding", + "completion", + "chat_completion" ] }, "inference._types.MistralServiceType": { @@ -84699,7 +84701,7 @@ "externalDocs": { "url": "https://docs.mistral.ai/getting-started/models/" }, - "description": "The name of the model to use for the inference task.\nRefer to the Mistral models documentation for the list of available text embedding models.", + "description": "The name of the model to use for the inference task.\nRefer to the Mistral models documentation for the list of available models.", "type": "string" }, "rate_limit": { @@ -84737,7 +84739,9 @@ "inference._types.TaskTypeMistral": { "type": "string", "enum": [ - "text_embedding" + "text_embedding", + "chat_completion", + "completion" ] }, "inference._types.OpenAITaskType": { diff --git a/output/openapi/elasticsearch-serverless-openapi.json b/output/openapi/elasticsearch-serverless-openapi.json index bc25f952ed..8ece9d0693 100644 --- a/output/openapi/elasticsearch-serverless-openapi.json +++ b/output/openapi/elasticsearch-serverless-openapi.json @@ -11842,13 +11842,13 @@ "inference" ], "summary": "Create a Mistral inference endpoint", - "description": "Creates an inference endpoint to perform an inference task with the `mistral` service.\n ##Required authorization\n* Cluster privileges: `manage_inference`", + "description": "Create an inference endpoint to perform an inference task with the `mistral` service.\n ##Required authorization\n* Cluster privileges: `manage_inference`", "operationId": "inference-put-mistral", "parameters": [ { "in": "path", "name": "task_type", - "description": "The task type.\nThe only valid task type for the model to perform is `text_embedding`.", + "description": "The type of the inference task that the model will perform.", "required": true, "deprecated": false, "schema": { @@ -54246,7 +54246,9 @@ "inference._types.MistralTaskType": { "type": "string", "enum": [ - "text_embedding" + "text_embedding", + "completion", + "chat_completion" ] }, "inference._types.MistralServiceType": { @@ -54273,7 +54275,7 @@ "externalDocs": { "url": "https://docs.mistral.ai/getting-started/models/" }, - "description": "The name of the model to use for the inference task.\nRefer to the Mistral models documentation for the list of available text embedding models.", + "description": "The name of the model to use for the inference task.\nRefer to the Mistral models documentation for the list of available models.", "type": "string" }, "rate_limit": { @@ -54311,7 +54313,9 @@ "inference._types.TaskTypeMistral": { "type": "string", "enum": [ - "text_embedding" + "text_embedding", + "chat_completion", + "completion" ] }, "inference._types.OpenAITaskType": { diff --git a/output/schema/schema.json b/output/schema/schema.json index 743c5af16b..e09a0f7340 100644 --- a/output/schema/schema.json +++ b/output/schema/schema.json @@ -10155,7 +10155,7 @@ "visibility": "public" } }, - "description": "Create a Mistral inference endpoint.\n\nCreates an inference endpoint to perform an inference task with the `mistral` service.", + "description": "Create a Mistral inference endpoint.\n\nCreate an inference endpoint to perform an inference task with the `mistral` service.", "docId": "inference-api-put-mistral", "docUrl": "https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-mistral", "name": "inference.put_mistral", @@ -157510,7 +157510,7 @@ } }, { - "description": "The name of the model to use for the inference task.\nRefer to the Mistral models documentation for the list of available text embedding models.", + "description": "The name of the model to use for the inference task.\nRefer to the Mistral models documentation for the list of available models.", "extDocId": "mistral-api-models", "extDocUrl": "https://docs.mistral.ai/getting-started/models/", "name": "model", @@ -157549,20 +157549,26 @@ "name": "MistralServiceType", "namespace": "inference._types" }, - "specLocation": "inference/_types/CommonTypes.ts#L1017-L1019" + "specLocation": "inference/_types/CommonTypes.ts#L1019-L1021" }, { "kind": "enum", "members": [ { "name": "text_embedding" + }, + { + "name": "completion" + }, + { + "name": "chat_completion" } ], "name": { "name": "MistralTaskType", "namespace": "inference._types" }, - "specLocation": "inference/_types/CommonTypes.ts#L1013-L1015" + "specLocation": "inference/_types/CommonTypes.ts#L1013-L1017" }, { "kind": "interface", @@ -157649,7 +157655,7 @@ } } ], - "specLocation": "inference/_types/CommonTypes.ts#L1021-L1063" + "specLocation": "inference/_types/CommonTypes.ts#L1023-L1065" }, { "kind": "enum", @@ -157662,7 +157668,7 @@ "name": "OpenAIServiceType", "namespace": "inference._types" }, - "specLocation": "inference/_types/CommonTypes.ts#L1079-L1081" + "specLocation": "inference/_types/CommonTypes.ts#L1081-L1083" }, { "kind": "interface", @@ -157684,7 +157690,7 @@ } } ], - "specLocation": "inference/_types/CommonTypes.ts#L1065-L1071" + "specLocation": "inference/_types/CommonTypes.ts#L1067-L1073" }, { "kind": "enum", @@ -157703,7 +157709,7 @@ "name": "OpenAITaskType", "namespace": "inference._types" }, - "specLocation": "inference/_types/CommonTypes.ts#L1073-L1077" + "specLocation": "inference/_types/CommonTypes.ts#L1075-L1079" }, { "kind": "interface", @@ -158231,13 +158237,19 @@ "members": [ { "name": "text_embedding" + }, + { + "name": "chat_completion" + }, + { + "name": "completion" } ], "name": { "name": "TaskTypeMistral", "namespace": "inference._types" }, - "specLocation": "inference/_types/TaskType.ts#L92-L94" + "specLocation": "inference/_types/TaskType.ts#L92-L96" }, { "kind": "enum", @@ -158256,7 +158268,7 @@ "name": "TaskTypeOpenAI", "namespace": "inference._types" }, - "specLocation": "inference/_types/TaskType.ts#L96-L100" + "specLocation": "inference/_types/TaskType.ts#L98-L102" }, { "kind": "enum", @@ -158272,7 +158284,7 @@ "name": "TaskTypeVoyageAI", "namespace": "inference._types" }, - "specLocation": "inference/_types/TaskType.ts#L102-L105" + "specLocation": "inference/_types/TaskType.ts#L104-L107" }, { "kind": "enum", @@ -158285,7 +158297,7 @@ "name": "TaskTypeWatsonx", "namespace": "inference._types" }, - "specLocation": "inference/_types/TaskType.ts#L107-L109" + "specLocation": "inference/_types/TaskType.ts#L109-L111" }, { "kind": "interface", @@ -158531,7 +158543,7 @@ } } ], - "specLocation": "inference/_types/CommonTypes.ts#L1083-L1114" + "specLocation": "inference/_types/CommonTypes.ts#L1085-L1116" }, { "kind": "enum", @@ -158544,7 +158556,7 @@ "name": "VoyageAIServiceType", "namespace": "inference._types" }, - "specLocation": "inference/_types/CommonTypes.ts#L1147-L1149" + "specLocation": "inference/_types/CommonTypes.ts#L1149-L1151" }, { "kind": "interface", @@ -158604,7 +158616,7 @@ } } ], - "specLocation": "inference/_types/CommonTypes.ts#L1116-L1140" + "specLocation": "inference/_types/CommonTypes.ts#L1118-L1142" }, { "kind": "enum", @@ -158620,7 +158632,7 @@ "name": "VoyageAITaskType", "namespace": "inference._types" }, - "specLocation": "inference/_types/CommonTypes.ts#L1142-L1145" + "specLocation": "inference/_types/CommonTypes.ts#L1144-L1147" }, { "kind": "interface", @@ -158708,7 +158720,7 @@ } } ], - "specLocation": "inference/_types/CommonTypes.ts#L1151-L1188" + "specLocation": "inference/_types/CommonTypes.ts#L1153-L1190" }, { "kind": "enum", @@ -158721,7 +158733,7 @@ "name": "WatsonxServiceType", "namespace": "inference._types" }, - "specLocation": "inference/_types/CommonTypes.ts#L1194-L1196" + "specLocation": "inference/_types/CommonTypes.ts#L1196-L1198" }, { "kind": "enum", @@ -158734,7 +158746,7 @@ "name": "WatsonxTaskType", "namespace": "inference._types" }, - "specLocation": "inference/_types/CommonTypes.ts#L1190-L1192" + "specLocation": "inference/_types/CommonTypes.ts#L1192-L1194" }, { "kind": "request", @@ -160985,7 +160997,7 @@ } ] }, - "description": "Create a Mistral inference endpoint.\n\nCreates an inference endpoint to perform an inference task with the `mistral` service.", + "description": "Create a Mistral inference endpoint.\n\nCreate an inference endpoint to perform an inference task with the `mistral` service.", "examples": { "PutMistralRequestExample1": { "description": "Run `PUT _inference/text_embedding/mistral-embeddings-test` to create a Mistral inference endpoint that performs a text embedding task.", @@ -161005,7 +161017,7 @@ }, "path": [ { - "description": "The task type.\nThe only valid task type for the model to perform is `text_embedding`.", + "description": "The type of the inference task that the model will perform.", "name": "task_type", "required": true, "type": { @@ -161030,7 +161042,7 @@ } ], "query": [], - "specLocation": "inference/put_mistral/PutMistralRequest.ts#L29-L72" + "specLocation": "inference/put_mistral/PutMistralRequest.ts#L29-L71" }, { "kind": "response", diff --git a/output/typescript/types.ts b/output/typescript/types.ts index 8f98612b1e..43858166d9 100644 --- a/output/typescript/types.ts +++ b/output/typescript/types.ts @@ -13819,7 +13819,7 @@ export interface InferenceMistralServiceSettings { export type InferenceMistralServiceType = 'mistral' -export type InferenceMistralTaskType = 'text_embedding' +export type InferenceMistralTaskType = 'text_embedding' | 'completion' | 'chat_completion' export interface InferenceOpenAIServiceSettings { api_key: string @@ -13903,7 +13903,7 @@ export type InferenceTaskTypeHuggingFace = 'text_embedding' export type InferenceTaskTypeJinaAi = 'text_embedding' | 'rerank' -export type InferenceTaskTypeMistral = 'text_embedding' +export type InferenceTaskTypeMistral = 'text_embedding' | 'chat_completion' | 'completion' export type InferenceTaskTypeOpenAI = 'text_embedding' | 'chat_completion' | 'completion' diff --git a/specification/inference/_types/CommonTypes.ts b/specification/inference/_types/CommonTypes.ts index 9bd0f96cf7..772d12a74d 100644 --- a/specification/inference/_types/CommonTypes.ts +++ b/specification/inference/_types/CommonTypes.ts @@ -999,7 +999,7 @@ export class MistralServiceSettings { max_input_tokens?: integer /** * The name of the model to use for the inference task. - * Refer to the Mistral models documentation for the list of available text embedding models. + * Refer to the Mistral models documentation for the list of available models. * @ext_doc_id mistral-api-models */ model: string @@ -1011,7 +1011,9 @@ export class MistralServiceSettings { } export enum MistralTaskType { - text_embedding + text_embedding, + completion, + chat_completion } export enum MistralServiceType { diff --git a/specification/inference/_types/TaskType.ts b/specification/inference/_types/TaskType.ts index 7c36323488..0e2a3807ea 100644 --- a/specification/inference/_types/TaskType.ts +++ b/specification/inference/_types/TaskType.ts @@ -90,7 +90,9 @@ export enum TaskTypeHuggingFace { } export enum TaskTypeMistral { - text_embedding + text_embedding, + chat_completion, + completion } export enum TaskTypeOpenAI { diff --git a/specification/inference/put_mistral/PutMistralRequest.ts b/specification/inference/put_mistral/PutMistralRequest.ts index 0bc9713a87..17508e1505 100644 --- a/specification/inference/put_mistral/PutMistralRequest.ts +++ b/specification/inference/put_mistral/PutMistralRequest.ts @@ -29,7 +29,7 @@ import { InferenceChunkingSettings } from '@inference/_types/Services' /** * Create a Mistral inference endpoint. * - * Creates an inference endpoint to perform an inference task with the `mistral` service. + * Create an inference endpoint to perform an inference task with the `mistral` service. * @rest_spec_name inference.put_mistral * @availability stack since=8.15.0 stability=stable visibility=public * @availability serverless stability=stable visibility=public @@ -45,8 +45,7 @@ export interface Request extends RequestBase { ] path_parts: { /** - * The task type. - * The only valid task type for the model to perform is `text_embedding`. + * The type of the inference task that the model will perform. */ task_type: MistralTaskType /**