This version of the OpenSearch documentation is no longer maintained. For the latest version, see the current documentation. For information about OpenSearch version maintenance, see Release Schedule and Maintenance Policy.
Creating connectors for third-party ML platforms
Machine Learning (ML) connectors provide the ability to integrate OpenSearch ML capabilities with third-party ML tools and platforms. Through connectors, OpenSearch can invoke these third-party endpoints to enrich query results and data pipelines.
You can provision connectors in two ways:
An external connector, saved in a connector index, which can be reused and shared with multiple remote models but requires access to both the model, the connector inside of OpenSearch, and the third party being accessed by the connector, such as OpenAI or SageMaker.
A local connector, saved in the model index, which can only be used with one remote model. Unlike a standalone connector, users only need access to the model itself to access an internal connector because the connection is established inside the model.
Supported connectors
As of OpenSearch 2.9, connectors have been tested for the following ML services, though it is possible to create connectors for other platforms not listed here:
- Amazon SageMaker allows you to host and manage the lifecycle of text-embedding models, powering semantic search queries in OpenSearch. When connected, Amazon SageMaker hosts your models and OpenSearch is used to query inferences. This benefits Amazon SageMaker users who value its functionality, such as model monitoring, serverless hosting, and workflow automation for continuous training and deployment.
- OpenAI ChatGPT enables you to invoke an OpenAI chat model from inside an OpenSearch cluster.
- Cohere allows you to use data from OpenSearch to power Cohere’s large language models.
All connectors consist of a JSON blueprint created by machine learning (ML) developers. The blueprint allows administrators and data scientists to make connections between OpenSearch and an AI service or model-serving technology.
You can find blueprints for each connector in the ML Commons repository.
If you want to build your own blueprint, see Building blueprints.
External connector
Admins are only required to enter their credential
settings, such as "openAI_key"
, for the service they are connecting to. All other parameters are defined within the blueprint.
The connector creation API, /_plugins/_ml/connectors/_create
, creates connections that allow users to deploy and register external models through OpenSearch. Using the endpoint
parameter, you can connect ML Commons to any supported ML tool using its specific API endpoint. For example, to connect to a ChatGPT model, you can connect using api.openai.com
, as shown in the following example:
POST /_plugins/_ml/connectors/_create
{
"name": "OpenAI Chat Connector",
"description": "The connector to public OpenAI model service for GPT 3.5",
"version": 1,
"protocol": "http",
"parameters": {
"endpoint": "api.openai.com",
"model": "gpt-3.5-turbo"
},
"credential": {
"openAI_key": "..."
},
"actions": [
{
"action_type": "predict",
"method": "POST",
"url": "https://${parameters.endpoint}/v1/chat/completions",
"headers": {
"Authorization": "Bearer ${credential.openAI_key}"
},
"request_body": "{ \"model\": \"${parameters.model}\", \"messages\": ${parameters.messages} }"
}
]
}
copy
If successful, the connector API responds with the connector_id
for the connection:
{
"connector_id": "a1eMb4kBJ1eYAeTMAljY"
}
With the returned connector_id
we can register a model that uses that connector:
POST /_plugins/_ml/models/_register
{
"name": "openAI-gpt-3.5-turbo",
"function_name": "remote",
"model_group_id": "lEFGL4kB4ubqQRzegPo2",
"description": "test model",
"connector_id": "a1eMb4kBJ1eYAeTMAljY"
}
Local connector
Admins are only required to enter their credential
settings, such as "openAI_key"
, for the service they are connecting to. All other parameters are defined within the blueprint.
To create an internal connector, add the connector
parameter to the Register model API, as shown in the following example:
POST /_plugins/_ml/models/_register
{
"name": "openAI-GPT-3.5: internal connector",
"function_name": "remote",
"model_group_id": "lEFGL4kB4ubqQRzegPo2",
"description": "test model",
"connector": {
"name": "OpenAI Connector",
"description": "The connector to public OpenAI model service for GPT 3.5",
"version": 1,
"protocol": "http",
"parameters": {
"endpoint": "api.openai.com",
"max_tokens": 7,
"temperature": 0,
"model": "text-davinci-003"
},
"credential": {
"openAI_key": "..."
},
"actions": [
{
"action_type": "predict",
"method": "POST",
"url": "https://${parameters.endpoint}/v1/completions",
"headers": {
"Authorization": "Bearer ${credential.openAI_key}"
},
"request_body": "{ \"model\": \"${parameters.model}\", \"prompt\": \"${parameters.prompt}\", \"max_tokens\": ${parameters.max_tokens}, \"temperature\": ${parameters.temperature} }"
}
]
}
}
}
Registering and deploying a connected model
After a connection has been created, use the connector_id
from the response to register and deploy a connected model.
To register a model, you have the following options:
- You can use
model_group_id
to register a model version to an existing model group. - If you do not use
model_group_id
, ML Commons creates a model with a new model group.
If you want to create a new model_group
, use the following example:
POST /_plugins/_ml/model_groups/_register
{
"name": "remote_model_group",
"description": "This is an example description"
}
ML Commons returns the following response:
{
"model_group_id": "wlcnb4kBJ1eYAeTMHlV6",
"status": "CREATED"
}
The following example registers a model named openAI-gpt-3.5-turbo
:
POST /_plugins/_ml/models/_register
{
"name": "openAI-gpt-3.5-turbo",
"function_name": "remote",
"model_group_id": "wlcnb4kBJ1eYAeTMHlV6",
"description": "test model",
"connector_id": "a1eMb4kBJ1eYAeTMAljY"
}
ML Commons returns the task_id
and registration status of the model:
{
"task_id": "cVeMb4kBJ1eYAeTMFFgj",
"status": "CREATED"
}
You can use the task_id
to find the model_id
, as shown the following example:
GET task request
GET /_plugins/_ml/tasks/cVeMb4kBJ1eYAeTMFFgj
GET task response
{
"model_id": "cleMb4kBJ1eYAeTMFFg4",
"task_type": "REGISTER_MODEL",
"function_name": "REMOTE",
"state": "COMPLETED",
"worker_node": [
"XPcXLV7RQoi5m8NI_jEOVQ"
],
"create_time": 1689793598499,
"last_update_time": 1689793598530,
"is_async": false
}
Lastly, use the model_id
to deploy the model:
Deploy model request
POST /_plugins/_ml/models/cleMb4kBJ1eYAeTMFFg4/_deploy
Deploy model response
{
"task_id": "vVePb4kBJ1eYAeTM7ljG",
"status": "CREATED"
}
Use the task_id
from the deploy model response to make sure the model deployment completes:
Verify deploy completion request
GET /_plugins/_ml/tasks/vVePb4kBJ1eYAeTM7ljG
Verify deploy completion response
{
"model_id": "cleMb4kBJ1eYAeTMFFg4",
"task_type": "DEPLOY_MODEL",
"function_name": "REMOTE",
"state": "COMPLETED",
"worker_node": [
"n-72khvBTBi3bnIIR8FTTw"
],
"create_time": 1689793851077,
"last_update_time": 1689793851101,
"is_async": true
}
After a successful deployment, you can test the model using the Predict API set in the connector’s action
settings, as shown in the following example:
POST /_plugins/_ml/models/cleMb4kBJ1eYAeTMFFg4/_predict
{
"parameters": {
"messages": [
{
"role": "system",
"content": "You are a helpful assistant."
},
{
"role": "user",
"content": "Hello!"
}
]
}
}
The Predict API returns inference results for the connected model, as shown in the following example response:
{
"inference_results": [
{
"output": [
{
"name": "response",
"dataAsMap": {
"id": "chatcmpl-7e6s5DYEutmM677UZokF9eH40dIY7",
"object": "chat.completion",
"created": 1689793889,
"model": "gpt-3.5-turbo-0613",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "Hello! How can I assist you today?"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 19,
"completion_tokens": 9,
"total_tokens": 28
}
}
}
]
}
]
}
Examples
The following example connector requests show how to create a connector with supported third-party tools.
OpenAI chat connector
The following example creates a standalone OpenAI chat connector. The same options can be used for an internal connector under the connector
parameter:
POST /_plugins/_ml/connectors/_create
{
"name": "OpenAI Chat Connector",
"description": "The connector to public OpenAI model service for GPT 3.5",
"version": 1,
"protocol": "http",
"parameters": {
"endpoint": "api.openai.com",
"model": "gpt-3.5-turbo"
},
"credential": {
"openAI_key": "..."
},
"actions": [
{
"action_type": "predict",
"method": "POST",
"url": "https://${parameters.endpoint}/v1/chat/completions",
"headers": {
"Authorization": "Bearer ${credential.openAI_key}"
},
"request_body": "{ \"model\": \"${parameters.model}\", \"messages\": ${parameters.messages} }"
}
]
}
After creating the connector, you can retrieve the task_id
and connector_id
to register and deploy the model and then use the Predict API, similarly to a standalone connector.
Amazon SageMaker
The following example creates a standalone Amazon SageMaker connector. The same options can be used for an internal connector under the connector
parameter:
POST /_plugins/_ml/connectors/_create
{
"name": "sagemaker: embedding",
"description": "Test connector for Sagemaker embedding model",
"version": 1,
"protocol": "aws_sigv4",
"credential": {
"access_key": "...",
"secret_key": "...",
"session_token": "..."
},
"parameters": {
"region": "us-west-2",
"service_name": "sagemaker"
},
"actions": [
{
"action_type": "predict",
"method": "POST",
"headers": {
"content-type": "application/json"
},
"url": "https://runtime.sagemaker.${parameters.region}.amazonaws.com/endpoints/lmi-model-2023-06-24-01-35-32-275/invocations",
"request_body": "[\"${parameters.inputs}\"]"
}
]
}
The credential
parameter contains the following options reserved for aws_sigv4
authentication:
access_key
: Required. Provides the access key for the AWS instance.secret_key
: Required. Provides the secret key for the AWS instance.session_token
: Optional. Provides a temporary set of credentials for the AWS instance.
The parameters
section requires the following options when using aws_sigv4
authentication:
region
: The AWS Region in which the AWS instance is located.service_name
: The name of the AWS service for the connector.
Cohere
The following example request creates a standalone Cohere connection:
POST /_plugins/_ml/connectors/_create
{
"name": "YOUR CONNECTOR NAME",
"description": "YOUR CONNECTOR DESCRIPTION",
"version": "YOUR CONNECTOR VERSION",
"protocol": "http",
"credential": {
"cohere_key": "ADD YOUR Cohere API KEY HERE"
},
"parameters": {
"model": "embed-english-v2.0",
"truncate": "END"
},
"actions": [
{
"action_type": "predict",
"method": "POST",
"url": "https://api.cohere.ai/v1/embed",
"headers": {
"Authorization": "Bearer ${credential.cohere_key}"
},
"request_body": "{ \"texts\": ${parameters.texts}, \"truncate\": \"${parameters.truncate}\", \"model\": \"${parameters.model}\" }"
}
]
}
copy
Next steps
- To learn more about using models in OpenSearch, see ML Framework.
- To learn more about model access control and model groups, see Model access control.