Use model groups to logically group models and control who gets access to them.
Register a model group using the register operation in the Model Group APIs, as shown in the
following example:
POST /_plugins/_ml/model_groups/_register
{
"name": "general pretrained models",
"description": "A model group for pretrained models hosted by OCI Search with OpenSearch"
}
Make note of the model_group_id returned in the response:
You have the following three options for semantic search:
Option 1: Register and deploy a pretrained
model hosted in OCI Search with OpenSearch using the steps
described in Using an OpenSearch Pretrained Model. This
option is the simplest to use, you don't need to configure any
additional IAM policies, and the
payload isn't as complex as the payload for the next option.
Option 2: Import, register and deploy an OpenSearch pretrained model
using the steps described in Custom Models. This includes uploading
the model file to an Object Storage bucket, and then specifying the
model file's Object Storage URL when you register the model.
Option 3: Create a Generative AI connector to register a remote
embedding model such as the cohere.embed-english-v3.0
model. For more information, see Conversational Search with OCI Generative AI.
This walkthrough shows you how to use option 1, a pretrained model.
To register a pretrained model, you need the following:
model_group_id: If you completed Step 1, this is the value
for model_group_id to the _register request.
Make note of the model_ID value returned in the response to use when
you deploy the model.
Step 4: Deploy the Model 🔗
After the register operation is completed for the model, you can deploy the model to
the cluster using the deploy operation of the Model APIs,
passing the model_ID from the Get operation response in the
previous step, as shown in the following example:
POST /_plugins/_ml/models/<embedding_model_ID>/_deploy
Make note of the task_id returned in the response, you can use the
task_id to check the status of the operation.
to check the status of the register operation, use the task_ID with
the Get operation of the Tasks APIs, as shown in the following
example:
GET /_plugins/_ml/tasks/<task_ID>
When the deploy operation is complete, the status value in the
response to the Get operation is COMPLETED.
Step 5: Create a k-NN Ingestion Pipeline 🔗
After the deploy operation is complete, create an ingestion pipeline using the
deployed model. The ingestion pipeline uses the deployed model to automatically
generate the embedding vectors for each document at ingestion time. The processor
handles everything for the embedding, so you only need to appropriately map the
expected text field in the document being converted into embedding. The following
example shows creating an ingestion pipeline:
If the ingestion pipeline was created, the following response is returned:
{
"acknowledged": true
}
Step 6: Create Index 🔗
Create an index using the ingestion pipeline created in the previous step. You can
use any of the available ANN engines in the index. The following example uses the
Lucene Engine:
The passage_text field for the create index matches the
passage_text field in the ingestion pipeline, this makes it so
the pipeline knows how to create embeddings and then map them to documents at
ingestion time.
To help you choose the engine you want to use, and the available configuration
parameters for those engines, see k-NN index and Approximate k-NN search.
The following is an example response for a successful index creation:
Ingest data into you index as shown in the following example:
POST /lucene-index/_doc/1
{
"<text_field_name>": "there are many sharks in the ocean"
}
POST /lucene-index/_doc/2
{
"<text_field_name>": "fishes must love swimming"
}
POST /lucene-index/_doc/3
{
"<text_field_name>": "summers are usually very hot"
}
POST /lucene-index/_doc/4
{
"<text_field_name>": "florida has a nice weather all year round"
}
You can see that the query question doesn't mention Florida, weather, nor
summer , however the model infers the semantic meaning of temperatures and
miami to return the most relevant answers.