Learn how to use the Inference API for semantic search.
For this example, you will need:
-
An Elastic deployment with minimum 4GB machine learning node
- We'll be using Elastic Cloud for this example (available with a free trial)
-
Elasticsearch 8.13 or above.
-
A paid Cohere account is required to use the Inference API with the Cohere service as the Cohere free trial API usage is limited.
If you don't have an Elastic Cloud deployment, sign up here for a free trial.
- Go to the Create deployment page
- Under Advanced settings, go to Machine Learning instances
- You'll need at least 4GB RAM per zone for this tutorial
- Select Create deployment
To get started, we'll need to connect to our Elastic deployment using the Python client (version 8.12.0 or above). Because we're using an Elastic Cloud deployment, we'll use the Cloud ID to identify our deployment.
First we need to pip install the following packages:
elasticsearch
Next, we need to import the modules we need. 🔐 NOTE: getpass enables us to securely prompt the user for credentials without echoing them to the terminal, or storing it in memory.
Now we can instantiate the Python Elasticsearch client.
First we prompt the user for their password and Cloud ID.
Then we create a client object that instantiates an instance of the Elasticsearch class.
Confirm that the client has connected with this test:
Refer to the documentation to learn how to connect to a self-managed deployment.
Read this page to learn how to connect using API keys.
Create the inference endpoint
Let's create the inference endpoint by using the Create inference API.
You'll need an Cohere API key for this that you can find in your Cohere account under the API keys section. A paid membership is required to complete the steps in this notebook as the Cohere free trial API usage is limited.
Create an ingest pipeline with an inference processor
Create an ingest pipeline with an inference processor by using the put_pipeline method. Reference the inference endpoint created above as the model_id to infer against the data that is being ingested in the pipeline.
Let's note a few important parameters from that API call:
inference: A processor that performs inference using a machine learning model.model_id: Specifies the ID of the inference endpoint to be used. In this example, the model ID is set tocohere_embeddings.input_output: Specifies input and output fields.input_field: Field name from which thedense_vectorrepresentation is created.output_field: Field name which contains inference results.
Create index
The mapping of the destination index – the index that contains the embeddings that the model will create based on your input text – must be created. The destination index must have a field with the dense_vector field type to index the output of the Cohere model.
Let's create an index named cohere-movie-embeddings with the mappings we need.
Insert Documents
Let's insert our example dataset of 12 movies. You need a paid Cohere account to complete this step, otherwise the documentation ingest will time out due to the API request rate limits.
Semantic search
After the dataset has been enriched with the embeddings, you can query the data using semantic search. Pass a query_vector_builder to the k-nearest neighbor (kNN) vector search API, and provide the query text and the model you have used to create the embeddings.
NOTE: The value of model_id in the query_vector_builder must match the value of inference_id you created in the first step.