You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: tutorial/markdown/generated/vector-search-cookbook/autovec-structured-autovec_langchain.md
+15-17Lines changed: 15 additions & 17 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -59,25 +59,25 @@ Now, before we actually create embeddings for the documents, we need to deploy a
59
59
## Selecting the Model
60
60
1. To select the model, you first need to navigate to the "<B>AI Services</B>" tab, then select "<B>Models</B>" and click on "<B>Deploy New Model</B>".
2. Enter the <B>model name</B>, and choose the model that you want to deploy. After selecting your model, choose the <B>model infrastructure</B> and <B>region</B> where the model will be deployed. **Ensure this matches your database cluster region.**
@@ -86,15 +86,15 @@ Now, we are at the step that will help us create the embeddings/vectors. To proc
86
86
87
87
1. For deploying the autovectorization, you need to go to the <B>`AI Services`</B> tab, then click on <B>`Workflows`</B>, and then click on <B>`Create New Workflow`</B>.
2. Start your workflow deployment by giving it a name and selecting where your data will be provided to the auto-vectorization service. There are currently 3 options: <B>`pre-processed data (JSON format) from Capella`</B>, <B>`pre-processed data (JSON format) from external sources (S3 buckets)`</B> and <B>`unstructured data from external sources (S3 buckets)`</B>. For this tutorial, we will choose the first option, which is pre-processed data from Capella.
3. Now, select the <B>`cluster`</B>, <B>`bucket`</B>, <B>`scope`</B> and <B>`collection`</B> from which you want to select the documents and get the data vectorized.
5. After choosing the type of mapping, you can optionally create a vector index on the new vector embedding field. While vector search will work without an index using brute force, creating an index is **highly recommended** for better performance, especially with larger datasets.
6. Below screenshot highlights the whole process which were mentioned above, and click next afterwards as shown below. We will be going ahead with the custom source field mappings for this tutorial.
- For this tutorial, capella based embedding model is used as can be seen in the image above. API credentials can be uploaded using the file downloaded during model setup section or it can be entered manually as well.
124
124
- Choices between private and insecure networking is available to choose.
@@ -128,18 +128,18 @@ Now, we are at the step that will help us create the embeddings/vectors. To proc
128
128
129
129
8. <B>`Workflow Summary`</B> will display all the necessary details of the workflow including `Data Source`, `Model Service` and `Billing Overview` as shown in image below.
After this step, your vector embeddings for the selected fields should be ready and you can checkout in your document schema a vector field should be there as highlighter below in the image.
In the next step, we will demonstrate how we can use the generated vectors to perform vector search.
145
145
@@ -326,5 +326,3 @@ The search results display the top 3 (or `k`) most semantically similar hotels b
326
326
327
327
> Your vector search pipeline is working if the returned documents feel meaningfully related to your natural language query—even when exact keywords do not match. Feel free to experiment with increasingly descriptive queries to observe the semantic power of the embeddings.
0 commit comments