Configuring an embedding connection
An embedding connection in Kyvos refers to the configuration that links the platform to an embedding model—a machine learning model that transforms textual or complex data into dense, numerical vectors called embeddings. This helps improve semantic search and conversational analytics by matching user intent, not just keywords, making it easier to find relevant data using natural language.
You can now specify the type of connection while uploading the custom provider.
After specifying the connection type, you can upload the required JAR and configuration files in a ZIP format corresponding to the selected type.
OpenAI
To configure the GenAI embedding connection, perform the following steps:
Parameter/Field | Description |
|---|---|
Connection Name | A unique name that identifies your GenAI connections. |
Provider | The name of the GenAI provider the system will use to generate output. Select the required provider from the list. |
Model | Select the name of the model for generating embeddings. |
EndPoint | Specify a unique identifier for the end user, which helps OpenAI monitor and detect abuse.
|
Prompt Token Limit | Specify maximum tokens allowed for prompt in single request for current model. |
Similarity Upper Threshold | Specify upper threshold to control the selection of records for Retrieval Augmented Generation (RAG). |
Similarity Lower Threshold | Specify lower threshold to control Retrieval Augmented Generation (RAG) record selection. |
Template Folder Path | Provide the folder path for templates. |
RAG Max Records | Specify the maximum number of records (RAG) required in a user prompt. |
Azure OpenAI
Parameter/Field | Description |
|---|---|
Connection Name | A unique name that identifies your GenAI connections. |
Provider | The name of the GenAI provider the system will use to generate output. Select the required provider from the list. |
Model | Select the name of the model for generating embeddings. |
EndPoint | Specify a unique identifier for the end user, which helps OpenAI monitor and detect abuse.
|
Prompt Token Limit | Specify maximum tokens allowed for prompt in single request for current model. |
Similarity Upper Threshold | Specify upper threshold to control the selection of records for Retrieval Augmented Generation (RAG). |
Similarity Lower Threshold | Specify lower threshold to control Retrieval Augmented Generation (RAG) record selection. |
Template Folder Path | Provide the folder path for templates. |
RAG Max Records | Specify the maximum number of records (RAG) required in a user prompt. |
AWS Bedrock
Parameter/Field | Description |
|---|---|
Connection Name | A unique name that identifies your GenAI connections. |
Provider | The name of the GenAI provider the system will use to generate output. Select the required provider from the list. |
Embedding Model | Select the name of the model for generating embeddings. |
EndPoint | Specify a unique identifier for the end user, which helps OpenAI monitor and detect abuse.
|
Embedding Prompt Token Limit | Specify maximum tokens allowed for prompt in single request for current model. |
Similarity Upper Threshold | Specify upper threshold to control the selection of records for Retrieval Augmented Generation (RAG). |
Similarity Lower Threshold | Specify lower threshold to control Retrieval Augmented Generation (RAG) record selection. |
AWS Region LLM | Provide the region for llm |
Template Folder Path | Provide the folder path for templates. |
RAG Max Records | Specify the maximum number of records (RAG) required in a user prompt. |