Configuring an embedding connection

Configuring an embedding connection

An embedding connection in Kyvos refers to the configuration that links the platform to an embedding model—a machine learning model that transforms textual or complex data into dense, numerical vectors called embeddings. This helps improve semantic search and conversational analytics by matching user intent, not just keywords, making it easier to find relevant data using natural language.

You can now specify the type of connection while uploading the custom provider.

After specifying the connection type, you can upload the required JAR and configuration files in a ZIP format corresponding to the selected type.

OpenAI

To configure the GenAI embedding connection, perform the following steps: 

Parameter/Field

Description

Parameter/Field

Description

Connection Name

A unique name that identifies your GenAI connections.

Provider

The name of the GenAI provider the system will use to generate output. Select the required provider from the list.

Model

Select the name of the model for generating embeddings.

EndPoint

Specify a unique identifier for the end user, which helps OpenAI monitor and detect abuse.

  • For OpenAI, provide the endpoint in the following format:
    /chat/completions

Prompt Token Limit

Specify maximum tokens allowed for prompt in single request for current model.

Similarity Upper Threshold

Specify upper threshold to control the selection of records for Retrieval Augmented Generation (RAG).

Similarity Lower Threshold

Specify lower threshold to control Retrieval Augmented Generation (RAG) record selection.

Template Folder Path

Provide the folder path for templates.

RAG Max Records

Specify the maximum number of records (RAG) required in a user prompt.

Azure OpenAI

Parameter/Field

Description

Parameter/Field

Description

Connection Name

A unique name that identifies your GenAI connections.

Provider

The name of the GenAI provider the system will use to generate output. Select the required provider from the list.

Model

Select the name of the model for generating embeddings.  

EndPoint

Specify a unique identifier for the end user, which helps OpenAI monitor and detect abuse.

  • For Azure OpenAI, provide the endpoint in the following format:
    {deployment-id}/embeddings?api-version=2024-10-21
    For example, /text-embedding-ada-002/embeddings?api-version=2023-05-15

Prompt Token Limit

Specify maximum tokens allowed for prompt in single request for current model.

Similarity Upper Threshold

Specify upper threshold to control the selection of records for Retrieval Augmented Generation (RAG).

Similarity Lower Threshold

Specify lower threshold to control Retrieval Augmented Generation (RAG) record selection.

Template Folder Path

Provide the folder path for templates.

RAG Max Records

Specify the maximum number of records (RAG) required in a user prompt.

AWS Bedrock

Parameter/Field

Description

Parameter/Field

Description

Connection Name

A unique name that identifies your GenAI connections.

Provider

The name of the GenAI provider the system will use to generate output. Select the required provider from the list.

Embedding Model

Select the name of the model for generating embeddings.

EndPoint

Specify a unique identifier for the end user, which helps OpenAI monitor and detect abuse.

Embedding Prompt Token Limit

Specify maximum tokens allowed for prompt in single request for current model.

Similarity Upper Threshold

Specify upper threshold to control the selection of records for Retrieval Augmented Generation (RAG).

Similarity Lower Threshold

Specify lower threshold to control Retrieval Augmented Generation (RAG) record selection.

AWS Region LLM

Provide the region for llm

Template Folder Path

Provide the folder path for templates.

RAG Max Records

Specify the maximum number of records (RAG) required in a user prompt.

Copyright Kyvos, Inc. 2025. All rights reserved.