Skip to main content

DataSource Chat AI

A Chat AI is an artificial intelligence system designed to engage in conversations with users through natural language. It leverages techniques such as natural language processing (NLP) and machine learning to understand, process, and generate human-like responses in real-time, enabling tasks like answering questions, providing recommendations, or holding interactive dialogues. Popular examples include virtual assistants and AI-driven customer support chatbots.

Supported models and sub-models

  1. OpenAI
    • GPT-3.5 Turbo
  2. Groq
    • Meta Llama 3 70B

Engaging in Practical Implementation

Open new Project in QuickIntegration Platform, and then follow these steps to get your flow working

  1. Click On the Connection Properties
  2. Select the DataSource Type from drop down
  3. Provide the Credentials
  4. Click on Submit to save the Credentials
  5. On the left side of the palette, you'll find the Configured Properties ready to be utilized within your API.
Chat AI configuration
FieldsDescriptionExample
DataSource NameDatasource Name which is configured in connections propertiesgroqds
DataSource TypeCHAT_AICHAT_AI
EnvironmentProvides a production environment where you can deploy applications and APIs publiclydev
URLA URL (Uniform Resource Locator) is a unique identifier used to locate a resource on the Internethttps://api.groq.com/openai
API TokenAn API token is a unique identifier used to authenticate and authorize access to an API, ensuring secure communication between a client and the server.<your-api-token>
Connection TimeoutThe maximum amount of time the driver will wait while attempting to establish a connection to the database.60000 (ms)
Read TimeoutThe maximum amount of time the client will wait for a response from the server after a connection is established60000 (ms)
ModelA model in the context of machine learning is an algorithm or mathematical structure that processes input data, identifies patterns, and makes predictions or decisions based on that data. It is trained using data to learn relationships and rules, which it then applies to new, unseen inputs.Groq
Sub-modelA sub-model in machine learning is a smaller or specialized component of a larger model, designed to handle a specific task or subset of the overall problem. It works in conjunction with other sub-models or the main model to contribute to the final prediction or decision-making process.Meta Llama 3 70B
OptionsDescriptionExample
Log ProbabilityThe log probability of a token is the logarithm of the probability assigned to that token. This is useful for numerical stability and for comparing relative probabilities, as it avoids issues with very small probabilities.true
TemperatureThe sampling temperature to use that controls the apparent creativity of generated completions. Higher values will make output more random while lower values will make results more focused and deterministic. It is not recommended to modify temperature and top_p for the same completions request as the interaction of these two settings is difficult to predict.0.8
Frequency PenaltyNumber between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model’s likelihood to repeat the same line verbatim.0.0
Max TokensThe maximum number of tokens to generate in the chat completion. The total length of input tokens and generated tokens is limited by the model’s context length.4096
UserA unique identifier representing your end-user, which can help OpenAI to monitor and detect abuse.-