Edit

Share via


Set up the data query for chat solution in Edge RAG Preview, enabled by Azure Arc

Configure data queries and model settings for your Edge RAG chat solution to optimize your chat results. Adjust search types, tune model parameters, and refine your chat experience in the Edge RAG developer portal.

Important

Edge RAG Preview, enabled by Azure Arc is currently in PREVIEW. See the Supplemental Terms of Use for Microsoft Azure Previews for legal terms that apply to Azure features that are in beta, preview, or otherwise not yet released into general availability.

Prerequisites

Before you begin:

Configure model settings

To get started, configure the model settings.

  1. Go to the local portal using the domain name provided at deployment and app registration. For example, https://arcrag.contoso.com.

  2. Sign in with developer credentials that have both "EdgeRAGDeveloper" and "EdgeRAGEndUser" roles assigned. If you have the right access configured, you're automatically redirected to the developer portal.

  3. Select Get Started.

  4. Select the Chat tab to get to the Chat playground.

  5. In the Data inferencing section, select the Type of search.

    Search type Description
    Hybrid text search Search that combines keyword (text) search and vector (contextual) search.
    Text search Search that queries exact words or phrases in documents.
    Vector search Search that queries contextual similarity rather than exact keyword matching.
    Hybrid multimodal search Search that combines multiple modalities, like text and image, simultaneously.
  6. Change the model parameters for Temperature, Top-N, Top-P, and others as needed.

    Screenshot of the data inferencing section of the chat window that allows you to change the model parameters.

  7. Review and update the system prompt as appropriate for your solution.

  8. Any changes that you make are applied when you submit a new question in the chat.

Test chat results

Next, test the chat endpoint.

  1. In the chat window, enter a question that uses a simple question and answer format. Queries that require summarization across multiple documents might not return accurate answers.

    Screenshot of the chat playground with a question in the chat window.

  2. (Optional) To refresh the chat playground and clear the chat history, select New chat.

  3. (Optional) Test the end user experience by using the chat solution app for Edge RAG.

View details to refine settings

Use the chat response details to analyze and fine-tune your model and search parameters to optimize your chat responses.

  1. Under the chat response, select View details.

    Screenshot that shows the view details option underneath the chat response.

  2. Use the chat details to understand the impact of the inferencing parameters on the language model's response to your question.

    Field Description
    LLM response Response from the large language model (LLM) for the corresponding question.
    User question Question asked by user.
    Parameters Parameters that are used to search content and generate LLM response.
    System prompt System prompt input set by Developer.
    Reranked chunks Shows search IDs by reranking score.
    LLM Input chunks Relevant chunks passed to LLM as retrieved content; the chunks are selected based on text strictness and image strictness.
    Search details Shows search details.
    Results from text search Results from textual search for a query; each result shows reranking score, search distance, text, file path, chunk ID, and last modified date.
    Results from vector search Results from semantic search for a query; each result shows reranking score, search distance, text, file path, chunk ID, and last modified date.
    Results from image search Results from image search for a query, each result shows reranking score, file path, last modified date.

    Screenshot that shows the details of the chat response including the search type, parameters, and more.

  3. To analyze the Details, select Copy to paste a JSON version of the text into a text editor.

  4. Tune the inferencing parameters to get the type of responses that you want for your ingested data.

Get the API endpoint

When you're satisfied with the solution, select on View the endpoint to get the API endpoint to use in your downstream applications.