LogoLogo
Tonic Validatetonic_validateDocs homeTonic.ai
  • Tonic Validate guide
  • About Tonic Validate
    • What is Tonic Validate?
    • Validate components and tools
    • Validate workflows
  • Getting started with Validate
    • Starting your Validate account
    • Setting up the Validate SDK
    • Quickstart example - create a run and log responses
    • Creating and revoking Validate API keys
  • About RAG metrics
    • About using metrics to evaluate a RAG system
    • RAG components summary
    • RAG metrics summary
    • RAG metrics reference
  • Benchmarks and projects
    • Managing benchmarks in Validate
    • Managing projects in Validate
  • Runs
    • Starting a Validate run
    • Viewing and managing runs
  • Production monitoring
    • Configuring your RAG system to send questions to Validate
    • Viewing the metric scores and logged questions
  • Code examples
    • End-to-end example using LlamaIndex
Powered by GitBook
On this page
  • Development project workflow
  • Create your benchmark (optional)
  • Create your project
  • Create a run
  • Review the run results
  • Update and iterate
  • Production monitoring project workflow
  • Create your project
  • Configure your RAG system to send questions to the project
  • View the results

Was this helpful?

Export as PDF
  1. About Tonic Validate

Validate workflows

Last updated 12 months ago

Was this helpful?

Development project workflow

The overall process to use a Tonic Validate development project to evaluate your RAG system consists of the following:

Create your benchmark (optional)

A Validate run analyzes a RAG system performance against a set of questions and optional ideal answers.

You can use the Validate application or SDK to add the benchmark to Validate.

Create your project

Create a run

The run configuration includes:

  • The project

  • The questions for to analyze the RAG performance. A Validate benchmark is one way to provide the question data.

  • Any metadata about the RAG data, such as the type of LLM, the embedder, or the retrieval algorithm

  • The metrics to calculate

Review the run results

Update and iterate

Based on the run results, you update the RAG system to improve the results, then create another run.

You compare the run results to see if your changes improved the quality of the answers.

Production monitoring project workflow

After you release your RAG system, you can use a Validate production monitoring project to track how well it answers user questions.

Create your project

Configure your RAG system to send questions to the project

  • Each question that a user asked

  • The answer that the RAG system provided

  • The context that the RAG system used

View the results

As it receives the questions, Validate generates metric scores.

One way to provide the questions and answers is to .

Next, use the Validate application to .

Use the Validate SDK to .

From the Validate application, .

Use the Validate application to .

In your RAG system, you to send the following to the production monitoring project:

In the Validate application, you can for the questions that Validate received from the RAG system.

You can also .

configure a benchmark in Validate
create a development project
create a run for the project
review the scores and metrics from the run
add a call to the Validate SDK
view a timeline of the average scores
view and filter the list of questions
create a production monitoring project
Overview diagram of a Validate development project workflow
Overview diagram of a Validate production monitoring project workflow