Evaluations in LLMKit

Evals! Evals! Evals!

Building any LLM powered product requires you to have a deep understanding of the quality of your prompts. You need to know how they are performing and you need to be able to assess how small changes are affecting the performance.


What Are Evaluations?

Evaluations let you systematically test your prompts by running them against a set of inputs and reviewing the outputs. They’re essential for:

  • Checking how your prompt performs across different scenarios.
  • Comparing multiple versions of a prompt.
  • Identifying weaknesses to improve.

In short, evaluations are your tool for ensuring your prompts are effective and ready for action.


Creating an Evaluation

Step 1: Navigate to Evals

From the main dashboard, head to the Evals tab. Select the prompt you want to create an eval for and click "Create new test eval".

Evals Step 1

Step 2: Define template field inputs

Name your eval and populate the template fields for your prompt. Click "Add eval" when complete.

Evals Step 2

Step 3: Create more evals

Create as many evals as you want to help cover your various testing scenarios.

Running an Evaluation

Once your evals are created, it's time to run them!

Step 1: Run the evals

With the same prompt selected, select "New eval run"

Evals Step 3

Step 2: Perform human evaluations

Once the evals are complete, you will see the inputs and outputs of your eval. Here you can select a 1-5 rating for each.

Evals Step 4

Step 3: View the results

Once you provide a rating for all the evals, you can see the results and track your prompt performance over time across versions.

Evals Step 5

Why Evaluations Matter

Evaluations streamline the process of perfecting your prompts. They offer:

  • Structured Testing: Test a range of inputs methodically.
  • Performance Tracking: See how changes improve your prompt over time.
  • Confidence: Validate your prompt before deploying it.