Introduction

Once you connect your model or upload your dataset, you can run a safety evaluation on it using Collinear AI’s suite of safety judges.

Steps to Create a Safety evaluation

Select Safety Evaluation

Click on “Safety” under “Select Evaluation Type”

Select a Judge

Select the type of safety judge you want to use

  1. CollinearGuard (rating): This is Collinear AI’s proprietary likert based (rating scale from 1 to 5) model for custom scoring criteria

Use the 5 rows to create a scoring criteria that corresponds to the score column on the left.

  1. CollinearGuard (classification): This is Collinear AI’s proprietary binary model for evaluating response, prompt and refusal safety.

Select one of options:

  1. Prompt - This evaluates the users input
  2. Response - This evaluates the LLM generated response
  3. Refusal - This evaluates when the AI declines to respond
  1. CollinearGuard Nano Categories: This is Collinear AI’s enhanced nano model that also outputs safety categories
  1. Llama Guard 3: This is Meta’s off the shelf safety judge

  2. Prompted Model: You can use any off the shelf model with a custom prompt You can select any model from the dropdown and customize the prompt template to suit your evaluation

Enter Judge Name

Name your judge

Enter Description

Enter a short description of your judge and click “Confirm” once you’re done

Finalize Run

Enter a name for your run and click “Confirm” once you’re done.