Creating a Safety Evaluation
Use the Collinear AI Platform to create a new safety evaluation
Introduction
Once you connect your model or upload your dataset, you can run a safety evaluation on it using Collinear AI’s suite of safety judges.
Steps to Create a Safety evaluation
Select Safety Evaluation
Click on “Safety” under “Select Evaluation Type”
Select a Judge
Select the type of safety judge you want to use
- CollinearGuard (rating): This is Collinear AI’s proprietary likert based (rating scale from 1 to 5) model for custom scoring criteria
Use the 5 rows to create a scoring criteria that corresponds to the score column on the left.
- CollinearGuard (classification): This is Collinear AI’s proprietary binary model for evaluating response, prompt and refusal safety.
Select one of options:
- Prompt - This evaluates the users input
- Response - This evaluates the LLM generated response
- Refusal - This evaluates when the AI declines to respond
- CollinearGuard Nano Categories: This is Collinear AI’s enhanced nano model that also outputs safety categories
-
Llama Guard 3: This is Meta’s off the shelf safety judge
-
Prompted Model: You can use any off the shelf model with a custom prompt You can select any model from the dropdown and customize the prompt template to suit your evaluation
Enter Judge Name
Name your judge
Enter Description
Enter a short description of your judge and click “Confirm” once you’re done
Finalize Run
Enter a name for your run and click “Confirm” once you’re done.