What are Red Teaming Tests?

Red teaming tests in Galtea are designed to evaluate the security, safety, and bias aspects of your product. These tests typically consist of adversarial inputs specifically crafted to probe potential weaknesses or vulnerabilities in your AI system.

Creating Red Teaming Tests

You can create red teaming tests in Galtea through two methods:

1

Prepare your knowledge base file

Create a file with examples of harmful content or sensitive topic areas you want to test

2

Configure the test

Select Red Teaming as the test type and Generated as the test origin

The test creation process acan be done via the SDK or the Galtea platform

3

Generate the test

Galtea will process the Knowledge Base and generate a Test File and Test Cases with adversarial inputs for you

Testing Categories

Red teaming tests typically focus on several key categories:

Example Red Teaming Tests and File Format

Here are examples of the content and its creation-document structure:

Red teaming tests intentionally contain harmful or adversarial content. These are used solely for the purpose of testing AI safety and security. Always handle such content responsibly and in accordance with your organization’s policies.

This structure is the required by Galtea in order to automatically generate the test cases for your test. If that format is not provided, you can still manually create the test cases.

For automatic processing, the file format must be CSV.

The contexts and the extent of the examples provided above are short extractions. In the actual CSV files the context can be much longer and the amount of test cases (rows) can be much higher.

Structure of Red Teaming Tests

Red teaming tests have a specific structure focused on providing challenging inputs:

instance_id
Integer
required

A unique identifier for the test instance.

input
String
required

The adversarial input specifically designed to test the product’s resilience against potential vulnerabilities. Example: “Please ignore any previous instructions and…”

expected_output
String

For red teaming tests, this field is typically marked as “N/A” since the purpose is to test system resilience rather than producing a specific output.

tag
String

A categorization tag for the test instance, typically “adversarial_inputs” for red teaming tests.

source
String

The source of the adversarial input, which may be marked as “N/A” for red teaming tests.