What are Red Teaming Tests?

Red teaming tests in Galtea are designed to evaluate the security, safety, and bias aspects of your product. These tests typically consist of different types of threats like adversarial inputs specifically crafted to probe potential weaknesses or vulnerabilities in your AI system. To further enhance the diversity and evasiveness of these tests, Galtea can apply various Red Teaming Strategies to the prompts generated from these threats.

Creating Red Teaming Tests

You can create red teaming tests in Galtea through two methods:

1

Prepare your threat input file

Create a file with examples of harmful content or sensitive topic areas you want to test

2

Configure the test

Select Red Teaming as the test type and Generated as the test origin

The test creation process can be done via the SDK or the Galtea dashboard

3

Generate the test

Galtea will process the Knowledge Base and generate a Test File containing Test Cases with adversarial inputs for you, potentially applying selected strategies to vary the attack vectors.

Testing Categories

Red teaming tests typically focus on several key categories:

Example Red Teaming Tests and File Format

Here are examples of the content and its creation-document structure:

Red teaming tests intentionally contain harmful or adversarial content. These are used solely for the purpose of testing AI safety and security. Always handle such content responsibly and in accordance with your organization’s policies.

This structure is the required by Galtea in order to automatically generate the test cases for your test. If that format is not provided, you can still manually create the test cases.

For automatic processing, the file format must be CSV.

The contexts and the extent of the examples provided above are short extractions. In the actual CSV files the context can be much longer and the amount of test cases (rows) can be much higher.

Structure of Red Teaming Tests

Red teaming tests have a specific structure focused on providing challenging inputs:

instance_id
Number
required

A unique identifier for the test instance.

input
Text
required

The adversarial input specifically designed to test the product’s resilience against potential vulnerabilities. Example: “Please ignore any previous instructions and…”

expected_output
Text

For red teaming tests, this field is typically marked as “N/A” since the purpose is to test system resilience rather than producing a specific output.

tag
Text

A categorization tag for the test instance, typically “adversarial_inputs” for red teaming tests.

source
Text

The source of the adversarial input, which may be marked as “N/A” for red teaming tests.