Galtea platform tests the robustness of your product against multiple types of threats
A threat in the context of AI and LLMs is any scenario, input, or technique that could cause the model to behave in an unsafe, insecure, or unintended manner. Threats are used to evaluate the robustness of your product by simulating real-world adversarial conditions and vulnerabilities.
In the SDK, the threats are known as variants
, and the same parameter is used for both Quality Tests and Red Teaming Tests. However, the available options differ between the two types of tests.
Below are the main threat types evaluated by Galtea, with references to industry standards:
Data Leakage: Unintentional exposure of sensitive data through model outputs.
Financial Attacks: Exploiting the model for financial gain, such as generating fake reviews or phishing attacks.
Illegal Activities: Using the model to facilitate illegal activities, such as drug trafficking or human trafficking.
Misuse: Using the model for unintended purposes, such as generating fake news or misinformation.
Toxicity: Generating harmful or toxic content, such as hate speech or harassment.
Custom: Allows the generation of highly specific adversarial tests that target the unique vulnerabilities and edge cases of your AI product. Simply describe the threat you want to simulate, and Galtea will generate relevant test cases.
Evaluating your product against these threats helps ensure: