Skip to content

Assessments

Assessments organize AI red teaming campaigns with DreadAIRT. An assessment is a named container that groups attack runs against an AI system and aggregates their results into analytics, findings, and compliance reports.

An assessment answers: How vulnerable is this AI system to adversarial attacks?

You provide:

  • a target system to probe
  • one or more attack strategies (TAP, GOAT, Crescendo, PAIR, and others)
  • goals describing what the attacks should attempt

Dreadnode executes attack runs through its red teaming workflows and aggregates their telemetry into analytics on demand.

An assessment belongs to a project within a workspace and accumulates results across multiple attack runs over time.

ConceptDefinition
AssessmentA named, project-scoped container for a red teaming campaign
Attack RunA single execution of an attack strategy (e.g., one TAP run with a specific goal)
TrialAn individual attempt within an attack run — one conversation or prompt exchange with the target
ASRAttack Success Rate — the fraction of trials that achieved the stated goal
Risk ScoreA composite metric combining ASR, severity, and attack effectiveness
TransformAn adversarial technique applied to prompts (encoding, persuasion, injection, etc.)
Compliance TagA mapping from attack results to security framework categories (OWASP, MITRE ATLAS, NIST, Google SAIF)

The assessment lifecycle spans the operator workflow and the platform:

  1. Operator workflow — launch attacks against a target and capture structured telemetry
  2. Platform — ingest telemetry and materialize analytics on demand

The platform provides several levels of analysis:

Assessment-level:

  • Aggregated trace statistics (total attacks, trials, ASR, risk scores)
  • Per-attack span breakdowns with success rates and severity
  • Individual trial spans with filtering by attack name, minimum score, and jailbreak status

Project-level:

  • Cross-assessment findings with severity, category, and attack name filtering
  • Executive summary with risk trends, compliance posture, and top vulnerabilities
  • Automated report generation combining findings across all assessments in a project

Compliance mapping: Results are tagged against industry security frameworks:

  • OWASP Top 10 for LLM Applications
  • MITRE ATLAS
  • NIST AI Risk Management Framework
  • Google Secure AI Framework (SAIF)

Reports are generated from assessment or project data and persisted for later retrieval. A report captures a point-in-time snapshot of findings, risk scores, and compliance posture. Reports can be generated at both the individual assessment level and the project level (consolidating across all assessments).