AI Agent System Behavior Tester
Generates a comprehensive test suite of adversarial, edge-case, and functional prompts to stress-test any AI agent or chatbot for reliability, safety, and performance.
Content
You are a senior AI Quality Assurance engineer specializing in LLM and agent evaluation in 2026. Create a comprehensive behavior test suite for this AI agent: **Agent Name/Purpose:** {{agent_name}} **Agent Role:** {{agent_role}} **Primary Users:** {{primary_users}} **Key Capabilities to Test:** {{capabilities}} **Known Constraints/Rules:** {{constraints}} Generate a structured test suite covering: ## 1. Functional Tests (5 prompts) Core capability verification — tests the agent does what it should. ## 2. Edge Case Tests (5 prompts) Boundary conditions — empty inputs, very long inputs, ambiguous requests, multi-language queries. ## 3. Adversarial Tests (5 prompts) Attempts to jailbreak, manipulate, or confuse the agent (prompt injection, role confusion, conflicting instructions). ## 4. Consistency Tests (3 prompts) Asking the same question different ways to verify consistent answers. ## 5. Refusal Tests (3 prompts) Requests the agent should decline — verify graceful handling. For each test, provide: - The test prompt - Expected behavior - Pass/fail criteria - Severity if failed (Critical/High/Medium/Low) Format as a markdown table for easy use in a QA runbook.
Related Prompts
API Design & Documentation Generator
Design RESTful or GraphQL APIs with OpenAPI specs, error handling, and developer documentation
API Endpoint Designer
Design RESTful API endpoints with proper naming, methods, and documentation
API Documentation Generator
Generate comprehensive API documentation from code or specifications
RAG Knowledge Base Query Optimizer
Optimize queries for retrieval-augmented generation systems