This file provides guidance to Claude Code (claude.ai/code) when working with code in this repository.
PentestGPT is an AI-powered autonomous penetration testing agent with a terminal user interface (TUI). It uses an agentic pipeline to solve CTF challenges, Hack The Box machines, and authorized security assessments.
Published at USENIX Security 2024: Paper
Stack: Python 3.12+, uv, Docker (Ubuntu 24.04), Textual (TUI), Rich (CLI), Agent SDK
# Development
uv sync # Install dependencies
uv run pentestgpt --target X # Run locally
# Testing
make test # Run all tests
make test-cov # Run tests with coverage
uv run pytest tests/test_controller.py -v # Run single test file
# Code Quality
make lint # Run ruff linter
make format # Format code with ruff
make typecheck # Run mypy type checking
make check # All checks (lint + typecheck)
# Docker Workflow
make install # Build Docker image
make connect # Connect to container (main usage)
make stop # Stop container
make clean-docker # Remove everything including configpentestgpt/interface/main.py- CLI entry, argument parsing, mode selection- Command:
pentestgpt --target <IP/URL> [--instruction "hint"] [--non-interactive] [--raw] [--debug]
- agent.py -
PentestAgent: Wraps the LLM agent, handles flag detection, logs to/workspace/pentestgpt-debug.log - backend.py -
AgentBackendinterface +ClaudeCodeBackendimplementation (framework-agnostic design) - controller.py -
AgentController: 5-state lifecycle (IDLE->RUNNING->PAUSED->COMPLETED->ERROR), pause/resume at message boundaries - events.py -
EventBus: Singleton pub/sub for TUI-agent decoupling (STATE_CHANGED, MESSAGE, TOOL, FLAG_FOUND events) - session.py -
SessionStore: File-based persistence in~/.pentestgpt/sessions/, supports session resumption - config.py - Pydantic settings with
.envfile support
- tui.py - Textual TUI app with real-time activity feed, F1 help, Ctrl+P pause, Ctrl+Q quit
- components/ - ActivityFeed, SplashScreen, tool-specific Renderers
- pentesting.py -
CTF_SYSTEM_PROMPT: CTF methodology, flag formats, persistence directives
- Event-Driven: TUI subscribes to EventBus; agent emits events for state changes, messages, flags
- Singletons:
EventBus.get(),get_global_tracer()for global access - Abstract Backend:
AgentBackendinterface allows swapping LLM backends - Flag Detection: Regex patterns in agent.py match
flag{},HTB{},CTF{}, 32-char hex
Tests use pytest with pytest-asyncio. Mock backends for unit tests.
uv run pytest tests/ -v # All tests
uv run pytest tests/test_controller.py -v # Single file
uv run pytest tests/test_controller.py::test_name # Single test- Non-root user:
pentesterwith sudo - Workdir:
/workspace(mounted from./workspace) - LLM config persisted in
claude-configvolume - Pre-installed: nmap, netcat, curl, wget, git, ripgrep, tmux
The previous multi-LLM version (v0.15) is archived in legacy/. It supports:
- OpenAI (GPT-4o, o3, o4-mini)
- Google Gemini
- Deepseek
- Ollama (local LLMs)
- GPT4All
To develop on the legacy version:
cd legacy
pip install -e .Use the standalone benchmark runner at benchmark/standalone-xbow-benchmark-runner/:
cd benchmark/standalone-xbow-benchmark-runner
python3 run_benchmarks.py --range 1-10 --pattern-flag # Run benchmarks 1-10
python3 run_benchmarks.py --all --pattern-flag # Run all 104 benchmarks
python3 run_benchmarks.py --retry-failed # Retry failed benchmarks
python3 run_benchmarks.py --dry-run --range 1-5 # Preview without executingSee benchmark/standalone-xbow-benchmark-runner/README.md for full documentation.
.
├── pentestgpt/ # Main package (agentic version)
│ ├── core/ # Agent, controller, events, session
│ ├── interface/ # TUI and CLI
│ ├── prompts/ # System prompts
│ ├── benchmark/ # Benchmark runner module
│ └── tools/ # Tool framework
├── benchmark/ # Benchmark suites
│ ├── xbow-validation-benchmarks/ # 104 XBOW benchmarks
│ └── standalone-xbow-benchmark-runner/ # Benchmark runner
├── tests/ # Test suite
├── workspace/ # Runtime workspace (Docker mount)
├── legacy/ # Archived v0.15 (multi-LLM)
├── Dockerfile # Ubuntu 24.04 container
├── docker-compose.yml # Container orchestration
└── Makefile # Development commands
When modifying code, ensure:
- Adherence to existing architecture and patterns
- Comprehensive tests for new features
- Ensure to run tests after changes, and do further updates to ensure code quality. Always keep the documentation up to date with any architectural changes. Also ensure all tests pass after modifications.