Settings Reference
This page documents configurable settings in the extension, organized by settings tab.
For tuning model spend, also see Token Usage & Cost Management.
AI Backend
Preferred Backend
Dropdown
codex-cli
Backend used for new chat sessions. List is filtered to locally available backends.
Codex CLI Command
Text
codex chat
Shell command used to launch Codex CLI.
Gemini CLI Command
Text
gemini --output-format text --model gemini-2.5-flash --yolo
Shell command used to launch Gemini CLI.
Claude CLI Command
Text
claude
Shell command used to launch Claude CLI.
OpenCode CLI Command
Text
opencode
Shell command used to launch OpenCode CLI.
Ollama URL
Text
http://127.0.0.1:11434
Base URL for Ollama API.
Ollama Model
Text
llama3.1
Model identifier sent to Ollama.
Ollama API Key
Text
(empty)
Optional bearer token for Ollama-compatible servers.
Ollama Extra Headers
Multiline
(empty)
Extra headers, one per line (Header: value).
Ollama Auto-Start
Toggle
On
Auto-start ollama serve when backend is selected.
Ollama Serve Command
Text
ollama serve
Command used for Ollama server startup.
Ollama Timeout
Number
120
Timeout in seconds (range: 30-3600).
Ollama Context Window
Number
8192
Context window size (range: 2048-128000).
LM Studio URL
Text
http://127.0.0.1:1234
Base URL for LM Studio API.
LM Studio Model
Text
lmstudio
Model identifier sent to LM Studio.
LM Studio API Key
Text
(empty)
Optional bearer token for OpenAI-compatible servers.
LM Studio Extra Headers
Multiline
(empty)
Extra headers, one per line (Header: value).
LM Studio Auto-Start
Toggle
On
Auto-start LM Studio server command.
LM Studio Server Command
Text
lms server start
Command used for LM Studio startup.
LM Studio Timeout
Number
120
Timeout in seconds (range: 30-3600).
OpenAI-Compatible URL
Text
(empty)
Base URL for OpenAI-compatible providers.
OpenAI-Compatible Model
Text
(empty)
Model identifier sent in requests.
OpenAI-Compatible API Key
Text
(empty)
Optional bearer token.
OpenAI-Compatible Extra Headers
Multiline
(empty)
Extra headers, one per line (Header: value).
OpenAI-Compatible Timeout
Number
120
Timeout in seconds (range: 30-3600).
Auto-Restart
Toggle
On
Automatically restart crashed CLI backends.
Agent Profile
Dropdown
pentester
Active profile from ~/.burp-ai-agent/AGENTS/*.md.
Privacy & Logging
Privacy Mode
Dropdown
OFF
Redaction policy (STRICT, BALANCED, OFF).
Determinism Mode
Toggle
Off
Stabilizes ordering for reproducible prompt bundles.
Host Anonymization Salt
Text
(auto-generated)
Secret used to produce stable host pseudonyms in STRICT mode.
Audit Logging
Toggle
Off
Enables JSONL audit log at ~/.burp-ai-agent/audit.jsonl.
MCP Server
Enable MCP
Toggle
On
Start MCP SSE server when extension loads.
Host
Text
127.0.0.1
Bind address.
Port
Number
9876
TCP port for MCP server.
External Access
Toggle
Off
Allow non-loopback connections. Requires TLS.
STDIO Bridge
Toggle
Off
Enable stdio transport in addition to SSE.
Token
Text
(auto-generated)
Bearer token for external access mode.
TLS Enabled
Toggle
Off
Enable HTTPS for MCP server.
Auto-Generate Certificate
Toggle
On
Generate self-signed PKCS12 keystore automatically.
Keystore Path
Text
(auto-populated)
Custom PKCS12 path for TLS mode.
Keystore Password
Text
(auto-generated)
Password used by custom keystore.
Max Concurrent Requests
Number
4
Max parallel MCP tool calls (range: 1-64).
Max Body Bytes
Number
2097152
Max body bytes returned by MCP tools.
Enable Unsafe Tools
Toggle
Off
Master switch for tools marked unsafe.
Burp Integration
Tool Toggles
Checkboxes
(varies)
Enable/disable MCP tools per category.
Passive AI Scanner
Enabled
Toggle
Off
Enable background passive analysis.
Rate Limit
Number
5
Minimum seconds between analysis requests (range: 1-60).
Scope Only
Toggle
On
Analyze only targets in Burp scope.
Max Size (KB)
Number
96
Maximum response size eligible for passive analysis (range: 16-1024).
Min Severity
Dropdown
LOW
Ignore findings below selected severity.
Endpoint dedup (min)
Number
30
Skip repeated equivalent endpoint analysis inside window (range: 1-240).
Response dedup (min)
Number
30
Skip repeated response-fingerprint analysis inside window (range: 1-240).
Prompt cache TTL (min)
Number
30
Reuse parsed AI results for identical prompts inside window (range: 1-240).
Prompt cache entries
Number
500
Maximum prompt-result cache entries (range: 50-5000).
Endpoint cache entries
Number
5000
Maximum endpoint dedup cache entries (range: 100-50000).
Fingerprint cache entries
Number
5000
Maximum response-fingerprint dedup cache entries (range: 100-50000).
Req body chars (AI)
Number
2000
Maximum request body chars included in passive AI metadata (range: 256-20000).
Resp body chars (AI)
Number
4000
Maximum response body chars included in passive AI metadata (range: 512-40000).
Max headers
Number
40
Maximum filtered headers included in passive AI metadata (range: 5-120).
Max params
Number
15
Maximum parameters included in passive AI metadata (range: 5-100).
Req body chars (manual)
Number
4000
Maximum request body chars included by manual context actions (range: 256-40000).
Resp body chars (manual)
Number
8000
Maximum response body chars included by manual context actions (range: 512-80000).
Manual context JSON
Toggle
On (compact)
Use compact JSON serialization for context-menu actions.
Active AI Scanner
Enabled
Toggle
Off
Enable active vulnerability testing.
Max Concurrent Scans
Number
3
Parallel active scans (range: 1-10).
Max Payloads per Point
Number
10
Payload variants per injection point (range: 1-50).
Timeout
Number
30
Timeout in seconds per scan request (range: 5-120).
Request Delay
Number
100
Delay between requests in ms (range: 0-5000).
Max Risk Level
Dropdown
SAFE
Maximum payload risk (SAFE, MODERATE, DANGEROUS).
Scope Only
Toggle
On
Scan only in-scope targets.
Scan Mode
Dropdown
FULL
Vulnerability-class strategy (BUG_BOUNTY, PENTEST, FULL).
Auto-Queue from Passive
Toggle
On
Queue high-confidence passive findings into active scanner.
Use Collaborator (OAST)
Toggle
Off
Enable out-of-band tests using Burp Collaborator.
Prompt Templates
Built-In Templates
Find Vulnerabilities
Text area
Prompt for Find vulnerabilities request action.
Analyze this request
Text area
Prompt for compact endpoint summary.
Explain JS
Text area
Prompt for JavaScript analysis.
Access Control
Text area
Prompt for authorization test planning.
Login Sequence
Text area
Prompt for login flow extraction.
Analyze this Issue
Text area
Prompt for scanner issue analysis.
Generate PoC & Validate
Text area
Prompt for proof-of-concept generation.
Impact & Severity
Text area
Prompt for impact and severity assessment.
Full Report
Text area
Prompt for complete report generation.
BountyPrompt Integration
Enable BountyPrompt actions
Toggle
Off
Enables curated BountyPrompt submenu actions in request/response context menus.
Prompt directory
Text
~/Tools/BountyPrompt/prompts
Directory containing BountyPrompt JSON files.
Auto-create issues
Toggle
On
Auto-create Burp issues for parsed findings from outputType = issue prompts.
Issue confidence threshold
Number
90
Minimum parsed confidence (0-100) required for issue creation.
Enabled prompt IDs
Multi-line text
Curated IDs
Allowlist of prompt IDs, comma- or newline-separated.
Default curated IDs:
API_Keys_Exposure_DetectionCSRF_Vulnerability_AssessmentSecurity_Headers_AnalysisVulnerable_Software_DetectionExtract_EndpointsVulnerable_File_Upload_Endpoint_DetectionWeb_Attack_SuggestionsSensitive_Error_Messages_Detection
File Locations
The extension stores runtime files under ~/.burp-ai-agent/.
~/.burp-ai-agent/audit.jsonl
JSONL audit entries with SHA-256 chaining.
~/.burp-ai-agent/bundles/
Prompt bundle snapshots for deterministic workflows.
~/.burp-ai-agent/contexts/
Context snapshot JSON files indexed by hash.
~/.burp-ai-agent/certs/
Auto-generated TLS artifacts for MCP server.
~/.burp-ai-agent/backends/
Drop-in backend JAR discovery directory.
~/.burp-ai-agent/AGENTS/
Agent profile markdown files.
~/.burp-ai-agent/AGENTS/default
Active profile marker file.
Runtime Constants
Some limits are operational constants rather than direct UI controls:
Active scanner queue max:
2000targets.Active dedup window:
1 hour.Passive analysis wait timeout:
90s.HTTP backend conversation history cap:
20messages and40000total characters (minimum 2 latest messages retained).CLI history cap:
10messages or20000characters.Large prompt threshold for Claude CLI fallback path:
32000characters.
Last updated
