Supported platforms
| Platform | Example URL | Notes |
|---|---|---|
| Ollama | http://localhost:11434/v1 | Local models, no API key needed |
| vLLM | http://localhost:8000/v1 | High-performance local serving |
| OpenRouter | https://openrouter.ai/api/v1 | Multi-provider gateway, requires API key |
| LocalAI | http://localhost:8080/v1 | Local models |
| LM Studio | http://localhost:1234/v1 | Desktop app for local models |
Setup
Option 1: Interactive setup
Option 2: CLI flags
Limit overrides
Local models often have smaller context windows than cloud APIs. Use these flags to prevent prompt truncation:| Flag | Default (cloud) | Suggested for local |
|---|---|---|
--max-prompt-chars | 400,000 | 50,000 - 150,000 |
--max-chunk-functions | 120 | 20 - 40 |
--max-output-tokens | 16,384 | 4,096 - 8,192 |
kong setup so they persist as defaults for your custom endpoint.
Endpoint probing
Kong validates that the endpoint is reachable before starting analysis. If your local server is down, you’ll get an error immediately instead of waiting for Ghidra to finish loading the binary:Cost tracking
Cost tracking is disabled for custom providers since Kong has no pricing data for arbitrary models. Token counts are still recorded and shown in the final stats:Further reading
- LLM Providers — cloud provider setup
- Setup Wizard — interactive configuration
- CLI Reference:
kong analyze— all flags

