AI Configuration
The AI Configuration category controls the AI assistant backend that powers the Ask AI feature and TalonStrike in Hawkra workspaces. You can choose between Google Gemini, Anthropic Claude, or a self-hosted local LLM server.
Settings Reference
| Setting | Key | Type | Default | Description |
|---|---|---|---|---|
| LLM Mode | llm_mode | Dropdown | gemini | gemini uses Google Gemini API. claude uses Anthropic Claude API. local uses a self-hosted LLM server. Legacy value cloud is accepted as an alias for gemini. |
| Gemini API Key | gemini_api_key | String | Empty | Google AI Studio API key. Required when LLM mode is gemini. Get one from Google AI Studio. |
| Gemini Model | gemini_model | Dropdown | gemini-2.0-flash | Gemini model to use. Has no effect when LLM mode is claude or local. |
| Anthropic API Key | anthropic_api_key | String | Empty | Anthropic API key. Required when LLM mode is claude. Get one from Anthropic Console. |
| Anthropic Model | anthropic_model | Dropdown | claude-sonnet-4-6 | Claude model to use. Has no effect when LLM mode is gemini or local. |
| Local LLM Server | local_llm_server | String | Empty | URL of your local LLM server. Required when LLM mode is local (e.g., http://ollama:11434). |
Gemini Model Options
| Model | Description |
|---|---|
| gemini-2.0-flash | Fast responses with good quality. Lowest API cost. Recommended for most use cases. |
| gemini-2.0-pro | Higher quality with deeper reasoning. Good balance for complex security analysis. |
| gemini-2.5-pro | Best quality. Ideal for complex multi-step analysis. |
Claude Model Options
| Model | Description |
|---|---|
| claude-sonnet-4-6 | Fast responses with excellent quality. Best balance of speed and capability. Recommended for most use cases. |
| claude-opus-4-6 | Most capable model with deepest reasoning. Ideal for complex multi-step security analysis. |
| claude-haiku-4-5 | Fastest and most cost-effective. Good for simple queries and high-volume usage. |
Setting Up a Local LLM with Ollama
Ollama is the recommended way to run a local LLM for Hawkra. It provides a simple API server that is compatible with Hawkra's local LLM integration.
Option 1: Ollama on the Same Host
If you want to run Ollama alongside Hawkra on the same server, add it to your Docker Compose configuration:
services:
ollama:
image: ollama/ollama
container_name: hawkra-ollama
volumes:
- ollama_data:/root/.ollama
ports:
- "11434:11434"
restart: unless-stopped
volumes:
ollama_data:
After starting the Ollama container, pull a model:
docker exec hawkra-ollama ollama pull llama3
Then configure Hawkra: set LLM Mode to local and Local LLM Server to http://ollama:11434. If Ollama is on a different Docker network, use the host machine's IP address instead of the container name.
Option 2: Ollama on a Separate Machine
- Install Ollama on the target machine following the instructions at ollama.com/download.
- Pull a model:
ollama pull llama3 - Ensure the Ollama server is accessible from your Hawkra server on port 11434.
- Configure Hawkra: set LLM Mode to
localand Local LLM Server tohttp://<ollama-server-ip>:11434.
Recommended Models
| Model | Size |
|---|---|
llama3 | 8B |
llama3:70b | 70B |
mistral | 7B |
mixtral | 8x7B |
Configuration via Environment Variables
| Setting | Environment Variable |
|---|---|
| LLM Mode | LLM_MODE |
| Gemini API Key | GEMINI_API_KEY |
| Gemini Model | GEMINI_MODEL |
| Anthropic API Key | ANTHROPIC_API_KEY |
| Anthropic Model | ANTHROPIC_MODEL |
| Local LLM Server | LOCAL_LLM_SERVER |