Files
html/api/ia-cascade-mechanics.json

652 lines
16 KiB
JSON

{
"generated_at": "2026-04-24T16:08:26.670867",
"version": "v1.0-cascade-mechanics",
"doctrine": "Doctrine 314 multi-server + Doctrine 186 WEVIA WEB IA autonomy + Doctrine 188 Blade MCP",
"summary": {
"total_claude_instances": 7,
"total_api_providers_free": 18,
"total_ollama_models": 7,
"total_gpu_free": 5,
"total_cdp_profiles": 8,
"blade_mcp_tools": 17,
"total_intents": 216,
"total_servers": 7,
"cascade_steps": 12,
"grand_total_agents": 52,
"monthly_cost": "0 EUR (cascade free + subscriptions existants)"
},
"layers": {
"1_claude_instances": [
{
"id": "opus",
"name": "Opus (claude.ai web)",
"type": "human-orchestrator",
"role": "Strategic architect, root cause",
"capability": "Long reasoning, multi-session memory via chat",
"access": "claude.ai (Yacine session)",
"cost": "Subscription",
"status": "retiring (doctrine #60)",
"color": "gold"
},
{
"id": "ambre",
"name": "Ambre",
"type": "human-agent",
"role": "File generation priority + cascade doctrine 167 UX",
"capability": "AMBRE 2026-04-21 file creation, UX doctrine cascade",
"access": "claude.ai (separate session)",
"cost": "Subscription",
"status": "active",
"color": "coral"
},
{
"id": "yanis",
"name": "Yanis (claude code)",
"type": "human-agent",
"role": "Commit push Gitea + GitHub, code reviewer",
"capability": "Git operations, code refactor, auto-sync",
"access": "Claude Code CLI + Gitea (yanis user)",
"cost": "Subscription",
"status": "active",
"color": "cyan"
},
{
"id": "yacine",
"name": "Yacine (owner)",
"type": "human-owner",
"role": "Product direction, decisions, NL chat driver",
"capability": "Strategic decisions, UX validation, vision",
"access": "All browsers + WEVIA chat + cockpit",
"cost": "0 (owner)",
"status": "active",
"color": "mint"
},
{
"id": "claude_code",
"name": "Claude Code (CLI)",
"type": "cli-agent",
"role": "File manipulation, code editing",
"capability": "Direct file I/O, shell exec, git",
"access": "Terminal local",
"cost": "Subscription",
"status": "active",
"color": "violet"
},
{
"id": "claude_2",
"name": "Claude 2 (UX/écrans/data)",
"type": "archived-session",
"role": "UX fixes 8 écrans + DB remapping + drill-down APIs",
"capability": "Historique: offer-engine, scout, lookalike, data-manager, scrapping-factory",
"access": "Session antérieure (17 fév 2026)",
"cost": "Subscription",
"status": "work integrated",
"color": "cyan"
},
{
"id": "claude_3",
"name": "Claude 3 (Account Creator/IA)",
"type": "archived-session",
"role": "Multi-provider accounts + IA providers tests live",
"capability": "16 cloud + 10 IA, O365 Graph API, FreeDNS, Cerebras/Groq/SambaNova",
"access": "Session antérieure (17 fév 2026)",
"cost": "Subscription",
"status": "work integrated",
"color": "violet"
}
],
"2_sovereign": [
{
"id": "sovereign_api",
"name": "Sovereign-API v3",
"port": 4000,
"type": "cascade_router",
"providers_active": 18,
"status": "LIVE",
"http": 200
}
],
"2b_api_providers": [
{
"id": "cerebras_fast",
"name": "Cerebras Fast",
"type": "api-free",
"model": "llama-3.3-70b",
"latency_ms": 429,
"cost": 0,
"tokens_quota": "fast",
"status": "LIVE"
},
{
"id": "cerebras_think",
"name": "Cerebras Think",
"type": "api-free",
"model": "reasoning",
"latency_ms": 800,
"cost": 0,
"tokens_quota": "slow",
"status": "LIVE"
},
{
"id": "groq",
"name": "Groq",
"type": "api-free",
"model": "llama-3.3-70b",
"latency_ms": 192,
"cost": 0,
"tokens_quota": "high",
"status": "LIVE"
},
{
"id": "cloudflare_ai",
"name": "Cloudflare AI",
"type": "api-free",
"model": "llama-3.1-8b",
"latency_ms": 600,
"cost": 0,
"status": "LIVE"
},
{
"id": "gemini",
"name": "Gemini (API)",
"type": "api-free",
"model": "gemini-2.5",
"latency_ms": 500,
"cost": 0,
"status": "LIVE"
},
{
"id": "sambanova",
"name": "SambaNova",
"type": "api-free",
"model": "Meta-Llama-3.3-70B",
"latency_ms": 800,
"cost": 0,
"status": "LIVE"
},
{
"id": "nvidia_nim",
"name": "NVIDIA NIM",
"type": "api-free",
"model": "llama-3.1-405b",
"latency_ms": 1200,
"cost": 0,
"status": "LIVE"
},
{
"id": "mistral",
"name": "Mistral",
"type": "api-free",
"model": "mistral-large",
"latency_ms": 700,
"cost": 0,
"status": "LIVE"
},
{
"id": "groq_oss",
"name": "Groq OSS",
"type": "api-free",
"model": "oss-models",
"latency_ms": 250,
"cost": 0,
"status": "LIVE"
},
{
"id": "hf_space",
"name": "HF Space",
"type": "api-free",
"model": "yace222/weval-brain-v4",
"latency_ms": 1500,
"cost": 0,
"status": "LIVE"
},
{
"id": "hf_router",
"name": "HF Router",
"type": "api-free",
"model": "routed",
"latency_ms": 1000,
"cost": 0,
"status": "LIVE"
},
{
"id": "openrouter",
"name": "OpenRouter",
"type": "api-free",
"model": "routed",
"latency_ms": 800,
"cost": 0,
"status": "LIVE"
},
{
"id": "github_models",
"name": "GitHub Models",
"type": "api-free",
"model": "multi",
"latency_ms": 900,
"cost": 0,
"status": "LIVE"
},
{
"id": "alibaba",
"name": "Alibaba Qwen",
"type": "api-free",
"model": "qwen-max",
"latency_ms": 700,
"cost": 0,
"status": "LIVE"
},
{
"id": "together",
"name": "Together AI",
"type": "api-free",
"model": "multi",
"latency_ms": 800,
"cost": 0,
"status": "LIVE"
},
{
"id": "cohere",
"name": "Cohere",
"type": "api-free",
"model": "command-r",
"latency_ms": 600,
"cost": 0,
"status": "LIVE"
},
{
"id": "zhipu",
"name": "Zhipu GLM",
"type": "api-free",
"model": "glm-4",
"latency_ms": 700,
"cost": 0,
"status": "LIVE"
},
{
"id": "replicate",
"name": "Replicate",
"type": "api-free",
"model": "multi",
"latency_ms": 1200,
"cost": 0,
"status": "LIVE"
}
],
"3_ollama_local": [
{
"id": "qwen2.5:32b",
"name": "Qwen 2.5 32B",
"type": "ollama-local",
"latency_ms": 2000,
"cost": 0,
"status": "LIVE"
},
{
"id": "huihui_llama3.2",
"name": "Huihui Llama 3.2 (abliterated)",
"type": "ollama-local",
"latency_ms": 1200,
"cost": 0,
"status": "LIVE"
},
{
"id": "weval-brain-v4",
"name": "weval-brain-v4 (fine-tuned)",
"type": "ollama-local",
"latency_ms": 1500,
"cost": 0,
"status": "LIVE"
},
{
"id": "llama3.2",
"name": "Llama 3.2",
"type": "ollama-local",
"latency_ms": 1000,
"cost": 0,
"status": "LIVE"
},
{
"id": "nomic-embed",
"name": "Nomic Embed",
"type": "ollama-local",
"latency_ms": 200,
"cost": 0,
"status": "LIVE"
},
{
"id": "qwen3:4b",
"name": "Qwen 3 4B",
"type": "ollama-local",
"latency_ms": 800,
"cost": 0,
"status": "LIVE"
},
{
"id": "all-minilm",
"name": "all-MiniLM",
"type": "ollama-local",
"latency_ms": 150,
"cost": 0,
"status": "LIVE"
}
],
"4_gpu_free": [
{
"id": "hf_zero_gpu",
"name": "HuggingFace Zero GPU",
"type": "gpu-free",
"user": "yace222",
"quota": "80h/week",
"hardware": "A100",
"cost": 0,
"status": "LIVE"
},
{
"id": "kaggle",
"name": "Kaggle Notebooks",
"type": "gpu-free",
"user": "yace222",
"quota": "30h/week",
"hardware": "T4 16GB",
"cost": 0,
"status": "LIVE",
"use_case": "Qwen2.5-3B fine-tune"
},
{
"id": "colab",
"name": "Google Colab",
"type": "gpu-free",
"user": "yace222",
"quota": "12h session",
"hardware": "T4",
"cost": 0,
"status": "LIVE"
},
{
"id": "render",
"name": "Render",
"type": "gpu-free",
"quota": "hobby",
"cost": 0,
"status": "STANDBY"
},
{
"id": "railway",
"name": "Railway",
"type": "gpu-free",
"quota": "free tier",
"cost": 0,
"status": "STANDBY"
}
],
"5_web_cdp": [
{
"id": "openai",
"name": "ChatGPT",
"port": 9222,
"type": "web-cdp",
"status": "RUNNING",
"pid": "2442177",
"cdp_listening": true,
"access_method": "Chrome DevTools Protocol (cookies Yacine)",
"cost": 0
},
{
"id": "anthropic",
"name": "Claude.ai",
"port": 9223,
"type": "web-cdp",
"status": "RUNNING",
"pid": "2407065",
"cdp_listening": true,
"access_method": "Chrome DevTools Protocol (cookies Yacine)",
"cost": 0
},
{
"id": "google",
"name": "Gemini",
"port": 9224,
"type": "web-cdp",
"status": "RUNNING",
"pid": "2440660",
"cdp_listening": true,
"access_method": "Chrome DevTools Protocol (cookies Yacine)",
"cost": 0
},
{
"id": "deepseek",
"name": "DeepSeek",
"port": 9225,
"type": "web-cdp",
"status": "RUNNING",
"pid": "2441158",
"cdp_listening": true,
"access_method": "Chrome DevTools Protocol (cookies Yacine)",
"cost": 0
},
{
"id": "mistral",
"name": "Mistral",
"port": 9226,
"type": "web-cdp",
"status": "RUNNING",
"pid": "2440889",
"cdp_listening": true,
"access_method": "Chrome DevTools Protocol (cookies Yacine)",
"cost": 0
},
{
"id": "poe",
"name": "Poe",
"port": 9227,
"type": "web-cdp",
"status": "RUNNING",
"pid": "2441620",
"cdp_listening": true,
"access_method": "Chrome DevTools Protocol (cookies Yacine)",
"cost": 0
},
{
"id": "perplexity",
"name": "Perplexity",
"port": 9228,
"type": "web-cdp",
"status": "RUNNING",
"pid": "2440818",
"cdp_listening": true,
"access_method": "Chrome DevTools Protocol (cookies Yacine)",
"cost": 0
},
{
"id": "hf",
"name": "HuggingFace",
"port": 9229,
"type": "web-cdp",
"status": "RUNNING",
"pid": "2441895",
"cdp_listening": true,
"access_method": "Chrome DevTools Protocol (cookies Yacine)",
"cost": 0
}
],
"6_blade_mcp": {
"id": "blade_mcp",
"name": "Blade IA (MCP port 8765)",
"type": "blade-mcp",
"tools_exposed": 17,
"endpoint": "http://localhost:8765/mcp/blade",
"token": "wevia_blade_mcp_20avr_k9f3m2x8n5q7p1",
"status": "LIVE",
"providers_supported": [
"chatgpt",
"claude",
"gemini",
"deepseek",
"mistral",
"poe",
"perplexity",
"hf"
],
"pattern": "ask_blade_<provider> → Chrome Yacine session déjà connecté",
"cost": 0
},
"7_routing": {
"nl_priority_intents": 184,
"total_intents": 216,
"dispatchers_ordered": [
"V182 master-public-guard",
"OPUS4-AUTOWIRE-EARLY-v2",
"V103-NATURAL-MULTI-AGENT-ROUTER (parallel/reconcile/bilan)",
"NL-PRIORITY PRE-DISPATCH (184 patterns)",
"OPUS5-STUB-DISPATCHER-v2 (wired-pending/intent-opus4-*.php)",
"OPUS_ROOT_CAUSE_GUARDS (SQL/git/CRM/memory/self-heal)",
"opus_persistent_intents + opus_mega_intents + opus_ux_audit"
]
},
"8_servers": [
{
"id": "s204",
"name": "S204 (MAIN)",
"role": "WEVIA Master + router + aggregator",
"ip": "204.168.152.13",
"cpu": "8vCPU",
"ram": "32GB",
"status": "LIVE"
},
{
"id": "s95",
"name": "S95 (ETHICA)",
"role": "WEVADS prod + Ethica HCPs + scrapers",
"ip": "95.216.167.89",
"cpu": "n/a",
"ram": "n/a",
"status": "LIVE",
"data": "171K HCPs live"
},
{
"id": "s151",
"name": "S151 (TRACKING)",
"role": "OVH tracking open.php",
"ip": "151.80.235.110",
"cpu": "n/a",
"ram": "n/a",
"status": "LIVE"
},
{
"id": "pmta_ser6",
"name": "PMTA SER_6",
"role": "Email sender",
"ip": "110.239.84.121",
"status": "LIVE"
},
{
"id": "pmta_ser7",
"name": "PMTA SER_7",
"role": "Email sender",
"ip": "110.239.65.64",
"status": "LIVE"
},
{
"id": "pmta_ser8",
"name": "PMTA SER_8",
"role": "Email sender",
"ip": "182.160.55.107",
"status": "LIVE"
},
{
"id": "pmta_ser9",
"name": "PMTA SER_9",
"role": "Email sender",
"ip": "110.239.86.68",
"status": "LIVE"
}
]
},
"cascade_flow": [
{
"step": 1,
"name": "Yacine types NL",
"actor": "Yacine",
"where": "WEVIA chat / cockpit",
"what": "Natural language query"
},
{
"step": 2,
"name": "V182 Guard",
"actor": "wevia-master-api.php",
"where": "S204",
"what": "Security guard (no public data leak)"
},
{
"step": 3,
"name": "NL-Priority matcher",
"actor": "184 patterns",
"where": "priority-intents-nl.json",
"what": "Fast regex match → direct shell exec"
},
{
"step": 4,
"name": "OPUS5-Stub dispatcher",
"actor": "216 intents",
"where": "wired-pending/intent-opus4-*.php",
"what": "Longest-match + safe prefix check → shell exec"
},
{
"step": 5,
"name": "Multi-agent decision",
"actor": "WEVIA Master",
"where": "Orchestrator",
"what": "Parallel dispatch if 'parallel|reconcile|bilan'"
},
{
"step": 6,
"name": "Cascade providers",
"actor": "Sovereign-API v3",
"where": "port 4000",
"what": "Try 18 free providers in fallback order"
},
{
"step": 7,
"name": "LLM exec",
"actor": "Cerebras/Groq/Sovereign",
"where": "Free providers",
"what": "Token generation 0€"
},
{
"step": 8,
"name": "Or Web Chrome",
"actor": "ask_blade_<provider>",
"where": "MCP 8765",
"what": "Chrome Yacine session logged-in"
},
{
"step": 9,
"name": "Or CDP local",
"actor": "ask_<provider>_web",
"where": "ports 9222-9229",
"what": "CDP fallback"
},
{
"step": 10,
"name": "Multi-server dispatch",
"actor": "wevia-dispatch.php",
"where": "S204+S95+S151+GPU",
"what": "Doctrine 314 parallel workers"
},
{
"step": 11,
"name": "Synthesis",
"actor": "WEVIA Master",
"where": "S204",
"what": "Aggregate + synthesize results"
},
{
"step": 12,
"name": "Response SSE",
"actor": "Chat UI",
"where": "User browser",
"what": "Stream response NL"
}
]
}