- Shell 59.4%
- Python 39.7%
- Dockerfile 0.9%
| .forgejo/workflows | ||
| .gitea/issue_template | ||
| docs | ||
| infrastructure | ||
| src/openproject-mcp | ||
| test | ||
| .gitignore | ||
| .gitleaks.toml | ||
| CLAUDE.md | ||
| LICENSE | ||
| PROJECT-STATE.md | ||
| README.md | ||
Generate One 🛸
AI infrastructure that answers to you. Self-hosted, sovereign, and GPU-accelerated on a single RTX 5090 server.
✨ Overview
Generate One is a full-stack AI platform running on a single dedicated server (RTX 5090 GPU, 32 GB VRAM). It provides LLM routing across 8 model tiers, graph-based persistent memory, vector search with 4-stream Reciprocal Rank Fusion, and 195+ tools exposed via the Model Context Protocol (MCP). All services are orchestrated through Coolify and routed via Traefik with Let's Encrypt TLS.
This is the monorepo root — it holds the infrastructure manifest, version tracking, CI/CD workflows, and cross-stack documentation. Individual service stacks live in their own repos.
🏗️ Architecture
┌──────────────────────────────────────────────────────────┐
│ Clients │
│ Claude Code · Chat UI │
└──────────────────────┬───────────────────────────────────┘
│
┌────────────▼─────────────┐
│ Traefik + Let's Encrypt │
│ (reverse proxy / TLS) │
└────┬──────────┬──────────┘
│ │ │
┌───────────▼──┐ ┌────▼─────┐ ┌▼────────────┐
│ g1-api │ │ g1-gpt │ │ g1-mcp │
│ Fusio │ │LibreChat │ │ MetaMCP │
│ REST API │ │ Chat UI │ │ 195+ tools │
└──────────────┘ └──────────┘ └─────────────┘
│
┌───────────────────▼──────────────────────────────┐
│ g1-llm │
│ LiteLLM (8 tiers) · vLLM (GPU) · Qdrant │
└──────────────────────┬───────────────────────────┘
│
┌───────────────────────▼──────────────────────────┐
│ g1-brain │
│ Neo4j · Graphiti · knowledge-mcp │
│ FalkorDB · reasoning-tools │
└──────────────────────────────────────────────────┘
📦 Stacks
| Stack | Domain | Purpose |
|---|---|---|
| g1-core | — | Shared PostgreSQL 16 + Valkey 9.0.1 |
| g1-llm | llm.generate.one | LiteLLM proxy, 3× vLLM (GPU), Qdrant |
| g1-brain | memory.generate.one | Neo4j, Graphiti, knowledge-mcp (4-stream RRF) |
| g1-mcp | mcp.generate.one | MetaMCP gateway, 9 namespaces, ~195 tools |
| g1-frontend | app.generate.one | Next.js apps: web, app, admin |
| g1-agent-backend | agent.generate.one | PydanticAI streaming agent API |
| g1-web | — | SearXNG metasearch, Crawl4AI scraping |
| g1-gpt | gpt.generate.one | LibreChat v0.8.2, MongoDB |
| g1-state | cms.generate.one | Directus CMS, Infisical secrets |
| g1-auth | auth.generate.one | Authentik SSO (12 OIDC providers) |
| g1-observe | observe.generate.one | Langfuse v3 LLM observability |
| g1-automation | n8n.generate.one | n8n workflow automation |
| g1-git | git.generate.one | Forgejo v14, Actions runner |
⚡ MCP Namespaces
| Namespace | Tools | Backend |
|---|---|---|
| g1-brain | 43 | graphiti-mcp, knowledge-mcp, reasoning-tools, logic-lm |
| g1-code | 112 | gitea-mcp, section-mcp |
| g1-web | ~15 | searxng-mcp, crawl4ai |
| g1-math | ~15 | sympy-mcp |
| g1-time | 5 | time-mcp |
| g1-presenter | ~5 | presenter-mcp |
| g1-project | 111 | worklenz-mcp |
🔧 Model Tiers
| Tier | Primary Model | Provider |
|---|---|---|
g1-llm |
K2-instruct | Groq → OpenRouter |
g1-llm-turbo |
GPT-OSS-120B | Groq → Cerebras → OpenRouter |
g1-llm-mini |
Qwen3-235B | Cerebras → OpenRouter |
g1-llm-code |
GLM-4.7 | Cerebras → OpenRouter |
g1-llm-pro |
K2.5 | OpenRouter |
g1-llm-micro |
Qwen3-8B-FP8 | Local vLLM → Cerebras |
g1-vlm |
Qwen3-VL-30B | OpenRouter |
🖥️ Infrastructure
| Component | Value |
|---|---|
| Server | au2404 / 161.184.162.156 |
| GPU | RTX 5090, 32 GB VRAM |
| GPU Budget | 90% VRAM (embedding 30% + reranker 15% + small model 45%) |
| Orchestration | Coolify (cloud-hosted) |
| Reverse Proxy | Traefik v3 + Let's Encrypt |
| Monitoring | Gatus (uptime) · Beszel (resources) · Langfuse (LLM quality) |
🔖 Version Manifest
Service versions are tracked in infrastructure/versions.yaml and validated weekly by the version-check CI workflow. Key pinned versions:
| Service | Version |
|---|---|
| PostgreSQL | 16 |
| Valkey | 9.0.1 |
| LiteLLM | 1.81.12 |
| Qdrant | v1.17.0 |
| vLLM | v0.16.0 |
| Neo4j | 5.26-community |
| MetaMCP | 2.4.22 |
| Forgejo | 14.0.2 |
| Langfuse | v3.158.0 |
| Authentik | 2026.2.0 |
📚 Documentation
Project documentation lives in Directus CMS — the canonical source of truth for skills, dispatches, decisions, and reports. Each stack repo contains its own compose files and README.
🛡️ Part of Generate One
Generate One — AI infrastructure that answers to you.
Self-hosted, sovereign AI platform. generate.one
Licensed under AGPL-3.0.