Create self-hosted-ai-stack-on-tx1 task

Task 9 migration.

Date: 2026-02-16
This commit is contained in:
2026-02-16 06:57:55 -06:00
parent 6e0375d9c3
commit 71059ef2ca

View File

@@ -0,0 +1,44 @@
# Self-Hosted AI Stack on TX1
**Status:** Blocked - Medical clearance
**Priority:** Tier 2 - Major Infrastructure
**Time:** 8-12 hours (3-4 active, rest downloads)
**Location:** TX1 Dallas
**Last Updated:** 2026-02-16
## Overview
Dual AI deployment: AnythingLLM (Michael/Meg, document-heavy) + Open WebUI (staff assistant). DERP backup, unlimited AI access, staff foundation.
## Architecture
**Primary: AnythingLLM** (ai.firefrostgaming.com)
- 1,000+ document libraries
- LanceDB vector database
- Workspace isolation (Operations, Pokerole, Brainstorming)
**Secondary: Open WebUI** (staff-ai.firefrostgaming.com)
- Lighter for staff wiki
- Chroma vector DB
- ChatGPT-like interface
## Phases
**Phase 1:** Deploy stack (1-2 hours)
**Phase 2:** Load models (6-8 hours overnight)
**Phase 3:** Document ingestion (2-3 hours active, 6-8 total)
## Models
- Qwen 2.5 Coder 72B (~40GB)
- Llama 3.3 70B (~40GB)
- Llama 3.2 Vision 11B (~7GB)
- Embeddings: all-MiniLM-L6-v2 (~400MB)
**Total:** ~150GB storage, ~110GB RAM when loaded
## Success Criteria
- ✅ Both stacks deployed
- ✅ Models loaded and operational
- ✅ Documents ingested (Ops, Pokerole, Brainstorming)
- ✅ DERP backup functional
**See:** deployment-plan.md for detailed phases
**Fire + Frost + Foundation** 💙🔥❄️