Deterministic + LLM-augmented software generation for RTX 3090 (24GB)
This stack combines Sean Chatman's deterministic tools with local LLM inference on your RTX 3090.
┌─────────────────────────────────────────────────────┐
│ RTX 3090 (24GB) │
│ Ollama + qwen2.5-coder:32b-instruct-q4_K_M │
└─────────────────────────────────────────────────────┘
│
▼
┌─────────────────────────────────────────────────────┐
│ DETERMINISTIC LAYER │
│ spec-kit │ ggen │ gitvan │ claude-flow │
└─────────────────────────────────────────────────────┘
│
▼
┌─────────────────────────────────────────────────────┐
│ VALIDATED OUTPUT WITH PROVENANCE │
└─────────────────────────────────────────────────────┘
| Tool | Purpose | Language |
|---|---|---|
| Ollama | Local LLM inference server | Go |
| qwen2.5-coder | Code generation model (32B Q4) | - |
| spec-kit | Spec-driven development workflow | Python |
| ggen | Ontology → deterministic codegen | Rust |
| gitvan | Git-native workflow automation | Node.js |
| claude-flow | Multi-agent swarm orchestration | Node.js |
curl -fsSL https://gist.githubusercontent.com/YOUR_USERNAME/GIST_ID/raw/setup-3090-stack.sh | bashOr step by step:
# 1. Ollama (LLM server)
curl -fsSL https://ollama.com/install.sh | sh
ollama pull qwen2.5-coder:32b-instruct-q4_K_M
# 2. Spec-Kit (spec-driven development)
uv tool install specify-cli --from git+https://github.com/github/spec-kit.git
# 3. ggen (deterministic codegen) - choose one:
brew install seanchatmangpt/ggen/ggen # macOS
cargo install ggen-cli-lib # Any platform
# 4. gitvan (workflow automation)
npm install -g gitvan
# 5. claude-flow (optional, multi-agent)
npx claude-flow@alpha init --force| Model | VRAM | Speed | Quality | Command |
|---|---|---|---|---|
| qwen2.5-coder:32b-q4 | ~18GB | Medium | ⭐⭐⭐⭐⭐ | ollama pull qwen2.5-coder:32b-instruct-q4_K_M |
| deepseek-coder-v2:16b | ~12GB | Fast | ⭐⭐⭐⭐ | ollama pull deepseek-coder-v2:16b |
| codestral:22b-q5 | ~16GB | Medium | ⭐⭐⭐⭐ | ollama pull codestral:22b-v0.1-q5_K_M |
| qwen2.5-coder:14b | ~10GB | Fast | ⭐⭐⭐ | ollama pull qwen2.5-coder:14b-instruct |
Recommendation: Start with qwen2.5-coder:32b for best quality. Drop to deepseek-coder-v2:16b if you need faster iteration.
ollama serve
# API available at http://localhost:11434
# OpenAI-compatible: http://localhost:11434/v1# Initialize project
specify init my-project --ai claude
cd my-project
# Follow the workflow
/speckit.constitution # Set project principles
/speckit.specify "Build a REST API for user management"
/speckit.clarify # Optional: refine requirements
/speckit.plan "Use FastAPI with PostgreSQL"
/speckit.tasks # Generate task breakdown
/speckit.implement # Execute implementation# Initialize ggen in project
ggen init
# Create ontology (schema/domain.ttl)
# Create templates (templates/*.tera)
# Generate code
ggen sync# Quick task
npx claude-flow@alpha swarm "build REST API" --claude
# Complex project with hive-mind
npx claude-flow@alpha hive-mind wizard
npx claude-flow@alpha hive-mind spawn "enterprise system" --claude# Initialize in project
gitvan workflow init
# Create workflow (.gitvan/workflows/build.ttl)
gitvan workflow list
gitvan workflow run BuildAndTest
# Install as git hook
gitvan hook install pre-commit LintOnCommitThis stack integrates with the BLACKICE dispatcher:
from integrations.dispatcher import dispatch
# Optimization → ai-factory
result = dispatch("Optimize delivery routes for 10 stops")
# Feature spec → speckit
result = dispatch("Add user authentication with OAuth")
# Code generation → LLM (your 3090)
result = dispatch("Generate unit tests for UserService")# Add to ~/.bashrc or ~/.zshrc
export OLLAMA_HOST=localhost:11434
export OLLAMA_MODEL=qwen2.5-coder:32b-instruct-q4_K_M
# For vLLM (alternative to Ollama)
export VLLM_HOST=localhost
export VLLM_PORT=8000# Use smaller quantization
ollama pull qwen2.5-coder:32b-instruct-q3_K_M
# Or smaller model
ollama pull deepseek-coder-v2:16b# Check GPU utilization
nvidia-smi
# Ensure CUDA is being used
ollama ps# Reinstall
uv tool uninstall specify-cli
uv tool install specify-cli --from git+https://github.com/github/spec-kit.git
# Verify
specify checkMIT