Files
microdao-daarion/QUICKSTART_PHASE3.md
Apple 744c149300
Some checks failed
Build and Deploy Docs / build-and-deploy (push) Has been cancelled
Add automated session logging system
- Created logs/ structure (sessions, operations, incidents)
- Added session-start/log/end scripts
- Installed Git hooks for auto-logging commits/pushes
- Added shell integration for zsh
- Created CHANGELOG.md
- Documented today's session (2026-01-10)
2026-01-10 04:53:17 -08:00

262 lines
4.7 KiB
Markdown
Raw Permalink Blame History

This file contains ambiguous Unicode characters
This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.
# ⚡ QUICKSTART: Phase 3 — LLM + Memory + Tools
**One-task start for real agent intelligence**
---
## 🎯 What Phase 3 Adds
| Before (Phase 2) | After (Phase 3) |
|------------------|-----------------|
| Mock LLM responses | Real GPT-4/DeepSeek/Local |
| No memory | RAG with vector search |
| No tools | Tool execution (projects, tasks, etc.) |
---
## 🚀 One-Command Start
```bash
# Copy Phase 3 master task
cat docs/tasks/PHASE3_MASTER_TASK.md | pbcopy
# Paste into Cursor AI
# Press Enter
# Wait ~2-3 hours
```
**Cursor will create:**
- ✅ llm-proxy (10 files)
- ✅ memory-orchestrator (9 files)
- ✅ toolcore (8 files)
- ✅ docker-compose updates
- ✅ agent-runtime integration
---
## 🔑 Prerequisites
### 1. OpenAI API Key (or Local LLM)
**Option A: OpenAI**
```bash
export OPENAI_API_KEY="sk-..."
```
**Option B: Local LLM (Ollama)**
```bash
# Install Ollama
curl https://ollama.ai/install.sh | sh
# Pull model
ollama pull qwen2.5:8b
# Run server
ollama serve
```
### 2. Vector Database
**Option A: pgvector (PostgreSQL extension)**
```sql
CREATE EXTENSION IF NOT EXISTS vector;
```
**Option B: Simple stub (Phase 3 OK)**
```
# Memory Orchestrator can work with simple PostgreSQL
# Vector search = stub for Phase 3
```
---
## 📦 After Implementation
### Start Services:
```bash
# If using existing start script
./scripts/start-phase2.sh # Existing services
# Start Phase 3 services
docker-compose -f docker-compose.phase3.yml up -d
```
### Test LLM Proxy:
```bash
curl -X POST http://localhost:7007/internal/llm/proxy \
-H "Content-Type: application/json" \
-d '{
"model": "gpt-4.1-mini",
"messages": [
{"role": "user", "content": "Hello!"}
],
"metadata": {
"agent_id": "agent:sofia",
"microdao_id": "microdao:daarion"
}
}'
# Expected: Real GPT-4 response!
```
### Test Memory:
```bash
# Query
curl -X POST http://localhost:7008/internal/agent-memory/query \
-H "Content-Type: application/json" \
-d '{
"agent_id": "agent:sofia",
"microdao_id": "microdao:daarion",
"query": "What did we discuss about Phase 3?",
"limit": 5
}'
```
### Test Tools:
```bash
# List tools
curl http://localhost:7009/internal/tools
# Call tool
curl -X POST http://localhost:7009/internal/tools/call \
-H "Content-Type: application/json" \
-d '{
"tool_id": "projects.list",
"agent_id": "agent:sofia",
"microdao_id": "microdao:daarion",
"args": {}
}'
```
---
## 🧪 E2E Test
### In Messenger UI:
**User:** "Sofia, що нового в проєкті DAARION?"
**Expected Agent Flow:**
1. ✅ Query memory (past discussions)
2. ✅ Call tool: projects.list
3. ✅ Build prompt with context
4. ✅ Call real LLM (GPT-4)
5. ✅ Post rich reply
**Sofia:** "В проєкті DAARION є кілька оновлень:
- Phase 2 Agent Integration завершено ✅
- Phase 3 LLM Proxy в розробці 🔄
- Додано 3 нові агенти
Хочете детальніше по якомусь пункту?"
---
## 📊 Service Status
```bash
# Check all Phase 3 services
docker ps | grep -E '(llm-proxy|memory-orchestrator|toolcore)'
# Check health
curl http://localhost:7007/health # LLM Proxy
curl http://localhost:7008/health # Memory
curl http://localhost:7009/health # Toolcore
# Check logs
docker logs -f llm-proxy
docker logs -f memory-orchestrator
docker logs -f toolcore
```
---
## 🎯 Success Indicators
After Phase 3:
- ✅ Agent uses real LLM (not keyword mock)
- ✅ Agent remembers conversations
- ✅ Agent can execute tools
- ✅ Responses are intelligent & contextual
- ✅ Latency still < 5s
---
## 🐛 Troubleshooting
### LLM Proxy not working?
```bash
# Check API key
docker logs llm-proxy | grep "OPENAI_API_KEY"
# Test provider directly
curl https://api.openai.com/v1/models \
-H "Authorization: Bearer $OPENAI_API_KEY"
```
### Memory not working?
```bash
# Check PostgreSQL connection
docker logs memory-orchestrator | grep "PostgreSQL"
# Check embeddings
docker logs memory-orchestrator | grep "embedding"
```
### Tools not working?
```bash
# Check registry loaded
curl http://localhost:7009/internal/tools
# Check permissions
docker logs toolcore | grep "allowed_agents"
```
---
## 📚 Documentation
- [PHASE3_MASTER_TASK.md](docs/tasks/PHASE3_MASTER_TASK.md) — Complete spec
- [PHASE3_READY.md](PHASE3_READY.md) — Overview
- [PHASE3_ROADMAP.md](docs/tasks/PHASE3_ROADMAP.md) — Detailed plan
---
## 🔜 Next Steps
After Phase 3 works:
1. Test with multiple agents
2. Add more tools (task.create, followup.create)
3. Tune memory relevance
4. Optimize LLM costs
5. Monitor usage
---
**Time to Start:** Copy PHASE3_MASTER_TASK.md into Cursor! 🚀
**Questions?** Check [PHASE3_READY.md](PHASE3_READY.md) first.