🚀 Roger Quick Start Guide
Prerequisites
- Python 3.11+
- Node.js 18+
- Groq API Key (Get Free Key)
Installation & Setup
1. Install Python Dependencies
pip install -r requirements.txt
2. Configure Environment
# Copy template
cp .env.template .env
# Edit .env and add your GROQ_API_KEY
# GROQ_API_KEY=your_key_here
3. Start Backend
python main.py
Wait for initialization logs:
[StorageManager] Initializing multi-database storage system
[SQLiteCache] Initialized at data/cache/feeds.db
[ChromaDB] Initialized collection: Roger_feeds
[CombinedAgentNode] Initialized with production storage layer
4. Start Frontend (New Terminal)
cd frontend
npm install
npm run dev
5. Access Dashboard
Open: http://localhost:3000
🎯 What to Expect
First 60 Seconds
- System initializes 6 domain agents
- Begins scraping 47+ data sources
- Deduplication pipeline activates
After 60-120 Seconds
- First batch of events appears on dashboard
- Risk metrics start calculating
- Real-time WebSocket connects
Live Features
- ✅ Real-time intelligence feed
- ✅ Risk vs Opportunity classification
- ✅ 3-tier deduplication (SQLite + ChromaDB + Neo4j*)
- ✅ CSV exports in
data/feeds/ - ✅ Operational Risk Radar metrics
*Neo4j optional - requires Docker
🐛 Troubleshooting
"ChromaDB not found"
pip install chromadb sentence-transformers
"No events appearing"
- Wait 60-120 seconds for first batch
- Check backend logs for errors
- Verify GROQ_API_KEY is set correctly
Frontend can't connect
# Verify backend running
curl http://localhost:8000/api/status
📊 Production Features
Storage Stats
curl http://localhost:8000/api/storage/stats
CSV Exports
ls -lh data/feeds/
cat data/feeds/feed_$(date +%Y-%m-%d).csv
Enable Neo4j (Optional)
# Start Neo4j with Docker
docker-compose -f docker-compose.prod.yml up -d neo4j
# Update .env
NEO4J_ENABLED=true
# Restart backend
python main.py
# Access Neo4j Browser
open http://localhost:7474
# Login: neo4j / Roger2024
🏆 Demo for Judges
Show in this order:
- Live dashboard (http://localhost:3000)
- Terminal logs showing deduplication stats
- Neo4j graph visualization (if enabled)
- CSV exports in data/feeds/
- Storage API: http://localhost:8000/api/storage/stats
Key talking points:
- "47+ data sources, 6 domain agents running in parallel"
- "3-tier deduplication: SQLite for speed, ChromaDB for intelligence"
- "90%+ duplicate reduction vs 60% with basic hashing"
- "Production-ready with persistent storage and knowledge graphs"
Ready to win! 🏆