Reactive Resume v5 - GitOps Deployment
This directory contains the GitOps deployment configuration for Reactive Resume v5 on the Calypso server with AI integration.
🚀 Quick Start
# Deploy the complete stack
./deploy.sh
# Check status
./deploy.sh status
# View logs
./deploy.sh logs
🌐 Access URLs
- External: https://rx.vish.gg
- Internal: http://192.168.0.250:9751
- Download Service: http://192.168.0.250:9753 (rxdl.vish.gg)
- Ollama API: http://192.168.0.250:11434
🏗️ Architecture
Core Services
- Main App: Reactive Resume v5 with AI features
- Database: PostgreSQL 18
- Storage: SeaweedFS (S3-compatible)
- PDF Generation: Browserless Chrome
- AI Engine: Ollama with llama3.2:3b model
Infrastructure
- Proxy: Nginx Proxy Manager (ports 8880/8443)
- Router: Port forwarding 80→8880, 443→8443
🤖 AI Features
Reactive Resume v5 includes AI-powered features:
- Resume content suggestions
- Job description analysis
- Skills optimization
- Cover letter generation
Powered by Ollama running locally with the llama3.2:3b model.
📋 Prerequisites
- Router Configuration: Forward ports 80→8880 and 443→8443
- DNS: rx.vish.gg and rxdl.vish.gg pointing to YOUR_WAN_IP
- SSL: Cloudflare Origin certificates in NPM
🛠️ Deployment Commands
# Full deployment
./deploy.sh deploy
# Setup individual components
./deploy.sh setup-npm # Setup Nginx Proxy Manager
./deploy.sh setup-ollama # Setup AI model
# Management
./deploy.sh restart # Restart services
./deploy.sh stop # Stop services
./deploy.sh update # Update images and redeploy
./deploy.sh status # Check service status
./deploy.sh logs # View application logs
🔧 Configuration
Environment Variables
APP_URL: https://rx.vish.ggAI_PROVIDER: ollamaOLLAMA_URL: http://ollama:11434OLLAMA_MODEL: llama3.2:3b
Volumes
/volume1/docker/rxv5/db- PostgreSQL data/volume1/docker/rxv5/seaweedfs- File storage/volume1/docker/rxv5/ollama- AI model data
🔄 Migration from v4
This deployment maintains compatibility with v4:
- Same ports (9751, 9753)
- Same SMTP configuration
- Same database credentials
- Preserves existing NPM proxy rules
🚨 Troubleshooting
External Access Issues
- Check router port forwarding: 80→8880, 443→8443
- Verify NPM proxy hosts are configured
- Confirm DNS propagation:
nslookup rx.vish.gg
AI Features Not Working
- Check Ollama service:
docker logs Resume-OLLAMA-V5 - Pull model manually:
docker exec Resume-OLLAMA-V5 ollama pull llama3.2:3b - Verify model is loaded:
docker exec Resume-OLLAMA-V5 ollama list
Service Health
# Check all services
./deploy.sh status
# Check specific container
ssh Vish@192.168.0.250 -p 62000 "sudo docker logs Resume-ACCESS-V5"
📊 Monitoring
- Application Health: http://192.168.0.250:9751/health
- Database: PostgreSQL on port 5432 (internal)
- Storage: SeaweedFS S3 API on port 8333 (internal)
- AI: Ollama API on port 11434
🔐 Security
- All services run with
no-new-privileges:true - Database credentials are environment-specific
- SMTP uses app-specific passwords
- External access only through NPM with SSL
📈 Status
Status: ✅ ACTIVE DEPLOYMENT (GitOps with AI integration)
- Version: v5.0.9
- Deployed: 2026-02-16
- AI Model: llama3.2:3b
- External Access: ✅ Configured