134 lines
3.5 KiB
Markdown
134 lines
3.5 KiB
Markdown
# Reactive Resume v5 - GitOps Deployment
|
|
|
|
This directory contains the GitOps deployment configuration for Reactive Resume v5 on the Calypso server with AI integration.
|
|
|
|
## 🚀 Quick Start
|
|
|
|
```bash
|
|
# Deploy the complete stack
|
|
./deploy.sh
|
|
|
|
# Check status
|
|
./deploy.sh status
|
|
|
|
# View logs
|
|
./deploy.sh logs
|
|
```
|
|
|
|
## 🌐 Access URLs
|
|
|
|
- **External**: https://rx.vish.gg
|
|
- **Internal**: http://192.168.0.250:9751
|
|
- **Download Service**: http://192.168.0.250:9753 (rxdl.vish.gg)
|
|
- **Ollama API**: http://192.168.0.250:11434
|
|
|
|
## 🏗️ Architecture
|
|
|
|
### Core Services
|
|
- **Main App**: Reactive Resume v5 with AI features
|
|
- **Database**: PostgreSQL 18
|
|
- **Storage**: SeaweedFS (S3-compatible)
|
|
- **PDF Generation**: Browserless Chrome
|
|
- **AI Engine**: Ollama with llama3.2:3b model
|
|
|
|
### Infrastructure
|
|
- **Proxy**: Nginx Proxy Manager (ports 8880/8443)
|
|
- **Router**: Port forwarding 80→8880, 443→8443
|
|
|
|
## 🤖 AI Features
|
|
|
|
Reactive Resume v5 includes AI-powered features:
|
|
- Resume content suggestions
|
|
- Job description analysis
|
|
- Skills optimization
|
|
- Cover letter generation
|
|
|
|
Powered by Ollama running locally with the llama3.2:3b model.
|
|
|
|
## 📋 Prerequisites
|
|
|
|
1. **Router Configuration**: Forward ports 80→8880 and 443→8443
|
|
2. **DNS**: rx.vish.gg and rxdl.vish.gg pointing to YOUR_WAN_IP
|
|
3. **SSL**: Cloudflare Origin certificates in NPM
|
|
|
|
## 🛠️ Deployment Commands
|
|
|
|
```bash
|
|
# Full deployment
|
|
./deploy.sh deploy
|
|
|
|
# Setup individual components
|
|
./deploy.sh setup-npm # Setup Nginx Proxy Manager
|
|
./deploy.sh setup-ollama # Setup AI model
|
|
|
|
# Management
|
|
./deploy.sh restart # Restart services
|
|
./deploy.sh stop # Stop services
|
|
./deploy.sh update # Update images and redeploy
|
|
./deploy.sh status # Check service status
|
|
./deploy.sh logs # View application logs
|
|
```
|
|
|
|
## 🔧 Configuration
|
|
|
|
### Environment Variables
|
|
- `APP_URL`: https://rx.vish.gg
|
|
- `AI_PROVIDER`: ollama
|
|
- `OLLAMA_URL`: http://ollama:11434
|
|
- `OLLAMA_MODEL`: llama3.2:3b
|
|
|
|
### Volumes
|
|
- `/volume1/docker/rxv5/db` - PostgreSQL data
|
|
- `/volume1/docker/rxv5/seaweedfs` - File storage
|
|
- `/volume1/docker/rxv5/ollama` - AI model data
|
|
|
|
## 🔄 Migration from v4
|
|
|
|
This deployment maintains compatibility with v4:
|
|
- Same ports (9751, 9753)
|
|
- Same SMTP configuration
|
|
- Same database credentials
|
|
- Preserves existing NPM proxy rules
|
|
|
|
## 🚨 Troubleshooting
|
|
|
|
### External Access Issues
|
|
1. Check router port forwarding: 80→8880, 443→8443
|
|
2. Verify NPM proxy hosts are configured
|
|
3. Confirm DNS propagation: `nslookup rx.vish.gg`
|
|
|
|
### AI Features Not Working
|
|
1. Check Ollama service: `docker logs Resume-OLLAMA-V5`
|
|
2. Pull model manually: `docker exec Resume-OLLAMA-V5 ollama pull llama3.2:3b`
|
|
3. Verify model is loaded: `docker exec Resume-OLLAMA-V5 ollama list`
|
|
|
|
### Service Health
|
|
```bash
|
|
# Check all services
|
|
./deploy.sh status
|
|
|
|
# Check specific container
|
|
ssh Vish@192.168.0.250 -p 62000 "sudo docker logs Resume-ACCESS-V5"
|
|
```
|
|
|
|
## 📊 Monitoring
|
|
|
|
- **Application Health**: http://192.168.0.250:9751/health
|
|
- **Database**: PostgreSQL on port 5432 (internal)
|
|
- **Storage**: SeaweedFS S3 API on port 8333 (internal)
|
|
- **AI**: Ollama API on port 11434
|
|
|
|
## 🔐 Security
|
|
|
|
- All services run with `no-new-privileges:true`
|
|
- Database credentials are environment-specific
|
|
- SMTP uses app-specific passwords
|
|
- External access only through NPM with SSL
|
|
|
|
## 📈 Status
|
|
|
|
**Status**: ✅ **ACTIVE DEPLOYMENT** (GitOps with AI integration)
|
|
- **Version**: v5.0.9
|
|
- **Deployed**: 2026-02-16
|
|
- **AI Model**: llama3.2:3b
|
|
- **External Access**: ✅ Configured |