Files
homelab-optimized/hosts/synology/calypso/reactive_resume_v5
Gitea Mirror Bot a63c410ce3
Some checks failed
Documentation / Deploy to GitHub Pages (push) Has been cancelled
Documentation / Build Docusaurus (push) Has been cancelled
Sanitized mirror from private repository - 2026-04-18 12:12:12 UTC
2026-04-18 12:12:12 +00:00
..

Reactive Resume v5 - GitOps Deployment

This directory contains the GitOps deployment configuration for Reactive Resume v5 on the Calypso server with AI integration.

🚀 Quick Start

# Deploy the complete stack
./deploy.sh

# Check status
./deploy.sh status

# View logs
./deploy.sh logs

🌐 Access URLs

🏗️ Architecture

Core Services

  • Main App: Reactive Resume v5 with AI features
  • Database: PostgreSQL 18
  • Storage: SeaweedFS (S3-compatible)
  • PDF Generation: Browserless Chrome
  • AI Engine: Ollama with llama3.2:3b model

Infrastructure

  • Proxy: Nginx Proxy Manager (ports 8880/8443)
  • Router: Port forwarding 80→8880, 443→8443

🤖 AI Features

Reactive Resume v5 includes AI-powered features:

  • Resume content suggestions
  • Job description analysis
  • Skills optimization
  • Cover letter generation

Powered by Ollama running locally with the llama3.2:3b model.

📋 Prerequisites

  1. Router Configuration: Forward ports 80→8880 and 443→8443
  2. DNS: rx.vish.gg and rxdl.vish.gg pointing to YOUR_WAN_IP
  3. SSL: Cloudflare Origin certificates in NPM

🛠️ Deployment Commands

# Full deployment
./deploy.sh deploy

# Setup individual components
./deploy.sh setup-npm      # Setup Nginx Proxy Manager
./deploy.sh setup-ollama   # Setup AI model

# Management
./deploy.sh restart        # Restart services
./deploy.sh stop          # Stop services
./deploy.sh update        # Update images and redeploy
./deploy.sh status        # Check service status
./deploy.sh logs          # View application logs

🔧 Configuration

Environment Variables

Volumes

  • /volume1/docker/rxv5/db - PostgreSQL data
  • /volume1/docker/rxv5/seaweedfs - File storage
  • /volume1/docker/rxv5/ollama - AI model data

🔄 Migration from v4

This deployment maintains compatibility with v4:

  • Same ports (9751, 9753)
  • Same SMTP configuration
  • Same database credentials
  • Preserves existing NPM proxy rules

🚨 Troubleshooting

External Access Issues

  1. Check router port forwarding: 80→8880, 443→8443
  2. Verify NPM proxy hosts are configured
  3. Confirm DNS propagation: nslookup rx.vish.gg

AI Features Not Working

  1. Check Ollama service: docker logs Resume-OLLAMA-V5
  2. Pull model manually: docker exec Resume-OLLAMA-V5 ollama pull llama3.2:3b
  3. Verify model is loaded: docker exec Resume-OLLAMA-V5 ollama list

Service Health

# Check all services
./deploy.sh status

# Check specific container
ssh Vish@192.168.0.250 -p 62000 "sudo docker logs Resume-ACCESS-V5"

📊 Monitoring

  • Application Health: http://192.168.0.250:9751/health
  • Database: PostgreSQL on port 5432 (internal)
  • Storage: SeaweedFS S3 API on port 8333 (internal)
  • AI: Ollama API on port 11434

🔐 Security

  • All services run with no-new-privileges:true
  • Database credentials are environment-specific
  • SMTP uses app-specific passwords
  • External access only through NPM with SSL

📈 Status

Status: ACTIVE DEPLOYMENT (GitOps with AI integration)

  • Version: v5.0.9
  • Deployed: 2026-02-16
  • AI Model: llama3.2:3b
  • External Access: Configured