Files
homelab-optimized/docs/services/individual/tdarr.md
Gitea Mirror Bot cd1988a21b
Some checks failed
Documentation / Deploy to GitHub Pages (push) Has been cancelled
Documentation / Build Docusaurus (push) Has been cancelled
Sanitized mirror from private repository - 2026-04-05 09:28:04 UTC
2026-04-05 09:28:04 +00:00

502 lines
20 KiB
Markdown
Raw Blame History

This file contains ambiguous Unicode characters
This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.
# Tdarr
**🟢 Media Service**
## 📋 Service Overview
| Property | Value |
|----------|-------|
| **Service Name** | tdarr |
| **Host** | Atlantis (Synology) - Server |
| **Category** | Media |
| **Difficulty** | 🟡 |
| **Docker Image** | `ghcr.io/haveagitgat/tdarr:latest` |
| **Compose File** | `hosts/synology/atlantis/arr-suite/docker-compose.yml` |
| **Directory** | `hosts/synology/atlantis/arr-suite` |
## 🎯 Purpose
Tdarr is a distributed transcoding system for automating media library optimization. It can automatically convert your media files to preferred codecs (like H.265/HEVC), remove unwanted audio tracks, and optimize file sizes while maintaining quality.
## 🖥️ Multi-Node Architecture
The Tdarr setup uses distributed worker nodes for parallel transcoding:
| Node | Host | Type | Hardware | Workers |
|------|------|------|----------|---------|
| **TdarrInternalNode** | Atlantis (Synology DS1621+) | CPU | AMD Ryzen | 1 transcode, 2 healthcheck |
| **NUC-QSV** | Proxmox LXC 103 | GPU | Intel QSV | 1 GPU transcode, 1 healthcheck |
| **Calypso-CPU** | Calypso (Synology DS723+) | CPU | AMD Ryzen R1600 | 2 transcode |
| **Guava-VAAPI** | Guava (TrueNAS Scale) | GPU | AMD Radeon 760M (VAAPI, hevc only) | 1 GPU transcode, 1 healthcheck |
### Node Configuration Files
- **Server**: `hosts/synology/atlantis/arr-suite/docker-compose.yml`
- **NUC-QSV**: `hosts/proxmox/lxc/tdarr-node/docker-compose.yaml`
- **Calypso-CPU**: `hosts/synology/calypso/tdarr-node/docker-compose.yaml`
- **Guava-VAAPI**: `hosts/truenas/guava/tdarr-node/docker-compose.yaml`
### Cache Path Configuration (Critical!)
All nodes **must** mount both `/temp` and `/cache` to the same cache directory to avoid path mismatch errors:
```yaml
volumes:
- /path/to/cache:/temp
- /path/to/cache:/cache # Both must point to same location!
```
## 🚀 Quick Start
### Prerequisites
- Docker and Docker Compose installed
- Media library accessible
- Fast storage for transcoding cache (NVMe recommended)
- Access to the host system (Atlantis)
### Deployment
```bash
# Navigate to service directory
cd hosts/synology/atlantis/arr-suite
# Start the service
docker-compose -f docker-compose.yml up -d tdarr
# Check service status
docker-compose -f docker-compose.yml ps
# View logs
docker-compose -f docker-compose.yml logs -f tdarr
```
## 🔧 Configuration
### Docker Compose Configuration
```yaml
tdarr:
image: ghcr.io/haveagitgat/tdarr:latest
container_name: tdarr
environment:
- PUID=1029
- PGID=100
- TZ=America/Los_Angeles
- UMASK=022
- serverIP=0.0.0.0
- serverPort=8266
- webUIPort=8265
- internalNode=true
- inContainer=true
- ffmpegVersion=6
- nodeName=TdarrInternalNode
volumes:
- /volume2/metadata/docker2/tdarr/server:/app/server
- /volume2/metadata/docker2/tdarr/configs:/app/configs
- /volume2/metadata/docker2/tdarr/logs:/app/logs
- /volume1/data/media:/media
- /volume3/usenet/tdarr_cache:/temp
ports:
- "8265:8265"
- "8266:8266"
networks:
media2_net:
ipv4_address: 172.24.0.15
security_opt:
- no-new-privileges:true
restart: always
```
### Environment Variables
| Variable | Value | Description |
|----------|-------|-------------|
| `PUID` | `1029` | User ID for file permissions |
| `PGID` | `100` | Group ID for file permissions |
| `TZ` | `America/Los_Angeles` | Timezone setting |
| `UMASK` | `022` | File permission mask |
| `serverIP` | `0.0.0.0` | Server bind address |
| `serverPort` | `8266` | Server communication port |
| `webUIPort` | `8265` | Web UI port |
| `internalNode` | `true` | Enable built-in transcoding node |
| `inContainer` | `true` | Running in container mode |
| `ffmpegVersion` | `6` | FFmpeg version to use |
| `nodeName` | `TdarrInternalNode` | Name for the internal node |
### Port Mappings
| Host Port | Container Port | Protocol | Purpose |
|-----------|----------------|----------|----------|
| 8265 | 8265 | TCP | Web UI |
| 8266 | 8266 | TCP | Server communication |
### Volume Mappings
| Host Path | Container Path | Type | Purpose |
|-----------|----------------|------|----------|
| `/volume2/metadata/docker2/tdarr/server` | `/app/server` | bind | Server data |
| `/volume2/metadata/docker2/tdarr/configs` | `/app/configs` | bind | Configuration files |
| `/volume2/metadata/docker2/tdarr/logs` | `/app/logs` | bind | Log files |
| `/volume1/data/media` | `/media` | bind | Media library |
| `/volume3/usenet/tdarr_cache` | `/temp` | bind | Transcode cache (NVMe) |
## 🌐 Access Information
| Interface | URL |
|-----------|-----|
| Web UI | `http://192.168.0.200:8265` |
| Server | `http://192.168.0.200:8266` |
## 🔒 Security Considerations
- ✅ Security options configured (no-new-privileges)
- ✅ Running with specific user/group IDs
- ⚠️ Ensure media permissions are correctly set
## 📊 Resource Requirements
### Recommended Resources
- **Minimum RAM**: 4GB (8GB+ recommended for transcoding)
- **Recommended RAM**: 8GB+
- **CPU**: 4+ cores (transcoding is CPU-intensive)
- **GPU**: Optional but highly recommended for hardware transcoding
- **Storage**: Fast NVMe for cache (improves transcode speed significantly)
### Resource Monitoring
```bash
docker stats tdarr
```
## 🔍 Health Monitoring
### Manual Health Checks
Each host requires a different method to access Docker. Synology NAS systems have Docker at `/usr/local/bin/docker` and require `sudo`:
```bash
# Check tdarr server (Atlantis - Synology)
ssh atlantis "sudo /usr/local/bin/docker ps --filter name=tdarr --format 'table {{.Names}}\t{{.Status}}\t{{.RunningFor}}'"
# Check node on Calypso (Synology)
ssh calypso "sudo /usr/local/bin/docker ps --filter name=tdarr --format 'table {{.Names}}\t{{.Status}}\t{{.RunningFor}}'"
# Check node on Guava (TrueNAS Scale) — user is in docker group, no sudo needed
ssh guava "docker ps --filter name=tdarr --format 'table {{.Names}}\t{{.Status}}\t{{.RunningFor}}'"
# Check node on NUC-QSV — runs inside Proxmox LXC 103, not on the NUC host directly
ssh pve "pct exec 103 -- docker ps --filter name=tdarr --format 'table {{.Names}}\t{{.Status}}\t{{.RunningFor}}'"
# Check web UI is responding
curl -s http://192.168.0.200:8265/api/v2/status
```
### Expected Container Names and Uptime
| Host | Container Name | SSH Alias | Notes |
|------|---------------|-----------|-------|
| Atlantis | `tdarr` | `atlantis` | Server + TdarrInternalNode |
| Calypso | `tdarr-node-calypso` | `calypso` | CPU-only node |
| Guava | `tdarr-node-guava` | `guava` | VAAPI GPU node |
| NUC (Proxmox LXC 103) | `tdarr-node` | `pve` + `pct exec 103` | Intel QSV GPU node |
### Portainer Access
- URL: `https://192.168.0.200:9443`
- Endpoint IDs: Atlantis=2, vish-concord-nuc=443395, Calypso=443397
- Note: Guava (TrueNAS) and NUC LXC 103 are not managed by Portainer
- NUC's Portainer endpoint (443395) shows the NUC host Docker — the tdarr-node runs in Proxmox LXC 103 with its own Docker socket, which is not visible there
### Node Worker Configuration
| Node | HC-CPU | TC-CPU | TC-GPU | Notes |
|------|--------|--------|--------|-------|
| TdarrInternalNode | 1 | 0 | 0 | Health checks only |
| NUC-QSV | 1 | 0 | 1 | Intel QSV GPU transcoding |
| Calypso-CPU | 1 | 0 | 0 | Health checks only (CPU transcode disabled) |
| Guava-VAAPI | 1 | 0 | 1 | AMD VAAPI GPU transcoding |
## 🚨 Troubleshooting
### Common Issues
**Transcoding stuck or slow**
- Check cache disk space: `df -h /volume3/usenet/tdarr_cache`
- Verify media permissions
- Check CPU/memory usage: `docker stats tdarr`
- Consider adding GPU transcoding
**Files not appearing**
- Verify media library path is correct
- Check file permissions (PUID/PGID)
- Scan library manually in Tdarr UI
**Node offline**
- Check container status on each host using the SSH commands in the Health Monitoring section above
- Atlantis/Calypso (Synology): `ssh <host> "sudo /usr/local/bin/docker logs tdarr-node-<name>"`
- Guava: `ssh guava "docker logs tdarr-node-guava"`
- NUC: `ssh pve "pct exec 103 -- docker logs tdarr-node"`
- All nodes lost connection when Tdarr server restarts — they reconnect automatically within ~30 seconds
**Node schedule showing all zeros (no workers active)**
- The per-node schedule overrides worker limits. If all 24 hours are set to 0 workers, nothing will process even if the base worker limits are configured.
- Fix via Tdarr UI (Nodes tab → schedule grid) or via API:
```bash
curl -s -X POST http://192.168.0.200:8265/api/v2/update-node \
-H 'Content-Type: application/json' \
-d '{"data":{"nodeID":"<NODE_ID>","nodeUpdates":{"schedule":[...]}}}'
```
**Transcodes complete but file not replaced (NFS nodes)**
- Remote nodes (Guava, Calypso) access media via NFS from Atlantis
- Atlantis NFS export for `/volume1/data` uses `all_squash,anonuid=1024` by default for 192.168.0.0/24, which maps all writes to uid 1024
- Media files are owned by uid 1029 — uid 1024 cannot write to them
- **Fix**: In DSM → Shared Folder → data → NFS Permissions, add a host-specific rule for the node's IP (e.g., 192.168.0.100 for Guava) with **No mapping** (no_all_squash) so the node can write as its actual UID
- The cache export (`/volume3/usenet`) already has `no_all_squash` so cache writes always work
**Flow edge missing — transcodes succeed but marked as error**
- If a flow plugin's output has no outgoing edge AND the working file is still in the transcode cache, Tdarr marks the job as "Transcode error"
- The fix: wire unconnected outputs to `setOriginalFile` — this resets the working file pointer to the original library path (not in cache), letting the job close cleanly as "Transcode success"
- See "Resetting errored files" in the Flows section to re-queue affected files
**Guava-VAAPI encoder limitations**
- AMD Radeon 760M (Phoenix) only supports `hevc_vaapi` encoding
- `h264_vaapi` and `av1_vaapi` both fail (exit code 228)
- The flow uses `hardwareType: auto` — Tdarr probes available hardware per node and selects the appropriate encoder automatically (VAAPI on Guava, QSV on NUC, CPU on others)
- Guava logs may show some encoder probe failures before settling on `hevc_vaapi` — this is expected
**Server/node version mismatch (Homarr widget ETIMEDOUT / nodes offline)**
- Tdarr server and all nodes must run the **same version**. A mismatch causes nodes to fail to connect, and the Homarr widget returns `ETIMEDOUT` because the server is effectively unreachable.
- The image tag is `:latest` on all compose files — but `docker pull` alone is not enough if the old container is still running the old image.
- **Fix**: Pull + stop + remove + recreate. On each affected host:
```bash
# On Atlantis (server)
ssh atlantis "sudo /usr/local/bin/docker pull ghcr.io/haveagitgat/tdarr:latest && \
sudo /usr/local/bin/docker stop tdarr && \
sudo /usr/local/bin/docker rm tdarr && \
sudo /usr/local/bin/docker compose -f /volume1/docker/arr-suite/docker-compose.yml up -d tdarr"
# On Calypso (node)
ssh calypso "sudo /usr/local/bin/docker pull ghcr.io/haveagitgat/tdarr_node:latest && \
sudo /usr/local/bin/docker stop tdarr-node-calypso && \
sudo /usr/local/bin/docker rm tdarr-node-calypso && \
sudo /usr/local/bin/docker compose -f /volume1/docker/tdarr-node/docker-compose.yaml up -d"
```
- After updating, all 4 nodes reconnect within ~30 seconds.
### Useful Commands
```bash
# View real-time logs (run on Atlantis via SSH)
ssh atlantis "sudo /usr/local/bin/docker logs -f tdarr"
# Restart server
ssh atlantis "sudo /usr/local/bin/docker restart tdarr"
# Restart a node
ssh calypso "sudo /usr/local/bin/docker restart tdarr-node-calypso"
ssh guava "docker restart tdarr-node-guava"
ssh pve "pct exec 103 -- docker restart tdarr-node"
# Update server image (pull + stop + rm + recreate to ensure new image is used)
ssh atlantis "sudo /usr/local/bin/docker pull ghcr.io/haveagitgat/tdarr:latest && \
sudo /usr/local/bin/docker stop tdarr && sudo /usr/local/bin/docker rm tdarr && \
sudo /usr/local/bin/docker compose -f /volume1/docker/arr-suite/docker-compose.yml up -d tdarr"
# Access server shell
ssh atlantis "sudo /usr/local/bin/docker exec -it tdarr /bin/bash"
```
## 🔄 Tdarr Flows
The system uses **Tdarr Flows** (instead of Classic Plugin Stack) for GPU/CPU fallback:
### Flow: "HEVC GPU with CPU Fallback"
Flow ID: `IMZomXmXOI` | DB: `flowsjsondb` on Atlantis
```
Input File
Begin Command (ffmpegCommandStart)
Set Video Encoder — Auto GPU, CRF/QP 20 (ffmpegCommandSetVideoEncoder)
↓ hardwareType: auto, hardwareEncoding: true
Execute (ffmpegCommandExecute)
Compare File Size Ratio (compareFileSizeRatio) ← 50100% of original
│ │
output 1: within range output 2: out of range
(file shrank → replace) (file grew → keep original)
↓ ↓
Replace Original File Set Original File
(resets working file to original,
discards transcode from cache,
completes as Transcode success)
```
**Flow Plugins:**
| Plugin | ID | Key Settings |
|--------|----|-------------|
| `inputFile` | `7IuuWhx9FF` | — |
| `ffmpegCommandStart` | `cmd_start` | — |
| `ffmpegCommandSetVideoEncoder` | `cmd_encoder` | `outputCodec: hevc`, `hardwareType: auto`, `hardwareEncoding: true`, `hardwareDecoding: true`, `ffmpegQuality: 20`, `forceEncoding: false` |
| `ffmpegCommandExecute` | `cmd_execute` | — |
| `compareFileSizeRatio` | `size_check` | `greaterThan: 50`, `lessThan: 100` |
| `replaceOriginalFile` | `JhCV_UZp7` | output 1 path (file smaller) |
| `setOriginalFile` | `keep_original` | output 2 path (file larger) |
**How it works:**
1. `hardwareType: auto` — Tdarr selects the best available encoder per node: QSV on NUC, VAAPI on Guava, CPU on Atlantis/Calypso
2. `ffmpegQuality: 20` — maps to `-global_quality 20` (QSV), `-qp 20` (VAAPI), `-crf 20` (CPU) — high quality, visually lossless
3. `forceEncoding: false` — files already in HEVC are skipped (marked Not Required)
4. `compareFileSizeRatio` — only replaces if output is 50100% of original size
5. `setOriginalFile` on output 2 — when the encode is larger, resets the working file pointer back to the original library file (not in cache), allowing the job to complete as "Transcode success" instead of "Transcode error". The oversized transcode in cache is discarded.
> **Note:** Do NOT leave output 2 of `compareFileSizeRatio` unconnected. Tdarr requires the final working file to not be in the cache — an unconnected output 2 leaves the transcode in the cache and Tdarr marks the job as "Transcode error" even though the original file was never replaced.
> **Warning:** Do NOT set `hardwareDecoding: false`. VAAPI encoding requires hardware context (`-hwaccel vaapi -hwaccel_device ...`) that the plugin only adds when `hardwareDecoding: true`. Disabling it breaks Guava entirely — all jobs fail immediately with empty video output. The NUC's QSV decode also works better with it enabled. To handle codecs the hardware decoder can't decode (e.g., AV1 on QSV), exclude them at the library level instead (see Library Codec Filter below).
**Modifying the flow:** The flow is stored directly in the SQLite DB. To update it:
```bash
ssh atlantis "python3 << 'PYEOF'
import sqlite3, json, time
db = '/volume2/metadata/docker2/tdarr/server/Tdarr/DB2/SQL/database.db'
# ... build new_flow dict ...
conn = sqlite3.connect(db)
cur = conn.cursor()
cur.execute('UPDATE flowsjsondb SET json_data = ?, timestamp = ? WHERE id = ?',
(json.dumps(new_flow), int(time.time()*1000), 'IMZomXmXOI'))
conn.commit()
conn.close()
PYEOF
"
ssh atlantis "sudo /usr/local/bin/docker restart tdarr"
```
> **Warning:** `checkFileSize` (the similarly-named built-in plugin) checks absolute file size (010000 GB by default), NOT relative to the original. It will pass every file and always replace. Use `compareFileSizeRatio` instead.
### Library Codec Filter
All three libraries (other, tv, anime) have a Classic Plugin Stack pre-filter that prevents certain codecs from being queued for transcoding:
| Library | `codecsToNotProcess` |
|---------|----------------------|
| other | `hevc,h265,av1` |
| tv | `hevc,h265,av1` |
| anime | `hevc,h265,av1` |
**Why AV1 is excluded:** AV1 is a newer, more efficient codec than HEVC. Converting AV1 → HEVC would produce larger files at the same quality. AV1 hardware decoding is also not supported on the NUC's Intel QSV, causing exit code 69 failures. AV1 files are left as-is.
To update the codec filter (e.g., to add a new codec to skip):
```bash
ssh atlantis "python3 << 'PYEOF'
import sqlite3, json, time
db = '/volume2/metadata/docker2/tdarr/server/Tdarr/DB2/SQL/database.db'
conn = sqlite3.connect(db)
cur = conn.cursor()
cur.execute('SELECT id, json_data FROM librarysettingsjsondb')
for row in cur.fetchall():
lib = json.loads(row[1])
for p in lib.get('pluginIDs', []):
if p.get('id') == 'Tdarr_Plugin_00td_filter_by_codec':
print(lib['name'], '->', p['InputsDB']['codecsToNotProcess'])
conn.close()
PYEOF
"
```
### Resetting errored files
After fixing a flow or filter issue, reset all "Transcode error" files back to Queued:
```bash
ssh atlantis "python3 << 'PYEOF'
import sqlite3
db = '/volume2/metadata/docker2/tdarr/server/Tdarr/DB2/SQL/database.db'
conn = sqlite3.connect(db)
cur = conn.cursor()
cur.execute(\"UPDATE filejsondb SET json_data = json_set(json_data, '$.TranscodeDecisionMaker', 'Queued') WHERE json_extract(json_data, '$.TranscodeDecisionMaker') = 'Transcode error'\")
print(f'Reset {cur.rowcount} files')
conn.commit()
conn.close()
PYEOF
"
```
### Enabling Flows on a Library
1. Go to **Libraries** → Select library
2. **Transcode Options** tab
3. Toggle **Flows: ON**
4. Toggle **Classic Plugin Stack: OFF**
5. Select the flow: "HEVC GPU with CPU Fallback"
## 🔧 Adding New Worker Nodes
### NFS Permissions for Remote Nodes
Remote nodes access media and cache via NFS from Atlantis. The NFS export settings on Atlantis must allow the node to write:
| Export | Path | Squash | Notes |
|--------|------|--------|-------|
| `/volume1/data` | Media library | `all_squash,anonuid=1024` (subnet default) | Must add per-host rule with **No mapping** for nodes that need write access |
| `/volume3/usenet` | Transcode cache | `no_all_squash` | Writable by all nodes |
To add write access for a new node: DSM → Control Panel → Shared Folder → data → Edit → NFS Permissions → Create rule for the node's IP with **No mapping**, **Read/Write**, **async**, **non-privileged ports allowed**.
### Adding a CPU-only Node (e.g., Synology NAS)
1. **Set up NFS mounts** to access media and cache:
```bash
# Create mount points
mkdir -p /mnt/atlantis_media /mnt/atlantis_cache
# Mount NFS shares
mount -t nfs 192.168.0.200:/volume1/data/media /mnt/atlantis_media -o rw,soft,nfsvers=3
mount -t nfs 192.168.0.200:/volume3/usenet/tdarr_cache /mnt/atlantis_cache -o rw,soft,nfsvers=3
```
2. **Create docker-compose.yaml**:
```yaml
services:
tdarr-node:
image: ghcr.io/haveagitgat/tdarr_node:latest
container_name: tdarr-node-<hostname>
environment:
- PUID=1029
- PGID=100
- TZ=America/Los_Angeles
- nodeName=<NodeName>
- serverIP=192.168.0.200
- serverPort=8266
- inContainer=true
- ffmpegVersion=6
volumes:
- ./configs:/app/configs
- ./logs:/app/logs
- /mnt/atlantis_media:/media
- /mnt/atlantis_cache:/temp
- /mnt/atlantis_cache:/cache
restart: always
```
3. **Configure workers** in Tdarr UI:
- Go to **Nodes** tab
- Set transcode CPU workers (2-4 recommended)
- Set healthcheck CPU workers (1-2 recommended)
### Adding a GPU Node (Intel QSV)
Same as above, but add device passthrough:
```yaml
devices:
- /dev/dri:/dev/dri # Intel QSV
```
## 📚 Additional Resources
- **Official Documentation**: [Tdarr Wiki](https://docs.tdarr.io/)
- **GitHub**: [HaveAGitGat/Tdarr](https://github.com/HaveAGitGat/Tdarr)
- **Discord**: Active community support
## 🔗 Related Services
Services REDACTED_APP_PASSWORD Tdarr:
- Plex
- Jellyfin
- Sonarr
- Radarr
---
*Last Updated*: 2026-03-10 (version sync troubleshooting, update procedure)
*Configuration Source*: `hosts/synology/atlantis/arr-suite/docker-compose.yml`