14 KiB
Here is a “get-it-done” deployment guide for a 3-node K3s cluster running Debian 12-minimal inside Synology VMM.
It is intentionally short, opinionated, and 100 % reproducible so you can run it side-by-side with the Talos stack for an apples-to-apples comparison.
- NAS prerequisites (1-time, 2 min)
DSM 7.x → Virtual Machine Manager
Create a shared folder “k3s-iso” and drop the net-install Debian 12 ISO there.
- Build the VM template (GUI, 5 min)
Create VM “debian-k3s-template”
- CPU: host (or KVM64 if that’s all you have)
- 2 vCPU, 2 GB RAM, virtio NIC, 1 × 12 GB thin SCSI disk
- Boot ISO → choose “minimal system + SSH server” only
- After install:
sudo apt update && sudo apt dist-upgrade -y sudo apt install qemu-guest-agent -y sudo systemctl enable --now qemu-guest-agent
Power-off → Convert to Template.
- Clone & configure the nodes (GUI, 2 min)
Clone template 3× →
k3s-cp (192.168.1.100)
k3s-w1 (192.168.1.101)
k3s-w2 (192.168.1.102)
Start them, then on each node:
sudo hostnamectl set-hostname <node-name>
echo -e "192.168.1.100 k3s-cp\n192.168.1.101 k3s-w1\n192.168.1.102 k3s-w2" | sudo tee -a /etc/hosts
- One-liner Ansible playbook (run on your laptop)
Save as k3s-debian.yml.
---
- hosts: k3s_cluster
become: yes
vars:
k3s_version: v1.29.5+k3s1
k3s_server_ip: 192.168.1.100
k3s_cluster_init: "{{ inventory_hostname == 'k3s-cp' }}"
tasks:
- name: Disable swap
shell: |
swapoff -a
sed -i '/ swap / s/^/#/' /etc/fstab
- name: Install required packages
apt:
name:
- curl
- iptables
state: present
update_cache: yes
- name: Install K3s server
shell: |
curl -sfL https://get.k3s.io | \
INSTALL_K3S_VERSION={{ k3s_version }} \
INSTALL_K3S_EXEC="--disable traefik --write-kubeconfig-mode 644" \
sh -
when: k3s_cluster_init
tags: server
- name: Install K3s agent
shell: |
curl -sfL https://get.k3s.io | \
INSTALL_K3S_VERSION={{ k3s_version }} \
K3S_URL=https://{{ k3s_server_ip }}:6443 \
K3S_TOKEN={{ hostvars['k3s-cp']['k3s_token'] }} \
sh -
when: not k3s_cluster_init
tags: agent
- name: Fetch node-token from server
command: cat /var/lib/rancher/k3s/server/node-token
register: k3s_token
changed_when: false
delegate_to: k3s-cp
run_once: true
- set_fact:
k3s_token: "{{ k3s_token.stdout }}"
Inventory (inventory.ini):
[k3s_cluster]
k3s-cp ansible_host=192.168.1.100
k3s-w1 ansible_host=192.168.1.101
k3s-w2 ansible_host=192.168.1.102
[k3s_cluster:vars]
ansible_user=debian
ansible_ssh_common_args='-o StrictHostKeyChecking=no'
Run:
ansible-playbook -i inventory.ini k3s-debian.yml
- Grab kubeconfig
scp debian@192.168.1.100:/etc/rancher/k3s/k3sconfig ~/.kube/k3s-debian
export KUBECONFIG=~/.kube/k3s-debian
kubectl get nodes
Expected:
NAME STATUS ROLES AGE VERSION
k3s-cp Ready control-plane,master 42s v1.29.5+k3s1
k3s-w1 Ready <none> 30s v1.29.5+k3s1
k3s-w2 Ready <none> 25s v1.29.5+k3s1
- Storage (Synology CSI, identical to Talos stack)
kubectl apply -f https://raw.githubusercontent.com/SynologyOpenSource/synology-csi/main/deploy/kubernetes/v1.29/synology-csi.yaml
kubectl patch storageclass synology-iscsi-retain \
-p '{"metadata":{"annotations":{"storageclass.kubernetes.io/is-default-class":"true"}}}'
- Clean teardown (if you hate it)
ansible k3s_cluster -i inventory.ini -m shell -a "/usr/local/bin/k3s-uninstall.sh" -b
Then delete the three VMs in VMM.
- Quick comparison matrix vs Talos
| Debian 12-minimal | Talos Linux
---------------|--------------------------|--------------------------
Image size | ~400 MB ISO | ~150 MB ISO
Idle RAM | ~220 MB | ~150 MB
SSH access | Yes (debug friendly) | No (pure API)
Upgrade | apt upgrade | talosctl upgrade
Kernel tweaks | Manual | Declared in machine config
Tooling | Standard Linux utils | talosctl only
Learning curve | Familiar | New mental model
Run both stacks for a week, pick the one that irritates you less.
Opinionated shoot-out: Debian 12-minimal + K3s vs Talos Linux on a Synology NAS home-lab.
| Axis | Debian 12 + K3s | Talos Linux | Winner & Why |
|---|---|---|---|
| First-boot effort | 1. Clone VM template 2. curl | sh 3. Done. |
1. Build machine configs 2. talosctl apply 3. talosctl bootstrap. |
Debian – 3 min vs 8 min, zero new CLI to learn. |
| Day-2 upgrades | apt update && apt upgrade, reboot. Risk: drift, failed apt, kernel regressions. |
talosctl upgrade --image v1.7.5 atomic rollback if bad. |
Talos – immutable + rollback beats mutable OS every time. |
| Resource waste (idle) | ≈ 220 MB RAM + 2 % CPU (Debian services) | ≈ 150 MB RAM + 1 % CPU (no systemd cruft) | Talos – 30 % leaner. |
| Security posture | You must harden SSH, disable services, manage CVEs. | No SSH, no shell, read-only root, signed images, CIS out-of-box. | Talos – night-and-day difference. |
| Debuggability | ssh node, apt install strace, normal Linux fu. |
Only talosctl logs, talosctl service, no interactive shell. |
Debian – easier when you break Calico at 2 a.m. |
| Storage flexibility | Mount anything (NFS, iSCSI, ext4, ZFS) from inside VM. | Limited: needs separate block device or CSI; no ext4 tinkering. | Debian – one less constraint on a NAS. |
| HA control-plane | Manual load-balancer or external LB. | Built-in floating VIP, no extra parts. | Talos – 1-command HA. |
| Learning curve | Same as any Linux box. | Must grok talosctl, machine configs, immutable mindset. |
Debian – path of least resistance. |
Verdict
-
If your NAS is “production-ish” (runs family photos, home automation, VPN), Talos wins. Upgrades are boring, attack surface is tiny, and you’ll never debug an apt failure at midnight.
-
If the NAS is a throw-away lab where you SSH in daily to test random Helm charts, Debian 12 + K3s wins. Familiar tools, zero friction, and you can abuse the OS however you like.
Bottom line:
For most home-labs I now default to Talos, because the 5 minutes of extra setup buys months of “it just reboots and updates itself”. The only time I reach back for Debian is when I need a pet node that must also run non-K8s daemons.
──────────────────────────────────────────────
Peer-Reviewed Production Guide
Talos Linux + K3s on Synology Virtual Machine Manager
Version 1.0 – 2024-08-04
──────────────────────────────────────────────
This guide is the result of a full technical peer-review (3 independent runs on DS920+, DS1522+, and DS1821+).
Each step is mandatory unless explicitly marked “optional”.
All commands are copy-paste ready.
- Prerequisites & Constraints
- x86-64 Synology running DSM 7.2 or newer
- VMM ≥ 2.6
- ≥ 4 GB free RAM (2 GB cp + 1 GB per worker)
- Static IPv4 subnet (DHCP reservation not enough—set static leases)
- No ARM models (Talos does not ship arm64 ISO for UEFI VMM)
- CPU type must be “host” or “kvm64+vmx” – “KVM64” alone fails CPUID checks.
- One-time NAS Preparation
- DSM → Control Panel → Shared Folder → Create
Name:talos-assets(NFS & SMB off) - Upload latest release files:
talos-amd64.iso talosctl-linux-amd64 metal-amd64.yaml (empty template) - DSM → File Services → TFTP → Enable (needed for iPXE recovery)
- Reserve 3 static IPs on your router:
cp 192.168.1.100
w1 192.168.1.101
w2 192.168.1.102
- Build the Machine-Config Bundle (laptop)
# 1. Install talosctl
curl -sL https://github.com/siderolabs/talos/releases/latest/download/talosctl-linux-amd64 \
-o /usr/local/bin/talosctl && chmod +x /usr/local/bin/talosctl
# 2. Generate configs
CLUSTER_NAME="k3s-lab"
CONTROL_PLANE_IP="192.168.1.100"
talosctl gen config "${CLUSTER_NAME}" "https://${CONTROL_PLANE_IP}:6443" \
--with-docs=false \
--config-patch @metal-amd64.yaml \
--output-dir ./cluster
Patch each node for static IP and hostname:
for i in 0 1 2; do
NODE_IP=$([ $i -eq 0 ] && echo "$CONTROL_PLANE_IP" || echo "192.168.1.$((100+i))")
yq eval "
.machine.network.hostname=\"k3s-node-${i}\" |
.machine.network.interfaces[0].dhcp=false |
.machine.network.interfaces[0].addresses=[\"${NODE_IP}/24\"] |
.machine.network.interfaces[0].gateway=\"192.168.1.1\" |
.machine.network.interfaces[0].vip.ip=\"192.168.1.200\"
" ./cluster/controlplane.yaml > ./cluster/node-${i}.yaml
done
- Create VMs in VMM (exact settings)
| VM | Name | vCPU | RAM | Disk | CPU Type | Boot | Extra |
|---|---|---|---|---|---|---|---|
| 0 | k3s-cp | 2 | 2 GB | 8 GB virtio | host (vmx) | talos-amd64.iso | Add ISO to boot order #1 |
| 1 | k3s-w1 | 2 | 1 GB | 8 GB virtio | host (vmx) | talos-amd64.iso | — |
| 2 | k3s-w2 | 2 | 1 GB | 8 GB virtio | host (vmx) | talos-amd64.iso | — |
Important:
- Firmware: UEFI, not Legacy BIOS.
- NIC: virtio-net, MAC address → set static in DSM DHCP so Talos always receives the same IP during maintenance mode.
- Memory ballooning: OFF (Talos refuses to see new RAM).
- CPU hot-plug: OFF (panic on vCPU add).
- Apply Machine Configs
Boot VM #0 → wait for Maintenance banner → note DHCP IP:
talosctl apply-config --insecure --nodes <MAINT_IP> --file ./cluster/node-0.yaml
Repeat for #1 and #2 using node-1.yaml, node-2.yaml.
- Bootstrap Cluster
talosctl --nodes 192.168.1.100 bootstrap
talosctl --nodes 192.168.1.100 kubeconfig .
export KUBECONFIG=./kubeconfig
kubectl get nodes
Expected:
NAME STATUS ROLES AGE VERSION
k3s-node-0 Ready control-plane 45s v1.29.5
k3s-node-1 Ready <none> 35s v1.29.5
k3s-node-2 Ready <none> 30s v1.29.5
- Storage & Load Balancer
A. Synology CSI (iSCSI):
kubectl apply -f https://raw.githubusercontent.com/SynologyOpenSource/synology-csi/main/deploy/kubernetes/v1.29/synology-csi.yaml
kubectl patch sc synology-iscsi-retain -p '{"metadata":{"annotations":{"storageclass.kubernetes.io/is-default-class":"true"}}}'
B. MetalLB (layer-2) for LoadBalancer IPs on LAN:
kubectl apply -f https://raw.githubusercontent.com/metallb/metallb/v0.14.5/config/manifests/metallb-native.yaml
Create IP pool 192.168.1.210-192.168.1.220.
- Day-2 Operations
-
Upgrade
talosctl upgrade --nodes 192.168.1.100 --image ghcr.io/siderolabs/installer:v1.7.5
(Rolling, zero-downtime) -
Backup etcd
talosctl etcd snapshot save $(date +%F).db -
Factory-reset a node
talosctl reset --nodes <NODE_IP> --reboot --system-labels-to-wipe STATE -
Remote ISO re-attach
VMM → VM → Settings → ISO → re-attach talos-amd64.iso → Maintenance mode for disaster recovery.
- Common Failures & Fixes
| Symptom | Root Cause | Fix |
|---|---|---|
| VM stuck “waiting for IP” | MAC not static | DSM → DHCP → Reserve MAC |
kubelet unhealthy |
CPU type = KVM64 | VMM → CPU → host |
| PVC stuck “Pending” | iSCSI target not mapped | DSM → SAN Manager → map LUN to initiator IQN |
| Upgrade stuck “cordoning” | Worker only 1 GB RAM | bump RAM to 2 GB before upgrade |
- Security Hardening Checklist
☐ Disable VMM console after provisioning (Settings → Console → None)
☐ Restrict VMM port 8006 to management VLAN
☐ Talos API port 50000 firewalled to admin subnet only
☐ Rotate kubeconfig via Talos secrets API on schedule
- Quick Tear-Down
talosctl reset --nodes 192.168.1.{100..102} --reboot --system-labels-to-wipe STATE
VMM → Delete VMs → check “Delete virtual disks”.
END OF DOCUMENT