Proxmox admins
4-node blade cluster, storage is backed by nas.k-space.ee. Will be moved to Supermicro 2U servers with NVMe+HDD CEPH in 2025-07.
Better self-service sometime™: https://git.k-space.ee/k-space/todo/issues/127
xyz.k-space.ee domain name!
Default SSH username is ubuntudebian.
Default IPv6 is currently a single-IP abnormality, fitting for the Zoo network. As long as it doesn't conflict with existing IPs, you are welcome to change it.
Used to:
If you remove or disable qemu-guest-agent:
nas ISO storage is shared and accessible to all PVE users. Integrity of images is not guaranteed!

nas storage → ISO Images.
Reserved for Proxmox VMs with IDs 100..199:
IPv4 193.40.103.100 .. 193.40.103.199
IPv6 2001:bb8:4008:20::100/64 .. 2001:bb8:4008:20::199/64
155 gets 193.40.103.155, 2001:bb8:4008:20::155.K-SPACE_VM_USER.ssh debian@[2001:bb8:4008:20::155] ← change to match configNote: CPUs are set higher to allow hotplugging to more, vCPU and resource limits are what matter.
/mnt/pve/nas/jc/)template.sh: create up-to-date templatesupgr: backup /etc, migrate VMs away, shut down nomigrate, upgrade and reboot. Confirm kube VMs are up 5m00s before proceeding to next node.toh: migrate VMs back. Run des to change configuration. May be replaced with HA.stalk_updates.sh: Enumerate guests on the host, check and report on updates. Was used to identify machines way out of date and (20.04) EoL. Should be refactored to modules based on tags.ha-manager crm-command node-maintenance enable $HOSTNAME
upgr # wrapper to do everything
ha-manager crm-command node-maintenance disable $HOSTNAME
/etc/apt/sources.list.d/,deb http://download.proxmox.com/debian/pve bullseye pve-no-subscription, make sure to update the codename.This has not happened in years.
Failed.
2021-02-21 17:44:11 migration status: completed
2021-02-21 17:44:14 # /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=pve1' root@2001:bb8:4008:21:172:21:20:1 qm unlock 156
2021-02-21 17:44:14 ERROR: failed to clear migrate lock: Configuration file 'nodes/pve1/qemu-server/156.conf' does not exist
2021-02-21 17:44:14 ERROR: migration finished with problems (duration 00:00:32)
TASK ERROR: migration problems
Go to the shell of the sender, run:
qm unlock <id of the vm stuck, ex 111>
If the command output is Configuration file 'nodes/pve2/qemu-server/156.conf' does not exist, run it on the recieving end instead. If that succeeds fine, you don't have to hibernate.
Hibernate the VM in question on the sender node. (broadcasts and sets/overwrites various doings of the VM, hibernate is the least costly)
Resume the VM.
Re-attempt migration. This attempt will fail. (cleanup)
Re-attempt migration.
Part of sweaty old documentation, no use with current tooling.
modprobe nbd max_part=8
qemu-nbd --connect /dev/nbd0 /mnt/pve/nas/images/9028/vm-9028-disk-0.raw
mkdir /mnt/cr
mount /dev/nbd0p2 /mnt/cr
mount -o bind /proc /mnt/cr/proc
mount -o bind /sys /mnt/cr/sys
mount -o bind /dev /mnt/cr/dev
mv /mnt/cr/etc/resolv.conf /mnt/cr_resolv.conf
cp /etc/resolv.conf /mnt/cr/etc/resolv.conf
chroot /mnt/cr
exit
umount /mnt/cr/{proc,sys,dev}
rm /mnt/cr/etc/resolv.conf
mv /mnt/cr_resolv.conf /mnt/cr/etc/resolv.conf
umount /mnt/cr
qemu-nbd --disconnect /dev/nbd0
rmmod nbd