Proxmox admins
4-node blade cluster, storage is backed by nas.k-space.ee. Will be moved to Supermicro 2U servers with NVMe+HDD CEPH in 2025-07.
Better self-service sometime™: https://git.k-space.ee/k-space/todo/issues/127
xyz.k-space.ee
domain name!Default SSH username is ubuntu
debian
.
Default IPv6 is currently a single-IP abnormality, fitting for the Zoo network. As long as it doesn't conflict with existing IPs, you are welcome to change it.
Used to:
If you remove or disable qemu-guest-agent
:
nas
ISO storage is shared and accessible to all PVE users. Integrity of images is not guaranteed!
nas
storage → ISO Images
.Reserved for Proxmox VMs with IDs 100..199:
IPv4 193.40.103.100
.. 193.40.103.199
IPv6 2001:bb8:4008:20::100/64
.. 2001:bb8:4008:20::199/64
155
gets 193.40.103.155
, 2001:bb8:4008:20::155
.K-SPACE_VM_USER
.ssh debian@[2001:bb8:4008:20::155]
← change to match configNote: CPUs are set higher to allow hotplugging to more, vCPU and resource limits are what matter.
/mnt/pve/nas/jc/
)template.sh
: create up-to-date templatesupgr
: backup /etc
, migrate VMs away, shut down nomigrate
, upgrade and reboot. Confirm kube VMs are up 5m00s before proceeding to next node.toh
: migrate VMs back. Run des
to change configuration. May be replaced with HA.stalk_updates.sh
: Enumerate guests on the host, check and report on updates. Was used to identify machines way out of date and (20.04) EoL. Should be refactored to modules based on tags.ha-manager crm-command node-maintenance enable $HOSTNAME
upgr # wrapper to do everything
ha-manager crm-command node-maintenance disable $HOSTNAME
/etc/apt/sources.list.d/
,deb http://download.proxmox.com/debian/pve bullseye pve-no-subscription
, make sure to update the codename.This has not happened in years.
Failed.
2021-02-21 17:44:11 migration status: completed
2021-02-21 17:44:14 # /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=pve1' root@2001:bb8:4008:21:172:21:20:1 qm unlock 156
2021-02-21 17:44:14 ERROR: failed to clear migrate lock: Configuration file 'nodes/pve1/qemu-server/156.conf' does not exist
2021-02-21 17:44:14 ERROR: migration finished with problems (duration 00:00:32)
TASK ERROR: migration problems
Go to the shell of the sender, run:
qm unlock <id of the vm stuck, ex 111>
If the command output is Configuration file 'nodes/pve2/qemu-server/156.conf' does not exist
, run it on the recieving end instead. If that succeeds fine, you don't have to hibernate.
Hibernate the VM in question on the sender node. (broadcasts and sets/overwrites various doings of the VM, hibernate is the least costly)
Resume the VM.
Re-attempt migration. This attempt will fail. (cleanup)
Re-attempt migration.
Part of sweaty old documentation, no use with current tooling.
modprobe nbd max_part=8
qemu-nbd --connect /dev/nbd0 /mnt/pve/nas/images/9028/vm-9028-disk-0.raw
mkdir /mnt/cr
mount /dev/nbd0p2 /mnt/cr
mount -o bind /proc /mnt/cr/proc
mount -o bind /sys /mnt/cr/sys
mount -o bind /dev /mnt/cr/dev
mv /mnt/cr/etc/resolv.conf /mnt/cr_resolv.conf
cp /etc/resolv.conf /mnt/cr/etc/resolv.conf
chroot /mnt/cr
exit
umount /mnt/cr/{proc,sys,dev}
rm /mnt/cr/etc/resolv.conf
mv /mnt/cr_resolv.conf /mnt/cr/etc/resolv.conf
umount /mnt/cr
qemu-nbd --disconnect /dev/nbd0
rmmod nbd