[PVE-User] qm migrate: strange output

Michael Rasmussen mir at miras.org
Tue Jun 20 19:01:15 CEST 2017


The former is for HA vm's the latter for non HA vm's

On June 20, 2017 6:19:36 PM GMT+02:00, Uwe Sauter <uwe.sauter.de at gmail.com> wrote:
>Hi all,
>
>usually when I update my PVE cluster I do it in a rolling fashion:
>1) empty one node from running VMs
>2) update & reboot that node
>3) go to next node
>4) migrate all running VMs to already updated node
>5) go to 2 until no more nodes need update
>
>For step 1 (or 4) I usually do:
>
># qm list
>      VMID NAME                 STATUS     MEM(MB)    BOOTDISK(GB) PID
>       106 test1                running    2048              32.00 4993
>       112 test2                running    1024              16.00 5218
>
># for i in 106 112; do qm migrate $i px-bravo-cluster --online; done
>
>Usually I get multiple lines like:
>
>Executing HA migrate for VM 106 to node px-bravo-cluster
>Executing HA migrate for VM 112 to node px-bravo-cluster
>
>But once in a while (in the last few days more often) I get:
>
>Jun 20 18:15:42 starting migration of VM 106 to node 'px-bravo-cluster'
>(169.254.42.49)
>Jun 20 18:15:42 copying disk images
>Jun 20 18:15:42 starting VM 106 on remote node 'px-bravo-cluster'
>Jun 20 18:15:46 start remote tunnel
>Jun 20 18:15:46 starting online/live migration on
>unix:/run/qemu-server/106.migrate
>Jun 20 18:15:46 migrate_set_speed: 8589934592
>Jun 20 18:15:46 migrate_set_downtime: 0.1
>Jun 20 18:15:46 set migration_caps
>Jun 20 18:15:46 set cachesize: 214748364
>Jun 20 18:15:46 start migrate command to
>unix:/run/qemu-server/106.migrate
>Jun 20 18:15:48 migration status: active (transferred 632787397,
>remaining 463048704), total 2156732416)
>Jun 20 18:15:48 migration xbzrle cachesize: 134217728 transferred 0
>pages 0 cachemiss 0 overflow 0
>Jun 20 18:15:50 migration speed: 512.00 MB/s - downtime 81 ms
>Jun 20 18:15:50 migration status: completed
>Jun 20 18:15:54 migration finished successfully (duration 00:00:13)
>Jun 20 18:15:55 starting migration of VM 112 to node 'px-bravo-cluster'
>(169.254.42.49)
>Jun 20 18:15:55 copying disk images
>Jun 20 18:15:55 starting VM 112 on remote node 'px-bravo-cluster'
>Jun 20 18:15:58 start remote tunnel
>Jun 20 18:15:58 starting online/live migration on
>unix:/run/qemu-server/112.migrate
>Jun 20 18:15:58 migrate_set_speed: 8589934592
>Jun 20 18:15:58 migrate_set_downtime: 0.1
>Jun 20 18:15:58 set migration_caps
>Jun 20 18:15:58 set cachesize: 107374182
>Jun 20 18:15:58 start migrate command to
>unix:/run/qemu-server/112.migrate
>Jun 20 18:16:00 migration status: active (transferred 876920642,
>remaining 143405056), total 1082990592)
>Jun 20 18:16:00 migration xbzrle cachesize: 67108864 transferred 0
>pages 0 cachemiss 0 overflow 0
>Jun 20 18:16:02 migration speed: 256.00 MB/s - downtime 66 ms
>Jun 20 18:16:02 migration status: completed
>Jun 20 18:16:06 migration finished successfully (duration 00:00:12)
>
>
>Can someone explain under which circumstances this output is displayed
>instead of just the short message that migration
>was started?
>
>
>Regards,
>
>	Uwe
>_______________________________________________
>pve-user mailing list
>pve-user at pve.proxmox.com
>https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user

-- 
Sent from my Android phone with K-9 Mail. Please excuse my brevity.

----

This mail was virus scanned and spam checked before delivery.
This mail is also DKIM signed. See header dkim-signature.





More information about the pve-user mailing list