[PVE-User] blue screen on VM windows2016 with 5.1

Vincent Malien v.malien at socofer.com
Wed Sep 5 12:29:50 CEST 2018


I didn't catch the error code because the blue  screen isn't readable in 
the GUI console, no minidump file and nothing in Event Viewer.
I will try to create a linux VM and stress it. I will also try to create 
a windows VM with the latest virtio-win-0.1.160

signature Cordialement.
Vincent MALIEN
Le 05/09/2018 à 11:39, Yannis Milios a écrit :
> Ok, have you managed to "catch" the error code of the BSOD ? That may lead
> you somewhere ...
> Also, try creating a temp Linux (or different Windows version VM) and
> stress it out a bit. Do they also crash ?
>
> It could also be something related to the virtio drivers or to the qemu
> version. Hopefully on of the above can help you narrow down the issue...
>
>
>
>
>
> On Wed, 5 Sep 2018 at 10:17, Vincent Malien <v.malien at socofer.com> wrote:
>
>> Thanks for your answer.
>> The underlying storage type is RAID5 on 6 SSD SAS samsung 850pro + 1 spare:
>> #  megasasctl
>> a0       AVAGO 3108 MegaRAID      encl:2 ldrv:1  batt:FAULT, charge failed
>> a0d0      4766GiB RAID 5   1x6  optimal
>> unconfigured:  a0e0s6
>> a0e0s0      953GiB  a0d0  online
>> a0e0s1      953GiB  a0d0  online
>> a0e0s2      953GiB  a0d0  online
>> a0e0s3      953GiB  a0d0  online
>> a0e0s4      953GiB  a0d0  online
>> a0e0s5      953GiB  a0d0  online
>> a0e0s6      953GiB        ready
>>
>> The Load average is around: 2.33,2.00,1.10
>> CPU usage is around: 7.34% of 40 CPU(s
>>
>> The BSOD frequency for VM 109: 02/08 03/08 04/08 06/08 08/08 but not
>> recently
>> for VM 110: 06/08 08/08 11/08 18/08 25/08 02/09 04/09
>>
>> the syslog around the last BSOD yesterday, (we tryed several time to
>> stop the VM):
>> Sep  4 15:47:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
>> Sep  4 15:47:00 pve2 systemd[1]: Started Proxmox VE replication runner.
>> Sep  4 15:47:06 pve2 pvedaemon[931]: <root at pam> successful auth for user
>> 'root at pam'
>> Sep  4 15:47:06 pve2 pvedaemon[4779]: <root at pam> successful auth for
>> user 'root at pam'
>> Sep  4 15:47:15 pve2 pveproxy[39786]: worker 3754 finished
>> Sep  4 15:47:15 pve2 pveproxy[39786]: starting 1 worker(s)
>> Sep  4 15:47:15 pve2 pveproxy[39786]: worker 7023 started
>> Sep  4 15:47:15 pve2 pvedaemon[931]: <root at pam> starting task
>> UPID:pve2:00001B70:32148B26:5B8E8CE3:vncproxy:110:root at pam:
>> Sep  4 15:47:15 pve2 pvedaemon[7024]: starting vnc proxy
>> UPID:pve2:00001B70:32148B26:5B8E8CE3:vncproxy:110:root at pam:
>> Sep  4 15:47:16 pve2 pveproxy[7022]: worker exit
>> Sep  4 15:47:29 pve2 pvedaemon[7039]: starting vnc proxy
>> UPID:pve2:00001B7F:321490AA:5B8E8CF1:vncproxy:109:root at pam:
>> Sep  4 15:47:29 pve2 pvedaemon[4545]: <root at pam> starting task
>> UPID:pve2:00001B7F:321490AA:5B8E8CF1:vncproxy:109:root at pam:
>> Sep  4 15:47:33 pve2 pvedaemon[4545]: <root at pam> end task
>> UPID:pve2:00001B7F:321490AA:5B8E8CF1:vncproxy:109:root at pam: OK
>> Sep  4 15:47:38 pve2 pvedaemon[931]: <root at pam> end task
>> UPID:pve2:00001B70:32148B26:5B8E8CE3:vncproxy:110:root at pam: OK
>> Sep  4 15:47:44 pve2 pvedaemon[4779]: <root at pam> starting task
>> UPID:pve2:00001BA0:32149646:5B8E8D00:vncproxy:110:root at pam:
>> Sep  4 15:47:44 pve2 pvedaemon[7072]: starting vnc proxy
>> UPID:pve2:00001BA0:32149646:5B8E8D00:vncproxy:110:root at pam:
>> Sep  4 15:48:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
>> Sep  4 15:48:00 pve2 systemd[1]: Started Proxmox VE replication runner.
>> Sep  4 15:48:36 pve2 pvedaemon[4779]: <root at pam> starting task
>> UPID:pve2:00001BF0:3214AACA:5B8E8D34:qmstart:110:root at pam:
>> Sep  4 15:48:36 pve2 pvedaemon[7152]: start VM 110:
>> UPID:pve2:00001BF0:3214AACA:5B8E8D34:qmstart:110:root at pam:
>> Sep  4 15:48:36 pve2 pvedaemon[7152]: VM 110 already running
>> Sep  4 15:48:36 pve2 pvedaemon[4779]: <root at pam> end task
>> UPID:pve2:00001BF0:3214AACA:5B8E8D34:qmstart:110:root at pam: VM 110
>> already running
>> Sep  4 15:48:37 pve2 pvedaemon[4779]: <root at pam> end task
>> UPID:pve2:00001BA0:32149646:5B8E8D00:vncproxy:110:root at pam: OK
>> Sep  4 15:48:37 pve2 pvedaemon[4545]: <root at pam> starting task
>> UPID:pve2:00001BF7:3214AB38:5B8E8D35:vncproxy:110:root at pam:
>> Sep  4 15:48:37 pve2 pvedaemon[7159]: starting vnc proxy
>> UPID:pve2:00001BF7:3214AB38:5B8E8D35:vncproxy:110:root at pam:
>> Sep  4 15:48:57 pve2 pvedaemon[4545]: <root at pam> end task
>> UPID:pve2:00001BF7:3214AB38:5B8E8D35:vncproxy:110:root at pam: OK
>> Sep  4 15:48:59 pve2 pvedaemon[4779]: <root at pam> starting task
>> UPID:pve2:00001C12:3214B388:5B8E8D4B:vncproxy:110:root at pam:
>> Sep  4 15:48:59 pve2 pvedaemon[7186]: starting vnc proxy
>> UPID:pve2:00001C12:3214B388:5B8E8D4B:vncproxy:110:root at pam:
>> Sep  4 15:49:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
>> Sep  4 15:49:00 pve2 systemd[1]: Started Proxmox VE replication runner.
>> Sep  4 15:50:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
>> Sep  4 15:50:00 pve2 systemd[1]: Started Proxmox VE replication runner.
>> Sep  4 15:51:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
>> Sep  4 15:51:00 pve2 systemd[1]: Started Proxmox VE replication runner.
>> Sep  4 15:52:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
>> Sep  4 15:52:00 pve2 systemd[1]: Started Proxmox VE replication runner.
>> Sep  4 15:52:43 pve2 pvedaemon[931]: <root at pam> starting task
>> UPID:pve2:00001D4D:32150B3B:5B8E8E2B:vncproxy:110:root at pam:
>> Sep  4 15:52:43 pve2 pvedaemon[7501]: starting vnc proxy
>> UPID:pve2:00001D4D:32150B3B:5B8E8E2B:vncproxy:110:root at pam:
>> Sep  4 15:52:52 pve2 rrdcached[1379]: flushing old values
>> Sep  4 15:52:52 pve2 rrdcached[1379]: rotating journals
>> Sep  4 15:52:52 pve2 rrdcached[1379]: started new journal
>> /var/lib/rrdcached/journal/rrd.journal.1536069172.270060
>> Sep  4 15:52:52 pve2 rrdcached[1379]: removing old journal
>> /var/lib/rrdcached/journal/rrd.journal.1536061972.270091
>> Sep  4 15:52:52 pve2 pvedaemon[931]: <root at pam> starting task
>> UPID:pve2:00001D56:32150EA8:5B8E8E34:qmreset:110:root at pam:
>> Sep  4 15:52:52 pve2 pvedaemon[931]: <root at pam> end task
>> UPID:pve2:00001D56:32150EA8:5B8E8E34:qmreset:110:root at pam: OK
>> Sep  4 15:53:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
>> Sep  4 15:53:00 pve2 systemd[1]: Started Proxmox VE replication runner.
>> Sep  4 15:53:57 pve2 pvedaemon[7600]: starting vnc proxy
>> UPID:pve2:00001DB0:321527E9:5B8E8E75:vncproxy:110:root at pam:
>> Sep  4 15:53:57 pve2 pvedaemon[4545]: <root at pam> starting task
>> UPID:pve2:00001DB0:321527E9:5B8E8E75:vncproxy:110:root at pam:
>> Sep  4 15:53:59 pve2 pvedaemon[4545]: <root at pam> end task
>> UPID:pve2:00001DB0:321527E9:5B8E8E75:vncproxy:110:root at pam: OK
>> Sep  4 15:54:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
>> Sep  4 15:54:00 pve2 systemd[1]: Started Proxmox VE replication runner.
>> Sep  4 15:54:08 pve2 pvedaemon[931]: <root at pam> starting task
>> UPID:pve2:00001DC1:32152C73:5B8E8E80:qmshutdown:110:root at pam:
>> Sep  4 15:54:08 pve2 pvedaemon[7617]: shutdown VM 110:
>> UPID:pve2:00001DC1:32152C73:5B8E8E80:qmshutdown:110:root at pam:
>> Sep  4 15:54:19 pve2 pveproxy[39786]: worker 4697 finished
>> Sep  4 15:54:19 pve2 pveproxy[39786]: starting 1 worker(s)
>> Sep  4 15:54:19 pve2 pveproxy[39786]: worker 7631 started
>> Sep  4 15:54:23 pve2 pveproxy[7630]: got inotify poll request in wrong
>> process - disabling inotify
>> Sep  4 15:55:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
>> Sep  4 15:55:00 pve2 systemd[1]: Started Proxmox VE replication runner.
>> Sep  4 15:56:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
>> Sep  4 15:56:00 pve2 systemd[1]: Started Proxmox VE replication runner.
>> Sep  4 15:56:20 pve2 pvedaemon[4545]: <root at pam> successful auth for
>> user 'root at pam'
>> Sep  4 15:57:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
>> Sep  4 15:57:00 pve2 systemd[1]: Started Proxmox VE replication runner.
>> Sep  4 15:57:20 pve2 pvedaemon[4545]: <root at pam> successful auth for
>> user 'root at pam'
>> Sep  4 15:57:20 pve2 pvedaemon[931]: <root at pam> successful auth for user
>> 'root at pam'
>> Sep  4 15:57:23 pve2 pveproxy[39786]: worker 5535 finished
>> Sep  4 15:57:23 pve2 pveproxy[39786]: starting 1 worker(s)
>> Sep  4 15:57:23 pve2 pveproxy[39786]: worker 7920 started
>> Sep  4 15:57:25 pve2 pveproxy[7919]: got inotify poll request in wrong
>> process - disabling inotify
>> Sep  4 15:57:25 pve2 pveproxy[7919]: worker exit
>> Sep  4 15:57:40 pve2 pvedaemon[4545]: <root at pam> starting task
>> UPID:pve2:00001F03:32157F30:5B8E8F54:vncproxy:110:root at pam:
>> Sep  4 15:57:40 pve2 pvedaemon[7939]: starting vnc proxy
>> UPID:pve2:00001F03:32157F30:5B8E8F54:vncproxy:110:root at pam:
>> Sep  4 15:58:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
>> Sep  4 15:58:00 pve2 systemd[1]: Started Proxmox VE replication runner.
>> Sep  4 15:59:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
>> Sep  4 15:59:00 pve2 systemd[1]: Started Proxmox VE replication runner.
>> Sep  4 15:59:38 pve2 pvedaemon[4545]: <root at pam> end task
>> UPID:pve2:00001F03:32157F30:5B8E8F54:vncproxy:110:root at pam: OK
>> Sep  4 15:59:43 pve2 pvedaemon[8114]: starting vnc proxy
>> UPID:pve2:00001FB2:3215AF0C:5B8E8FCF:vncproxy:110:root at pam:
>> Sep  4 15:59:43 pve2 pvedaemon[4779]: <root at pam> starting task
>> UPID:pve2:00001FB2:3215AF0C:5B8E8FCF:vncproxy:110:root at pam:
>> Sep  4 15:59:44 pve2 pvedaemon[4779]: <root at pam> end task
>> UPID:pve2:00001FB2:3215AF0C:5B8E8FCF:vncproxy:110:root at pam: OK
>>
>> the kern.log:
>> Sep  4 15:47:06 pve2 pvedaemon[4779]: <root at pam> successful auth for
>> user 'root at pam'
>> Sep  4 15:47:15 pve2 pvedaemon[931]: <root at pam> starting task
>> UPID:pve2:00001B70:32148B26:5B8E8CE3:vncproxy:110:root at pam:
>> Sep  4 15:47:29 pve2 pvedaemon[4545]: <root at pam> starting task
>> UPID:pve2:00001B7F:321490AA:5B8E8CF1:vncproxy:109:root at pam:
>> Sep  4 15:47:33 pve2 pvedaemon[4545]: <root at pam> end task
>> UPID:pve2:00001B7F:321490AA:5B8E8CF1:vncproxy:109:root at pam: OK
>> Sep  4 15:47:38 pve2 pvedaemon[931]: <root at pam> end task
>> UPID:pve2:00001B70:32148B26:5B8E8CE3:vncproxy:110:root at pam: OK
>> Sep  4 15:47:44 pve2 pvedaemon[4779]: <root at pam> starting task
>> UPID:pve2:00001BA0:32149646:5B8E8D00:vncproxy:110:root at pam:
>> Sep  4 15:48:36 pve2 pvedaemon[4779]: <root at pam> starting task
>> UPID:pve2:00001BF0:3214AACA:5B8E8D34:qmstart:110:root at pam:
>> Sep  4 15:48:36 pve2 pvedaemon[4779]: <root at pam> end task
>> UPID:pve2:00001BF0:3214AACA:5B8E8D34:qmstart:110:root at pam: VM 110
>> already running
>> Sep  4 15:48:37 pve2 pvedaemon[4779]: <root at pam> end task
>> UPID:pve2:00001BA0:32149646:5B8E8D00:vncproxy:110:root at pam: OK
>> Sep  4 15:48:37 pve2 pvedaemon[4545]: <root at pam> starting task
>> UPID:pve2:00001BF7:3214AB38:5B8E8D35:vncproxy:110:root at pam:
>> Sep  4 15:48:57 pve2 pvedaemon[4545]: <root at pam> end task
>> UPID:pve2:00001BF7:3214AB38:5B8E8D35:vncproxy:110:root at pam: OK
>> Sep  4 15:48:59 pve2 pvedaemon[4779]: <root at pam> starting task
>> UPID:pve2:00001C12:3214B388:5B8E8D4B:vncproxy:110:root at pam:
>> Sep  4 15:52:43 pve2 pvedaemon[931]: <root at pam> starting task
>> UPID:pve2:00001D4D:32150B3B:5B8E8E2B:vncproxy:110:root at pam:
>> Sep  4 15:52:52 pve2 pvedaemon[931]: <root at pam> starting task
>> UPID:pve2:00001D56:32150EA8:5B8E8E34:qmreset:110:root at pam:
>> Sep  4 15:52:52 pve2 pvedaemon[931]: <root at pam> end task
>> UPID:pve2:00001D56:32150EA8:5B8E8E34:qmreset:110:root at pam: OK
>> Sep  4 15:53:57 pve2 pvedaemon[4545]: <root at pam> starting task
>> UPID:pve2:00001DB0:321527E9:5B8E8E75:vncproxy:110:root at pam:
>> Sep  4 15:53:59 pve2 pvedaemon[4545]: <root at pam> end task
>> UPID:pve2:00001DB0:321527E9:5B8E8E75:vncproxy:110:root at pam: OK
>> Sep  4 15:54:08 pve2 pvedaemon[931]: <root at pam> starting task
>> UPID:pve2:00001DC1:32152C73:5B8E8E80:qmshutdown:110:root at pam:
>> Sep  4 15:56:20 pve2 pvedaemon[4545]: <root at pam> successful auth for
>> user 'root at pam'
>> Sep  4 15:57:20 pve2 pvedaemon[4545]: <root at pam> successful auth for
>> user 'root at pam'
>> Sep  4 15:57:20 pve2 pvedaemon[931]: <root at pam> successful auth for user
>> 'root at pam'
>> Sep  4 15:57:40 pve2 pvedaemon[4545]: <root at pam> starting task
>> UPID:pve2:00001F03:32157F30:5B8E8F54:vncproxy:110:root at pam:
>> Sep  4 15:59:38 pve2 pvedaemon[4545]: <root at pam> end task
>> UPID:pve2:00001F03:32157F30:5B8E8F54:vncproxy:110:root at pam: OK
>> Sep  4 15:59:43 pve2 pvedaemon[4779]: <root at pam> starting task
>> UPID:pve2:00001FB2:3215AF0C:5B8E8FCF:vncproxy:110:root at pam:
>> Sep  4 15:59:44 pve2 pvedaemon[4779]: <root at pam> end task
>> UPID:pve2:00001FB2:3215AF0C:5B8E8FCF:vncproxy:110:root at pam: OK
>> I don't see any thing in syslog & kern.log
>> I haven't try switching the vdisk temporarily to IDE, but I don't know
>> what else to do...
>> signature Cordialement.
>> Vincent MALIEN
>> Le 05/09/2018 à 09:33, Yannis Milios a écrit :
>>> If both VMs fail with a BSOD, then definitely something must be wrong
>>> somewhere.
>>> Win2016 is supported in PVE 5+, so don't think it's necessary to upgrade
>> to
>>> a newer version.
>>> I would focus my attention on  any potential hardware issues on the
>> actual
>>> host (RAM,Storage etc).
>>> What's your underlying storage type (RAID,SSD,HDD) ? What are the load
>>> average values on the host ?
>>> Any clues in the Syslog ? Have you tried switching the vdisk temporarily
>> to
>>> IDE (even though, I don't think that will help in your case).
>>>
>>>
>>>
>>> On Wed, 5 Sep 2018 at 08:04, Vincent Malien <v.malien at socofer.com>
>> wrote:
>>>> Hi pve users,
>>>> I run 2 VM using windows 2016 witch do often blue screen and today this
>>>> message: guest has not initialize the display (yet)
>>>> here is my config:
>>>> proxmox-ve: 5.1-25 (running kernel: 4.13.4-1-pve)
>>>> pve-manager: 5.1-35 (running version: 5.1-35/722cc488)
>>>> pve-kernel-4.13.4-1-pve: 4.13.4-25
>>>> libpve-http-server-perl: 2.0-6
>>>> lvm2: 2.02.168-pve6
>>>> corosync: 2.4.2-pve3
>>>> libqb0: 1.0.1-1
>>>> pve-cluster: 5.0-15
>>>> qemu-server: 5.0-17
>>>> pve-firmware: 2.0-3
>>>> libpve-common-perl: 5.0-20
>>>> libpve-guest-common-perl: 2.0-13
>>>> libpve-access-control: 5.0-7
>>>> libpve-storage-perl: 5.0-16
>>>> pve-libspice-server1: 0.12.8-3
>>>> vncterm: 1.5-2
>>>> pve-docs: 5.1-12
>>>> pve-qemu-kvm: 2.9.1-2
>>>> pve-container: 2.0-17
>>>> pve-firewall: 3.0-3
>>>> pve-ha-manager: 2.0-3
>>>> ksm-control-daemon: 1.2-2
>>>> glusterfs-client: 3.8.8-1
>>>> lxc-pve: 2.1.0-2
>>>> lxcfs: 2.0.7-pve4
>>>> criu: 2.11.1-1~bpo90
>>>> novnc-pve: 0.6-4
>>>> smartmontools: 6.5+svn4324-1
>>>> zfsutils-linux: 0.7.2-pve1~bpo90
>>>>
>>>> qm config of 1VM:
>>>> agent: 1
>>>> bootdisk: scsi0
>>>> cores: 4
>>>> ide0: none,media=cdrom
>>>> memory: 12288
>>>> name: srverp
>>>> net0: virtio=F2:30:F0:DE:09:1F,bridge=vmbr0
>>>> numa: 0
>>>> ostype: win10
>>>> scsi0: local-lvm:vm-110-disk-1,discard=on,size=500G
>>>> scsihw: virtio-scsi-pci
>>>> smbios1: uuid=51c201a6-cd20-488c-9c89-f3f0fe4abd06
>>>> sockets: 1
>>>>
>>>> virtio is virtio-win-0.1.141
>>>> I checked the VM disk with windows tool, no error.
>>>> should I upgrade to 5.2 or some thing else?
>>>>
>>>> --
>>>> Cordialement.
>>>> Vincent MALIEN
>>>> /12 Avenue Yves Farge
>>>> BP 20258
>>>> 37702 St Pierre des Corps cedex 2/
>>>> _______________________________________________
>>>> pve-user mailing list
>>>> pve-user at pve.proxmox.com
>>>> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>>>>
>>> _______________________________________________
>>> pve-user mailing list
>>> pve-user at pve.proxmox.com
>>> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>> _______________________________________________
>> pve-user mailing list
>> pve-user at pve.proxmox.com
>> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>>
> _______________________________________________
> pve-user mailing list
> pve-user at pve.proxmox.com
> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user




More information about the pve-user mailing list