[PVE-User] blue screen on VM windows2016 with 5.1

Vincent Malien v.malien at socofer.com
Wed Sep 5 11:17:36 CEST 2018


Thanks for your answer.
The underlying storage type is RAID5 on 6 SSD SAS samsung 850pro + 1 spare:
#  megasasctl
a0       AVAGO 3108 MegaRAID      encl:2 ldrv:1  batt:FAULT, charge failed
a0d0      4766GiB RAID 5   1x6  optimal
unconfigured:  a0e0s6
a0e0s0      953GiB  a0d0  online
a0e0s1      953GiB  a0d0  online
a0e0s2      953GiB  a0d0  online
a0e0s3      953GiB  a0d0  online
a0e0s4      953GiB  a0d0  online
a0e0s5      953GiB  a0d0  online
a0e0s6      953GiB        ready

The Load average is around: 2.33,2.00,1.10
CPU usage is around: 7.34% of 40 CPU(s

The BSOD frequency for VM 109: 02/08 03/08 04/08 06/08 08/08 but not 
recently
for VM 110: 06/08 08/08 11/08 18/08 25/08 02/09 04/09

the syslog around the last BSOD yesterday, (we tryed several time to 
stop the VM):
Sep  4 15:47:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
Sep  4 15:47:00 pve2 systemd[1]: Started Proxmox VE replication runner.
Sep  4 15:47:06 pve2 pvedaemon[931]: <root at pam> successful auth for user 
'root at pam'
Sep  4 15:47:06 pve2 pvedaemon[4779]: <root at pam> successful auth for 
user 'root at pam'
Sep  4 15:47:15 pve2 pveproxy[39786]: worker 3754 finished
Sep  4 15:47:15 pve2 pveproxy[39786]: starting 1 worker(s)
Sep  4 15:47:15 pve2 pveproxy[39786]: worker 7023 started
Sep  4 15:47:15 pve2 pvedaemon[931]: <root at pam> starting task 
UPID:pve2:00001B70:32148B26:5B8E8CE3:vncproxy:110:root at pam:
Sep  4 15:47:15 pve2 pvedaemon[7024]: starting vnc proxy 
UPID:pve2:00001B70:32148B26:5B8E8CE3:vncproxy:110:root at pam:
Sep  4 15:47:16 pve2 pveproxy[7022]: worker exit
Sep  4 15:47:29 pve2 pvedaemon[7039]: starting vnc proxy 
UPID:pve2:00001B7F:321490AA:5B8E8CF1:vncproxy:109:root at pam:
Sep  4 15:47:29 pve2 pvedaemon[4545]: <root at pam> starting task 
UPID:pve2:00001B7F:321490AA:5B8E8CF1:vncproxy:109:root at pam:
Sep  4 15:47:33 pve2 pvedaemon[4545]: <root at pam> end task 
UPID:pve2:00001B7F:321490AA:5B8E8CF1:vncproxy:109:root at pam: OK
Sep  4 15:47:38 pve2 pvedaemon[931]: <root at pam> end task 
UPID:pve2:00001B70:32148B26:5B8E8CE3:vncproxy:110:root at pam: OK
Sep  4 15:47:44 pve2 pvedaemon[4779]: <root at pam> starting task 
UPID:pve2:00001BA0:32149646:5B8E8D00:vncproxy:110:root at pam:
Sep  4 15:47:44 pve2 pvedaemon[7072]: starting vnc proxy 
UPID:pve2:00001BA0:32149646:5B8E8D00:vncproxy:110:root at pam:
Sep  4 15:48:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
Sep  4 15:48:00 pve2 systemd[1]: Started Proxmox VE replication runner.
Sep  4 15:48:36 pve2 pvedaemon[4779]: <root at pam> starting task 
UPID:pve2:00001BF0:3214AACA:5B8E8D34:qmstart:110:root at pam:
Sep  4 15:48:36 pve2 pvedaemon[7152]: start VM 110: 
UPID:pve2:00001BF0:3214AACA:5B8E8D34:qmstart:110:root at pam:
Sep  4 15:48:36 pve2 pvedaemon[7152]: VM 110 already running
Sep  4 15:48:36 pve2 pvedaemon[4779]: <root at pam> end task 
UPID:pve2:00001BF0:3214AACA:5B8E8D34:qmstart:110:root at pam: VM 110 
already running
Sep  4 15:48:37 pve2 pvedaemon[4779]: <root at pam> end task 
UPID:pve2:00001BA0:32149646:5B8E8D00:vncproxy:110:root at pam: OK
Sep  4 15:48:37 pve2 pvedaemon[4545]: <root at pam> starting task 
UPID:pve2:00001BF7:3214AB38:5B8E8D35:vncproxy:110:root at pam:
Sep  4 15:48:37 pve2 pvedaemon[7159]: starting vnc proxy 
UPID:pve2:00001BF7:3214AB38:5B8E8D35:vncproxy:110:root at pam:
Sep  4 15:48:57 pve2 pvedaemon[4545]: <root at pam> end task 
UPID:pve2:00001BF7:3214AB38:5B8E8D35:vncproxy:110:root at pam: OK
Sep  4 15:48:59 pve2 pvedaemon[4779]: <root at pam> starting task 
UPID:pve2:00001C12:3214B388:5B8E8D4B:vncproxy:110:root at pam:
Sep  4 15:48:59 pve2 pvedaemon[7186]: starting vnc proxy 
UPID:pve2:00001C12:3214B388:5B8E8D4B:vncproxy:110:root at pam:
Sep  4 15:49:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
Sep  4 15:49:00 pve2 systemd[1]: Started Proxmox VE replication runner.
Sep  4 15:50:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
Sep  4 15:50:00 pve2 systemd[1]: Started Proxmox VE replication runner.
Sep  4 15:51:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
Sep  4 15:51:00 pve2 systemd[1]: Started Proxmox VE replication runner.
Sep  4 15:52:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
Sep  4 15:52:00 pve2 systemd[1]: Started Proxmox VE replication runner.
Sep  4 15:52:43 pve2 pvedaemon[931]: <root at pam> starting task 
UPID:pve2:00001D4D:32150B3B:5B8E8E2B:vncproxy:110:root at pam:
Sep  4 15:52:43 pve2 pvedaemon[7501]: starting vnc proxy 
UPID:pve2:00001D4D:32150B3B:5B8E8E2B:vncproxy:110:root at pam:
Sep  4 15:52:52 pve2 rrdcached[1379]: flushing old values
Sep  4 15:52:52 pve2 rrdcached[1379]: rotating journals
Sep  4 15:52:52 pve2 rrdcached[1379]: started new journal 
/var/lib/rrdcached/journal/rrd.journal.1536069172.270060
Sep  4 15:52:52 pve2 rrdcached[1379]: removing old journal 
/var/lib/rrdcached/journal/rrd.journal.1536061972.270091
Sep  4 15:52:52 pve2 pvedaemon[931]: <root at pam> starting task 
UPID:pve2:00001D56:32150EA8:5B8E8E34:qmreset:110:root at pam:
Sep  4 15:52:52 pve2 pvedaemon[931]: <root at pam> end task 
UPID:pve2:00001D56:32150EA8:5B8E8E34:qmreset:110:root at pam: OK
Sep  4 15:53:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
Sep  4 15:53:00 pve2 systemd[1]: Started Proxmox VE replication runner.
Sep  4 15:53:57 pve2 pvedaemon[7600]: starting vnc proxy 
UPID:pve2:00001DB0:321527E9:5B8E8E75:vncproxy:110:root at pam:
Sep  4 15:53:57 pve2 pvedaemon[4545]: <root at pam> starting task 
UPID:pve2:00001DB0:321527E9:5B8E8E75:vncproxy:110:root at pam:
Sep  4 15:53:59 pve2 pvedaemon[4545]: <root at pam> end task 
UPID:pve2:00001DB0:321527E9:5B8E8E75:vncproxy:110:root at pam: OK
Sep  4 15:54:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
Sep  4 15:54:00 pve2 systemd[1]: Started Proxmox VE replication runner.
Sep  4 15:54:08 pve2 pvedaemon[931]: <root at pam> starting task 
UPID:pve2:00001DC1:32152C73:5B8E8E80:qmshutdown:110:root at pam:
Sep  4 15:54:08 pve2 pvedaemon[7617]: shutdown VM 110: 
UPID:pve2:00001DC1:32152C73:5B8E8E80:qmshutdown:110:root at pam:
Sep  4 15:54:19 pve2 pveproxy[39786]: worker 4697 finished
Sep  4 15:54:19 pve2 pveproxy[39786]: starting 1 worker(s)
Sep  4 15:54:19 pve2 pveproxy[39786]: worker 7631 started
Sep  4 15:54:23 pve2 pveproxy[7630]: got inotify poll request in wrong 
process - disabling inotify
Sep  4 15:55:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
Sep  4 15:55:00 pve2 systemd[1]: Started Proxmox VE replication runner.
Sep  4 15:56:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
Sep  4 15:56:00 pve2 systemd[1]: Started Proxmox VE replication runner.
Sep  4 15:56:20 pve2 pvedaemon[4545]: <root at pam> successful auth for 
user 'root at pam'
Sep  4 15:57:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
Sep  4 15:57:00 pve2 systemd[1]: Started Proxmox VE replication runner.
Sep  4 15:57:20 pve2 pvedaemon[4545]: <root at pam> successful auth for 
user 'root at pam'
Sep  4 15:57:20 pve2 pvedaemon[931]: <root at pam> successful auth for user 
'root at pam'
Sep  4 15:57:23 pve2 pveproxy[39786]: worker 5535 finished
Sep  4 15:57:23 pve2 pveproxy[39786]: starting 1 worker(s)
Sep  4 15:57:23 pve2 pveproxy[39786]: worker 7920 started
Sep  4 15:57:25 pve2 pveproxy[7919]: got inotify poll request in wrong 
process - disabling inotify
Sep  4 15:57:25 pve2 pveproxy[7919]: worker exit
Sep  4 15:57:40 pve2 pvedaemon[4545]: <root at pam> starting task 
UPID:pve2:00001F03:32157F30:5B8E8F54:vncproxy:110:root at pam:
Sep  4 15:57:40 pve2 pvedaemon[7939]: starting vnc proxy 
UPID:pve2:00001F03:32157F30:5B8E8F54:vncproxy:110:root at pam:
Sep  4 15:58:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
Sep  4 15:58:00 pve2 systemd[1]: Started Proxmox VE replication runner.
Sep  4 15:59:00 pve2 systemd[1]: Starting Proxmox VE replication runner...
Sep  4 15:59:00 pve2 systemd[1]: Started Proxmox VE replication runner.
Sep  4 15:59:38 pve2 pvedaemon[4545]: <root at pam> end task 
UPID:pve2:00001F03:32157F30:5B8E8F54:vncproxy:110:root at pam: OK
Sep  4 15:59:43 pve2 pvedaemon[8114]: starting vnc proxy 
UPID:pve2:00001FB2:3215AF0C:5B8E8FCF:vncproxy:110:root at pam:
Sep  4 15:59:43 pve2 pvedaemon[4779]: <root at pam> starting task 
UPID:pve2:00001FB2:3215AF0C:5B8E8FCF:vncproxy:110:root at pam:
Sep  4 15:59:44 pve2 pvedaemon[4779]: <root at pam> end task 
UPID:pve2:00001FB2:3215AF0C:5B8E8FCF:vncproxy:110:root at pam: OK

the kern.log:
Sep  4 15:47:06 pve2 pvedaemon[4779]: <root at pam> successful auth for 
user 'root at pam'
Sep  4 15:47:15 pve2 pvedaemon[931]: <root at pam> starting task 
UPID:pve2:00001B70:32148B26:5B8E8CE3:vncproxy:110:root at pam:
Sep  4 15:47:29 pve2 pvedaemon[4545]: <root at pam> starting task 
UPID:pve2:00001B7F:321490AA:5B8E8CF1:vncproxy:109:root at pam:
Sep  4 15:47:33 pve2 pvedaemon[4545]: <root at pam> end task 
UPID:pve2:00001B7F:321490AA:5B8E8CF1:vncproxy:109:root at pam: OK
Sep  4 15:47:38 pve2 pvedaemon[931]: <root at pam> end task 
UPID:pve2:00001B70:32148B26:5B8E8CE3:vncproxy:110:root at pam: OK
Sep  4 15:47:44 pve2 pvedaemon[4779]: <root at pam> starting task 
UPID:pve2:00001BA0:32149646:5B8E8D00:vncproxy:110:root at pam:
Sep  4 15:48:36 pve2 pvedaemon[4779]: <root at pam> starting task 
UPID:pve2:00001BF0:3214AACA:5B8E8D34:qmstart:110:root at pam:
Sep  4 15:48:36 pve2 pvedaemon[4779]: <root at pam> end task 
UPID:pve2:00001BF0:3214AACA:5B8E8D34:qmstart:110:root at pam: VM 110 
already running
Sep  4 15:48:37 pve2 pvedaemon[4779]: <root at pam> end task 
UPID:pve2:00001BA0:32149646:5B8E8D00:vncproxy:110:root at pam: OK
Sep  4 15:48:37 pve2 pvedaemon[4545]: <root at pam> starting task 
UPID:pve2:00001BF7:3214AB38:5B8E8D35:vncproxy:110:root at pam:
Sep  4 15:48:57 pve2 pvedaemon[4545]: <root at pam> end task 
UPID:pve2:00001BF7:3214AB38:5B8E8D35:vncproxy:110:root at pam: OK
Sep  4 15:48:59 pve2 pvedaemon[4779]: <root at pam> starting task 
UPID:pve2:00001C12:3214B388:5B8E8D4B:vncproxy:110:root at pam:
Sep  4 15:52:43 pve2 pvedaemon[931]: <root at pam> starting task 
UPID:pve2:00001D4D:32150B3B:5B8E8E2B:vncproxy:110:root at pam:
Sep  4 15:52:52 pve2 pvedaemon[931]: <root at pam> starting task 
UPID:pve2:00001D56:32150EA8:5B8E8E34:qmreset:110:root at pam:
Sep  4 15:52:52 pve2 pvedaemon[931]: <root at pam> end task 
UPID:pve2:00001D56:32150EA8:5B8E8E34:qmreset:110:root at pam: OK
Sep  4 15:53:57 pve2 pvedaemon[4545]: <root at pam> starting task 
UPID:pve2:00001DB0:321527E9:5B8E8E75:vncproxy:110:root at pam:
Sep  4 15:53:59 pve2 pvedaemon[4545]: <root at pam> end task 
UPID:pve2:00001DB0:321527E9:5B8E8E75:vncproxy:110:root at pam: OK
Sep  4 15:54:08 pve2 pvedaemon[931]: <root at pam> starting task 
UPID:pve2:00001DC1:32152C73:5B8E8E80:qmshutdown:110:root at pam:
Sep  4 15:56:20 pve2 pvedaemon[4545]: <root at pam> successful auth for 
user 'root at pam'
Sep  4 15:57:20 pve2 pvedaemon[4545]: <root at pam> successful auth for 
user 'root at pam'
Sep  4 15:57:20 pve2 pvedaemon[931]: <root at pam> successful auth for user 
'root at pam'
Sep  4 15:57:40 pve2 pvedaemon[4545]: <root at pam> starting task 
UPID:pve2:00001F03:32157F30:5B8E8F54:vncproxy:110:root at pam:
Sep  4 15:59:38 pve2 pvedaemon[4545]: <root at pam> end task 
UPID:pve2:00001F03:32157F30:5B8E8F54:vncproxy:110:root at pam: OK
Sep  4 15:59:43 pve2 pvedaemon[4779]: <root at pam> starting task 
UPID:pve2:00001FB2:3215AF0C:5B8E8FCF:vncproxy:110:root at pam:
Sep  4 15:59:44 pve2 pvedaemon[4779]: <root at pam> end task 
UPID:pve2:00001FB2:3215AF0C:5B8E8FCF:vncproxy:110:root at pam: OK
I don't see any thing in syslog & kern.log
I haven't try switching the vdisk temporarily to IDE, but I don't know 
what else to do...
signature Cordialement.
Vincent MALIEN
Le 05/09/2018 à 09:33, Yannis Milios a écrit :
> If both VMs fail with a BSOD, then definitely something must be wrong
> somewhere.
> Win2016 is supported in PVE 5+, so don't think it's necessary to upgrade to
> a newer version.
> I would focus my attention on  any potential hardware issues on the actual
> host (RAM,Storage etc).
> What's your underlying storage type (RAID,SSD,HDD) ? What are the load
> average values on the host ?
> Any clues in the Syslog ? Have you tried switching the vdisk temporarily to
> IDE (even though, I don't think that will help in your case).
>
>
>
> On Wed, 5 Sep 2018 at 08:04, Vincent Malien <v.malien at socofer.com> wrote:
>
>> Hi pve users,
>> I run 2 VM using windows 2016 witch do often blue screen and today this
>> message: guest has not initialize the display (yet)
>> here is my config:
>> proxmox-ve: 5.1-25 (running kernel: 4.13.4-1-pve)
>> pve-manager: 5.1-35 (running version: 5.1-35/722cc488)
>> pve-kernel-4.13.4-1-pve: 4.13.4-25
>> libpve-http-server-perl: 2.0-6
>> lvm2: 2.02.168-pve6
>> corosync: 2.4.2-pve3
>> libqb0: 1.0.1-1
>> pve-cluster: 5.0-15
>> qemu-server: 5.0-17
>> pve-firmware: 2.0-3
>> libpve-common-perl: 5.0-20
>> libpve-guest-common-perl: 2.0-13
>> libpve-access-control: 5.0-7
>> libpve-storage-perl: 5.0-16
>> pve-libspice-server1: 0.12.8-3
>> vncterm: 1.5-2
>> pve-docs: 5.1-12
>> pve-qemu-kvm: 2.9.1-2
>> pve-container: 2.0-17
>> pve-firewall: 3.0-3
>> pve-ha-manager: 2.0-3
>> ksm-control-daemon: 1.2-2
>> glusterfs-client: 3.8.8-1
>> lxc-pve: 2.1.0-2
>> lxcfs: 2.0.7-pve4
>> criu: 2.11.1-1~bpo90
>> novnc-pve: 0.6-4
>> smartmontools: 6.5+svn4324-1
>> zfsutils-linux: 0.7.2-pve1~bpo90
>>
>> qm config of 1VM:
>> agent: 1
>> bootdisk: scsi0
>> cores: 4
>> ide0: none,media=cdrom
>> memory: 12288
>> name: srverp
>> net0: virtio=F2:30:F0:DE:09:1F,bridge=vmbr0
>> numa: 0
>> ostype: win10
>> scsi0: local-lvm:vm-110-disk-1,discard=on,size=500G
>> scsihw: virtio-scsi-pci
>> smbios1: uuid=51c201a6-cd20-488c-9c89-f3f0fe4abd06
>> sockets: 1
>>
>> virtio is virtio-win-0.1.141
>> I checked the VM disk with windows tool, no error.
>> should I upgrade to 5.2 or some thing else?
>>
>> --
>> Cordialement.
>> Vincent MALIEN
>> /12 Avenue Yves Farge
>> BP 20258
>> 37702 St Pierre des Corps cedex 2/
>> _______________________________________________
>> pve-user mailing list
>> pve-user at pve.proxmox.com
>> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>>
> _______________________________________________
> pve-user mailing list
> pve-user at pve.proxmox.com
> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user




More information about the pve-user mailing list