summaryrefslogtreecommitdiffstats
path: root/results/classifier/zero-shot/118/all/1207686
diff options
context:
space:
mode:
Diffstat (limited to 'results/classifier/zero-shot/118/all/1207686')
-rw-r--r--results/classifier/zero-shot/118/all/1207686273
1 files changed, 273 insertions, 0 deletions
diff --git a/results/classifier/zero-shot/118/all/1207686 b/results/classifier/zero-shot/118/all/1207686
new file mode 100644
index 00000000..7ca4ccce
--- /dev/null
+++ b/results/classifier/zero-shot/118/all/1207686
@@ -0,0 +1,273 @@
+permissions: 0.972
+vnc: 0.959
+user-level: 0.951
+network: 0.949
+peripherals: 0.946
+debug: 0.942
+PID: 0.940
+semantic: 0.932
+register: 0.930
+socket: 0.930
+device: 0.928
+performance: 0.923
+architecture: 0.920
+KVM: 0.917
+boot: 0.916
+kernel: 0.897
+mistranslation: 0.892
+assembly: 0.889
+files: 0.878
+graphic: 0.870
+virtual: 0.867
+risc-v: 0.860
+hypervisor: 0.847
+ppc: 0.813
+VMM: 0.807
+arm: 0.794
+x86: 0.793
+TCG: 0.791
+i386: 0.414
+
+qemu-1.4.0 and onwards, linux kernel 3.2.x, heavy I/O leads to kernel_hung_tasks_timout_secs message and unresponsive qemu-process
+
+Hi,
+
+after some testing I tried to narrow down a problem, which was initially reported by some users.
+Seen on different distros - debian 7.1, ubuntu 12.04 LTS, IPFire-2.3 as reported by now.
+
+All using some flavour of linux-3.2.x kernel.
+
+Tried e.g. under Ubuntu an upgrade to "Linux 3.8.0-27-generic x86_64" which solves the problem.
+Problem could be triggert with some workload ala:
+
+spew -v --raw -P -t -i 3 -b 4k -p random -B 4k 1G /tmp/doof.dat
+and in parallel do some apt-get install/remove/whatever.
+
+That results in a somewhat stuck qemu-session with the bad "kernel_hung_task..." messages.
+
+A typical command-line is as follows:
+
+/usr/local/qemu-1.6.0/bin/qemu-system-x86_64 -usbdevice tablet -enable-kvm -daemonize -pidfile /var/run/qemu-server/760.pid -monitor unix:/var/run/qemu-server/760.mon,server,nowait -vnc unix:/var/run/qemu-server/760.vnc,password -qmp unix:/var/run/qemu-server/760.qmp,server,nowait -nodefaults -serial none -parallel none -device virtio-net-pci,mac=00:F1:70:00:2F:80,netdev=vlan0d0 -netdev type=tap,id=vlan0d0,ifname=tap760i0d0,script=/etc/fcms/add_if.sh,downscript=/etc/fcms/downscript.sh -name 1155823384-4 -m 512 -vga cirrus -k de -smp sockets=1,cores=1 -device virtio-blk-pci,drive=virtio0 -drive format=raw,file=rbd:1155823384/vm-760-disk-1.rbd:rbd_cache=false,cache=writeback,if=none,id=virtio0,media=disk,index=0,aio=native -drive format=raw,file=rbd:1155823384/vm-760-swap-1.rbd:rbd_cache=false,cache=writeback,if=virtio,media=disk,index=1,aio=native -drive if=ide,media=cdrom,id=ide1-cd0,readonly=on -drive if=ide,media=cdrom,id=ide1-cd1,readonly=on -boot order=dc
+
+no "system_reset", "sendkey ctrl-alt-delete" or "q" in monitoring-session is accepted, need to hard-kill the process.
+
+Please give any advice on what to do for tracing/debugging, because the number of tickets here are raising, and noone knows, what users are doing inside their VM.
+
+Kind regards,
+
+Oliver Francke.
+
+On Fri, Aug 02, 2013 at 09:58:29AM -0000, Oliver Francke wrote:
+> after some testing I tried to narrow down a problem, which was initially reported by some users.
+> Seen on different distros - debian 7.1, ubuntu 12.04 LTS, IPFire-2.3 as reported by now.
+>
+> All using some flavour of linux-3.2.x kernel.
+>
+> Tried e.g. under Ubuntu an upgrade to "Linux 3.8.0-27-generic x86_64" which solves the problem.
+
+Is that a guest kernel upgrade?
+
+> Problem could be triggert with some workload ala:
+>
+> spew -v --raw -P -t -i 3 -b 4k -p random -B 4k 1G /tmp/doof.dat
+> and in parallel do some apt-get install/remove/whatever.
+>
+> That results in a somewhat stuck qemu-session with the bad
+> "kernel_hung_task..." messages.
+>
+> A typical command-line is as follows:
+>
+> /usr/local/qemu-1.6.0/bin/qemu-system-x86_64 -usbdevice tablet -enable-
+> kvm -daemonize -pidfile /var/run/qemu-server/760.pid -monitor
+> unix:/var/run/qemu-server/760.mon,server,nowait -vnc unix:/var/run/qemu-
+> server/760.vnc,password -qmp unix:/var/run/qemu-
+> server/760.qmp,server,nowait -nodefaults -serial none -parallel none
+> -device virtio-net-pci,mac=00:F1:70:00:2F:80,netdev=vlan0d0 -netdev
+> type=tap,id=vlan0d0,ifname=tap760i0d0,script=/etc/fcms/add_if.sh,downscript=/etc/fcms/downscript.sh
+> -name 1155823384-4 -m 512 -vga cirrus -k de -smp sockets=1,cores=1
+> -device virtio-blk-pci,drive=virtio0 -drive
+> format=raw,file=rbd:1155823384/vm-760-disk-1.rbd:rbd_cache=false,cache=writeback,if=none,id=virtio0,media=disk,index=0,aio=native
+> -drive
+> format=raw,file=rbd:1155823384/vm-760-swap-1.rbd:rbd_cache=false,cache=writeback,if=virtio,media=disk,index=1,aio=native
+> -drive if=ide,media=cdrom,id=ide1-cd0,readonly=on -drive
+> if=ide,media=cdrom,id=ide1-cd1,readonly=on -boot order=dc
+>
+> no "system_reset", "sendkey ctrl-alt-delete" or "q" in monitoring-
+> session is accepted, need to hard-kill the process.
+
+Yesterday I saw a possibly related report on IRC. It was a Windows
+guest running under OpenStack with images on Ceph.
+
+They reported that the QEMU process would lock up - ping would not work
+and their management tools showed 0 CPU activity for the guest.
+However, they were able to "kick" the guest by taking a VNC screenshot
+(I think). Then it would come back to life.
+
+If you have a Linux guest that is reporting kernel_hung_task, then it
+could be a similar scenario.
+
+Please confirm that the hung task message is from inside the guest.
+
+If you are able to reproduce this and have an alternative non-Ceph
+storage pool, please try that since Ceph is common to both these bug
+reports.
+
+Stefan
+
+
+Hi Stefan,
+
+Am 02.08.2013 um 17:24 schrieb Stefan Hajnoczi <email address hidden>:
+
+> On Fri, Aug 02, 2013 at 09:58:29AM -0000, Oliver Francke wrote:
+>> after some testing I tried to narrow down a problem, which was initially reported by some users.
+>> Seen on different distros - debian 7.1, ubuntu 12.04 LTS, IPFire-2.3 as reported by now.
+>>
+>> All using some flavour of linux-3.2.x kernel.
+>>
+>> Tried e.g. under Ubuntu an upgrade to "Linux 3.8.0-27-generic x86_64" which solves the problem.
+>
+> Is that a guest kernel upgrade?
+
+yeah, sorry if that was not clear enough.
+
+>
+>> Problem could be triggert with some workload ala:
+>>
+>> spew -v --raw -P -t -i 3 -b 4k -p random -B 4k 1G /tmp/doof.dat
+>> and in parallel do some apt-get install/remove/whatever.
+>>
+>> That results in a somewhat stuck qemu-session with the bad
+>> "kernel_hung_task..." messages.
+>>
+>> A typical command-line is as follows:
+>>
+>> /usr/local/qemu-1.6.0/bin/qemu-system-x86_64 -usbdevice tablet -enable-
+>> kvm -daemonize -pidfile /var/run/qemu-server/760.pid -monitor
+>> unix:/var/run/qemu-server/760.mon,server,nowait -vnc unix:/var/run/qemu-
+>> server/760.vnc,password -qmp unix:/var/run/qemu-
+>> server/760.qmp,server,nowait -nodefaults -serial none -parallel none
+>> -device virtio-net-pci,mac=00:F1:70:00:2F:80,netdev=vlan0d0 -netdev
+>> type=tap,id=vlan0d0,ifname=tap760i0d0,script=/etc/fcms/add_if.sh,downscript=/etc/fcms/downscript.sh
+>> -name 1155823384-4 -m 512 -vga cirrus -k de -smp sockets=1,cores=1
+>> -device virtio-blk-pci,drive=virtio0 -drive
+>> format=raw,file=rbd:1155823384/vm-760-disk-1.rbd:rbd_cache=false,cache=writeback,if=none,id=virtio0,media=disk,index=0,aio=native
+>> -drive
+>> format=raw,file=rbd:1155823384/vm-760-swap-1.rbd:rbd_cache=false,cache=writeback,if=virtio,media=disk,index=1,aio=native
+>> -drive if=ide,media=cdrom,id=ide1-cd0,readonly=on -drive
+>> if=ide,media=cdrom,id=ide1-cd1,readonly=on -boot order=dc
+>>
+>> no "system_reset", "sendkey ctrl-alt-delete" or "q" in monitoring-
+>> session is accepted, need to hard-kill the process.
+>
+> Yesterday I saw a possibly related report on IRC. It was a Windows
+> guest running under OpenStack with images on Ceph.
+>
+> They reported that the QEMU process would lock up - ping would not work
+> and their management tools showed 0 CPU activity for the guest.
+> However, they were able to "kick" the guest by taking a VNC screenshot
+> (I think). Then it would come back to life.
+>
+> If you have a Linux guest that is reporting kernel_hung_task, then it
+> could be a similar scenario.
+>
+> Please confirm that the hung task message is from inside the guest.
+>
+
+confirmed.
+
+> If you are able to reproduce this and have an alternative non-Ceph
+> storage pool, please try that since Ceph is common to both these bug
+> reports.
+>
+
+I can reproduce it with: kernel 3.2.something + qemu-1.[456] ( never spent much time on 1.3) and high I/O.
+I took this VM later this day and converted it to local-storage-qcow2, no prob with any kernel. I already asked on ceph-users-list for assistance, especially from Josh ( if he's not on summer holiday ;) )
+
+What is strange, I have a session via VNC-console opened and have a loop ala:
+
+while true; do apt-get install -y ntp libopts25; apt-get remove -y ntp-libopts25; done
+and and parallel spew as described, the apt-"session" dies and one can see the hung_task-thingy, but I still can restart the spew-test.
+Just for completeness.
+
+Thnx for you attention,
+
+Oliver.
+
+> Stefan
+>
+> --
+> You received this bug notification because you are subscribed to the bug
+> report.
+> https://bugs.launchpad.net/bugs/1207686
+>
+> Title:
+> qemu-1.4.0 and onwards, linux kernel 3.2.x, heavy I/O leads to
+> kernel_hung_tasks_timout_secs message and unresponsive qemu-process
+>
+> Status in QEMU:
+> New
+>
+> Bug description:
+> Hi,
+>
+> after some testing I tried to narrow down a problem, which was initially reported by some users.
+> Seen on different distros - debian 7.1, ubuntu 12.04 LTS, IPFire-2.3 as reported by now.
+>
+> All using some flavour of linux-3.2.x kernel.
+>
+> Tried e.g. under Ubuntu an upgrade to "Linux 3.8.0-27-generic x86_64" which solves the problem.
+> Problem could be triggert with some workload ala:
+>
+> spew -v --raw -P -t -i 3 -b 4k -p random -B 4k 1G /tmp/doof.dat
+> and in parallel do some apt-get install/remove/whatever.
+>
+> That results in a somewhat stuck qemu-session with the bad
+> "kernel_hung_task..." messages.
+>
+> A typical command-line is as follows:
+>
+> /usr/local/qemu-1.6.0/bin/qemu-system-x86_64 -usbdevice tablet
+> -enable-kvm -daemonize -pidfile /var/run/qemu-server/760.pid -monitor
+> unix:/var/run/qemu-server/760.mon,server,nowait -vnc unix:/var/run
+> /qemu-server/760.vnc,password -qmp unix:/var/run/qemu-
+> server/760.qmp,server,nowait -nodefaults -serial none -parallel none
+> -device virtio-net-pci,mac=00:F1:70:00:2F:80,netdev=vlan0d0 -netdev
+> type=tap,id=vlan0d0,ifname=tap760i0d0,script=/etc/fcms/add_if.sh,downscript=/etc/fcms/downscript.sh
+> -name 1155823384-4 -m 512 -vga cirrus -k de -smp sockets=1,cores=1
+> -device virtio-blk-pci,drive=virtio0 -drive
+> format=raw,file=rbd:1155823384/vm-760-disk-1.rbd:rbd_cache=false,cache=writeback,if=none,id=virtio0,media=disk,index=0,aio=native
+> -drive
+> format=raw,file=rbd:1155823384/vm-760-swap-1.rbd:rbd_cache=false,cache=writeback,if=virtio,media=disk,index=1,aio=native
+> -drive if=ide,media=cdrom,id=ide1-cd0,readonly=on -drive
+> if=ide,media=cdrom,id=ide1-cd1,readonly=on -boot order=dc
+>
+> no "system_reset", "sendkey ctrl-alt-delete" or "q" in monitoring-
+> session is accepted, need to hard-kill the process.
+>
+> Please give any advice on what to do for tracing/debugging, because
+> the number of tickets here are raising, and noone knows, what users
+> are doing inside their VM.
+>
+> Kind regards,
+>
+> Oliver Francke.
+>
+> To manage notifications about this bug go to:
+> https://bugs.launchpad.net/qemu/+bug/1207686/+subscriptions
+
+
+
+Hi,
+
+opened a ticket with the ceph-guys, and it turned out to be a bug in "librados aio flush".
+
+With latest "wip-librados-aio-flush (bobtail)" I got no error even with _very_ high load.
+
+Thnx for the attention ;)
+
+Oliver.
+
+
+Closing as "Invalid" since this was not a QEMU bug according to comment #3.
+