summary refs log tree commit diff stats
path: root/gitlab/issues/target_missing/host_missing/accel_missing/1889.toml
diff options
context:
space:
mode:
authorChristian Krinitsin <mail@krinitsin.com>2025-05-30 16:52:07 +0200
committerChristian Krinitsin <mail@krinitsin.com>2025-05-30 16:52:17 +0200
commit9260319e7411ff8281700a532caa436f40120ec4 (patch)
tree2f6bfe5f3458dd49d328d3a9eb508595450adec0 /gitlab/issues/target_missing/host_missing/accel_missing/1889.toml
parent225caa38269323af1bfc2daadff5ec8bd930747f (diff)
downloademulator-bug-study-9260319e7411ff8281700a532caa436f40120ec4.tar.gz
emulator-bug-study-9260319e7411ff8281700a532caa436f40120ec4.zip
gitlab scraper: download in toml and text format
Diffstat (limited to 'gitlab/issues/target_missing/host_missing/accel_missing/1889.toml')
-rw-r--r--gitlab/issues/target_missing/host_missing/accel_missing/1889.toml55
1 files changed, 0 insertions, 55 deletions
diff --git a/gitlab/issues/target_missing/host_missing/accel_missing/1889.toml b/gitlab/issues/target_missing/host_missing/accel_missing/1889.toml
deleted file mode 100644
index 935b21ab..00000000
--- a/gitlab/issues/target_missing/host_missing/accel_missing/1889.toml
+++ /dev/null
@@ -1,55 +0,0 @@
-id = 1889
-title = "IO delays on live migration lv initialization"
-state = "opened"
-created_at = "2023-09-14T08:19:44.547Z"
-closed_at = "n/a"
-labels = []
-url = "https://gitlab.com/qemu-project/qemu/-/issues/1889"
-host-os = "Debian 11"
-host-arch = "x86"
-qemu-version = "QEMU emulator version 7.2.0 (pve-qemu-kvm_7.2.0-8)"
-guest-os = "Debian 11"
-guest-arch = "x86"
-description = """Hi,
-
-When I live migrate a VM via Proxmox and the destination is an LVM thin pool I see that at the start of copying the disk it's first initialized.
-
-This leads the thin volume to be directly 100% allocated which needs to be discarded afterwards. Not ideal but ....
-
-The more annoying thing is that this initialization step used 100% of disk IO. In iotop I see it writing over 1000MB/sec. The nasty side effect is that other VM's on that host are negatively affected. It's not completely locked up, I can ssh in and look around, but storage intensive things see more delay. With e.g. http requests timing out. And even a simple ls command could take 10+ seconds which is normally instant.
-
-
-I've previously reported it on the [proxmox forum](https://forum.proxmox.com/threads/io-delays-on-live-migration-lv-initialization.132296/#post-582050) but the call was made that this is behavior from Qemu.
-
-> The zeroing happens, because of what QEMU does when the discard option is enabled:
-
-
-When I disable discard for the VM disk I can see that it's not pre-initialized during migration, but not having that defeats the purpose of having an lvm thin pool.
-
-For the (disk) migration itself I can set a bandwidth limit ... could we do something similar for initialization?
-
-
-Even better would be to not initialize at all when using LVM thin. As far as I understand it the new blocks allocated by lvm thin should always be empty."""
-reproduce = """1. Migrate a vm with a large disk
-2. look in iotop on the new host, would be see more write IO then the network could handle.. just before the disk content is transferred.
-3. look in another VM on the destination host, reading from disk would be significantly slower then normal."""
-additional = """An example VM config
-```
-agent: 1,fstrim_cloned_disks=1
-balloon: 512
-bootdisk: scsi0
-cores: 6
-ide2: none,media=cdrom
-memory: 8196
-name: ...
-net0: virtio=...,bridge=...
-numa: 0
-onboot: 1
-ostype: l26
-scsi0: thin_pool_hwraid:vm-301-disk-0,discard=on,format=raw,size=16192M
-scsi1: thin_pool_hwraid:vm-301-disk-1,discard=on,format=raw,size=26G
-scsihw: virtio-scsi-pci
-serial0: socket
-smbios1: uuid=...
-sockets: 1
-```"""