From cddf7dfa5a6e92a9057e7f6afae5d9b970585e6f Mon Sep 17 00:00:00 2001 From: Christian Krinitsin Date: Sun, 1 Jun 2025 12:57:38 +0000 Subject: change mail-ids --- mailinglist/downloader.py | 4 +- mailinglist/output_mailinglist/0001467 | 95 - mailinglist/output_mailinglist/02364653 | 358 ++ mailinglist/output_mailinglist/0247400 | 1481 ------ mailinglist/output_mailinglist/02572177 | 416 ++ mailinglist/output_mailinglist/04472277 | 571 +++ mailinglist/output_mailinglist/0504199 | 590 --- mailinglist/output_mailinglist/05479587 | 78 + mailinglist/output_mailinglist/0804350 | 7443 ------------------------------- mailinglist/output_mailinglist/0891566 | 395 -- mailinglist/output_mailinglist/0966902 | 34 - mailinglist/output_mailinglist/1067127 | 149 - mailinglist/output_mailinglist/11357571 | 42 + mailinglist/output_mailinglist/11933524 | 1120 +++++ mailinglist/output_mailinglist/1195866 | 237 - mailinglist/output_mailinglist/12360755 | 291 ++ mailinglist/output_mailinglist/1267916 | 1873 -------- mailinglist/output_mailinglist/12869209 | 83 + mailinglist/output_mailinglist/13442371 | 364 ++ mailinglist/output_mailinglist/1398669 | 780 ---- mailinglist/output_mailinglist/1412913 | 2895 ------------ mailinglist/output_mailinglist/14488057 | 706 +++ mailinglist/output_mailinglist/1452608 | 73 - mailinglist/output_mailinglist/14887122 | 253 ++ mailinglist/output_mailinglist/16056596 | 93 + mailinglist/output_mailinglist/16201167 | 95 + mailinglist/output_mailinglist/16228234 | 1839 ++++++++ mailinglist/output_mailinglist/1693040 | 1056 ----- mailinglist/output_mailinglist/17743720 | 766 ++++ mailinglist/output_mailinglist/2047990 | 994 ----- mailinglist/output_mailinglist/21221931 | 323 ++ mailinglist/output_mailinglist/21247035 | 1316 ++++++ mailinglist/output_mailinglist/22219210 | 38 + mailinglist/output_mailinglist/2308923 | 230 - mailinglist/output_mailinglist/23270873 | 687 +++ mailinglist/output_mailinglist/23300761 | 308 ++ mailinglist/output_mailinglist/23448582 | 260 ++ mailinglist/output_mailinglist/2393649 | 339 -- mailinglist/output_mailinglist/2409210 | 413 -- mailinglist/output_mailinglist/24190340 | 2051 +++++++++ mailinglist/output_mailinglist/24930826 | 28 + mailinglist/output_mailinglist/2537817 | 527 --- mailinglist/output_mailinglist/2562302 | 144 - mailinglist/output_mailinglist/25842545 | 197 + mailinglist/output_mailinglist/25892827 | 1072 +++++ mailinglist/output_mailinglist/26095107 | 153 + mailinglist/output_mailinglist/2609717 | 4934 -------------------- mailinglist/output_mailinglist/26430026 | 160 + mailinglist/output_mailinglist/28596630 | 108 + mailinglist/output_mailinglist/2880487 | 182 - mailinglist/output_mailinglist/30680944 | 590 +++ mailinglist/output_mailinglist/31349848 | 149 + mailinglist/output_mailinglist/3223447 | 194 - mailinglist/output_mailinglist/3242247 | 401 -- mailinglist/output_mailinglist/32484936 | 218 + mailinglist/output_mailinglist/33802194 | 4934 ++++++++++++++++++++ mailinglist/output_mailinglist/3457423 | 35 - mailinglist/output_mailinglist/3501174 | 2788 ------------ mailinglist/output_mailinglist/35170175 | 516 +++ mailinglist/output_mailinglist/36568044 | 4576 +++++++++++++++++++ mailinglist/output_mailinglist/3749377 | 358 -- mailinglist/output_mailinglist/3825088 | 516 --- mailinglist/output_mailinglist/3847403 | 78 - mailinglist/output_mailinglist/3886413 | 28 - mailinglist/output_mailinglist/4158985 | 1475 ------ mailinglist/output_mailinglist/42226390 | 182 + mailinglist/output_mailinglist/42613410 | 144 + mailinglist/output_mailinglist/42974450 | 424 ++ mailinglist/output_mailinglist/4314117 | 706 --- mailinglist/output_mailinglist/43643137 | 533 +++ mailinglist/output_mailinglist/4412535 | 343 -- mailinglist/output_mailinglist/46572227 | 401 ++ mailinglist/output_mailinglist/4774720 | 323 -- mailinglist/output_mailinglist/4800759 | 364 -- mailinglist/output_mailinglist/48245039 | 525 +++ mailinglist/output_mailinglist/4938208 | 1839 -------- mailinglist/output_mailinglist/4970412 | 83 - mailinglist/output_mailinglist/5057521 | 766 ---- mailinglist/output_mailinglist/50773216 | 105 + mailinglist/output_mailinglist/51610399 | 303 ++ mailinglist/output_mailinglist/5215275 | 533 --- mailinglist/output_mailinglist/5321072 | 416 -- mailinglist/output_mailinglist/53568181 | 73 + mailinglist/output_mailinglist/5362491 | 93 - mailinglist/output_mailinglist/5373318 | 687 --- mailinglist/output_mailinglist/5396868 | 218 - mailinglist/output_mailinglist/5443005 | 571 --- mailinglist/output_mailinglist/55247116 | 1305 ++++++ mailinglist/output_mailinglist/55367348 | 527 +++ mailinglist/output_mailinglist/55753058 | 288 ++ mailinglist/output_mailinglist/55961334 | 34 + mailinglist/output_mailinglist/56309929 | 175 + mailinglist/output_mailinglist/56937788 | 339 ++ mailinglist/output_mailinglist/57195159 | 310 ++ mailinglist/output_mailinglist/57231878 | 237 + mailinglist/output_mailinglist/5745618 | 150 - mailinglist/output_mailinglist/57756589 | 1416 ++++++ mailinglist/output_mailinglist/5798945 | 38 - mailinglist/output_mailinglist/5843372 | 2051 --------- mailinglist/output_mailinglist/5912779 | 310 -- mailinglist/output_mailinglist/5933279 | 4576 ------------------- mailinglist/output_mailinglist/59540920 | 371 ++ mailinglist/output_mailinglist/60339453 | 56 + mailinglist/output_mailinglist/6117378 | 26 - mailinglist/output_mailinglist/6156219 | 1416 ------ mailinglist/output_mailinglist/6178292 | 253 -- mailinglist/output_mailinglist/62179944 | 26 + mailinglist/output_mailinglist/6257722 | 711 --- mailinglist/output_mailinglist/6355518 | 525 --- mailinglist/output_mailinglist/63565653 | 44 + mailinglist/output_mailinglist/6416205 | 56 - mailinglist/output_mailinglist/64322995 | 49 + mailinglist/output_mailinglist/64571620 | 780 ++++ mailinglist/output_mailinglist/6531392 | 288 -- mailinglist/output_mailinglist/65781993 | 2788 ++++++++++++ mailinglist/output_mailinglist/66743673 | 359 ++ mailinglist/output_mailinglist/6739993 | 260 -- mailinglist/output_mailinglist/67821138 | 194 + mailinglist/output_mailinglist/6866700 | 49 - mailinglist/output_mailinglist/68897003 | 711 +++ mailinglist/output_mailinglist/6983580 | 424 -- mailinglist/output_mailinglist/6998781 | 1072 ----- mailinglist/output_mailinglist/70021271 | 7443 +++++++++++++++++++++++++++++++ mailinglist/output_mailinglist/70294255 | 1056 +++++ mailinglist/output_mailinglist/70416488 | 1174 +++++ mailinglist/output_mailinglist/70868267 | 35 + mailinglist/output_mailinglist/7143139 | 121 - mailinglist/output_mailinglist/71456293 | 1481 ++++++ mailinglist/output_mailinglist/73660729 | 26 + mailinglist/output_mailinglist/7427991 | 308 -- mailinglist/output_mailinglist/74466963 | 1873 ++++++++ mailinglist/output_mailinglist/74545755 | 339 ++ mailinglist/output_mailinglist/74715356 | 121 + mailinglist/output_mailinglist/7639274 | 1305 ------ mailinglist/output_mailinglist/7647456 | 105 - mailinglist/output_mailinglist/7658242 | 1120 ----- mailinglist/output_mailinglist/7711787 | 160 - mailinglist/output_mailinglist/7733130 | 42 - mailinglist/output_mailinglist/7837801 | 108 - mailinglist/output_mailinglist/7960594 | 153 - mailinglist/output_mailinglist/79834768 | 404 ++ mailinglist/output_mailinglist/8019995 | 26 - mailinglist/output_mailinglist/80570214 | 395 ++ mailinglist/output_mailinglist/80604314 | 1475 ++++++ mailinglist/output_mailinglist/80615920 | 343 ++ mailinglist/output_mailinglist/8109943 | 404 -- mailinglist/output_mailinglist/81775929 | 230 + mailinglist/output_mailinglist/8511484 | 291 -- mailinglist/output_mailinglist/85542195 | 115 + mailinglist/output_mailinglist/8566429 | 44 - mailinglist/output_mailinglist/8621822 | 371 -- mailinglist/output_mailinglist/8627146 | 359 -- mailinglist/output_mailinglist/8653736 | 115 - mailinglist/output_mailinglist/8691137 | 175 - mailinglist/output_mailinglist/8720260 | 339 -- mailinglist/output_mailinglist/88225572 | 2895 ++++++++++++ mailinglist/output_mailinglist/88281850 | 276 ++ mailinglist/output_mailinglist/8874178 | 197 - mailinglist/output_mailinglist/92957605 | 413 ++ mailinglist/output_mailinglist/95154278 | 150 + mailinglist/output_mailinglist/96782458 | 994 +++++ mailinglist/output_mailinglist/9777608 | 276 -- mailinglist/output_mailinglist/9818783 | 303 -- mailinglist/output_mailinglist/9840852 | 1174 ----- mailinglist/output_mailinglist/9937102 | 143 - mailinglist/output_mailinglist/9948366 | 1316 ------ mailinglist/output_mailinglist/99674399 | 143 + 167 files changed, 58841 insertions(+), 58841 deletions(-) delete mode 100644 mailinglist/output_mailinglist/0001467 create mode 100644 mailinglist/output_mailinglist/02364653 delete mode 100644 mailinglist/output_mailinglist/0247400 create mode 100644 mailinglist/output_mailinglist/02572177 create mode 100644 mailinglist/output_mailinglist/04472277 delete mode 100644 mailinglist/output_mailinglist/0504199 create mode 100644 mailinglist/output_mailinglist/05479587 delete mode 100644 mailinglist/output_mailinglist/0804350 delete mode 100644 mailinglist/output_mailinglist/0891566 delete mode 100644 mailinglist/output_mailinglist/0966902 delete mode 100644 mailinglist/output_mailinglist/1067127 create mode 100644 mailinglist/output_mailinglist/11357571 create mode 100644 mailinglist/output_mailinglist/11933524 delete mode 100644 mailinglist/output_mailinglist/1195866 create mode 100644 mailinglist/output_mailinglist/12360755 delete mode 100644 mailinglist/output_mailinglist/1267916 create mode 100644 mailinglist/output_mailinglist/12869209 create mode 100644 mailinglist/output_mailinglist/13442371 delete mode 100644 mailinglist/output_mailinglist/1398669 delete mode 100644 mailinglist/output_mailinglist/1412913 create mode 100644 mailinglist/output_mailinglist/14488057 delete mode 100644 mailinglist/output_mailinglist/1452608 create mode 100644 mailinglist/output_mailinglist/14887122 create mode 100644 mailinglist/output_mailinglist/16056596 create mode 100644 mailinglist/output_mailinglist/16201167 create mode 100644 mailinglist/output_mailinglist/16228234 delete mode 100644 mailinglist/output_mailinglist/1693040 create mode 100644 mailinglist/output_mailinglist/17743720 delete mode 100644 mailinglist/output_mailinglist/2047990 create mode 100644 mailinglist/output_mailinglist/21221931 create mode 100644 mailinglist/output_mailinglist/21247035 create mode 100644 mailinglist/output_mailinglist/22219210 delete mode 100644 mailinglist/output_mailinglist/2308923 create mode 100644 mailinglist/output_mailinglist/23270873 create mode 100644 mailinglist/output_mailinglist/23300761 create mode 100644 mailinglist/output_mailinglist/23448582 delete mode 100644 mailinglist/output_mailinglist/2393649 delete mode 100644 mailinglist/output_mailinglist/2409210 create mode 100644 mailinglist/output_mailinglist/24190340 create mode 100644 mailinglist/output_mailinglist/24930826 delete mode 100644 mailinglist/output_mailinglist/2537817 delete mode 100644 mailinglist/output_mailinglist/2562302 create mode 100644 mailinglist/output_mailinglist/25842545 create mode 100644 mailinglist/output_mailinglist/25892827 create mode 100644 mailinglist/output_mailinglist/26095107 delete mode 100644 mailinglist/output_mailinglist/2609717 create mode 100644 mailinglist/output_mailinglist/26430026 create mode 100644 mailinglist/output_mailinglist/28596630 delete mode 100644 mailinglist/output_mailinglist/2880487 create mode 100644 mailinglist/output_mailinglist/30680944 create mode 100644 mailinglist/output_mailinglist/31349848 delete mode 100644 mailinglist/output_mailinglist/3223447 delete mode 100644 mailinglist/output_mailinglist/3242247 create mode 100644 mailinglist/output_mailinglist/32484936 create mode 100644 mailinglist/output_mailinglist/33802194 delete mode 100644 mailinglist/output_mailinglist/3457423 delete mode 100644 mailinglist/output_mailinglist/3501174 create mode 100644 mailinglist/output_mailinglist/35170175 create mode 100644 mailinglist/output_mailinglist/36568044 delete mode 100644 mailinglist/output_mailinglist/3749377 delete mode 100644 mailinglist/output_mailinglist/3825088 delete mode 100644 mailinglist/output_mailinglist/3847403 delete mode 100644 mailinglist/output_mailinglist/3886413 delete mode 100644 mailinglist/output_mailinglist/4158985 create mode 100644 mailinglist/output_mailinglist/42226390 create mode 100644 mailinglist/output_mailinglist/42613410 create mode 100644 mailinglist/output_mailinglist/42974450 delete mode 100644 mailinglist/output_mailinglist/4314117 create mode 100644 mailinglist/output_mailinglist/43643137 delete mode 100644 mailinglist/output_mailinglist/4412535 create mode 100644 mailinglist/output_mailinglist/46572227 delete mode 100644 mailinglist/output_mailinglist/4774720 delete mode 100644 mailinglist/output_mailinglist/4800759 create mode 100644 mailinglist/output_mailinglist/48245039 delete mode 100644 mailinglist/output_mailinglist/4938208 delete mode 100644 mailinglist/output_mailinglist/4970412 delete mode 100644 mailinglist/output_mailinglist/5057521 create mode 100644 mailinglist/output_mailinglist/50773216 create mode 100644 mailinglist/output_mailinglist/51610399 delete mode 100644 mailinglist/output_mailinglist/5215275 delete mode 100644 mailinglist/output_mailinglist/5321072 create mode 100644 mailinglist/output_mailinglist/53568181 delete mode 100644 mailinglist/output_mailinglist/5362491 delete mode 100644 mailinglist/output_mailinglist/5373318 delete mode 100644 mailinglist/output_mailinglist/5396868 delete mode 100644 mailinglist/output_mailinglist/5443005 create mode 100644 mailinglist/output_mailinglist/55247116 create mode 100644 mailinglist/output_mailinglist/55367348 create mode 100644 mailinglist/output_mailinglist/55753058 create mode 100644 mailinglist/output_mailinglist/55961334 create mode 100644 mailinglist/output_mailinglist/56309929 create mode 100644 mailinglist/output_mailinglist/56937788 create mode 100644 mailinglist/output_mailinglist/57195159 create mode 100644 mailinglist/output_mailinglist/57231878 delete mode 100644 mailinglist/output_mailinglist/5745618 create mode 100644 mailinglist/output_mailinglist/57756589 delete mode 100644 mailinglist/output_mailinglist/5798945 delete mode 100644 mailinglist/output_mailinglist/5843372 delete mode 100644 mailinglist/output_mailinglist/5912779 delete mode 100644 mailinglist/output_mailinglist/5933279 create mode 100644 mailinglist/output_mailinglist/59540920 create mode 100644 mailinglist/output_mailinglist/60339453 delete mode 100644 mailinglist/output_mailinglist/6117378 delete mode 100644 mailinglist/output_mailinglist/6156219 delete mode 100644 mailinglist/output_mailinglist/6178292 create mode 100644 mailinglist/output_mailinglist/62179944 delete mode 100644 mailinglist/output_mailinglist/6257722 delete mode 100644 mailinglist/output_mailinglist/6355518 create mode 100644 mailinglist/output_mailinglist/63565653 delete mode 100644 mailinglist/output_mailinglist/6416205 create mode 100644 mailinglist/output_mailinglist/64322995 create mode 100644 mailinglist/output_mailinglist/64571620 delete mode 100644 mailinglist/output_mailinglist/6531392 create mode 100644 mailinglist/output_mailinglist/65781993 create mode 100644 mailinglist/output_mailinglist/66743673 delete mode 100644 mailinglist/output_mailinglist/6739993 create mode 100644 mailinglist/output_mailinglist/67821138 delete mode 100644 mailinglist/output_mailinglist/6866700 create mode 100644 mailinglist/output_mailinglist/68897003 delete mode 100644 mailinglist/output_mailinglist/6983580 delete mode 100644 mailinglist/output_mailinglist/6998781 create mode 100644 mailinglist/output_mailinglist/70021271 create mode 100644 mailinglist/output_mailinglist/70294255 create mode 100644 mailinglist/output_mailinglist/70416488 create mode 100644 mailinglist/output_mailinglist/70868267 delete mode 100644 mailinglist/output_mailinglist/7143139 create mode 100644 mailinglist/output_mailinglist/71456293 create mode 100644 mailinglist/output_mailinglist/73660729 delete mode 100644 mailinglist/output_mailinglist/7427991 create mode 100644 mailinglist/output_mailinglist/74466963 create mode 100644 mailinglist/output_mailinglist/74545755 create mode 100644 mailinglist/output_mailinglist/74715356 delete mode 100644 mailinglist/output_mailinglist/7639274 delete mode 100644 mailinglist/output_mailinglist/7647456 delete mode 100644 mailinglist/output_mailinglist/7658242 delete mode 100644 mailinglist/output_mailinglist/7711787 delete mode 100644 mailinglist/output_mailinglist/7733130 delete mode 100644 mailinglist/output_mailinglist/7837801 delete mode 100644 mailinglist/output_mailinglist/7960594 create mode 100644 mailinglist/output_mailinglist/79834768 delete mode 100644 mailinglist/output_mailinglist/8019995 create mode 100644 mailinglist/output_mailinglist/80570214 create mode 100644 mailinglist/output_mailinglist/80604314 create mode 100644 mailinglist/output_mailinglist/80615920 delete mode 100644 mailinglist/output_mailinglist/8109943 create mode 100644 mailinglist/output_mailinglist/81775929 delete mode 100644 mailinglist/output_mailinglist/8511484 create mode 100644 mailinglist/output_mailinglist/85542195 delete mode 100644 mailinglist/output_mailinglist/8566429 delete mode 100644 mailinglist/output_mailinglist/8621822 delete mode 100644 mailinglist/output_mailinglist/8627146 delete mode 100644 mailinglist/output_mailinglist/8653736 delete mode 100644 mailinglist/output_mailinglist/8691137 delete mode 100644 mailinglist/output_mailinglist/8720260 create mode 100644 mailinglist/output_mailinglist/88225572 create mode 100644 mailinglist/output_mailinglist/88281850 delete mode 100644 mailinglist/output_mailinglist/8874178 create mode 100644 mailinglist/output_mailinglist/92957605 create mode 100644 mailinglist/output_mailinglist/95154278 create mode 100644 mailinglist/output_mailinglist/96782458 delete mode 100644 mailinglist/output_mailinglist/9777608 delete mode 100644 mailinglist/output_mailinglist/9818783 delete mode 100644 mailinglist/output_mailinglist/9840852 delete mode 100644 mailinglist/output_mailinglist/9937102 delete mode 100644 mailinglist/output_mailinglist/9948366 create mode 100644 mailinglist/output_mailinglist/99674399 (limited to 'mailinglist') diff --git a/mailinglist/downloader.py b/mailinglist/downloader.py index ebe4d2d7..12673e7c 100755 --- a/mailinglist/downloader.py +++ b/mailinglist/downloader.py @@ -63,13 +63,13 @@ def main(): # existing thread re_match = match(r'(?i)^re:\s*(.*)', text) # matches 'Re:' if re_match: - title_hash = str(hash(re_match.group(1).strip()))[1:8] + title_hash = str(hash(re_match.group(1).strip()))[1:9] if path.exists(f"output_mailinglist/{title_hash}"): process_thread(urljoin(url, href), title_hash) continue # new thread - title_hash = str(hash(text.strip()))[1:8] + title_hash = str(hash(text.strip()))[1:9] if path.exists(f"output_mailinglist/{title_hash}"): print(f"ERROR: {title_hash} should not exist!") continue diff --git a/mailinglist/output_mailinglist/0001467 b/mailinglist/output_mailinglist/0001467 deleted file mode 100644 index 0bd1e609..00000000 --- a/mailinglist/output_mailinglist/0001467 +++ /dev/null @@ -1,95 +0,0 @@ -[BUG] Qemu abort with error "kvm_mem_ioeventfd_add: error adding ioeventfd: File exists (17)" - -Hi list, - -When I did some tests in my virtual domain with live-attached virtio deivces, I -got a coredump file of Qemu. - -The error print from qemu is "kvm_mem_ioeventfd_add: error adding ioeventfd: -File exists (17)". -And the call trace in the coredump file displays as below: -#0 0x0000ffff89acecc8 in ?? () from /usr/lib64/libc.so.6 -#1 0x0000ffff89a8acbc in raise () from /usr/lib64/libc.so.6 -#2 0x0000ffff89a78d2c in abort () from /usr/lib64/libc.so.6 -#3 0x0000aaaabd7ccf1c in kvm_mem_ioeventfd_add (listener=, -section=, match_data=, data=, -e=) at ../accel/kvm/kvm-all.c:1607 -#4 0x0000aaaabd6e0304 in address_space_add_del_ioeventfds (fds_old_nb=164, -fds_old=0xffff5c80a1d0, fds_new_nb=160, fds_new=0xffff5c565080, -as=0xaaaabdfa8810 ) - at ../softmmu/memory.c:795 -#5 address_space_update_ioeventfds (as=0xaaaabdfa8810 ) -at ../softmmu/memory.c:856 -#6 0x0000aaaabd6e24d8 in memory_region_commit () at ../softmmu/memory.c:1113 -#7 0x0000aaaabd6e25c4 in memory_region_transaction_commit () at -../softmmu/memory.c:1144 -#8 0x0000aaaabd394eb4 in pci_bridge_update_mappings -(br=br@entry=0xaaaae755f7c0) at ../hw/pci/pci_bridge.c:248 -#9 0x0000aaaabd394f4c in pci_bridge_write_config (d=0xaaaae755f7c0, -address=44, val=, len=4) at ../hw/pci/pci_bridge.c:272 -#10 0x0000aaaabd39a928 in rp_write_config (d=0xaaaae755f7c0, address=44, -val=128, len=4) at ../hw/pci-bridge/pcie_root_port.c:39 -#11 0x0000aaaabd6df328 in memory_region_write_accessor (mr=0xaaaae63898d0, -addr=65580, value=, size=4, shift=, -mask=, attrs=...) at ../softmmu/memory.c:494 -#12 0x0000aaaabd6dcb6c in access_with_adjusted_size (addr=addr@entry=65580, -value=value@entry=0xffff817adc78, size=size@entry=4, access_size_min=, access_size_max=, - access_fn=access_fn@entry=0xaaaabd6df284 , -mr=mr@entry=0xaaaae63898d0, attrs=attrs@entry=...) at ../softmmu/memory.c:556 -#13 0x0000aaaabd6e0dc8 in memory_region_dispatch_write -(mr=mr@entry=0xaaaae63898d0, addr=65580, data=, op=MO_32, -attrs=attrs@entry=...) at ../softmmu/memory.c:1534 -#14 0x0000aaaabd6d0574 in flatview_write_continue (fv=fv@entry=0xffff5c02da00, -addr=addr@entry=275146407980, attrs=attrs@entry=..., -ptr=ptr@entry=0xffff8aa8c028, len=len@entry=4, - addr1=, l=, mr=mr@entry=0xaaaae63898d0) at -/usr/src/debug/qemu-6.2.0-226.aarch64/include/qemu/host-utils.h:165 -#15 0x0000aaaabd6d4584 in flatview_write (len=4, buf=0xffff8aa8c028, attrs=..., -addr=275146407980, fv=0xffff5c02da00) at ../softmmu/physmem.c:3375 -#16 address_space_write (as=, addr=275146407980, attrs=..., -buf=buf@entry=0xffff8aa8c028, len=4) at ../softmmu/physmem.c:3467 -#17 0x0000aaaabd6d462c in address_space_rw (as=, addr=, attrs=..., attrs@entry=..., buf=buf@entry=0xffff8aa8c028, len=, is_write=) - at ../softmmu/physmem.c:3477 -#18 0x0000aaaabd7cf6e8 in kvm_cpu_exec (cpu=cpu@entry=0xaaaae625dfd0) at -../accel/kvm/kvm-all.c:2970 -#19 0x0000aaaabd7d09bc in kvm_vcpu_thread_fn (arg=arg@entry=0xaaaae625dfd0) at -../accel/kvm/kvm-accel-ops.c:49 -#20 0x0000aaaabd94ccd8 in qemu_thread_start (args=) at -../util/qemu-thread-posix.c:559 - - -By printing more info in the coredump file, I found that the addr of -fds_old[146] and fds_new[146] are same, but fds_old[146] belonged to a -live-attached virtio-scsi device while fds_new[146] was owned by another -live-attached virtio-net. -The reason why addr conflicted was then been found from vm's console log. Just -before qemu aborted, the guest kernel crashed and kdump.service booted the -dump-capture kernel where re-alloced address for the devices. -Because those virtio devices were live-attached after vm creating, different -addr may been assigned to them in the dump-capture kernel: - -the initial kernel booting log: -[ 1.663297] pci 0000:00:02.1: BAR 14: assigned [mem 0x11900000-0x11afffff] -[ 1.664560] pci 0000:00:02.1: BAR 15: assigned [mem -0x8001800000-0x80019fffff 64bit pref] - -the dump-capture kernel booting log: -[ 1.845211] pci 0000:00:02.0: BAR 14: assigned [mem 0x11900000-0x11bfffff] -[ 1.846542] pci 0000:00:02.0: BAR 15: assigned [mem -0x8001800000-0x8001afffff 64bit pref] - - -I think directly aborting the qemu process may not be the best choice in this -case cuz it will interrupt the work of kdump.service so that failed to generate -memory dump of the crashed guest kernel. -Perhaps, IMO, the error could be simply ignored in this case and just let kdump -to reboot the system after memory-dump finishing, but I failed to find a -suitable judgment in the codes. - -Any solution for this problem? Hope I can get some helps here. - -Hao - diff --git a/mailinglist/output_mailinglist/02364653 b/mailinglist/output_mailinglist/02364653 new file mode 100644 index 00000000..dd26adfd --- /dev/null +++ b/mailinglist/output_mailinglist/02364653 @@ -0,0 +1,358 @@ +[Qemu-devel] [BUG] Inappropriate size of target_sigset_t + +Hello, Peter, Laurent, + +While working on another problem yesterday, I think I discovered a +long-standing bug in QEMU Linux user mode: our target_sigset_t structure is +eight times smaller as it should be! + +In this code segment from syscalls_def.h: + +#ifdef TARGET_MIPS +#define TARGET_NSIG 128 +#else +#define TARGET_NSIG 64 +#endif +#define TARGET_NSIG_BPW TARGET_ABI_BITS +#define TARGET_NSIG_WORDS (TARGET_NSIG / TARGET_NSIG_BPW) + +typedef struct { + abi_ulong sig[TARGET_NSIG_WORDS]; +} target_sigset_t; + +... TARGET_ABI_BITS should be replaced by eight times smaller constant (in +fact, semantically, we need TARGET_ABI_BYTES, but it is not defined) (what is +needed is actually "a byte per signal" in target_sigset_t, and we allow "a bit +per signal"). + +All this probably sounds to you like something impossible, since this code is +in QEMU "since forever", but I checked everything, and the bug seems real. I +wish you can prove me wrong. + +I just wanted to let you know about this, given the sensitive timing of current +softfreeze, and the fact that I won't be able to do more investigation on this +in coming weeks, since I am busy with other tasks, but perhaps you can analyze +and do something which you consider appropriate. + +Yours, +Aleksandar + +Le 03/07/2019 à 21:46, Aleksandar Markovic a écrit : +> +Hello, Peter, Laurent, +> +> +While working on another problem yesterday, I think I discovered a +> +long-standing bug in QEMU Linux user mode: our target_sigset_t structure is +> +eight times smaller as it should be! +> +> +In this code segment from syscalls_def.h: +> +> +#ifdef TARGET_MIPS +> +#define TARGET_NSIG 128 +> +#else +> +#define TARGET_NSIG 64 +> +#endif +> +#define TARGET_NSIG_BPW TARGET_ABI_BITS +> +#define TARGET_NSIG_WORDS (TARGET_NSIG / TARGET_NSIG_BPW) +> +> +typedef struct { +> +abi_ulong sig[TARGET_NSIG_WORDS]; +> +} target_sigset_t; +> +> +... TARGET_ABI_BITS should be replaced by eight times smaller constant (in +> +fact, semantically, we need TARGET_ABI_BYTES, but it is not defined) (what is +> +needed is actually "a byte per signal" in target_sigset_t, and we allow "a +> +bit per signal"). +TARGET_NSIG is divided by TARGET_ABI_BITS which gives you the number of +abi_ulong words we need in target_sigset_t. + +> +All this probably sounds to you like something impossible, since this code is +> +in QEMU "since forever", but I checked everything, and the bug seems real. I +> +wish you can prove me wrong. +> +> +I just wanted to let you know about this, given the sensitive timing of +> +current softfreeze, and the fact that I won't be able to do more +> +investigation on this in coming weeks, since I am busy with other tasks, but +> +perhaps you can analyze and do something which you consider appropriate. +If I compare with kernel, it looks good: + +In Linux: + + arch/mips/include/uapi/asm/signal.h + + #define _NSIG 128 + #define _NSIG_BPW (sizeof(unsigned long) * 8) + #define _NSIG_WORDS (_NSIG / _NSIG_BPW) + + typedef struct { + unsigned long sig[_NSIG_WORDS]; + } sigset_t; + +_NSIG_BPW is 8 * 8 = 64 on MIPS64 or 4 * 8 = 32 on MIPS + +In QEMU: + +TARGET_NSIG_BPW is TARGET_ABI_BITS which is TARGET_LONG_BITS which is +64 on MIPS64 and 32 on MIPS. + +I think there is no problem. + +Thanks, +Laurent + +From: Laurent Vivier +> +If I compare with kernel, it looks good: +> +... +> +I think there is no problem. +Sure, thanks for such fast response - again, I am glad if you are right. +However, for some reason, glibc (and musl too) define sigset_t differently than +kernel. Please take a look. I am not sure if this is covered fine in our code. + +Yours, +Aleksandar + +> +Thanks, +> +Laurent + +On Wed, 3 Jul 2019 at 21:20, Aleksandar Markovic wrote: +> +> +From: Laurent Vivier +> +> If I compare with kernel, it looks good: +> +> ... +> +> I think there is no problem. +> +> +Sure, thanks for such fast response - again, I am glad if you are right. +> +However, for some reason, glibc (and musl too) define sigset_t differently +> +than kernel. Please take a look. I am not sure if this is covered fine in our +> +code. +Yeah, the libc definitions of sigset_t don't match the +kernel ones (this is for obscure historical reasons IIRC). +We're providing implementations of the target +syscall interface, so our target_sigset_t should be the +target kernel's version (and the target libc's version doesn't +matter to us). On the other hand we will be using the +host libc version, I think, so a little caution is required +and it's possible we have some bugs in our code. + +thanks +-- PMM + +> +From: Peter Maydell +> +> +On Wed, 3 Jul 2019 at 21:20, Aleksandar Markovic wrote: +> +> +> +> From: Laurent Vivier +> +> > If I compare with kernel, it looks good: +> +> > ... +> +> > I think there is no problem. +> +> +> +> Sure, thanks for such fast response - again, I am glad if you are right. +> +> However, for some reason, glibc (and musl too) define sigset_t differently +> +> than kernel. Please take a look. I am not sure if this is covered fine in +> +> our code. +> +> +Yeah, the libc definitions of sigset_t don't match the +> +kernel ones (this is for obscure historical reasons IIRC). +> +We're providing implementations of the target +> +syscall interface, so our target_sigset_t should be the +> +target kernel's version (and the target libc's version doesn't +> +matter to us). On the other hand we will be using the +> +host libc version, I think, so a little caution is required +> +and it's possible we have some bugs in our code. +OK, I gather than this is not something that requires our immediate attention +(for 4.1), but we can analyze it later on. + +Thanks for response!! + +Sincerely, +Aleksandar + +> +thanks +> +-- PMM + +Le 03/07/2019 à 22:28, Peter Maydell a écrit : +> +On Wed, 3 Jul 2019 at 21:20, Aleksandar Markovic wrote: +> +> +> +> From: Laurent Vivier +> +>> If I compare with kernel, it looks good: +> +>> ... +> +>> I think there is no problem. +> +> +> +> Sure, thanks for such fast response - again, I am glad if you are right. +> +> However, for some reason, glibc (and musl too) define sigset_t differently +> +> than kernel. Please take a look. I am not sure if this is covered fine in +> +> our code. +> +> +Yeah, the libc definitions of sigset_t don't match the +> +kernel ones (this is for obscure historical reasons IIRC). +> +We're providing implementations of the target +> +syscall interface, so our target_sigset_t should be the +> +target kernel's version (and the target libc's version doesn't +> +matter to us). On the other hand we will be using the +> +host libc version, I think, so a little caution is required +> +and it's possible we have some bugs in our code. +It's why we need host_to_target_sigset_internal() and +target_to_host_sigset_internal() that translates bits and bytes between +guest kernel interface and host libc interface. + +void host_to_target_sigset_internal(target_sigset_t *d, + const sigset_t *s) +{ + int i; + target_sigemptyset(d); + for (i = 1; i <= TARGET_NSIG; i++) { + if (sigismember(s, i)) { + target_sigaddset(d, host_to_target_signal(i)); + } + } +} + +void target_to_host_sigset_internal(sigset_t *d, + const target_sigset_t *s) +{ + int i; + sigemptyset(d); + for (i = 1; i <= TARGET_NSIG; i++) { + if (target_sigismember(s, i)) { + sigaddset(d, target_to_host_signal(i)); + } + } +} + +Thanks, +Laurent + +Hi Aleksandar, + +On Wed, Jul 3, 2019 at 12:48 PM Aleksandar Markovic + wrote: +> +#define TARGET_NSIG_BPW TARGET_ABI_BITS +> +#define TARGET_NSIG_WORDS (TARGET_NSIG / TARGET_NSIG_BPW) +> +> +typedef struct { +> +abi_ulong sig[TARGET_NSIG_WORDS]; +> +} target_sigset_t; +> +> +... TARGET_ABI_BITS should be replaced by eight times smaller constant (in +> +fact, +> +semantically, we need TARGET_ABI_BYTES, but it is not defined) (what is needed +> +is actually "a byte per signal" in target_sigset_t, and we allow "a bit per +> +signal"). +Why do we need a byte per target signal, if the functions in linux-user/signal.c +operate with bits? + +-- +Thanks. +-- Max + +> +Why do we need a byte per target signal, if the functions in +> +linux-user/signal.c +> +operate with bits? +Max, + +I did not base my findings on code analysis, but on dumping size/offsets of +elements of some structures, as they are emulated in QEMU, and in real systems. +So, I can't really answer your question. + +Yours, +Aleksandar + +> +-- +> +Thanks. +> +-- Max + diff --git a/mailinglist/output_mailinglist/0247400 b/mailinglist/output_mailinglist/0247400 deleted file mode 100644 index ae9db277..00000000 --- a/mailinglist/output_mailinglist/0247400 +++ /dev/null @@ -1,1481 +0,0 @@ -[Qemu-devel][bug] qemu crash when migrate vm and vm's disks - -When migrate vm and vm’s disks target host qemu crash due to an invalid free. -#0  object_unref (obj=0x1000) at /qemu-2.12/rpmbuild/BUILD/qemu-2.12/qom/object.c:920 -#1  0x0000560434d79e79 in memory_region_unref (mr=) -at /qemu-2.12/rpmbuild/BUILD/qemu-2.12/memory.c:1730 -#2  flatview_destroy (view=0x560439653880) at /qemu-2.12/rpmbuild/BUILD/qemu-2.12/memory.c:292 -#3  0x000056043514dfbe in call_rcu_thread (opaque=) -at /qemu-2.12/rpmbuild/BUILD/qemu-2.12/util/rcu.c:284 -#4  0x00007fbc2b36fe25 in start_thread () from /lib64/libpthread.so.0 -#5  0x00007fbc2b099bad in clone () from /lib64/libc.so.6 -test base qemu-2.12.0 -, -but use lastest qemu(v6.0.0-rc2) also reproduce. -As follow patch can resolve this problem: -https://lists.gnu.org/archive/html/qemu-devel/2018-07/msg02272.html -Steps to reproduce: -(1) Create VM (virsh define) -(2) Add 64 virtio scsi disks -(3) migrate vm and vm’disks -------------------------------------------------------------------------------------------------------------------------------------- -本邮件及其附件含有新华三集团的保密信息,仅限于发送给上面地址中列出 -的个人或群组。禁止任何其他人以任何形式使用(包括但不限于全部或部分地泄露、复制、 -或散发)本邮件中的信息。如果您错收了本邮件,请您立即电话或邮件通知发件人并删除本 -邮件! -This e-mail and its attachments contain confidential information from New H3C, which is -intended only for the person or entity whose address is listed above. Any use of the -information contained herein in any way (including, but not limited to, total or partial -disclosure, reproduction, or dissemination) by persons other than the intended -recipient(s) is prohibited. If you receive this e-mail in error, please notify the sender -by phone or email immediately and delete it! - -* Yuchen (yu.chen@h3c.com) wrote: -> -When migrate vm and vm’s disks target host qemu crash due to an invalid free. -> -> -#0 object_unref (obj=0x1000) at -> -/qemu-2.12/rpmbuild/BUILD/qemu-2.12/qom/object.c:920 -> -#1 0x0000560434d79e79 in memory_region_unref (mr=) -> -at /qemu-2.12/rpmbuild/BUILD/qemu-2.12/memory.c:1730 -> -#2 flatview_destroy (view=0x560439653880) at -> -/qemu-2.12/rpmbuild/BUILD/qemu-2.12/memory.c:292 -> -#3 0x000056043514dfbe in call_rcu_thread (opaque=) -> -at /qemu-2.12/rpmbuild/BUILD/qemu-2.12/util/rcu.c:284 -> -#4 0x00007fbc2b36fe25 in start_thread () from /lib64/libpthread.so.0 -> -#5 0x00007fbc2b099bad in clone () from /lib64/libc.so.6 -> -> -test base qemu-2.12.0,but use lastest qemu(v6.0.0-rc2) also reproduce. -Interesting. - -> -As follow patch can resolve this problem: -> -https://lists.gnu.org/archive/html/qemu-devel/2018-07/msg02272.html -That's a pci/rcu change; ccing Paolo and Micahel. - -> -Steps to reproduce: -> -(1) Create VM (virsh define) -> -(2) Add 64 virtio scsi disks -Is that hot adding the disks later, or are they included in the VM at -creation? -Can you provide a libvirt XML example? - -> -(3) migrate vm and vm’disks -What do you mean by 'and vm disks' - are you doing a block migration? - -Dave - -> -------------------------------------------------------------------------------------------------------------------------------------- -> -本邮件及其附件含有新华三集团的保密信息,仅限于发送给上面地址中列出 -> -的个人或群组。禁止任何其他人以任何形式使用(包括但不限于全部或部分地泄露、复制、 -> -或散发)本邮件中的信息。如果您错收了本邮件,请您立即电话或邮件通知发件人并删除本 -> -邮件! -> -This e-mail and its attachments contain confidential information from New -> -H3C, which is -> -intended only for the person or entity whose address is listed above. Any use -> -of the -> -information contained herein in any way (including, but not limited to, total -> -or partial -> -disclosure, reproduction, or dissemination) by persons other than the intended -> -recipient(s) is prohibited. If you receive this e-mail in error, please -> -notify the sender -> -by phone or email immediately and delete it! --- -Dr. David Alan Gilbert / dgilbert@redhat.com / Manchester, UK - -> ------邮件原件----- -> -发件人: Dr. David Alan Gilbert [ -mailto:dgilbert@redhat.com -] -> -发送时间: 2021å¹´4月8日 19:27 -> -收件人: yuchen (Cloud) ; pbonzini@redhat.com; -> -mst@redhat.com -> -抄送: qemu-devel@nongnu.org -> -主题: Re: [Qemu-devel][bug] qemu crash when migrate vm and vm's disks -> -> -* Yuchen (yu.chen@h3c.com) wrote: -> -> When migrate vm and vm’s disks target host qemu crash due to an invalid -> -free. -> -> -> -> #0 object_unref (obj=0x1000) at -> -> /qemu-2.12/rpmbuild/BUILD/qemu-2.12/qom/object.c:920 -> -> #1 0x0000560434d79e79 in memory_region_unref (mr=) -> -> at /qemu-2.12/rpmbuild/BUILD/qemu-2.12/memory.c:1730 -> -> #2 flatview_destroy (view=0x560439653880) at -> -> /qemu-2.12/rpmbuild/BUILD/qemu-2.12/memory.c:292 -> -> #3 0x000056043514dfbe in call_rcu_thread (opaque=) -> -> at /qemu-2.12/rpmbuild/BUILD/qemu-2.12/util/rcu.c:284 -> -> #4 0x00007fbc2b36fe25 in start_thread () from /lib64/libpthread.so.0 -> -> #5 0x00007fbc2b099bad in clone () from /lib64/libc.so.6 -> -> -> -> test base qemu-2.12.0,but use lastest qemu(v6.0.0-rc2) also reproduce. -> -> -Interesting. -> -> -> As follow patch can resolve this problem: -> -> -https://lists.gnu.org/archive/html/qemu-devel/2018-07/msg02272.html -> -> -That's a pci/rcu change; ccing Paolo and Micahel. -> -> -> Steps to reproduce: -> -> (1) Create VM (virsh define) -> -> (2) Add 64 virtio scsi disks -> -> -Is that hot adding the disks later, or are they included in the VM at -> -creation? -> -Can you provide a libvirt XML example? -> -Include disks in the VM at creation - -vm disks xml (only virtio scsi disks): - - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - -
- - - - - -
- - - - -
- - - -vm disks xml (only virtio disks): - - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - - -
- - - - -
- - - -> -> (3) migrate vm and vm’disks -> -> -What do you mean by 'and vm disks' - are you doing a block migration? -> -Yes, block migration. -In fact, only migration domain also reproduced. - -> -Dave -> -> -> ---------------------------------------------------------------------- -> -> --------------------------------------------------------------- -> -Dr. David Alan Gilbert / dgilbert@redhat.com / Manchester, UK -------------------------------------------------------------------------------------------------------------------------------------- -本邮件及其附件含有新华三集团的保密信息,仅限于发送给上面地址中列出 -的个人或群组。禁止任何其他人以任何形式使用(包括但不限于全部或部分地泄露、复制、 -或散发)本邮件中的信息。如果您错收了本邮件,请您立即电话或邮件通知发件人并删除本 -邮件! -This e-mail and its attachments contain confidential information from New H3C, -which is -intended only for the person or entity whose address is listed above. Any use -of the -information contained herein in any way (including, but not limited to, total -or partial -disclosure, reproduction, or dissemination) by persons other than the intended -recipient(s) is prohibited. If you receive this e-mail in error, please notify -the sender -by phone or email immediately and delete it! - diff --git a/mailinglist/output_mailinglist/02572177 b/mailinglist/output_mailinglist/02572177 new file mode 100644 index 00000000..e7c0731e --- /dev/null +++ b/mailinglist/output_mailinglist/02572177 @@ -0,0 +1,416 @@ +[Qemu-devel] 答复: Re: [BUG]COLO failover hang + +hi. + + +I test the git qemu master have the same problem. + + +(gdb) bt + + +#0 qio_channel_socket_readv (ioc=0x7f65911b4e50, iov=0x7f64ef3fd880, niov=1, +fds=0x0, nfds=0x0, errp=0x0) at io/channel-socket.c:461 + + +#1 0x00007f658e4aa0c2 in qio_channel_read (address@hidden, address@hidden "", +address@hidden, address@hidden) at io/channel.c:114 + + +#2 0x00007f658e3ea990 in channel_get_buffer (opaque=<optimized out>, +buf=0x7f65907cb838 "", pos=<optimized out>, size=32768) at +migration/qemu-file-channel.c:78 + + +#3 0x00007f658e3e97fc in qemu_fill_buffer (f=0x7f65907cb800) at +migration/qemu-file.c:295 + + +#4 0x00007f658e3ea2e1 in qemu_peek_byte (address@hidden, address@hidden) at +migration/qemu-file.c:555 + + +#5 0x00007f658e3ea34b in qemu_get_byte (address@hidden) at +migration/qemu-file.c:568 + + +#6 0x00007f658e3ea552 in qemu_get_be32 (address@hidden) at +migration/qemu-file.c:648 + + +#7 0x00007f658e3e66e5 in colo_receive_message (f=0x7f65907cb800, +address@hidden) at migration/colo.c:244 + + +#8 0x00007f658e3e681e in colo_receive_check_message (f=<optimized out>, +address@hidden, address@hidden) + + + at migration/colo.c:264 + + +#9 0x00007f658e3e740e in colo_process_incoming_thread (opaque=0x7f658eb30360 +<mis_current.31286>) at migration/colo.c:577 + + +#10 0x00007f658be09df3 in start_thread () from /lib64/libpthread.so.0 + + +#11 0x00007f65881983ed in clone () from /lib64/libc.so.6 + + +(gdb) p ioc->name + + +$2 = 0x7f658ff7d5c0 "migration-socket-incoming" + + +(gdb) p ioc->features Do not support QIO_CHANNEL_FEATURE_SHUTDOWN + + +$3 = 0 + + + + + +(gdb) bt + + +#0 socket_accept_incoming_migration (ioc=0x7fdcceeafa90, condition=G_IO_IN, +opaque=0x7fdcceeafa90) at migration/socket.c:137 + + +#1 0x00007fdcc6966350 in g_main_dispatch (context=<optimized out>) at +gmain.c:3054 + + +#2 g_main_context_dispatch (context=<optimized out>, address@hidden) at +gmain.c:3630 + + +#3 0x00007fdccb8a6dcc in glib_pollfds_poll () at util/main-loop.c:213 + + +#4 os_host_main_loop_wait (timeout=<optimized out>) at util/main-loop.c:258 + + +#5 main_loop_wait (address@hidden) at util/main-loop.c:506 + + +#6 0x00007fdccb526187 in main_loop () at vl.c:1898 + + +#7 main (argc=<optimized out>, argv=<optimized out>, envp=<optimized out>) at +vl.c:4709 + + +(gdb) p ioc->features + + +$1 = 6 + + +(gdb) p ioc->name + + +$2 = 0x7fdcce1b1ab0 "migration-socket-listener" + + + + + +May be socket_accept_incoming_migration should call +qio_channel_set_feature(ioc, QIO_CHANNEL_FEATURE_SHUTDOWN)?? + + + + + +thank you. + + + + + + + + + + + + + + + +原始邮件 + + + +发件人: address@hidden +收件人:王广10165992 address@hidden +抄送人: address@hidden address@hidden +日 期 :2017å¹´03月16日 14:46 +主 题 :Re: [Qemu-devel] COLO failover hang + + + + + + + +On 03/15/2017 05:06 PM, wangguang wrote: +> am testing QEMU COLO feature described here [QEMU +> Wiki]( +http://wiki.qemu-project.org/Features/COLO +). +> +> When the Primary Node panic,the Secondary Node qemu hang. +> hang at recvmsg in qio_channel_socket_readv. +> And I run { 'execute': 'nbd-server-stop' } and { "execute": +> "x-colo-lost-heartbeat" } in Secondary VM's +> monitor,the Secondary Node qemu still hang at recvmsg . +> +> I found that the colo in qemu is not complete yet. +> Do the colo have any plan for development? + +Yes, We are developing. You can see some of patch we pushing. + +> Has anyone ever run it successfully? Any help is appreciated! + +In our internal version can run it successfully, +The failover detail you can ask Zhanghailiang for help. +Next time if you have some question about COLO, +please cc me and zhanghailiang address@hidden + + +Thanks +Zhang Chen + + +> +> +> +> centos7.2+qemu2.7.50 +> (gdb) bt +> #0 0x00007f3e00cc86ad in recvmsg () from /lib64/libpthread.so.0 +> #1 0x00007f3e0332b738 in qio_channel_socket_readv (ioc=<optimized out>, +> iov=<optimized out>, niov=<optimized out>, fds=0x0, nfds=0x0, errp=0x0) at +> io/channel-socket.c:497 +> #2 0x00007f3e03329472 in qio_channel_read (address@hidden, +> address@hidden "", address@hidden, +> address@hidden) at io/channel.c:97 +> #3 0x00007f3e032750e0 in channel_get_buffer (opaque=<optimized out>, +> buf=0x7f3e05910f38 "", pos=<optimized out>, size=32768) at +> migration/qemu-file-channel.c:78 +> #4 0x00007f3e0327412c in qemu_fill_buffer (f=0x7f3e05910f00) at +> migration/qemu-file.c:257 +> #5 0x00007f3e03274a41 in qemu_peek_byte (address@hidden, +> address@hidden) at migration/qemu-file.c:510 +> #6 0x00007f3e03274aab in qemu_get_byte (address@hidden) at +> migration/qemu-file.c:523 +> #7 0x00007f3e03274cb2 in qemu_get_be32 (address@hidden) at +> migration/qemu-file.c:603 +> #8 0x00007f3e03271735 in colo_receive_message (f=0x7f3e05910f00, +> address@hidden) at migration/colo.c:215 +> #9 0x00007f3e0327250d in colo_wait_handle_message (errp=0x7f3d62bfaa48, +> checkpoint_request=<synthetic pointer>, f=<optimized out>) at +> migration/colo.c:546 +> #10 colo_process_incoming_thread (opaque=0x7f3e067245e0) at +> migration/colo.c:649 +> #11 0x00007f3e00cc1df3 in start_thread () from /lib64/libpthread.so.0 +> #12 0x00007f3dfc9c03ed in clone () from /lib64/libc.so.6 +> +> +> +> +> +> -- +> View this message in context: +http://qemu.11.n7.nabble.com/COLO-failover-hang-tp473250.html +> Sent from the Developer mailing list archive at Nabble.com. +> +> +> +> + +-- +Thanks +Zhang Chen + +Hi,Wang. + +You can test this branch: +https://github.com/coloft/qemu/tree/colo-v5.1-developing-COLO-frame-v21-with-shared-disk +and please follow wiki ensure your own configuration correctly. +http://wiki.qemu-project.org/Features/COLO +Thanks + +Zhang Chen + + +On 03/21/2017 03:27 PM, address@hidden wrote: +hi. + +I test the git qemu master have the same problem. + +(gdb) bt +#0 qio_channel_socket_readv (ioc=0x7f65911b4e50, iov=0x7f64ef3fd880, +niov=1, fds=0x0, nfds=0x0, errp=0x0) at io/channel-socket.c:461 +#1 0x00007f658e4aa0c2 in qio_channel_read +(address@hidden, address@hidden "", +address@hidden, address@hidden) at io/channel.c:114 +#2 0x00007f658e3ea990 in channel_get_buffer (opaque=<optimized out>, +buf=0x7f65907cb838 "", pos=<optimized out>, size=32768) at +migration/qemu-file-channel.c:78 +#3 0x00007f658e3e97fc in qemu_fill_buffer (f=0x7f65907cb800) at +migration/qemu-file.c:295 +#4 0x00007f658e3ea2e1 in qemu_peek_byte (address@hidden, +address@hidden) at migration/qemu-file.c:555 +#5 0x00007f658e3ea34b in qemu_get_byte (address@hidden) at +migration/qemu-file.c:568 +#6 0x00007f658e3ea552 in qemu_get_be32 (address@hidden) at +migration/qemu-file.c:648 +#7 0x00007f658e3e66e5 in colo_receive_message (f=0x7f65907cb800, +address@hidden) at migration/colo.c:244 +#8 0x00007f658e3e681e in colo_receive_check_message (f=<optimized +out>, address@hidden, +address@hidden) +at migration/colo.c:264 +#9 0x00007f658e3e740e in colo_process_incoming_thread +(opaque=0x7f658eb30360 <mis_current.31286>) at migration/colo.c:577 +#10 0x00007f658be09df3 in start_thread () from /lib64/libpthread.so.0 + +#11 0x00007f65881983ed in clone () from /lib64/libc.so.6 + +(gdb) p ioc->name + +$2 = 0x7f658ff7d5c0 "migration-socket-incoming" + +(gdb) p ioc->features Do not support QIO_CHANNEL_FEATURE_SHUTDOWN + +$3 = 0 + + +(gdb) bt +#0 socket_accept_incoming_migration (ioc=0x7fdcceeafa90, +condition=G_IO_IN, opaque=0x7fdcceeafa90) at migration/socket.c:137 +#1 0x00007fdcc6966350 in g_main_dispatch (context=<optimized out>) at +gmain.c:3054 +#2 g_main_context_dispatch (context=<optimized out>, +address@hidden) at gmain.c:3630 +#3 0x00007fdccb8a6dcc in glib_pollfds_poll () at util/main-loop.c:213 +#4 os_host_main_loop_wait (timeout=<optimized out>) at +util/main-loop.c:258 +#5 main_loop_wait (address@hidden) at +util/main-loop.c:506 +#6 0x00007fdccb526187 in main_loop () at vl.c:1898 +#7 main (argc=<optimized out>, argv=<optimized out>, envp=<optimized +out>) at vl.c:4709 +(gdb) p ioc->features + +$1 = 6 + +(gdb) p ioc->name + +$2 = 0x7fdcce1b1ab0 "migration-socket-listener" +May be socket_accept_incoming_migration should +call qio_channel_set_feature(ioc, QIO_CHANNEL_FEATURE_SHUTDOWN)?? +thank you. + + + + + +原始邮件 +address@hidden; +*收件人:*王广10165992;address@hidden; +address@hidden;address@hidden; +*日 期 :*2017å¹´03月16日 14:46 +*主 题 :**Re: [Qemu-devel] COLO failover hang* + + + + +On 03/15/2017 05:06 PM, wangguang wrote: +> am testing QEMU COLO feature described here [QEMU +> Wiki]( +http://wiki.qemu-project.org/Features/COLO +). +> +> When the Primary Node panic,the Secondary Node qemu hang. +> hang at recvmsg in qio_channel_socket_readv. +> And I run { 'execute': 'nbd-server-stop' } and { "execute": +> "x-colo-lost-heartbeat" } in Secondary VM's +> monitor,the Secondary Node qemu still hang at recvmsg . +> +> I found that the colo in qemu is not complete yet. +> Do the colo have any plan for development? + +Yes, We are developing. You can see some of patch we pushing. + +> Has anyone ever run it successfully? Any help is appreciated! + +In our internal version can run it successfully, +The failover detail you can ask Zhanghailiang for help. +Next time if you have some question about COLO, +please cc me and zhanghailiang address@hidden + + +Thanks +Zhang Chen + + +> +> +> +> centos7.2+qemu2.7.50 +> (gdb) bt +> #0 0x00007f3e00cc86ad in recvmsg () from /lib64/libpthread.so.0 +> #1 0x00007f3e0332b738 in qio_channel_socket_readv (ioc=<optimized out>, +> iov=<optimized out>, niov=<optimized out>, fds=0x0, nfds=0x0, errp=0x0) at +> io/channel-socket.c:497 +> #2 0x00007f3e03329472 in qio_channel_read (address@hidden, +> address@hidden "", address@hidden, +> address@hidden) at io/channel.c:97 +> #3 0x00007f3e032750e0 in channel_get_buffer (opaque=<optimized out>, +> buf=0x7f3e05910f38 "", pos=<optimized out>, size=32768) at +> migration/qemu-file-channel.c:78 +> #4 0x00007f3e0327412c in qemu_fill_buffer (f=0x7f3e05910f00) at +> migration/qemu-file.c:257 +> #5 0x00007f3e03274a41 in qemu_peek_byte (address@hidden, +> address@hidden) at migration/qemu-file.c:510 +> #6 0x00007f3e03274aab in qemu_get_byte (address@hidden) at +> migration/qemu-file.c:523 +> #7 0x00007f3e03274cb2 in qemu_get_be32 (address@hidden) at +> migration/qemu-file.c:603 +> #8 0x00007f3e03271735 in colo_receive_message (f=0x7f3e05910f00, +> address@hidden) at migration/colo.c:215 +> #9 0x00007f3e0327250d in colo_wait_handle_message (errp=0x7f3d62bfaa48, +> checkpoint_request=<synthetic pointer>, f=<optimized out>) at +> migration/colo.c:546 +> #10 colo_process_incoming_thread (opaque=0x7f3e067245e0) at +> migration/colo.c:649 +> #11 0x00007f3e00cc1df3 in start_thread () from /lib64/libpthread.so.0 +> #12 0x00007f3dfc9c03ed in clone () from /lib64/libc.so.6 +> +> +> +> +> +> -- +> View this message in context: +http://qemu.11.n7.nabble.com/COLO-failover-hang-tp473250.html +> Sent from the Developer mailing list archive at Nabble.com. +> +> +> +> + +-- +Thanks +Zhang Chen +-- +Thanks +Zhang Chen + diff --git a/mailinglist/output_mailinglist/04472277 b/mailinglist/output_mailinglist/04472277 new file mode 100644 index 00000000..48fea251 --- /dev/null +++ b/mailinglist/output_mailinglist/04472277 @@ -0,0 +1,571 @@ +[BUG][KVM_SET_USER_MEMORY_REGION] KVM_SET_USER_MEMORY_REGION failed + +Hi all, +I start a VM in openstack, and openstack use libvirt to start qemu VM, but now log show this ERROR. +Is there any one know this? +The ERROR log from /var/log/libvirt/qemu/instance-0000000e.log +``` +2023-03-14T10:09:17.674114Z qemu-system-x86_64: kvm_set_user_memory_region: KVM_SET_USER_MEMORY_REGION failed, slot=4, start=0xfffffffffe000000, size=0x2000: Invalid argument +kvm_set_phys_mem: error registering slot: Invalid argument +2023-03-14 10:09:18.198+0000: shutting down, reason=crashed +``` +The xml file +``` +root@c1c2:~# cat /etc/libvirt/qemu/instance-0000000e.xml + + +  instance-0000000e +  ff91d2dc-69a1-43ef-abde-c9e4e9a0305b +  +    +      +      provider-instance +      2023-03-14 10:09:13 +      +        64 +        1 +        0 +        0 +        1 +      +      +        admin +        admin +      +      +      +        +          +        +      +    +  +  65536 +  65536 +  1 +  +    +      OpenStack Foundation +      OpenStack Nova +      25.1.0 +      ff91d2dc-69a1-43ef-abde-c9e4e9a0305b +      ff91d2dc-69a1-43ef-abde-c9e4e9a0305b +      Virtual Machine +    +  +  +    hvm +    +    +  +  +    +    +    +  +  +    +  +  +    +    +    +  +  destroy +  restart +  destroy +  +    /usr/bin/qemu-system-x86_64 +    +      +      +      +     
+    +    +     
+    +    +    +      +      +       
+      +     
+    +    +      +      +        +      +    +    +      +      +    +    +     
+    +    +    +    +      +    +  Â