From 4b927bc37359dec23f67d3427fc982945f24f404 Mon Sep 17 00:00:00 2001 From: Christian Krinitsin Date: Wed, 21 May 2025 21:21:26 +0200 Subject: add gitlab issues in toml format --- .../host_missing/accel_missing/1532.toml | 513 +++++++++++++++++++++ 1 file changed, 513 insertions(+) create mode 100644 gitlab/issues/target_missing/host_missing/accel_missing/1532.toml (limited to 'gitlab/issues/target_missing/host_missing/accel_missing/1532.toml') diff --git a/gitlab/issues/target_missing/host_missing/accel_missing/1532.toml b/gitlab/issues/target_missing/host_missing/accel_missing/1532.toml new file mode 100644 index 000000000..3307ca349 --- /dev/null +++ b/gitlab/issues/target_missing/host_missing/accel_missing/1532.toml @@ -0,0 +1,513 @@ +id = 1532 +title = "libivrtd fork qemu to create vm ,which start with ceph rbd device, after vm status:runing , the qemu stuck at booting from hard disk...." +state = "closed" +created_at = "2023-03-06T03:16:16.415Z" +closed_at = "2023-03-08T09:57:11.880Z" +labels = [] +url = "https://gitlab.com/qemu-project/qemu/-/issues/1532" +host-os = "- OS/kernel version:" +host-arch = "- QEMU flavor:" +qemu-version = "- QEMU command line:" +guest-os = "- OS/kernel version:" +guest-arch = "-- emulated environment" +description = """[root@ceph-client ceph]# virsh list --all + Id Name State +---------------------------------------------------- + 19 c7_ceph running + +the vm qemu stuck at booting from hard disk.....""" +reproduce = """1. use ceph-deploy deploy a ceph distribute storage, which use to store vm's qcow2 files,this ceph has 3 osd node +2. refer the link https://docs.ceph.com/en/quincy/rbd/libvirt/ create a ceph user :client.libvirt +3. import a exists qcow2 file into ceph libvit-pool, then start vm + +[root@ceph-1 ~]# ceph -s + cluster: + id: 3fbbf51f-88fd-4883-9f24-595bf853c5f2 + health: HEALTH_OK + + services: + mon: 1 daemons, quorum ceph-1 + mgr: ceph-1(active) + osd: 3 osds: 3 up, 3 in + + data: + pools: 1 pools, 128 pgs + objects: 940 objects, 3.6 GiB + usage: 31 GiB used, 209 GiB / 240 GiB avail + pgs: 128 active+clean + +[root@ceph-1 ~]#ceph auth ls +client.libvirt +\tkey: AQD/XwFkq7kHMhAA1OmPtKPVno6gjmZleOevOA== +\tcaps: [mon] allow r +\tcaps: [osd] allow class-read object_prefix rbd_children, allow rwx pool=libvirt-pool + +[root@ceph-client ceph]# cat ceph.conf +[global] +fsid = 3fbbf51f-88fd-4883-9f24-595bf853c5f2 +mon_initial_members = ceph-1 +mon_host = 172.24.193.62 +auth_cluster_required = cephx +auth_service_required = cephx +auth_client_required = cephx + +osd_pool_default_size = 2 +[root@ceph-client ceph]# + +[root@ceph-client ceph]# virsh start c7_ceph +Domain c7_ceph started + +[root@ceph-client ceph]# +[root@ceph-client ceph]# virsh list --all + Id Name State +---------------------------------------------------- + 19 c7_ceph running + + + /usr/local/qemu-3.0/bin/qemu-system-x86_64 + + + + + + + + + + + +
+ + +======================== +[root@ceph-client ceph]# cat /run/libvirt/qemu/c7_ceph.xml + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + Broadwell + + + + + + c7_ceph + ff08671e-824c-4939-80ec-602235c0662e + 4194304 + 4194304 + 2 + + /machine + + + hvm + + + + + + + + Broadwell + + + + + + + + + + + + + + + destroy + restart + destroy + + + + + + /usr/local/qemu-3.0/bin/qemu-system-x86_64 + + + + + + + + + + + + + + + + + +
+ + + +
+ + + + +
+ + + + +
+ + + + +
+ + + + + + +
+ + + +
+ + + +
+ + + + + + + +
+ + + + + + + + + + + + + + + + + +
+ + + +
+ + + + + + + + + + +