[PVE-User] PVE 5.4: cannot move disk image to Ceph

Mark Adams mark at openvs.co.uk
Fri Sep 6 12:09:16 CEST 2019


Is it potentially an issue with having the same pool name on 2 different
ceph clusters?

is there a vm-112-disk-0 on vdisks_cluster2?

On Fri, 6 Sep 2019, 12:45 Uwe Sauter, <uwe.sauter.de at gmail.com> wrote:

> Hello Alwin,
>
> Am 06.09.19 um 11:32 schrieb Alwin Antreich:
> > Hello Uwe,
> >
> > On Fri, Sep 06, 2019 at 10:41:18AM +0200, Uwe Sauter wrote:
> >> Hi,
> >>
> >> I'm having trouble moving a disk image to Ceph. Moving between local
> disks and NFS share is working.
> >>
> >> The error given is:
> >>
> >> ########
> >> create full clone of drive scsi0
> (aurel-cluster1-VMs:112/vm-112-disk-0.qcow2)
> >> rbd: create error: (17) File exists
> >> TASK ERROR: storage migration failed: error with cfs lock
> 'storage-vdisks_vm': rbd create vm-112-disk-0' error: rbd: create error:
> >> (17) File exists
> >> ########
> > Can you see anything in the ceph logs? And on what version (pveversion
> > -v) are you on?
>
> Nothing obvious in the logs. The cluster is healthy
>
> root at px-bravo-cluster:~# ceph status
>   cluster:
>     id:     982484e6-69bf-490c-9b3a-942a179e759b
>     health: HEALTH_OK
>
>   services:
>     mon: 3 daemons, quorum
> px-alpha-cluster,px-bravo-cluster,px-charlie-cluster
>     mgr: px-alpha-cluster(active), standbys: px-bravo-cluster,
> px-charlie-cluster
>     osd: 9 osds: 9 up, 9 in
>
>   data:
>     pools:   1 pools, 128 pgs
>     objects: 14.76k objects, 56.0GiB
>     usage:   163GiB used, 3.99TiB / 4.15TiB avail
>     pgs:     128 active+clean
>
>   io:
>     client:   2.31KiB/s wr, 0op/s rd, 0op/s wr
>
> I'm on a fully up-to-date PVE 5.4 (all three nodes).
>
> root at px-bravo-cluster:~# pveversion -v
> proxmox-ve: 5.4-2 (running kernel: 4.15.18-20-pve)
> pve-manager: 5.4-13 (running version: 5.4-13/aee6f0ec)
> pve-kernel-4.15: 5.4-8
> pve-kernel-4.15.18-20-pve: 4.15.18-46
> pve-kernel-4.15.18-19-pve: 4.15.18-45
> ceph: 12.2.12-pve1
> corosync: 2.4.4-pve1
> criu: 2.11.1-1~bpo90
> glusterfs-client: 3.8.8-1
> ksm-control-daemon: 1.2-2
> libjs-extjs: 6.0.1-2
> libpve-access-control: 5.1-12
> libpve-apiclient-perl: 2.0-5
> libpve-common-perl: 5.0-54
> libpve-guest-common-perl: 2.0-20
> libpve-http-server-perl: 2.0-14
> libpve-storage-perl: 5.0-44
> libqb0: 1.0.3-1~bpo9
> lvm2: 2.02.168-pve6
> lxc-pve: 3.1.0-6
> lxcfs: 3.0.3-pve1
> novnc-pve: 1.0.0-3
> proxmox-widget-toolkit: 1.0-28
> pve-cluster: 5.0-38
> pve-container: 2.0-40
> pve-docs: 5.4-2
> pve-edk2-firmware: 1.20190312-1
> pve-firewall: 3.0-22
> pve-firmware: 2.0-7
> pve-ha-manager: 2.0-9
> pve-i18n: 1.1-4
> pve-libspice-server1: 0.14.1-2
> pve-qemu-kvm: 3.0.1-4
> pve-xtermjs: 3.12.0-1
> qemu-server: 5.0-54
> smartmontools: 6.5+svn4324-1
> spiceterm: 3.0-5
> vncterm: 1.5-3
> zfsutils-linux: 0.7.13-pve1~bpo2
>
>
>
> >>
> >> but this is not true:
> >>
> >> ########
> >> root at px-bravo-cluster:~# rbd -p vdisks ls
> >> vm-106-disk-0
> >> vm-113-disk-0
> >> vm-113-disk-1
> >> vm-113-disk-2
> >> vm-118-disk-0
> >> vm-119-disk-0
> >> vm-120-disk-0
> >> vm-125-disk-0
> >> vm-125-disk-1
> >> ########
> > Can you create the image by hand (rbd -p rbd create vm-112-disk-0 --size
> > 1G)? And (rbd -p rbd rm vm-112-disk-0) for delete, ofc.
>
> root at px-bravo-cluster:~# rbd -p vdisks create vm-112-disk-0 --size 1G
> rbd: create error: (17) File exists
> 2019-09-06 11:35:20.943998 7faf704660c0 -1 librbd: rbd image vm-112-disk-0
> already exists
>
> root at px-bravo-cluster:~# rbd -p vdisks create test --size 1G
>
> root at px-bravo-cluster:~# rbd -p vdisks ls
> test
> vm-106-disk-0
> vm-113-disk-0
> vm-113-disk-1
> vm-113-disk-2
> vm-118-disk-0
> vm-119-disk-0
> vm-120-disk-0
> vm-125-disk-0
> vm-125-disk-1
>
> root at px-bravo-cluster:~# rbd -p vdisks rm test
> Removing image: 100% complete...done.
>
> root at px-bravo-cluster:~# rbd -p vdisks rm vm-112-disk-0
> 2019-09-06 11:36:07.570749 7eff7cff9700 -1 librbd::image::OpenRequest:
> failed to retreive immutable metadata: (2) No such file or
> directory
> Removing image: 0% complete...failed.
> rbd: delete error: (2) No such file or directory
>
>
> >
> >>
> >> Here is the relevant part of my storage.cfg:
> >>
> >> ########
> >> nfs: aurel-cluster1-VMs
> >>      export /backup/proxmox-infra/VMs
> >>      path /mnt/pve/aurel-cluster1-VMs
> >>      server X.X.X.X
> >>      content images
> >>      options vers=4.2
> >>
> >>
> >> rbd: vdisks_vm
> >>      content images
> >>      krbd 0
> >>      pool vdisks
> >> ########
> > Is this the complete storage.cfg?
>
> No, only the parts that are relevant for this particular move. Here's the
> complete file:
>
> ########
> rbd: vdisks_vm
>         content images
>         krbd 0
>         pool vdisks
>
> dir: local-hdd
>         path /mnt/local
>         content images,iso
>         nodes px-alpha-cluster,px-bravo-cluster,px-charlie-cluster
>         shared 0
>
> nfs: aurel-cluster1-daily
>         export /backup/proxmox-infra/daily
>         path /mnt/pve/aurel-cluster1-daily
>         server X.X.X.X
>         content backup
>         maxfiles 30
>         options vers=4.2
>
> nfs: aurel-cluster1-weekly
>         export /backup/proxmox-infra/weekly
>         path /mnt/pve/aurel-cluster1-weekly
>         server X.X.X.X
>         content backup
>         maxfiles 30
>         options vers=4.2
>
> nfs: aurel-cluster1-VMs
>         export /backup/proxmox-infra/VMs
>         path /mnt/pve/aurel-cluster1-VMs
>         server X.X.X.X
>         content images
>         options vers=4.2
>
> nfs: aurel-cluster2-daily
>         export /backup/proxmox-infra2/daily
>         path /mnt/pve/aurel-cluster2-daily
>         server X.X.X.X
>         content backup
>         maxfiles 30
>         options vers=4.2
>
> nfs: aurel-cluster2-weekly
>         export /backup/proxmox-infra2/weekly
>         path /mnt/pve/aurel-cluster2-weekly
>         server X.X.X.X
>         content backup
>         maxfiles 30
>         options vers=4.2
>
> nfs: aurel-cluster2-VMs
>         export /backup/proxmox-infra2/VMs
>         path /mnt/pve/aurel-cluster2-VMs
>         server X.X.X.X
>         content images
>         options vers=4.2
>
> dir: local
>         path /var/lib/vz
>         content snippets,vztmpl,images,rootdir,iso
>         maxfiles 0
>
> rbd: vdisks_cluster2
>         content images
>         krbd 0
>         monhost px-golf-cluster, px-hotel-cluster, px-india-cluster
>         pool vdisks
>         username admin
> ########
>
> Thanks,
>
>         Uwe
>
> > --
> > Cheers,
> > Alwin
> >
>
> _______________________________________________
> pve-user mailing list
> pve-user at pve.proxmox.com
> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>



More information about the pve-user mailing list