[PVE-User] PVE 5.4: cannot move disk image to Ceph
Uwe Sauter
uwe.sauter.de at gmail.com
Fri Sep 6 11:44:10 CEST 2019
Hello Alwin,
Am 06.09.19 um 11:32 schrieb Alwin Antreich:
> Hello Uwe,
>
> On Fri, Sep 06, 2019 at 10:41:18AM +0200, Uwe Sauter wrote:
>> Hi,
>>
>> I'm having trouble moving a disk image to Ceph. Moving between local disks and NFS share is working.
>>
>> The error given is:
>>
>> ########
>> create full clone of drive scsi0 (aurel-cluster1-VMs:112/vm-112-disk-0.qcow2)
>> rbd: create error: (17) File exists
>> TASK ERROR: storage migration failed: error with cfs lock 'storage-vdisks_vm': rbd create vm-112-disk-0' error: rbd: create error:
>> (17) File exists
>> ########
> Can you see anything in the ceph logs? And on what version (pveversion
> -v) are you on?
Nothing obvious in the logs. The cluster is healthy
root at px-bravo-cluster:~# ceph status
cluster:
id: 982484e6-69bf-490c-9b3a-942a179e759b
health: HEALTH_OK
services:
mon: 3 daemons, quorum px-alpha-cluster,px-bravo-cluster,px-charlie-cluster
mgr: px-alpha-cluster(active), standbys: px-bravo-cluster, px-charlie-cluster
osd: 9 osds: 9 up, 9 in
data:
pools: 1 pools, 128 pgs
objects: 14.76k objects, 56.0GiB
usage: 163GiB used, 3.99TiB / 4.15TiB avail
pgs: 128 active+clean
io:
client: 2.31KiB/s wr, 0op/s rd, 0op/s wr
I'm on a fully up-to-date PVE 5.4 (all three nodes).
root at px-bravo-cluster:~# pveversion -v
proxmox-ve: 5.4-2 (running kernel: 4.15.18-20-pve)
pve-manager: 5.4-13 (running version: 5.4-13/aee6f0ec)
pve-kernel-4.15: 5.4-8
pve-kernel-4.15.18-20-pve: 4.15.18-46
pve-kernel-4.15.18-19-pve: 4.15.18-45
ceph: 12.2.12-pve1
corosync: 2.4.4-pve1
criu: 2.11.1-1~bpo90
glusterfs-client: 3.8.8-1
ksm-control-daemon: 1.2-2
libjs-extjs: 6.0.1-2
libpve-access-control: 5.1-12
libpve-apiclient-perl: 2.0-5
libpve-common-perl: 5.0-54
libpve-guest-common-perl: 2.0-20
libpve-http-server-perl: 2.0-14
libpve-storage-perl: 5.0-44
libqb0: 1.0.3-1~bpo9
lvm2: 2.02.168-pve6
lxc-pve: 3.1.0-6
lxcfs: 3.0.3-pve1
novnc-pve: 1.0.0-3
proxmox-widget-toolkit: 1.0-28
pve-cluster: 5.0-38
pve-container: 2.0-40
pve-docs: 5.4-2
pve-edk2-firmware: 1.20190312-1
pve-firewall: 3.0-22
pve-firmware: 2.0-7
pve-ha-manager: 2.0-9
pve-i18n: 1.1-4
pve-libspice-server1: 0.14.1-2
pve-qemu-kvm: 3.0.1-4
pve-xtermjs: 3.12.0-1
qemu-server: 5.0-54
smartmontools: 6.5+svn4324-1
spiceterm: 3.0-5
vncterm: 1.5-3
zfsutils-linux: 0.7.13-pve1~bpo2
>>
>> but this is not true:
>>
>> ########
>> root at px-bravo-cluster:~# rbd -p vdisks ls
>> vm-106-disk-0
>> vm-113-disk-0
>> vm-113-disk-1
>> vm-113-disk-2
>> vm-118-disk-0
>> vm-119-disk-0
>> vm-120-disk-0
>> vm-125-disk-0
>> vm-125-disk-1
>> ########
> Can you create the image by hand (rbd -p rbd create vm-112-disk-0 --size
> 1G)? And (rbd -p rbd rm vm-112-disk-0) for delete, ofc.
root at px-bravo-cluster:~# rbd -p vdisks create vm-112-disk-0 --size 1G
rbd: create error: (17) File exists
2019-09-06 11:35:20.943998 7faf704660c0 -1 librbd: rbd image vm-112-disk-0 already exists
root at px-bravo-cluster:~# rbd -p vdisks create test --size 1G
root at px-bravo-cluster:~# rbd -p vdisks ls
test
vm-106-disk-0
vm-113-disk-0
vm-113-disk-1
vm-113-disk-2
vm-118-disk-0
vm-119-disk-0
vm-120-disk-0
vm-125-disk-0
vm-125-disk-1
root at px-bravo-cluster:~# rbd -p vdisks rm test
Removing image: 100% complete...done.
root at px-bravo-cluster:~# rbd -p vdisks rm vm-112-disk-0
2019-09-06 11:36:07.570749 7eff7cff9700 -1 librbd::image::OpenRequest: failed to retreive immutable metadata: (2) No such file or
directory
Removing image: 0% complete...failed.
rbd: delete error: (2) No such file or directory
>
>>
>> Here is the relevant part of my storage.cfg:
>>
>> ########
>> nfs: aurel-cluster1-VMs
>> export /backup/proxmox-infra/VMs
>> path /mnt/pve/aurel-cluster1-VMs
>> server X.X.X.X
>> content images
>> options vers=4.2
>>
>>
>> rbd: vdisks_vm
>> content images
>> krbd 0
>> pool vdisks
>> ########
> Is this the complete storage.cfg?
No, only the parts that are relevant for this particular move. Here's the complete file:
########
rbd: vdisks_vm
content images
krbd 0
pool vdisks
dir: local-hdd
path /mnt/local
content images,iso
nodes px-alpha-cluster,px-bravo-cluster,px-charlie-cluster
shared 0
nfs: aurel-cluster1-daily
export /backup/proxmox-infra/daily
path /mnt/pve/aurel-cluster1-daily
server X.X.X.X
content backup
maxfiles 30
options vers=4.2
nfs: aurel-cluster1-weekly
export /backup/proxmox-infra/weekly
path /mnt/pve/aurel-cluster1-weekly
server X.X.X.X
content backup
maxfiles 30
options vers=4.2
nfs: aurel-cluster1-VMs
export /backup/proxmox-infra/VMs
path /mnt/pve/aurel-cluster1-VMs
server X.X.X.X
content images
options vers=4.2
nfs: aurel-cluster2-daily
export /backup/proxmox-infra2/daily
path /mnt/pve/aurel-cluster2-daily
server X.X.X.X
content backup
maxfiles 30
options vers=4.2
nfs: aurel-cluster2-weekly
export /backup/proxmox-infra2/weekly
path /mnt/pve/aurel-cluster2-weekly
server X.X.X.X
content backup
maxfiles 30
options vers=4.2
nfs: aurel-cluster2-VMs
export /backup/proxmox-infra2/VMs
path /mnt/pve/aurel-cluster2-VMs
server X.X.X.X
content images
options vers=4.2
dir: local
path /var/lib/vz
content snippets,vztmpl,images,rootdir,iso
maxfiles 0
rbd: vdisks_cluster2
content images
krbd 0
monhost px-golf-cluster, px-hotel-cluster, px-india-cluster
pool vdisks
username admin
########
Thanks,
Uwe
> --
> Cheers,
> Alwin
>
More information about the pve-user
mailing list