[PVE-User] Migration error!

Gilberto Nunes gilberto.nunes32 at gmail.com
Fri Aug 25 15:11:53 CEST 2017


So.. One of the premise of the ZFS Replication volume, is to replicate
local volume to another node.
Or am I wrong?


Obrigado

Cordialmente


Gilberto Ferreira

Consultor TI Linux | IaaS Proxmox, CloudStack, KVM | Zentyal Server |
Zimbra Mail Server

(47) 3025-5907
(47) 99676-7530

Skype: gilberto.nunes36


konnectati.com.br <http://www.konnectati.com.br/>


https://www.youtube.com/watch?v=dsiTPeNWcSE


2017-08-25 10:07 GMT-03:00 Fábio Rabelo <fabio at fabiorabelo.wiki.br>:

> this entry :
>
> 2017-08-25 09:24:44 can't migrate local disk 'stg:vm-100-disk-1': can't
> live migrate attached local disks without with-local-disks option
>
> Seems to be the responsable .
>
> Local disk ?
>
> where this image are stored ?
>
>
> Fábio Rabelo
>
> 2017-08-25 9:36 GMT-03:00 Gilberto Nunes <gilberto.nunes32 at gmail.com>:
> > If I turn off the VM, migrate goes on.
> > But make offline migration is out of the question!!!
> >
> >
> >
> > 2017-08-25 9:28 GMT-03:00 Gilberto Nunes <gilberto.nunes32 at gmail.com>:
> >
> >> Hi again
> >>
> >> I try remove all replication jobs and image files from target node...
> >> Still get critical error:
> >>
> >> qm migrate 100 prox02 --online
> >> 2017-08-25 09:24:43 starting migration of VM 100 to node 'prox02'
> >> (10.1.1.20)
> >> 2017-08-25 09:24:44 found local disk 'stg:vm-100-disk-1' (in current VM
> >> config)
> >> 2017-08-25 09:24:44 can't migrate local disk 'stg:vm-100-disk-1': can't
> >> live migrate attached local disks without with-local-disks option
> >> 2017-08-25 09:24:44 ERROR: Failed to sync data - can't migrate VM -
> check
> >> log
> >> 2017-08-25 09:24:44 aborting phase 1 - cleanup resources
> >> 2017-08-25 09:24:44 ERROR: migration aborted (duration 00:00:02): Failed
> >> to sync data - can't migrate VM - check log
> >> migration aborted
> >> prox01:~# qm migrate 100 prox02 --online --with-local-disks
> >> 2017-08-25 09:24:58 starting migration of VM 100 to node 'prox02'
> >> (10.1.1.20)
> >> 2017-08-25 09:24:58 found local disk 'stg:vm-100-disk-1' (in current VM
> >> config)
> >> 2017-08-25 09:24:58 copying disk images
> >> 2017-08-25 09:24:58 ERROR: Failed to sync data - can't live migrate VM
> >> with replicated volumes
> >> 2017-08-25 09:24:58 aborting phase 1 - cleanup resources
> >> 2017-08-25 09:24:58 ERROR: migration aborted (duration 00:00:01): Failed
> >> to sync data - can't live migrate VM with replicated volumes
> >> migration aborted
> >> prox01:~# pvesr status
> >> JobID      Enabled    Target                           LastSync
> >>   NextSync   Duration  FailCount State
> >> 100-0      Yes        local/prox02          2017-08-25_09:25:01
> >>  2017-08-25_12:00:00  15.200315          0 OK
> >>
> >> Somebody help me!
> >>
> >> Cheers
> >>
> >>
> >>
> >>
> >> 2017-08-24 9:55 GMT-03:00 Gilberto Nunes <gilberto.nunes32 at gmail.com>:
> >>
> >>> Well...
> >>> I will try it
> >>>
> >>> Thanks
> >>>
> >>>
> >>>
> >>>
> >>> 2017-08-24 4:37 GMT-03:00 Dominik Csapak <d.csapak at proxmox.com>:
> >>>
> >>>> On 08/23/2017 08:50 PM, Gilberto Nunes wrote:
> >>>>
> >>>>> more info:
> >>>>>
> >>>>>
> >>>>> pvesr status
> >>>>> JobID      Enabled    Target                           LastSync
> >>>>> NextSync   Duration  FailCount State
> >>>>> 100-0      Yes        local/prox01                            -
> >>>>>   2017-08-23_15:55:04   3.151884          1 command 'set -o pipefail
> &&
> >>>>> pvesm export local-zfs:vm-100-disk-1 zfs - -with-snapshots 1
> -snapshot
> >>>>> __replicate_100-0_1503514204__ | /usr/bin/cstream -t 1024000000 |
> >>>>> /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=prox01'
> root at 10.1.1.10
> >>>>> --
> >>>>> pvesm import local-zfs:vm-100-disk-1 zfs - -with-snapshots 1' failed:
> >>>>> exit
> >>>>> code 255
> >>>>> 100-1      Yes        local/prox02                            -
> >>>>>   2017-08-23_15:55:01   3.089044          1 command 'set -o pipefail
> &&
> >>>>> pvesm export local-zfs:vm-100-disk-1 zfs - -with-snapshots 1
> -snapshot
> >>>>> __replicate_100-1_1503514201__ | /usr/bin/cstream -t 1024000000 |
> >>>>> /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=prox02'
> root at 10.1.1.20
> >>>>> --
> >>>>> pvesm import local-zfs:vm-100-disk-1 zfs - -with-snapshots 1' failed:
> >>>>> exit
> >>>>> code 255
> >>>>>
> >>>>>
> >>>>>
> >>>> according to this output, no lastsync was completed, so i guess the
> >>>> replication did never work, so the migration will also not worK?
> >>>>
> >>>> i would remove all replication jobs (maybe with -force, via
> commandline)
> >>>> delete all images of this vm from all nodes where the vm *not* is at
> the
> >>>> moment (afaics from prox01 and prox02, as the vm is currently on
> prox03)
> >>>>
> >>>> then add the replication again wait for it to complete (verify with
> >>>> pvesr status) and try again to migrate
> >>>>
> >>>> _______________________________________________
> >>>> pve-user mailing list
> >>>> pve-user at pve.proxmox.com
> >>>> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> >>>>
> >>>
> >>>
> >>
> > _______________________________________________
> > pve-user mailing list
> > pve-user at pve.proxmox.com
> > https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> _______________________________________________
> pve-user mailing list
> pve-user at pve.proxmox.com
> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>



More information about the pve-user mailing list