[PVE-User] Migration error!

Gilberto Nunes gilberto.nunes32 at gmail.com
Fri Aug 25 16:14:00 CEST 2017


Yeah!
That's explain the fact storage replication works fine here with a CT!
Thanks for remind!

2017-08-25 11:11 GMT-03:00 Yannis Milios <yannis.milios at gmail.com>:

> My understanding is that in pvesr the live migration of guest vm is not
> supported:
>
> "Virtual guest with active replication cannot currently use online
> migration. Offline migration is supported in general"
>
> On Fri, 25 Aug 2017 at 16:48, Fábio Rabelo <fabio at fabiorabelo.wiki.br>
> wrote:
>
> > Sorry .... my knowledge do not go beyond here ...
> >
> > I abandoned shared storage years ago for lack of trustworthy
> >
> >
> > Fábio Rabelo
> >
> > 2017-08-25 10:22 GMT-03:00 Gilberto Nunes <gilberto.nunes32 at gmail.com>:
> > > According to the design model of Proxmox Storage Replication, there is
> a
> > > schedule to make the sync.
> > > And of course, I set up the VM, I have scheduled the sync and for
> finish.
> > > But still stuck!
> > >
> > >
> > >
> > >
> > >
> > > 2017-08-25 10:19 GMT-03:00 Fábio Rabelo <fabio at fabiorabelo.wiki.br>:
> > >
> > >> I never used zfs on Linux .
> > >>
> > >> But, in the Solaris OS family, this replication must be set up
> > beforehand
> > >> ...
> > >>
> > >> Someone with some milestone with zfs on linux can confirm or deny that
> > ??
> > >>
> > >>
> > >> Fábio Rabelo
> > >>
> > >> 2017-08-25 10:11 GMT-03:00 Gilberto Nunes <gilberto.nunes32 at gmail.com
> >:
> > >> > So.. One of the premise of the ZFS Replication volume, is to
> replicate
> > >> > local volume to another node.
> > >> > Or am I wrong?
> > >> >
> > >> >
> > >> > Obrigado
> > >> >
> > >> > Cordialmente
> > >> >
> > >> >
> > >> > Gilberto Ferreira
> > >> >
> > >> > Consultor TI Linux | IaaS Proxmox, CloudStack, KVM | Zentyal Server
> |
> > >> > Zimbra Mail Server
> > >> >
> > >> > (47) 3025-5907
> > >> > (47) 99676-7530
> > >> >
> > >> > Skype: gilberto.nunes36
> > >> >
> > >> >
> > >> > konnectati.com.br <http://www.konnectati.com.br/>
> > >> >
> > >> >
> > >> > https://www.youtube.com/watch?v=dsiTPeNWcSE
> > >> >
> > >> >
> > >> > 2017-08-25 10:07 GMT-03:00 Fábio Rabelo <fabio at fabiorabelo.wiki.br
> >:
> > >> >
> > >> >> this entry :
> > >> >>
> > >> >> 2017-08-25 09:24:44 can't migrate local disk 'stg:vm-100-disk-1':
> > can't
> > >> >> live migrate attached local disks without with-local-disks option
> > >> >>
> > >> >> Seems to be the responsable .
> > >> >>
> > >> >> Local disk ?
> > >> >>
> > >> >> where this image are stored ?
> > >> >>
> > >> >>
> > >> >> Fábio Rabelo
> > >> >>
> > >> >> 2017-08-25 9:36 GMT-03:00 Gilberto Nunes <
> gilberto.nunes32 at gmail.com
> > >:
> > >> >> > If I turn off the VM, migrate goes on.
> > >> >> > But make offline migration is out of the question!!!
> > >> >> >
> > >> >> >
> > >> >> >
> > >> >> > 2017-08-25 9:28 GMT-03:00 Gilberto Nunes <
> > gilberto.nunes32 at gmail.com
> > >> >:
> > >> >> >
> > >> >> >> Hi again
> > >> >> >>
> > >> >> >> I try remove all replication jobs and image files from target
> > node...
> > >> >> >> Still get critical error:
> > >> >> >>
> > >> >> >> qm migrate 100 prox02 --online
> > >> >> >> 2017-08-25 09:24:43 starting migration of VM 100 to node
> 'prox02'
> > >> >> >> (10.1.1.20)
> > >> >> >> 2017-08-25 09:24:44 found local disk 'stg:vm-100-disk-1' (in
> > current
> > >> VM
> > >> >> >> config)
> > >> >> >> 2017-08-25 09:24:44 can't migrate local disk
> 'stg:vm-100-disk-1':
> > >> can't
> > >> >> >> live migrate attached local disks without with-local-disks
> option
> > >> >> >> 2017-08-25 09:24:44 ERROR: Failed to sync data - can't migrate
> VM
> > -
> > >> >> check
> > >> >> >> log
> > >> >> >> 2017-08-25 09:24:44 aborting phase 1 - cleanup resources
> > >> >> >> 2017-08-25 09:24:44 ERROR: migration aborted (duration
> 00:00:02):
> > >> Failed
> > >> >> >> to sync data - can't migrate VM - check log
> > >> >> >> migration aborted
> > >> >> >> prox01:~# qm migrate 100 prox02 --online --with-local-disks
> > >> >> >> 2017-08-25 09:24:58 starting migration of VM 100 to node
> 'prox02'
> > >> >> >> (10.1.1.20)
> > >> >> >> 2017-08-25 09:24:58 found local disk 'stg:vm-100-disk-1' (in
> > current
> > >> VM
> > >> >> >> config)
> > >> >> >> 2017-08-25 09:24:58 copying disk images
> > >> >> >> 2017-08-25 09:24:58 ERROR: Failed to sync data - can't live
> > migrate
> > >> VM
> > >> >> >> with replicated volumes
> > >> >> >> 2017-08-25 09:24:58 aborting phase 1 - cleanup resources
> > >> >> >> 2017-08-25 09:24:58 ERROR: migration aborted (duration
> 00:00:01):
> > >> Failed
> > >> >> >> to sync data - can't live migrate VM with replicated volumes
> > >> >> >> migration aborted
> > >> >> >> prox01:~# pvesr status
> > >> >> >> JobID      Enabled    Target                           LastSync
> > >> >> >>   NextSync   Duration  FailCount State
> > >> >> >> 100-0      Yes        local/prox02          2017-08-25_09:25:01
> > >> >> >>  2017-08-25_12:00:00  15.200315          0 OK
> > >> >> >>
> > >> >> >> Somebody help me!
> > >> >> >>
> > >> >> >> Cheers
> > >> >> >>
> > >> >> >>
> > >> >> >>
> > >> >> >>
> > >> >> >> 2017-08-24 9:55 GMT-03:00 Gilberto Nunes <
> > gilberto.nunes32 at gmail.com
> > >> >:
> > >> >> >>
> > >> >> >>> Well...
> > >> >> >>> I will try it
> > >> >> >>>
> > >> >> >>> Thanks
> > >> >> >>>
> > >> >> >>>
> > >> >> >>>
> > >> >> >>>
> > >> >> >>> 2017-08-24 4:37 GMT-03:00 Dominik Csapak <d.csapak at proxmox.com
> >:
> > >> >> >>>
> > >> >> >>>> On 08/23/2017 08:50 PM, Gilberto Nunes wrote:
> > >> >> >>>>
> > >> >> >>>>> more info:
> > >> >> >>>>>
> > >> >> >>>>>
> > >> >> >>>>> pvesr status
> > >> >> >>>>> JobID      Enabled    Target
>  LastSync
> > >> >> >>>>> NextSync   Duration  FailCount State
> > >> >> >>>>> 100-0      Yes        local/prox01
>   -
> > >> >> >>>>>   2017-08-23_15:55:04   3.151884          1 command 'set -o
> > >> pipefail
> > >> >> &&
> > >> >> >>>>> pvesm export local-zfs:vm-100-disk-1 zfs - -with-snapshots 1
> > >> >> -snapshot
> > >> >> >>>>> __replicate_100-0_1503514204__ | /usr/bin/cstream -t
> > 1024000000 |
> > >> >> >>>>> /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=prox01'
> > >> >> root at 10.1.1.10
> > >> >> >>>>> --
> > >> >> >>>>> pvesm import local-zfs:vm-100-disk-1 zfs - -with-snapshots 1'
> > >> failed:
> > >> >> >>>>> exit
> > >> >> >>>>> code 255
> > >> >> >>>>> 100-1      Yes        local/prox02
>   -
> > >> >> >>>>>   2017-08-23_15:55:01   3.089044          1 command 'set -o
> > >> pipefail
> > >> >> &&
> > >> >> >>>>> pvesm export local-zfs:vm-100-disk-1 zfs - -with-snapshots 1
> > >> >> -snapshot
> > >> >> >>>>> __replicate_100-1_1503514201__ | /usr/bin/cstream -t
> > 1024000000 |
> > >> >> >>>>> /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=prox02'
> > >> >> root at 10.1.1.20
> > >> >> >>>>> --
> > >> >> >>>>> pvesm import local-zfs:vm-100-disk-1 zfs - -with-snapshots 1'
> > >> failed:
> > >> >> >>>>> exit
> > >> >> >>>>> code 255
> > >> >> >>>>>
> > >> >> >>>>>
> > >> >> >>>>>
> > >> >> >>>> according to this output, no lastsync was completed, so i
> guess
> > the
> > >> >> >>>> replication did never work, so the migration will also not
> worK?
> > >> >> >>>>
> > >> >> >>>> i would remove all replication jobs (maybe with -force, via
> > >> >> commandline)
> > >> >> >>>> delete all images of this vm from all nodes where the vm *not*
> > is
> > >> at
> > >> >> the
> > >> >> >>>> moment (afaics from prox01 and prox02, as the vm is currently
> on
> > >> >> prox03)
> > >> >> >>>>
> > >> >> >>>> then add the replication again wait for it to complete (verify
> > with
> > >> >> >>>> pvesr status) and try again to migrate
> > >> >> >>>>
> > >> >> >>>> _______________________________________________
> > >> >> >>>> pve-user mailing list
> > >> >> >>>> pve-user at pve.proxmox.com
> > >> >> >>>> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> > >> >> >>>>
> > >> >> >>>
> > >> >> >>>
> > >> >> >>
> > >> >> > _______________________________________________
> > >> >> > pve-user mailing list
> > >> >> > pve-user at pve.proxmox.com
> > >> >> > https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> > >> >> _______________________________________________
> > >> >> pve-user mailing list
> > >> >> pve-user at pve.proxmox.com
> > >> >> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> > >> >>
> > >> > _______________________________________________
> > >> > pve-user mailing list
> > >> > pve-user at pve.proxmox.com
> > >> > https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> > >> _______________________________________________
> > >> pve-user mailing list
> > >> pve-user at pve.proxmox.com
> > >> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> > >>
> > > _______________________________________________
> > > pve-user mailing list
> > > pve-user at pve.proxmox.com
> > > https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> > _______________________________________________
> > pve-user mailing list
> > pve-user at pve.proxmox.com
> > https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> >
> --
> Sent from Gmail Mobile
> _______________________________________________
> pve-user mailing list
> pve-user at pve.proxmox.com
> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>



More information about the pve-user mailing list