[PVE-User] Migration error!

Gilberto Nunes gilberto.nunes32 at gmail.com
Fri Aug 25 14:36:34 CEST 2017


If I turn off the VM, migrate goes on.
But make offline migration is out of the question!!!



2017-08-25 9:28 GMT-03:00 Gilberto Nunes <gilberto.nunes32 at gmail.com>:

> Hi again
>
> I try remove all replication jobs and image files from target node...
> Still get critical error:
>
> qm migrate 100 prox02 --online
> 2017-08-25 09:24:43 starting migration of VM 100 to node 'prox02'
> (10.1.1.20)
> 2017-08-25 09:24:44 found local disk 'stg:vm-100-disk-1' (in current VM
> config)
> 2017-08-25 09:24:44 can't migrate local disk 'stg:vm-100-disk-1': can't
> live migrate attached local disks without with-local-disks option
> 2017-08-25 09:24:44 ERROR: Failed to sync data - can't migrate VM - check
> log
> 2017-08-25 09:24:44 aborting phase 1 - cleanup resources
> 2017-08-25 09:24:44 ERROR: migration aborted (duration 00:00:02): Failed
> to sync data - can't migrate VM - check log
> migration aborted
> prox01:~# qm migrate 100 prox02 --online --with-local-disks
> 2017-08-25 09:24:58 starting migration of VM 100 to node 'prox02'
> (10.1.1.20)
> 2017-08-25 09:24:58 found local disk 'stg:vm-100-disk-1' (in current VM
> config)
> 2017-08-25 09:24:58 copying disk images
> 2017-08-25 09:24:58 ERROR: Failed to sync data - can't live migrate VM
> with replicated volumes
> 2017-08-25 09:24:58 aborting phase 1 - cleanup resources
> 2017-08-25 09:24:58 ERROR: migration aborted (duration 00:00:01): Failed
> to sync data - can't live migrate VM with replicated volumes
> migration aborted
> prox01:~# pvesr status
> JobID      Enabled    Target                           LastSync
>   NextSync   Duration  FailCount State
> 100-0      Yes        local/prox02          2017-08-25_09:25:01
>  2017-08-25_12:00:00  15.200315          0 OK
>
> Somebody help me!
>
> Cheers
>
>
>
>
> 2017-08-24 9:55 GMT-03:00 Gilberto Nunes <gilberto.nunes32 at gmail.com>:
>
>> Well...
>> I will try it
>>
>> Thanks
>>
>>
>>
>>
>> 2017-08-24 4:37 GMT-03:00 Dominik Csapak <d.csapak at proxmox.com>:
>>
>>> On 08/23/2017 08:50 PM, Gilberto Nunes wrote:
>>>
>>>> more info:
>>>>
>>>>
>>>> pvesr status
>>>> JobID      Enabled    Target                           LastSync
>>>> NextSync   Duration  FailCount State
>>>> 100-0      Yes        local/prox01                            -
>>>>   2017-08-23_15:55:04   3.151884          1 command 'set -o pipefail &&
>>>> pvesm export local-zfs:vm-100-disk-1 zfs - -with-snapshots 1 -snapshot
>>>> __replicate_100-0_1503514204__ | /usr/bin/cstream -t 1024000000 |
>>>> /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=prox01' root at 10.1.1.10
>>>> --
>>>> pvesm import local-zfs:vm-100-disk-1 zfs - -with-snapshots 1' failed:
>>>> exit
>>>> code 255
>>>> 100-1      Yes        local/prox02                            -
>>>>   2017-08-23_15:55:01   3.089044          1 command 'set -o pipefail &&
>>>> pvesm export local-zfs:vm-100-disk-1 zfs - -with-snapshots 1 -snapshot
>>>> __replicate_100-1_1503514201__ | /usr/bin/cstream -t 1024000000 |
>>>> /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=prox02' root at 10.1.1.20
>>>> --
>>>> pvesm import local-zfs:vm-100-disk-1 zfs - -with-snapshots 1' failed:
>>>> exit
>>>> code 255
>>>>
>>>>
>>>>
>>> according to this output, no lastsync was completed, so i guess the
>>> replication did never work, so the migration will also not worK?
>>>
>>> i would remove all replication jobs (maybe with -force, via commandline)
>>> delete all images of this vm from all nodes where the vm *not* is at the
>>> moment (afaics from prox01 and prox02, as the vm is currently on prox03)
>>>
>>> then add the replication again wait for it to complete (verify with
>>> pvesr status) and try again to migrate
>>>
>>> _______________________________________________
>>> pve-user mailing list
>>> pve-user at pve.proxmox.com
>>> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>>>
>>
>>
>



More information about the pve-user mailing list