[PVE-User] Migration error!
Gilberto Nunes
gilberto.nunes32 at gmail.com
Fri Aug 25 14:28:10 CEST 2017
Hi again
I try remove all replication jobs and image files from target node...
Still get critical error:
qm migrate 100 prox02 --online
2017-08-25 09:24:43 starting migration of VM 100 to node 'prox02'
(10.1.1.20)
2017-08-25 09:24:44 found local disk 'stg:vm-100-disk-1' (in current VM
config)
2017-08-25 09:24:44 can't migrate local disk 'stg:vm-100-disk-1': can't
live migrate attached local disks without with-local-disks option
2017-08-25 09:24:44 ERROR: Failed to sync data - can't migrate VM - check
log
2017-08-25 09:24:44 aborting phase 1 - cleanup resources
2017-08-25 09:24:44 ERROR: migration aborted (duration 00:00:02): Failed to
sync data - can't migrate VM - check log
migration aborted
prox01:~# qm migrate 100 prox02 --online --with-local-disks
2017-08-25 09:24:58 starting migration of VM 100 to node 'prox02'
(10.1.1.20)
2017-08-25 09:24:58 found local disk 'stg:vm-100-disk-1' (in current VM
config)
2017-08-25 09:24:58 copying disk images
2017-08-25 09:24:58 ERROR: Failed to sync data - can't live migrate VM with
replicated volumes
2017-08-25 09:24:58 aborting phase 1 - cleanup resources
2017-08-25 09:24:58 ERROR: migration aborted (duration 00:00:01): Failed to
sync data - can't live migrate VM with replicated volumes
migration aborted
prox01:~# pvesr status
JobID Enabled Target LastSync
NextSync Duration FailCount State
100-0 Yes local/prox02 2017-08-25_09:25:01
2017-08-25_12:00:00 15.200315 0 OK
Somebody help me!
Cheers
2017-08-24 9:55 GMT-03:00 Gilberto Nunes <gilberto.nunes32 at gmail.com>:
> Well...
> I will try it
>
> Thanks
>
>
>
>
> 2017-08-24 4:37 GMT-03:00 Dominik Csapak <d.csapak at proxmox.com>:
>
>> On 08/23/2017 08:50 PM, Gilberto Nunes wrote:
>>
>>> more info:
>>>
>>>
>>> pvesr status
>>> JobID Enabled Target LastSync
>>> NextSync Duration FailCount State
>>> 100-0 Yes local/prox01 -
>>> 2017-08-23_15:55:04 3.151884 1 command 'set -o pipefail &&
>>> pvesm export local-zfs:vm-100-disk-1 zfs - -with-snapshots 1 -snapshot
>>> __replicate_100-0_1503514204__ | /usr/bin/cstream -t 1024000000 |
>>> /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=prox01' root at 10.1.1.10
>>> --
>>> pvesm import local-zfs:vm-100-disk-1 zfs - -with-snapshots 1' failed:
>>> exit
>>> code 255
>>> 100-1 Yes local/prox02 -
>>> 2017-08-23_15:55:01 3.089044 1 command 'set -o pipefail &&
>>> pvesm export local-zfs:vm-100-disk-1 zfs - -with-snapshots 1 -snapshot
>>> __replicate_100-1_1503514201__ | /usr/bin/cstream -t 1024000000 |
>>> /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=prox02' root at 10.1.1.20
>>> --
>>> pvesm import local-zfs:vm-100-disk-1 zfs - -with-snapshots 1' failed:
>>> exit
>>> code 255
>>>
>>>
>>>
>> according to this output, no lastsync was completed, so i guess the
>> replication did never work, so the migration will also not worK?
>>
>> i would remove all replication jobs (maybe with -force, via commandline)
>> delete all images of this vm from all nodes where the vm *not* is at the
>> moment (afaics from prox01 and prox02, as the vm is currently on prox03)
>>
>> then add the replication again wait for it to complete (verify with pvesr
>> status) and try again to migrate
>>
>> _______________________________________________
>> pve-user mailing list
>> pve-user at pve.proxmox.com
>> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>>
>
>
More information about the pve-user
mailing list