[PVE-User] Create proxmox cluster / storage question.
Leandro Roggerone
leandro at tecnetmza.com.ar
Fri Feb 28 15:25:43 CET 2020
Dear Eneko:
When you talk about redundacy and availability, do you want...
- HA? (automatic restart of VMs in the other node in case one server fails)
- Be able to move VMs from one server to the other "fast" (without
copying the disks)?
Yes , this is what I want.
So regarding my original layout from my 5.5Tb storage:
Im using 1T for LVM , 1TB for LVM-thin and 3.5 TB unassigned space, is it
ok to use this unassigned space for a ceph ?
Can I set it later ? with server on production ?
Other:
Using an NFS system, means to have an external server running a file sistem?
So you should have at least two servers for the cluster and one for the
file system?
It seems to me that using ceph has a better redundancy plan an it is easier
to deploy since I only need two servers. (am i right?).
Thanks!
El vie., 28 feb. 2020 a las 11:06, Eneko Lacunza (<elacunza at binovo.es>)
escribió:
> Hola Leandro,
>
> El 28/2/20 a las 14:43, Leandro Roggerone escribió:
> > Regarding your question , what is the tarjet use for this server.
> > I have a dell R610 with 6 drive bays.
> > Today I have 4 (2TB) drives in Raid5 , resulting a 5.5TB capacity.
> > I will add 2 ssd drives later in raid1 for applications that need more
> read
> > speed.
> > The purpose for this server is to run proxmox with some VMs for external
> > and internal access.
> > Im planning to build a second server and create a cluster just to have
> more
> > redundancy and availability.
> >
> > I would like to set all I can now that server is not in production and
> > minimize risk later.
> > Thats why im asking so many questions.
> Asking is good, but we need info to be able to help you ;)
>
> When you talk about redundacy and availability, do you want...
> - HA? (automatic restart of VMs in the other node in case one server fails)
> - Be able to move VMs from one server to the other "fast" (without
> copying the disks)?
>
> If your answer is yes to any of the previous questions, you have to look
> at using a NFS server or deploying Ceph.
>
> If it's no, then we can talk about local storage in your servers. What
> RAID card do you have in that server? Does it have write cache (non
> volatile of battery-backed) If it doesn't have such, RAID5 could prove
> slow (eat quite CPU), I suggest you use 2xRAID1 or a RAID10 setup. Also,
> please bear in mind that RAID5 with "big" disks is considered quite
> unsecure (risk of having a second disk failure during recovery is high).
>
> Saludos
> Eneko
> > Regards.
> > Leandro.
> >
> >
> > El vie., 28 feb. 2020 a las 5:49, Eneko Lacunza (<elacunza at binovo.es>)
> > escribió:
> >
> >> Hola Leandro,
> >>
> >> El 27/2/20 a las 17:29, Leandro Roggerone escribió:
> >>> Hi guys , i'm still tunning my 5.5 Tb server.
> >>> While setting storage options during install process, I set 2000 for hd
> >>> size, so I have 3.5 TB free to assign later.
> >>>
> >>> my layout is as follows:
> >>> root at pve:~# lsblk
> >>> NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
> >>> sda 8:0 0 5.5T 0 disk
> >>> ├─sda1 8:1 0 1007K 0 part
> >>> ├─sda2 8:2 0 512M 0 part
> >>> └─sda3 8:3 0 2T 0 part
> >>> ├─pve-swap 253:0 0 8G 0 lvm [SWAP]
> >>> ├─pve-root 253:1 0 1T 0 lvm /
> >>> ├─pve-data_tmeta 253:2 0 9G 0 lvm
> >>> │ └─pve-data 253:4 0 949.6G 0 lvm
> >>> └─pve-data_tdata 253:3 0 949.6G 0 lvm
> >>> └─pve-data 253:4 0 949.6G 0 lvm
> >>> sr0 11:0 1 1024M 0 rom
> >>>
> >>> My question is:
> >>> Is it possible to expand sda3 partition later without service outage ?
> >>> Is it possible to expand pve group on sda3 partition ?
> >> You don't need to expand sda3 really. You can just create a new
> >> partition, create a new PV with it and add the new PV to pve VG.
> >>
> >>> In case to create a proxmox cluster, what should I do with that 3.5 TB
> >> free
> >>> ?
> >> I don't know really how to reply to this. If you're building a cluster,
> >> I suggest you configure some kind of shared storage; NFS server or Ceph
> >> cluster for example.
> >>
> >>> Is there a best partition type suited for this ? Can I do it without
> >>> service outage?
> >> For what?
> >>
> >>> I have not any service running yet , so I can experiment what it takes.
> >>> Any thought about this would be great.
> >> Maybe you can start telling us your target use for this server/cluster.
> >> Also some detailed spec of the server would help; for example does it
> >> have a RAID card with more than one disk, or you're using a 6TB single
> >> disk?
> >>
> >> Cheers
> >> Eneko
> >>
> >> --
> >> Zuzendari Teknikoa / Director Técnico
> >> Binovo IT Human Project, S.L.
> >> Telf. 943569206
> >> Astigarragako bidea 2, 2º izq. oficina 11; 20180 Oiartzun (Gipuzkoa)
> >> www.binovo.es
> >>
> >> _______________________________________________
> >> pve-user mailing list
> >> pve-user at pve.proxmox.com
> >> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> >>
> > _______________________________________________
> > pve-user mailing list
> > pve-user at pve.proxmox.com
> > https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>
>
> --
> Zuzendari Teknikoa / Director Técnico
> Binovo IT Human Project, S.L.
> Telf. 943569206
> Astigarragako bidea 2, 2º izq. oficina 11; 20180 Oiartzun (Gipuzkoa)
> www.binovo.es
>
> _______________________________________________
> pve-user mailing list
> pve-user at pve.proxmox.com
> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>
More information about the pve-user
mailing list