[pve-devel] multipath problems
    Alexandre DERUMIER 
    aderumier at odiso.com
       
    Wed Aug 22 08:00:32 CEST 2012
    
    
  
Hi Dietmar, I'll try to help you.
what is your hardware config ?( cpu/ram/disk).
what's your raid level ? (raid-z, mirrors)
do you use sas or sata disk?
some quick performnce tuning,through gui:
:2000/settings/preferences/?this_form=sys_form
zfs_no_cacheflush=yes
Sys_zfs_vdev_max_pending = 1 for sata, = 10 for sas
Also if you want to have more sequential bandwith, you can try to use big block size for zvol, like 128K. (I setup 4K by default in my module).
about nexenta caches:
For read, nexenta use 2cache layer, 1rst layer ram, 2nd layer(optionnal) ssd (named l2arc).
for write, nexenta can use ssd or nvram devices (slog devices), to write datas randomly and flush them sequentially to disks each Xseconds.
I never do hdparm test, so I can't compare, but maybe can you try with fio ?
random read :
fio --filename=/dev/dm-20 --direct=1 --rw=randread --bs=4k --size=20G --numjobs=200 --runtime=60 --group_reporting --name=file1
randomwrite  
fio --filename=/dev/dm-20 --direct=1 --rw=randwrite --bs=4k --size=20G --numjobs=200 --runtime=60 --group_reporting --name=file1
seq read
fio --filename=/dev/dm-20 --direct=1 --rw=read --bs=1m --size=20G --numjobs=200 --runtime=60 --group_reporting --name=file1 
seq write: 
fio --filename=/dev/dm-20 --direct=1 --rw=write --bs=1m --size=5G --numjobs=200 --runtime=60 --group_reporting --name=file1 
----- Mail original ----- 
De: "Dietmar Maurer" <dietmar at proxmox.com> 
À: pve-devel at pve.proxmox.com 
Envoyé: Mercredi 22 Août 2012 07:28:08 
Objet: [pve-devel] multipath problems 
Hi all, 
I am just setting up a nexenta storage (community edition) and try to use multipath on the proxmox side: 
# multipath -l 
mpath1 (3600144f0001e58febc44503399600001) dm-3 NEXENTA,NEXENTASTOR 
size=100G features='1 queue_if_no_path' hwhandler='0' wp=rw 
`-+- policy='round-robin 0' prio=-3 status=active 
|- 5:0:0:0 sdd 8:48 active undef running 
|- 3:0:0:0 sdb 8:16 active undef running 
`- 4:0:0:0 sdc 8:32 active undef running 
I assume above output is OK so far. 
The strange thing is that we get very poor performance. When I test a 
single iscsi drive I get: 
# hdparm -t /dev/sdc 
/dev/sdc: 
Timing buffered disk reads: 338 MB in 3.00 seconds = 112.48 MB/sec 
root at dell1:~# hdparm -t /dev/sdb 
/dev/sdb: 
Timing buffered disk reads: 222 MB in 3.01 seconds = 73.69 MB/sec 
root at dell1:~# hdparm -t /dev/sdd 
/dev/sdd: 
Timing buffered disk reads: 262 MB in 3.01 seconds = 87.07 MB/sec 
When I test the multipath drive I get: 
# hdparm -t /dev/mapper/mpath1 
/dev/mapper/mpath1: 
Timing buffered disk reads: 266 MB in 3.02 seconds = 88.04 MB/sec 
So there is no speed gain. I already checked that traffic is routed to different network cards. 
Any idea why it is that slow? 
- Dietmar 
_______________________________________________ 
pve-devel mailing list 
pve-devel at pve.proxmox.com 
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-devel 
-- 
-- 
	
Alexandre D e rumier 
Ingénieur Systèmes et Réseaux 
Fixe : 03 20 68 88 85 
Fax : 03 20 68 90 88 
45 Bvd du Général Leclerc 59100 Roubaix 
12 rue Marivaux 75002 Paris 
    
    
More information about the pve-devel
mailing list