[pve-devel] Quorum problems with NICs Intel of 10 Gb/s and VMsturns off

Alexandre DERUMIER aderumier at odiso.com
Tue Jan 6 04:37:33 CET 2015


>>And about of your suggestion:
>>-A PVEFW-HOST-IN -s yournetwork/24 -p udp -m addrtype --dst-type 
>>MULTICAST -m udp --dport 5404:5405 -j RETURN

Note that this is the rules for HOST-IN,
if you want to adapt you can do

-A FORWARD -s yournetwork/24 -p udp -m addrtype --dst-type MULTICAST -m udp --dport 5404:5405 -j DROP

>>1) ¿Such rule will avoid the cluster communication to the VMs?
>>2) ¿Such rule will not prejudice the normal use of the igmp protocol own of 
the Windows systems in the VMs?

this will block multicast traffic, on udp port 5404:5405 (corosync default port), from your source network.

>>3) If both answers are correct, where i should put the rule that you 
suggest?

Currently it's not possible to do it with proxmox firewall,
but you can add it in rc.local for example.

iptables -A FORWARD -s yournetwork/24 -p udp -m addrtype --dst-type MULTICAST -m udp --dport 5404:5405 -j DROP

Proxmox firewall don't override custom rules


----- Mail original -----
De: "Cesar Peschiera" <brain at click.com.py>
À: "aderumier" <aderumier at odiso.com>, "dietmar" <dietmar at proxmox.com>
Cc: "pve-devel" <pve-devel at pve.proxmox.com>
Envoyé: Mardi 6 Janvier 2015 00:09:17
Objet: Re: [pve-devel] Quorum problems with NICs Intel of 10 Gb/s and VMsturns off

Hi to all 

Recently i have tested in the company that igmp is necessary for the VMs 
(tested with tcpdump), the company has Windows Servers as VMs and several 
Windows systems as workstations in the local network, so i can tell you that 
i need to have the protocol igmp enabled in some VMs for that the Windows 
systems in the company work perfectly. 

And about of your suggestion: 
-A PVEFW-HOST-IN -s yournetwork/24 -p udp -m addrtype --dst-type 
MULTICAST -m udp --dport 5404:5405 -j RETURN 

I would like to do some questions: 
1) ¿Such rule will avoid the cluster communication to the VMs? 
2) ¿Such rule will not prejudice the normal use of the igmp protocol own of 
the Windows systems in the VMs? 
3) If both answers are correct, where i should put the rule that you 
suggest? 


----- Original Message ----- 
From: "Alexandre DERUMIER" <aderumier at odiso.com> 
To: "dietmar" <dietmar at proxmox.com> 
Cc: "pve-devel" <pve-devel at pve.proxmox.com> 
Sent: Monday, January 05, 2015 6:18 AM 
Subject: Re: [pve-devel] Quorum problems with NICs Intel of 10 Gb/s and 
VMsturns off 


>>>Following rule on your pve nodes should prevent igmp packages flooding 
>>>your bridge: 
>>>iptables -t filter -A FORWARD -i vmbr0 -p igmp -j DROP 
>>> 
>>>If something happens you can remove the rule this way: 
>>>iptables -t filter -D FORWARD -i vmbr0 -p igmp -j DROP 
> 
> Just be carefull that it'll block all igmp, so if you need multicast 
> inside your vms, 
> I'll block it too. 
> 
> Currently, we have a default rule for IN|OUT for host communication 
> 
> -A PVEFW-HOST-IN -s yournetwork/24 -p udp -m addrtype --dst-type 
> MULTICAST -m udp --dport 5404:5405 -j RETURN 
> to open multicast between nodes. 
> 
> Bit indeed, currently, in proxmox firewall, we can't define global rule in 
> FORWARD. 
> 
> 
> 
> 
> @Dietmar: maybe can we add a default drop rule in -A PVEFW-FORWARD, to 
> drop multicast traffic from host ? 
> 
> Or maybe better, allow to create rules at datacenter level, and put them 
> in -A PVEFW-FORWARD ? 
> 
> 
> 
> ----- Mail original ----- 
> De: "datanom.net" <mir at datanom.net> 
> À: "pve-devel" <pve-devel at pve.proxmox.com> 
> Envoyé: Dimanche 4 Janvier 2015 03:34:57 
> Objet: Re: [pve-devel] Quorum problems with NICs Intel of 10 Gb/s and 
> VMsturns off 
> 
> On Sat, 3 Jan 2015 21:32:54 -0300 
> "Cesar Peschiera" <brain at click.com.py> wrote: 
> 
>> 
>> Now in the switch i have igmp snooping disabled, but i want to avoid 
>> flooding the entire VLAN and the VMs 
>> 
> Following rule on your pve nodes should prevent igmp packages flooding 
> your bridge: 
> iptables -t filter -A FORWARD -i vmbr0 -p igmp -j DROP 
> 
> If something happens you can remove the rule this way: 
> iptables -t filter -D FORWARD -i vmbr0 -p igmp -j DROP 
> 
> PS. Your SPF for click.com.py is configured wrong: 
> Received-SPF: softfail (click.com.py ... _spf.copaco.com.py: Sender is 
> not authorized by default to use 'brain at click.com.py' in 'mfrom' 
> identity, however domain is not currently prepared for false failures 
> (mechanism '~all' matched)) receiver=mail1.copaco.com.py; 
> identity=mailfrom; envelope-from="brain at click.com.py"; helo=gerencia; 
> client-ip=190.23.61.163 
> Received-SPF: softfail (click.com.py ... _spf.copaco.com.py: Sender is 
> not authorized by default to use 'brain at click.com.py' in 'mfrom' 
> identity, however domain is not currently prepared for false failures 
> (mechanism '~all' matched)) receiver=mail1.copaco.com.py; 
> identity=mailfrom; envelope-from="brain at click.com.py"; helo=gerencia; 
> client-ip=190.23.61.163 
> Received-SPF: softfail (click.com.py ... _spf.copaco.com.py: Sender is 
> not authorized by default to use 'brain at click.com.py' in 'mfrom' 
> identity, however domain is not currently prepared for false failures 
> (mechanism '~all' matched)) receiver=mail1.copaco.com.py; 
> identity=mailfrom; envelope-from="brain at click.com.py"; helo=gerencia; 
> client-ip=190.23.61.163 
> -- 
> Hilsen/Regards 
> Michael Rasmussen 
> 
> Get my public GnuPG keys: 
> michael <at> rasmussen <dot> cc 
> http://pgp.mit.edu:11371/pks/lookup?op=get&search=0xD3C9A00E 
> mir <at> datanom <dot> net 
> http://pgp.mit.edu:11371/pks/lookup?op=get&search=0xE501F51C 
> mir <at> miras <dot> org 
> http://pgp.mit.edu:11371/pks/lookup?op=get&search=0xE3E80917 
> -------------------------------------------------------------- 
> /usr/games/fortune -es says: 
> Why does a hearse horse snicker, hauling a lawyer away? 
> -- Carl Sandburg 
> 
> _______________________________________________ 
> pve-devel mailing list 
> pve-devel at pve.proxmox.com 
> http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-devel 
> _______________________________________________ 
> pve-devel mailing list 
> pve-devel at pve.proxmox.com 
> http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-devel 
> 


More information about the pve-devel mailing list