[Pacemaker] fence_xvm / fence_virtd problem

Digimer lists at alteeve.ca
Wed Jun 19 04:40:01 EDT 2013


The key was fine, but you nailed it on the mcast address. I thought it 
had to match the mcast used by the cluster.

Many thanks! Now I can get back to learning. :)

digimer

On 06/18/2013 11:53 PM, Andrew Beekhof wrote:
> Try the default multicast address perhaps?
>
> address = "225.0.0.12";
>
> You checked /etc/cluster/fence_xvm.key match on both machines I assume?
>
> On 17/06/2013, at 4:09 AM, Digimer <lists at alteeve.ca> wrote:
>
>> Guest's firewall is off entirely, as is selinux.
>>
>> On 06/16/2013 02:25 AM, Vladislav Bogdanov wrote:
>>> 16.06.2013 06:19, Digimer wrote:
>>>> Tried allowing everything into the host from the bridge. No luck...
>>>
>>> I meant guest's firewall.
>>>
>>>> Tried with;
>>>>
>>>> iptables -I INPUT -i virbr0 -j ACCEPT
>>>>
>>>> new rules;
>>>>
>>>> ====
>>>> # Generated by iptables-save v1.4.16.2 on Sat Jun 15 23:17:15 2013
>>>> *nat
>>>> :PREROUTING ACCEPT [397:94492]
>>>> :INPUT ACCEPT [34:7520]
>>>> :OUTPUT ACCEPT [709:61932]
>>>> :POSTROUTING ACCEPT [681:59876]
>>>> -A POSTROUTING -s 192.168.122.0/24 ! -d 192.168.122.0/24 -p tcp -j
>>>> MASQUERADE --to-ports 1024-65535
>>>> -A POSTROUTING -s 192.168.122.0/24 ! -d 192.168.122.0/24 -p udp -j
>>>> MASQUERADE --to-ports 1024-65535
>>>> -A POSTROUTING -s 192.168.122.0/24 ! -d 192.168.122.0/24 -j MASQUERADE
>>>> COMMIT
>>>> # Completed on Sat Jun 15 23:17:15 2013
>>>> # Generated by iptables-save v1.4.16.2 on Sat Jun 15 23:17:15 2013
>>>> *mangle
>>>> :PREROUTING ACCEPT [52594:46854362]
>>>> :INPUT ACCEPT [38456:45153011]
>>>> :FORWARD ACCEPT [13667:1595902]
>>>> :OUTPUT ACCEPT [28298:2665665]
>>>> :POSTROUTING ACCEPT [42095:4288350]
>>>> -A POSTROUTING -o virbr0 -p udp -m udp --dport 68 -j CHECKSUM
>>>> --checksum-fill
>>>> COMMIT
>>>> # Completed on Sat Jun 15 23:17:15 2013
>>>> # Generated by iptables-save v1.4.16.2 on Sat Jun 15 23:17:15 2013
>>>> *filter
>>>> :INPUT ACCEPT [69:8423]
>>>> :FORWARD ACCEPT [0:0]
>>>> :OUTPUT ACCEPT [130:12757]
>>>> -A INPUT -i virbr0 -j ACCEPT
>>>> -A INPUT -i virbr0 -p udp -m udp --dport 53 -j ACCEPT
>>>> -A INPUT -i virbr0 -p tcp -m tcp --dport 53 -j ACCEPT
>>>> -A INPUT -i virbr0 -p udp -m udp --dport 67 -j ACCEPT
>>>> -A INPUT -i virbr0 -p tcp -m tcp --dport 67 -j ACCEPT
>>>> -A FORWARD -d 192.168.122.0/24 -o virbr0 -m conntrack --ctstate
>>>> RELATED,ESTABLISHED -j ACCEPT
>>>> -A FORWARD -s 192.168.122.0/24 -i virbr0 -j ACCEPT
>>>> -A FORWARD -i virbr0 -o virbr0 -j ACCEPT
>>>> -A FORWARD -o virbr0 -j REJECT --reject-with icmp-port-unreachable
>>>> -A FORWARD -i virbr0 -j REJECT --reject-with icmp-port-unreachable
>>>> COMMIT
>>>> # Completed on Sat Jun 15 23:17:15 2013
>>>> ====
>>>>
>>>> On 06/15/2013 06:04 PM, Vladislav Bogdanov wrote:
>>>>> 15.06.2013 20:26, Digimer wrote:
>>>>>> Ah, I think it's a problem with the firewall rules on the host. Not sure
>>>>>> how to fix it though...
>>>>>
>>>>> You probably need to open port 1229/tcp in filter INPUT on virtual
>>>>> cluster members. That is where fence_xvm listens for a connection from
>>>>> fence_virtd after it sends multicast request iirc. Design deficiency,
>>>>> fence_xvm works only one copy on a system at a time, it listens for tcp
>>>>> connection on a predefined port.
>>>>>
>>>>>>
>>>>>> lemass:/home/digimer# iptables-save
>>>>>> # Generated by iptables-save v1.4.16.2 on Sat Jun 15 13:26:33 2013
>>>>>> *nat
>>>>>> :PREROUTING ACCEPT [246583:89552160]
>>>>>> :INPUT ACCEPT [2335:362026]
>>>>>> :OUTPUT ACCEPT [11740:741351]
>>>>>> :POSTROUTING ACCEPT [11706:738225]
>>>>>> -A POSTROUTING -s 192.168.122.0/24 ! -d 192.168.122.0/24 -p tcp -j
>>>>>> MASQUERADE --to-ports 1024-65535
>>>>>> -A POSTROUTING -s 192.168.122.0/24 ! -d 192.168.122.0/24 -p udp -j
>>>>>> MASQUERADE --to-ports 1024-65535
>>>>>> -A POSTROUTING -s 192.168.122.0/24 ! -d 192.168.122.0/24 -j MASQUERADE
>>>>>> COMMIT
>>>>>> # Completed on Sat Jun 15 13:26:33 2013
>>>>>> # Generated by iptables-save v1.4.16.2 on Sat Jun 15 13:26:33 2013
>>>>>> *mangle
>>>>>> :PREROUTING ACCEPT [3250861:2486027770]
>>>>>> :INPUT ACCEPT [2557761:1301267981]
>>>>>> :FORWARD ACCEPT [444644:1094901100]
>>>>>> :OUTPUT ACCEPT [1919457:2636518995]
>>>>>> :POSTROUTING ACCEPT [2364615:3731498365]
>>>>>> -A POSTROUTING -o virbr0 -p udp -m udp --dport 68 -j CHECKSUM
>>>>>> --checksum-fill
>>>>>> COMMIT
>>>>>> # Completed on Sat Jun 15 13:26:33 2013
>>>>>> # Generated by iptables-save v1.4.16.2 on Sat Jun 15 13:26:33 2013
>>>>>> *filter
>>>>>> :INPUT ACCEPT [2557761:1301267981]
>>>>>> :FORWARD ACCEPT [0:0]
>>>>>> :OUTPUT ACCEPT [1919457:2636518995]
>>>>>> -A INPUT -i virbr0 -p udp -m udp --dport 53 -j ACCEPT
>>>>>> -A INPUT -i virbr0 -p tcp -m tcp --dport 53 -j ACCEPT
>>>>>> -A INPUT -i virbr0 -p udp -m udp --dport 67 -j ACCEPT
>>>>>> -A INPUT -i virbr0 -p tcp -m tcp --dport 67 -j ACCEPT
>>>>>> -A FORWARD -d 192.168.122.0/24 -o virbr0 -m conntrack --ctstate
>>>>>> RELATED,ESTABLISHED -j ACCEPT
>>>>>> -A FORWARD -s 192.168.122.0/24 -i virbr0 -j ACCEPT
>>>>>> -A FORWARD -i virbr0 -o virbr0 -j ACCEPT
>>>>>> -A FORWARD -o virbr0 -j REJECT --reject-with icmp-port-unreachable
>>>>>> -A FORWARD -i virbr0 -j REJECT --reject-with icmp-port-unreachable
>>>>>> COMMIT
>>>>>> # Completed on Sat Jun 15 13:26:33 2013
>>>>>>
>>>>>> digimer
>>>>>>
>>>>>> On 06/15/2013 01:09 PM, Digimer wrote:
>>>>>>> Hi all,
>>>>>>>
>>>>>>>      I'm trying to play with pacemaker on fedora 19 (pre-release) and
>>>>>>> I am
>>>>>>> having trouble getting the guests to talk to the host.
>>>>>>>
>>>>>>>    From the host, I can run;
>>>>>>>
>>>>>>> lemass:/home/digimer# fence_xvm -o list
>>>>>>> pcmk1                83f6abdc-bb48-d794-4aca-13f091f32c8b on
>>>>>>> pcmk2                2d778455-de7d-a9fa-994c-69d7b079fda8 on
>>>>>>>
>>>>>>> I can fence the guests from the host as well. However, I can not get
>>>>>>> the
>>>>>>> list (or fence) from the quests;
>>>>>>>
>>>>>>> [root at pcmk1 ~]# fence_xvm -o list
>>>>>>> Timed out waiting for response
>>>>>>> Operation failed
>>>>>>>
>>>>>>> I suspect a multicast issue, but so far as I can tell, multicast is
>>>>>>> enabled on the bridge;
>>>>>>>
>>>>>>> lemass:/home/digimer# ifconfig
>>>>>>> virbr0: flags=4163<UP,BROADCAST,RUNNING,MULTICAST>  mtu 1500
>>>>>>>            inet 192.168.122.1  netmask 255.255.255.0  broadcast
>>>>>>> 192.168.122.255
>>>>>>>            ether 52:54:00:da:90:a1  txqueuelen 0  (Ethernet)
>>>>>>>            RX packets 103858  bytes 8514464 (8.1 MiB)
>>>>>>>            RX errors 0  dropped 0  overruns 0  frame 0
>>>>>>>            TX packets 151988  bytes 177742562 (169.5 MiB)
>>>>>>>            TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0
>>>>>>>
>>>>>>> vnet0: flags=4163<UP,BROADCAST,RUNNING,MULTICAST>  mtu 1500
>>>>>>>            inet6 fe80::fc54:ff:feed:3701  prefixlen 64  scopeid
>>>>>>> 0x20<link>
>>>>>>>            ether fe:54:00:ed:37:01  txqueuelen 500  (Ethernet)
>>>>>>>            RX packets 212828  bytes 880551892 (839.7 MiB)
>>>>>>>            RX errors 0  dropped 0  overruns 0  frame 0
>>>>>>>            TX packets 225430  bytes 182955760 (174.4 MiB)
>>>>>>>            TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0
>>>>>>>
>>>>>>> vnet1: flags=4163<UP,BROADCAST,RUNNING,MULTICAST>  mtu 1500
>>>>>>>            inet6 fe80::fc54:ff:fe45:e9ae  prefixlen 64  scopeid
>>>>>>> 0x20<link>
>>>>>>>            ether fe:54:00:45:e9:ae  txqueuelen 500  (Ethernet)
>>>>>>>            RX packets 4840  bytes 587902 (574.1 KiB)
>>>>>>>            RX errors 0  dropped 0  overruns 0  frame 0
>>>>>>>            TX packets 7495  bytes 899578 (878.4 KiB)
>>>>>>>            TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0
>>>>>>>
>>>>>>> I tried specifying the mcast address and port without success.
>>>>>>>
>>>>>>> The host's config is:
>>>>>>>
>>>>>>> lemass:/home/digimer# cat /etc/fence_virt.conf
>>>>>>> backends {
>>>>>>>        libvirt {
>>>>>>>            uri = "qemu:///system";
>>>>>>>        }
>>>>>>>
>>>>>>> }
>>>>>>>
>>>>>>> listeners {
>>>>>>>        multicast {
>>>>>>>            port = "1229";
>>>>>>>            family = "ipv4";
>>>>>>>            interface = "virbr0";
>>>>>>>            address = "239.192.214.190";
>>>>>>>            key_file = "/etc/cluster/fence_xvm.key";
>>>>>>>        }
>>>>>>>
>>>>>>> }
>>>>>>>
>>>>>>> fence_virtd {
>>>>>>>        module_path = "/usr/lib64/fence-virt";
>>>>>>>        backend = "libvirt";
>>>>>>>        listener = "multicast";
>>>>>>> }
>>>>>>>
>>>>>>> The cluster forms and corosync is using multicast, so I am not sure if
>>>>>>> mcast really is the problem.
>>>>>>>
>>>>>>> Any tips/help?
>>>>>>>
>>>>>>> Thanks!
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>
>>>>> Project Home: http://www.clusterlabs.org
>>>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>
>>>>
>>>>
>>>
>>
>>
>> --
>> Digimer
>> Papers and Projects: https://alteeve.ca/w/
>> What if the cure for cancer is trapped in the mind of a person without access to education?
>>
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>


-- 
Digimer
Papers and Projects: https://alteeve.ca/w/
What if the cure for cancer is trapped in the mind of a person without 
access to education?




More information about the Pacemaker mailing list