[ClusterLabs] CentOS 7 nfs exports erroring

Steve Dainard sdainard at spd1.com
Thu May 28 16:51:55 UTC 2015


Hello,

I'm configuring a cluster which maps and mounts ceph rbd's, then exports
each mount over nfs, mostly following Sebastien's post here:
http://www.sebastien-han.fr/blog/2012/07/06/nfs-over-rbd/ but there are
some differences using pacemaker 1.2 with systemd.

The issue I'm running into is I can get ceph working without issue, but the
nfs exports error with:

# pcs status
Cluster name: nfs
Last updated: Thu May 28 09:21:13 2015
Last change: Wed May 27 17:47:06 2015
Stack: corosync
Current DC: node1 (1) - partition with quorum
Version: 1.1.12-a14efad
3 Nodes configured
24 Resources configured


Online: [ node1 node2 node3 ]

Full list of resources:

 Resource Group: group_rbd_fs_nfs_vip
     rbd_vol1   (ocf::ceph:rbd.in):     Started node1
     ...
     rbd_vol8   (ocf::ceph:rbd.in):     Started node1
     fs_vol1    (ocf::heartbeat:Filesystem):    Started node1
     ...
     fs_vol8    (ocf::heartbeat:Filesystem):    Started node1
     export_vol1        (ocf::heartbeat:exportfs):      Stopped
     ...
     export_vol8        (ocf::heartbeat:exportfs):      Stopped

*Failed actions:*
*    export_vol1_start_0 on node1 'unknown error' (1): call=262,
status=complete, exit-reason='none', last-rc-change='Wed May 27 17:42:37
2015', queued=0ms, exec=56ms*
*    export_vol1_start_0 on node2 'unknown error' (1): call=196,
status=complete, exit-reason='none', last-rc-change='Wed May 27 17:43:04
2015', queued=0ms, exec=63ms*
*    export_vol1_start_0 on node3 'unknown error' (1): call=196,
status=complete, exit-reason='none', last-rc-change='Wed May 27 17:43:27
2015', queued=0ms, exec=69ms*


PCSD Status:
  node1: Online
  node2: Online
  node3: Online

Daemon Status:
  corosync: active/disabled
  pacemaker: active/disabled
  pcsd: active/enabled

I thought this was an issue with the nfsd kernel module not loading, so I
manually loaded it on each host, but no change.

I'm also wondering if there's an error with my export resources config:
  Resource: export_vol1 (class=ocf provider=heartbeat type=exportfs)
   Attributes: directory=/mnt/vol1 clientspec=10.0.231.0/255.255.255.0
options=rw,no_subtree_check,no_root_squash fsid=1
   Operations: stop interval=0s timeout=120 (export_vol1-stop-timeout-120)
               monitor interval=10s timeout=20s
(export_vol1-monitor-interval-10s)
               start interval=0 timeout=40s (export_vol1-start-interval-0)
  Resource: export_vol2 (class=ocf provider=heartbeat type=exportfs)
   Attributes: directory=/mnt/vol2 clientspec=10.0.231.0/255.255.255.0
options=rw,no_subtree_check,no_root_squash fsid=2
   Operations: stop interval=0s timeout=120 (export_vol2-stop-timeout-120)
               monitor interval=10s timeout=20s
(export_vol2-monitor-interval-10s)
               start interval=0 timeout=40s (export_vol2-start-interval-0)
... (8 in total)


exportfs is throwing an error, and even more odd its breaking the nfs
subnet auth into IP's (10.0.231.103,10.0.231.100) which I have no idea
where its getting those IP addresses from.
Logs: http://pastebin.com/xEX2L7m1

/etc/exportfs:
/mnt/vol1       10.0.231.0/24(rw,no_subtree_check,no_root_squash)
...
/mnt/vol8       10.0.231.0/24(rw,no_subtree_check,no_root_squash)

# getenforce
Permissive

# systemctl status nfs
nfs-server.service - NFS server and services
   Loaded: loaded (/usr/lib/systemd/system/nfs-server.service; disabled)
   Active: inactive (dead)

Thanks,
Steve
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20150528/b8c330f0/attachment-0003.html>


More information about the Users mailing list