There is a new version of this tutorial available for Fedora 13.

Creating An NFS-Like Standalone Storage Server With GlusterFS On Fedora 12 - Page 2

This tutorial exists for these OS versions

On this page

  1. 3 Setting Up The GlusterFS Client
  2. 4 Links

3 Setting Up The GlusterFS Client

There's a GlusterFS client rpm package for Fedora 12, but the problem with it is that you will get errors like df: `/mnt/glusterfs': Software caused connection abort or df: `/mnt/glusterfs': Transport endpoint is not connected when you try to access the GlusterFS share. That's why we build the GlusterFS client from the sources to avoid these problems.

Before we build the GlusterFS client, we install its prerequisites:

yum groupinstall 'Development Tools'
yum groupinstall 'Development Libraries'
yum install libibverbs-devel fuse-devel

Then we download the GlusterFS 2.0.9 sources (please note that this is the same version that is installed on the server!) and build GlusterFS as follows:

cd /tmp
tar xvfz glusterfs-2.0.9.tar.gz
cd glusterfs-2.0.9

At the end of the ./configure command, you should see something like this:

GlusterFS configure summary
FUSE client        : yes
Infiniband verbs   : yes
epoll IO multiplex : yes
Berkeley-DB        : yes
libglusterfsclient : yes
argp-standalone    : no

[root@client1 glusterfs-2.0.9]#
make && make install

Check the GlusterFS version afterwards (should be 2.0.9):

glusterfs --version
[root@client1 glusterfs-2.0.9]# glusterfs --version
glusterfs 2.0.9 built on Feb 19 2010 19:20:46
Repository revision: v2.0.9
Copyright (c) 2006-2009 Gluster Inc. <>
You may redistribute copies of GlusterFS under the terms of the GNU General Public License.
[root@client1 glusterfs-2.0.9]#

Then we create the following two directories:

mkdir /mnt/glusterfs
mkdir /etc/glusterfs

Next we create the file /etc/glusterfs/glusterfs.vol:

vi /etc/glusterfs/glusterfs.vol
volume remote
  type protocol/client
  option transport-type tcp
  option remote-host # can be IP or hostname
  option remote-subvolume brick

volume writebehind
  type performance/write-behind
  option window-size 4MB
  subvolumes remote

volume cache
  type performance/io-cache
  option cache-size 512MB
  subvolumes writebehind

Make sure you use the correct server hostname or IP address in the option remote-host line!

That's it! Now we can mount the GlusterFS filesystem to /mnt/glusterfs with one of the following two commands:

glusterfs -f /etc/glusterfs/glusterfs.vol /mnt/glusterfs


mount -t glusterfs /etc/glusterfs/glusterfs.vol /mnt/glusterfs

You should now see the new share in the outputs of...

[root@client1 glusterfs-2.0.9]# mount
/dev/mapper/vg_server2-lv_root on / type ext4 (rw)
proc on /proc type proc (rw)
sysfs on /sys type sysfs (rw)
devpts on /dev/pts type devpts (rw,gid=5,mode=620)
tmpfs on /dev/shm type tmpfs (rw)
/dev/sda1 on /boot type ext4 (rw)
none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)
sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw)
/etc/glusterfs/glusterfs.vol on /mnt/glusterfs type fuse.glusterfs (rw,allow_other,default_permissions,max_read=131072)
[root@client1 glusterfs-2.0.9]#

... and...

df -h
[root@client1 glusterfs-2.0.9]# df -h
Filesystem            Size  Used Avail Use% Mounted on
                       29G  2.6G   25G  10% /
tmpfs                 185M     0  185M   0% /dev/shm
/dev/sda1             194M   23M  161M  13% /boot
                       29G  2.7G   25G  10% /mnt/glusterfs
[root@client1 glusterfs-2.0.9]#

Instead of mounting the GlusterFS share manually on the client, you could modify /etc/fstab so that the share gets mounted automatically when the client boots.

Open /etc/fstab and append the following line:

vi /etc/fstab
/etc/glusterfs/glusterfs.vol  /mnt/glusterfs  glusterfs  defaults  0  0

To test if your modified /etc/fstab is working, reboot the client:


After the reboot, you should find the share in the outputs of...

df -h

... and...



Share this page:

0 Comment(s)