Gluster FS is a scalable, distributed file system that aggregates disk storage resources from multiple servers into a single global namespace.


  • Scales to several petabytes
  • Handles thousands of clients
  • Uses commodity hardware
  • Allows optimization for different workloads
  • Open Source
  • Gluster FS Architecture:

    Gluster File System| Overview  

    Setup GlusterFS Cluster in Oracle Cloud Infrastructure

    Setting up OCI Virtual Machine Instances

    1. Create two virtual machine instances in OCI. You can create virtual machine instances in OCI using the console or REST apis or any automation tool of your choice. The instructions in this blog work on centos 7 operating system for both VM instances.
    2. You need to make sure that the two instances that you have created are able to talk to each other. If the two instances are in the same subnet, they should be able to talk to each other automatically. if they are in separate subnets edit the security rule and open the security list to enable the instances in the two subnets to access each other.
    3. In the interest of this blog, I have named my instances instance01 and instance02.
    4. Make an entry in the /etc/hosts file with the IP and hostname of both VM instances as shown below. The IP's should be resolvable using the hostnames. Change the IP in the following to your server IP's. instance01 instance02
    5. Create two storage volumes in oci (using oci console) and attach one to each OCI VM instance.
    6. Instance Port Configuration

      Open the below ports using security rules in both subnet security lists as well the server firewall if it is enabled

      111 - 24007 GlusterFS Daemon
      24008 - GlusterFS Management
      38465 to 38467 - GlusterFS NFS service
      49152 to 60000 - Depends on number of bricks

      If you are using Oracle Linux, then follow below additional steps highlighted in purple else skip this one

      • Edit file /etc/yum.conf and add a line assumeyes=1 at the last
      • Create a file under /etc/yum.repos.d/ by the name of Gluster.repo and add below content in it

      • On all servers - Login as root user and execute below commands-
      • yum update -y
        yum install xfsprogs attr -y
        yum install centos-release-gluster -y
        yum install glusterfs-server glusterfs-client glusterfs-common -y
        vi /etc/sysconfig/selinux and disabled it
        systemctl enable glusterd
        service glusterd start
        firewall-cmd --zone=public --add-port=24007-24009/tcp --add-port=49152-49251/tcp --permanent
        firewall-cmd --reload

      Create Mount points for OCI Block storage volumes

      sudo mkfs -t xfs /dev/sdb

      You need to do the following in both the OCI virtual machine instances. Format the volume to xfs.sdb is the name of the oci block storage volumes. You can list the available devices using lsblk command.

      Create a mount directory named /gshare and mount the formatted volume

      sudo mkdir /gshare sudo mount /share /dev/sdb

      Add the mount to /etc/fstab

      /dev/sdb /gshare xfs defaults,nofail 0

      GlusterFS installation on Virtual machine instances

      You need to perform the steps below on both virtual machine instances

      1. Create a GlusterFs repo.
        sudo vi /etc/yum.repos.d/Gluster.repo
        Add and enable the following repo
        [gluster38]name=Gluster 3.8baseurl=$basearch/gluster-3.8/gpgcheck=0enabled=1
      2. Install the glusterFS server on the VM instances
        sudo yum install glusterfs-server -y
      3. Start the glusterd service
        [opc@instance01 ~]$ sudo systemctl start glusterd
        [opc@instance01 ~]$ sudo systemctl status glusterd
        Output: glusterd.service - GlusterFS, a clustered file-system serverLoaded: loaded (/usr/lib/systemd/system/glusterd.service;
        vendor preset: enabled)Active: active (running) since Sat 2020-02-29 03:04:09 GMT;
        8s agoDocs: man:glusterd(8)Process: 6556 ExecStart=/usr/sbin/glusterd -p /var/run/ --log-level $LOG_LEVEL $GLUSTERD_OPTIONS (code=exited, status=0/SUCCESS) Main PID: 6557 (glusterd)
        CGroup: /system.slice/glusterd.service 6557 /usr/sbin/glusterd -p /var/run/ --log-level INF...Feb 29 03:04:09 instance01 systemd[1]: Starting GlusterFS, a clustered file.....Feb 29 03:04:09 instance01 systemd[1]: Started GlusterFS, a clustered file-...r
        Configure the GlusterFS server just installed on both VM instances
        From instance01 execute the following command to create a trusted storage pool with instance02.
        [opc@instance01 ~]$ sudo gluster peer probe instance02
        peer probe: success.
        2. Check the peer status using the following command
        [opc@instance01 ~]$ sudo gluster peer status
        Number of Peers: 1Hostname: instance02Uuid: ec83924b-877e-4fe8-b58a-b7541285999fState: Peer in Cluster (Connected)
        3. Create a data directory on gshare mount on both virtual machine instances
        [opc@instance01 ~]$ sudo mkdir /gshare/data
      4. Create a GlusterFS HA shared volume
        [opc@instance01 ~]$ sudo gluster volume create gdata replica 2 instance01:/gshare/data instance02:/gshare/data
      5. Start the gdata volume
        [opc@instance01 ~]$ sudo gluster volume start gdata
      6. Create a third virtual machine instance in OCI. You can create virtual machine instances in OCI using the console or REST apis or any automation tool of your choice. The virtual machines instance is also based on Centos 7 operating system. Once the instance is up and running, open the oci console and navigate to this instance and look at the public ip of the newly created instance. The initial three digits of the ip address gives you the CIDR range that needs to be enabled in this step. (e.g. 152* as in this example)
        Set the volume permissions for gdata volume for client access. Here I am using 152.* CIDR. You need to replace it based on your network range of your GlusterFS client instance subnet
        [opc@instance01 ~]$ sudo gluster volume set gdata auth.allow "152.*"
      7. To get all the info about the volume, execute the following command
        [opc@instance01 ~]$ sudo gluster volume info gdata
        Volume Name: gdataType: ReplicateVolume ID: 699ec5a2-55c1-43a2-94c1-2a8779f32d4cStatus: StartedSnapshot Count: 0Number of Bricks: 1 x 2 = 2Transport-type: tcpBricks:Brick1: instance01:/gshare/dataBrick2: instance02:/gshare/dataOptions Reconfigured:auth.allow: 152.*transport.address-family: inetstorage.fips-mode-rchecksum: onnfs.disable: onperformance.client-io-threads: off Start the GlusterFS Client Setup on third instance created in step7 above
        1. Enable fuse kernel module
          [opc@instanceglusterclient ~]$ sudo modprobe fuse
        2. Install all the glusterFS client dependencies
          [opc@instanceglusterclient ~]$ sudo yum install fuse fuse-libs openib libibverbs -y
        3. Install the GlusterFS client
          [opc@instanceglusterclient ~]$ sudo yum install glusterfs-client -y

        GlusterFS Client Configuration

        The data will be replicated if you are writing from the GlusterFS client. You can mount the GlusterFS volume to any number of clients. A client machine is not part of the glusterFS cluster. It is the machine in which you want to mount the replicated volume.

        1. Create a client mount directory
          [opc@instanceglusterclient ~]$ sudo mkdir /gfdata
        2. Mount gfdata directory to the glusterFS replicated volume
          [opc@instanceglusterclient ~]$ sudo mount -t glusterfs instance01:/gdata /gfdata
          You are all done !!!! . Your two node GlusterFS is working now !!!!
        3. About Vast Edge - Simplified Cloud Solutions

          Founded in the year 2004, Vast Edge is a leading IT Consulting Company and Global Service provider of Business Intelligence, big data analytics, cloud ERP, IoT platform, enterprise backup, and disaster recovery, Blockchain, Cassandra, AI/ML, and Integration solutions. We perform Application modernization, follow continuous change management, and implement advanced practices while working on the key areas of enhancing methodologies of project development.

          Get a Free Assessment from our cloud experts.

    Enter Details

    Email Company Name

    Get in touch with us

    Hello! 👋 How can we help you today?