This tutorial will take you through the provisioning and testing of a multinode GlusterFS cluster on GCE.
First you will need to download some credentials and set some environment variables in order to authenticate with GCP.
gcloud iam service-accounts create ansible-gluster --display-name ansible-gluster
export SA_EMAIL=$(gcloud iam service-accounts list --filter="displayName:ansible-gluster" --format='value(email)')
export PROJECT=$(gcloud info --format='value(config.project)')
gcloud projects add-iam-policy-binding $PROJECT --role roles/compute.storageAdmin --member serviceAccount:$SA_EMAIL
gcloud projects add-iam-policy-binding $PROJECT --role roles/compute.instanceAdmin.v1 --member serviceAccount:$SA_EMAIL
gcloud projects add-iam-policy-binding $PROJECT --role roles/compute.networkAdmin --member serviceAccount:$SA_EMAIL
gcloud projects add-iam-policy-binding $PROJECT --role roles/compute.securityAdmin --member serviceAccount:$SA_EMAIL
gcloud iam service-accounts keys create ansible-gluster-sa.json --iam-account $SA_EMAIL
You will also need to ensure that you have your local machine's SSH key uploaded to your project's metadata.
ls ~/.ssh/id_rsa.pub || ssh-keygen -N ""
gcloud compute project-info describe --format=json | jq -r '.commonInstanceMetadata.items[] | select(.key == "sshKeys") | .value' > sshKeys.pub
echo "$USER:$(cat ~/.ssh/id_rsa.pub)" >> sshKeys.pub
gcloud compute project-info add-metadata --metadata-from-file sshKeys=sshKeys.pub
In order to run the playbook you will need to install Ansible and Libcloud as follows:
pip install --user "ansible==2.2.2" "apache-libcloud==1.5.0"
Once those dependencies have been installed, clone the repository and enter the directory:
git clone https://github.com/GoogleCloudPlatform/compute-ansible-gluster
cd compute-ansible-gluster
In the repository you will find a gluster.yml
file that contains parameters at
the top for configuring your cluster.
machine_type
: The size of your each machine in the cluster, more cores will give you
better network performance. The full list of machine types can be found
here.hosts
: This list defines both the names of your host and your total cluster
sizedisk_type
: The type of disk to use for Gluster's data volume. This can be
either pd-ssd or pd-standard.disk_size
: The size of the data disk that will be attached to each instance
in the cluster. Keep in mind that the IOPS and throughput performance of each
disk scales with its size. More info on persistent disk performance can be
found here.zone
: This defines the specific zone with a region that the cluster will be
provisioned in.In order to create Gluster volumes at provisioning time you can define the
volumes
list variable. Each element in the list is a dictionary with the
following keys that describe the type of volume to create in the cluster.
name
: Identifier for your volumetype
: Gluster volume type. Examples are stripe 3
, replica 3
, or stripe 3 replica 3
. Can also be left blank for a distributed volume (default).parameters
: List of parameters for the volume.
Full
listhosts
: The list of hosts that this volume will be created on.Volumes will be mounted at /mnt/NAME by default. The mount path can be changed
by setting the global variable mount_point
.
Once you have configured your cluster specifications you can provision it by running the Ansible playbook as follows:
export GCE_EMAIL=$SA_EMAIL
export GCE_PROJECT=$PROJECT
export GCE_CREDENTIALS_FILE_PATH=../ansible-gluster-sa.json
export ANSIBLE_HOST_KEY_CHECKING=False
export PATH=~/.local/bin:$PATH
ansible-playbook -i hosts gluster.yml
cat /tmp/gluster-client-*/*/tmp/* # view your results
In order to mount your Gluster volume, launch an instance in your network and then run the following:
sudo yum install glusterfs-client
sudo mount -t glusterfs <ip-of-a-machine-in-your-cluster>:/share /mnt
For a quick example, you can run the gluster-clients.yml playbook which will take care of the provisioning a client for you and run a benchmark with fio.
export ANSIBLE_HOST_KEY_CHECKING=False
ansible-playbook -i hosts gluster-clients.yml
These playbooks also setup the Stackdriver and Logging agents on the instances. You can checkout your logs here and your monitoring here.
You can use the Ansible playbooks to turn down your machines by running the following:
ansible-playbook -i hosts gluster.yml gluster-clients.yml -e state=absent