Constellation-Labs / euclid-development-environment

Quickly spin up basic dev environments to build local metagraph projects with our extendable base.
Apache License 2.0
5 stars 12 forks source link

Welcome to Euclid Development Environment

Dependencies

Docker

Cargo

Ansible

JQ

YQ

First Steps

Understanding the folder structure

With the euclid-development-environment cloned, you'll see the following structure

- infra
- scripts
- source
- euclid.json

let's see what each of these directories represents:

Infra

This directory contains infrastructure related to the running of Euclid.

Custom Metagraph Base Image Dockerfile

You can customize your metagraph-base-image Dockerfile by creating a new file called Dockerfile under the infra/docker/custom/metagraph-base-image directory. This allows the build process to load and use your custom Dockerfile, enabling you to add additional features, such as external databases or tools.

This customization feature allows you to tailor your base image to include the specific features you require. Ensure that your custom Dockerfile continues to meet the primary goals of the original file: creating directories, building JARs, and moving the necessary files for execution, such as genesis.snapshot.

Scripts

Thats the "home" of hydra script, here you'll find the hydra and hydra-update (deprecated) scripts.

Source

Here is the home of the local codebase and required files for each layer.

euclid.json

Here is the hydra configuration file, there you can set the p12 file names and your GITHUB_TOKEN. It's required to fill the GitHub token here to run the hydra script

Hydra scripts options

Run the following command to list all the possibilities of the hydra script

./hydra -h

you should see something like this:

USAGE: hydra <COMMAND>

COMMANDS:
  install                           Installs a local framework and detaches project
  install-template                  Installs a project from templates
  build                             Build containers
  start-genesis                     Start containers from the genesis snapshot (erasing history) [aliases: start_genesis]
  start-rollback                    Start containers from the last snapshot (maintaining history) [aliases: start_rollback]
  stop                              Stop containers
  destroy                           Destroy containers
  purge                             Destroy containers and images
  status                            Check the status of the containers
  remote-deploy                     Remotely deploy to cloud instances using Ansible [aliases: remote_deploy]
  remote-start                      Remotely start the metagraph on cloud instances using Ansible [aliases: remote_start]
  remote-status                     Check the status of the remote nodes
  remote-logs                       Get the logs from the remote hosts
  remote-snapshot-fee-config        Get the remote snapshot fee config
  update                            Update Euclid
  logs                              Get the logs from containers
  install-monitoring-service        Download the metagraph-monitoring-service (https://github.com/Constellation-Labs/metagraph-monitoring-service) [aliases: install_monitoring_service]
  remote-deploy-monitoring-service  Deploy the metagraph-monitoring-service to remote host [aliases: remote_deploy_monitoring_service]
  remote-start-monitoring-service   Start the metagraph-monitoring-service on remote host [aliases: remote_start_monitoring_service]

TIP: You can use the same -h in each command listed above to see the accepted parameters

Building

Let's start with the build command. This command could be used simply this way:

./hydra build   

This script has some parameters such as --no_cache (run without previous cache), --run (automatically run after build).

Starting

We have the options start-genesis and start-rollback to start the containers. This option will fail case you didn't build the images yet.

./hydra start-genesis
./hydra start-rollback   

Stopping

We have the option stop to stop the containers. You can call the option this way:

./hydra stop   

Destroying

We have the option destroy to destroy the containers. You can call the option this way:

./hydra destroy   

We also have the purge option to destroy the containers and clean all images

./hydra purge   

Status

We have the option status to show the containers status. You can call the option this way:

./hydra status   

Installing

We have the option install to remove the link with remote git. You can call the option this way:

./hydra install   

You can import a metagraph template from custom examples by using the following command:

./hydra install-template

By default, we use the Metagraph Examples repository. You should provide the template name when running this command. To list the templates available to install, type:

./hydra install-template --list

Logs

We have the option logs to show the logs of nodes per container and layer. You can call the option this way:

./hydra logs :container_name :layer_name   

Update

We have the option update to update the Euclid. You can call the option this way:

./hydra update   

NOTE: FOR ALL OPTIONS ABOVE YOU CAN USE -h TO CHECK THE AVAILABLE PARAMETERS

Let's build

After understanding the folder structure, we can start build our containers.

NOTE: Make sure to fill your GITHUB_TOKEN on euclid.json file before start

Move your terminal to directory /scripts, home of the hydra script.

  cd scripts/

We need to install argc to run the script, here is the doc of argc

cargo install argc

Then run the following to build your containers

./hydra build

After the end of this step, run the following:

./hydra start-genesis

After the end of start-genesis, you should see something like this:

######################### METAGRAPH INFO #########################

Metagraph ID: :your_metagraph_id

Container metagraph-node-1 URLs
Global L0: http://localhost:9000/node/info
Metagraph L0: http://localhost:9200/node/info
Currency L1: http://localhost:9300/node/info
Data L1: http://localhost:9400/node/info

Container metagraph-node-2 URLs
Metagraph L0: http://localhost:9210/node/info
Currency L1: http://localhost:9310/node/info
Data L1: http://localhost:9410/node/info

Container metagraph-node-3 URLs
Metagraph L0: http://localhost:9220/node/info
Currency L1: http://localhost:9320/node/info
Data L1: http://localhost:9420/node/info

Clusters URLs
Global L0: http://localhost:9000/cluster/info
Metagraph L0: http://localhost:9200/cluster/info
Currency L1: http://localhost:9300/cluster/info
Data L1: http://localhost:9400/cluster/info

####################################################################

You can now access the URLs and see that your containers are working properly

You can also call the hydra option

./hydra status

Grafana

We have a Grafana container that can monitor your nodes. To enable this feature, modify the following field in euclid.json to true:

start_grafana_container=true

After updating this field, a Grafana container will be constructed when you start the services. You can access this tool at the following URL: http://localhost:3000/. The initial login credentials are:

username: admin
password: admin

You'll be requested to update the password after your first login

In this tool we have 2 dashboards, you can access them on Dashboard section

NOTE: This monitoring feature is distinct from remote monitoring. It displays data from your nodes on Dashboards, allowing you to check various metrics. However, it does not perform restarts or any other operations.

Deployment

Configuring, deploying, and starting remote node instances is supported through Ansible playbooks. The default settings deploy to three node instances via SSH which host all layers of your metagraph project (gL0, mL0, cL1, dL1). Two hydra methods are available to help with the deployment process: hydra remote-deploy and hydra remote-start. Prior to running these methods, remote host information must be configured in infra/ansible/remote/hosts.ansible.yml.

By default, we use the default directory for the SSH file, which is ~/.ssh/id_rsa. However, you can change it to your preferred SSH file directory. You can find instructions on how to generate your SSH file here.

Ansible functions more effectively with .pem key files. If you possess a .ppk key file, you can utilize these instructions to convert it to .pem.

If your file contains a password, you will be prompted to enter it to proceed with remote operations.

Host Configuration

To run your metagraph remotely, you'll need remote server instances - 3 instances for the default configuration. These hosts should be running either ubuntu-20.04 or ubuntu-22.04. It's recommended that each host meets the following minimum requirements:

You can choose your preferred platform for hosting your instances, such as AWS or DigitalOcean. After creating your hosts, you'll need to provide the following information in the hosts.ansible.yml file:

P12 Files

P12 files contain the public/private key pair identifying each node (peerID) and should be located in the source/p12-files directory by default. The file-name, key-alias, and password should be specified in the euclid.json file under the p12_files section. By default, Euclid comes with three example files: token-key.p12, token-key-1.p12, and token-key-2.p12. NOTE: Before deploying, be sure to replace these example files with your own, as these files are public and their credentials are shared.

NOTE: If deploying to MainNet, ensure that your peerIDs are registered and present on the metagraph seedlist. Otherwise, the metagraph startup will fail because the network will reject the snapshots.

Network Selection

Currently, there are two networks available for running your metagraph: IntegrationNet, and MainNet. You need to specify the network on which your metagraph will run in the euclid.json file under deploy -> network -> name.

GL0 Node Configuration

The deploy script does not deploy the gl0 node. It's recommended to use nodectl to build your gl0 node. Information on installing nodectl can be found here. Nodectl helps manage gl0 nodes by providing tools such as auto-upgrade and auto-restart which keep the node online in the case of a disconnection or network upgrade. Using these features is highly recommended for the stability of your metagraph.

NOTE: Your GL0 node must be up and running before deploying your metagraph. You can use the same host to run all four layers: gl0, ml0, cl1, and dl1.

hydra remote-deploy

This method configures remote instances with all the necessary dependencies to run a metagraph, including Java, Scala, and required build tools. The Ansible playbook used for this process can be found and edited in infra/ansible/playbooks/deploy.ansible.yml. It also creates all required directories on the remote hosts, and creates or updates metagraph files to match your local Euclid environment. Specifically, it creates the following directories:

Each directory will be created with cl-keytool.jar, cl-wallet.jar, and a P12 file for the instance. Additionally, they contain the following:

In code/metagraph-l0:

Parameters

If you run the command ./hydra remote-deploy -h you should see something like

Remotely deploy to cloud instances using Ansible

USAGE: hydra remote-deploy [OPTIONS]

OPTIONS:
      --force_genesis                                  Force metagraph to deploy as genesis
  -h, --help   

NOTE: You'll be prompted to proceed with force_genesis since it will delete all the previous data

hydra remote-start

This method initiates the remote startup of your metagraph in one of the available networks: integrationnet or mainnet. The network should be set in euclid.json under deploy -> network

To begin the remote startup of the metagraph, we utilize the parameters configured in euclid.json (network, gl0_node -> ip, gl0_node -> id, gl0_node -> public_port, ansible -> hosts, and ansible -> playbooks -> start). The startup process unfolds as follows:

  1. Termination of any processes currently running on the metagraph ports, which by default are 7000 for ml0, 8000 for cl1, and 9000 for dl1 (you can change on hosts.ansible.yml).
  2. Relocation of any existing logs to a folder named archived-logs, residing within each layer directory: metagraph-l0, currency-l1, and data-l1.
  3. Initiation of the metagraph-l0 layer, with node-1 designated as the genesis node.
  4. Initial startup as genesis, transitioning to rollback for subsequent executions. To force a genesis startup, utilize the --force_genesis flag with the hydra remote-start command. This will move the current data directory to a folder named archived-data and restart the metagraph from the first snapshot.
  5. Detection of missing files required for layer execution, such as :your_file.p12 and metagraph-l0.jar, triggering an error and halting execution.
  6. Following the initiation of metagraph-l0, the l1 layers, namely currency-l1 and data-l1, are started. These layers only started if present in your project.

After the script completes execution, you can verify if your metagraph is generating snapshots by checking the block explorer of the selected network:

You can verify if the cluster was successfully built by accessing the following URL:

http://{your_host_ip}:{your_layer_port}/cluster/info

Replace:

Each layer directory on every node contains a folder named logs. You can monitor and track your metagraph logs by running:

tail -f logs/app.log

NOTE: Don't forget to add your hosts' information, such as host, user, and SSH key file, to your infra/ansible/remote/hosts.ansible.yml file.

Parameters

If you run the command ./hydra remote-start -h you should see something like

Remotely start the metagraph on cloud instances using Ansible

USAGE: hydra remote-start [OPTIONS]

OPTIONS:
      --force_genesis                                    Force metagraph to run as genesis
      --force_owner_message                              Force to send owner message
      --force_staking_message                            Force to send staking message
  -h, --help    

NOTE: You'll be prompted to proceed with force_genesis since it will delete all the previous data

hydra remote-status

This method will return the status of your remote hosts. You should see the following:

################################## Node 1 ##################################
Metagraph L0
URL: http://:your_node_ip:your_port/node/info
State: :state
Host: :host
Public port: :your_port
P2P port: :your_port
Peer id: :peerId

Currency L1
URL: http://:your_node_ip:your_port/node/info
State: :state
Host: :host
Public port: :your_port
P2P port: :your_port
Peer id: :peerId

Data L1
URL: http://:your_node_ip:your_port/node/info
State: :state
Host: :host
Public port: :your_port
P2P port: :your_port
Peer id: :peerId

################################## Node 2 ##################################
Metagraph L0
URL: http://:your_node_ip:your_port/node/info
State: :state
Host: :host
Public port: :your_port
P2P port: :your_port
Peer id: :peerId

Currency L1
URL: http://:your_node_ip:your_port/node/info
State: :state
Host: :host
Public port: :your_port
P2P port: :your_port
Peer id: :peerId

Data L1
URL: http://:your_node_ip:your_port/node/info
State: :state
Host: :host
Public port: :your_port
P2P port: :your_port
Peer id: :peerId

################################## Node 3 ##################################
Metagraph L0
URL: http://:your_node_ip:your_port/node/info
State: :state
Host: :host
Public port: :your_port
P2P port: :your_port
Peer id: :peerId

Currency L1
URL: http://:your_node_ip:your_port/node/info
State: :state
Host: :host
Public port: :your_port
P2P port: :your_port
Peer id: :peerId

Data L1
URL: http://:your_node_ip:your_port/node/info
State: :state
Host: :host
Public port: :your_port
P2P port: :your_port
Peer id: :peerId

Remote Monitoring

We have introduced a tool in version v0.10.0 that can monitor your metagraph and restart it if necessary.

Introduction

This service monitors your metagraph and performs restarts as necessary. It is deployed using Ansible on a remote Ubuntu host. Commands such as install-monitoring-service, remote-deploy-monitoring-service, and remote-start-monitoring-service will be further explained in subsequent sections.

The service is developed using NodeJS, and all necessary dependencies are installed on your remote instance during deployment.

Running in the background with PM2, the service initiates checks at intervals specified in the configuration under the field: check_healthy_interval_in_minutes. It evaluates the health of the metagraph based on predefined and customizable restart-conditions, detailed in the metagraph-monitoring-service repository. For example, if an unhealthy node is detected, the service triggers a restart.

To restart a node or layer, the service first attempts to stop any running processes (referring to the layer). This operation requires sudo privileges without a password requirement (refer to this document for instructions on setting up password-less sudo). In addition to terminating processes, log files from the node are moved to the code/restart_logs directory, which may also require sudo privileges.

After these steps, the service restarts the node or layer and reintegrates it into the cluster.

Installation

This tool it's not default to Euclid, so you need to install this service. To do this you need to run the following:

hydra install-monitoring-service

this command to creates a monitoring project in your source directory, which will be named metagraph-monitoring-service

To use this feature, we need to know the informations about the remote host that will be used as monitoring. So, you need to populate the file infra/ansible/remote/hosts.ansible.yml under the monitoring section. You should provide a user that has sudo privileges without requiring a password. Refer to this document to learn how to enable password-less sudo for a user.

Monitoring Configuration

Before deploying to remote instances, you need to configure your monitoring by editing the file config/config.json. When you run the install command, some fields will be auto-populated based on the euclid.json file, which includes:

NOTE: You must provide your SSH key file that has access to each node. It is recommended to place this under the config directory. Ensure that this file has access to the node and that the user you've provided also has sudo privileges without a password.

Customize Monitoring

Learn how to customize your monitoring by checking the repositories:

Deploying Monitoring

Once you've configured your metagraph monitoring, deploy it to the remote host with:

hydra remote-deploy-monitoring-service

This command sends your current monitoring service from euclid to your remote instance and downloads all necessary dependencies.

Starting Monitoring

After deployment, start your monitoring with:

hydra remote-start-monitoring-service

To force a complete restart of your metagraph, use:

hydra remote-start-monitoring-service --force-restart