Welcome!
The community writes code to help people see the power of blockchains, understand true DeFi and support amazing networks like PulseChain.
These scripts will help automate your setup of a validator node running on the PulseChain Mainnet. Since it is a fork of Ethereum 2.0, most all of the methods and guidance can easily be re-worked for setting up validators on the Ethereum side as well. These scripts work on both your own hardware as well as cloud servers.
Setting up and running a validator server requires basic knowledge of Linux command line which you can learn here.
Please read ALL the instructions as they will explain and tell you how to run the scripts and any caveats.
To download these scripts on your server, you need git and then you can do git clone https://github.com/rhmaxdotorg/pulsechain-validator.git
.
After you download the code, you may need to chmod +x *.sh
to make all the scripts executable and able to run on the system.
To simply install the validator software, use pulsechain-validator-setup.sh. All the other scripts are just extra bells and whistles to support more features and maintenance. See this video for a quick demo of how to run the script.
Once you’re finished running the setup script, go to AFTER RUNNING THE SCRIPT section to complete the process and get your validator on the network.
Check out the FAQ or binge the 75+ validator videos that will probably answer any questions you have about how validators work, setup process, maintenance and more.
Scripts and guidance available include...
The setup script installs pre-reqs, golang, rust, go-pulse (geth fork) and lighthouse on a fresh, clean Ubuntu Linux OS for getting a PulseChain Mainnet Validator Node setup and running with Geth (go-pulse) and Lighthouse clients.
Clients are running on the same machine and not in docker containers (such as the method other scripts use). There are advantages and disadvantages to containerizing the clients vs running them on the host OS. You can also automate deployments with Terraform on AWS along with many other software packages. However, these scripts are meant to make it pretty easy to spin up and tear down new validator machines once you have the hardware up with access to a fresh install of Ubuntu Linux and try to make the process as seamless as possible.
There are other helper scripts that do various things, check the notes for each one specifically for more info.
You can run pulsechain-validator-setup.sh to setup your validator clients and monitoring-setup.sh afterwards to install the graphs and monitoring software.
Note: the pulsechain validator setup script doesn't install monitoring/metrics packages, however a script to do that is provided. It would need to run the validator setup script AND THEN run the monitoring-setup.sh script provided. Do not run the monitoring script before installing your validator clients. See details in the Grafana or Prometheus section.
Also check out the introductory blog post on Becoming a PulseChain Validator for a simple breakdown of how the process works as well as the detailed setup walkthrough video.
Check out these videos for further explanations and code walkthroughs.
More videos
How validators work
Profitability
$ ./pulsechain-validator-setup.sh [0x...YOUR NETWORK FEE ADDRESS] [12.89...YOUR SERVER IP ADDRESS]
Note: run the script with only the fee address and IP address, without the brackets (they are just included above for demonstration purposes).
NETWORK FEE ADDRESS is the FEE_RECIPIENT value for --suggested-fee-recipient to a wallet address you want to recieve priority fees from users whenever your validator proposes a new block (else it goes to the burn address)
SERVER_IP_ADDRESS to your validator server's IP address
Note: you may get prompted throughout the process to hit [Enter] for OK and continue the process at least once, so it's not meant to be a completely unattended install, but it could be depending on your environment setup.
Note: You can use choose to use either the same wallet or two different wallets for your fee address and withdrawal address. Some people might like their fees going to a hot wallet and withdrawals to a cold wallet, which in case they would use two different wallets.
If you encounter errors running the script and want to run the script again, use the Reset the Validator BEFORE running it over and over again.
Just make sure you know what you're doing and manually edit the reset script to bypass the "I don't know what I'm doing" check. It's very straightforward, just read the code, acknowledge you know what the script it doing and change I_KNOW_WHAT_I_AM_DOING=false to true to get it to run.
Tested on Ubuntu 22.04 Linux running as a non-root user with sudo privileges.
Desktop or Server edition works, however Desktop edition is recommended as it provides a GUI and is easier to navigate for beginners. LTS (long term support) versions are often more stable, which is better for servers and things you want to know are well tested and reliable.
The consensus on the minimum recommended requirements to run a validator seem to be 32gb RAM, 2TB disk and plenty of processing power (quadcore, xeon/ryzen, 4-8 vCPUs and such). These can come in the form of buying or building your own server and paying an upfront cost, utilities and maintenance OR renting a server from a VPS/cloud provider such as Amazon AWS (M2.2Xlarge server) or Digital Ocean 32GB Droplet and paying monthly to use their platform and resources. Both have advantages and disadvantages as well as varying time, monetary and management costs.
Could you get by with an old PC under your desk with a $50 battery backup? Maybe, but that would not be recommended. I'd rather not skimp on hardware for things that I would plan to run for years and pay for the peace of mind of not worrying about what I'm going to do if X fails one day, wishing I'd started with stronger foundations. If you try and do it right the first time, you might save a lot of time and headache.
It's recommended you have new hardware that meets or exceeds the minimum recommended requirements to give yourself the chance to have the best experience being a validator. There's also the PulseChain validator hardware store you can check out for more ready-to-go options at https://www.validatorstore.com.
The script automates a roughly estimated ~85% of what it takes to get the validator configured, but there's still a few manual steps you need to do to complete the setup: generate your keys in a secure environment, copy and import them on your validator server and once your clients are fully synced, make the deposit to activate your validator on the network.
Steps 1-4
1) Generate your keys (on another, secure machine) 2) Copy and import your keys on the validator 3) Start the validator client 4) Make your deposit(s)
Environment and hardware options for key generation
Generate validator keys with deposit tool and import them into Lighthouse
Run the staking deposit client (ON A DIFFERENT MACHINE, see notes below for details)
$ sudo apt install -y python3-pip
$ git clone https://gitlab.com/pulsechaincom/staking-deposit-cli.git
$ cd staking-deposit-cli && pip3 install -r requirements.txt && sudo python3 setup.py install
$ ./deposit.sh new-mnemonic --chain=pulsechain --eth1_withdrawal_address=0x... (ENTER THE CORRECT WALLET ADDRESS TO WITHDRAWAL YOUR FUNDS)
If you get an error after running the first command, saying that it can't find python3-pip (such as when you're booting from USB to run Ubuntu OS), you can update the apt config and that should fix it.
sudo add-apt-repository universe && sudo apt update
It is VERY IMPORTANT that you use a withdrawal wallet address that you have access to and is SECURE for a long time. Otherwise you may lose all your deposit funds.
Note: You can use choose to use either the same wallet or two different wallets for your fee address and withdrawal address. Some people might like their fees going to a hot wallet and withdrawals to a cold wallet, which in case they would use two different wallets.
Then follow the instructions from there, copy them over to the validator and import into lighthouse AS THE NODE USER (not the 'ubuntu' user on ec2) as described below.
If you are setting this up on a home server, you can copy the validator keys onto a USB drive and then plug the USB drive into the validator when you're ready to copy. If logged in to the validator server locally and using the Desktop, after plugging in the drive you can open a window to view files and folders on the drive and see the validator_keys folder. If you open another window that shows the Home directory, you can drag (or right click and copy) the validator_keys folder from the USB drive to the Home directory.
Once you have copied the validator_keys folder into the Home directory (for the user you're logged in as, which may be the ubuntu user or otherwise, as we're not using the node user for import operations yet), these are the next commands to get through the process.
$ sudo cp -R /home/ubuntu/validator_keys /home/node/validator_keys
$ sudo chown -R node:node /home/node/validator_keys
$ sudo -u node bash
(as node user)
$ cd ~
$ /opt/lighthouse/lighthouse/lh account validator import --directory ~/validator_keys --network=pulsechain
enter password to import validator(s)
(exit and back as ubuntu user)
Note: generate your keys on a different, secure machine (NOT on the validator server) and transfer them over for import. See the Security section for more references on why this is important. AWS even offers a free tier option that allows you to spin up and use VMs basically for free for a certain period of time, so you could use that for quick and easy tiny VMs running Ubuntu Linux (not beefy enough to be a validator, but fine for small tasks and learning).
You can use the scp
command to copy validator keys over the network (encrypted), USB stick (if hardware is local, not vps/cloud) OR use this base64 encoding trick for a copy and paste style solution such as the following. Note: this is advanced and you need to pay attention to be successful with it. If you're not confident you can do it, better to use scp or USB methods.
On disposable VM, live CD or otherwise emphemeral filesystem
sudo apt install -y unzip zip
zip -r validator_keys.zip validator_keys
base64 -w0 validator_keys.zip > validator_keys.b64
cat validator_keys.b64 (and copy the output)
On your validator server
cat > validator_keys.b64 <<EOF
Paste the output
[Enter] + type “EOF” + [Enter]
base64 -d validator_keys.b64 > validator_keys.zip
unzip validator_keys.zip
Also see Validator Key Generation and Management for more guidance.
Start the beacon and validator clients
$ sudo systemctl daemon-reload
$ sudo systemctl enable lighthouse-beacon lighthouse-validator
$ sudo systemctl start lighthouse-beacon lighthouse-validator
If you want to look at lighthouse debug logs (similar to geth)
$ journalctl -u lighthouse-beacon.service (with -f to get the latest logs OR without it to get the beginning logs)
$ journalctl -u lighthouse-validator.service
Once the blockchain clients are synced, you can make your 32m PLS deposit (per validator)
You can have multiple on one machine. The deposit is made @ https://launchpad.pulsechain.com to get your validator activated and participating on the network.
If you do the deposit before the clients are fully synced and ready to go, then you risk penalities as your validator would join the network, but due to not being synced, unable to participate in validator duties (until it's fully synced).
Once on the network, you check use Beacon Explorer to check for Active/Inactive status, Effectiveness and Rewards. See this post for how to view stats for validators.
Now let's get validating! @rhmaximalist
In general, if you happen to be stuck syncing at N% or even at 100% but not sure why you're not seeing the validator make attestations, sometimes a reboot can help. You might even see "Error during attestation routine" or "Error updating deposit contract cache" in the logs and notice your validator isn't making attestations. You can try rebooting the server itself if restarting services doesn't work, which might take more then once even in a day or so time period.
Check service logs to see "Successfully published attestations" and use the beacon explorer to confirm your validator(s) are making attestations again.
curl -s http://localhost:8545 -H "Content-Type: application/json" --data '{"jsonrpc":"2.0","method":"eth_syncing","params":[],"id":67}' | jq
You should see a similar output below from the command above
{
"jsonrpc": "2.0",
"id": 67,
"result": {
"currentBlock": "0xffe4e3", // THIS IS WHERE YOU ARE
"highestBlock": "0xffe8fa", // THIS IS WHERE YOU’RE GOING
[full output was truncated for brevity]
}
}
So you can compare the current with the highest to see how far you are from being fully sync’d. Or is result=false, you are sync'd.
curl -s http://localhost:8545 -H "Content-Type: application/json" --data "{\"jsonrpc\":\"2.0\",\"method\":\"eth_syncing\",\"params\":[],\"id\":67}" | jq
You should see a similar output below from the command above
{
"jsonrpc": "2.0",
"id": 67,
"result": false
}
$ curl -s http://localhost:5052/lighthouse/ui/health | jq
{
"data": {
"total_memory": XXXX,
"free_memory": XXXX,
"used_memory": XXXX,
"os_version": "Linux XXXX",
"host_name": "XXXX",
"network_name": "XXXX",
"network_bytes_total_received": XXXX,
"network_bytes_total_transmit": XXXX,
"nat_open": true,
"connected_peers": 0, // PROBLEM
"sync_state": "Synced"
[full output was truncated for brevity]
}
}
$ curl -s http://localhost:5052/lighthouse/syncing | jq
{
"data": "Synced"
}
$ sudo systemctl status geth lighthouse-beacon lighthouse-validator
● geth.service - Geth (Go-Pulse)
Loaded: loaded (/etc/systemd/system/geth.service; enabled; vendor preset: enabled)
[some output truncated for brevity]
Apr 00 19:30:20 server geth[126828]: INFO Unindexed transactions blocks=1 txs=56 tail=14,439,524 elapsed=2.966ms
Apr 00 19:30:30 server geth[126828]: INFO Imported new potential chain segment blocks=1 txs=35 mgas=1.577 elapsed=21.435ms mgasps=73.569 number=16,789,524 hash=xxxxd7..xxx>
Apr 00 19:30:30 server geth[126828]: INFO Chain head was updated number=16,789,xxx hash=xxxxd7..cdxxxx root=xxxx9c..03xxxx elapsed=1.345514ms
Apr 00 19:30:30 server geth[126828]: INFO Unindexed transactions blocks=1 txs=96 tail=14,439,xxx elapsed=4.618ms
● lighthouse-beacon.service - Lighthouse Beacon
Loaded: loaded (/etc/systemd/system/lighthouse-beacon.service; enabled; vendor preset: enabled)
[some output truncated for brevity]
Apr 00 19:30:05 server lighthouse[126782]: INFO Synced slot: 300xxx, block: 0x8355…xxxx, epoch: 93xx, finalized_epoch: 93xx, finalized_root: 0x667f…707b, exec_>
Apr 00 19:30:10 server lighthouse[126782]: INFO New block received root: 0xxxxxxxxxf5e1364e34de345ab72bf1632e814915eb3fdc888e5b83aaxxxxxxxx, slot: 300061
Apr 00 19:30:15 server lighthouse[126782]: INFO Synced slot: 300xxx, block: 0x681e…xxxx, epoch: 93xx, finalized_epoch: 93xx, finalized_root: 0x667f…707b, exec_>
● lighthouse-validator.service - Lighthouse Validator
Loaded: loaded (/etc/systemd/system/lighthouse-validator.service; enabled; vendor preset: enabled)
[some output truncated for brevity]
Apr 00 19:30:05 server lighthouse[126779]: Apr 06 19:30:05.000 INFO Connected to beacon node(s) synced: X, available: X, total: X, service: notifier
Apr 00 19:30:05 server lighthouse[126779]: INFO All validators active slot: 300xxx, epoch: 93xx, total_validators: X, active_validators: X, current_epoch_proposers: 0, servic>
For example, journalctl
will let you look at each client's debug logs. Recommend it with the -f
option to get the latest logs.
$ journalctl -u geth.service
$ journalctl -u lighthouse-beacon.service
$ journalctl -u lighthouse-validator.service -f (with -f to get the latest logs OR without it get the beginning logs)
You can also run the htop
command to check the CPU and memory usage of your validator server.
This helper script deletes all your validator data so you can try the setup again if you want a fresh install or feel like you made an error.
Be careful! It deletes and resets things, so read the code and make sure you understand what it does before using it.
Just some nice-to-haves if you're using Digital Ocean or AWS Cloud for your validator server.
It pulls updates from Gitlab, rebuilds the clients and restarts the services back again. Only supports Geth and Lighthouse.
Important Notice If you used the setup script and you made a new validator within the first 6-7 weeks of PulseChain, prior to early July 2023, the update script will not work for your server. It requires either a quick rebuild of the validator with new changes in the setup script OR git environment update before it can properly upgrade the clients. See this post which has the guidance for going through the process.
Update: If you would like to backup/restore the current clients as a safety measure, see the section below Backup clients helper before running the update script
Running the script is as simple as ./update-client.sh
. If you get a message to modify the script to make sure you understand the process, because the validator will be offline for likely 1 hour while the updates are taking place, make sure you understand and are OK with that. This is just a precaution to make sure you understand what you're doing before running the script. If all is good, you can do the following.
pico reset-validator.sh
and change line 6
from I_KNOW_WHAT_I_AM_DOING=false
to I_KNOW_WHAT_I_AM_DOING=true
, then ctrl+x
to exit, it will ask you to save so say y
and Enter
to save the changes.
And then ./update-client.sh
to start the process.
Once the process completes, you can verify the updated version of Geth and Lighthouse with these commands (respectively).
$ /opt/geth/build/bin/geth version
Geth
Version: 3.0.1-pulse-stable
$ sudo -u node bash -c "/opt/lighthouse/lighthouse/lh --version"
Lighthouse Lighthouse-Pulse/v2.4.0-2b37ea4
Optionally, the script backup-clients.sh
can be run with either option [backup | restore]
and it should be run before the update-client.sh
script, if performing the backup
or after an update in case the new binaries are not working properly by running it with the restore
parameter.
backup
: Will copy the currently running Geth and Lighthouse binaries into a backup folder in the node's user $HOME
home directory. This option is intended to be run before running the update script which pulls and builds the latest clientsrestore
Will replace the newly built binaries with the ones backed up (previous versions)Rollbacks
backup
: To rollback running this command simply remove the backed up binaries or the whole folder sudo -u node bash -c "rm -rf /home/node/backup"
(Run as node user)restore
: To rollback and run again the latest clients, the best thing to do is to re-run the update-client.sh
script againThis one allows you to update the network fee recipient and server IP address for Lighthouse.
These were specified during the initial PulseChain validator setup, however both of them may change for you over time, so the script allows you to easily update them and restart the clients.
$ ./update-fee-ip-addr.sh [0x...YOUR NETWORK FEE ADDRESS] [12.89...YOUR SERVER IP ADDRESS]
Enables the RPC interface so you can use your own node in Metamask (support for Firefox only). Running your own node and using it can be used for testing purposes, not relying on public servers or bypassing slow, rate limited services by "doing it yourself".
Do not expose your RPC publicly unless you know what you're doing. This script helps you more securely expose it to your own local environment for your own use.
If your RPC is...
Add your server to Metamask
Click the Network drop-down, then Add Network and Add a Network Manaully.
Now you can use your own node for transactions on the network that your validator is participating in.
Takes a backup snapshot of blockchain data on a fully synced validator so it can be copied over and used to bootstrap a new validator. Clients must be stopped until the snapshot completes, afterwards they will be restarted so the validator can resume normal operation.
It's recommended to use get up to speed on "resumable terminals" such as tmux and use it when you're doing long-running operations such as snapshots. This mitigates unnecessary failures such as a disconnection from the server, causing the process to be interrupted, and requiring a re-run of the script.
After running the script, copy the geth.tar.xz and lighthouse.tar.xz (compressed blockchain data, kinda like ZIP files) over to the new validator server (see scp demo below OR use a USB stick).
$ scp -i key.pem geth.tar.xz ubuntu@new-validator-server-ip:/home/ubuntu
$ scp -i key.pem lighthouse.tar.xz ubuntu@new-validator-server-ip:/home/ubuntu
Copying over the network could take anywhere from 1 hour to a few hours (depending on the bandwidth of your server's network).
Then you can run the following commands ON THE NEW SERVER
$ sudo systemctl stop geth lighthouse-beacon lighthouse-validator
$ tar -xJf geth.tar.xz
$ tar -xJf lighthouse.tar.xz
$ sudo cp -Rf opt /
$ sudo chown -R node:node /opt
$ sudo systemctl start geth lighthouse-beacon lighthouse-validator
The geth.tar.xz file is likely going to be > 100gb and the lighthouse compressed file probably smaller, but prepare for ~200gb of data total for the transfer. It can take 4-6 hours to decompress these blockchain data files as well, for example...
$ date; tar -xJf geth.tar.xz; date
03:32:56 UTC 2023
07:50:01 UTC 2023
Note: this should work fine for Ethereum too as it's just copying the blockchain data directories for Geth and Lighthouse, but the scenario is technically untested. Also, this relies on the new validator setup (which you are copying the snapshot to) to be setup with this repo's setup script.
Allows you to prune geth blockchain data to reduce disk space usage on the validator. Erigon does this automatically, but for maintenance pruning Geth regularly (quarterly or bi-yearly) is recommended to avoid the disk filling up.
You can also setup a cron job to do this automatically every quarter or 6 months, otherwise if you don't do the maintence, depending on your disk size, it can fill up and cause your validator to stop working properly.
References
AWS also offers a free tier option that allows you to spin up Linux VMs for free for a certain period of time, so you could use that for quick and easy tiny VMs running Ubuntu Linux. They are not beefy enough to be a validator, so that's not an option, but they are fine for small tasks and learning. You just need to sign up for an account and follow the instructions in the above document, except choose Free Tier options instead of the validator hardware configuration as described.
DO seems to be a lot cheaper per month to run validator servers vs AWS which is why many people may choose it as a cloud provider rather than AWS.
The monitoring-setup.sh and reset-monitoring.sh automate most of the setup for grafana and prometheus as well as let you reset (or remove) the monitoring, respectively.
You need to run the validator setup script FIRST and then use the monitoring setup script to "upgrade" the install with monitoring.
After running the monitoring setup script, you must finish the configuration at the Grafana portal and import the dashboards.
The standard config assumes you are not sharing the validator server with other people (local user accounts). Otherwise, it’s recommended for security reasons to set up further authentication on the monitoring services. TL;DR you should be the only one with remote access to your validator server, so ensure your keys and passwords are safe and do not share them with anyone for any reason.
If you are accessing your validator on the cloud or remote server (not logged into a monitor connected to the validator at home) You can setup grafana for secure access externally as opposed to the less secure way of forwarding port 3000 on the firewall and open it up to the world, which could put your server at risk next time Grafana has a security bug that anyone interested enough can exploit.
ssh -i key.pem -N ubuntu@validator-server-IP -L 8080:localhost:3000
Then open a browser window on your computer and login to grafana yourself without exposing it externally to the world. Magic, huh!
Go to http://localhost:8080
and login with admin/admin (as the initial username/password). It will then ask you to set a new password, make it a good one.
If you logged into it with a monitor connected to the validator (not accessing your validator from remote location)
Then you don't need to forward any ports or use SSH as you can open a browser on the validator (don't use it for generally surfing the internet, only for specific purposes like checking metrics) and go to http://localhost:3030
(not 8080) and with the same default login credentials of admin/admin, then set a new strong password.
Now to continue the monitoring setup...
In the lower left bar area, click the gear box -> Data Sources -> Add Data Source.
Use your mouse cursor to hover over the Dashboards icon (upper left bar area, 4 squares icon).
This Staking Dashboard was forked from the Ethereum-based one for PulseChain and has really good stats!
Staking Dashboard (one of the best ones)
There are also Dashboards for client stats, such as...
Geth
Lighthouse VC
Lighthouse Beacon
Now you can browse the Dashboards and see various stats and data!
Also see the guides below for additional help (scripts were mostly based on those instructions)
You can also setup email alerts on Grafana. See guide at the link below.
Basic Security Model for PulseChain/Ethereum Validator
Network Security
If running a validator in the cloud, there's already isolation away from your home network and the devices connected to it. However, if running a validator on your home network, the game is to keep attackers off of your home network. This is much easier when you're not inviting the public to connect to your server that sits on your network at home, but with validators, you're naturally exposing infrastructure running on your own network, which may be the same one you connect your personal devices to as well.
Recommended that if running a validator at home, you isolate it from everything else on your home network using another router into the mix, cascading routers or using VLANs and other kinds of network isolation or "guest" networks.
See references below for more information.
Validator Security AMA
References
There are ports that need to be exposed to the Internet for your validator to operate, specially for Geth and Lighthouse it's TCP/UDP ports 30303 and 9000 respectively. There are two common ways to control the firewall on your network: the Linux server and the network (such as your router or gateway to the Internet).
On the Linux server, you can open ports like this (as seen in the code).
# firewall rules to allow go-pulse and lighthouse services
sudo ufw allow 30303/tcp
sudo ufw allow 30303/udp
sudo ufw allow 9000/tcp
sudo ufw allow 9000/udp
This depends on your router device and model, so you'll need to research how to open ports on your specific networking device.
Security groups (firewall)
XX.YY.0.0/16
(replace XX.YY with first two numbers of your public IP address, assuming you only plan to access the server from that location)Analogous to AWS cloud network settings and IP restrictions as you need appropriate.
You can add graffiti flags for simple, one-line messages or ask Lighthouse to read from a file to express yourself in text for blocks your validator helps create. By default, the script does not set graffiti flags, but you can do this manually by editing the Lighthouse service files and adding in the flags and values that you want.
Check out the Lighthouse manual page on graffiti for instructions on how it works.
Example
$ sudo pico /etc/systemd/system/lighthouse-beacon.service
add something like this to the ExecStart= command (line 12)
--graffiti "richard heart was right"
$ sudo systemctl daemon-reload
$ sudo systemctl restart lighthouse-beacon
These instructions are only meant for use on Testnet and have not been tested on Mainnet, so only use them on Testnet until further testing and confirmation.
Be EXTRA CAREFUL as mistakes here can cost you funds and you must use these instructions at your own risk and hold yourself fully accountable for control and actions with your own funds, just like in the other parts of crypto.
There are full withdrawals and partial withdrawals. This section will focus on the full withdrawal and validator exit process.
If you set a withdrawal address when generating your validator keys, you can check on the launchpad withdrawal page to verify withdrawals are enabled and then exit your validator (see process below).
If you didn't set a withdrawal address when generating your validator keys, you need to "upgrade your keys" (generate BLSToExecution JSON) using the staking deposit client and broadcast it via the Launchpad, which as of now is unavailable. Will update with further instructions as this feature to support the scenario becomes available. Then, you can exit your validator from the network.
TREAT THIS AS IF YOU ARE GENERATING YOUR VALIDATOR KEYS + SEED WORDS
PERFORM IT ON A DIFFERENT, SECURE MACHINE (not your validator server)
Find the validator index for the specific validator you want to initiate an exit and withdrawal. Check on the beacon explorer with the validator’s public key (for example, 4369).
Download the latest staking deposit
$ git clone https://gitlab.com/pulsechaincom/staking-deposit-cli.git
$ ./deposit.sh generate-bls-to-execution-change
English
pulsechain
(Enter seed words)
0
(Enter each validator index as currently shown on the network)
You can open your deposit JSON file and copy each validator’s public key into the beacon explorer one at a time to get each index, but they should be sequential, like 4369, 4370, 4371, etc OR just the launchpad as it will show you validator index and withdrawal_credentials.
(Enter each withdrawal_credentials for each validator which is also in the deposit JSON file)
(Enter your execution address, aka withdrawal address, the wallet you’ve securely, made sure you and only you have access to and want to send the deposited PLS from the network back into)
DOUBLE CONFIRM YOU HAVE ACCESS TO THIS WALLET ADDRESS
It cannot be changed and if you typo something here, TWICE, YOUR FUNDS WILL BE UNRECOVERABLE.
Once you’ve double checked, enter the address again
Now your bls_to_execution_change JSON file is in the newly created bls_to_execution_changes folder.
Note: this was tested on Testnet V4 but has not been rigorously tested on PulseChain mainnet. See references for more info.
You can broadcast the change using your Lighthouse client (for the specific validator you want to exit and initiate withdrawal).
$ sudo -u node bash
$ /opt/lighthouse/lighthouse/lh account validator exit --network pulsechain --keystore ~/.lighthouse/pulsechain/validators/0x…(the validator public key you want to exit)/keystore-(specific for your setup)...json
Enter the keystore password
Enter the exit phrase described @ https://lighthouse-book.sigmaprime.io/voluntary-exit.html
“Successfully validated and published voluntary exit for validator 0x...” – and we can check it’s status on the beacon explorer
"Waiting for voluntary exit to be accepted into the beacon chain..."
"Voluntary exit has been accepted into the beacon chain, but not yet finalized. Finalization may take several minutes or longer. Before finalization there is a low probability that the exit may be reverted."
https://beacon.pulsechain.com/validator/0x...
And you can see it’s going from Active to Exit (pulsing green).
Once it's exited, you have to wait for Withdrawals to become available.
This validator has exited the system during epoch 5369 and is no longer validating.
There is no need to keep the validator running anymore. Funds will be withdrawable after epoch 5555.
References
You can do various types of backups.
For the full server backup, see the Home section below.
For blockchain data backup only, check out the Snapshot helper script and guidance.
For snapshotting your disk on AWS cloud server, see the Cloud section below.
You can use various tools on Linux to make scheduled backups to another disk OR another server.
Most use rsync, cron jobs or programs like Timeshift to automate the process.
Using Timeshift to backup the server (eg. on a 1TB or 2TB big USB or external hard drive)
Further Guides and References
Snapshots
You should see in green...
It will be in Pending status for a while before the process completes (could be a few hours).
Using a Snapshot
You now have a new server with a hard disk volume based on the snapshot of the other server, yay!
There are many ways to get uptime monitoring alerts for your validator, some more complex and invasive than others. You can setup a mail server and detect outage condition X or Y, probably even monitor Grafana for alerts. There are pros and cons using your validator server to do this vs leveraging external services.
One light-weight way to get alerts if your validator "goes offline" is to use an external service to simply check and see if at least (1) condition has been met for a properly functioning validator server. In this example, we're using UptimeRobot and alerting if our server fails to meet a condition such as the Geth client on port 30303 is unreachable.
This will catch "validator is down" scenarios such as...
Be aware that this isn't meant to catch all of them. However, it should work for common downtime situations such as hardware failures, server crashes or networking issues.
First, you'll need to sign up @ www.uptimerobot.com and then on the dashboard, choose Add New Monitor.
Friendly Name: Validator 30303 IP: [validator IP address] Port: Custom Port Port: 30303 URL: ping
See the “Monitor Created” notification in green
Go to monitor detail
Friendly name: Validator 30303
Click on the eye icon to see your public status page The URL will look something like https://stats.uptimerobot.com/XYZ369abc5555.
Testing notification setup
1) WAIT 24-36hrs before testing for outage alerts, service may not pick up outages or be very effective prior to 24-36hrs period AFTER setting up alerts for your validator 2) “Send test notification” to try it out and see what it sends to your inbox 3) Or “sudo systemctl stop geth” and see what it does with a real outage (NOTE: this may cause some penalties to your validator and affect uptime % stats obviously, but should be minimal)
Now if your validator goes offline and meets the conditions we're checking for (port 30303 is reachable), you should get an email from UptimeRobot and also see updates on the UptimeRobot dashboard and status page.
Specs and preferences vary between who you talk to, but at least 32gb ram and a beefy i7 or server-based processor and 2TB SSD hard disk. In the cloud, this roughly translates into a M2.2XLarge EC2 instance + 2TB disk.
It depends on your bandwidth, server specs and the if the network is running smoothly, but you should expect anywhere from 24 - 96hrs for a validator node to sync. Then 12-36hrs afterwards for your validator to be activated and added to the network. So it can take 2-3 days for the full process.
Yes, you can run as many validators as you want on your server. Only caveat being that if you plan to run 100+, you may want to double the specs (at least memory) on your hardware to account for any additional resource usage. If you plan on running 1 or 10 or even 50, the minimum recommended hardware specs will probably work.
The setup script has no dependencies on the number of validators you run, it simply installs the clients and when you generate your validator keys with the staking deposit tool, there you choose the specific number you want to run. It could be 1, 5, 10 or 100. Then, when you import your keys to Lighthouse, you will import each key and it will configure the client to run that number of validators.
Check out the Adding more validators guide @ CoinCashew.
Let's say you have generated your keys for 1 validator and want to add 2 more. It would look something like this for pulsechain.
$ ./deposit.sh existing-mnemonic --validator_start_index 1 --num_validators 2 --chain pulsechain --eth1_withdrawal_address 0x..(MAKE SURE YOU WILL ALWAYS HAVE ACCESS TO THIS WALLET)
You can also check out instructions here.
Look at your deposit JSON file to get the list of your validator(s) public keys, then check https://beacon.pulsechain.com/validator/ + your validator's public key which each one that you want to check the stats on.
For example this validator's stats: https://beacon.pulsechain.com/validator/8001503cd43190b01aaa444d966a41ddb95c140e4910bb00ad638a4c020bc3a070612f318e3372109f33e40e7c268b0b
You can also setup Monitoring with Grafana and Prometheus (see guidance in above sections).
Did your server's IP address change? If so, update lighthouse beacon service file @ /etc/systemd/system/lighthouse-beacon.service or use the update-fee-ip-addr.sh script to update both the fee address + server IP address.
Did your network/firewall role change? Make sure the required client ports are accessible.
What is your status on the beacon explorer? Active, Pending, Exited or something else? If not active, it may be a client issue which you can debug with the steps discussed in the Debugging section.
Are your clients fully synced? They must be synced, talking to each other and talking to the network for the validator to work properly.
Depends on if you're using your own hardware or the cloud. For example, you could build or buy your own hardware for initial cost of $1k-$2k and then pay for electricity it uses from running 24/7 each month. Or you can rent a server in the Amazon AWS cloud for an (estimated) few hundred dollars per month. Both ways have advantages and disadvantages.
You could getting wrong votes from time to time or otherwise which is just a natural part of participating in the system. Also, you could be seeing this because it's withdrawaling rewards in batches automatically to your withdrawal address.
Also, you can learn more about the different kinds of validator rewards here.
References
There's only a couple maintenance items that can be completed manually (or by running the supporting scripts in the repo) or setup to run automatically with a cronjob, for example.
Again, there are scripts in the repo that make this easy. Check this tweet for information on pruning the validator.
This could depend on the number of validators, but generally speaking 50mbps seems to be fine for most cases, eg. 1-10 validators.
10+ or 100+ validators may benefit from 100mbps+ connection, but its always better to have too fast of a connection vs too slow. Gigabit is great if available and practical for situation.
Yes, by default the script will use mainnet settings, however if you edit the script and comment out mainnet and uncomment the testnet parameters, you can run a testnet v4 validator.
First, change the client chain flags.
# chain flags
GETH_MAINNET_CHAIN="pulsechain"
LIGHTHOUSE_MAINNET_CHAIN="pulsechain"
GETH_TESTNET_CHAIN="pulsechain-testnet-v4"
LIGHTHOUSE_TESTNET_CHAIN="pulsechain_testnet_v4"
# default=mainnet, comment/uncomment to switch to testnet
GETH_CHAIN=$GETH_MAINNET_CHAIN
LIGHTHOUSE_CHAIN=$LIGHTHOUSE_MAINNET_CHAIN
#GETH_CHAIN=$GETH_TESTNET_CHAIN
#LIGHTHOUSE_CHAIN=$LIGHTHOUSE_TESTNET_CHAIN
Then, change the checkpoint URL.
# checkpoint urls
LIGHTHOUSE_MAINNET_CHECKPOINT_URL="https://checkpoint.pulsechain.com"
LIGHTHOUSE_TESTNET_CHECKPOINT_URL="https://checkpoint.v4.testnet.pulsechain.com"
LIGHTHOUSE_CHECKPOINT_URL=$LIGHTHOUSE_MAINNET_CHECKPOINT_URL
#LIGHTHOUSE_CHECKPOINT_URL=$LIGHTHOUSE_TESTNET_CHECKPOINT_URL
So, to run testnet instead of mainnet, the testnet options should be enabled and look something like this.
GETH_CHAIN=$GETH_TESTNET_CHAIN
LIGHTHOUSE_CHAIN=$LIGHTHOUSE_TESTNET_CHAIN
LIGHTHOUSE_CHECKPOINT_URL=$LIGHTHOUSE_TESTNET_CHECKPOINT_URL
Binge the Validator Playlist. It covers just about every question and scenario you could run into during setup and understanding operation.