Closed wpank closed 4 years ago
Pasting my comments from riot:
I just confirmed that this can happen due to the changes introduced in the keystore API for remote signing. https://github.com/paritytech/substrate/pull/6178/files#diff-233fff7c89a6b4ca96ca4ee31c62088cL385.this would never fail previously. The GRANDPA voter didn't crash, it failed to sign a message and it deliberately took the system down. I think this is the correct behavior as not being able to sign messages is a critical error, and I think it's better to take the node down rather than pretend it is working. In order to trigger the issue I setup an invalid session key through the RPC API (generated a sr25519 keypair and set it as grandpa key which is ed25519). IMO the root issue is that the key was accepted by the server whereas it should have been rejected, otherwise we only run into the problem whenever we try to sign something.
I have no clue how/if this related to authority discovery. We need more info from the people that have had this error.
Also some more anecdotes:
How likely the last two are is very much up in the air.
I think we can close this now. Seems to be fixed in v0.8.22 and we don't have new reports.
2022-02-15 23:39:58.380 INFO main sc_cli::runner: Substrate Node
2022-02-15 23:39:58.380 INFO main sc_cli::runner: ✌️ version 4.0.0-dev-b53da9f-x86_64-linux-gnu
2022-02-15 23:39:58.380 INFO main sc_cli::runner: ❤️ by Substrate DevHub https://github.com/substrate-developer-hub, 2017-2022
2022-02-15 23:39:58.380 INFO main sc_cli::runner: 📋 Chain specification: My Custom Testnet
2022-02-15 23:39:58.380 INFO main sc_cli::runner: 🏷 Node name: MyNode01
2022-02-15 23:39:58.380 INFO main sc_cli::runner: 👤 Role: AUTHORITY
2022-02-15 23:39:58.380 INFO main sc_cli::runner: 💾 Database: RocksDb at /tmp/node01/chains/local_testnet/db/full
2022-02-15 23:39:58.380 INFO main sc_cli::runner: ⛓ Native runtime: node-template-100 (node-template-1.tx1.au1)
2022-02-15 23:39:58.645 WARN main sc_service::config: Using default protocol ID "sup" because none is configured in the chain specs
2022-02-15 23:39:58.646 INFO main sub-libp2p: 🏷 Local node identity is: 12D3KooWB7x4EruG852DyxwigtTDsJ82wMV7c1uG5waJE2VxMcpo
2022-02-15 23:39:58.647 INFO main sc_service::builder: 📦 Highest known block at #0
2022-02-15 23:39:58.648 INFO tokio-runtime-worker substrate_prometheus_endpoint: 〽️ Prometheus exporter started at 127.0.0.1:9615
2022-02-15 23:39:58.650 INFO main parity_ws: Listening for new connections on 127.0.0.1:9945.
2022-02-15 23:39:59.326 ERROR tokio-runtime-worker afg: GRANDPA voter error: Signing("Failed to sign GRANDPA vote for round 1 targetting 0x3d3b1dae280f17e0f419618adee764457579da9dc2050250f4ec914cc3e27323")
2022-02-15 23:39:59.326 ERROR tokio-runtime-worker sc_service::task_manager: Essential task grandpa-voter
failed. Shutting down service.
Error: Service(Other("Essential task failed."))
I was following this tutorial: https://docs.substrate.io/tutorials/v3/private-network/ Private Network failed to run after inserting a grandpa key...
Everyone. Please help me.
When I run my substrate node server by running this command "./target/release/node-template --dev --ws-external --rpc-external", server works well and start generating blocks first. But after few hours, my server is been stopped and also stopped generating blocks automatically. How can I solve this problem? Here is Terminal Error. Best Regards.
I also had a similar error.
After deleting the chain data in --base-path, it was executed normally.
The command below may help.
./target/release/node-template purge-chain --base-path /tmp/node01 --chain local -y
/tmp/node01
<-- --base-path value used when starting the node
In addition,
If the password entered when generating the Sr25519 and Ed25519 keys is incorrectly entered when starting the node, the following error occurs.
GRANDPA voter error: could not sign outgoing message: Failed to sign GRANDPA vote for round 1 targetting
if Sr25519 instead Ed25519 for generating gp keys you will get this error
More than one Validators in in the Kusama Validator lounge reported their nodes failing with the logs saying something like this: