Closed gituser closed 4 years ago
Seems this issue also affects v2.7.2 as well (from gitter):
MinwooJ @MinwooJ Aug 28 09:00
Hello~
I'm running a ropesteen testnet through Parity (v2.7.X), and it's stuck syncing at block #8579999.
But when you look at etherscan, it's currently sinking from #8569345, and some other explorer's sync is stuck at #8579999, like me.
Ropsten's chain seems to have divided, so I'd like to know which is right.
This affected all my parity nodes. sudo parity --chain ropsten db kill
and subsequent default resync did not fix the issue. Stalling again at block 8579999.
Currently attempting a default resync with openethereum v3.0.0-stable
Both of my attempts with v3.0.0-stable
resynced to #8579999:
--warp-barrier 8563500
I do confirm the issue, no new block after #8579999 on docker image: parity/parity:v2.5.13-stable
Any OpenEthereum's developers reaction?
Ping @adria0 @rakita @vorot93 @seunlanlege @denisgranha @sorpaas
Hello, sorry for not responding, to be honest, I don't have anything to add.
There were multiple reorgs, first one was 194 blocks at #8568394 biggest one was 11606 blocks at #8568394.
I tried to sync it with pruning DB and as guys said it gets stuck, archive node should be able to do this big reorg. Maybe it is connected with warping, I got warped from #8575000 and this snapshot is probably made on abandon fork.
I am currently syncing without warping to see what happens (I am currently at #5780341), will update when finish.
I hope it will resolved quickly... ETH Mainnet gas price is crazy for test T.T
https://ropsten.etherscan.io/blocks_forked?p=8
One reorg at 8568394
was 11606
in depth
. 8568394 + 11606
== 8580000
.
@holiman so what to do in order to get parity node working on ropsten again?
any news guys?
Just in time The first takeaway, as you already know, this was a 51% attack on the Ropsten test network and they succeeded to revert two days of worth of blocks and the attacker chain became valid. Just to say it, this is a test network, and these things can happen there.
Okay now, for things related to OE:
What you can do:
--no-warp
and this passed for me without a problem. Here is log--warp-barrier 8690000
with this you will force skip problematic snapshot (but for the next few days I not sure how many peers will have valid new snapshot), in future as new snapshots are generated this force skip will not be needed.@rakita Thank you for your suggestions. I completed a --warp-barrier 8563500
reync on Saturday alas no success. Will try my luck with --no-warp
today
If you know any nodes that have a post-attack snapshot, that would be great to get to use as --reserved-peers to get a successful warp sync.
@kepikoi --warp--barrier
at that block number probably take 8575000
. What barrier does is enforce that Snapshot must be greater than the warp barrier if any
@rakita
I had parity ropsten's backup from block #7857894 and resync didn't work for me with warp disabled and tracing = on
I have in my parity_testnet.toml:
[network]
# Parity will listen for connections on port 50315.
port = 30303
# Parity will connect only to public network IP addresses.
allow_ips = "public"
# Min Peers
min_peers = 25
# Max Peers
max_peers = 50
# Disable serving light peers.
no_serve_light = true
# Parity will sync by downloading latest state first. Node will be operational in couple minutes.
warp = false
[footprint]
# Compute and Store tracing data. (Enables trace_* APIs).
tracing = "on"
# Prune old state data. Maintains journal overlay - fast but extra 50MB of memory used.
pruning = "fast"
# Will keep up to 1024 old state entries.
pruning_history = 12000
# Will keep up to 128 MB old state entries.
pruning_memory = 128
# Will keep up to 4096MB data in Database cache.
cache_size_db = 4096
I've set now bootnodes from geth https://github.com/ethereum/go-ethereum/blob/master/params/bootnodes.go#L37
and pruning_history = 32000
will see how it go.
I've noticed you have in your logs:
UTC Starting Parity-Ethereum/v2.5.13-stable-b5695c1d7-20200824/x86_64-linux-gnu/rustc1.45.2
whilst I have parity-v2.5.13-stable-253ff3f-20191231
built from source, did you apply additional commits to your build?
NVM, found: https://github.com/vorot93/oe_playground/commit/b5695c1d73cac2e9d445905c6d13985a96fcafae
I've managed to get in sync.
Here is what I've used:
first you need your communication port accessible outside, default port is 30303, if you're running behind NAT, you need to add into your config, where 1.2.3.4
is your outside IP:
[network]
port=30303
nat = "extip:1.2.3.4"
and also forward this port 30303
on your router / iptables to your machine
add bootnodes (these are from geth):
# Override the bootnodes from selected chain file.
[network]
bootnodes = ["enode://30b7ab30a01c124a6cceca36863ece12c4f5fa68e3ba9b0b51407ccc002eeed3b3102d20a88f1c1d3c3154e2449317b8ef95090e77b312d5cc39354f86d5d606@52.176.7.10:30303", "enode://865a63255b3bb68023b6bffd5095118fcc13e79dcf014fe4e47e065c350c7cc72af2e53eff895f11ba1bbb6a2b33271c1116ee870f266618eadfc2e78aa7349c@52.176.100.77:30303"]
add warp=false
[network]
# Parity will sync by downloading latest state first. Node will be operational in couple minutes.
warp = false
also increase pruning_history
to 20000, not sure it might work with lower value, but just to be sure
[footprint]
# Compute and Store tracing data. (Enables trace_* APIs).
tracing = "on"
# Prune old state data. Maintains journal overlay - fast but extra 50MB of memory used.
pruning = "fast"
# Will keep up to 1024 old state entries.
pruning_history = 20000
pruning_memory = 128
cache_size_db = 4096
NOTE: parity will require a LOT of memory in this configuration about ~30GB or more in my case with `v2.5.13`.
Probably you can change certain value back to defaults or lower them after the problem is mitigated.
@gituser not sure what happened there. I think it helped that you used different bootnodes
Sorry to bother @gituser @rakita but how did you manage sync so fast? I'm on a 8 core / 32GB / premium SSD VM and looking at this graph (diff to reference node) it could take a week to --no-warp
Ropsten
Some relevant parts from my config:
[parity]
mode = "active"
chain = "ropsten"
[network]
min_peers = 25
max_peers = 50
max_pending_peers = 50
[footprint]
tracing = "off"
pruning = "fast"
db_compaction = "ssd"
cache_size = 4096
@rakita trying to warp sync on another machine for two days with --warp-barrier 8690000
but just like you assumed it stalled lacking a complete snapshot
2020-09-04 07:49:42 UTC IO Worker #0 INFO import Syncing snapshot 2870/3834 #0 23/25 peers 920 bytes chain 100 KiB db 0 bytes queue 172 KiB sync RPC: 0 co
nn, 0 req/s, 16 µs
2020-09-04 07:49:47 UTC IO Worker #3 INFO import Syncing snapshot 2870/3834 #0 23/25 peers 920 bytes chain 100 KiB db 0 bytes queue 172 KiB sync RPC: 0 co
nn, 0 req/s, 16 µs
2020-09-04 07:49:52 UTC IO Worker #0 INFO import Syncing snapshot 2870/3834 #0 23/25 peers 920 bytes chain 100 KiB db 0 bytes queue 172 KiB sync RPC: 0 co
nn, 0 req/s, 16 µs
2020-09-04 07:49:57 UTC IO Worker #2 INFO import Syncing snapshot 2870/3834 #0 23/25 peers 920 bytes chain 100 KiB db 0 bytes queue 172 KiB sync RPC: 0 co
nn, 0 req/s, 16 µs
@kepikoi I had a backup previously from block #7857894 that's why it was much faster.
I also exported ropsten's recent blocks and state into: snap and rlp files I can share them somewhere probably for everyone.
$ du -sh * --total
58G ropsten_blocks_20200903.rlp
12G ropsten_snapshot_20200903.snap
70G total
@kepikoi also try adding bootnodes
into your config and you might try increasing buffers a little for:
# Override the bootnodes from selected chain file.
[network]
bootnodes = ["enode://30b7ab30a01c124a6cceca36863ece12c4f5fa68e3ba9b0b51407ccc002eeed3b3102d20a88f1c1d3c3154e2449317b8ef95090e77b312d5cc39354f86d5d606@52.176.7.10:30303", "enode://865a63255b3bb68023b6bffd5095118fcc13e79dcf014fe4e47e065c350c7cc72af2e53eff895f11ba1bbb6a2b33271c1116ee870f266618eadfc2e78aa7349c@52.176.100.77:30303"]
[footprint]
# Will keep up to 1024 old state entries.
pruning_history = 20000
# Will keep up to 128 MB old state entries.
pruning_memory = 128
# Will keep up to 1024MB data in Blockchain cache.
cache_size_blocks = 1024
# Will keep up to 512MB of blocks in block import queue.
cache_size_queue = 512
# Will keep up to 256MB data in State cache.
cache_size_state = 256
Sorry to clutter the thread, but will this issue eventually be "fixed" in a way that makes it easy for me to get my node going again or am I going to have to roll up my sleeves?
Eventually, there will be nodes that have post-attack snapshots and that you will be able to use for warp sync, and then things should get back to normal. If anyone knows the enode-URI for such nodes, please let me know, I'd love to use such a snapshot to get my warp/fast-sync node back up.
@noconsulate no you can't rollback.
It's by design if you have fast node (that means you only keep only certain amount of state history, by default 64 last states), you can't rollback easily.
If you run an archive node you can rollback to any block and start syncing from it again.
Fortunately it's no big deal to wipe the database and sync all over again because Ropsten is relatively small. How will I know when there are post-attack snapshots? Since this happened a week ago is it safe to assume now is a good time to sync?
I had resync from genesis with warp=false
during 5 days. but it eventually failed and stuck on 8579999.
so I dont think now is a good time to sync.
but I had made DB snapshot at 8556199. so I remove stuck DB and retry sync.
My node restart sync with option pruning_history = 20000
at this time.
and eventually jump 8579999.
I'm happy. thank you @gituser
Tried the recommendations from above over the weekend. Still no luck :(
v2.6.8-beta
node using @gituser's bootnodes config again and stalled at block #8579999no-warp
resynced my v3.0.0-stable
node all over again and stalled at block #8579999Will try the backup route next...
I deleted my database and re-synced with out any special parameters. I set it and forgot about it and now it's stuck on #8579999 once again.
version OpenEthereum/v3.0.1-stable-8ca8089-20200601/x86_64-unknown-linux-gnu/rustc1.43.1
This should work:
Add these options to the [network]
section in the config.toml
file:
[network]
reserved_peers = "<path to peers file .e.g /config/peers >"
reserved_only = true
min_peers = 1
max_peers = 25
warp = false
This should force the sync to be only from the nodes specified in the peers file.
Create a corresponding peers
file with the following nodes:
enode://8a4bc82041f560873d48f9f0d6754677096880195fc2eab3d57783370483efca8a36009d3fd87e4ee497ee7286425b886a16e6047f876b863666e57276ef8aad@35.246.127.164:30303
enode://74b86aed3ff9b23c8e423b673293044ec5b5525b43c237ac35a8634ea051dfd3cf4365c15133c8064c713684cd6a9911c15b32a7323f9db1d70f01706a682501@35.246.26.6:30303
enode://8c5131f577ee602ccaad5e5f600011c024d43d33e6af6f8a89ed26cbfadd7efe903a8e426dde4071b857e0838a2efa29ae3b268f8b55acd59e72d4a919673cbc@13.251.47.174:30303
The nodes above have the correct state (same as ropsten.etherscan) so your node should start syncing the correct state from those peers.
After the node is fully synced - you can remove the reserved_peers
and reserved_only
options from the configuration file and delete the peers file.
I finally got a snapshot at block #8645000 today, so my sync with --warp-barrier=8580000 succeeded and I should have a useful node again, even though I miss the event logs that were helpful in testing, but more important to actually have a node that works :)
Here I added a commit that will mark block 8,563,601 from the original chain as invalid and force the client to take the current chain. Now when you sync without warp it will go to current chain or block syncing if you dont have peers with current chain.
Thank you @nostdm , your solution helped!
any news? it still not work
@trongcauhcmus
Try this solution https://github.com/openethereum/openethereum/issues/11862#issuecomment-688521370
i tried many times. this is the log:
2020-10-12 03:54:56 UTC IO Worker #1 INFO import 0/ 1 peers 74 KiB chain 54 MiB db 0 bytes queue 5 MiB sync RPC: 0 conn, 0 req/s, 32 µs
2020-10-12 03:55:21 UTC IO Worker #3 INFO import Syncing #8579999 0x0b22…26bc 0.00 blk/s 0.0 tx/s 0.0 Mgas/s 0+ 0 Qed #8596041 1/ 1 peers 74 KiB chain 54 MiB db 0 bytes queue 5 MiB sync RPC: 0 conn, 1 req/s, 32 µs
2020-10-12 03:55:26 UTC IO Worker #1 INFO import Syncing #8579999 0x0b22…26bc 0.00 blk/s 0.0 tx/s 0.0 Mgas/s 0+ 0 Qed #8596041 1/ 1 peers 74 KiB chain 54 MiB db 0 bytes queue 5 MiB sync RPC: 0 conn, 0 req/s, 32 µs
2020-10-12 03:55:46 UTC IO Worker #3 INFO import Syncing #8579999 0x0b22…26bc 0.00 blk/s 0.0 tx/s 0.0 Mgas/s 0+ 0 Qed #8596041 1/ 1 peers 74 KiB chain 54 MiB db 0 bytes queue 5 MiB sync RPC: 0 conn, 0 req/s, 32 µs
2020-10-12 03:56:16 UTC IO Worker #2 INFO import 0/ 1 peers 74 KiB chain 54 MiB db 0 bytes queue 5 MiB sync RPC: 0 conn, 0 req/s, 32 µs
2020-10-12 03:56:46 UTC IO Worker #1 INFO import 0/ 1 peers 74 KiB chain 54 MiB db 0 bytes queue 5 MiB sync RPC: 0 conn, 0 req/s, 54 µs
2020-10-12 03:56:56 UTC IO Worker #0 INFO import Syncing #8579999 0x0b22…26bc 0.00 blk/s 0.0 tx/s 0.0 Mgas/s 0+ 0 Qed #8596041 1/ 1 peers 74 KiB chain 54 MiB db 0 bytes queue 5 MiB sync RPC: 0 conn, 0 req/s, 54 µs
2020-10-12 03:57:26 UTC IO Worker #1 INFO import Syncing #8579999 0x0b22…26bc 0.00 blk/s 0.0 tx/s 0.0 Mgas/s 0+ 0 Qed #8596041 1/ 1 peers 74 KiB chain 54 MiB db 0 bytes queue 5 MiB sync RPC: 0 conn, 0 req/s, 54 µs
2020-10-12 03:57:56 UTC IO Worker #2 INFO import Syncing #8579999 0x0b22…26bc 0.00 blk/s 0.0 tx/s 0.0 Mgas/s 0+ 0 Qed #8596041 1/ 1 peers 74 KiB chain 54 MiB db 0 bytes queue 5 MiB sync RPC: 0 conn, 0 req/s, 54 µs
2020-10-12 03:58:31 UTC IO Worker #2 INFO import 0/ 1 peers 74 KiB chain 54 MiB db 0 bytes queue 5 MiB sync RPC: 0 conn, 0 req/s, 54 µs
this is my config:
[parity]
chain = "ropsten"
[network]
# Parity will listen for connections on port 30304.
reserved_peers = "/data/peers.conf"
reserved_only = true
min_peers = 1
max_peers = 25
warp = false
[rpc]
disable = false
port = 8545
interface = "10.xx.0.xx"
cors = ["test-parity"]
apis = ["web3", "eth", "pubsub", "net", "parity", "parity_pubsub", "traces", "rpc", "shh", "shh_pubsub", "personal", "parity_accounts", "parity_set", "signer"]
hosts = ["all"]
[footprint]
tracing = "on"
# Prune old state data. Maintains journal overlay - fast but extra 50MB of memory used.
pruning = "fast"
# # Will keep up to 1024 old state entries.
pruning_history = 50000
# # Will keep up to 128 MB old state entries.
pruning_memory = 128
# # Will keep up to 4096MB data in Database cache.
cache_size_db = 4096
[misc]
logging = "own_tx=trace, rpc=trace, txqueue=trace"
log_file = "/var/log/parity.log"
color = true
Can you help me check is everything ok?
Do we need to clear database and resync?
@trongcauhcmus Yes, you need to clear your database, add the config lines from this comment https://github.com/openethereum/openethereum/issues/11862#issuecomment-688521370 and resync.
@trongcauhcmus yes you need to clear DB and start again. Another option other what gituser mentioned is if you are can build from source, to use this commit that invalidates the first block in bad chain: https://github.com/openethereum/openethereum/issues/11862#issuecomment-690948102
and ether way use --warp-barrier=8850000
option to skip to the newest snapshot.
There is nothing more to be done here, hence I am closing this.
thank you @gituser @rakita
Seems ropsten network had a split at block https://ropsten.etherscan.io/block/8563600
Parity v2.5.13 is stuck at block #8579999 while according to https://ropsten.etherscan.io last block is #8578042
Parity somehow went into forked chain at block #8563600 (the hash of this block differs from the one at ropsten.etherscan.io, whilst previous block #8563599 matches the hash).
From the parity logs:
1) initially it imports correct block with correct hash 0xeda7...6a26 which matches https://ropsten.etherscan.io/block/8563600
2) and then happens a reorg:
It might be a similar issue as happened on Ethereum Classic some time ago https://github.com/openethereum/openethereum/issues/11843
I'm running parity with
--pruning-history 1024
so can't even rollback as the latest block went way past 1024 and I'm not sure that full re-sync will help in this case.Any ideas how to fix this issue?