Open kogeler opened 1 year ago
Kusama DB (1.1TB): https://storage.googleapis.com/debug-releases/debug/debug-paritydb.tar
@kogeler Woud it be possible to get SSH access to the machine as well?
@arkpar I think it is not possible because it's just a pod in a k8s cluster. Yes, it uses a dedicated k8s node, but it isn't trivial to connect to the runtime environment. I shutdown the pod and made a snapshot of the GCP disk to upload the copy of the DB.
Once the index file stops fitting in memory, hash index search becomes a major bottleneck. There will be some improvenets in the next parity-db release, but ultimately this will be resolve with #199, which will remove index lookups for trie nodes.
@kogeler Could you then upload a copy of rocksdb database as well? For some reference testing.
@arkpar Do you need a full-synced archive copy of rocksdb?
@kogeler Fully synced or around the same block as the parity-db snapshot (15m -ish). If it is too much trouble, I can probably sync one myself, even though it will take a few days.
@arkpar You can download our public periodic DB snapshots using the manual
@arkpar Are there any updates?
We are working on a major feature (#199) that will resolve this. It will take a few weaks to land in substrate/polkadot.
@arkpar Are there any updates about this issue?
Substrate integration is a work in progress that can be tracked here: https://github.com/paritytech/polkadot-sdk/pull/3386
I tested the Rococo chain.
To sync a Rococo paritydb archive node, it took 13 days, whereas it took just 40 hours on rocksdb. The hardware is the same.
Block height paritydb:
Block height rocksdb:
role: full(archive) binary: docker pull parity/polkadot:v0.9.42 instance: GCP - t2d-standard-4 disk: GCP - SSD persistent disk OS: Container-Optimized OS from Google kernel: 5.10.162+
CLI flags:
I'm trying to sync backup nodes from scratch. I have 8 nodes (Kusama, Polkadot, archive, prune, rocksdb, paritydb).
I use the same instances, regions, and CLI flags. All nodes have 100 peers (in 75/out 25).
2 archive rocksdb nodes (Kusama, Pollkadot) synced in a couple of days. But 2 archive paritydb nodes (Kusama, Pollkadot) have been syncing for 1,5 weeks. At some point (around 15M blocks), the sync rate decreased quickly. Now it is less than 0 blocks/second. Restars don't help. It looks like an issue of paritydb.
The current state is: Kusama - target=#18852634 (100 peers), best: #15387848 (0x117e…c8a0), finalized #15387648 (0x065b…4684), ⬇ 705.8kiB/s ⬆ 461.9kiB/s Polkadot - target=#16463661 (100 peers), best: #15045441 (0x9fc3…0db0), finalized #15045402 (0xfef9…401b), ⬇ 134.9kiB/s ⬆ 125.2kiB/s
The disk sub-system is overloaded: 15k iops and 100MB/s by reading.