Closed Bidon15 closed 1 year ago
Given #9155, is this issue still relevant?
@Bidon15 Further to @thanethomson's comment, is the bad behaviour seen in v0.37.x (or in a later release of v0.34.x)?
Note that all releases in the v0.35.x
branch have been retracted as @thanethomson is pointing out.
If this bad behaviour is only seen in v0.35.x
we will close this issue. Feel free to re-open in the future if you see it happening in v0.34.x
, v0.37.x
or any future release.
Hey @sergio-mena
On v0.34.x the issue is not reproducible. Can't say for 0.37.x
Cool, thanks @Bidon15. We'll close this issue for now then. If you see this happen again, please feel free to reopen it.
Introduction
At Celestia we are relying on the ability to generate and broadcast large transactions, hence using testground, we emulated the scenario when a set of validators in the network can generate and broadcast transaction up to 500kb each that will be included in the next block (e.g. 3 validators generating and broadcasting 500kb each results in a ~1.5mb next block size)
We would like to demonstrate below how the same test scenario and environment produced different outcome for tendermint v0.35.6(*) and latest downgraded v0.34.20
(*) - We have added two ABCI++ methods for our needs and this change https://github.com/celestiaorg/celestia-core/pull/793
Environment
Testground Network Configuration
Bandwidth: 100 and 256Mib
Latency: 0ms
Config.toml for each of the validator
Mempool
Consensus
RPC
Notes:
Test Scenario
Pre-Requisites:
(**) - this means that we are communicating with the BL of the node as if we are node operators and using CLI commands
Steps:
Expected Results:
timed out waiting for tx to be included in a block
Actual Results
v0.35.x
The whole chain get stuck in rounds for 5 minutes without either:
v0.34.x
After successful downgrade to latest version of v0.34.x, we observed:
More Info:
Logs from testground and each of the validators can be found in this issue: https://github.com/celestiaorg/celestia-app/issues/563
In addition, we are continuing the investigation on our side to have an understanding if the root cause might be on our fork and provide more data: https://github.com/celestiaorg/celestia-core/issues/814