One of our testnet nodes has seemingly lost connection to all peers and cannot reconnect. As a result, the node was not getting new blocks. It remained in this state for several hours, until we restarted the service and then it was able to catch up normally.
This is the second time it happens to us, just don't remember if the first one was also on testnet or mainnet.
Geth
Version: 5.5.0-mainnet
Architecture: amd64
Go Version: go1.22.4
Operating System: linux
Expected behaviour
Node should be able to find/reconnect to peers and keep syncing
Actual behaviour
Node loses connection to peers and doesn't find/reconnect to any peer until l2geth is restarted
Steps to reproduce the behaviour
N/A
Backtrace
Node was importing new chain segments until it wasn't:
Jun 27 12:35:59 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:35:59.513] Imported new chain segment blocks=1 txs=1 mgas=0.021 elapsed=2.303ms mgasps=9.117 >
Jun 27 12:36:01 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:36:01.668] Imported new chain segment blocks=1 txs=1 mgas=0.021 elapsed=3.030ms mgasps=6.930 >
Jun 27 12:36:04 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:36:04.990] Imported new chain segment blocks=1 txs=1 mgas=0.021 elapsed=2.535ms mgasps=8.282 >
Jun 27 12:36:07 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:36:07.064] Imported new chain segment blocks=1 txs=1 mgas=0.021 elapsed=2.270ms mgasps=9.251 >
Jun 27 12:36:07 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:36:07.811] Deep froze chain segment blocks=21 elapsed=13.866ms number=4,904,702 hash=f83edb..bc5>
Jun 27 12:36:08 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:36:08.596] Looking for peers peercount=0 tried=340 static=0
Jun 27 12:36:18 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:36:18.901] Looking for peers peercount=0 tried=257 static=0
Jun 27 12:36:28 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:36:28.927] Looking for peers peercount=0 tried=411 static=0
Jun 27 12:36:38 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:36:38.936] Looking for peers peercount=0 tried=359 static=0
Jun 27 12:36:48 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:36:48.981] Looking for peers peercount=0 tried=293 static=0
Jun 27 12:36:59 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:36:59.114] Looking for peers peercount=0 tried=359 static=0
Jun 27 12:37:09 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:37:09.127] Looking for peers peercount=0 tried=353 static=0
Jun 27 12:37:19 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:37:19.403] Looking for peers peercount=0 tried=274 static=0
Jun 27 12:37:29 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:37:29.422] Looking for peers peercount=0 tried=352 static=0
Jun 27 12:37:39 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:37:39.440] Looking for peers peercount=0 tried=306 static=0
Jun 27 12:37:49 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:37:49.441] Looking for peers peercount=1 tried=335 static=0
Jun 27 12:37:59 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:37:59.481] Looking for peers peercount=0 tried=263 static=0
Jun 27 12:38:09 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:38:09.534] Looking for peers peercount=0 tried=318 static=0
Jun 27 12:38:19 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:38:19.556] Looking for peers peercount=0 tried=309 static=0
Jun 27 12:38:29 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:38:29.680] Looking for peers peercount=0 tried=341 static=0
Jun 27 12:38:39 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:38:39.681] Looking for peers peercount=0 tried=262 static=0
Jun 27 12:38:49 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:38:49.681] Looking for peers peercount=0 tried=321 static=0
Jun 27 12:38:59 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:38:59.700] Looking for peers peercount=0 tried=344 static=0
Jun 27 12:39:09 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:39:09.701] Looking for peers peercount=0 tried=424 static=0
Jun 27 12:39:10 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:39:10.401] Sync rollup events progress update latestProcessedBlock=6,197,837
Jun 27 12:39:19 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:39:19.705] Looking for peers peercount=0 tried=216 static=0
Jun 27 12:39:29 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:39:29.938] Looking for peers peercount=0 tried=322 static=0
Jun 27 12:39:39 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:39:39.938] Looking for peers peercount=0 tried=226 static=0
Jun 27 12:39:49 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:39:49.975] Looking for peers peercount=0 tried=238 static=0
Jun 27 12:39:59 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:39:59.976] Looking for peers peercount=0 tried=262 static=0
Jun 27 12:40:09 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:40:09.989] Looking for peers peercount=0 tried=313 static=0
Jun 27 12:40:20 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:40:20.147] Looking for peers peercount=0 tried=360 static=0
Jun 27 12:40:30 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:40:30.151] Looking for peers peercount=0 tried=315 static=0
Jun 27 12:40:40 t-scroll-sepolia-01 l2geth[3256645]: INFO [06-27|12:40:40.168] Looking for peers peercount=0 tried=192 static=0
System information
One of our testnet nodes has seemingly lost connection to all peers and cannot reconnect. As a result, the node was not getting new blocks. It remained in this state for several hours, until we restarted the service and then it was able to catch up normally. This is the second time it happens to us, just don't remember if the first one was also on testnet or mainnet.
Expected behaviour
Node should be able to find/reconnect to peers and keep syncing
Actual behaviour
Node loses connection to peers and doesn't find/reconnect to any peer until
l2geth
is restartedSteps to reproduce the behaviour
N/A
Backtrace
Node was importing new chain segments until it wasn't: