-
Notifications
You must be signed in to change notification settings - Fork 1.1k
Description
I set the session keys of Polkadot validator 123kFHVth2udmM79sn3RPQ81HukrQWCxA1vmTWkGHSvkR4k1 in session 12093 from one active node to another running node with healthy status, syncing, with the --validator flag, behind a well-configured network that had been tested before. The node instance was supposed to take over in session 12095, yet it started missing attestations at the beginning of the session, with no indication of an error in the logs. Pre-restart logs, after the beginning of the session, can be founds here: pre_restart.log
I restarted the node after noticing the missed attestations. It started submitting attestations after the restart, giving a number of error logs similar to the following:
2025-12-02 15:50:24 Cluster has too many pending statements, something wrong with our connection to our group peers
Restart might be needed if validator gets 0 backing rewards for more than 3-4 consecutive sessions pending_statements={ValidatorIndex(529): {(ValidatorIndex(529), CompactStatement::Seconded(0x5471da2120a51e38fd7fb652d24305499396722a21bad6cb85c685f38938651c)), (ValidatorIndex(529), CompactStatement::Seconded(0xdfecdd79a1cc9fcafc464d40f805d879aa30e891e4a54f66a36dbdff5895bb7f))}, ValidatorIndex(527): {(ValidatorIndex(529), CompactStatement::Seconded(0xdfecdd79a1cc9fcafc464d40f805d879aa30e891e4a54f66a36dbdff5895bb7f)), (ValidatorIndex(529), CompactStatement::Seconded(0x5471da2120a51e38fd7fb652d24305499396722a21bad6cb85c685f38938651c))}, ValidatorIndex(528): {(ValidatorIndex(529), CompactStatement::Seconded(0xdfecdd79a1cc9fcafc464d40f805d879aa30e891e4a54f66a36dbdff5895bb7f)), (ValidatorIndex(529), CompactStatement::Seconded(0x5471da2120a51e38fd7fb652d24305499396722a21bad6cb85c685f38938651c))}, ValidatorIndex(525): {(ValidatorIndex(529), CompactStatement::Seconded(0xdfecdd79a1cc9fcafc464d40f805d879aa30e891e4a54f66a36dbdff5895bb7f)), (ValidatorIndex(529), CompactStatement::Seconded(0x5471da2120a51e38fd7fb652d24305499396722a21bad6cb85c685f38938651c))}, ValidatorIndex(526): {(ValidatorIndex(529), CompactStatement::Seconded(0x5471da2120a51e38fd7fb652d24305499396722a21bad6cb85c685f38938651c)), (ValidatorIndex(529), CompactStatement::Seconded(0xdfecdd79a1cc9fcafc464d40f805d879aa30e891e4a54f66a36dbdff5895bb7f))}} parent_hash=0x2228e9b450570432932cf327ab2914a6bca89ccf27de2b1e3aaf33ed85a44ced
No other measure was taken other than the restart, and this same node instance with the same configuration had been in the active set multiple times prior to this session key rotation. Post-restart logs can be found here: post_restart.log
There seems to be no indication of an error in the pre-restart logs, when the node was already in the active set. I cannot tell whether the node was aware of the issue pre-restart, but identifying the root cause would be possible if the logs had an indication/explanation for the condition causing the issue.