Skip to content

Polkadot validator misses attestations with no error logs until a restart #10504

@kukabi

Description

@kukabi

I set the session keys of Polkadot validator 123kFHVth2udmM79sn3RPQ81HukrQWCxA1vmTWkGHSvkR4k1 in session 12093 from one active node to another running node with healthy status, syncing, with the --validator flag, behind a well-configured network that had been tested before. The node instance was supposed to take over in session 12095, yet it started missing attestations at the beginning of the session, with no indication of an error in the logs. Pre-restart logs, after the beginning of the session, can be founds here: pre_restart.log

I restarted the node after noticing the missed attestations. It started submitting attestations after the restart, giving a number of error logs similar to the following:

2025-12-02 15:50:24 Cluster has too many pending statements, something wrong with our connection to our group peers
				Restart might be needed if validator gets 0 backing rewards for more than 3-4 consecutive sessions pending_statements={ValidatorIndex(529): {(ValidatorIndex(529), CompactStatement::Seconded(0x5471da2120a51e38fd7fb652d24305499396722a21bad6cb85c685f38938651c)), (ValidatorIndex(529), CompactStatement::Seconded(0xdfecdd79a1cc9fcafc464d40f805d879aa30e891e4a54f66a36dbdff5895bb7f))}, ValidatorIndex(527): {(ValidatorIndex(529), CompactStatement::Seconded(0xdfecdd79a1cc9fcafc464d40f805d879aa30e891e4a54f66a36dbdff5895bb7f)), (ValidatorIndex(529), CompactStatement::Seconded(0x5471da2120a51e38fd7fb652d24305499396722a21bad6cb85c685f38938651c))}, ValidatorIndex(528): {(ValidatorIndex(529), CompactStatement::Seconded(0xdfecdd79a1cc9fcafc464d40f805d879aa30e891e4a54f66a36dbdff5895bb7f)), (ValidatorIndex(529), CompactStatement::Seconded(0x5471da2120a51e38fd7fb652d24305499396722a21bad6cb85c685f38938651c))}, ValidatorIndex(525): {(ValidatorIndex(529), CompactStatement::Seconded(0xdfecdd79a1cc9fcafc464d40f805d879aa30e891e4a54f66a36dbdff5895bb7f)), (ValidatorIndex(529), CompactStatement::Seconded(0x5471da2120a51e38fd7fb652d24305499396722a21bad6cb85c685f38938651c))}, ValidatorIndex(526): {(ValidatorIndex(529), CompactStatement::Seconded(0x5471da2120a51e38fd7fb652d24305499396722a21bad6cb85c685f38938651c)), (ValidatorIndex(529), CompactStatement::Seconded(0xdfecdd79a1cc9fcafc464d40f805d879aa30e891e4a54f66a36dbdff5895bb7f))}} parent_hash=0x2228e9b450570432932cf327ab2914a6bca89ccf27de2b1e3aaf33ed85a44ced

No other measure was taken other than the restart, and this same node instance with the same configuration had been in the active set multiple times prior to this session key rotation. Post-restart logs can be found here: post_restart.log

There seems to be no indication of an error in the pre-restart logs, when the node was already in the active set. I cannot tell whether the node was aware of the issue pre-restart, but identifying the root cause would be possible if the logs had an indication/explanation for the condition causing the issue.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions