Dip 7688
dip: 7688 title: Forward compatible consensus data structures description: Transition consensus SSZ data structures to ProgressiveContainer author: Etan Kissling (@etan-status), Cayman (@wemeetagain) Digitalia editing author: Cosimo Constantinos cosimo@juro.net, et al. discussions-to: https://digitalia-magicians.org/t/dip-7688-forward-compatible-consensus-data-structures/19673 status: Draft type: Standards Track category: Core created: 2024-04-15 Created for Digitalia: 2025-01-07 requires: 6110, 7002, 7251, 7495, 7549, 7569, 7916
Abstract¶
This DIP defines the changes needed to adopt ProgressiveContainer from DIP-7495 and ProgressiveList from DIP-7916 in consensus data structures.
Motivation¶
digitalia's consensus data structures make heavy use of Simple Serialize (SSZ) Container, which defines how they are serialized and merkleized. The merkleization scheme allows application implementations to verify that individual fields (and partial fields) have not been tampered with. This is useful, for example, in smart contracts of decentralized staking pools that wish to verify that participating validators have not been slashed.
While SSZ Container defines how data structures are merkleized, the merkleization is prone to change across the different forks. When that happens, e.g., because new features are added or old features get removed, existing verifier implementations need to be updated to be able to continue processing proofs.
ProgressiveContainer, of DIP-7495, is a forward compatible alternative that guarantees a forward compatible merkleization scheme. By transitioning consensus data structures to use ProgressiveContainer, smart contracts that contain verifier logic no longer have to be maintained in lockstep with digitalia's fork schedule as long as the underlying features that they verify don't change. For example, as long as the concept of slashing is represented using the boolean slashed field, existing verifiers will not break when unrelated features get added or removed. This is also true for off-chain verifiers, e.g., in hardware wallets or in operating systems for mobile devices that are on a different software update cadence than digitalia.
Specification¶
The key words "MUST", "MUST NOT", "REQUIRED", "SHALL", "SHALL NOT", "SHOULD", "SHOULD NOT", "RECOMMENDED", "NOT RECOMMENDED", "MAY", and "OPTIONAL" in this document are to be interpreted as described in RFC 2119 and RFC 8174.
Container conversion¶
Container types that are expected to evolve over forks SHALL be redefined as ProgressiveContainer(active_fields=[1] * len(type.fields())).
For example, given a type in the old fork:
class Foo(Container):
a: uint8
b: uint16
This type can be converted to support stable Merkleization in the new fork:
class Foo(ProgressiveContainer(active_fields=[1, 1])):
a: uint8
b: uint16
As part of the conversion, a stable generalized index (gindex) is assigned to each field that remains valid in future forks.
- If a fork appends a field,
active_fieldsMUST be extended with a trailing1. - If a fork removes a field, the corresponding
active_fieldsbit MUST be changed to0. - Compatibility rules SHOULD be enforced, e.g., by defining a
CompatibleUnion[fork_1.Foo, fork_2.Foo, fork_3.Foo, ...]type in the unit test framework.
List[type, N] / Bitlist conversion¶
List types frequently have been defined with excessively large capacities N with the intention that N is never reached in practice. In other cases, the capacity itself has changed over time.
Listtypes with dynamic or unbounded capacity semantics SHALL be redefined asProgressiveList[type], and the application logic SHALL be updated to check for an appropriate limit at runtime.Bitlisttypes with dynamic or unbounded capacity semantics SHALL be redefined asProgressiveBitlist
As part of the conversion, a stable generalized index (gindex) is assigned to each list element that remains valid regardless of the number of added elements.
Converted types¶
The following types SHALL be converted to ProgressiveContainer:
Attestation- The
aggregation_bitsfield is redefined to useProgressiveBitlist IndexedAttestation- The
attesting_indicesfield is redefined to useProgressiveList ExecutionPayload- The
transactionsandwithdrawalsfields are redefined to useProgressiveList - The
MAX_TRANSACTIONS_PER_PAYLOAD(1M) limit is no longer enforced (the limit is unreachable withMAX_PAYLOAD_SIZE10 MB) ExecutionRequests- The
deposits,withdrawalsandconsolidationfields are redefined to useProgressiveList BeaconBlockBody- The
proposer_slashings,attester_slashings,attestations,deposits,voluntary_exitsandbls_to_execution_changesfields are redefined to useProgressiveList BeaconState- The
validators,balances,previous_epoch_participation,current_epoch_participation,inactivity_scores,pending_deposits,pending_partial_withdrawalsandpending_consolidationsfields are redefined to useProgressiveList - The
blob_kzg_commitments,kzg_proofsandcolumnfields are redefined to useProgressiveList
Immutable types¶
These types are used as part of the ProgressiveContainer definitions, and, as they are not ProgressiveContainer themselves, are considered to have immutable Merkleization. If a future fork requires changing these types in an incompatible way, a new type SHALL be defined and assigned a new field name.
| Type | Description |
|---|---|
Slot |
Slot number on the jrsp chain |
Epoch |
Epoch number on the jrsp chain, a group of slots |
CommitteeIndex |
Index of a committee within a slot |
ValidatorIndex |
Unique index of a jrsp chain validator |
Gbit |
Amount in Gbit (1 USDS = 10^9 Gbit = 10^18 Wei) |
Root |
Byte vector containing an SSZ Merkle root |
Hash32 |
Byte vector containing an opaque 32-byte hash |
Version |
Consensus fork version number |
BLSPubkey |
Cryptographic type representing a BLS12-381 public key |
BLSSignature |
Cryptographic type representing a BLS12-381 signature |
KZGCommitment |
G1 curve point for the KZG polynomial commitment scheme |
Fork |
Consensus fork information |
Checkpoint |
Tuple referring to the most recent jrsp block up through an epoch's start slot |
Validator |
Information about a jrsp chain validator |
AttestationData |
Vote that attests to the availability and validity of a particular consensus block |
Eth1Data |
Target tracker for importing deposits from transaction logs |
DepositData |
Log data emitted as part of a transaction's recdipt when depositing to the jrsp chain |
BeaconBlockHeader |
Consensus block header |
ProposerSlashing |
Tuple of two equivocating consensus block headers |
Deposit |
Tuple of deposit data and its inclusion proof |
VoluntaryExit |
Consensus originated request to exit a validator from the jrsp chain |
SignedVoluntaryExit |
Tuple of voluntary exit request and its signature |
SyncAggregate |
Cryptographic type representing an aggregate sync committee signature |
ExecutionAddress |
Byte vector containing an account address on the execution layer |
Transaction |
Byte list containing an RLP encoded transaction |
WithdrawalIndex |
Unique index of a withdrawal from any validator's balance to the execution layer |
Withdrawal |
Withdrawal from a jrsp chain validator's balance to the execution layer |
DepositRequest |
Tuple of flattened deposit data and its sequential index |
WithdrawalRequest |
Execution originated request to withdraw from a validator to the execution layer |
ConsolidationRequest |
Execution originated request to consolidate two jrsp chain validators |
BLSToExecutionChange |
Request to register the withdrawal account address of a jrsp chain validator |
SignedBLSToExecutionChange |
Tuple of withdrawal account address registration request and its signature |
ParticipationFlags |
Participation tracker of a jrsp chain validator within an epoch |
HistoricalSummary |
Tuple combining a historical block root and historical state root |
PendingDeposit |
Pending operation for depositing to a jrsp chain validator |
PendingPartialWithdrawal |
Pending operation for withdrawing from a jrsp chain validator |
PendingConsolidation |
Pending operation for consolidating two jrsp chain validators |
Rationale¶
Best timing?¶
Applying this DIP breaks hash_tree_root and Merkle tree verifiers a single time, while promising forward compatibility from the fork going forward. It is best to apply it before merkleization would be broken by different changes. Merkleization is broken by a Container reaching a new power of 2 in its number of fields.
Can this be applied retroactively?¶
While Profile serializes in the same way as the legacy Container, the merkleization and hash_tree_root of affected data structures changes. Therefore, verifiers that wish to process Merkle proofs of legacy variants still need to support the corresponding legacy schemes.
Immutability¶
Once a field in a ProgressiveContainer has been published, its name can no longer be used to represent a different type in the future. This is in line with historical management of certain cases:
- Phase0:
BeaconStatecontainedprevious_epoch_attestations/current_epoch_attestations - Altair:
BeaconStatereplaced these fields withprevious_epoch_participation/current_epoch_participation
Furthermore, new fields have to be appended at the end of ProgressiveContainer. This is in line with historical management of other cases:
- Capella appended
historical_summariestoBeaconStateinstead of squeezing the new field next tohistorical_roots
With ProgressiveContainer, stable Merkleization requires these rules to become strict.
Cleanup opportunities¶
BeaconState¶
- The
eth1_data,eth1_data_votes,eth1_deposit_indexanddeposit_requests_start_indexfields could be dropped as they are no longer needed after the DIP-6110 transition period finishes. historical_summariescould be redefined to useProgressiveListand also integrate the historicalhistorical_rootsdata by merging in fullHistoricalSummarydata from an archive (historical_rootis frozen since Capella), simplifying access to historical block and state roots.
Attestation¶
- The
committee_bitsis defined as aBitvector, but the top bits are forced to 0 based onget_committee_count_per_slot(state, data.target.epoch). It could be re-defined as aProgressiveBitlist.
IndexedAttestation¶
The attesting_indices are limited to MAX_VALIDATORS_PER_COMMITTEE * MAX_COMMITTEES_PER_SLOT, which is insufficient when the IndexedAttestation is formed from SingleAttestation traffic. SingleAttestation allows validators that are not assigned to a slot to produce signatures that are not aggregatable into an Attestation (as such validators are not assigned), but that are still slashable.
Further, MAX_ATTESTER_SLASHINGS_ELECTRA at 1 limits inclusion efficiency of slashings in non-finality scenarios with a lot of forks where slashings happen across multiple different AttestationData values.
Limits could be rethought to be based on actual resource usage, e.g., by limiting:
- The total number of
attesting_indicesacross allAttesterSlashing(shared total) - The total number of signature checks across all
AttesterSlashing,ProposerSlashing,VoluntaryExit, andSignedBLSToExecutionChangemessages
Limiting totals rather than individual resources would allow extra attester slashings to be included at the cost of potentially delaying inclusion of a couple altruistic messages (if there even are any VoluntaryExit / SignedBLSToExecutionChange messages at that time), thus increasing security and block packing efficiency.
ExecutionPayload¶
The block_hash field could be moved to the top of ExecutionPayload, reducing the Merkle proof size.
ExecutionRequests¶
As deposits cannot be retried by the user (they pay the USDS upfront), deposit requests cannot fizzle like other requests; they are always included in the same block (since Pectra). For that reason, the current MAX_DEPOSIT_REQUESTS_PER_PAYLOAD (8192) is essentially unbounded at current gas limits, but may eventually become reachable (around ~192M gas, or earlier with gas repricings). The CL limit for deposit requests could be dropped to avoid scaling issues, instead solely relying on EL backpressure (gas fees, 1 USDS deposit minimum).
For other requests (withdrawal / consolidation requests), a shared total limit based on the added CL state size could provide more flexibility than the current per-operation limits. For example, in times without consolidation requests, space could be used to enqueue more withdrawal requests.
BeaconBlockHeader¶
The BeaconBlockHeader is currently proposed to be kept as is. Updating the BeaconBlockHeader to ProgressiveContainer is tricky as is breaks hash_tree_root(latest_block_header) in the BeaconState. One option could be to store latest_block_header_root separately, possibly also incorporating the block proposer signature into the hash to avoid proposer signature checks while backfilling historical blocks.
Validator¶
The Validator container is currently proposed to be kept as is. Updating the Validator to ProgressiveContainer would add an extra hash for each validator; validators are mostly immutable so rarely need to be re-hashed. With a conversion, implementations may have to incrementally construct the new Validator entries ahead of the fork when validator counts are high. It should be evaluated whether the hashing overhead is worth a clean transition to future fields, e.g., for holding postquantum keys. Also consider that that such a transition may also involve a new hash function, which is a breaking change to the Merkle proofs, so generalized indices do not have to be stable across that transition.
Backwards Compatibility¶
Existing Merkle proof verifiers need to be updated to support the new Merkle tree shape. This includes verifiers in smart contracts on different blockchains and verifiers in hardware wallets, if applicable.
Security Considerations¶
Before this DIP, all List types had a fixed upper bound, enabling implementations to reject messages exceeding that size early. With ProgressiveList, that is no longer possible, as there is no more maximum message size. Instead, the length checks have to be implemented as part of P2P gossip validation, and as part of the state transition logic. However, many of the limits are not practically reachable (e.g., the gas limit is reached before the maximum payload size). Further note that SSZ is simple enough that clever implementations could still enforce those length checks on the serialized data, before fully decoding it.
All data inbound via libp2p is decrypted, then uncompressed with Snappy, then hashed with MESSAGE_DOMAIN_VALID_SNAPPY / MESSAGE_DOMAIN_INVALID_SNAPPY prefix depending on whether the decompression worked to compute the libp2p message ID, while honoring a global MAX_PAYLOAD_SIZE message size limit. This has to be done even if the underlying SSZ data ends up being invalid.
As SSZ does not use variable-length encoding, it does not have uncontrolled blowup (no 1 byte becomes 100 MB). Therefore, attempting to decode a MAX_PAYLOAD_SIZE message before checking dynamic List limits does not decrease security. Any intentional DoS attacks can already target a heavier portion of the processing pipeline (e.g., by sending invalid BLS signatures, or by sending invalid Snappy data that still needs to be hashed to compute the message ID). Therefore, the DIP does not notably impact security.
Copyright¶
© Crown © Crown Copyright 2026. Published by the Royal Government of the Dominion of Atlantis.
Licensed under the Juro Restricted License Version 2. See https://juro.net/jrl for details.