 Documentation
      ¶
      Documentation
      ¶
    
    
  
    
  
    Index ¶
- Constants
- func InitAll(metrics module.CacheMetrics, db *badger.DB) *storage.All
- func InitPublic(opts badger.Options) (*badger.DB, error)
- func InitSecret(opts badger.Options) (*badger.DB, error)
- type Batch
- type BatchBuilder
- type Cache
- type ChunksQueue
- type Cleaner
- type ClusterBlocks
- type ClusterPayloads
- type Headers
- func (h *Headers) BlockIDByHeight(height uint64) (flow.Identifier, error)
- func (h *Headers) ByBlockID(blockID flow.Identifier) (*flow.Header, error)
- func (h *Headers) ByHeight(height uint64) (*flow.Header, error)
- func (h *Headers) ByParentID(parentID flow.Identifier) ([]*flow.Header, error)
- func (h *Headers) Exists(blockID flow.Identifier) (bool, error)
- func (h *Headers) FindHeaders(filter func(header *flow.Header) bool) ([]flow.Header, error)
- func (h *Headers) RollbackExecutedBlock(header *flow.Header) error
- func (h *Headers) Store(header *flow.Header) error
 
- type RecoverablePrivateBeaconKeyStateMachine
- func (ds *RecoverablePrivateBeaconKeyStateMachine) CommitMyBeaconPrivateKey(epochCounter uint64, commit *flow.EpochCommit) error
- func (ds *RecoverablePrivateBeaconKeyStateMachine) GetDKGState(epochCounter uint64) (flow.DKGState, error)
- func (ds *RecoverablePrivateBeaconKeyStateMachine) InsertMyBeaconPrivateKey(epochCounter uint64, key crypto.PrivateKey) error
- func (ds *RecoverablePrivateBeaconKeyStateMachine) IsDKGStarted(epochCounter uint64) (bool, error)
- func (ds *RecoverablePrivateBeaconKeyStateMachine) RetrieveMyBeaconPrivateKey(epochCounter uint64) (key crypto.PrivateKey, safe bool, err error)
- func (ds *RecoverablePrivateBeaconKeyStateMachine) SetDKGState(epochCounter uint64, newState flow.DKGState) error
- func (ds *RecoverablePrivateBeaconKeyStateMachine) UnsafeRetrieveMyBeaconPrivateKey(epochCounter uint64) (crypto.PrivateKey, error)
- func (ds *RecoverablePrivateBeaconKeyStateMachine) UpsertMyBeaconPrivateKey(epochCounter uint64, key crypto.PrivateKey, commit *flow.EpochCommit) error
 
- type ResultApprovals
- func (r *ResultApprovals) ByChunk(resultID flow.Identifier, chunkIndex uint64) (*flow.ResultApproval, error)
- func (r *ResultApprovals) ByID(approvalID flow.Identifier) (*flow.ResultApproval, error)
- func (r *ResultApprovals) Index(resultID flow.Identifier, chunkIndex uint64, approvalID flow.Identifier) error
- func (r *ResultApprovals) Store(approval *flow.ResultApproval) error
 
Constants ¶
const DefaultCacheSize = uint(1000)
    const JobQueueChunksQueue = "JobQueueChunksQueue"
    Variables ¶
This section is empty.
Functions ¶
func InitAll ¶ added in v0.12.0
func InitAll(metrics module.CacheMetrics, db *badger.DB) *storage.All
func InitPublic ¶ added in v0.22.4
func InitPublic(opts badger.Options) (*badger.DB, error)
InitPublic initializes a public database by checking and setting the database type marker. If an existing, inconsistent type marker is set, this method will return an error. Once a database type marker has been set using these methods, the type cannot be changed.
func InitSecret ¶ added in v0.22.4
func InitSecret(opts badger.Options) (*badger.DB, error)
InitSecret initializes a secrets database by checking and setting the database type marker. If an existing, inconsistent type marker is set, this method will return an error. Once a database type marker has been set using these methods, the type cannot be changed.
Types ¶
type Batch ¶ added in v0.15.1
type Batch struct {
	// contains filtered or unexported fields
}
    func NewBatch ¶ added in v0.15.1
func NewBatch(db BatchBuilder) *Batch
type BatchBuilder ¶ added in v0.32.0
type BatchBuilder interface {
	NewWriteBatch() *badger.WriteBatch
}
    type Cache ¶
type Cache[K comparable, V any] struct { // contains filtered or unexported fields }
func (*Cache[K, V]) Get ¶
Get will try to retrieve the resource from cache first, and then from the injected. During normal operations, the following error returns are expected:
- `storage.ErrNotFound` if key is unknown.
func (*Cache[K, V]) Insert ¶ added in v0.16.1
func (c *Cache[K, V]) Insert(key K, resource V)
Insert will add a resource directly to the cache with the given ID
func (*Cache[K, V]) IsCached ¶ added in v0.31.0
IsCached returns true if the key exists in the cache. It DOES NOT check whether the key exists in the underlying data store.
type ChunksQueue ¶ added in v0.15.0
type ChunksQueue struct {
	// contains filtered or unexported fields
}
    ChunksQueue stores a queue of chunk locators that assigned to me to verify. Job consumers can read the locators as job from the queue by index. Chunk locators stored in this queue are unique.
func NewChunkQueue ¶ added in v0.15.0
func NewChunkQueue(db *badger.DB) *ChunksQueue
NewChunkQueue will initialize the underlying badger database of chunk locator queue.
func (*ChunksQueue) AtIndex ¶ added in v0.15.0
func (q *ChunksQueue) AtIndex(index uint64) (*chunks.Locator, error)
AtIndex returns the chunk locator stored at the given index in the queue.
func (*ChunksQueue) Init ¶ added in v0.15.0
func (q *ChunksQueue) Init(defaultIndex uint64) (bool, error)
Init initializes chunk queue's latest index with the given default index.
func (*ChunksQueue) LatestIndex ¶ added in v0.15.0
func (q *ChunksQueue) LatestIndex() (uint64, error)
LatestIndex returns the index of the latest chunk locator stored in the queue.
func (*ChunksQueue) StoreChunkLocator ¶ added in v0.15.0
func (q *ChunksQueue) StoreChunkLocator(locator *chunks.Locator) (bool, error)
StoreChunkLocator stores a new chunk locator that assigned to me to the job queue. A true will be returned, if the locator was new. A false will be returned, if the locator was duplicate.
type Cleaner ¶
Cleaner uses component.ComponentManager to implement module.Startable and module.ReadyDoneAware to run an internal goroutine which run badger value log garbage collection at a semi-regular interval. The Cleaner exists for 2 reasons:
- Run GC frequently enough that each GC is relatively inexpensive
- Avoid GC being synchronized across all nodes. Since in the happy path, all nodes have very similar database load patterns, without intervention they are likely to schedule GC at the same time, which can cause temporary consensus halts.
func NewCleaner ¶
func NewCleaner(log zerolog.Logger, db *badger.DB, metrics module.CleanerMetrics, interval time.Duration) *Cleaner
NewCleaner returns a cleaner that runs the badger value log garbage collection once every `interval` duration if an interval of zero is passed in, we will not run the GC at all.
type ClusterBlocks ¶
type ClusterBlocks struct {
	// contains filtered or unexported fields
}
    ClusterBlocks implements a simple block storage around a badger DB.
func NewClusterBlocks ¶
func NewClusterBlocks(db *badger.DB, chainID flow.ChainID, headers *Headers, payloads *ClusterPayloads) *ClusterBlocks
func (*ClusterBlocks) ByHeight ¶
func (b *ClusterBlocks) ByHeight(height uint64) (*cluster.Block, error)
func (*ClusterBlocks) ByID ¶
func (b *ClusterBlocks) ByID(blockID flow.Identifier) (*cluster.Block, error)
type ClusterPayloads ¶
type ClusterPayloads struct {
	// contains filtered or unexported fields
}
    ClusterPayloads implements storage of block payloads for collection node cluster consensus.
func NewClusterPayloads ¶
func NewClusterPayloads(cacheMetrics module.CacheMetrics, db *badger.DB) *ClusterPayloads
func (*ClusterPayloads) ByBlockID ¶
func (cp *ClusterPayloads) ByBlockID(blockID flow.Identifier) (*cluster.Payload, error)
func (*ClusterPayloads) Store ¶
func (cp *ClusterPayloads) Store(blockID flow.Identifier, payload *cluster.Payload) error
type Headers ¶
type Headers struct {
	// contains filtered or unexported fields
}
    Headers implements a simple read-only header storage around a badger DB.
func NewHeaders ¶
func NewHeaders(collector module.CacheMetrics, db *badger.DB) *Headers
func (*Headers) BlockIDByHeight ¶ added in v0.25.15
func (h *Headers) BlockIDByHeight(height uint64) (flow.Identifier, error)
BlockIDByHeight returns the block ID that is finalized at the given height. It is an optimized version of `ByHeight` that skips retrieving the block. Expected errors during normal operations:
- `storage.ErrNotFound` if no finalized block is known at given height.
func (*Headers) ByParentID ¶
func (*Headers) Exists ¶ added in v0.31.0
func (h *Headers) Exists(blockID flow.Identifier) (bool, error)
Exists returns true if a header with the given ID has been stored. No errors are expected during normal operation.
func (*Headers) FindHeaders ¶
func (*Headers) RollbackExecutedBlock ¶ added in v0.24.5
RollbackExecutedBlock update the executed block header to the given header. only useful for execution node to roll back executed block height
type RecoverablePrivateBeaconKeyStateMachine ¶ added in v0.39.0
type RecoverablePrivateBeaconKeyStateMachine struct {
	// contains filtered or unexported fields
}
    RecoverablePrivateBeaconKeyStateMachine stores state information about in-progress and completed DKGs, including computed keys. Must be instantiated using secrets database. On the happy path, each consensus committee member takes part in the DKG, and after successfully finishing the DKG protocol it obtains a random beacon private key, which is stored in the database along with DKG state flow.DKGStateCompleted. If for any reason the DKG fails, then the private key will be nil and DKG state is set to flow.DKGStateFailure. When the epoch recovery takes place, we need to query the last valid beacon private key for the current replica and also set it for use during the Recovery Epoch, otherwise replicas won't be able to vote for blocks during the Recovery Epoch. CAUTION: This implementation heavily depends on atomic Badger transactions with interleaved reads and writes for correctness.
func NewRecoverableRandomBeaconStateMachine ¶ added in v0.39.0
func NewRecoverableRandomBeaconStateMachine(collector module.CacheMetrics, db *badger.DB, myNodeID flow.Identifier) (*RecoverablePrivateBeaconKeyStateMachine, error)
NewRecoverableRandomBeaconStateMachine returns the RecoverablePrivateBeaconKeyStateMachine implementation backed by Badger DB. No errors are expected during normal operations.
func (*RecoverablePrivateBeaconKeyStateMachine) CommitMyBeaconPrivateKey ¶ added in v0.39.0
func (ds *RecoverablePrivateBeaconKeyStateMachine) CommitMyBeaconPrivateKey(epochCounter uint64, commit *flow.EpochCommit) error
CommitMyBeaconPrivateKey commits the previously inserted random beacon private key for an epoch. Effectively, this method transitions the state machine into the flow.RandomBeaconKeyCommitted state if the current state is flow.DKGStateCompleted. The caller needs to supply the flow.EpochCommit as evidence that the stored key is valid for the specified epoch. Repeated calls for the same epoch are accepted (idempotent operation), if and only if the provided EpochCommit confirms the already committed key. No errors are expected during normal operations.
func (*RecoverablePrivateBeaconKeyStateMachine) GetDKGState ¶ added in v0.39.0
func (ds *RecoverablePrivateBeaconKeyStateMachine) GetDKGState(epochCounter uint64) (flow.DKGState, error)
GetDKGState retrieves the current state of the state machine for the given epoch. If an error is returned, the state is undefined meaning that state machine is in initial state Error returns:
- storage.ErrNotFound - if there is no state stored for given epoch, meaning the state machine is in initial state.
func (*RecoverablePrivateBeaconKeyStateMachine) InsertMyBeaconPrivateKey ¶ added in v0.39.0
func (ds *RecoverablePrivateBeaconKeyStateMachine) InsertMyBeaconPrivateKey(epochCounter uint64, key crypto.PrivateKey) error
InsertMyBeaconPrivateKey stores the random beacon private key for an epoch and transitions the state machine into the flow.DKGStateCompleted state.
CAUTION: these keys are stored before they are validated against the canonical key vector and may not be valid for use in signing. Use storage.SafeBeaconKeys interface to guarantee only keys safe for signing are returned. Error returns:
- storage.ErrAlreadyExists - if there is already a key stored for given epoch.
- storage.InvalidDKGStateTransitionError - if the requested state transition is invalid.
func (*RecoverablePrivateBeaconKeyStateMachine) IsDKGStarted ¶ added in v0.39.0
func (ds *RecoverablePrivateBeaconKeyStateMachine) IsDKGStarted(epochCounter uint64) (bool, error)
IsDKGStarted checks whether the DKG has been started for the given epoch. No errors expected during normal operation.
func (*RecoverablePrivateBeaconKeyStateMachine) RetrieveMyBeaconPrivateKey ¶ added in v0.39.0
func (ds *RecoverablePrivateBeaconKeyStateMachine) RetrieveMyBeaconPrivateKey(epochCounter uint64) (key crypto.PrivateKey, safe bool, err error)
RetrieveMyBeaconPrivateKey retrieves my beacon private key for the given epoch, only if my key has been confirmed valid and safe for use.
Returns:
- (key, true, nil) if the key is present and confirmed valid
- (nil, false, nil) if the key has been marked invalid or unavailable -> no beacon key will ever be available for the epoch in this case
- (nil, false, storage.ErrNotFound) if the DKG has not ended
- (nil, false, error) for any unexpected exception
func (*RecoverablePrivateBeaconKeyStateMachine) SetDKGState ¶ added in v0.39.0
func (ds *RecoverablePrivateBeaconKeyStateMachine) SetDKGState(epochCounter uint64, newState flow.DKGState) error
SetDKGState performs a state transition for the Random Beacon Recoverable State Machine. Some state transitions may not be possible using this method. For instance, we might not be able to enter flow.DKGStateCompleted state directly from flow.DKGStateStarted, even if such transition is valid. The reason for this is that some states require additional data to be processed by the state machine before the transition can be made. For such cases there are dedicated methods that should be used, ex. InsertMyBeaconPrivateKey and UpsertMyBeaconPrivateKey, which allow to store the needed data and perform the transition in one atomic operation. Error returns:
- storage.InvalidDKGStateTransitionError - if the requested state transition is invalid.
func (*RecoverablePrivateBeaconKeyStateMachine) UnsafeRetrieveMyBeaconPrivateKey ¶ added in v0.39.0
func (ds *RecoverablePrivateBeaconKeyStateMachine) UnsafeRetrieveMyBeaconPrivateKey(epochCounter uint64) (crypto.PrivateKey, error)
UnsafeRetrieveMyBeaconPrivateKey retrieves the random beacon private key for an epoch.
CAUTION: these keys were stored before they are validated against the canonical key vector and may not be valid for use in signing. Use storage.SafeBeaconKeys interface to guarantee only keys safe for signing are returned Error returns:
- storage.ErrNotFound - if there is no key stored for given epoch.
func (*RecoverablePrivateBeaconKeyStateMachine) UpsertMyBeaconPrivateKey ¶ added in v0.39.0
func (ds *RecoverablePrivateBeaconKeyStateMachine) UpsertMyBeaconPrivateKey(epochCounter uint64, key crypto.PrivateKey, commit *flow.EpochCommit) error
UpsertMyBeaconPrivateKey overwrites the random beacon private key for the epoch that recovers the protocol from Epoch Fallback Mode. The resulting state of this method call is flow.RandomBeaconKeyCommitted. State transitions are allowed if and only if the current state is not equal to flow.RandomBeaconKeyCommitted. Repeated calls for the same epoch are idempotent, if and only if the provided EpochCommit confirms the already committed key (error otherwise). No errors are expected during normal operations.
type ResultApprovals ¶ added in v0.14.0
type ResultApprovals struct {
	// contains filtered or unexported fields
}
    ResultApprovals implements persistent storage for result approvals.
func NewResultApprovals ¶ added in v0.14.0
func NewResultApprovals(collector module.CacheMetrics, db *badger.DB) *ResultApprovals
func (*ResultApprovals) ByChunk ¶ added in v0.14.0
func (r *ResultApprovals) ByChunk(resultID flow.Identifier, chunkIndex uint64) (*flow.ResultApproval, error)
ByChunk retrieves a ResultApproval by result ID and chunk index. The ResultApprovals store is only used within a verification node, where it is assumed that there is never more than one approval per chunk.
func (*ResultApprovals) ByID ¶ added in v0.14.0
func (r *ResultApprovals) ByID(approvalID flow.Identifier) (*flow.ResultApproval, error)
ByID retrieves a ResultApproval by its ID
func (*ResultApprovals) Index ¶ added in v0.14.0
func (r *ResultApprovals) Index(resultID flow.Identifier, chunkIndex uint64, approvalID flow.Identifier) error
Index indexes a ResultApproval by chunk (ResultID + chunk index). operation is idempotent (repeated calls with the same value are equivalent to just calling the method once; still the method succeeds on each call).
func (*ResultApprovals) Store ¶ added in v0.14.0
func (r *ResultApprovals) Store(approval *flow.ResultApproval) error
Store stores a ResultApproval