Documentation
¶
Index ¶
- Variables
- func SetIteratorDebug(enabled bool)
- type BackendDB
- type Batch
- func (b *Batch) Close() error
- func (b *Batch) Delete(key []byte) error
- func (b *Batch) DeleteView(key []byte) error
- func (b *Batch) GetByteSize() (int, error)
- func (b *Batch) Replay(fn func(batch.Entry) error) error
- func (b *Batch) Reset()
- func (b *Batch) Set(key, value []byte) error
- func (b *Batch) SetOps(ops []batch.Entry) error
- func (b *Batch) SetView(key, value []byte) error
- func (b *Batch) Write() error
- func (b *Batch) WriteSync() error
- type DB
- func (db *DB) AcquireSnapshot() *Snapshot
- func (db *DB) Checkpoint() error
- func (db *DB) Close() error
- func (db *DB) CompactionAssist()
- func (db *DB) Delete(key []byte) error
- func (db *DB) DeleteRange(start, end []byte) error
- func (db *DB) DeleteSync(key []byte) error
- func (db *DB) Drain() error
- func (db *DB) Get(key []byte) ([]byte, error)
- func (db *DB) GetAppend(key, dst []byte) ([]byte, error)
- func (db *DB) GetMany(keys [][]byte) ([][]byte, error)
- func (db *DB) GetManyParallelPlan(keyCount int) (workers int, parallel bool)
- func (db *DB) GetUnsafe(key []byte) ([]byte, error)
- func (db *DB) Has(key []byte) (bool, error)
- func (db *DB) Iterator(start, end []byte) (merging.Iterator, error)
- func (db *DB) MaintenancePhase() MaintenancePhase
- func (db *DB) NewBatch() *Batch
- func (db *DB) NewBatchWithSize(size int) *Batch
- func (db *DB) Print() error
- func (db *DB) QueueBacklogBytes() int64
- func (db *DB) ReadValueLogRecord(ptr page.ValuePtr) ([]byte, error)
- func (db *DB) ReserveValueLogRIDs(count int) (uint64, error)
- func (db *DB) ReverseIterator(start, end []byte) (merging.Iterator, error)
- func (db *DB) Set(key, value []byte) error
- func (db *DB) SetDictStore(store DictStore)
- func (db *DB) SetMaintenancePhase(phase MaintenancePhase)
- func (db *DB) SetSync(key, value []byte) error
- func (db *DB) SetTemplateStore(store template.Store)
- func (db *DB) StartAutoCheckpoint(interval time.Duration, maxWALBytes int64, idleInterval time.Duration)
- func (db *DB) Stats() map[string]string
- func (db *DB) TriggerAutoCheckpoint()
- func (db *DB) TriggerFlush()
- func (db *DB) ValueLogRetainedPaths() []string
- type DictStore
- type MaintenancePhase
- type Options
- type Snapshot
- func (s *Snapshot) Close() error
- func (s *Snapshot) Get(key []byte) ([]byte, error)
- func (s *Snapshot) GetAppend(key, dst []byte) ([]byte, error)
- func (s *Snapshot) GetEntry(key []byte) (node.LeafEntry, error)
- func (s *Snapshot) GetEntryExact(key []byte) (node.LeafEntry, error)
- func (s *Snapshot) GetUnsafe(key []byte) ([]byte, error)
- func (s *Snapshot) Has(key []byte) (bool, error)
- func (s *Snapshot) Pager() *pager.Pager
- func (s *Snapshot) State() *backenddb.DBState
- type VlogAutotuneBenchMode
- type VlogAutotuneBenchRequest
- type VlogAutotuneBenchResult
- type VlogAutotuneBenchSegment
- type VlogAutotuneBenchSegmentResult
Constants ¶
This section is empty.
Variables ¶
var ErrBatchClosed = fmt.Errorf("batch has been written or closed")
var ErrKeyEmpty = fmt.Errorf("key cannot be empty")
var ErrMemtableFull = fmt.Errorf("memtable full")
var ErrUnsafeOptions = fmt.Errorf("unsafe options require AllowUnsafe")
var ErrValueNil = fmt.Errorf("value cannot be nil")
Functions ¶
func SetIteratorDebug ¶
func SetIteratorDebug(enabled bool)
SetIteratorDebug toggles attaching debug metadata to iterators returned by CachingDB.Iterator. It is intended for benchmarking/diagnostics.
Types ¶
type BackendDB ¶
type BackendDB interface {
Get(key []byte) ([]byte, error)
GetUnsafe(key []byte) ([]byte, error)
GetAppend(key, dst []byte) ([]byte, error)
Has(key []byte) (bool, error)
Iterator(start, end []byte) (iterator.UnsafeIterator, error)
ReverseIterator(start, end []byte) (iterator.UnsafeIterator, error)
NewBatch() batch.Interface
Close() error
Print() error
Stats() map[string]string
}
BackendDB defines the subset of treedb.DB needed by CachingDB.
type Batch ¶
type Batch struct {
// contains filtered or unexported fields
}
func (*Batch) DeleteView ¶
DeleteView records a Delete without copying key bytes. Callers must treat key as immutable until the batch is written or closed.
func (*Batch) GetByteSize ¶
func (*Batch) Reset ¶
func (b *Batch) Reset()
Reset clears the batch for reuse without closing it.
This intentionally keeps internal buffers to avoid per-batch allocations in callers that frequently reset (e.g. geth benchmarks).
type DB ¶
type DB struct {
// contains filtered or unexported fields
}
func (*DB) AcquireSnapshot ¶ added in v0.4.0
AcquireSnapshot returns a cached snapshot that includes queued memtable writes.
func (*DB) Checkpoint ¶
Checkpoint forces a durable backend boundary and trims the WAL so long-running cached-mode runs do not accumulate unbounded `wal/` growth.
It blocks writers while it:
- rotates the current mutable memtable (if non-empty),
- rotates to a fresh WAL segment,
- flushes all queued memtables with backend sync,
- forces a backend sync boundary (even if the queue is empty),
- removes all older WAL segments (keeping only the currently-open one).
func (*DB) CompactionAssist ¶
func (db *DB) CompactionAssist()
CompactionAssist performs bounded flush work when backpressure triggers. It is intended to be called by background maintenance (e.g. index compaction) so that flush debt does not grow unbounded in the absence of foreground writes.
func (*DB) DeleteRange ¶
DeleteRange deletes all keys in the range [start, end).
When WAL is disabled and the backend is empty, a full-range delete can be satisfied by clearing the in-memory layers without enumerating keys.
func (*DB) DeleteSync ¶
func (*DB) Drain ¶
Drain flushes all currently buffered writes (mutable + queued memtables) to the backend. It is intended for maintenance operations that require a fully materialized backend state (e.g. index vacuum).
Drain does not provide mutual exclusion against concurrent writers; callers should ensure no writes occur concurrently if they require a fully drained state.
func (*DB) GetAppend ¶
GetAppend appends the value for the key to dst and returns the new slice. If the key is not found, it returns dst and ErrKeyNotFound.
func (*DB) GetMany ¶ added in v0.4.0
GetMany returns safe copies of values for keys.
Missing keys are returned as nil entries with no error.
func (*DB) GetManyParallelPlan ¶ added in v0.4.0
GetManyParallelPlan reports a safe upper-bound scheduling plan for GetMany.
When cache hit-rates are high, backend calls may involve fewer keys than the provided key count. This method intentionally reports based on the input upper-bound to stay conservative for concurrency budgeting at callers.
func (*DB) MaintenancePhase ¶ added in v0.4.0
func (db *DB) MaintenancePhase() MaintenancePhase
func (*DB) NewBatchWithSize ¶
func (*DB) QueueBacklogBytes ¶
QueueBacklogBytes returns the current queued memtable backlog in bytes.
func (*DB) ReadValueLogRecord ¶ added in v0.4.0
ReadValueLogRecord reads a raw value-log record for diagnostics and tooling.
func (*DB) ReserveValueLogRIDs ¶ added in v0.4.0
ReserveValueLogRIDs reserves a contiguous RID range from the cached-mode shared allocator and returns the starting RID.
func (*DB) ReverseIterator ¶
func (*DB) SetDictStore ¶ added in v0.2.0
SetDictStore installs the dictionary store for current-ID freezing.
func (*DB) SetMaintenancePhase ¶ added in v0.4.0
func (db *DB) SetMaintenancePhase(phase MaintenancePhase)
func (*DB) SetTemplateStore ¶ added in v0.3.0
SetTemplateStore installs the template store used for template compression.
func (*DB) StartAutoCheckpoint ¶
func (db *DB) StartAutoCheckpoint(interval time.Duration, maxWALBytes int64, idleInterval time.Duration)
StartAutoCheckpoint enables a background loop that periodically forces a durable boundary and trims cached-mode WAL segments. When idleInterval > 0, it also triggers an opportunistic checkpoint after a period of write-idleness.
interval > 0 enables periodic checkpoints. maxWALBytes is a safety cap: if > 0, the loop will attempt to checkpoint when the effective WAL bytes exceed this cap. maxWALBytes <= 0 disables the size trigger.
This does not make each individual write durable; it bounds the window of unsynced writes for long-running workloads.
func (*DB) TriggerAutoCheckpoint ¶
func (db *DB) TriggerAutoCheckpoint()
TriggerAutoCheckpoint schedules a best-effort immediate auto-checkpoint pass.
func (*DB) TriggerFlush ¶
func (db *DB) TriggerFlush()
TriggerFlush schedules a background flush pass (best-effort).
func (*DB) ValueLogRetainedPaths ¶ added in v0.4.0
ValueLogRetainedPaths returns a best-effort snapshot of retained value-log segment paths currently pinned by cached-mode pointer lifecycle tracking.
type DictStore ¶ added in v0.2.0
type DictStore interface {
GetCurrent(ctx context.Context) (uint64, error)
GetDictBytes(ctx context.Context, dictID uint64) ([]byte, error)
}
DictStore provides access to the current dictionary ID for write freezing.
type MaintenancePhase ¶ added in v0.4.0
type MaintenancePhase uint8
const ( MaintenancePhaseSteady MaintenancePhase = iota MaintenancePhaseRestore MaintenancePhaseCatchUp )
type Options ¶
type Options struct {
FlushThreshold int64
// MemtableMode selects the in-memory write buffer implementation.
// Supported: "skiplist", "hash_sorted", "btree", "append_only", "adaptive".
// Use "adaptive" or "adaptive:<mode>" to switch per-rotation based on workload.
MemtableMode string
// MemtableShards controls the number of mutable memtable shards. Values <= 0
// use a default derived from GOMAXPROCS. The count is rounded down to a power
// of two.
MemtableShards int
// DomainIngressWorkers enables experimental domain-local write ingress queues.
// Values <= 0 keep the legacy direct caller write path.
DomainIngressWorkers int
// DomainIngressQueueSize configures the per-worker ingress queue length.
// Values <= 0 use a default.
DomainIngressQueueSize int
// Legacy backpressure knob: queue length limit.
// 0 uses the default (4). <0 disables writer backpressure entirely.
MaxQueuedMemtables int
// Adaptive backpressure knobs (seconds/bytes). If any of these are non-zero,
// the caching layer uses backlog-bytes thresholds instead of queue length.
SlowdownBacklogSeconds float64
StopBacklogSeconds float64
MaxBacklogBytes int64
// Writer flush assist limits when backpressure triggers.
WriterFlushMaxMemtables int
WriterFlushMaxDuration time.Duration
// FlushBuildConcurrency controls how many goroutines may be used to build a
// combined flush batch from multiple immutable memtables. Values <= 1 disable
// parallelism.
FlushBuildConcurrency int
// FlushBuildMinEntries gates the parallel build path by total entries.
// Values <= 0 use a default of 16k.
FlushBuildMinEntries int
// FlushBuildMinUnits gates the parallel build path by number of queued units.
// Values <= 0 use a default of 2.
FlushBuildMinUnits int
// FlushBuildChunkCap controls the maximum entries per build chunk.
// Values < 0 use the fixed default of 8192, 0 enables adaptive chunk sizing,
// and values > 0 set a fixed cap.
FlushBuildChunkCap int
// FlushBuildChunkTargetBytes controls adaptive chunk sizing (bytes per chunk).
// Values <= 0 use a default of 2MiB.
FlushBuildChunkTargetBytes int
// FlushBuildChunkMinBytes clamps adaptive chunk sizes (minimum bytes).
// Values <= 0 use a default of 1MiB.
FlushBuildChunkMinBytes int
// FlushBuildChunkMaxBytes clamps adaptive chunk sizes (maximum bytes).
// Values <= 0 use a default of 4MiB.
FlushBuildChunkMaxBytes int
// FlushBuildPrefetchUnits controls how many memtables to start building ahead
// of the consumer. Values <= 0 use FlushBuildConcurrency.
FlushBuildPrefetchUnits int
// FlushBackendMaxEntries caps how many operations are buffered into a single
// backend batch before committing it and continuing with a fresh batch.
//
// This increases backend commit cadence during very large flushes, which can
// reduce index.db high-watermark growth under small KeepRecent windows by
// making retired pages eligible for reuse sooner.
//
// 0 uses the internal default. Negative disables chunking (single backend
// commit per flush).
FlushBackendMaxEntries int
// FlushBackendMaxBatches caps how many intermediate backend commits a single
// flush may emit. This bounds zipper/apply overhead when FlushBackendMaxEntries
// is very small relative to the flush size.
//
// 0 uses the internal default. Negative disables the cap.
FlushBackendMaxBatches int
// DisableWAL disables the redo/journal log while keeping the value log enabled.
DisableWAL bool
// JournalLanes controls the number of active commit/value log lanes
// (0=GOMAXPROCS-aware default).
// Max supported lanes is 255; value-log segment sequence per lane is capped at 8,388,607.
JournalLanes int
// WALMaxSegmentBytes caps the size of a single WAL segment payload.
// 0 uses the default limit.
WALMaxSegmentBytes int64
// JournalCompression enables best-effort zstd compression for journal/commitlog
// segments (metadata only). The writer only keeps compressed bytes when they
// are smaller than the raw payload, so compression never causes size
// amplification.
JournalCompression bool
// RelaxedSync disables fsync on Sync operations.
RelaxedSync bool
// ValueLogPointerThreshold controls when WAL/vlog pointers are used.
// Values <= 0 use a default threshold. In relaxed durability modes, the
// default is smaller to avoid catastrophic update-heavy cliffs at large key
// counts by pushing moderate values into the value log.
ValueLogPointerThreshold int
// IndexOuterLeavesInValueLog stores B+Tree leaf pages (the pages containing
// key/value entries) in the persistent value log instead of index.db.
IndexOuterLeavesInValueLog bool
// ValueLogDomainInlineThresholds configures optional per-domain overrides
// for inline-vs-pointer placement. Longest-prefix match wins.
ValueLogDomainInlineThresholds []backenddb.ValueLogDomainThreshold
// ValueLogRawWritevMinAvgBytes controls raw grouped-frame writev usage for
// the value log.
//
// 0 uses adaptive mode (no average-bytes floor); values >0 require average
// payload bytes/record to meet this floor before raw writev is considered.
ValueLogRawWritevMinAvgBytes int
// ValueLogRawWritevMinBatchRecords controls the minimum grouped records before
// raw writev is considered for value-log appends.
//
// Values <=0 use a default of 8.
ValueLogRawWritevMinBatchRecords int
// ValueLogCompression selects value-log compression behavior:
// 0=default(unset; normalized to auto by TreeDB Open), 1=off, 2=block,
// 3=dict, 4=auto.
ValueLogCompression uint8
// ValueLogBlockCodec selects block codec when block compression is enabled:
// 0=snappy, 1=lz4.
ValueLogBlockCodec uint8
// ValueLogBlockTargetCompressedBytes controls block-mode grouped frame K
// adaptation target (0=default).
ValueLogBlockTargetCompressedBytes int
// ValueLogIncompressibleHoldBytes configures auto-mode incompressible hold
// window bytes (0=default).
ValueLogIncompressibleHoldBytes int
// ValueLogIncompressibleProbeBytes configures auto-mode hold probe interval
// bytes (0=default).
ValueLogIncompressibleProbeBytes int
// ValueLogAutoPolicy controls auto-mode dict-vs-block bias:
// 0=balanced, 1=throughput, 2=size.
ValueLogAutoPolicy uint8
// ValueLogMaxSegmentBytes caps the size of a single value-log segment file.
// 0 disables the cap.
//
// This is an internal safety knob used by experimental index encodings
// (e.g. packed on-disk ValuePtr) that require value-log offsets stay within a
// smaller representable range.
ValueLogMaxSegmentBytes int64
// ValueLogGenerationPolicy selects generational mode:
// 0=default(unset; normalized to hot_warm_cold by Open),
// 1=off, 2=hot_warm_cold.
ValueLogGenerationPolicy uint8
// ValueLogGenerationLeafSegmentTargetBytes configures leaf_vlog segment target size.
ValueLogGenerationLeafSegmentTargetBytes int64
// ValueLogGenerationHotSegmentTargetBytes configures hot segment target size.
ValueLogGenerationHotSegmentTargetBytes int64
// ValueLogGenerationWarmSegmentTargetBytes configures warm segment target size.
ValueLogGenerationWarmSegmentTargetBytes int64
// ValueLogGenerationColdSegmentTargetBytes configures cold segment target size.
ValueLogGenerationColdSegmentTargetBytes int64
// ValueLogRewriteBudgetBytesPerSec configures incremental rewrite byte budget.
ValueLogRewriteBudgetBytesPerSec int64
// ValueLogRewriteBudgetRecordsPerSec configures incremental rewrite record budget.
ValueLogRewriteBudgetRecordsPerSec int
// ValueLogRewriteTriggerStaleRatioPPM triggers rewrite by stale/live ratio.
ValueLogRewriteTriggerStaleRatioPPM uint32
// ValueLogRewriteTriggerTotalBytes triggers rewrite by retained bytes.
ValueLogRewriteTriggerTotalBytes int64
// ValueLogRewriteTriggerChurnPerSec triggers rewrite by churn rate.
ValueLogRewriteTriggerChurnPerSec int64
// ValueLogRewriteMinSegmentAge gates online rewrite to source segments that
// are at least this old.
//
// 0 uses the implementation default.
ValueLogRewriteMinSegmentAge time.Duration
// ForceValueLogPointers stores all values out-of-line in the value log.
ForceValueLogPointers bool
// DisableReadChecksum skips CRC verification on value-log reads.
DisableReadChecksum bool
// AllowUnsafe acknowledges unsafe durability options.
// When false, Open will reject DisableWAL or RelaxedSync.
AllowUnsafe bool
// MaxValueLogRetainedBytes emits a warning once retained value-log bytes
// reach or exceed this threshold and also acts as the soft retained-byte
// trigger for background prune scheduling (0 disables both).
MaxValueLogRetainedBytes int64
// MaxValueLogRetainedBytesHard disables value-log pointers for new large
// values once retained bytes exceed this threshold (0 disables the cap).
MaxValueLogRetainedBytesHard int64
// ValueLogDictTrain configures background dictionary training for value-log frame compression.
// TrainBytes <= 0 disables training.
ValueLogDictTrain compression.TrainConfig
// ValueLogDictMaxK clamps the maximum group size (K) used for dict-compressed
// value-log frames. Values <= 0 use the default (32).
ValueLogDictMaxK int
// ValueLogDictClassMode controls dictionary-state partitioning:
// 0=single shared dict stream, 1=split outer-leaf vs single-value streams.
ValueLogDictClassMode uint8
// ValueLogDictFrameEncodeLevel controls the zstd encoder level used for
// dict-compressed value-log frames. Values <= 0 use SpeedFastest.
ValueLogDictFrameEncodeLevel zstd.EncoderLevel
// ValueLogDictFrameEnableEntropy enables entropy coding for dict-compressed
// frames (higher ratio, lower throughput).
ValueLogDictFrameEnableEntropy bool
// ValueLogDictAdaptiveRatio enables adaptive pause of dict compression when payload ratios degrade.
// 0 disables.
ValueLogDictAdaptiveRatio float64
// ValueLogDictMetricsWindowBytes controls the metrics window size (0=default).
ValueLogDictMetricsWindowBytes int
// ValueLogDictMetricsMinRecords is a minimum record count before pausing (0=default).
ValueLogDictMetricsMinRecords int
// ValueLogDictMetricsPauseBytes controls pause duration in bytes (0=default).
ValueLogDictMetricsPauseBytes int
// ValueLogDictIncompressibleHoldBytes enables classifier-driven hold mode for
// high-entropy streams. While hold mode is active, dict compression attempts
// and trainer collection are bypassed until hold bytes are consumed.
//
// 0 uses profile/default hold configuration; <0 explicitly disables hold
// mode and opts out of profile defaults.
ValueLogDictIncompressibleHoldBytes int
// ValueLogDictProbeIntervalBytes controls periodic probe attempts while
// incompressible hold mode is active.
//
// Values <=0 use a default derived from hold bytes.
ValueLogDictProbeIntervalBytes int
// ValueLogDictMinPayloadSavingsRatio rejects newly trained dictionaries whose
// payload ratio does not improve by at least this fraction (0 uses a
// throughput-oriented default: 0.02 normally, 0.05 with force pointers or
// WAL disabled).
ValueLogDictMinPayloadSavingsRatio float64
// ValueLogCompressionAutotune configures the wall-time value-log compression autotuner.
// Cached mode only (value log enabled by default).
ValueLogCompressionAutotune valuelog.AutotuneOptions
// ValueLogTemplateMode controls template-based compression for value-log values.
ValueLogTemplateMode template.Mode
// ValueLogTemplateConfig controls template creation and encoding behavior.
ValueLogTemplateConfig template.Config
// ValueLogTemplateReadStrict controls strict template decode behavior.
ValueLogTemplateReadStrict bool
// NotifyError is an optional hook for background maintenance failures.
NotifyError func(error)
}
type Snapshot ¶ added in v0.4.0
type Snapshot struct {
// contains filtered or unexported fields
}
Snapshot is a consistent point-in-time view of the cached+backend TreeDB state.
Snapshot isolation in cached mode is implemented by rotating any pending writes out of the mutable memtables into the immutable queue, then reading from:
- the frozen queue (stored oldest-to-newest, scanned newest-to-oldest)
- the backend snapshot
Mutable memtables are intentionally ignored so that writes after AcquireSnapshot are not visible through the snapshot.
Snapshot pointers are single-use: after Close returns, callers must discard the pointer and treat further use as invalid.
func (*Snapshot) GetEntryExact ¶ added in v0.4.0
type VlogAutotuneBenchMode ¶ added in v0.2.0
type VlogAutotuneBenchMode string
VlogAutotuneBenchMode controls the deterministic bench mode.
const ( VlogAutotuneBenchOff VlogAutotuneBenchMode = "off" VlogAutotuneBenchNoDictFixed VlogAutotuneBenchMode = "no_dict_fixed" VlogAutotuneBenchDictFixed VlogAutotuneBenchMode = "dict_fixed" VlogAutotuneBenchAutotune VlogAutotuneBenchMode = "autotune" VlogAutotuneBenchTemplate VlogAutotuneBenchMode = "template_fixed" )
type VlogAutotuneBenchRequest ¶ added in v0.2.0
type VlogAutotuneBenchRequest struct {
Mode VlogAutotuneBenchMode
FixedK int
Segments []VlogAutotuneBenchSegment
}
type VlogAutotuneBenchResult ¶ added in v0.2.0
type VlogAutotuneBenchResult struct {
Mode VlogAutotuneBenchMode
Segments []VlogAutotuneBenchSegmentResult
RawBytes uint64
StoredBytes uint64
WallTimeNs int64
ThroughputMB float64
TrainerStats compression.TrainerStats
}
func RunVlogAutotuneBench ¶ added in v0.2.0
func RunVlogAutotuneBench(req VlogAutotuneBenchRequest) (*VlogAutotuneBenchResult, error)
type VlogAutotuneBenchSegment ¶ added in v0.2.0
type VlogAutotuneBenchSegmentResult ¶ added in v0.2.0
type VlogAutotuneBenchSegmentResult struct {
Name string
RawBytes uint64
StoredBytes uint64
WallTimeNs int64
ThroughputRawMBps float64
AttemptedFrac float64
KeptFrac float64
ObservedRatio float64
FramesTotal uint64
FramesAttempted uint64
FramesKept uint64
EncodeNsTotal int64
IoNsTotal int64
State string
DictID uint64
DictHash uint64
HistoryBytes int
K int
PublishOrderingOK bool
TrainerProfileOK bool
TrainerProfileK int
TrainerProfileHash uint64
}
Source Files
¶
- backpressure.go
- bytes_to_string_unsafe.go
- db.go
- debug_memtable_rotate.go
- debug_vlog_maint.go
- debug_vlog_shape.go
- debug_vlog_timings.go
- expvar_stats.go
- lane.go
- leaf_page_log.go
- leafref_live_ids.go
- log_writer.go
- process_memory_linux.go
- snapshot.go
- value_log_iterator.go
- vlog_autotune.go
- vlog_autotune_bench.go
- vlog_autotune_metrics.go
- vlog_compression_selector.go
- vlog_dict.go
- vlog_generation_state.go
- vlog_queue_metrics.go
- vlog_writer_caps.go