cosmos-sdk/store
Emmanuel T Odeke 5399e72f32
perf: store/cachekv: avoid a map lookup if unnecessary, clear maps fast (#10486)
We can shave off some milliseconds, but also cut down some Megabytes of
RAM consumed by only requesting from the cache if needed, but also using
the map clearing idiom which is recognized by the compiler to make fast
code.

Noticed in profiles from Tharsis' Ethermint per https://github.com/tharsis/ethermint/issues/710

- Before
* Memory profiles
```shell
   19.50MB    19.50MB    134:	store.cache = make(map[string]*cValue)
   18.50MB    18.50MB    135:	store.deleted = make(map[string]struct{})
   15.50MB    15.50MB    136:	store.unsortedCache = make(map[string]struct{})
```

* CPU profiles
```go
         .          .    118:	// TODO: Consider allowing usage of Batch, which would allow the write to
         .          .    119:	// at least happen atomically.
     150ms      150ms    120:	for _, key := range keys {
     220ms      3.64s    121:		cacheValue := store.cache[key]
         .          .    122:
         .          .    123:		switch {
         .      250ms    124:		case store.isDeleted(key):
         .          .    125:			store.parent.Delete([]byte(key))
     210ms      210ms    126:		case cacheValue.value == nil:
         .          .    127:			// Skip, it already doesn't exist in parent.
         .          .    128:		default:
     240ms     27.94s    129:			store.parent.Set([]byte(key), cacheValue.value)
         .          .    130:		}
         .          .    131:	}

...

      10ms       60ms    134:	store.cache = make(map[string]*cValue)
         .       40ms    135:	store.deleted = make(map[string]struct{})
         .       50ms    136:	store.unsortedCache = make(map[string]struct{})
         .      110ms    137:	store.sortedCache = dbm.NewMemDB()
```

- After
* Memory profiles
```shell
         .          .    130:	// Clear the cache using the map clearing idiom
         .          .    131:	// and not allocating fresh objects.
         .          .    132:	// Please see https://bencher.orijtech.com/perfclinic/mapclearing/
         .          .    133:	for key := range store.cache {
         .          .    134:		delete(store.cache, key)
         .          .    135:	}
         .          .    136:	for key := range store.deleted {
         .          .    137:		delete(store.deleted, key)
         .          .    138:	}
         .          .    139:	for key := range store.unsortedCache {
         .          .    140:		delete(store.unsortedCache, key)
         .          .    141:	}
```

* CPU profiles
```shell
         .          .    111:	// TODO: Consider allowing usage of Batch, which would allow the write to
         .          .    112:	// at least happen atomically.
     110ms      110ms    113:	for _, key := range keys {
         .      210ms    114:		if store.isDeleted(key) {
         .          .    115:			// We use []byte(key) instead of conv.UnsafeStrToBytes because we cannot
         .          .    116:			// be sure if the underlying store might do a save with the byteslice or
         .          .    117:			// not. Once we get confirmation that .Delete is guaranteed not to
         .          .    118:			// save the byteslice, then we can assume only a read-only copy is sufficient.
         .          .    119:			store.parent.Delete([]byte(key))
         .          .    120:			continue
         .          .    121:		}
         .          .    122:
      50ms      2.45s    123:		cacheValue := store.cache[key]
     910ms      920ms    124:		if cacheValue.value != nil {
         .          .    125:			// It already exists in the parent, hence delete it.
     120ms     29.56s    126:			store.parent.Set([]byte(key), cacheValue.value)
         .          .    127:		}
         .          .    128:	}
         .          .    129:
         .          .    130:	// Clear the cache using the map clearing idiom
         .          .    131:	// and not allocating fresh objects.
         .          .    132:	// Please see https://bencher.orijtech.com/perfclinic/mapclearing/
         .      210ms    133:	for key := range store.cache {
         .          .    134:		delete(store.cache, key)
         .          .    135:	}
         .       10ms    136:	for key := range store.deleted {
         .          .    137:		delete(store.deleted, key)
         .          .    138:	}
         .      170ms    139:	for key := range store.unsortedCache {
         .          .    140:		delete(store.unsortedCache, key)
         .          .    141:	}
         .      260ms    142:	store.sortedCache = dbm.NewMemDB()
         .       10ms    143:}

```

Fixes #10487
Updates https://github.com/tharsis/ethermint/issues/710
2021-11-08 23:49:13 +00:00
..
cache all: ensure b.ReportAllocs() in all the benchmarks (#8460) 2021-01-27 23:52:08 -08:00
cachekv perf: store/cachekv: avoid a map lookup if unnecessary, clear maps fast (#10486) 2021-11-08 23:49:13 +00:00
cachemulti feat: ADR-038 Part 2: StreamingService interface, file writing implementation, and configuration (#8664) 2021-10-24 21:37:37 +00:00
dbadapter ADR-038 Part 1: WriteListener, listen.KVStore, MultiStore and KVStore updates (#8551) 2021-03-30 16:13:51 -04:00
gaskv perf: Remove more telemetry ops, update docs (#10334) 2021-10-11 08:23:42 +00:00
iavl feat!: support debug trace QueryResult (#9576) 2021-07-08 09:25:40 +00:00
internal store/internal: validate keys before calling ProofsFromMap (#9235) 2021-05-02 15:53:59 -07:00
listenkv codec: Rename codec and marshaler interfaces (#9226) 2021-04-29 10:46:22 +00:00
mem ADR-038 Part 1: WriteListener, listen.KVStore, MultiStore and KVStore updates (#8551) 2021-03-30 16:13:51 -04:00
prefix ADR-038 Part 1: WriteListener, listen.KVStore, MultiStore and KVStore updates (#8551) 2021-03-30 16:13:51 -04:00
rootmulti feat: ADR-040: Implement KV Store with decoupled storage and SMT (#9892) 2021-10-19 11:58:06 +00:00
streaming style: lint go and markdown (#10060) 2021-10-30 13:43:04 +00:00
tracekv ADR-038 Part 1: WriteListener, listen.KVStore, MultiStore and KVStore updates (#8551) 2021-03-30 16:13:51 -04:00
transient Merge PR #7265: Tendermint Block Pruning 2020-09-14 10:12:49 -04:00
types chore: Add "Since:" on proto doc comments (#10434) 2021-10-27 14:13:35 +00:00
v2 style: lint go and markdown (#10060) 2021-10-30 13:43:04 +00:00
README.md chore: add markdownlint to lint commands (#9353) 2021-05-27 15:31:04 +00:00
firstlast.go types: add kv type (#6897) 2020-07-30 14:53:02 +00:00
reexport.go Merge PR #6475: Pruning Refactor 2020-06-22 16:31:33 -04:00
store.go Merge PR #6475: Pruning Refactor 2020-06-22 16:31:33 -04:00

README.md

Store

CacheKV

cachekv.Store is a wrapper KVStore which provides buffered writing / cached reading functionalities over the underlying KVStore.

type Store struct {
    cache map[string]cValue
    parent types.KVStore
}

Get

Store.Get() checks Store.cache first in order to find if there is any cached value associated with the key. If the value exists, the function returns it. If not, the function calls Store.parent.Get(), sets the key-value pair to the Store.cache, and returns it.

Set

Store.Set() sets the key-value pair to the Store.cache. cValue has the field dirty bool which indicates whether the cached value is different from the underlying value. When Store.Set() cache new pair, the cValue.dirty is set true so when Store.Write() is called it can be written to the underlying store.

Iterator

Store.Iterator() have to traverse on both caches items and the original items. In Store.iterator(), two iterators are generated for each of them, and merged. memIterator is essentially a slice of the KVPairs, used for cached items. mergeIterator is a combination of two iterators, where traverse happens ordered on both iterators.

CacheMulti

cachemulti.Store is a wrapper MultiStore which provides buffered writing / cached reading functionalities over the underlying MutliStore

type Store struct {
    db types.CacheKVStore
    stores map[types.StoreKey] types.CacheWrap
}

cachemulti.Store branches all substores in its constructor and hold them in Store.stores. Store.GetKVStore() returns the store from Store.stores, and Store.Write() recursively calls CacheWrap.Write() on the substores.

DBAdapter

dbadapter.Store is a adapter for dbm.DB making it fulfilling the KVStore interface.

type Store struct {
    dbm.DB
}

dbadapter.Store embeds dbm.DB, so most of the KVStore interface functions are implemented. The other functions(mostly miscellaneous) are manually implemented.

IAVL

iavl.Store is a base-layer self-balancing merkle tree. It is guaranteed that

  1. Get & set operations are O(log n), where n is the number of elements in the tree
  2. Iteration efficiently returns the sorted elements within the range
  3. Each tree version is immutable and can be retrieved even after a commit(depending on the pruning settings)

Specification and implementation of IAVL tree can be found in [https://github.com/tendermint/iavl].

GasKV

gaskv.Store is a wrapper KVStore which provides gas consuming functionalities over the underlying KVStore.

type Store struct {
    gasMeter types.GasMeter
    gasConfig types.GasConfig
    parent types.KVStore
}

When each KVStore methods are called, gaskv.Store automatically consumes appropriate amount of gas depending on the Store.gasConfig.

Prefix

prefix.Store is a wrapper KVStore which provides automatic key-prefixing functionalities over the underlying KVStore.

type Store struct {
    parent types.KVStore
    prefix []byte
}

When Store.{Get, Set}() is called, the store forwards the call to its parent, with the key prefixed with the Store.prefix.

When Store.Iterator() is called, it does not simply prefix the Store.prefix, since it does not work as intended. In that case, some of the elements are traversed even they are not starting with the prefix.

RootMulti

rootmulti.Store is a base-layer MultiStore where multiple KVStore can be mounted on it and retrieved via object-capability keys. The keys are memory addresses, so it is impossible to forge the key unless an object is a valid owner(or a receiver) of the key, according to the object capability principles.

TraceKV

tracekv.Store is a wrapper KVStore which provides operation tracing functionalities over the underlying KVStore.

type Store struct {
    parent types.KVStore
    writer io.Writer
    context types.TraceContext
}

When each KVStore methods are called, tracekv.Store automatically logs traceOperation to the Store.writer.

type traceOperation struct {
    Operation operation
    Key string
    Value string
    Metadata map[string]interface{}
}

traceOperation.Metadata is filled with Store.context when it is not nil. TraceContext is a map[string]interface{}.

Transient

transient.Store is a base-layer KVStore which is automatically discarded at the end of the block.

type Store struct {
    dbadapter.Store
}

Store.Store is a dbadapter.Store with a dbm.NewMemDB(). All KVStore methods are reused. When Store.Commit() is called, new dbadapter.Store is assigned, discarding previous reference and making it garbage collected.