/* * Copyright 2017 Dgraph Labs, Inc. and Contributors * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /* Adapted from RocksDB inline skiplist. Key differences: - No optimization for sequential inserts (no "prev"). - No custom comparator. - Support overwrites. This requires care when we see the same key when inserting. For RocksDB or LevelDB, overwrites are implemented as a newer sequence number in the key, so there is no need for values. We don't intend to support versioning. In-place updates of values would be more efficient. - We discard all non-concurrent code. - We do not support Splices. This simplifies the code a lot. - No AllocateNode or other pointer arithmetic. - We combine the findLessThan, findGreaterOrEqual, etc into one function. */ package skl import ( "math" "math/rand" "sync/atomic" "unsafe" "github.com/dgraph-io/badger/y" ) const ( maxHeight = 20 heightIncrease = math.MaxUint32 / 3 ) // MaxNodeSize is the memory footprint of a node of maximum height. const MaxNodeSize = int(unsafe.Sizeof(node{})) type node struct { // Multiple parts of the value are encoded as a single uint64 so that it // can be atomically loaded and stored: // value offset: uint32 (bits 0-31) // value size : uint16 (bits 32-47) value uint64 // A byte slice is 24 bytes. We are trying to save space here. keyOffset uint32 // Immutable. No need to lock to access key. keySize uint16 // Immutable. No need to lock to access key. // Height of the tower. height uint16 // Most nodes do not need to use the full height of the tower, since the // probability of each successive level decreases exponentially. Because // these elements are never accessed, they do not need to be allocated. // Therefore, when a node is allocated in the arena, its memory footprint // is deliberately truncated to not include unneeded tower elements. // // All accesses to elements should use CAS operations, with no need to lock. tower [maxHeight]uint32 } // Skiplist maps keys to values (in memory) type Skiplist struct { height int32 // Current height. 1 <= height <= kMaxHeight. CAS. head *node ref int32 arena *Arena } // IncrRef increases the refcount func (s *Skiplist) IncrRef() { atomic.AddInt32(&s.ref, 1) } // DecrRef decrements the refcount, deallocating the Skiplist when done using it func (s *Skiplist) DecrRef() { newRef := atomic.AddInt32(&s.ref, -1) if newRef > 0 { return } s.arena.reset() // Indicate we are closed. Good for testing. Also, lets GC reclaim memory. Race condition // here would suggest we are accessing skiplist when we are supposed to have no reference! s.arena = nil } func (s *Skiplist) valid() bool { return s.arena != nil } func newNode(arena *Arena, key []byte, v y.ValueStruct, height int) *node { // The base level is already allocated in the node struct. offset := arena.putNode(height) node := arena.getNode(offset) node.keyOffset = arena.putKey(key) node.keySize = uint16(len(key)) node.height = uint16(height) node.value = encodeValue(arena.putVal(v), v.EncodedSize()) return node } func encodeValue(valOffset uint32, valSize uint16) uint64 { return uint64(valSize)<<32 | uint64(valOffset) } func decodeValue(value uint64) (valOffset uint32, valSize uint16) { valOffset = uint32(value) valSize = uint16(value >> 32) return } // NewSkiplist makes a new empty skiplist, with a given arena size func NewSkiplist(arenaSize int64) *Skiplist { arena := newArena(arenaSize) head := newNode(arena, nil, y.ValueStruct{}, maxHeight) return &Skiplist{ height: 1, head: head, arena: arena, ref: 1, } } func (s *node) getValueOffset() (uint32, uint16) { value := atomic.LoadUint64(&s.value) return decodeValue(value) } func (s *node) key(arena *Arena) []byte { return arena.getKey(s.keyOffset, s.keySize) } func (s *node) setValue(arena *Arena, v y.ValueStruct) { valOffset := arena.putVal(v) value := encodeValue(valOffset, v.EncodedSize()) atomic.StoreUint64(&s.value, value) } func (s *node) getNextOffset(h int) uint32 { return atomic.LoadUint32(&s.tower[h]) } func (s *node) casNextOffset(h int, old, val uint32) bool { return atomic.CompareAndSwapUint32(&s.tower[h], old, val) } // Returns true if key is strictly > n.key. // If n is nil, this is an "end" marker and we return false. //func (s *Skiplist) keyIsAfterNode(key []byte, n *node) bool { // y.AssertTrue(n != s.head) // return n != nil && y.CompareKeys(key, n.key) > 0 //} func randomHeight() int { h := 1 for h < maxHeight && rand.Uint32() <= heightIncrease { h++ } return h } func (s *Skiplist) getNext(nd *node, height int) *node { return s.arena.getNode(nd.getNextOffset(height)) } // findNear finds the node near to key. // If less=true, it finds rightmost node such that node.key < key (if allowEqual=false) or // node.key <= key (if allowEqual=true). // If less=false, it finds leftmost node such that node.key > key (if allowEqual=false) or // node.key >= key (if allowEqual=true). // Returns the node found. The bool returned is true if the node has key equal to given key. func (s *Skiplist) findNear(key []byte, less bool, allowEqual bool) (*node, bool) { x := s.head level := int(s.getHeight() - 1) for { // Assume x.key < key. next := s.getNext(x, level) if next == nil { // x.key < key < END OF LIST if level > 0 { // Can descend further to iterate closer to the end. level-- continue } // Level=0. Cannot descend further. Let's return something that makes sense. if !less { return nil, false } // Try to return x. Make sure it is not a head node. if x == s.head { return nil, false } return x, false } nextKey := next.key(s.arena) cmp := y.CompareKeys(key, nextKey) if cmp > 0 { // x.key < next.key < key. We can continue to move right. x = next continue } if cmp == 0 { // x.key < key == next.key. if allowEqual { return next, true } if !less { // We want >, so go to base level to grab the next bigger note. return s.getNext(next, 0), false } // We want <. If not base level, we should go closer in the next level. if level > 0 { level-- continue } // On base level. Return x. if x == s.head { return nil, false } return x, false } // cmp < 0. In other words, x.key < key < next. if level > 0 { level-- continue } // At base level. Need to return something. if !less { return next, false } // Try to return x. Make sure it is not a head node. if x == s.head { return nil, false } return x, false } } // findSpliceForLevel returns (outBefore, outAfter) with outBefore.key <= key <= outAfter.key. // The input "before" tells us where to start looking. // If we found a node with the same key, then we return outBefore = outAfter. // Otherwise, outBefore.key < key < outAfter.key. func (s *Skiplist) findSpliceForLevel(key []byte, before *node, level int) (*node, *node) { for { // Assume before.key < key. next := s.getNext(before, level) if next == nil { return before, next } nextKey := next.key(s.arena) cmp := y.CompareKeys(key, nextKey) if cmp == 0 { // Equality case. return next, next } if cmp < 0 { // before.key < key < next.key. We are done for this level. return before, next } before = next // Keep moving right on this level. } } func (s *Skiplist) getHeight() int32 { return atomic.LoadInt32(&s.height) } // Put inserts the key-value pair. func (s *Skiplist) Put(key []byte, v y.ValueStruct) { // Since we allow overwrite, we may not need to create a new node. We might not even need to // increase the height. Let's defer these actions. listHeight := s.getHeight() var prev [maxHeight + 1]*node var next [maxHeight + 1]*node prev[listHeight] = s.head next[listHeight] = nil for i := int(listHeight) - 1; i >= 0; i-- { // Use higher level to speed up for current level. prev[i], next[i] = s.findSpliceForLevel(key, prev[i+1], i) if prev[i] == next[i] { prev[i].setValue(s.arena, v) return } } // We do need to create a new node. height := randomHeight() x := newNode(s.arena, key, v, height) // Try to increase s.height via CAS. listHeight = s.getHeight() for height > int(listHeight) { if atomic.CompareAndSwapInt32(&s.height, listHeight, int32(height)) { // Successfully increased skiplist.height. break } listHeight = s.getHeight() } // We always insert from the base level and up. After you add a node in base level, we cannot // create a node in the level above because it would have discovered the node in the base level. for i := 0; i < height; i++ { for { if prev[i] == nil { y.AssertTrue(i > 1) // This cannot happen in base level. // We haven't computed prev, next for this level because height exceeds old listHeight. // For these levels, we expect the lists to be sparse, so we can just search from head. prev[i], next[i] = s.findSpliceForLevel(key, s.head, i) // Someone adds the exact same key before we are able to do so. This can only happen on // the base level. But we know we are not on the base level. y.AssertTrue(prev[i] != next[i]) } nextOffset := s.arena.getNodeOffset(next[i]) x.tower[i] = nextOffset if prev[i].casNextOffset(i, nextOffset, s.arena.getNodeOffset(x)) { // Managed to insert x between prev[i] and next[i]. Go to the next level. break } // CAS failed. We need to recompute prev and next. // It is unlikely to be helpful to try to use a different level as we redo the search, // because it is unlikely that lots of nodes are inserted between prev[i] and next[i]. prev[i], next[i] = s.findSpliceForLevel(key, prev[i], i) if prev[i] == next[i] { y.AssertTruef(i == 0, "Equality can happen only on base level: %d", i) prev[i].setValue(s.arena, v) return } } } } // Empty returns if the Skiplist is empty. func (s *Skiplist) Empty() bool { return s.findLast() == nil } // findLast returns the last element. If head (empty list), we return nil. All the find functions // will NEVER return the head nodes. func (s *Skiplist) findLast() *node { n := s.head level := int(s.getHeight()) - 1 for { next := s.getNext(n, level) if next != nil { n = next continue } if level == 0 { if n == s.head { return nil } return n } level-- } } // Get gets the value associated with the key. It returns a valid value if it finds equal or earlier // version of the same key. func (s *Skiplist) Get(key []byte) y.ValueStruct { n, _ := s.findNear(key, false, true) // findGreaterOrEqual. if n == nil { return y.ValueStruct{} } nextKey := s.arena.getKey(n.keyOffset, n.keySize) if !y.SameKey(key, nextKey) { return y.ValueStruct{} } valOffset, valSize := n.getValueOffset() vs := s.arena.getVal(valOffset, valSize) vs.Version = y.ParseTs(nextKey) return vs } // NewIterator returns a skiplist iterator. You have to Close() the iterator. func (s *Skiplist) NewIterator() *Iterator { s.IncrRef() return &Iterator{list: s} } // MemSize returns the size of the Skiplist in terms of how much memory is used within its internal // arena. func (s *Skiplist) MemSize() int64 { return s.arena.size() } // Iterator is an iterator over skiplist object. For new objects, you just // need to initialize Iterator.list. type Iterator struct { list *Skiplist n *node } // Close frees the resources held by the iterator func (s *Iterator) Close() error { s.list.DecrRef() return nil } // Valid returns true iff the iterator is positioned at a valid node. func (s *Iterator) Valid() bool { return s.n != nil } // Key returns the key at the current position. func (s *Iterator) Key() []byte { return s.list.arena.getKey(s.n.keyOffset, s.n.keySize) } // Value returns value. func (s *Iterator) Value() y.ValueStruct { valOffset, valSize := s.n.getValueOffset() return s.list.arena.getVal(valOffset, valSize) } // Next advances to the next position. func (s *Iterator) Next() { y.AssertTrue(s.Valid()) s.n = s.list.getNext(s.n, 0) } // Prev advances to the previous position. func (s *Iterator) Prev() { y.AssertTrue(s.Valid()) s.n, _ = s.list.findNear(s.Key(), true, false) // find <. No equality allowed. } // Seek advances to the first entry with a key >= target. func (s *Iterator) Seek(target []byte) { s.n, _ = s.list.findNear(target, false, true) // find >=. } // SeekForPrev finds an entry with key <= target. func (s *Iterator) SeekForPrev(target []byte) { s.n, _ = s.list.findNear(target, true, true) // find <=. } // SeekToFirst seeks position at the first entry in list. // Final state of iterator is Valid() iff list is not empty. func (s *Iterator) SeekToFirst() { s.n = s.list.getNext(s.list.head, 0) } // SeekToLast seeks position at the last entry in list. // Final state of iterator is Valid() iff list is not empty. func (s *Iterator) SeekToLast() { s.n = s.list.findLast() } // UniIterator is a unidirectional memtable iterator. It is a thin wrapper around // Iterator. We like to keep Iterator as before, because it is more powerful and // we might support bidirectional iterators in the future. type UniIterator struct { iter *Iterator reversed bool } // NewUniIterator returns a UniIterator. func (s *Skiplist) NewUniIterator(reversed bool) *UniIterator { return &UniIterator{ iter: s.NewIterator(), reversed: reversed, } } // Next implements y.Interface func (s *UniIterator) Next() { if !s.reversed { s.iter.Next() } else { s.iter.Prev() } } // Rewind implements y.Interface func (s *UniIterator) Rewind() { if !s.reversed { s.iter.SeekToFirst() } else { s.iter.SeekToLast() } } // Seek implements y.Interface func (s *UniIterator) Seek(key []byte) { if !s.reversed { s.iter.Seek(key) } else { s.iter.SeekForPrev(key) } } // Key implements y.Interface func (s *UniIterator) Key() []byte { return s.iter.Key() } // Value implements y.Interface func (s *UniIterator) Value() y.ValueStruct { return s.iter.Value() } // Valid implements y.Interface func (s *UniIterator) Valid() bool { return s.iter.Valid() } // Close implements y.Interface (and frees up the iter's resources) func (s *UniIterator) Close() error { return s.iter.Close() }