d28f005552
Fix limit for databases other than sqlite go mod tidy && go mod vendor Remove unneeded break statements Make everything work with the new xorm version Fix xorm logging Fix lint Fix redis init Fix using id field Fix database init for testing Change default database log level Add xorm logger Use const for postgres go mod tidy Merge branch 'master' into update/xorm # Conflicts: # go.mod # go.sum # vendor/modules.txt go mod vendor Fix loading fixtures for postgres Go mod vendor1 Update xorm to version 1 Co-authored-by: kolaente <k@knt.li> Reviewed-on: https://kolaente.dev/vikunja/api/pulls/323
1139 lines
27 KiB
Go
1139 lines
27 KiB
Go
// Copyright (c) 2012, Suryandaru Triandana <syndtr@gmail.com>
|
|
// All rights reserved.
|
|
//
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file.
|
|
|
|
package table
|
|
|
|
import (
|
|
"encoding/binary"
|
|
"fmt"
|
|
"io"
|
|
"sort"
|
|
"strings"
|
|
"sync"
|
|
|
|
"github.com/golang/snappy"
|
|
|
|
"github.com/syndtr/goleveldb/leveldb/cache"
|
|
"github.com/syndtr/goleveldb/leveldb/comparer"
|
|
"github.com/syndtr/goleveldb/leveldb/errors"
|
|
"github.com/syndtr/goleveldb/leveldb/filter"
|
|
"github.com/syndtr/goleveldb/leveldb/iterator"
|
|
"github.com/syndtr/goleveldb/leveldb/opt"
|
|
"github.com/syndtr/goleveldb/leveldb/storage"
|
|
"github.com/syndtr/goleveldb/leveldb/util"
|
|
)
|
|
|
|
// Reader errors.
|
|
var (
|
|
ErrNotFound = errors.ErrNotFound
|
|
ErrReaderReleased = errors.New("leveldb/table: reader released")
|
|
ErrIterReleased = errors.New("leveldb/table: iterator released")
|
|
)
|
|
|
|
// ErrCorrupted describes error due to corruption. This error will be wrapped
|
|
// with errors.ErrCorrupted.
|
|
type ErrCorrupted struct {
|
|
Pos int64
|
|
Size int64
|
|
Kind string
|
|
Reason string
|
|
}
|
|
|
|
func (e *ErrCorrupted) Error() string {
|
|
return fmt.Sprintf("leveldb/table: corruption on %s (pos=%d): %s", e.Kind, e.Pos, e.Reason)
|
|
}
|
|
|
|
func max(x, y int) int {
|
|
if x > y {
|
|
return x
|
|
}
|
|
return y
|
|
}
|
|
|
|
type block struct {
|
|
bpool *util.BufferPool
|
|
bh blockHandle
|
|
data []byte
|
|
restartsLen int
|
|
restartsOffset int
|
|
}
|
|
|
|
func (b *block) seek(cmp comparer.Comparer, rstart, rlimit int, key []byte) (index, offset int, err error) {
|
|
index = sort.Search(b.restartsLen-rstart-(b.restartsLen-rlimit), func(i int) bool {
|
|
offset := int(binary.LittleEndian.Uint32(b.data[b.restartsOffset+4*(rstart+i):]))
|
|
offset++ // shared always zero, since this is a restart point
|
|
v1, n1 := binary.Uvarint(b.data[offset:]) // key length
|
|
_, n2 := binary.Uvarint(b.data[offset+n1:]) // value length
|
|
m := offset + n1 + n2
|
|
return cmp.Compare(b.data[m:m+int(v1)], key) > 0
|
|
}) + rstart - 1
|
|
if index < rstart {
|
|
// The smallest key is greater-than key sought.
|
|
index = rstart
|
|
}
|
|
offset = int(binary.LittleEndian.Uint32(b.data[b.restartsOffset+4*index:]))
|
|
return
|
|
}
|
|
|
|
func (b *block) restartIndex(rstart, rlimit, offset int) int {
|
|
return sort.Search(b.restartsLen-rstart-(b.restartsLen-rlimit), func(i int) bool {
|
|
return int(binary.LittleEndian.Uint32(b.data[b.restartsOffset+4*(rstart+i):])) > offset
|
|
}) + rstart - 1
|
|
}
|
|
|
|
func (b *block) restartOffset(index int) int {
|
|
return int(binary.LittleEndian.Uint32(b.data[b.restartsOffset+4*index:]))
|
|
}
|
|
|
|
func (b *block) entry(offset int) (key, value []byte, nShared, n int, err error) {
|
|
if offset >= b.restartsOffset {
|
|
if offset != b.restartsOffset {
|
|
err = &ErrCorrupted{Reason: "entries offset not aligned"}
|
|
}
|
|
return
|
|
}
|
|
v0, n0 := binary.Uvarint(b.data[offset:]) // Shared prefix length
|
|
v1, n1 := binary.Uvarint(b.data[offset+n0:]) // Key length
|
|
v2, n2 := binary.Uvarint(b.data[offset+n0+n1:]) // Value length
|
|
m := n0 + n1 + n2
|
|
n = m + int(v1) + int(v2)
|
|
if n0 <= 0 || n1 <= 0 || n2 <= 0 || offset+n > b.restartsOffset {
|
|
err = &ErrCorrupted{Reason: "entries corrupted"}
|
|
return
|
|
}
|
|
key = b.data[offset+m : offset+m+int(v1)]
|
|
value = b.data[offset+m+int(v1) : offset+n]
|
|
nShared = int(v0)
|
|
return
|
|
}
|
|
|
|
func (b *block) Release() {
|
|
b.bpool.Put(b.data)
|
|
b.bpool = nil
|
|
b.data = nil
|
|
}
|
|
|
|
type dir int
|
|
|
|
const (
|
|
dirReleased dir = iota - 1
|
|
dirSOI
|
|
dirEOI
|
|
dirBackward
|
|
dirForward
|
|
)
|
|
|
|
type blockIter struct {
|
|
tr *Reader
|
|
block *block
|
|
blockReleaser util.Releaser
|
|
releaser util.Releaser
|
|
key, value []byte
|
|
offset int
|
|
// Previous offset, only filled by Next.
|
|
prevOffset int
|
|
prevNode []int
|
|
prevKeys []byte
|
|
restartIndex int
|
|
// Iterator direction.
|
|
dir dir
|
|
// Restart index slice range.
|
|
riStart int
|
|
riLimit int
|
|
// Offset slice range.
|
|
offsetStart int
|
|
offsetRealStart int
|
|
offsetLimit int
|
|
// Error.
|
|
err error
|
|
}
|
|
|
|
func (i *blockIter) sErr(err error) {
|
|
i.err = err
|
|
i.key = nil
|
|
i.value = nil
|
|
i.prevNode = nil
|
|
i.prevKeys = nil
|
|
}
|
|
|
|
func (i *blockIter) reset() {
|
|
if i.dir == dirBackward {
|
|
i.prevNode = i.prevNode[:0]
|
|
i.prevKeys = i.prevKeys[:0]
|
|
}
|
|
i.restartIndex = i.riStart
|
|
i.offset = i.offsetStart
|
|
i.dir = dirSOI
|
|
i.key = i.key[:0]
|
|
i.value = nil
|
|
}
|
|
|
|
func (i *blockIter) isFirst() bool {
|
|
switch i.dir {
|
|
case dirForward:
|
|
return i.prevOffset == i.offsetRealStart
|
|
case dirBackward:
|
|
return len(i.prevNode) == 1 && i.restartIndex == i.riStart
|
|
}
|
|
return false
|
|
}
|
|
|
|
func (i *blockIter) isLast() bool {
|
|
switch i.dir {
|
|
case dirForward, dirBackward:
|
|
return i.offset == i.offsetLimit
|
|
}
|
|
return false
|
|
}
|
|
|
|
func (i *blockIter) First() bool {
|
|
if i.err != nil {
|
|
return false
|
|
} else if i.dir == dirReleased {
|
|
i.err = ErrIterReleased
|
|
return false
|
|
}
|
|
|
|
if i.dir == dirBackward {
|
|
i.prevNode = i.prevNode[:0]
|
|
i.prevKeys = i.prevKeys[:0]
|
|
}
|
|
i.dir = dirSOI
|
|
return i.Next()
|
|
}
|
|
|
|
func (i *blockIter) Last() bool {
|
|
if i.err != nil {
|
|
return false
|
|
} else if i.dir == dirReleased {
|
|
i.err = ErrIterReleased
|
|
return false
|
|
}
|
|
|
|
if i.dir == dirBackward {
|
|
i.prevNode = i.prevNode[:0]
|
|
i.prevKeys = i.prevKeys[:0]
|
|
}
|
|
i.dir = dirEOI
|
|
return i.Prev()
|
|
}
|
|
|
|
func (i *blockIter) Seek(key []byte) bool {
|
|
if i.err != nil {
|
|
return false
|
|
} else if i.dir == dirReleased {
|
|
i.err = ErrIterReleased
|
|
return false
|
|
}
|
|
|
|
ri, offset, err := i.block.seek(i.tr.cmp, i.riStart, i.riLimit, key)
|
|
if err != nil {
|
|
i.sErr(err)
|
|
return false
|
|
}
|
|
i.restartIndex = ri
|
|
i.offset = max(i.offsetStart, offset)
|
|
if i.dir == dirSOI || i.dir == dirEOI {
|
|
i.dir = dirForward
|
|
}
|
|
for i.Next() {
|
|
if i.tr.cmp.Compare(i.key, key) >= 0 {
|
|
return true
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
func (i *blockIter) Next() bool {
|
|
if i.dir == dirEOI || i.err != nil {
|
|
return false
|
|
} else if i.dir == dirReleased {
|
|
i.err = ErrIterReleased
|
|
return false
|
|
}
|
|
|
|
if i.dir == dirSOI {
|
|
i.restartIndex = i.riStart
|
|
i.offset = i.offsetStart
|
|
} else if i.dir == dirBackward {
|
|
i.prevNode = i.prevNode[:0]
|
|
i.prevKeys = i.prevKeys[:0]
|
|
}
|
|
for i.offset < i.offsetRealStart {
|
|
key, value, nShared, n, err := i.block.entry(i.offset)
|
|
if err != nil {
|
|
i.sErr(i.tr.fixErrCorruptedBH(i.block.bh, err))
|
|
return false
|
|
}
|
|
if n == 0 {
|
|
i.dir = dirEOI
|
|
return false
|
|
}
|
|
i.key = append(i.key[:nShared], key...)
|
|
i.value = value
|
|
i.offset += n
|
|
}
|
|
if i.offset >= i.offsetLimit {
|
|
i.dir = dirEOI
|
|
if i.offset != i.offsetLimit {
|
|
i.sErr(i.tr.newErrCorruptedBH(i.block.bh, "entries offset not aligned"))
|
|
}
|
|
return false
|
|
}
|
|
key, value, nShared, n, err := i.block.entry(i.offset)
|
|
if err != nil {
|
|
i.sErr(i.tr.fixErrCorruptedBH(i.block.bh, err))
|
|
return false
|
|
}
|
|
if n == 0 {
|
|
i.dir = dirEOI
|
|
return false
|
|
}
|
|
i.key = append(i.key[:nShared], key...)
|
|
i.value = value
|
|
i.prevOffset = i.offset
|
|
i.offset += n
|
|
i.dir = dirForward
|
|
return true
|
|
}
|
|
|
|
func (i *blockIter) Prev() bool {
|
|
if i.dir == dirSOI || i.err != nil {
|
|
return false
|
|
} else if i.dir == dirReleased {
|
|
i.err = ErrIterReleased
|
|
return false
|
|
}
|
|
|
|
var ri int
|
|
if i.dir == dirForward {
|
|
// Change direction.
|
|
i.offset = i.prevOffset
|
|
if i.offset == i.offsetRealStart {
|
|
i.dir = dirSOI
|
|
return false
|
|
}
|
|
ri = i.block.restartIndex(i.restartIndex, i.riLimit, i.offset)
|
|
i.dir = dirBackward
|
|
} else if i.dir == dirEOI {
|
|
// At the end of iterator.
|
|
i.restartIndex = i.riLimit
|
|
i.offset = i.offsetLimit
|
|
if i.offset == i.offsetRealStart {
|
|
i.dir = dirSOI
|
|
return false
|
|
}
|
|
ri = i.riLimit - 1
|
|
i.dir = dirBackward
|
|
} else if len(i.prevNode) == 1 {
|
|
// This is the end of a restart range.
|
|
i.offset = i.prevNode[0]
|
|
i.prevNode = i.prevNode[:0]
|
|
if i.restartIndex == i.riStart {
|
|
i.dir = dirSOI
|
|
return false
|
|
}
|
|
i.restartIndex--
|
|
ri = i.restartIndex
|
|
} else {
|
|
// In the middle of restart range, get from cache.
|
|
n := len(i.prevNode) - 3
|
|
node := i.prevNode[n:]
|
|
i.prevNode = i.prevNode[:n]
|
|
// Get the key.
|
|
ko := node[0]
|
|
i.key = append(i.key[:0], i.prevKeys[ko:]...)
|
|
i.prevKeys = i.prevKeys[:ko]
|
|
// Get the value.
|
|
vo := node[1]
|
|
vl := vo + node[2]
|
|
i.value = i.block.data[vo:vl]
|
|
i.offset = vl
|
|
return true
|
|
}
|
|
// Build entries cache.
|
|
i.key = i.key[:0]
|
|
i.value = nil
|
|
offset := i.block.restartOffset(ri)
|
|
if offset == i.offset {
|
|
ri--
|
|
if ri < 0 {
|
|
i.dir = dirSOI
|
|
return false
|
|
}
|
|
offset = i.block.restartOffset(ri)
|
|
}
|
|
i.prevNode = append(i.prevNode, offset)
|
|
for {
|
|
key, value, nShared, n, err := i.block.entry(offset)
|
|
if err != nil {
|
|
i.sErr(i.tr.fixErrCorruptedBH(i.block.bh, err))
|
|
return false
|
|
}
|
|
if offset >= i.offsetRealStart {
|
|
if i.value != nil {
|
|
// Appends 3 variables:
|
|
// 1. Previous keys offset
|
|
// 2. Value offset in the data block
|
|
// 3. Value length
|
|
i.prevNode = append(i.prevNode, len(i.prevKeys), offset-len(i.value), len(i.value))
|
|
i.prevKeys = append(i.prevKeys, i.key...)
|
|
}
|
|
i.value = value
|
|
}
|
|
i.key = append(i.key[:nShared], key...)
|
|
offset += n
|
|
// Stop if target offset reached.
|
|
if offset >= i.offset {
|
|
if offset != i.offset {
|
|
i.sErr(i.tr.newErrCorruptedBH(i.block.bh, "entries offset not aligned"))
|
|
return false
|
|
}
|
|
|
|
break
|
|
}
|
|
}
|
|
i.restartIndex = ri
|
|
i.offset = offset
|
|
return true
|
|
}
|
|
|
|
func (i *blockIter) Key() []byte {
|
|
if i.err != nil || i.dir <= dirEOI {
|
|
return nil
|
|
}
|
|
return i.key
|
|
}
|
|
|
|
func (i *blockIter) Value() []byte {
|
|
if i.err != nil || i.dir <= dirEOI {
|
|
return nil
|
|
}
|
|
return i.value
|
|
}
|
|
|
|
func (i *blockIter) Release() {
|
|
if i.dir != dirReleased {
|
|
i.tr = nil
|
|
i.block = nil
|
|
i.prevNode = nil
|
|
i.prevKeys = nil
|
|
i.key = nil
|
|
i.value = nil
|
|
i.dir = dirReleased
|
|
if i.blockReleaser != nil {
|
|
i.blockReleaser.Release()
|
|
i.blockReleaser = nil
|
|
}
|
|
if i.releaser != nil {
|
|
i.releaser.Release()
|
|
i.releaser = nil
|
|
}
|
|
}
|
|
}
|
|
|
|
func (i *blockIter) SetReleaser(releaser util.Releaser) {
|
|
if i.dir == dirReleased {
|
|
panic(util.ErrReleased)
|
|
}
|
|
if i.releaser != nil && releaser != nil {
|
|
panic(util.ErrHasReleaser)
|
|
}
|
|
i.releaser = releaser
|
|
}
|
|
|
|
func (i *blockIter) Valid() bool {
|
|
return i.err == nil && (i.dir == dirBackward || i.dir == dirForward)
|
|
}
|
|
|
|
func (i *blockIter) Error() error {
|
|
return i.err
|
|
}
|
|
|
|
type filterBlock struct {
|
|
bpool *util.BufferPool
|
|
data []byte
|
|
oOffset int
|
|
baseLg uint
|
|
filtersNum int
|
|
}
|
|
|
|
func (b *filterBlock) contains(filter filter.Filter, offset uint64, key []byte) bool {
|
|
i := int(offset >> b.baseLg)
|
|
if i < b.filtersNum {
|
|
o := b.data[b.oOffset+i*4:]
|
|
n := int(binary.LittleEndian.Uint32(o))
|
|
m := int(binary.LittleEndian.Uint32(o[4:]))
|
|
if n < m && m <= b.oOffset {
|
|
return filter.Contains(b.data[n:m], key)
|
|
} else if n == m {
|
|
return false
|
|
}
|
|
}
|
|
return true
|
|
}
|
|
|
|
func (b *filterBlock) Release() {
|
|
b.bpool.Put(b.data)
|
|
b.bpool = nil
|
|
b.data = nil
|
|
}
|
|
|
|
type indexIter struct {
|
|
*blockIter
|
|
tr *Reader
|
|
slice *util.Range
|
|
// Options
|
|
fillCache bool
|
|
}
|
|
|
|
func (i *indexIter) Get() iterator.Iterator {
|
|
value := i.Value()
|
|
if value == nil {
|
|
return nil
|
|
}
|
|
dataBH, n := decodeBlockHandle(value)
|
|
if n == 0 {
|
|
return iterator.NewEmptyIterator(i.tr.newErrCorruptedBH(i.tr.indexBH, "bad data block handle"))
|
|
}
|
|
|
|
var slice *util.Range
|
|
if i.slice != nil && (i.blockIter.isFirst() || i.blockIter.isLast()) {
|
|
slice = i.slice
|
|
}
|
|
return i.tr.getDataIterErr(dataBH, slice, i.tr.verifyChecksum, i.fillCache)
|
|
}
|
|
|
|
// Reader is a table reader.
|
|
type Reader struct {
|
|
mu sync.RWMutex
|
|
fd storage.FileDesc
|
|
reader io.ReaderAt
|
|
cache *cache.NamespaceGetter
|
|
err error
|
|
bpool *util.BufferPool
|
|
// Options
|
|
o *opt.Options
|
|
cmp comparer.Comparer
|
|
filter filter.Filter
|
|
verifyChecksum bool
|
|
|
|
dataEnd int64
|
|
metaBH, indexBH, filterBH blockHandle
|
|
indexBlock *block
|
|
filterBlock *filterBlock
|
|
}
|
|
|
|
func (r *Reader) blockKind(bh blockHandle) string {
|
|
switch bh.offset {
|
|
case r.metaBH.offset:
|
|
return "meta-block"
|
|
case r.indexBH.offset:
|
|
return "index-block"
|
|
case r.filterBH.offset:
|
|
if r.filterBH.length > 0 {
|
|
return "filter-block"
|
|
}
|
|
}
|
|
return "data-block"
|
|
}
|
|
|
|
func (r *Reader) newErrCorrupted(pos, size int64, kind, reason string) error {
|
|
return &errors.ErrCorrupted{Fd: r.fd, Err: &ErrCorrupted{Pos: pos, Size: size, Kind: kind, Reason: reason}}
|
|
}
|
|
|
|
func (r *Reader) newErrCorruptedBH(bh blockHandle, reason string) error {
|
|
return r.newErrCorrupted(int64(bh.offset), int64(bh.length), r.blockKind(bh), reason)
|
|
}
|
|
|
|
func (r *Reader) fixErrCorruptedBH(bh blockHandle, err error) error {
|
|
if cerr, ok := err.(*ErrCorrupted); ok {
|
|
cerr.Pos = int64(bh.offset)
|
|
cerr.Size = int64(bh.length)
|
|
cerr.Kind = r.blockKind(bh)
|
|
return &errors.ErrCorrupted{Fd: r.fd, Err: cerr}
|
|
}
|
|
return err
|
|
}
|
|
|
|
func (r *Reader) readRawBlock(bh blockHandle, verifyChecksum bool) ([]byte, error) {
|
|
data := r.bpool.Get(int(bh.length + blockTrailerLen))
|
|
if _, err := r.reader.ReadAt(data, int64(bh.offset)); err != nil && err != io.EOF {
|
|
return nil, err
|
|
}
|
|
|
|
if verifyChecksum {
|
|
n := bh.length + 1
|
|
checksum0 := binary.LittleEndian.Uint32(data[n:])
|
|
checksum1 := util.NewCRC(data[:n]).Value()
|
|
if checksum0 != checksum1 {
|
|
r.bpool.Put(data)
|
|
return nil, r.newErrCorruptedBH(bh, fmt.Sprintf("checksum mismatch, want=%#x got=%#x", checksum0, checksum1))
|
|
}
|
|
}
|
|
|
|
switch data[bh.length] {
|
|
case blockTypeNoCompression:
|
|
data = data[:bh.length]
|
|
case blockTypeSnappyCompression:
|
|
decLen, err := snappy.DecodedLen(data[:bh.length])
|
|
if err != nil {
|
|
r.bpool.Put(data)
|
|
return nil, r.newErrCorruptedBH(bh, err.Error())
|
|
}
|
|
decData := r.bpool.Get(decLen)
|
|
decData, err = snappy.Decode(decData, data[:bh.length])
|
|
r.bpool.Put(data)
|
|
if err != nil {
|
|
r.bpool.Put(decData)
|
|
return nil, r.newErrCorruptedBH(bh, err.Error())
|
|
}
|
|
data = decData
|
|
default:
|
|
r.bpool.Put(data)
|
|
return nil, r.newErrCorruptedBH(bh, fmt.Sprintf("unknown compression type %#x", data[bh.length]))
|
|
}
|
|
return data, nil
|
|
}
|
|
|
|
func (r *Reader) readBlock(bh blockHandle, verifyChecksum bool) (*block, error) {
|
|
data, err := r.readRawBlock(bh, verifyChecksum)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
restartsLen := int(binary.LittleEndian.Uint32(data[len(data)-4:]))
|
|
b := &block{
|
|
bpool: r.bpool,
|
|
bh: bh,
|
|
data: data,
|
|
restartsLen: restartsLen,
|
|
restartsOffset: len(data) - (restartsLen+1)*4,
|
|
}
|
|
return b, nil
|
|
}
|
|
|
|
func (r *Reader) readBlockCached(bh blockHandle, verifyChecksum, fillCache bool) (*block, util.Releaser, error) {
|
|
if r.cache != nil {
|
|
var (
|
|
err error
|
|
ch *cache.Handle
|
|
)
|
|
if fillCache {
|
|
ch = r.cache.Get(bh.offset, func() (size int, value cache.Value) {
|
|
var b *block
|
|
b, err = r.readBlock(bh, verifyChecksum)
|
|
if err != nil {
|
|
return 0, nil
|
|
}
|
|
return cap(b.data), b
|
|
})
|
|
} else {
|
|
ch = r.cache.Get(bh.offset, nil)
|
|
}
|
|
if ch != nil {
|
|
b, ok := ch.Value().(*block)
|
|
if !ok {
|
|
ch.Release()
|
|
return nil, nil, errors.New("leveldb/table: inconsistent block type")
|
|
}
|
|
return b, ch, err
|
|
} else if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
}
|
|
|
|
b, err := r.readBlock(bh, verifyChecksum)
|
|
return b, b, err
|
|
}
|
|
|
|
func (r *Reader) readFilterBlock(bh blockHandle) (*filterBlock, error) {
|
|
data, err := r.readRawBlock(bh, true)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
n := len(data)
|
|
if n < 5 {
|
|
return nil, r.newErrCorruptedBH(bh, "too short")
|
|
}
|
|
m := n - 5
|
|
oOffset := int(binary.LittleEndian.Uint32(data[m:]))
|
|
if oOffset > m {
|
|
return nil, r.newErrCorruptedBH(bh, "invalid data-offsets offset")
|
|
}
|
|
b := &filterBlock{
|
|
bpool: r.bpool,
|
|
data: data,
|
|
oOffset: oOffset,
|
|
baseLg: uint(data[n-1]),
|
|
filtersNum: (m - oOffset) / 4,
|
|
}
|
|
return b, nil
|
|
}
|
|
|
|
func (r *Reader) readFilterBlockCached(bh blockHandle, fillCache bool) (*filterBlock, util.Releaser, error) {
|
|
if r.cache != nil {
|
|
var (
|
|
err error
|
|
ch *cache.Handle
|
|
)
|
|
if fillCache {
|
|
ch = r.cache.Get(bh.offset, func() (size int, value cache.Value) {
|
|
var b *filterBlock
|
|
b, err = r.readFilterBlock(bh)
|
|
if err != nil {
|
|
return 0, nil
|
|
}
|
|
return cap(b.data), b
|
|
})
|
|
} else {
|
|
ch = r.cache.Get(bh.offset, nil)
|
|
}
|
|
if ch != nil {
|
|
b, ok := ch.Value().(*filterBlock)
|
|
if !ok {
|
|
ch.Release()
|
|
return nil, nil, errors.New("leveldb/table: inconsistent block type")
|
|
}
|
|
return b, ch, err
|
|
} else if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
}
|
|
|
|
b, err := r.readFilterBlock(bh)
|
|
return b, b, err
|
|
}
|
|
|
|
func (r *Reader) getIndexBlock(fillCache bool) (b *block, rel util.Releaser, err error) {
|
|
if r.indexBlock == nil {
|
|
return r.readBlockCached(r.indexBH, true, fillCache)
|
|
}
|
|
return r.indexBlock, util.NoopReleaser{}, nil
|
|
}
|
|
|
|
func (r *Reader) getFilterBlock(fillCache bool) (*filterBlock, util.Releaser, error) {
|
|
if r.filterBlock == nil {
|
|
return r.readFilterBlockCached(r.filterBH, fillCache)
|
|
}
|
|
return r.filterBlock, util.NoopReleaser{}, nil
|
|
}
|
|
|
|
func (r *Reader) newBlockIter(b *block, bReleaser util.Releaser, slice *util.Range, inclLimit bool) *blockIter {
|
|
bi := &blockIter{
|
|
tr: r,
|
|
block: b,
|
|
blockReleaser: bReleaser,
|
|
// Valid key should never be nil.
|
|
key: make([]byte, 0),
|
|
dir: dirSOI,
|
|
riStart: 0,
|
|
riLimit: b.restartsLen,
|
|
offsetStart: 0,
|
|
offsetRealStart: 0,
|
|
offsetLimit: b.restartsOffset,
|
|
}
|
|
if slice != nil {
|
|
if slice.Start != nil {
|
|
if bi.Seek(slice.Start) {
|
|
bi.riStart = b.restartIndex(bi.restartIndex, b.restartsLen, bi.prevOffset)
|
|
bi.offsetStart = b.restartOffset(bi.riStart)
|
|
bi.offsetRealStart = bi.prevOffset
|
|
} else {
|
|
bi.riStart = b.restartsLen
|
|
bi.offsetStart = b.restartsOffset
|
|
bi.offsetRealStart = b.restartsOffset
|
|
}
|
|
}
|
|
if slice.Limit != nil {
|
|
if bi.Seek(slice.Limit) && (!inclLimit || bi.Next()) {
|
|
bi.offsetLimit = bi.prevOffset
|
|
bi.riLimit = bi.restartIndex + 1
|
|
}
|
|
}
|
|
bi.reset()
|
|
if bi.offsetStart > bi.offsetLimit {
|
|
bi.sErr(errors.New("leveldb/table: invalid slice range"))
|
|
}
|
|
}
|
|
return bi
|
|
}
|
|
|
|
func (r *Reader) getDataIter(dataBH blockHandle, slice *util.Range, verifyChecksum, fillCache bool) iterator.Iterator {
|
|
b, rel, err := r.readBlockCached(dataBH, verifyChecksum, fillCache)
|
|
if err != nil {
|
|
return iterator.NewEmptyIterator(err)
|
|
}
|
|
return r.newBlockIter(b, rel, slice, false)
|
|
}
|
|
|
|
func (r *Reader) getDataIterErr(dataBH blockHandle, slice *util.Range, verifyChecksum, fillCache bool) iterator.Iterator {
|
|
r.mu.RLock()
|
|
defer r.mu.RUnlock()
|
|
|
|
if r.err != nil {
|
|
return iterator.NewEmptyIterator(r.err)
|
|
}
|
|
|
|
return r.getDataIter(dataBH, slice, verifyChecksum, fillCache)
|
|
}
|
|
|
|
// NewIterator creates an iterator from the table.
|
|
//
|
|
// Slice allows slicing the iterator to only contains keys in the given
|
|
// range. A nil Range.Start is treated as a key before all keys in the
|
|
// table. And a nil Range.Limit is treated as a key after all keys in
|
|
// the table.
|
|
//
|
|
// WARNING: Any slice returned by interator (e.g. slice returned by calling
|
|
// Iterator.Key() or Iterator.Key() methods), its content should not be modified
|
|
// unless noted otherwise.
|
|
//
|
|
// The returned iterator is not safe for concurrent use and should be released
|
|
// after use.
|
|
//
|
|
// Also read Iterator documentation of the leveldb/iterator package.
|
|
func (r *Reader) NewIterator(slice *util.Range, ro *opt.ReadOptions) iterator.Iterator {
|
|
r.mu.RLock()
|
|
defer r.mu.RUnlock()
|
|
|
|
if r.err != nil {
|
|
return iterator.NewEmptyIterator(r.err)
|
|
}
|
|
|
|
fillCache := !ro.GetDontFillCache()
|
|
indexBlock, rel, err := r.getIndexBlock(fillCache)
|
|
if err != nil {
|
|
return iterator.NewEmptyIterator(err)
|
|
}
|
|
index := &indexIter{
|
|
blockIter: r.newBlockIter(indexBlock, rel, slice, true),
|
|
tr: r,
|
|
slice: slice,
|
|
fillCache: !ro.GetDontFillCache(),
|
|
}
|
|
return iterator.NewIndexedIterator(index, opt.GetStrict(r.o, ro, opt.StrictReader))
|
|
}
|
|
|
|
func (r *Reader) find(key []byte, filtered bool, ro *opt.ReadOptions, noValue bool) (rkey, value []byte, err error) {
|
|
r.mu.RLock()
|
|
defer r.mu.RUnlock()
|
|
|
|
if r.err != nil {
|
|
err = r.err
|
|
return
|
|
}
|
|
|
|
indexBlock, rel, err := r.getIndexBlock(true)
|
|
if err != nil {
|
|
return
|
|
}
|
|
defer rel.Release()
|
|
|
|
index := r.newBlockIter(indexBlock, nil, nil, true)
|
|
defer index.Release()
|
|
|
|
if !index.Seek(key) {
|
|
if err = index.Error(); err == nil {
|
|
err = ErrNotFound
|
|
}
|
|
return
|
|
}
|
|
|
|
dataBH, n := decodeBlockHandle(index.Value())
|
|
if n == 0 {
|
|
r.err = r.newErrCorruptedBH(r.indexBH, "bad data block handle")
|
|
return nil, nil, r.err
|
|
}
|
|
|
|
// The filter should only used for exact match.
|
|
if filtered && r.filter != nil {
|
|
filterBlock, frel, ferr := r.getFilterBlock(true)
|
|
if ferr == nil {
|
|
if !filterBlock.contains(r.filter, dataBH.offset, key) {
|
|
frel.Release()
|
|
return nil, nil, ErrNotFound
|
|
}
|
|
frel.Release()
|
|
} else if !errors.IsCorrupted(ferr) {
|
|
return nil, nil, ferr
|
|
}
|
|
}
|
|
|
|
data := r.getDataIter(dataBH, nil, r.verifyChecksum, !ro.GetDontFillCache())
|
|
if !data.Seek(key) {
|
|
data.Release()
|
|
if err = data.Error(); err != nil {
|
|
return
|
|
}
|
|
|
|
// The nearest greater-than key is the first key of the next block.
|
|
if !index.Next() {
|
|
if err = index.Error(); err == nil {
|
|
err = ErrNotFound
|
|
}
|
|
return
|
|
}
|
|
|
|
dataBH, n = decodeBlockHandle(index.Value())
|
|
if n == 0 {
|
|
r.err = r.newErrCorruptedBH(r.indexBH, "bad data block handle")
|
|
return nil, nil, r.err
|
|
}
|
|
|
|
data = r.getDataIter(dataBH, nil, r.verifyChecksum, !ro.GetDontFillCache())
|
|
if !data.Next() {
|
|
data.Release()
|
|
if err = data.Error(); err == nil {
|
|
err = ErrNotFound
|
|
}
|
|
return
|
|
}
|
|
}
|
|
|
|
// Key doesn't use block buffer, no need to copy the buffer.
|
|
rkey = data.Key()
|
|
if !noValue {
|
|
if r.bpool == nil {
|
|
value = data.Value()
|
|
} else {
|
|
// Value does use block buffer, and since the buffer will be
|
|
// recycled, it need to be copied.
|
|
value = append([]byte{}, data.Value()...)
|
|
}
|
|
}
|
|
data.Release()
|
|
return
|
|
}
|
|
|
|
// Find finds key/value pair whose key is greater than or equal to the
|
|
// given key. It returns ErrNotFound if the table doesn't contain
|
|
// such pair.
|
|
// If filtered is true then the nearest 'block' will be checked against
|
|
// 'filter data' (if present) and will immediately return ErrNotFound if
|
|
// 'filter data' indicates that such pair doesn't exist.
|
|
//
|
|
// The caller may modify the contents of the returned slice as it is its
|
|
// own copy.
|
|
// It is safe to modify the contents of the argument after Find returns.
|
|
func (r *Reader) Find(key []byte, filtered bool, ro *opt.ReadOptions) (rkey, value []byte, err error) {
|
|
return r.find(key, filtered, ro, false)
|
|
}
|
|
|
|
// FindKey finds key that is greater than or equal to the given key.
|
|
// It returns ErrNotFound if the table doesn't contain such key.
|
|
// If filtered is true then the nearest 'block' will be checked against
|
|
// 'filter data' (if present) and will immediately return ErrNotFound if
|
|
// 'filter data' indicates that such key doesn't exist.
|
|
//
|
|
// The caller may modify the contents of the returned slice as it is its
|
|
// own copy.
|
|
// It is safe to modify the contents of the argument after Find returns.
|
|
func (r *Reader) FindKey(key []byte, filtered bool, ro *opt.ReadOptions) (rkey []byte, err error) {
|
|
rkey, _, err = r.find(key, filtered, ro, true)
|
|
return
|
|
}
|
|
|
|
// Get gets the value for the given key. It returns errors.ErrNotFound
|
|
// if the table does not contain the key.
|
|
//
|
|
// The caller may modify the contents of the returned slice as it is its
|
|
// own copy.
|
|
// It is safe to modify the contents of the argument after Find returns.
|
|
func (r *Reader) Get(key []byte, ro *opt.ReadOptions) (value []byte, err error) {
|
|
r.mu.RLock()
|
|
defer r.mu.RUnlock()
|
|
|
|
if r.err != nil {
|
|
err = r.err
|
|
return
|
|
}
|
|
|
|
rkey, value, err := r.find(key, false, ro, false)
|
|
if err == nil && r.cmp.Compare(rkey, key) != 0 {
|
|
value = nil
|
|
err = ErrNotFound
|
|
}
|
|
return
|
|
}
|
|
|
|
// OffsetOf returns approximate offset for the given key.
|
|
//
|
|
// It is safe to modify the contents of the argument after Get returns.
|
|
func (r *Reader) OffsetOf(key []byte) (offset int64, err error) {
|
|
r.mu.RLock()
|
|
defer r.mu.RUnlock()
|
|
|
|
if r.err != nil {
|
|
err = r.err
|
|
return
|
|
}
|
|
|
|
indexBlock, rel, err := r.readBlockCached(r.indexBH, true, true)
|
|
if err != nil {
|
|
return
|
|
}
|
|
defer rel.Release()
|
|
|
|
index := r.newBlockIter(indexBlock, nil, nil, true)
|
|
defer index.Release()
|
|
if index.Seek(key) {
|
|
dataBH, n := decodeBlockHandle(index.Value())
|
|
if n == 0 {
|
|
r.err = r.newErrCorruptedBH(r.indexBH, "bad data block handle")
|
|
return
|
|
}
|
|
offset = int64(dataBH.offset)
|
|
return
|
|
}
|
|
err = index.Error()
|
|
if err == nil {
|
|
offset = r.dataEnd
|
|
}
|
|
return
|
|
}
|
|
|
|
// Release implements util.Releaser.
|
|
// It also close the file if it is an io.Closer.
|
|
func (r *Reader) Release() {
|
|
r.mu.Lock()
|
|
defer r.mu.Unlock()
|
|
|
|
if closer, ok := r.reader.(io.Closer); ok {
|
|
closer.Close()
|
|
}
|
|
if r.indexBlock != nil {
|
|
r.indexBlock.Release()
|
|
r.indexBlock = nil
|
|
}
|
|
if r.filterBlock != nil {
|
|
r.filterBlock.Release()
|
|
r.filterBlock = nil
|
|
}
|
|
r.reader = nil
|
|
r.cache = nil
|
|
r.bpool = nil
|
|
r.err = ErrReaderReleased
|
|
}
|
|
|
|
// NewReader creates a new initialized table reader for the file.
|
|
// The fi, cache and bpool is optional and can be nil.
|
|
//
|
|
// The returned table reader instance is safe for concurrent use.
|
|
func NewReader(f io.ReaderAt, size int64, fd storage.FileDesc, cache *cache.NamespaceGetter, bpool *util.BufferPool, o *opt.Options) (*Reader, error) {
|
|
if f == nil {
|
|
return nil, errors.New("leveldb/table: nil file")
|
|
}
|
|
|
|
r := &Reader{
|
|
fd: fd,
|
|
reader: f,
|
|
cache: cache,
|
|
bpool: bpool,
|
|
o: o,
|
|
cmp: o.GetComparer(),
|
|
verifyChecksum: o.GetStrict(opt.StrictBlockChecksum),
|
|
}
|
|
|
|
if size < footerLen {
|
|
r.err = r.newErrCorrupted(0, size, "table", "too small")
|
|
return r, nil
|
|
}
|
|
|
|
footerPos := size - footerLen
|
|
var footer [footerLen]byte
|
|
if _, err := r.reader.ReadAt(footer[:], footerPos); err != nil && err != io.EOF {
|
|
return nil, err
|
|
}
|
|
if string(footer[footerLen-len(magic):footerLen]) != magic {
|
|
r.err = r.newErrCorrupted(footerPos, footerLen, "table-footer", "bad magic number")
|
|
return r, nil
|
|
}
|
|
|
|
var n int
|
|
// Decode the metaindex block handle.
|
|
r.metaBH, n = decodeBlockHandle(footer[:])
|
|
if n == 0 {
|
|
r.err = r.newErrCorrupted(footerPos, footerLen, "table-footer", "bad metaindex block handle")
|
|
return r, nil
|
|
}
|
|
|
|
// Decode the index block handle.
|
|
r.indexBH, n = decodeBlockHandle(footer[n:])
|
|
if n == 0 {
|
|
r.err = r.newErrCorrupted(footerPos, footerLen, "table-footer", "bad index block handle")
|
|
return r, nil
|
|
}
|
|
|
|
// Read metaindex block.
|
|
metaBlock, err := r.readBlock(r.metaBH, true)
|
|
if err != nil {
|
|
if errors.IsCorrupted(err) {
|
|
r.err = err
|
|
return r, nil
|
|
}
|
|
return nil, err
|
|
}
|
|
|
|
// Set data end.
|
|
r.dataEnd = int64(r.metaBH.offset)
|
|
|
|
// Read metaindex.
|
|
metaIter := r.newBlockIter(metaBlock, nil, nil, true)
|
|
for metaIter.Next() {
|
|
key := string(metaIter.Key())
|
|
if !strings.HasPrefix(key, "filter.") {
|
|
continue
|
|
}
|
|
fn := key[7:]
|
|
if f0 := o.GetFilter(); f0 != nil && f0.Name() == fn {
|
|
r.filter = f0
|
|
} else {
|
|
for _, f0 := range o.GetAltFilters() {
|
|
if f0.Name() == fn {
|
|
r.filter = f0
|
|
break
|
|
}
|
|
}
|
|
}
|
|
if r.filter != nil {
|
|
filterBH, n := decodeBlockHandle(metaIter.Value())
|
|
if n == 0 {
|
|
continue
|
|
}
|
|
r.filterBH = filterBH
|
|
// Update data end.
|
|
r.dataEnd = int64(filterBH.offset)
|
|
break
|
|
}
|
|
}
|
|
metaIter.Release()
|
|
metaBlock.Release()
|
|
|
|
// Cache index and filter block locally, since we don't have global cache.
|
|
if cache == nil {
|
|
r.indexBlock, err = r.readBlock(r.indexBH, true)
|
|
if err != nil {
|
|
if errors.IsCorrupted(err) {
|
|
r.err = err
|
|
return r, nil
|
|
}
|
|
return nil, err
|
|
}
|
|
if r.filter != nil {
|
|
r.filterBlock, err = r.readFilterBlock(r.filterBH)
|
|
if err != nil {
|
|
if !errors.IsCorrupted(err) {
|
|
return nil, err
|
|
}
|
|
|
|
// Don't use filter then.
|
|
r.filter = nil
|
|
}
|
|
}
|
|
}
|
|
|
|
return r, nil
|
|
}
|