d28f005552
Fix limit for databases other than sqlite go mod tidy && go mod vendor Remove unneeded break statements Make everything work with the new xorm version Fix xorm logging Fix lint Fix redis init Fix using id field Fix database init for testing Change default database log level Add xorm logger Use const for postgres go mod tidy Merge branch 'master' into update/xorm # Conflicts: # go.mod # go.sum # vendor/modules.txt go mod vendor Fix loading fixtures for postgres Go mod vendor1 Update xorm to version 1 Co-authored-by: kolaente <k@knt.li> Reviewed-on: https://kolaente.dev/vikunja/api/pulls/323
524 lines
13 KiB
Go
524 lines
13 KiB
Go
// Copyright 2011 The LevelDB-Go Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style
|
|
// license that can be found in the LICENSE file.
|
|
|
|
// Taken from: https://code.google.com/p/leveldb-go/source/browse/leveldb/record/record.go?r=1d5ccbe03246da926391ee12d1c6caae054ff4b0
|
|
// License, authors and contributors informations can be found at bellow URLs respectively:
|
|
// https://code.google.com/p/leveldb-go/source/browse/LICENSE
|
|
// https://code.google.com/p/leveldb-go/source/browse/AUTHORS
|
|
// https://code.google.com/p/leveldb-go/source/browse/CONTRIBUTORS
|
|
|
|
// Package journal reads and writes sequences of journals. Each journal is a stream
|
|
// of bytes that completes before the next journal starts.
|
|
//
|
|
// When reading, call Next to obtain an io.Reader for the next journal. Next will
|
|
// return io.EOF when there are no more journals. It is valid to call Next
|
|
// without reading the current journal to exhaustion.
|
|
//
|
|
// When writing, call Next to obtain an io.Writer for the next journal. Calling
|
|
// Next finishes the current journal. Call Close to finish the final journal.
|
|
//
|
|
// Optionally, call Flush to finish the current journal and flush the underlying
|
|
// writer without starting a new journal. To start a new journal after flushing,
|
|
// call Next.
|
|
//
|
|
// Neither Readers or Writers are safe to use concurrently.
|
|
//
|
|
// Example code:
|
|
// func read(r io.Reader) ([]string, error) {
|
|
// var ss []string
|
|
// journals := journal.NewReader(r, nil, true, true)
|
|
// for {
|
|
// j, err := journals.Next()
|
|
// if err == io.EOF {
|
|
// break
|
|
// }
|
|
// if err != nil {
|
|
// return nil, err
|
|
// }
|
|
// s, err := ioutil.ReadAll(j)
|
|
// if err != nil {
|
|
// return nil, err
|
|
// }
|
|
// ss = append(ss, string(s))
|
|
// }
|
|
// return ss, nil
|
|
// }
|
|
//
|
|
// func write(w io.Writer, ss []string) error {
|
|
// journals := journal.NewWriter(w)
|
|
// for _, s := range ss {
|
|
// j, err := journals.Next()
|
|
// if err != nil {
|
|
// return err
|
|
// }
|
|
// if _, err := j.Write([]byte(s)), err != nil {
|
|
// return err
|
|
// }
|
|
// }
|
|
// return journals.Close()
|
|
// }
|
|
//
|
|
// The wire format is that the stream is divided into 32KiB blocks, and each
|
|
// block contains a number of tightly packed chunks. Chunks cannot cross block
|
|
// boundaries. The last block may be shorter than 32 KiB. Any unused bytes in a
|
|
// block must be zero.
|
|
//
|
|
// A journal maps to one or more chunks. Each chunk has a 7 byte header (a 4
|
|
// byte checksum, a 2 byte little-endian uint16 length, and a 1 byte chunk type)
|
|
// followed by a payload. The checksum is over the chunk type and the payload.
|
|
//
|
|
// There are four chunk types: whether the chunk is the full journal, or the
|
|
// first, middle or last chunk of a multi-chunk journal. A multi-chunk journal
|
|
// has one first chunk, zero or more middle chunks, and one last chunk.
|
|
//
|
|
// The wire format allows for limited recovery in the face of data corruption:
|
|
// on a format error (such as a checksum mismatch), the reader moves to the
|
|
// next block and looks for the next full or first chunk.
|
|
package journal
|
|
|
|
import (
|
|
"encoding/binary"
|
|
"fmt"
|
|
"io"
|
|
|
|
"github.com/syndtr/goleveldb/leveldb/errors"
|
|
"github.com/syndtr/goleveldb/leveldb/storage"
|
|
"github.com/syndtr/goleveldb/leveldb/util"
|
|
)
|
|
|
|
// These constants are part of the wire format and should not be changed.
|
|
const (
|
|
fullChunkType = 1
|
|
firstChunkType = 2
|
|
middleChunkType = 3
|
|
lastChunkType = 4
|
|
)
|
|
|
|
const (
|
|
blockSize = 32 * 1024
|
|
headerSize = 7
|
|
)
|
|
|
|
type flusher interface {
|
|
Flush() error
|
|
}
|
|
|
|
// ErrCorrupted is the error type that generated by corrupted block or chunk.
|
|
type ErrCorrupted struct {
|
|
Size int
|
|
Reason string
|
|
}
|
|
|
|
func (e *ErrCorrupted) Error() string {
|
|
return fmt.Sprintf("leveldb/journal: block/chunk corrupted: %s (%d bytes)", e.Reason, e.Size)
|
|
}
|
|
|
|
// Dropper is the interface that wrap simple Drop method. The Drop
|
|
// method will be called when the journal reader dropping a block or chunk.
|
|
type Dropper interface {
|
|
Drop(err error)
|
|
}
|
|
|
|
// Reader reads journals from an underlying io.Reader.
|
|
type Reader struct {
|
|
// r is the underlying reader.
|
|
r io.Reader
|
|
// the dropper.
|
|
dropper Dropper
|
|
// strict flag.
|
|
strict bool
|
|
// checksum flag.
|
|
checksum bool
|
|
// seq is the sequence number of the current journal.
|
|
seq int
|
|
// buf[i:j] is the unread portion of the current chunk's payload.
|
|
// The low bound, i, excludes the chunk header.
|
|
i, j int
|
|
// n is the number of bytes of buf that are valid. Once reading has started,
|
|
// only the final block can have n < blockSize.
|
|
n int
|
|
// last is whether the current chunk is the last chunk of the journal.
|
|
last bool
|
|
// err is any accumulated error.
|
|
err error
|
|
// buf is the buffer.
|
|
buf [blockSize]byte
|
|
}
|
|
|
|
// NewReader returns a new reader. The dropper may be nil, and if
|
|
// strict is true then corrupted or invalid chunk will halt the journal
|
|
// reader entirely.
|
|
func NewReader(r io.Reader, dropper Dropper, strict, checksum bool) *Reader {
|
|
return &Reader{
|
|
r: r,
|
|
dropper: dropper,
|
|
strict: strict,
|
|
checksum: checksum,
|
|
last: true,
|
|
}
|
|
}
|
|
|
|
var errSkip = errors.New("leveldb/journal: skipped")
|
|
|
|
func (r *Reader) corrupt(n int, reason string, skip bool) error {
|
|
if r.dropper != nil {
|
|
r.dropper.Drop(&ErrCorrupted{n, reason})
|
|
}
|
|
if r.strict && !skip {
|
|
r.err = errors.NewErrCorrupted(storage.FileDesc{}, &ErrCorrupted{n, reason})
|
|
return r.err
|
|
}
|
|
return errSkip
|
|
}
|
|
|
|
// nextChunk sets r.buf[r.i:r.j] to hold the next chunk's payload, reading the
|
|
// next block into the buffer if necessary.
|
|
func (r *Reader) nextChunk(first bool) error {
|
|
for {
|
|
if r.j+headerSize <= r.n {
|
|
checksum := binary.LittleEndian.Uint32(r.buf[r.j+0 : r.j+4])
|
|
length := binary.LittleEndian.Uint16(r.buf[r.j+4 : r.j+6])
|
|
chunkType := r.buf[r.j+6]
|
|
unprocBlock := r.n - r.j
|
|
if checksum == 0 && length == 0 && chunkType == 0 {
|
|
// Drop entire block.
|
|
r.i = r.n
|
|
r.j = r.n
|
|
return r.corrupt(unprocBlock, "zero header", false)
|
|
}
|
|
if chunkType < fullChunkType || chunkType > lastChunkType {
|
|
// Drop entire block.
|
|
r.i = r.n
|
|
r.j = r.n
|
|
return r.corrupt(unprocBlock, fmt.Sprintf("invalid chunk type %#x", chunkType), false)
|
|
}
|
|
r.i = r.j + headerSize
|
|
r.j = r.j + headerSize + int(length)
|
|
if r.j > r.n {
|
|
// Drop entire block.
|
|
r.i = r.n
|
|
r.j = r.n
|
|
return r.corrupt(unprocBlock, "chunk length overflows block", false)
|
|
} else if r.checksum && checksum != util.NewCRC(r.buf[r.i-1:r.j]).Value() {
|
|
// Drop entire block.
|
|
r.i = r.n
|
|
r.j = r.n
|
|
return r.corrupt(unprocBlock, "checksum mismatch", false)
|
|
}
|
|
if first && chunkType != fullChunkType && chunkType != firstChunkType {
|
|
chunkLength := (r.j - r.i) + headerSize
|
|
r.i = r.j
|
|
// Report the error, but skip it.
|
|
return r.corrupt(chunkLength, "orphan chunk", true)
|
|
}
|
|
r.last = chunkType == fullChunkType || chunkType == lastChunkType
|
|
return nil
|
|
}
|
|
|
|
// The last block.
|
|
if r.n < blockSize && r.n > 0 {
|
|
if !first {
|
|
return r.corrupt(0, "missing chunk part", false)
|
|
}
|
|
r.err = io.EOF
|
|
return r.err
|
|
}
|
|
|
|
// Read block.
|
|
n, err := io.ReadFull(r.r, r.buf[:])
|
|
if err != nil && err != io.EOF && err != io.ErrUnexpectedEOF {
|
|
return err
|
|
}
|
|
if n == 0 {
|
|
if !first {
|
|
return r.corrupt(0, "missing chunk part", false)
|
|
}
|
|
r.err = io.EOF
|
|
return r.err
|
|
}
|
|
r.i, r.j, r.n = 0, 0, n
|
|
}
|
|
}
|
|
|
|
// Next returns a reader for the next journal. It returns io.EOF if there are no
|
|
// more journals. The reader returned becomes stale after the next Next call,
|
|
// and should no longer be used. If strict is false, the reader will returns
|
|
// io.ErrUnexpectedEOF error when found corrupted journal.
|
|
func (r *Reader) Next() (io.Reader, error) {
|
|
r.seq++
|
|
if r.err != nil {
|
|
return nil, r.err
|
|
}
|
|
r.i = r.j
|
|
for {
|
|
if err := r.nextChunk(true); err == nil {
|
|
break
|
|
} else if err != errSkip {
|
|
return nil, err
|
|
}
|
|
}
|
|
return &singleReader{r, r.seq, nil}, nil
|
|
}
|
|
|
|
// Reset resets the journal reader, allows reuse of the journal reader. Reset returns
|
|
// last accumulated error.
|
|
func (r *Reader) Reset(reader io.Reader, dropper Dropper, strict, checksum bool) error {
|
|
r.seq++
|
|
err := r.err
|
|
r.r = reader
|
|
r.dropper = dropper
|
|
r.strict = strict
|
|
r.checksum = checksum
|
|
r.i = 0
|
|
r.j = 0
|
|
r.n = 0
|
|
r.last = true
|
|
r.err = nil
|
|
return err
|
|
}
|
|
|
|
type singleReader struct {
|
|
r *Reader
|
|
seq int
|
|
err error
|
|
}
|
|
|
|
func (x *singleReader) Read(p []byte) (int, error) {
|
|
r := x.r
|
|
if r.seq != x.seq {
|
|
return 0, errors.New("leveldb/journal: stale reader")
|
|
}
|
|
if x.err != nil {
|
|
return 0, x.err
|
|
}
|
|
if r.err != nil {
|
|
return 0, r.err
|
|
}
|
|
for r.i == r.j {
|
|
if r.last {
|
|
return 0, io.EOF
|
|
}
|
|
x.err = r.nextChunk(false)
|
|
if x.err != nil {
|
|
if x.err == errSkip {
|
|
x.err = io.ErrUnexpectedEOF
|
|
}
|
|
return 0, x.err
|
|
}
|
|
}
|
|
n := copy(p, r.buf[r.i:r.j])
|
|
r.i += n
|
|
return n, nil
|
|
}
|
|
|
|
func (x *singleReader) ReadByte() (byte, error) {
|
|
r := x.r
|
|
if r.seq != x.seq {
|
|
return 0, errors.New("leveldb/journal: stale reader")
|
|
}
|
|
if x.err != nil {
|
|
return 0, x.err
|
|
}
|
|
if r.err != nil {
|
|
return 0, r.err
|
|
}
|
|
for r.i == r.j {
|
|
if r.last {
|
|
return 0, io.EOF
|
|
}
|
|
x.err = r.nextChunk(false)
|
|
if x.err != nil {
|
|
if x.err == errSkip {
|
|
x.err = io.ErrUnexpectedEOF
|
|
}
|
|
return 0, x.err
|
|
}
|
|
}
|
|
c := r.buf[r.i]
|
|
r.i++
|
|
return c, nil
|
|
}
|
|
|
|
// Writer writes journals to an underlying io.Writer.
|
|
type Writer struct {
|
|
// w is the underlying writer.
|
|
w io.Writer
|
|
// seq is the sequence number of the current journal.
|
|
seq int
|
|
// f is w as a flusher.
|
|
f flusher
|
|
// buf[i:j] is the bytes that will become the current chunk.
|
|
// The low bound, i, includes the chunk header.
|
|
i, j int
|
|
// buf[:written] has already been written to w.
|
|
// written is zero unless Flush has been called.
|
|
written int
|
|
// first is whether the current chunk is the first chunk of the journal.
|
|
first bool
|
|
// pending is whether a chunk is buffered but not yet written.
|
|
pending bool
|
|
// err is any accumulated error.
|
|
err error
|
|
// buf is the buffer.
|
|
buf [blockSize]byte
|
|
}
|
|
|
|
// NewWriter returns a new Writer.
|
|
func NewWriter(w io.Writer) *Writer {
|
|
f, _ := w.(flusher)
|
|
return &Writer{
|
|
w: w,
|
|
f: f,
|
|
}
|
|
}
|
|
|
|
// fillHeader fills in the header for the pending chunk.
|
|
func (w *Writer) fillHeader(last bool) {
|
|
if w.i+headerSize > w.j || w.j > blockSize {
|
|
panic("leveldb/journal: bad writer state")
|
|
}
|
|
if last {
|
|
if w.first {
|
|
w.buf[w.i+6] = fullChunkType
|
|
} else {
|
|
w.buf[w.i+6] = lastChunkType
|
|
}
|
|
} else {
|
|
if w.first {
|
|
w.buf[w.i+6] = firstChunkType
|
|
} else {
|
|
w.buf[w.i+6] = middleChunkType
|
|
}
|
|
}
|
|
binary.LittleEndian.PutUint32(w.buf[w.i+0:w.i+4], util.NewCRC(w.buf[w.i+6:w.j]).Value())
|
|
binary.LittleEndian.PutUint16(w.buf[w.i+4:w.i+6], uint16(w.j-w.i-headerSize))
|
|
}
|
|
|
|
// writeBlock writes the buffered block to the underlying writer, and reserves
|
|
// space for the next chunk's header.
|
|
func (w *Writer) writeBlock() {
|
|
_, w.err = w.w.Write(w.buf[w.written:])
|
|
w.i = 0
|
|
w.j = headerSize
|
|
w.written = 0
|
|
}
|
|
|
|
// writePending finishes the current journal and writes the buffer to the
|
|
// underlying writer.
|
|
func (w *Writer) writePending() {
|
|
if w.err != nil {
|
|
return
|
|
}
|
|
if w.pending {
|
|
w.fillHeader(true)
|
|
w.pending = false
|
|
}
|
|
_, w.err = w.w.Write(w.buf[w.written:w.j])
|
|
w.written = w.j
|
|
}
|
|
|
|
// Close finishes the current journal and closes the writer.
|
|
func (w *Writer) Close() error {
|
|
w.seq++
|
|
w.writePending()
|
|
if w.err != nil {
|
|
return w.err
|
|
}
|
|
w.err = errors.New("leveldb/journal: closed Writer")
|
|
return nil
|
|
}
|
|
|
|
// Flush finishes the current journal, writes to the underlying writer, and
|
|
// flushes it if that writer implements interface{ Flush() error }.
|
|
func (w *Writer) Flush() error {
|
|
w.seq++
|
|
w.writePending()
|
|
if w.err != nil {
|
|
return w.err
|
|
}
|
|
if w.f != nil {
|
|
w.err = w.f.Flush()
|
|
return w.err
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// Reset resets the journal writer, allows reuse of the journal writer. Reset
|
|
// will also closes the journal writer if not already.
|
|
func (w *Writer) Reset(writer io.Writer) (err error) {
|
|
w.seq++
|
|
if w.err == nil {
|
|
w.writePending()
|
|
err = w.err
|
|
}
|
|
w.w = writer
|
|
w.f, _ = writer.(flusher)
|
|
w.i = 0
|
|
w.j = 0
|
|
w.written = 0
|
|
w.first = false
|
|
w.pending = false
|
|
w.err = nil
|
|
return
|
|
}
|
|
|
|
// Next returns a writer for the next journal. The writer returned becomes stale
|
|
// after the next Close, Flush or Next call, and should no longer be used.
|
|
func (w *Writer) Next() (io.Writer, error) {
|
|
w.seq++
|
|
if w.err != nil {
|
|
return nil, w.err
|
|
}
|
|
if w.pending {
|
|
w.fillHeader(true)
|
|
}
|
|
w.i = w.j
|
|
w.j = w.j + headerSize
|
|
// Check if there is room in the block for the header.
|
|
if w.j > blockSize {
|
|
// Fill in the rest of the block with zeroes.
|
|
for k := w.i; k < blockSize; k++ {
|
|
w.buf[k] = 0
|
|
}
|
|
w.writeBlock()
|
|
if w.err != nil {
|
|
return nil, w.err
|
|
}
|
|
}
|
|
w.first = true
|
|
w.pending = true
|
|
return singleWriter{w, w.seq}, nil
|
|
}
|
|
|
|
type singleWriter struct {
|
|
w *Writer
|
|
seq int
|
|
}
|
|
|
|
func (x singleWriter) Write(p []byte) (int, error) {
|
|
w := x.w
|
|
if w.seq != x.seq {
|
|
return 0, errors.New("leveldb/journal: stale writer")
|
|
}
|
|
if w.err != nil {
|
|
return 0, w.err
|
|
}
|
|
n0 := len(p)
|
|
for len(p) > 0 {
|
|
// Write a block, if it is full.
|
|
if w.j == blockSize {
|
|
w.fillHeader(false)
|
|
w.writeBlock()
|
|
if w.err != nil {
|
|
return 0, w.err
|
|
}
|
|
w.first = false
|
|
}
|
|
// Copy bytes into the buffer.
|
|
n := copy(w.buf[w.j:], p)
|
|
w.j += n
|
|
p = p[n:]
|
|
}
|
|
return n0, nil
|
|
}
|