summaryrefslogtreecommitdiff
path: root/vendor/github.com/klauspost/compress/zstd/blockdec.go
diff options
context:
space:
mode:
Diffstat (limited to 'vendor/github.com/klauspost/compress/zstd/blockdec.go')
-rw-r--r--vendor/github.com/klauspost/compress/zstd/blockdec.go731
1 files changed, 0 insertions, 731 deletions
diff --git a/vendor/github.com/klauspost/compress/zstd/blockdec.go b/vendor/github.com/klauspost/compress/zstd/blockdec.go
deleted file mode 100644
index 9c28840c3..000000000
--- a/vendor/github.com/klauspost/compress/zstd/blockdec.go
+++ /dev/null
@@ -1,731 +0,0 @@
-// Copyright 2019+ Klaus Post. All rights reserved.
-// License information can be found in the LICENSE file.
-// Based on work by Yann Collet, released under BSD License.
-
-package zstd
-
-import (
- "bytes"
- "encoding/binary"
- "errors"
- "fmt"
- "hash/crc32"
- "io"
- "os"
- "path/filepath"
- "sync"
-
- "github.com/klauspost/compress/huff0"
- "github.com/klauspost/compress/zstd/internal/xxhash"
-)
-
-type blockType uint8
-
-//go:generate stringer -type=blockType,literalsBlockType,seqCompMode,tableIndex
-
-const (
- blockTypeRaw blockType = iota
- blockTypeRLE
- blockTypeCompressed
- blockTypeReserved
-)
-
-type literalsBlockType uint8
-
-const (
- literalsBlockRaw literalsBlockType = iota
- literalsBlockRLE
- literalsBlockCompressed
- literalsBlockTreeless
-)
-
-const (
- // maxCompressedBlockSize is the biggest allowed compressed block size (128KB)
- maxCompressedBlockSize = 128 << 10
-
- compressedBlockOverAlloc = 16
- maxCompressedBlockSizeAlloc = 128<<10 + compressedBlockOverAlloc
-
- // Maximum possible block size (all Raw+Uncompressed).
- maxBlockSize = (1 << 21) - 1
-
- maxMatchLen = 131074
- maxSequences = 0x7f00 + 0xffff
-
- // We support slightly less than the reference decoder to be able to
- // use ints on 32 bit archs.
- maxOffsetBits = 30
-)
-
-var (
- huffDecoderPool = sync.Pool{New: func() interface{} {
- return &huff0.Scratch{}
- }}
-
- fseDecoderPool = sync.Pool{New: func() interface{} {
- return &fseDecoder{}
- }}
-)
-
-type blockDec struct {
- // Raw source data of the block.
- data []byte
- dataStorage []byte
-
- // Destination of the decoded data.
- dst []byte
-
- // Buffer for literals data.
- literalBuf []byte
-
- // Window size of the block.
- WindowSize uint64
-
- err error
-
- // Check against this crc, if hasCRC is true.
- checkCRC uint32
- hasCRC bool
-
- // Frame to use for singlethreaded decoding.
- // Should not be used by the decoder itself since parent may be another frame.
- localFrame *frameDec
-
- sequence []seqVals
-
- async struct {
- newHist *history
- literals []byte
- seqData []byte
- seqSize int // Size of uncompressed sequences
- fcs uint64
- }
-
- // Block is RLE, this is the size.
- RLESize uint32
-
- Type blockType
-
- // Is this the last block of a frame?
- Last bool
-
- // Use less memory
- lowMem bool
-}
-
-func (b *blockDec) String() string {
- if b == nil {
- return "<nil>"
- }
- return fmt.Sprintf("Steam Size: %d, Type: %v, Last: %t, Window: %d", len(b.data), b.Type, b.Last, b.WindowSize)
-}
-
-func newBlockDec(lowMem bool) *blockDec {
- b := blockDec{
- lowMem: lowMem,
- }
- return &b
-}
-
-// reset will reset the block.
-// Input must be a start of a block and will be at the end of the block when returned.
-func (b *blockDec) reset(br byteBuffer, windowSize uint64) error {
- b.WindowSize = windowSize
- tmp, err := br.readSmall(3)
- if err != nil {
- println("Reading block header:", err)
- return err
- }
- bh := uint32(tmp[0]) | (uint32(tmp[1]) << 8) | (uint32(tmp[2]) << 16)
- b.Last = bh&1 != 0
- b.Type = blockType((bh >> 1) & 3)
- // find size.
- cSize := int(bh >> 3)
- maxSize := maxCompressedBlockSizeAlloc
- switch b.Type {
- case blockTypeReserved:
- return ErrReservedBlockType
- case blockTypeRLE:
- if cSize > maxCompressedBlockSize || cSize > int(b.WindowSize) {
- if debugDecoder {
- printf("rle block too big: csize:%d block: %+v\n", uint64(cSize), b)
- }
- return ErrWindowSizeExceeded
- }
- b.RLESize = uint32(cSize)
- if b.lowMem {
- maxSize = cSize
- }
- cSize = 1
- case blockTypeCompressed:
- if debugDecoder {
- println("Data size on stream:", cSize)
- }
- b.RLESize = 0
- maxSize = maxCompressedBlockSizeAlloc
- if windowSize < maxCompressedBlockSize && b.lowMem {
- maxSize = int(windowSize) + compressedBlockOverAlloc
- }
- if cSize > maxCompressedBlockSize || uint64(cSize) > b.WindowSize {
- if debugDecoder {
- printf("compressed block too big: csize:%d block: %+v\n", uint64(cSize), b)
- }
- return ErrCompressedSizeTooBig
- }
- // Empty compressed blocks must at least be 2 bytes
- // for Literals_Block_Type and one for Sequences_Section_Header.
- if cSize < 2 {
- return ErrBlockTooSmall
- }
- case blockTypeRaw:
- if cSize > maxCompressedBlockSize || cSize > int(b.WindowSize) {
- if debugDecoder {
- printf("rle block too big: csize:%d block: %+v\n", uint64(cSize), b)
- }
- return ErrWindowSizeExceeded
- }
-
- b.RLESize = 0
- // We do not need a destination for raw blocks.
- maxSize = -1
- default:
- panic("Invalid block type")
- }
-
- // Read block data.
- if _, ok := br.(*byteBuf); !ok && cap(b.dataStorage) < cSize {
- // byteBuf doesn't need a destination buffer.
- if b.lowMem || cSize > maxCompressedBlockSize {
- b.dataStorage = make([]byte, 0, cSize+compressedBlockOverAlloc)
- } else {
- b.dataStorage = make([]byte, 0, maxCompressedBlockSizeAlloc)
- }
- }
- b.data, err = br.readBig(cSize, b.dataStorage)
- if err != nil {
- if debugDecoder {
- println("Reading block:", err, "(", cSize, ")", len(b.data))
- printf("%T", br)
- }
- return err
- }
- if cap(b.dst) <= maxSize {
- b.dst = make([]byte, 0, maxSize+1)
- }
- return nil
-}
-
-// sendEOF will make the decoder send EOF on this frame.
-func (b *blockDec) sendErr(err error) {
- b.Last = true
- b.Type = blockTypeReserved
- b.err = err
-}
-
-// Close will release resources.
-// Closed blockDec cannot be reset.
-func (b *blockDec) Close() {
-}
-
-// decodeBuf
-func (b *blockDec) decodeBuf(hist *history) error {
- switch b.Type {
- case blockTypeRLE:
- if cap(b.dst) < int(b.RLESize) {
- if b.lowMem {
- b.dst = make([]byte, b.RLESize)
- } else {
- b.dst = make([]byte, maxCompressedBlockSize)
- }
- }
- b.dst = b.dst[:b.RLESize]
- v := b.data[0]
- for i := range b.dst {
- b.dst[i] = v
- }
- hist.appendKeep(b.dst)
- return nil
- case blockTypeRaw:
- hist.appendKeep(b.data)
- return nil
- case blockTypeCompressed:
- saved := b.dst
- // Append directly to history
- if hist.ignoreBuffer == 0 {
- b.dst = hist.b
- hist.b = nil
- } else {
- b.dst = b.dst[:0]
- }
- err := b.decodeCompressed(hist)
- if debugDecoder {
- println("Decompressed to total", len(b.dst), "bytes, hash:", xxhash.Sum64(b.dst), "error:", err)
- }
- if hist.ignoreBuffer == 0 {
- hist.b = b.dst
- b.dst = saved
- } else {
- hist.appendKeep(b.dst)
- }
- return err
- case blockTypeReserved:
- // Used for returning errors.
- return b.err
- default:
- panic("Invalid block type")
- }
-}
-
-func (b *blockDec) decodeLiterals(in []byte, hist *history) (remain []byte, err error) {
- // There must be at least one byte for Literals_Block_Type and one for Sequences_Section_Header
- if len(in) < 2 {
- return in, ErrBlockTooSmall
- }
-
- litType := literalsBlockType(in[0] & 3)
- var litRegenSize int
- var litCompSize int
- sizeFormat := (in[0] >> 2) & 3
- var fourStreams bool
- var literals []byte
- switch litType {
- case literalsBlockRaw, literalsBlockRLE:
- switch sizeFormat {
- case 0, 2:
- // Regenerated_Size uses 5 bits (0-31). Literals_Section_Header uses 1 byte.
- litRegenSize = int(in[0] >> 3)
- in = in[1:]
- case 1:
- // Regenerated_Size uses 12 bits (0-4095). Literals_Section_Header uses 2 bytes.
- litRegenSize = int(in[0]>>4) + (int(in[1]) << 4)
- in = in[2:]
- case 3:
- // Regenerated_Size uses 20 bits (0-1048575). Literals_Section_Header uses 3 bytes.
- if len(in) < 3 {
- println("too small: litType:", litType, " sizeFormat", sizeFormat, len(in))
- return in, ErrBlockTooSmall
- }
- litRegenSize = int(in[0]>>4) + (int(in[1]) << 4) + (int(in[2]) << 12)
- in = in[3:]
- }
- case literalsBlockCompressed, literalsBlockTreeless:
- switch sizeFormat {
- case 0, 1:
- // Both Regenerated_Size and Compressed_Size use 10 bits (0-1023).
- if len(in) < 3 {
- println("too small: litType:", litType, " sizeFormat", sizeFormat, len(in))
- return in, ErrBlockTooSmall
- }
- n := uint64(in[0]>>4) + (uint64(in[1]) << 4) + (uint64(in[2]) << 12)
- litRegenSize = int(n & 1023)
- litCompSize = int(n >> 10)
- fourStreams = sizeFormat == 1
- in = in[3:]
- case 2:
- fourStreams = true
- if len(in) < 4 {
- println("too small: litType:", litType, " sizeFormat", sizeFormat, len(in))
- return in, ErrBlockTooSmall
- }
- n := uint64(in[0]>>4) + (uint64(in[1]) << 4) + (uint64(in[2]) << 12) + (uint64(in[3]) << 20)
- litRegenSize = int(n & 16383)
- litCompSize = int(n >> 14)
- in = in[4:]
- case 3:
- fourStreams = true
- if len(in) < 5 {
- println("too small: litType:", litType, " sizeFormat", sizeFormat, len(in))
- return in, ErrBlockTooSmall
- }
- n := uint64(in[0]>>4) + (uint64(in[1]) << 4) + (uint64(in[2]) << 12) + (uint64(in[3]) << 20) + (uint64(in[4]) << 28)
- litRegenSize = int(n & 262143)
- litCompSize = int(n >> 18)
- in = in[5:]
- }
- }
- if debugDecoder {
- println("literals type:", litType, "litRegenSize:", litRegenSize, "litCompSize:", litCompSize, "sizeFormat:", sizeFormat, "4X:", fourStreams)
- }
- if litRegenSize > int(b.WindowSize) || litRegenSize > maxCompressedBlockSize {
- return in, ErrWindowSizeExceeded
- }
-
- switch litType {
- case literalsBlockRaw:
- if len(in) < litRegenSize {
- println("too small: litType:", litType, " sizeFormat", sizeFormat, "remain:", len(in), "want:", litRegenSize)
- return in, ErrBlockTooSmall
- }
- literals = in[:litRegenSize]
- in = in[litRegenSize:]
- //printf("Found %d uncompressed literals\n", litRegenSize)
- case literalsBlockRLE:
- if len(in) < 1 {
- println("too small: litType:", litType, " sizeFormat", sizeFormat, "remain:", len(in), "want:", 1)
- return in, ErrBlockTooSmall
- }
- if cap(b.literalBuf) < litRegenSize {
- if b.lowMem {
- b.literalBuf = make([]byte, litRegenSize, litRegenSize+compressedBlockOverAlloc)
- } else {
- b.literalBuf = make([]byte, litRegenSize, maxCompressedBlockSize+compressedBlockOverAlloc)
- }
- }
- literals = b.literalBuf[:litRegenSize]
- v := in[0]
- for i := range literals {
- literals[i] = v
- }
- in = in[1:]
- if debugDecoder {
- printf("Found %d RLE compressed literals\n", litRegenSize)
- }
- case literalsBlockTreeless:
- if len(in) < litCompSize {
- println("too small: litType:", litType, " sizeFormat", sizeFormat, "remain:", len(in), "want:", litCompSize)
- return in, ErrBlockTooSmall
- }
- // Store compressed literals, so we defer decoding until we get history.
- literals = in[:litCompSize]
- in = in[litCompSize:]
- if debugDecoder {
- printf("Found %d compressed literals\n", litCompSize)
- }
- huff := hist.huffTree
- if huff == nil {
- return in, errors.New("literal block was treeless, but no history was defined")
- }
- // Ensure we have space to store it.
- if cap(b.literalBuf) < litRegenSize {
- if b.lowMem {
- b.literalBuf = make([]byte, 0, litRegenSize+compressedBlockOverAlloc)
- } else {
- b.literalBuf = make([]byte, 0, maxCompressedBlockSize+compressedBlockOverAlloc)
- }
- }
- var err error
- // Use our out buffer.
- huff.MaxDecodedSize = litRegenSize
- if fourStreams {
- literals, err = huff.Decoder().Decompress4X(b.literalBuf[:0:litRegenSize], literals)
- } else {
- literals, err = huff.Decoder().Decompress1X(b.literalBuf[:0:litRegenSize], literals)
- }
- // Make sure we don't leak our literals buffer
- if err != nil {
- println("decompressing literals:", err)
- return in, err
- }
- if len(literals) != litRegenSize {
- return in, fmt.Errorf("literal output size mismatch want %d, got %d", litRegenSize, len(literals))
- }
-
- case literalsBlockCompressed:
- if len(in) < litCompSize {
- println("too small: litType:", litType, " sizeFormat", sizeFormat, "remain:", len(in), "want:", litCompSize)
- return in, ErrBlockTooSmall
- }
- literals = in[:litCompSize]
- in = in[litCompSize:]
- // Ensure we have space to store it.
- if cap(b.literalBuf) < litRegenSize {
- if b.lowMem {
- b.literalBuf = make([]byte, 0, litRegenSize+compressedBlockOverAlloc)
- } else {
- b.literalBuf = make([]byte, 0, maxCompressedBlockSize+compressedBlockOverAlloc)
- }
- }
- huff := hist.huffTree
- if huff == nil || (hist.dict != nil && huff == hist.dict.litEnc) {
- huff = huffDecoderPool.Get().(*huff0.Scratch)
- if huff == nil {
- huff = &huff0.Scratch{}
- }
- }
- var err error
- if debugDecoder {
- println("huff table input:", len(literals), "CRC:", crc32.ChecksumIEEE(literals))
- }
- huff, literals, err = huff0.ReadTable(literals, huff)
- if err != nil {
- println("reading huffman table:", err)
- return in, err
- }
- hist.huffTree = huff
- huff.MaxDecodedSize = litRegenSize
- // Use our out buffer.
- if fourStreams {
- literals, err = huff.Decoder().Decompress4X(b.literalBuf[:0:litRegenSize], literals)
- } else {
- literals, err = huff.Decoder().Decompress1X(b.literalBuf[:0:litRegenSize], literals)
- }
- if err != nil {
- println("decoding compressed literals:", err)
- return in, err
- }
- // Make sure we don't leak our literals buffer
- if len(literals) != litRegenSize {
- return in, fmt.Errorf("literal output size mismatch want %d, got %d", litRegenSize, len(literals))
- }
- // Re-cap to get extra size.
- literals = b.literalBuf[:len(literals)]
- if debugDecoder {
- printf("Decompressed %d literals into %d bytes\n", litCompSize, litRegenSize)
- }
- }
- hist.decoders.literals = literals
- return in, nil
-}
-
-// decodeCompressed will start decompressing a block.
-func (b *blockDec) decodeCompressed(hist *history) error {
- in := b.data
- in, err := b.decodeLiterals(in, hist)
- if err != nil {
- return err
- }
- err = b.prepareSequences(in, hist)
- if err != nil {
- return err
- }
- if hist.decoders.nSeqs == 0 {
- b.dst = append(b.dst, hist.decoders.literals...)
- return nil
- }
- before := len(hist.decoders.out)
- err = hist.decoders.decodeSync(hist.b[hist.ignoreBuffer:])
- if err != nil {
- return err
- }
- if hist.decoders.maxSyncLen > 0 {
- hist.decoders.maxSyncLen += uint64(before)
- hist.decoders.maxSyncLen -= uint64(len(hist.decoders.out))
- }
- b.dst = hist.decoders.out
- hist.recentOffsets = hist.decoders.prevOffset
- return nil
-}
-
-func (b *blockDec) prepareSequences(in []byte, hist *history) (err error) {
- if debugDecoder {
- printf("prepareSequences: %d byte(s) input\n", len(in))
- }
- // Decode Sequences
- // https://github.com/facebook/zstd/blob/dev/doc/zstd_compression_format.md#sequences-section
- if len(in) < 1 {
- return ErrBlockTooSmall
- }
- var nSeqs int
- seqHeader := in[0]
- switch {
- case seqHeader < 128:
- nSeqs = int(seqHeader)
- in = in[1:]
- case seqHeader < 255:
- if len(in) < 2 {
- return ErrBlockTooSmall
- }
- nSeqs = int(seqHeader-128)<<8 | int(in[1])
- in = in[2:]
- case seqHeader == 255:
- if len(in) < 3 {
- return ErrBlockTooSmall
- }
- nSeqs = 0x7f00 + int(in[1]) + (int(in[2]) << 8)
- in = in[3:]
- }
- if nSeqs == 0 && len(in) != 0 {
- // When no sequences, there should not be any more data...
- if debugDecoder {
- printf("prepareSequences: 0 sequences, but %d byte(s) left on stream\n", len(in))
- }
- return ErrUnexpectedBlockSize
- }
-
- var seqs = &hist.decoders
- seqs.nSeqs = nSeqs
- if nSeqs > 0 {
- if len(in) < 1 {
- return ErrBlockTooSmall
- }
- br := byteReader{b: in, off: 0}
- compMode := br.Uint8()
- br.advance(1)
- if debugDecoder {
- printf("Compression modes: 0b%b", compMode)
- }
- if compMode&3 != 0 {
- return errors.New("corrupt block: reserved bits not zero")
- }
- for i := uint(0); i < 3; i++ {
- mode := seqCompMode((compMode >> (6 - i*2)) & 3)
- if debugDecoder {
- println("Table", tableIndex(i), "is", mode)
- }
- var seq *sequenceDec
- switch tableIndex(i) {
- case tableLiteralLengths:
- seq = &seqs.litLengths
- case tableOffsets:
- seq = &seqs.offsets
- case tableMatchLengths:
- seq = &seqs.matchLengths
- default:
- panic("unknown table")
- }
- switch mode {
- case compModePredefined:
- if seq.fse != nil && !seq.fse.preDefined {
- fseDecoderPool.Put(seq.fse)
- }
- seq.fse = &fsePredef[i]
- case compModeRLE:
- if br.remain() < 1 {
- return ErrBlockTooSmall
- }
- v := br.Uint8()
- br.advance(1)
- if seq.fse == nil || seq.fse.preDefined {
- seq.fse = fseDecoderPool.Get().(*fseDecoder)
- }
- symb, err := decSymbolValue(v, symbolTableX[i])
- if err != nil {
- printf("RLE Transform table (%v) error: %v", tableIndex(i), err)
- return err
- }
- seq.fse.setRLE(symb)
- if debugDecoder {
- printf("RLE set to 0x%x, code: %v", symb, v)
- }
- case compModeFSE:
- if debugDecoder {
- println("Reading table for", tableIndex(i))
- }
- if seq.fse == nil || seq.fse.preDefined {
- seq.fse = fseDecoderPool.Get().(*fseDecoder)
- }
- err := seq.fse.readNCount(&br, uint16(maxTableSymbol[i]))
- if err != nil {
- println("Read table error:", err)
- return err
- }
- err = seq.fse.transform(symbolTableX[i])
- if err != nil {
- println("Transform table error:", err)
- return err
- }
- if debugDecoder {
- println("Read table ok", "symbolLen:", seq.fse.symbolLen)
- }
- case compModeRepeat:
- seq.repeat = true
- }
- if br.overread() {
- return io.ErrUnexpectedEOF
- }
- }
- in = br.unread()
- }
- if debugDecoder {
- println("Literals:", len(seqs.literals), "hash:", xxhash.Sum64(seqs.literals), "and", seqs.nSeqs, "sequences.")
- }
-
- if nSeqs == 0 {
- if len(b.sequence) > 0 {
- b.sequence = b.sequence[:0]
- }
- return nil
- }
- br := seqs.br
- if br == nil {
- br = &bitReader{}
- }
- if err := br.init(in); err != nil {
- return err
- }
-
- if err := seqs.initialize(br, hist, b.dst); err != nil {
- println("initializing sequences:", err)
- return err
- }
- // Extract blocks...
- if false && hist.dict == nil {
- fatalErr := func(err error) {
- if err != nil {
- panic(err)
- }
- }
- fn := fmt.Sprintf("n-%d-lits-%d-prev-%d-%d-%d-win-%d.blk", hist.decoders.nSeqs, len(hist.decoders.literals), hist.recentOffsets[0], hist.recentOffsets[1], hist.recentOffsets[2], hist.windowSize)
- var buf bytes.Buffer
- fatalErr(binary.Write(&buf, binary.LittleEndian, hist.decoders.litLengths.fse))
- fatalErr(binary.Write(&buf, binary.LittleEndian, hist.decoders.matchLengths.fse))
- fatalErr(binary.Write(&buf, binary.LittleEndian, hist.decoders.offsets.fse))
- buf.Write(in)
- os.WriteFile(filepath.Join("testdata", "seqs", fn), buf.Bytes(), os.ModePerm)
- }
-
- return nil
-}
-
-func (b *blockDec) decodeSequences(hist *history) error {
- if cap(b.sequence) < hist.decoders.nSeqs {
- if b.lowMem {
- b.sequence = make([]seqVals, 0, hist.decoders.nSeqs)
- } else {
- b.sequence = make([]seqVals, 0, 0x7F00+0xffff)
- }
- }
- b.sequence = b.sequence[:hist.decoders.nSeqs]
- if hist.decoders.nSeqs == 0 {
- hist.decoders.seqSize = len(hist.decoders.literals)
- return nil
- }
- hist.decoders.windowSize = hist.windowSize
- hist.decoders.prevOffset = hist.recentOffsets
-
- err := hist.decoders.decode(b.sequence)
- hist.recentOffsets = hist.decoders.prevOffset
- return err
-}
-
-func (b *blockDec) executeSequences(hist *history) error {
- hbytes := hist.b
- if len(hbytes) > hist.windowSize {
- hbytes = hbytes[len(hbytes)-hist.windowSize:]
- // We do not need history anymore.
- if hist.dict != nil {
- hist.dict.content = nil
- }
- }
- hist.decoders.windowSize = hist.windowSize
- hist.decoders.out = b.dst[:0]
- err := hist.decoders.execute(b.sequence, hbytes)
- if err != nil {
- return err
- }
- return b.updateHistory(hist)
-}
-
-func (b *blockDec) updateHistory(hist *history) error {
- if len(b.data) > maxCompressedBlockSize {
- return fmt.Errorf("compressed block size too large (%d)", len(b.data))
- }
- // Set output and release references.
- b.dst = hist.decoders.out
- hist.recentOffsets = hist.decoders.prevOffset
-
- if b.Last {
- // if last block we don't care about history.
- println("Last block, no history returned")
- hist.b = hist.b[:0]
- return nil
- } else {
- hist.append(b.dst)
- if debugDecoder {
- println("Finished block with ", len(b.sequence), "sequences. Added", len(b.dst), "to history, now length", len(hist.b))
- }
- }
- hist.decoders.out, hist.decoders.literals = nil, nil
-
- return nil
-}