William Petit
a5c67c29d0
Some checks failed
arcad/edge/pipeline/head There was a failure building this commit
164 lines
3.8 KiB
Go
164 lines
3.8 KiB
Go
package zim
|
|
|
|
import (
|
|
"bytes"
|
|
"encoding/binary"
|
|
"io"
|
|
"os"
|
|
"sync"
|
|
|
|
"github.com/pkg/errors"
|
|
)
|
|
|
|
type CompressedBlobReader struct {
|
|
reader *Reader
|
|
decoderFactory BlobDecoderFactory
|
|
|
|
clusterStartOffset uint64
|
|
clusterEndOffset uint64
|
|
blobIndex uint32
|
|
blobSize int
|
|
readOffset uint64
|
|
|
|
loadCluster sync.Once
|
|
loadClusterErr error
|
|
|
|
data []byte
|
|
closed bool
|
|
}
|
|
|
|
// Size implements BlobReader.
|
|
func (r *CompressedBlobReader) Size() (int64, error) {
|
|
if err := r.loadClusterData(); err != nil {
|
|
return 0, errors.WithStack(err)
|
|
}
|
|
|
|
return int64(len(r.data)), nil
|
|
}
|
|
|
|
// Close implements io.ReadCloser.
|
|
func (r *CompressedBlobReader) Close() error {
|
|
clear(r.data)
|
|
r.closed = true
|
|
return nil
|
|
}
|
|
|
|
// Read implements io.ReadCloser.
|
|
func (r *CompressedBlobReader) Read(p []byte) (int, error) {
|
|
if err := r.loadClusterData(); err != nil {
|
|
return 0, errors.WithStack(err)
|
|
}
|
|
|
|
length := len(p)
|
|
remaining := len(r.data) - int(r.readOffset)
|
|
if length > remaining {
|
|
length = remaining
|
|
}
|
|
|
|
chunk := make([]byte, length)
|
|
|
|
copy(chunk, r.data[r.readOffset:int(r.readOffset)+length])
|
|
copy(p, chunk)
|
|
|
|
if length == remaining {
|
|
return length, io.EOF
|
|
}
|
|
|
|
r.readOffset += uint64(length)
|
|
|
|
return length, nil
|
|
}
|
|
|
|
func (r *CompressedBlobReader) loadClusterData() error {
|
|
if r.closed {
|
|
return errors.WithStack(os.ErrClosed)
|
|
}
|
|
|
|
r.loadCluster.Do(func() {
|
|
compressedData := make([]byte, r.clusterEndOffset-r.clusterStartOffset)
|
|
if err := r.reader.readRange(int64(r.clusterStartOffset+1), compressedData); err != nil {
|
|
r.loadClusterErr = errors.WithStack(err)
|
|
return
|
|
}
|
|
|
|
blobBuffer := bytes.NewBuffer(compressedData)
|
|
|
|
decoder, err := r.decoderFactory(blobBuffer)
|
|
if err != nil {
|
|
r.loadClusterErr = errors.WithStack(err)
|
|
return
|
|
}
|
|
|
|
defer decoder.Close()
|
|
|
|
uncompressedData, err := io.ReadAll(decoder)
|
|
if err != nil {
|
|
r.loadClusterErr = errors.WithStack(err)
|
|
return
|
|
}
|
|
|
|
var (
|
|
blobStart uint64
|
|
blobEnd uint64
|
|
)
|
|
|
|
if r.blobSize == 8 {
|
|
blobStart64, err := readUint64(uncompressedData[r.blobIndex*uint32(r.blobSize):r.blobIndex*uint32(r.blobSize)+uint32(r.blobSize)], binary.LittleEndian)
|
|
if err != nil {
|
|
r.loadClusterErr = errors.WithStack(err)
|
|
return
|
|
}
|
|
|
|
blobStart = blobStart64
|
|
|
|
blobEnd64, err := readUint64(uncompressedData[r.blobIndex*uint32(r.blobSize)+uint32(r.blobSize):r.blobIndex*uint32(r.blobSize)+uint32(r.blobSize)+uint32(r.blobSize)], binary.LittleEndian)
|
|
if err != nil {
|
|
r.loadClusterErr = errors.WithStack(err)
|
|
return
|
|
}
|
|
|
|
blobEnd = blobEnd64
|
|
} else {
|
|
blobStart32, err := readUint32(uncompressedData[r.blobIndex*uint32(r.blobSize):r.blobIndex*uint32(r.blobSize)+uint32(r.blobSize)], binary.LittleEndian)
|
|
if err != nil {
|
|
r.loadClusterErr = errors.WithStack(err)
|
|
return
|
|
}
|
|
|
|
blobStart = uint64(blobStart32)
|
|
|
|
blobEnd32, err := readUint32(uncompressedData[r.blobIndex*uint32(r.blobSize)+uint32(r.blobSize):r.blobIndex*uint32(r.blobSize)+uint32(r.blobSize)+uint32(r.blobSize)], binary.LittleEndian)
|
|
if err != nil {
|
|
r.loadClusterErr = errors.WithStack(err)
|
|
return
|
|
}
|
|
|
|
blobEnd = uint64(blobEnd32)
|
|
}
|
|
|
|
r.data = make([]byte, blobEnd-blobStart)
|
|
copy(r.data, uncompressedData[blobStart:blobEnd])
|
|
})
|
|
if r.loadClusterErr != nil {
|
|
return errors.WithStack(r.loadClusterErr)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
type BlobDecoderFactory func(io.Reader) (io.ReadCloser, error)
|
|
|
|
func NewCompressedBlobReader(reader *Reader, decoderFactory BlobDecoderFactory, clusterStartOffset, clusterEndOffset uint64, blobIndex uint32, blobSize int) *CompressedBlobReader {
|
|
return &CompressedBlobReader{
|
|
reader: reader,
|
|
decoderFactory: decoderFactory,
|
|
clusterStartOffset: clusterStartOffset,
|
|
clusterEndOffset: clusterEndOffset,
|
|
blobIndex: blobIndex,
|
|
blobSize: blobSize,
|
|
readOffset: 0,
|
|
}
|
|
}
|
|
|
|
var _ BlobReader = &UncompressedBlobReader{}
|