Платформа ЦРНП "Мирокод" для разработки проектов
https://git.mirocod.ru
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
252 lines
5.9 KiB
252 lines
5.9 KiB
package flate |
|
|
|
import ( |
|
"io" |
|
"math" |
|
) |
|
|
|
const ( |
|
maxStatelessBlock = math.MaxInt16 |
|
|
|
slTableBits = 13 |
|
slTableSize = 1 << slTableBits |
|
slTableShift = 32 - slTableBits |
|
) |
|
|
|
type statelessWriter struct { |
|
dst io.Writer |
|
closed bool |
|
} |
|
|
|
func (s *statelessWriter) Close() error { |
|
if s.closed { |
|
return nil |
|
} |
|
s.closed = true |
|
// Emit EOF block |
|
return StatelessDeflate(s.dst, nil, true) |
|
} |
|
|
|
func (s *statelessWriter) Write(p []byte) (n int, err error) { |
|
err = StatelessDeflate(s.dst, p, false) |
|
if err != nil { |
|
return 0, err |
|
} |
|
return len(p), nil |
|
} |
|
|
|
func (s *statelessWriter) Reset(w io.Writer) { |
|
s.dst = w |
|
s.closed = false |
|
} |
|
|
|
// NewStatelessWriter will do compression but without maintaining any state |
|
// between Write calls. |
|
// There will be no memory kept between Write calls, |
|
// but compression and speed will be suboptimal. |
|
// Because of this, the size of actual Write calls will affect output size. |
|
func NewStatelessWriter(dst io.Writer) io.WriteCloser { |
|
return &statelessWriter{dst: dst} |
|
} |
|
|
|
// StatelessDeflate allows to compress directly to a Writer without retaining state. |
|
// When returning everything will be flushed. |
|
func StatelessDeflate(out io.Writer, in []byte, eof bool) error { |
|
var dst tokens |
|
bw := newHuffmanBitWriter(out) |
|
if eof && len(in) == 0 { |
|
// Just write an EOF block. |
|
// Could be faster... |
|
bw.writeStoredHeader(0, true) |
|
bw.flush() |
|
return bw.err |
|
} |
|
|
|
for len(in) > 0 { |
|
todo := in |
|
if len(todo) > maxStatelessBlock { |
|
todo = todo[:maxStatelessBlock] |
|
} |
|
in = in[len(todo):] |
|
// Compress |
|
statelessEnc(&dst, todo) |
|
isEof := eof && len(in) == 0 |
|
|
|
if dst.n == 0 { |
|
bw.writeStoredHeader(len(todo), isEof) |
|
if bw.err != nil { |
|
return bw.err |
|
} |
|
bw.writeBytes(todo) |
|
} else if int(dst.n) > len(todo)-len(todo)>>4 { |
|
// If we removed less than 1/16th, huffman compress the block. |
|
bw.writeBlockHuff(isEof, todo, false) |
|
} else { |
|
bw.writeBlockDynamic(&dst, isEof, todo, false) |
|
} |
|
if bw.err != nil { |
|
return bw.err |
|
} |
|
dst.Reset() |
|
} |
|
if !eof { |
|
// Align. |
|
bw.writeStoredHeader(0, false) |
|
} |
|
bw.flush() |
|
return bw.err |
|
} |
|
|
|
func hashSL(u uint32) uint32 { |
|
return (u * 0x1e35a7bd) >> slTableShift |
|
} |
|
|
|
func load3216(b []byte, i int16) uint32 { |
|
// Help the compiler eliminate bounds checks on the read so it can be done in a single read. |
|
b = b[i:] |
|
b = b[:4] |
|
return uint32(b[0]) | uint32(b[1])<<8 | uint32(b[2])<<16 | uint32(b[3])<<24 |
|
} |
|
|
|
func load6416(b []byte, i int16) uint64 { |
|
// Help the compiler eliminate bounds checks on the read so it can be done in a single read. |
|
b = b[i:] |
|
b = b[:8] |
|
return uint64(b[0]) | uint64(b[1])<<8 | uint64(b[2])<<16 | uint64(b[3])<<24 | |
|
uint64(b[4])<<32 | uint64(b[5])<<40 | uint64(b[6])<<48 | uint64(b[7])<<56 |
|
} |
|
|
|
func statelessEnc(dst *tokens, src []byte) { |
|
const ( |
|
inputMargin = 12 - 1 |
|
minNonLiteralBlockSize = 1 + 1 + inputMargin |
|
) |
|
|
|
type tableEntry struct { |
|
offset int16 |
|
} |
|
|
|
var table [slTableSize]tableEntry |
|
|
|
// This check isn't in the Snappy implementation, but there, the caller |
|
// instead of the callee handles this case. |
|
if len(src) < minNonLiteralBlockSize { |
|
// We do not fill the token table. |
|
// This will be picked up by caller. |
|
dst.n = uint16(len(src)) |
|
return |
|
} |
|
|
|
s := int16(1) |
|
nextEmit := int16(0) |
|
// sLimit is when to stop looking for offset/length copies. The inputMargin |
|
// lets us use a fast path for emitLiteral in the main loop, while we are |
|
// looking for copies. |
|
sLimit := int16(len(src) - inputMargin) |
|
|
|
// nextEmit is where in src the next emitLiteral should start from. |
|
cv := load3216(src, s) |
|
|
|
for { |
|
const skipLog = 5 |
|
const doEvery = 2 |
|
|
|
nextS := s |
|
var candidate tableEntry |
|
for { |
|
nextHash := hashSL(cv) |
|
candidate = table[nextHash] |
|
nextS = s + doEvery + (s-nextEmit)>>skipLog |
|
if nextS > sLimit || nextS <= 0 { |
|
goto emitRemainder |
|
} |
|
|
|
now := load6416(src, nextS) |
|
table[nextHash] = tableEntry{offset: s} |
|
nextHash = hashSL(uint32(now)) |
|
|
|
if cv == load3216(src, candidate.offset) { |
|
table[nextHash] = tableEntry{offset: nextS} |
|
break |
|
} |
|
|
|
// Do one right away... |
|
cv = uint32(now) |
|
s = nextS |
|
nextS++ |
|
candidate = table[nextHash] |
|
now >>= 8 |
|
table[nextHash] = tableEntry{offset: s} |
|
|
|
if cv == load3216(src, candidate.offset) { |
|
table[nextHash] = tableEntry{offset: nextS} |
|
break |
|
} |
|
cv = uint32(now) |
|
s = nextS |
|
} |
|
|
|
// A 4-byte match has been found. We'll later see if more than 4 bytes |
|
// match. But, prior to the match, src[nextEmit:s] are unmatched. Emit |
|
// them as literal bytes. |
|
for { |
|
// Invariant: we have a 4-byte match at s, and no need to emit any |
|
// literal bytes prior to s. |
|
|
|
// Extend the 4-byte match as long as possible. |
|
t := candidate.offset |
|
l := int16(matchLen(src[s+4:], src[t+4:]) + 4) |
|
|
|
// Extend backwards |
|
for t > 0 && s > nextEmit && src[t-1] == src[s-1] { |
|
s-- |
|
t-- |
|
l++ |
|
} |
|
if nextEmit < s { |
|
emitLiteral(dst, src[nextEmit:s]) |
|
} |
|
|
|
// Save the match found |
|
dst.AddMatchLong(int32(l), uint32(s-t-baseMatchOffset)) |
|
s += l |
|
nextEmit = s |
|
if nextS >= s { |
|
s = nextS + 1 |
|
} |
|
if s >= sLimit { |
|
goto emitRemainder |
|
} |
|
|
|
// We could immediately start working at s now, but to improve |
|
// compression we first update the hash table at s-2 and at s. If |
|
// another emitCopy is not our next move, also calculate nextHash |
|
// at s+1. At least on GOARCH=amd64, these three hash calculations |
|
// are faster as one load64 call (with some shifts) instead of |
|
// three load32 calls. |
|
x := load6416(src, s-2) |
|
o := s - 2 |
|
prevHash := hashSL(uint32(x)) |
|
table[prevHash] = tableEntry{offset: o} |
|
x >>= 16 |
|
currHash := hashSL(uint32(x)) |
|
candidate = table[currHash] |
|
table[currHash] = tableEntry{offset: o + 2} |
|
|
|
if uint32(x) != load3216(src, candidate.offset) { |
|
cv = uint32(x >> 8) |
|
s++ |
|
break |
|
} |
|
} |
|
} |
|
|
|
emitRemainder: |
|
if int(nextEmit) < len(src) { |
|
// If nothing was added, don't encode literals. |
|
if dst.n == 0 { |
|
return |
|
} |
|
emitLiteral(dst, src[nextEmit:]) |
|
} |
|
}
|
|
|