summaryrefslogtreecommitdiffhomepage
path: root/pkg/compressio/compressio_test.go
diff options
context:
space:
mode:
Diffstat (limited to 'pkg/compressio/compressio_test.go')
-rw-r--r--pkg/compressio/compressio_test.go290
1 files changed, 290 insertions, 0 deletions
diff --git a/pkg/compressio/compressio_test.go b/pkg/compressio/compressio_test.go
new file mode 100644
index 000000000..86dc47e44
--- /dev/null
+++ b/pkg/compressio/compressio_test.go
@@ -0,0 +1,290 @@
+// Copyright 2018 The gVisor Authors.
+//
+// Licensed under the Apache License, Version 2.0 (the "License");
+// you may not use this file except in compliance with the License.
+// You may obtain a copy of the License at
+//
+// http://www.apache.org/licenses/LICENSE-2.0
+//
+// Unless required by applicable law or agreed to in writing, software
+// distributed under the License is distributed on an "AS IS" BASIS,
+// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+// See the License for the specific language governing permissions and
+// limitations under the License.
+
+package compressio
+
+import (
+ "bytes"
+ "compress/flate"
+ "encoding/base64"
+ "fmt"
+ "io"
+ "math/rand"
+ "runtime"
+ "testing"
+ "time"
+)
+
+type harness interface {
+ Errorf(format string, v ...interface{})
+ Fatalf(format string, v ...interface{})
+ Logf(format string, v ...interface{})
+}
+
+func initTest(t harness, size int) []byte {
+ // Set number of processes to number of CPUs.
+ runtime.GOMAXPROCS(runtime.NumCPU())
+
+ // Construct synthetic data. We do this by encoding random data with
+ // base64. This gives a high level of entropy, but still quite a bit of
+ // structure, to give reasonable compression ratios (~75%).
+ var buf bytes.Buffer
+ bufW := base64.NewEncoder(base64.RawStdEncoding, &buf)
+ bufR := rand.New(rand.NewSource(0))
+ if _, err := io.CopyN(bufW, bufR, int64(size)); err != nil {
+ t.Fatalf("unable to seed random data: %v", err)
+ }
+ return buf.Bytes()
+}
+
+type testOpts struct {
+ Name string
+ Data []byte
+ NewWriter func(*bytes.Buffer) (io.Writer, error)
+ NewReader func(*bytes.Buffer) (io.Reader, error)
+ PreCompress func()
+ PostCompress func()
+ PreDecompress func()
+ PostDecompress func()
+ CompressIters int
+ DecompressIters int
+ CorruptData bool
+}
+
+func doTest(t harness, opts testOpts) {
+ // Compress.
+ var compressed bytes.Buffer
+ compressionStartTime := time.Now()
+ if opts.PreCompress != nil {
+ opts.PreCompress()
+ }
+ if opts.CompressIters <= 0 {
+ opts.CompressIters = 1
+ }
+ for i := 0; i < opts.CompressIters; i++ {
+ compressed.Reset()
+ w, err := opts.NewWriter(&compressed)
+ if err != nil {
+ t.Errorf("%s: NewWriter got err %v, expected nil", opts.Name, err)
+ }
+ if _, err := io.Copy(w, bytes.NewBuffer(opts.Data)); err != nil {
+ t.Errorf("%s: compress got err %v, expected nil", opts.Name, err)
+ return
+ }
+ closer, ok := w.(io.Closer)
+ if ok {
+ if err := closer.Close(); err != nil {
+ t.Errorf("%s: got err %v, expected nil", opts.Name, err)
+ return
+ }
+ }
+ }
+ if opts.PostCompress != nil {
+ opts.PostCompress()
+ }
+ compressionTime := time.Since(compressionStartTime)
+ compressionRatio := float32(compressed.Len()) / float32(len(opts.Data))
+
+ // Decompress.
+ var decompressed bytes.Buffer
+ decompressionStartTime := time.Now()
+ if opts.PreDecompress != nil {
+ opts.PreDecompress()
+ }
+ if opts.DecompressIters <= 0 {
+ opts.DecompressIters = 1
+ }
+ if opts.CorruptData {
+ b := compressed.Bytes()
+ b[rand.Intn(len(b))]++
+ }
+ for i := 0; i < opts.DecompressIters; i++ {
+ decompressed.Reset()
+ r, err := opts.NewReader(bytes.NewBuffer(compressed.Bytes()))
+ if err != nil {
+ if opts.CorruptData {
+ continue
+ }
+ t.Errorf("%s: NewReader got err %v, expected nil", opts.Name, err)
+ return
+ }
+ if _, err := io.Copy(&decompressed, r); (err != nil) != opts.CorruptData {
+ t.Errorf("%s: decompress got err %v unexpectly", opts.Name, err)
+ return
+ }
+ }
+ if opts.PostDecompress != nil {
+ opts.PostDecompress()
+ }
+ decompressionTime := time.Since(decompressionStartTime)
+
+ if opts.CorruptData {
+ return
+ }
+
+ // Verify.
+ if decompressed.Len() != len(opts.Data) {
+ t.Errorf("%s: got %d bytes, expected %d", opts.Name, decompressed.Len(), len(opts.Data))
+ }
+ if !bytes.Equal(opts.Data, decompressed.Bytes()) {
+ t.Errorf("%s: got mismatch, expected match", opts.Name)
+ if len(opts.Data) < 32 { // Don't flood the logs.
+ t.Errorf("got %v, expected %v", decompressed.Bytes(), opts.Data)
+ }
+ }
+
+ t.Logf("%s: compression time %v, ratio %2.2f, decompression time %v",
+ opts.Name, compressionTime, compressionRatio, decompressionTime)
+}
+
+var hashKey = []byte("01234567890123456789012345678901")
+
+func TestCompress(t *testing.T) {
+ rand.Seed(time.Now().Unix())
+
+ var (
+ data = initTest(t, 10*1024*1024)
+ data0 = data[:0]
+ data1 = data[:1]
+ data2 = data[:11]
+ data3 = data[:16]
+ data4 = data[:]
+ )
+
+ for _, data := range [][]byte{data0, data1, data2, data3, data4} {
+ for _, blockSize := range []uint32{1, 4, 1024, 4 * 1024, 16 * 1024} {
+ // Skip annoying tests; they just take too long.
+ if blockSize <= 16 && len(data) > 16 {
+ continue
+ }
+
+ for _, key := range [][]byte{nil, hashKey} {
+ for _, corruptData := range []bool{false, true} {
+ if key == nil && corruptData {
+ // No need to test corrupt data
+ // case when not doing hashing.
+ continue
+ }
+ // Do the compress test.
+ doTest(t, testOpts{
+ Name: fmt.Sprintf("len(data)=%d, blockSize=%d, key=%s, corruptData=%v", len(data), blockSize, string(key), corruptData),
+ Data: data,
+ NewWriter: func(b *bytes.Buffer) (io.Writer, error) {
+ return NewWriter(b, key, blockSize, flate.BestSpeed)
+ },
+ NewReader: func(b *bytes.Buffer) (io.Reader, error) {
+ return NewReader(b, key)
+ },
+ CorruptData: corruptData,
+ })
+ }
+ }
+ }
+
+ // Do the vanilla test.
+ doTest(t, testOpts{
+ Name: fmt.Sprintf("len(data)=%d, vanilla flate", len(data)),
+ Data: data,
+ NewWriter: func(b *bytes.Buffer) (io.Writer, error) {
+ return flate.NewWriter(b, flate.BestSpeed)
+ },
+ NewReader: func(b *bytes.Buffer) (io.Reader, error) {
+ return flate.NewReader(b), nil
+ },
+ })
+ }
+}
+
+const (
+ benchDataSize = 600 * 1024 * 1024
+)
+
+func benchmark(b *testing.B, compress bool, hash bool, blockSize uint32) {
+ b.StopTimer()
+ b.SetBytes(benchDataSize)
+ data := initTest(b, benchDataSize)
+ compIters := b.N
+ decompIters := b.N
+ if compress {
+ decompIters = 0
+ } else {
+ compIters = 0
+ }
+ key := hashKey
+ if !hash {
+ key = nil
+ }
+ doTest(b, testOpts{
+ Name: fmt.Sprintf("compress=%t, hash=%t, len(data)=%d, blockSize=%d", compress, hash, len(data), blockSize),
+ Data: data,
+ PreCompress: b.StartTimer,
+ PostCompress: b.StopTimer,
+ NewWriter: func(b *bytes.Buffer) (io.Writer, error) {
+ return NewWriter(b, key, blockSize, flate.BestSpeed)
+ },
+ NewReader: func(b *bytes.Buffer) (io.Reader, error) {
+ return NewReader(b, key)
+ },
+ CompressIters: compIters,
+ DecompressIters: decompIters,
+ })
+}
+
+func BenchmarkCompressNoHash64K(b *testing.B) {
+ benchmark(b, true, false, 64*1024)
+}
+
+func BenchmarkCompressHash64K(b *testing.B) {
+ benchmark(b, true, true, 64*1024)
+}
+
+func BenchmarkDecompressNoHash64K(b *testing.B) {
+ benchmark(b, false, false, 64*1024)
+}
+
+func BenchmarkDecompressHash64K(b *testing.B) {
+ benchmark(b, false, true, 64*1024)
+}
+
+func BenchmarkCompressNoHash1M(b *testing.B) {
+ benchmark(b, true, false, 1024*1024)
+}
+
+func BenchmarkCompressHash1M(b *testing.B) {
+ benchmark(b, true, true, 1024*1024)
+}
+
+func BenchmarkDecompressNoHash1M(b *testing.B) {
+ benchmark(b, false, false, 1024*1024)
+}
+
+func BenchmarkDecompressHash1M(b *testing.B) {
+ benchmark(b, false, true, 1024*1024)
+}
+
+func BenchmarkCompressNoHash16M(b *testing.B) {
+ benchmark(b, true, false, 16*1024*1024)
+}
+
+func BenchmarkCompressHash16M(b *testing.B) {
+ benchmark(b, true, true, 16*1024*1024)
+}
+
+func BenchmarkDecompressNoHash16M(b *testing.B) {
+ benchmark(b, false, false, 16*1024*1024)
+}
+
+func BenchmarkDecompressHash16M(b *testing.B) {
+ benchmark(b, false, true, 16*1024*1024)
+}