package storage import ( "bytes" "compress/gzip" "io" "io/ioutil" "os" "testing" ) func TestDuplicateFail(t *testing.T) { e := []Entry{ Entry{ Type: FileType, Name: "./hurr.txt", Payload: []byte("abcde"), }, Entry{ Type: FileType, Name: "./hurr.txt", Payload: []byte("deadbeef"), }, Entry{ Type: FileType, Name: "hurr.txt", // slightly different path, same file though Payload: []byte("deadbeef"), }, } buf := []byte{} b := bytes.NewBuffer(buf) jp := NewJSONPacker(b) if _, err := jp.AddEntry(e[0]); err != nil { t.Error(err) } if _, err := jp.AddEntry(e[1]); err != ErrDuplicatePath { t.Errorf("expected failure on duplicate path") } if _, err := jp.AddEntry(e[2]); err != ErrDuplicatePath { t.Errorf("expected failure on duplicate path") } } func TestJSONPackerUnpacker(t *testing.T) { e := []Entry{ Entry{ Type: SegmentType, Payload: []byte("how"), }, Entry{ Type: SegmentType, Payload: []byte("y'all"), }, Entry{ Type: FileType, Name: "./hurr.txt", Payload: []byte("deadbeef"), }, Entry{ Type: SegmentType, Payload: []byte("doin"), }, } buf := []byte{} b := bytes.NewBuffer(buf) func() { jp := NewJSONPacker(b) for i := range e { if _, err := jp.AddEntry(e[i]); err != nil { t.Error(err) } } }() // >> packer_test.go:43: uncompressed: 266 //t.Errorf("uncompressed: %d", len(b.Bytes())) b = bytes.NewBuffer(b.Bytes()) entries := Entries{} func() { jup := NewJSONUnpacker(b) for { entry, err := jup.Next() if err != nil { if err == io.EOF { break } t.Error(err) } entries = append(entries, *entry) t.Logf("got %#v", entry) } }() if len(entries) != len(e) { t.Errorf("expected %d entries, got %d", len(e), len(entries)) } } // you can use a compress Reader/Writer and make nice savings. // // For these two tests that are using the same set, it the difference of 266 // bytes uncompressed vs 138 bytes compressed. func TestGzip(t *testing.T) { e := []Entry{ Entry{ Type: SegmentType, Payload: []byte("how"), }, Entry{ Type: SegmentType, Payload: []byte("y'all"), }, Entry{ Type: FileType, Name: "./hurr.txt", Payload: []byte("deadbeef"), }, Entry{ Type: SegmentType, Payload: []byte("doin"), }, } buf := []byte{} b := bytes.NewBuffer(buf) gzW := gzip.NewWriter(b) jp := NewJSONPacker(gzW) for i := range e { if _, err := jp.AddEntry(e[i]); err != nil { t.Error(err) } } gzW.Close() // >> packer_test.go:99: compressed: 138 //t.Errorf("compressed: %d", len(b.Bytes())) b = bytes.NewBuffer(b.Bytes()) gzR, err := gzip.NewReader(b) if err != nil { t.Fatal(err) } entries := Entries{} func() { jup := NewJSONUnpacker(gzR) for { entry, err := jup.Next() if err != nil { if err == io.EOF { break } t.Error(err) } entries = append(entries, *entry) t.Logf("got %#v", entry) } }() if len(entries) != len(e) { t.Errorf("expected %d entries, got %d", len(e), len(entries)) } } func BenchmarkGetPut(b *testing.B) { e := []Entry{ Entry{ Type: SegmentType, Payload: []byte("how"), }, Entry{ Type: SegmentType, Payload: []byte("y'all"), }, Entry{ Type: FileType, Name: "./hurr.txt", Payload: []byte("deadbeef"), }, Entry{ Type: SegmentType, Payload: []byte("doin"), }, } b.RunParallel(func(pb *testing.PB) { for pb.Next() { func() { fh, err := ioutil.TempFile("", "tar-split.") if err != nil { b.Fatal(err) } defer os.Remove(fh.Name()) defer fh.Close() jp := NewJSONPacker(fh) for i := range e { if _, err := jp.AddEntry(e[i]); err != nil { b.Fatal(err) } } fh.Sync() up := NewJSONUnpacker(fh) for { _, err := up.Next() if err != nil { if err == io.EOF { break } b.Fatal(err) } } }() } }) }