1
0
Fork 0
mirror of https://github.com/vbatts/go-mtree.git synced 2024-11-25 17:45:39 +00:00

Merge pull request #92 from vbatts/keyword

Keyword overhaul
This commit is contained in:
Vincent Batts 2016-11-17 21:58:04 -05:00 committed by GitHub
commit f51a3381de
21 changed files with 820 additions and 648 deletions

View file

@ -3,10 +3,10 @@ BUILD := gomtree
CWD := $(shell pwd) CWD := $(shell pwd)
SOURCE_FILES := $(shell find . -type f -name "*.go") SOURCE_FILES := $(shell find . -type f -name "*.go")
default: validation build default: build validation
.PHONY: validation .PHONY: validation
validation: test lint vet .cli.test validation: .test .lint .vet .cli.test
.PHONY: test .PHONY: test
test: .test test: .test

View file

@ -6,16 +6,20 @@ package mtree
// //
// This is equivalent to creating a new DirectoryHierarchy with Walk(root, nil, // This is equivalent to creating a new DirectoryHierarchy with Walk(root, nil,
// keywords) and then doing a Compare(dh, newDh, keywords). // keywords) and then doing a Compare(dh, newDh, keywords).
func Check(root string, dh *DirectoryHierarchy, keywords []string) ([]InodeDelta, error) { func Check(root string, dh *DirectoryHierarchy, keywords []Keyword) ([]InodeDelta, error) {
if keywords == nil { if keywords == nil {
keywords = CollectUsedKeywords(dh) used := dh.UsedKeywords()
newDh, err := Walk(root, nil, used)
if err != nil {
return nil, err
}
return Compare(dh, newDh, used)
} }
newDh, err := Walk(root, nil, keywords) newDh, err := Walk(root, nil, keywords)
if err != nil { if err != nil {
return nil, err return nil, err
} }
// TODO: Handle tar_time, if necessary. // TODO: Handle tar_time, if necessary.
return Compare(dh, newDh, keywords) return Compare(dh, newDh, keywords)
} }
@ -23,9 +27,9 @@ func Check(root string, dh *DirectoryHierarchy, keywords []string) ([]InodeDelta
// TarCheck is the tar equivalent of checking a file hierarchy spec against a // TarCheck is the tar equivalent of checking a file hierarchy spec against a
// tar stream to determine if files have been changed. This is precisely // tar stream to determine if files have been changed. This is precisely
// equivalent to Compare(dh, tarDH, keywords). // equivalent to Compare(dh, tarDH, keywords).
func TarCheck(tarDH, dh *DirectoryHierarchy, keywords []string) ([]InodeDelta, error) { func TarCheck(tarDH, dh *DirectoryHierarchy, keywords []Keyword) ([]InodeDelta, error) {
if keywords == nil { if keywords == nil {
keywords = CollectUsedKeywords(dh) return Compare(dh, tarDH, dh.UsedKeywords())
} }
return Compare(dh, tarDH, keywords) return Compare(dh, tarDH, keywords)
} }

View file

@ -69,14 +69,14 @@ func TestCheckKeywords(t *testing.T) {
t.Fatal(err) t.Fatal(err)
} }
if len(res) != 1 { if len(res) != 1 {
t.Errorf("expected to get 1 delta on changed mtimes, but did not") t.Fatal("expected to get 1 delta on changed mtimes, but did not")
} }
if res[0].Type() != Modified { if res[0].Type() != Modified {
t.Errorf("expected to get modified delta on changed mtimes, but did not") t.Errorf("expected to get modified delta on changed mtimes, but did not")
} }
// Check again, but only sha1 and mode. This ought to pass. // Check again, but only sha1 and mode. This ought to pass.
res, err = Check(dir, dh, []string{"sha1", "mode"}) res, err = Check(dir, dh, []Keyword{"sha1", "mode"})
if err != nil { if err != nil {
t.Fatal(err) t.Fatal(err)
} }
@ -200,7 +200,7 @@ func TestTarTime(t *testing.T) {
t.Fatal(err) t.Fatal(err)
} }
keywords := CollectUsedKeywords(dh) keywords := dh.UsedKeywords()
// make sure "time" keyword works // make sure "time" keyword works
_, err = Check(dir, dh, keywords) _, err = Check(dir, dh, keywords)

View file

@ -29,6 +29,251 @@ var (
flVersion = flag.Bool("version", false, "display the version of this tool") flVersion = flag.Bool("version", false, "display the version of this tool")
) )
func main() {
// so that defers cleanly exec
if err := app(); err != nil {
log.Fatal(err)
}
}
func app() error {
flag.Parse()
if *flDebug {
os.Setenv("DEBUG", "1")
}
if *flVersion {
fmt.Printf("%s :: %s\n", mtree.AppName, mtree.Version)
return nil
}
// -list-keywords
if *flListKeywords {
fmt.Println("Available keywords:")
for k := range mtree.KeywordFuncs {
fmt.Print(" ")
fmt.Print(k)
if mtree.Keyword(k).Default() {
fmt.Print(" (default)")
}
if !mtree.Keyword(k).Bsd() {
fmt.Print(" (not upstream)")
}
fmt.Print("\n")
}
return nil
}
// --result-format
formatFunc, ok := formats[*flResultFormat]
if !ok {
return fmt.Errorf("invalid output format: %s", *flResultFormat)
}
var (
err error
tmpKeywords []mtree.Keyword
currentKeywords []mtree.Keyword
)
// -k <keywords>
if *flUseKeywords != "" {
tmpKeywords = splitKeywordsArg(*flUseKeywords)
if !mtree.InKeywordSlice("type", tmpKeywords) {
tmpKeywords = append([]mtree.Keyword{"type"}, tmpKeywords...)
}
} else {
if *flTar != "" {
tmpKeywords = mtree.DefaultTarKeywords[:]
} else {
tmpKeywords = mtree.DefaultKeywords[:]
}
}
// -K <keywords>
if *flAddKeywords != "" {
for _, kw := range splitKeywordsArg(*flAddKeywords) {
if !mtree.InKeywordSlice(kw, tmpKeywords) {
tmpKeywords = append(tmpKeywords, kw)
}
}
}
// -bsd-keywords
if *flBsdKeywords {
for _, k := range tmpKeywords {
if mtree.Keyword(k).Bsd() {
currentKeywords = append(currentKeywords, k)
} else {
fmt.Fprintf(os.Stderr, "INFO: ignoring %q as it is not an upstream keyword\n", k)
}
}
} else {
currentKeywords = tmpKeywords
}
// Check mutual exclusivity of keywords.
// TODO(cyphar): Abstract this inside keywords.go.
if mtree.InKeywordSlice("tar_time", currentKeywords) && mtree.InKeywordSlice("time", currentKeywords) {
return fmt.Errorf("tar_time and time are mutually exclusive keywords")
}
// If we're doing a comparison, we always are comparing between a spec and
// state DH. If specDh is nil, we are generating a new one.
var (
specDh *mtree.DirectoryHierarchy
stateDh *mtree.DirectoryHierarchy
specKeywords []mtree.Keyword
)
// -f <file>
if *flFile != "" && !*flCreate {
// load the hierarchy, if we're not creating a new spec
fh, err := os.Open(*flFile)
if err != nil {
return err
}
specDh, err = mtree.ParseSpec(fh)
fh.Close()
if err != nil {
return err
}
// We can't check against more fields than in the specKeywords list, so
// currentKeywords can only have a subset of specKeywords.
specKeywords = specDh.UsedKeywords()
}
// -list-used
if *flListUsedKeywords {
if specDh == nil {
return fmt.Errorf("no specification provided. please provide a validation manifest")
}
if *flResultFormat == "json" {
// if they're asking for json, give it to them
data := map[string][]mtree.Keyword{*flFile: specKeywords}
buf, err := json.MarshalIndent(data, "", " ")
if err != nil {
return err
}
fmt.Println(string(buf))
} else {
fmt.Printf("Keywords used in [%s]:\n", *flFile)
for _, kw := range specKeywords {
fmt.Printf(" %s", kw)
if _, ok := mtree.KeywordFuncs[kw]; !ok {
fmt.Print(" (unsupported)")
}
fmt.Printf("\n")
}
}
return nil
}
if specKeywords != nil {
// If we didn't actually change the set of keywords, we can just use specKeywords.
if *flUseKeywords == "" && *flAddKeywords == "" {
currentKeywords = specKeywords
}
for _, keyword := range currentKeywords {
// As always, time is a special case.
// TODO: Fix that.
if (keyword == "time" && mtree.InKeywordSlice("tar_time", specKeywords)) || (keyword == "tar_time" && mtree.InKeywordSlice("time", specKeywords)) {
continue
}
}
}
// -p and -T are mutually exclusive
if *flPath != "" && *flTar != "" {
return fmt.Errorf("options -T and -p are mutually exclusive")
}
// -p <path>
var rootPath = "."
if *flPath != "" {
rootPath = *flPath
}
// -T <tar file>
if *flTar != "" {
var input io.Reader
if *flTar == "-" {
input = os.Stdin
} else {
fh, err := os.Open(*flTar)
if err != nil {
return err
}
defer fh.Close()
input = fh
}
ts := mtree.NewTarStreamer(input, currentKeywords)
if _, err := io.Copy(ioutil.Discard, ts); err != nil && err != io.EOF {
return err
}
if err := ts.Close(); err != nil {
return err
}
var err error
stateDh, err = ts.Hierarchy()
if err != nil {
return err
}
} else {
// with a root directory
stateDh, err = mtree.Walk(rootPath, nil, currentKeywords)
if err != nil {
return err
}
}
// -c
if *flCreate {
fh := os.Stdout
if *flFile != "" {
fh, err = os.Create(*flFile)
if err != nil {
return err
}
}
// output stateDh
stateDh.WriteTo(fh)
return nil
}
// This is a validation.
if specDh != nil && stateDh != nil {
var res []mtree.InodeDelta
res, err = mtree.Compare(specDh, stateDh, currentKeywords)
if err != nil {
return err
}
if res != nil {
if isTarSpec(specDh) || *flTar != "" {
res = filterMissingKeywords(res)
}
//if len(res) > 0 {
//return fmt.Errorf("unexpected missing keywords: %d", len(res))
//}
out := formatFunc(res)
if _, err := os.Stdout.Write([]byte(out)); err != nil {
return err
}
}
} else {
return fmt.Errorf("neither validating or creating a manifest. Please provide additional arguments")
}
return nil
}
var formats = map[string]func([]mtree.InodeDelta) string{ var formats = map[string]func([]mtree.InodeDelta) string{
// Outputs the errors in the BSD format. // Outputs the errors in the BSD format.
"bsd": func(d []mtree.InodeDelta) string { "bsd": func(d []mtree.InodeDelta) string {
@ -146,301 +391,10 @@ func isTarSpec(spec *mtree.DirectoryHierarchy) bool {
return false return false
} }
func main() { func splitKeywordsArg(str string) []mtree.Keyword {
flag.Parse() keywords := []mtree.Keyword{}
for _, kw := range strings.Fields(strings.Replace(str, ",", " ", -1)) {
if *flDebug { keywords = append(keywords, mtree.KeywordSynonym(kw))
os.Setenv("DEBUG", "1")
}
// so that defers cleanly exec
// TODO: Switch everything to being inside a function, to remove the need for isErr.
var isErr bool
defer func() {
if isErr {
os.Exit(1)
}
}()
if *flVersion {
fmt.Printf("%s :: %s\n", os.Args[0], mtree.Version)
return
}
// -list-keywords
if *flListKeywords {
fmt.Println("Available keywords:")
for k := range mtree.KeywordFuncs {
fmt.Print(" ")
fmt.Print(k)
if mtree.Keyword(k).Default() {
fmt.Print(" (default)")
}
if !mtree.Keyword(k).Bsd() {
fmt.Print(" (not upstream)")
}
fmt.Print("\n")
}
return
}
// --result-format
formatFunc, ok := formats[*flResultFormat]
if !ok {
log.Printf("invalid output format: %s", *flResultFormat)
isErr = true
return
}
var (
err error
tmpKeywords []string
currentKeywords []string
)
// -k <keywords>
if *flUseKeywords != "" {
tmpKeywords = splitKeywordsArg(*flUseKeywords)
if !inSlice("type", tmpKeywords) {
tmpKeywords = append([]string{"type"}, tmpKeywords...)
}
} else {
if *flTar != "" {
tmpKeywords = mtree.DefaultTarKeywords[:]
} else {
tmpKeywords = mtree.DefaultKeywords[:]
}
}
// -K <keywords>
if *flAddKeywords != "" {
for _, kw := range splitKeywordsArg(*flAddKeywords) {
if !inSlice(kw, tmpKeywords) {
tmpKeywords = append(tmpKeywords, kw)
}
}
}
// -bsd-keywords
if *flBsdKeywords {
for _, k := range tmpKeywords {
if mtree.Keyword(k).Bsd() {
currentKeywords = append(currentKeywords, k)
} else {
fmt.Fprintf(os.Stderr, "INFO: ignoring %q as it is not an upstream keyword\n", k)
}
}
} else {
currentKeywords = tmpKeywords
}
// Check mutual exclusivity of keywords.
// TODO(cyphar): Abstract this inside keywords.go.
if inSlice("tar_time", currentKeywords) && inSlice("time", currentKeywords) {
log.Printf("tar_time and time are mutually exclusive keywords")
isErr = true
return
}
// If we're doing a comparison, we always are comparing between a spec and
// state DH. If specDh is nil, we are generating a new one.
var (
specDh *mtree.DirectoryHierarchy
stateDh *mtree.DirectoryHierarchy
specKeywords []string
)
// -f <file>
if *flFile != "" && !*flCreate {
// load the hierarchy, if we're not creating a new spec
fh, err := os.Open(*flFile)
if err != nil {
log.Println(err)
isErr = true
return
}
specDh, err = mtree.ParseSpec(fh)
fh.Close()
if err != nil {
log.Println(err)
isErr = true
return
}
// We can't check against more fields than in the specKeywords list, so
// currentKeywords can only have a subset of specKeywords.
specKeywords = mtree.CollectUsedKeywords(specDh)
}
// -list-used
if *flListUsedKeywords {
if specDh == nil {
log.Println("no specification provided. please provide a validation manifest")
isErr = true
return
}
if *flResultFormat == "json" {
// if they're asking for json, give it to them
data := map[string][]string{*flFile: specKeywords}
buf, err := json.MarshalIndent(data, "", " ")
if err != nil {
defer os.Exit(1)
isErr = true
return
}
fmt.Println(string(buf))
} else {
fmt.Printf("Keywords used in [%s]:\n", *flFile)
for _, kw := range specKeywords {
fmt.Printf(" %s", kw)
if _, ok := mtree.KeywordFuncs[kw]; !ok {
fmt.Print(" (unsupported)")
}
fmt.Printf("\n")
}
}
return
}
if specKeywords != nil {
// If we didn't actually change the set of keywords, we can just use specKeywords.
if *flUseKeywords == "" && *flAddKeywords == "" {
currentKeywords = specKeywords
}
for _, keyword := range currentKeywords {
// As always, time is a special case.
// TODO: Fix that.
if (keyword == "time" && inSlice("tar_time", specKeywords)) || (keyword == "tar_time" && inSlice("time", specKeywords)) {
continue
}
if !inSlice(keyword, specKeywords) {
log.Printf("cannot verify keywords not in mtree specification: %s\n", keyword)
isErr = true
}
if isErr {
return
}
}
}
// -p and -T are mutually exclusive
if *flPath != "" && *flTar != "" {
log.Println("options -T and -p are mutually exclusive")
isErr = true
return
}
// -p <path>
var rootPath = "."
if *flPath != "" {
rootPath = *flPath
}
// -T <tar file>
if *flTar != "" {
var input io.Reader
if *flTar == "-" {
input = os.Stdin
} else {
fh, err := os.Open(*flTar)
if err != nil {
log.Println(err)
isErr = true
return
}
defer fh.Close()
input = fh
}
ts := mtree.NewTarStreamer(input, currentKeywords)
if _, err := io.Copy(ioutil.Discard, ts); err != nil && err != io.EOF {
log.Println(err)
isErr = true
return
}
if err := ts.Close(); err != nil {
log.Println(err)
isErr = true
return
}
var err error
stateDh, err = ts.Hierarchy()
if err != nil {
log.Println(err)
isErr = true
return
}
} else {
// with a root directory
stateDh, err = mtree.Walk(rootPath, nil, currentKeywords)
if err != nil {
log.Println(err)
isErr = true
return
}
}
// -c
if *flCreate {
fh := os.Stdout
if *flFile != "" {
fh, err = os.Create(*flFile)
if err != nil {
log.Println(err)
isErr = true
return
}
}
// output stateDh
stateDh.WriteTo(fh)
return
}
// This is a validation.
if specDh != nil && stateDh != nil {
var res []mtree.InodeDelta
res, err = mtree.Compare(specDh, stateDh, currentKeywords)
if err != nil {
log.Println(err)
isErr = true
return
}
if res != nil {
if isTarSpec(specDh) || *flTar != "" {
res = filterMissingKeywords(res)
}
if len(res) > 0 {
defer os.Exit(1)
}
out := formatFunc(res)
if _, err := os.Stdout.Write([]byte(out)); err != nil {
log.Println(err)
isErr = true
return
}
}
} else {
log.Println("neither validating or creating a manifest. Please provide additional arguments")
isErr = true
return
} }
} return keywords
func splitKeywordsArg(str string) []string {
return strings.Fields(strings.Replace(str, ",", " ", -1))
}
func inSlice(a string, list []string) bool {
for _, b := range list {
if b == a {
return true
}
}
return false
} }

View file

@ -123,7 +123,7 @@ func (i InodeDelta) String() string {
// returned with InodeDelta.Diff(). // returned with InodeDelta.Diff().
type KeyDelta struct { type KeyDelta struct {
diff DifferenceType diff DifferenceType
name string name Keyword
old string old string
new string new string
} }
@ -136,7 +136,7 @@ func (k KeyDelta) Type() DifferenceType {
// Name returns the name (the key) of the KeyDeltaVal entry in the // Name returns the name (the key) of the KeyDeltaVal entry in the
// DirectoryHierarchy. // DirectoryHierarchy.
func (k KeyDelta) Name() string { func (k KeyDelta) Name() Keyword {
return k.name return k.name
} }
@ -164,7 +164,7 @@ func (k KeyDelta) New() *string {
func (k KeyDelta) MarshalJSON() ([]byte, error) { func (k KeyDelta) MarshalJSON() ([]byte, error) {
return json.Marshal(struct { return json.Marshal(struct {
Type DifferenceType `json:"type"` Type DifferenceType `json:"type"`
Name string `json:"name"` Name Keyword `json:"name"`
Old string `json:"old"` Old string `json:"old"`
New string `json:"new"` New string `json:"new"`
}{ }{
@ -184,11 +184,17 @@ func compareEntry(oldEntry, newEntry Entry) ([]KeyDelta, error) {
New *KeyVal New *KeyVal
} }
diffs := map[string]*stateT{} diffs := map[Keyword]*stateT{}
oldKeys := oldEntry.AllKeys()
newKeys := newEntry.AllKeys()
// Fill the map with the old keys first. // Fill the map with the old keys first.
for _, kv := range oldEntry.AllKeys() { for _, kv := range oldKeys {
key := kv.Keyword() key := kv.Keyword()
// only add this diff if the new keys has this keyword
if key != "tar_time" && key != "time" && HasKeyword(newKeys, key) == emptyKV {
continue
}
// Cannot take &kv because it's the iterator. // Cannot take &kv because it's the iterator.
copy := new(KeyVal) copy := new(KeyVal)
@ -202,8 +208,12 @@ func compareEntry(oldEntry, newEntry Entry) ([]KeyDelta, error) {
} }
// Then fill the new keys. // Then fill the new keys.
for _, kv := range newEntry.AllKeys() { for _, kv := range newKeys {
key := kv.Keyword() key := kv.Keyword()
// only add this diff if the old keys has this keyword
if key != "tar_time" && key != "time" && HasKeyword(oldKeys, key) == emptyKV {
continue
}
// Cannot take &kv because it's the iterator. // Cannot take &kv because it's the iterator.
copy := new(KeyVal) copy := new(KeyVal)
@ -218,7 +228,7 @@ func compareEntry(oldEntry, newEntry Entry) ([]KeyDelta, error) {
// We need a full list of the keys so we can deal with different keyvalue // We need a full list of the keys so we can deal with different keyvalue
// orderings. // orderings.
var kws []string var kws []Keyword
for kw := range diffs { for kw := range diffs {
kws = append(kws, kw) kws = append(kws, kw)
} }
@ -226,7 +236,7 @@ func compareEntry(oldEntry, newEntry Entry) ([]KeyDelta, error) {
// If both tar_time and time were specified in the set of keys, we have to // If both tar_time and time were specified in the set of keys, we have to
// mess with the diffs. This is an unfortunate side-effect of tar archives. // mess with the diffs. This is an unfortunate side-effect of tar archives.
// TODO(cyphar): This really should be abstracted inside keywords.go // TODO(cyphar): This really should be abstracted inside keywords.go
if inSlice("tar_time", kws) && inSlice("time", kws) { if InKeywordSlice("tar_time", kws) && InKeywordSlice("time", kws) {
// Delete "time". // Delete "time".
timeStateT := diffs["time"] timeStateT := diffs["time"]
delete(diffs, "time") delete(diffs, "time")
@ -312,19 +322,13 @@ func compareEntry(oldEntry, newEntry Entry) ([]KeyDelta, error) {
// //
// NB: The order of the parameters matters (old, new) because Extra and // NB: The order of the parameters matters (old, new) because Extra and
// Missing are considered as different discrepancy types. // Missing are considered as different discrepancy types.
func Compare(oldDh, newDh *DirectoryHierarchy, keys []string) ([]InodeDelta, error) { func Compare(oldDh, newDh *DirectoryHierarchy, keys []Keyword) ([]InodeDelta, error) {
// Represents the new and old states for an entry. // Represents the new and old states for an entry.
type stateT struct { type stateT struct {
Old *Entry Old *Entry
New *Entry New *Entry
} }
// Make dealing with the keys mapping easier.
keySet := map[string]struct{}{}
for _, key := range keys {
keySet[key] = struct{}{}
}
// To deal with different orderings of the entries, use a path-keyed // To deal with different orderings of the entries, use a path-keyed
// map to make sure we don't start comparing unrelated entries. // map to make sure we don't start comparing unrelated entries.
diffs := map[string]*stateT{} diffs := map[string]*stateT{}
@ -405,7 +409,7 @@ func Compare(oldDh, newDh *DirectoryHierarchy, keys []string) ([]InodeDelta, err
if keys != nil { if keys != nil {
var filterChanged []KeyDelta var filterChanged []KeyDelta
for _, keyDiff := range changed { for _, keyDiff := range changed {
if _, ok := keySet[keyDiff.name]; ok { if InKeywordSlice(keyDiff.name, keys) {
filterChanged = append(filterChanged, keyDiff) filterChanged = append(filterChanged, keyDiff)
} }
} }

View file

@ -325,7 +325,7 @@ func TestCompareKeys(t *testing.T) {
} }
// Compare. // Compare.
diffs, err := Compare(old, new, []string{"size"}) diffs, err := Compare(old, new, []Keyword{"size"})
if err != nil { if err != nil {
t.Fatal(err) t.Fatal(err)
} }

View file

@ -21,7 +21,7 @@ type Entry struct {
Pos int // order in the spec Pos int // order in the spec
Raw string // file or directory name Raw string // file or directory name
Name string // file or directory name Name string // file or directory name
Keywords []string // TODO(vbatts) maybe a keyword typed set of values? Keywords []KeyVal // TODO(vbatts) maybe a keyword typed set of values?
Type EntryType Type EntryType
} }
@ -94,23 +94,20 @@ func (e Entry) String() string {
if e.Type == DotDotType { if e.Type == DotDotType {
return e.Name return e.Name
} }
if e.Type == SpecialType || e.Type == FullType || inSlice("type=dir", e.Keywords) { if e.Type == SpecialType || e.Type == FullType || inKeyValSlice("type=dir", e.Keywords) {
return fmt.Sprintf("%s %s", e.Name, strings.Join(e.Keywords, " ")) return fmt.Sprintf("%s %s", e.Name, strings.Join(KeyValToString(e.Keywords), " "))
} }
return fmt.Sprintf(" %s %s", e.Name, strings.Join(e.Keywords, " ")) return fmt.Sprintf(" %s %s", e.Name, strings.Join(KeyValToString(e.Keywords), " "))
} }
// AllKeys returns the full set of KeyVals for the given entry, based on the // AllKeys returns the full set of KeyVal for the given entry, based on the
// /set keys as well as the entry-local keys. Entry-local keys always take // /set keys as well as the entry-local keys. Entry-local keys always take
// precedence. // precedence.
func (e Entry) AllKeys() KeyVals { func (e Entry) AllKeys() []KeyVal {
var kv KeyVals
if e.Set != nil { if e.Set != nil {
kv = MergeSet(e.Set.Keywords, e.Keywords) return MergeKeyValSet(e.Set.Keywords, e.Keywords)
} else {
kv = NewKeyVals(e.Keywords)
} }
return kv return e.Keywords
} }
// EntryType are the formats of lines in an mtree spec file // EntryType are the formats of lines in an mtree spec file

View file

@ -26,26 +26,23 @@ func (dh DirectoryHierarchy) WriteTo(w io.Writer) (n int64, err error) {
return sum, nil return sum, nil
} }
// CollectUsedKeywords collects and returns all the keywords used in a // UsedKeywords collects and returns all the keywords used in a
// a DirectoryHierarchy // a DirectoryHierarchy
func CollectUsedKeywords(dh *DirectoryHierarchy) []string { func (dh DirectoryHierarchy) UsedKeywords() []Keyword {
if dh != nil { usedkeywords := []Keyword{}
usedkeywords := []string{} for _, e := range dh.Entries {
for _, e := range dh.Entries { switch e.Type {
switch e.Type { case FullType, RelativeType, SpecialType:
case FullType, RelativeType, SpecialType: if e.Type != SpecialType || e.Name == "/set" {
if e.Type != SpecialType || e.Name == "/set" { kvs := e.Keywords
kvs := e.Keywords for _, kv := range kvs {
for _, kv := range kvs { kw := KeyVal(kv).Keyword()
kw := KeyVal(kv).Keyword() if !InKeywordSlice(kw, usedkeywords) {
if !inSlice(kw, usedkeywords) { usedkeywords = append(usedkeywords, KeywordSynonym(string(kw)))
usedkeywords = append(usedkeywords, kw)
}
} }
} }
} }
} }
return usedkeywords
} }
return nil return usedkeywords
} }

38
hierarchy_test.go Normal file
View file

@ -0,0 +1,38 @@
package mtree
import (
"strings"
"testing"
)
var checklist = []struct {
blob string
set []Keyword
}{
{blob: `
# machine: bananaboat
# tree: .git
# date: Wed Nov 16 14:54:17 2016
# .
/set type=file nlink=1 mode=0664 uid=1000 gid=100
. size=4096 type=dir mode=0755 nlink=8 time=1479326055.423853146
.COMMIT_EDITMSG.un~ size=1006 mode=0644 time=1479325423.450468662 sha1digest=dead0face
.TAG_EDITMSG.un~ size=1069 mode=0600 time=1471362316.801317529 sha256digest=dead0face
`, set: []Keyword{"size", "mode", "time", "sha256digest"}},
}
func TestUsedKeywords(t *testing.T) {
for i, item := range checklist {
dh, err := ParseSpec(strings.NewReader(item.blob))
if err != nil {
t.Error(err)
}
used := dh.UsedKeywords()
for _, k := range item.set {
if !InKeywordSlice(k, used) {
t.Errorf("%d: expected to find %q in %q", i, k, used)
}
}
}
}

168
keywordfunc.go Normal file
View file

@ -0,0 +1,168 @@
package mtree
import (
"archive/tar"
"crypto/md5"
"crypto/sha1"
"crypto/sha256"
"crypto/sha512"
"fmt"
"hash"
"io"
"os"
"golang.org/x/crypto/ripemd160"
)
// KeywordFunc is the type of a function called on each file to be included in
// a DirectoryHierarchy, that will produce the string output of the keyword to
// be included for the file entry. Otherwise, empty string.
// io.Reader `r` is to the file stream for the file payload. While this
// function takes an io.Reader, the caller needs to reset it to the beginning
// for each new KeywordFunc
type KeywordFunc func(path string, info os.FileInfo, r io.Reader) (KeyVal, error)
var (
// KeywordFuncs is the map of all keywords (and the functions to produce them)
KeywordFuncs = map[Keyword]KeywordFunc{
"size": sizeKeywordFunc, // The size, in bytes, of the file
"type": typeKeywordFunc, // The type of the file
"time": timeKeywordFunc, // The last modification time of the file
"link": linkKeywordFunc, // The target of the symbolic link when type=link
"uid": uidKeywordFunc, // The file owner as a numeric value
"gid": gidKeywordFunc, // The file group as a numeric value
"nlink": nlinkKeywordFunc, // The number of hard links the file is expected to have
"uname": unameKeywordFunc, // The file owner as a symbolic name
"mode": modeKeywordFunc, // The current file's permissions as a numeric (octal) or symbolic value
"cksum": cksumKeywordFunc, // The checksum of the file using the default algorithm specified by the cksum(1) utility
"md5": hasherKeywordFunc("md5digest", md5.New), // The MD5 message digest of the file
"md5digest": hasherKeywordFunc("md5digest", md5.New), // A synonym for `md5`
"rmd160": hasherKeywordFunc("ripemd160digest", ripemd160.New), // The RIPEMD160 message digest of the file
"rmd160digest": hasherKeywordFunc("ripemd160digest", ripemd160.New), // A synonym for `rmd160`
"ripemd160digest": hasherKeywordFunc("ripemd160digest", ripemd160.New), // A synonym for `rmd160`
"sha1": hasherKeywordFunc("sha1digest", sha1.New), // The SHA1 message digest of the file
"sha1digest": hasherKeywordFunc("sha1digest", sha1.New), // A synonym for `sha1`
"sha256": hasherKeywordFunc("sha256digest", sha256.New), // The SHA256 message digest of the file
"sha256digest": hasherKeywordFunc("sha256digest", sha256.New), // A synonym for `sha256`
"sha384": hasherKeywordFunc("sha384digest", sha512.New384), // The SHA384 message digest of the file
"sha384digest": hasherKeywordFunc("sha384digest", sha512.New384), // A synonym for `sha384`
"sha512": hasherKeywordFunc("sha512digest", sha512.New), // The SHA512 message digest of the file
"sha512digest": hasherKeywordFunc("sha512digest", sha512.New), // A synonym for `sha512`
"flags": flagsKeywordFunc, // NOTE: this is a noop, but here to support the presence of the "flags" keyword.
// This is not an upstreamed keyword, but used to vary from "time", as tar
// archives do not store nanosecond precision. So comparing on "time" will
// be only seconds level accurate.
"tar_time": tartimeKeywordFunc, // The last modification time of the file, from a tar archive mtime
// This is not an upstreamed keyword, but a needed attribute for file validation.
// The pattern for this keyword key is prefixed by "xattr." followed by the extended attribute "namespace.key".
// The keyword value is the SHA1 digest of the extended attribute's value.
// In this way, the order of the keys does not matter, and the contents of the value is not revealed.
"xattr": xattrKeywordFunc,
"xattrs": xattrKeywordFunc,
}
)
var (
modeKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
permissions := info.Mode().Perm()
if os.ModeSetuid&info.Mode() > 0 {
permissions |= (1 << 11)
}
if os.ModeSetgid&info.Mode() > 0 {
permissions |= (1 << 10)
}
if os.ModeSticky&info.Mode() > 0 {
permissions |= (1 << 9)
}
return KeyVal(fmt.Sprintf("mode=%#o", permissions)), nil
}
sizeKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
if sys, ok := info.Sys().(*tar.Header); ok {
if sys.Typeflag == tar.TypeSymlink {
return KeyVal(fmt.Sprintf("size=%d", len(sys.Linkname))), nil
}
}
return KeyVal(fmt.Sprintf("size=%d", info.Size())), nil
}
cksumKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
if !info.Mode().IsRegular() {
return emptyKV, nil
}
sum, _, err := cksum(r)
if err != nil {
return emptyKV, err
}
return KeyVal(fmt.Sprintf("cksum=%d", sum)), nil
}
hasherKeywordFunc = func(name string, newHash func() hash.Hash) KeywordFunc {
return func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
if !info.Mode().IsRegular() {
return emptyKV, nil
}
h := newHash()
if _, err := io.Copy(h, r); err != nil {
return emptyKV, err
}
return KeyVal(fmt.Sprintf("%s=%x", KeywordSynonym(name), h.Sum(nil))), nil
}
}
tartimeKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
return KeyVal(fmt.Sprintf("tar_time=%d.%9.9d", info.ModTime().Unix(), 0)), nil
}
timeKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
tSec := info.ModTime().Unix()
tNano := info.ModTime().Nanosecond()
return KeyVal(fmt.Sprintf("time=%d.%9.9d", tSec, tNano)), nil
}
linkKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
if sys, ok := info.Sys().(*tar.Header); ok {
if sys.Linkname != "" {
linkname, err := Vis(sys.Linkname)
if err != nil {
return emptyKV, err
}
return KeyVal(fmt.Sprintf("link=%s", linkname)), nil
}
return emptyKV, nil
}
if info.Mode()&os.ModeSymlink != 0 {
str, err := os.Readlink(path)
if err != nil {
return emptyKV, err
}
linkname, err := Vis(str)
if err != nil {
return emptyKV, err
}
return KeyVal(fmt.Sprintf("link=%s", linkname)), nil
}
return emptyKV, nil
}
typeKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
if info.Mode().IsDir() {
return "type=dir", nil
}
if info.Mode().IsRegular() {
return "type=file", nil
}
if info.Mode()&os.ModeSocket != 0 {
return "type=socket", nil
}
if info.Mode()&os.ModeSymlink != 0 {
return "type=link", nil
}
if info.Mode()&os.ModeNamedPipe != 0 {
return "type=fifo", nil
}
if info.Mode()&os.ModeDevice != 0 {
if info.Mode()&os.ModeCharDevice != 0 {
return "type=char", nil
}
return "type=device", nil
}
return emptyKV, nil
}
)

View file

@ -1,55 +1,97 @@
package mtree package mtree
import ( import (
"archive/tar"
"crypto/md5"
"crypto/sha1"
"crypto/sha256"
"crypto/sha512"
"fmt" "fmt"
"hash"
"io"
"os"
"strings" "strings"
"golang.org/x/crypto/ripemd160"
) )
// KeywordFunc is the type of a function called on each file to be included in
// a DirectoryHierarchy, that will produce the string output of the keyword to
// be included for the file entry. Otherwise, empty string.
// io.Reader `r` is to the file stream for the file payload. While this
// function takes an io.Reader, the caller needs to reset it to the beginning
// for each new KeywordFunc
type KeywordFunc func(path string, info os.FileInfo, r io.Reader) (string, error)
// Keyword is the string name of a keyword, with some convenience functions for // Keyword is the string name of a keyword, with some convenience functions for
// determining whether it is a default or bsd standard keyword. // determining whether it is a default or bsd standard keyword.
type Keyword string type Keyword string
// Default returns whether this keyword is in the default set of keywords // Default returns whether this keyword is in the default set of keywords
func (k Keyword) Default() bool { func (k Keyword) Default() bool {
return inSlice(string(k), DefaultKeywords) return InKeywordSlice(k, DefaultKeywords)
} }
// Bsd returns whether this keyword is in the upstream FreeBSD mtree(8) // Bsd returns whether this keyword is in the upstream FreeBSD mtree(8)
func (k Keyword) Bsd() bool { func (k Keyword) Bsd() bool {
return inSlice(string(k), BsdKeywords) return InKeywordSlice(k, BsdKeywords)
}
// Synonym returns the canonical name for this keyword. This is provides the
// same functionality as KeywordSynonym()
func (k Keyword) Synonym() Keyword {
return KeywordSynonym(string(k))
}
// InKeywordSlice checks for the presence of `a` in `list`
func InKeywordSlice(a Keyword, list []Keyword) bool {
for _, b := range list {
if b == a {
return true
}
}
return false
}
func inKeyValSlice(a KeyVal, list []KeyVal) bool {
for _, b := range list {
if b == a {
return true
}
}
return false
}
// ToKeywords makes a list of Keyword from a list of string
func ToKeywords(list []string) []Keyword {
ret := make([]Keyword, len(list))
for i := range list {
ret[i] = Keyword(list[i])
}
return ret
}
// FromKeywords makes a list of string from a list of Keyword
func FromKeywords(list []Keyword) []string {
ret := make([]string, len(list))
for i := range list {
ret[i] = string(list[i])
}
return ret
}
// KeyValToString constructs a list of string from the list of KeyVal
func KeyValToString(list []KeyVal) []string {
ret := make([]string, len(list))
for i := range list {
ret[i] = string(list[i])
}
return ret
}
// StringToKeyVals constructs a list of KeyVal from the list of strings, like "keyword=value"
func StringToKeyVals(list []string) []KeyVal {
ret := make([]KeyVal, len(list))
for i := range list {
ret[i] = KeyVal(list[i])
}
return ret
} }
// KeyVal is a "keyword=value" // KeyVal is a "keyword=value"
type KeyVal string type KeyVal string
// Keyword is the mapping to the available keywords // Keyword is the mapping to the available keywords
func (kv KeyVal) Keyword() string { func (kv KeyVal) Keyword() Keyword {
if !strings.Contains(string(kv), "=") { if !strings.Contains(string(kv), "=") {
return "" return Keyword("")
} }
chunks := strings.SplitN(strings.TrimSpace(string(kv)), "=", 2)[0] chunks := strings.SplitN(strings.TrimSpace(string(kv)), "=", 2)[0]
if !strings.Contains(chunks, ".") { if !strings.Contains(chunks, ".") {
return chunks return Keyword(chunks)
} }
return strings.SplitN(chunks, ".", 2)[0] return Keyword(strings.SplitN(chunks, ".", 2)[0])
} }
// KeywordSuffix is really only used for xattr, as the keyword is a prefix to // KeywordSuffix is really only used for xattr, as the keyword is a prefix to
@ -78,7 +120,7 @@ func (kv KeyVal) ChangeValue(newval string) string {
return fmt.Sprintf("%s=%s", kv.Keyword(), newval) return fmt.Sprintf("%s=%s", kv.Keyword(), newval)
} }
// KeyValEqual returns whether two KeyVals are equivalent. This takes // KeyValEqual returns whether two KeyVal are equivalent. This takes
// care of certain odd cases such as tar_mtime, and should be used over // care of certain odd cases such as tar_mtime, and should be used over
// using == comparisons directly unless you really know what you're // using == comparisons directly unless you really know what you're
// doing. // doing.
@ -87,35 +129,49 @@ func KeyValEqual(a, b KeyVal) bool {
return a.Keyword() == b.Keyword() && a.Value() == b.Value() return a.Keyword() == b.Keyword() && a.Value() == b.Value()
} }
// keywordSelector takes an array of "keyword=value" and filters out that only the set of words // keyvalSelector takes an array of KeyVal ("keyword=value") and filters out that only the set of keywords
func keywordSelector(keyval, words []string) []string { func keyvalSelector(keyval []KeyVal, keyset []Keyword) []KeyVal {
retList := []string{} retList := []KeyVal{}
for _, kv := range keyval { for _, kv := range keyval {
if inSlice(KeyVal(kv).Keyword(), words) { if InKeywordSlice(kv.Keyword(), keyset) {
retList = append(retList, kv) retList = append(retList, kv)
} }
} }
return retList return retList
} }
// NewKeyVals constructs a list of KeyVal from the list of strings, like "keyword=value" func keyValDifference(this, that []KeyVal) []KeyVal {
func NewKeyVals(keyvals []string) KeyVals { if len(this) == 0 {
kvs := make(KeyVals, len(keyvals)) return that
for i := range keyvals {
kvs[i] = KeyVal(keyvals[i])
} }
return kvs diff := []KeyVal{}
for _, kv := range this {
if !inKeyValSlice(kv, that) {
diff = append(diff, kv)
}
}
return diff
}
func keyValCopy(set []KeyVal) []KeyVal {
ret := make([]KeyVal, len(set))
for i := range set {
ret[i] = set[i]
}
return ret
} }
// KeyVals is a list of KeyVal
type KeyVals []KeyVal
// Has the "keyword" present in the list of KeyVal, and returns the // Has the "keyword" present in the list of KeyVal, and returns the
// corresponding KeyVal, else an empty string. // corresponding KeyVal, else an empty string.
func (kvs KeyVals) Has(keyword string) KeyVal { func Has(keyvals []KeyVal, keyword string) KeyVal {
for i := range kvs { return HasKeyword(keyvals, Keyword(keyword))
if kvs[i].Keyword() == keyword { }
return kvs[i]
// HasKeyword the "keyword" present in the list of KeyVal, and returns the
// corresponding KeyVal, else an empty string.
func HasKeyword(keyvals []KeyVal, keyword Keyword) KeyVal {
for i := range keyvals {
if keyvals[i].Keyword() == keyword {
return keyvals[i]
} }
} }
return emptyKV return emptyKV
@ -125,20 +181,27 @@ var emptyKV = KeyVal("")
// MergeSet takes the current setKeyVals, and then applies the entryKeyVals // MergeSet takes the current setKeyVals, and then applies the entryKeyVals
// such that the entry's values win. The union is returned. // such that the entry's values win. The union is returned.
func MergeSet(setKeyVals, entryKeyVals []string) KeyVals { func MergeSet(setKeyVals, entryKeyVals []string) []KeyVal {
retList := NewKeyVals(append([]string{}, setKeyVals...)) retList := StringToKeyVals(setKeyVals)
eKVs := NewKeyVals(entryKeyVals) eKVs := StringToKeyVals(entryKeyVals)
seenKeywords := []string{} return MergeKeyValSet(retList, eKVs)
}
// MergeKeyValSet does a merge of the two sets of KeyVal, and the KeyVal of
// entryKeyVals win when there is a duplicate Keyword.
func MergeKeyValSet(setKeyVals, entryKeyVals []KeyVal) []KeyVal {
retList := keyValCopy(setKeyVals)
seenKeywords := []Keyword{}
for i := range retList { for i := range retList {
word := retList[i].Keyword() word := retList[i].Keyword()
if ekv := eKVs.Has(word); ekv != emptyKV { if ekv := HasKeyword(entryKeyVals, word); ekv != emptyKV {
retList[i] = ekv retList[i] = ekv
} }
seenKeywords = append(seenKeywords, word) seenKeywords = append(seenKeywords, word)
} }
for i := range eKVs { for i := range entryKeyVals {
if !inSlice(eKVs[i].Keyword(), seenKeywords) { if !InKeywordSlice(entryKeyVals[i].Keyword(), seenKeywords) {
retList = append(retList, eKVs[i]) retList = append(retList, entryKeyVals[i])
} }
} }
return retList return retList
@ -147,7 +210,7 @@ func MergeSet(setKeyVals, entryKeyVals []string) KeyVals {
var ( var (
// DefaultKeywords has the several default keyword producers (uid, gid, // DefaultKeywords has the several default keyword producers (uid, gid,
// mode, nlink, type, size, mtime) // mode, nlink, type, size, mtime)
DefaultKeywords = []string{ DefaultKeywords = []Keyword{
"size", "size",
"type", "type",
"uid", "uid",
@ -160,7 +223,7 @@ var (
// DefaultTarKeywords has keywords that should be used when creating a manifest from // DefaultTarKeywords has keywords that should be used when creating a manifest from
// an archive. Currently, evaluating the # of hardlinks has not been implemented yet // an archive. Currently, evaluating the # of hardlinks has not been implemented yet
DefaultTarKeywords = []string{ DefaultTarKeywords = []Keyword{
"size", "size",
"type", "type",
"uid", "uid",
@ -171,7 +234,7 @@ var (
} }
// BsdKeywords is the set of keywords that is only in the upstream FreeBSD mtree // BsdKeywords is the set of keywords that is only in the upstream FreeBSD mtree
BsdKeywords = []string{ BsdKeywords = []Keyword{
"cksum", "cksum",
"device", "device",
"flags", // this one is really mostly BSD specific ... "flags", // this one is really mostly BSD specific ...
@ -205,151 +268,36 @@ var (
} }
// SetKeywords is the default set of keywords calculated for a `/set` SpecialType // SetKeywords is the default set of keywords calculated for a `/set` SpecialType
SetKeywords = []string{ SetKeywords = []Keyword{
"uid", "uid",
"gid", "gid",
} }
// KeywordFuncs is the map of all keywords (and the functions to produce them)
KeywordFuncs = map[string]KeywordFunc{
"size": sizeKeywordFunc, // The size, in bytes, of the file
"type": typeKeywordFunc, // The type of the file
"time": timeKeywordFunc, // The last modification time of the file
"link": linkKeywordFunc, // The target of the symbolic link when type=link
"uid": uidKeywordFunc, // The file owner as a numeric value
"gid": gidKeywordFunc, // The file group as a numeric value
"nlink": nlinkKeywordFunc, // The number of hard links the file is expected to have
"uname": unameKeywordFunc, // The file owner as a symbolic name
"mode": modeKeywordFunc, // The current file's permissions as a numeric (octal) or symbolic value
"cksum": cksumKeywordFunc, // The checksum of the file using the default algorithm specified by the cksum(1) utility
"md5": hasherKeywordFunc("md5digest", md5.New), // The MD5 message digest of the file
"md5digest": hasherKeywordFunc("md5digest", md5.New), // A synonym for `md5`
"rmd160": hasherKeywordFunc("ripemd160digest", ripemd160.New), // The RIPEMD160 message digest of the file
"rmd160digest": hasherKeywordFunc("ripemd160digest", ripemd160.New), // A synonym for `rmd160`
"ripemd160digest": hasherKeywordFunc("ripemd160digest", ripemd160.New), // A synonym for `rmd160`
"sha1": hasherKeywordFunc("sha1digest", sha1.New), // The SHA1 message digest of the file
"sha1digest": hasherKeywordFunc("sha1digest", sha1.New), // A synonym for `sha1`
"sha256": hasherKeywordFunc("sha256digest", sha256.New), // The SHA256 message digest of the file
"sha256digest": hasherKeywordFunc("sha256digest", sha256.New), // A synonym for `sha256`
"sha384": hasherKeywordFunc("sha384digest", sha512.New384), // The SHA384 message digest of the file
"sha384digest": hasherKeywordFunc("sha384digest", sha512.New384), // A synonym for `sha384`
"sha512": hasherKeywordFunc("sha512digest", sha512.New), // The SHA512 message digest of the file
"sha512digest": hasherKeywordFunc("sha512digest", sha512.New), // A synonym for `sha512`
"flags": flagsKeywordFunc, // NOTE: this is a noop, but here to support the presence of the "flags" keyword.
// This is not an upstreamed keyword, but used to vary from "time", as tar
// archives do not store nanosecond precision. So comparing on "time" will
// be only seconds level accurate.
"tar_time": tartimeKeywordFunc, // The last modification time of the file, from a tar archive mtime
// This is not an upstreamed keyword, but a needed attribute for file validation.
// The pattern for this keyword key is prefixed by "xattr." followed by the extended attribute "namespace.key".
// The keyword value is the SHA1 digest of the extended attribute's value.
// In this way, the order of the keys does not matter, and the contents of the value is not revealed.
"xattr": xattrKeywordFunc,
"xattrs": xattrKeywordFunc,
}
) )
var ( // KeywordSynonym returns the canonical name for keywords that have synonyms,
modeKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) { // and just returns the name provided if there is no synonym. In this way it
permissions := info.Mode().Perm() // ought to be safe to wrap any keyword name.
if os.ModeSetuid&info.Mode() > 0 { func KeywordSynonym(name string) Keyword {
permissions |= (1 << 11) var retname string
} switch name {
if os.ModeSetgid&info.Mode() > 0 { case "md5":
permissions |= (1 << 10) retname = "md5digest"
} case "rmd160":
if os.ModeSticky&info.Mode() > 0 { retname = "ripemd160digest"
permissions |= (1 << 9) case "rmd160digest":
} retname = "ripemd160digest"
return fmt.Sprintf("mode=%#o", permissions), nil case "sha1":
retname = "sha1digest"
case "sha256":
retname = "sha256digest"
case "sha384":
retname = "sha384digest"
case "sha512":
retname = "sha512digest"
case "xattrs":
retname = "xattr"
default:
retname = name
} }
sizeKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) { return Keyword(retname)
if sys, ok := info.Sys().(*tar.Header); ok { }
if sys.Typeflag == tar.TypeSymlink {
return fmt.Sprintf("size=%d", len(sys.Linkname)), nil
}
}
return fmt.Sprintf("size=%d", info.Size()), nil
}
cksumKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) {
if !info.Mode().IsRegular() {
return "", nil
}
sum, _, err := cksum(r)
if err != nil {
return "", err
}
return fmt.Sprintf("cksum=%d", sum), nil
}
hasherKeywordFunc = func(name string, newHash func() hash.Hash) KeywordFunc {
return func(path string, info os.FileInfo, r io.Reader) (string, error) {
if !info.Mode().IsRegular() {
return "", nil
}
h := newHash()
if _, err := io.Copy(h, r); err != nil {
return "", err
}
return fmt.Sprintf("%s=%x", name, h.Sum(nil)), nil
}
}
tartimeKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) {
return fmt.Sprintf("tar_time=%d.%9.9d", info.ModTime().Unix(), 0), nil
}
timeKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) {
tSec := info.ModTime().Unix()
tNano := info.ModTime().Nanosecond()
return fmt.Sprintf("time=%d.%9.9d", tSec, tNano), nil
}
linkKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) {
if sys, ok := info.Sys().(*tar.Header); ok {
if sys.Linkname != "" {
linkname, err := Vis(sys.Linkname)
if err != nil {
return "", err
}
return fmt.Sprintf("link=%s", linkname), nil
}
return "", nil
}
if info.Mode()&os.ModeSymlink != 0 {
str, err := os.Readlink(path)
if err != nil {
return "", err
}
linkname, err := Vis(str)
if err != nil {
return "", err
}
return fmt.Sprintf("link=%s", linkname), nil
}
return "", nil
}
typeKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) {
if info.Mode().IsDir() {
return "type=dir", nil
}
if info.Mode().IsRegular() {
return "type=file", nil
}
if info.Mode()&os.ModeSocket != 0 {
return "type=socket", nil
}
if info.Mode()&os.ModeSymlink != 0 {
return "type=link", nil
}
if info.Mode()&os.ModeNamedPipe != 0 {
return "type=fifo", nil
}
if info.Mode()&os.ModeDevice != 0 {
if info.Mode()&os.ModeCharDevice != 0 {
return "type=char", nil
}
return "type=device", nil
}
return "", nil
}
)

View file

@ -12,46 +12,46 @@ import (
) )
var ( var (
flagsKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) { flagsKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
// ideally this will pull in from here https://www.freebsd.org/cgi/man.cgi?query=chflags&sektion=2 // ideally this will pull in from here https://www.freebsd.org/cgi/man.cgi?query=chflags&sektion=2
return "", nil return emptyKV, nil
} }
unameKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) { unameKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
if hdr, ok := info.Sys().(*tar.Header); ok { if hdr, ok := info.Sys().(*tar.Header); ok {
return fmt.Sprintf("uname=%s", hdr.Uname), nil return KeyVal(fmt.Sprintf("uname=%s", hdr.Uname)), nil
} }
stat := info.Sys().(*syscall.Stat_t) stat := info.Sys().(*syscall.Stat_t)
u, err := user.LookupId(fmt.Sprintf("%d", stat.Uid)) u, err := user.LookupId(fmt.Sprintf("%d", stat.Uid))
if err != nil { if err != nil {
return "", err return emptyKV, err
} }
return fmt.Sprintf("uname=%s", u.Username), nil return KeyVal(fmt.Sprintf("uname=%s", u.Username)), nil
} }
uidKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) { uidKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
if hdr, ok := info.Sys().(*tar.Header); ok { if hdr, ok := info.Sys().(*tar.Header); ok {
return fmt.Sprintf("uid=%d", hdr.Uid), nil return KeyVal(fmt.Sprintf("uid=%d", hdr.Uid)), nil
} }
stat := info.Sys().(*syscall.Stat_t) stat := info.Sys().(*syscall.Stat_t)
return fmt.Sprintf("uid=%d", stat.Uid), nil return KeyVal(fmt.Sprintf("uid=%d", stat.Uid)), nil
} }
gidKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) { gidKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
if hdr, ok := info.Sys().(*tar.Header); ok { if hdr, ok := info.Sys().(*tar.Header); ok {
return fmt.Sprintf("gid=%d", hdr.Gid), nil return KeyVal(fmt.Sprintf("gid=%d", hdr.Gid)), nil
} }
if stat, ok := info.Sys().(*syscall.Stat_t); ok { if stat, ok := info.Sys().(*syscall.Stat_t); ok {
return fmt.Sprintf("gid=%d", stat.Gid), nil return KeyVal(fmt.Sprintf("gid=%d", stat.Gid)), nil
} }
return "", nil return emptyKV, nil
} }
nlinkKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) { nlinkKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
if stat, ok := info.Sys().(*syscall.Stat_t); ok { if stat, ok := info.Sys().(*syscall.Stat_t); ok {
return fmt.Sprintf("nlink=%d", stat.Nlink), nil return KeyVal(fmt.Sprintf("nlink=%d", stat.Nlink)), nil
} }
return "", nil return emptyKV, nil
} }
xattrKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) { xattrKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
return "", nil return emptyKV, nil
} }
) )

View file

@ -17,71 +17,71 @@ import (
var ( var (
// this is bsd specific https://www.freebsd.org/cgi/man.cgi?query=chflags&sektion=2 // this is bsd specific https://www.freebsd.org/cgi/man.cgi?query=chflags&sektion=2
flagsKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) { flagsKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
return "", nil return emptyKV, nil
} }
unameKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) { unameKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
if hdr, ok := info.Sys().(*tar.Header); ok { if hdr, ok := info.Sys().(*tar.Header); ok {
return fmt.Sprintf("uname=%s", hdr.Uname), nil return KeyVal(fmt.Sprintf("uname=%s", hdr.Uname)), nil
} }
stat := info.Sys().(*syscall.Stat_t) stat := info.Sys().(*syscall.Stat_t)
u, err := user.LookupId(fmt.Sprintf("%d", stat.Uid)) u, err := user.LookupId(fmt.Sprintf("%d", stat.Uid))
if err != nil { if err != nil {
return "", err return emptyKV, err
} }
return fmt.Sprintf("uname=%s", u.Username), nil return KeyVal(fmt.Sprintf("uname=%s", u.Username)), nil
} }
uidKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) { uidKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
if hdr, ok := info.Sys().(*tar.Header); ok { if hdr, ok := info.Sys().(*tar.Header); ok {
return fmt.Sprintf("uid=%d", hdr.Uid), nil return KeyVal(fmt.Sprintf("uid=%d", hdr.Uid)), nil
} }
stat := info.Sys().(*syscall.Stat_t) stat := info.Sys().(*syscall.Stat_t)
return fmt.Sprintf("uid=%d", stat.Uid), nil return KeyVal(fmt.Sprintf("uid=%d", stat.Uid)), nil
} }
gidKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) { gidKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
if hdr, ok := info.Sys().(*tar.Header); ok { if hdr, ok := info.Sys().(*tar.Header); ok {
return fmt.Sprintf("gid=%d", hdr.Gid), nil return KeyVal(fmt.Sprintf("gid=%d", hdr.Gid)), nil
} }
if stat, ok := info.Sys().(*syscall.Stat_t); ok { if stat, ok := info.Sys().(*syscall.Stat_t); ok {
return fmt.Sprintf("gid=%d", stat.Gid), nil return KeyVal(fmt.Sprintf("gid=%d", stat.Gid)), nil
} }
return "", nil return emptyKV, nil
} }
nlinkKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) { nlinkKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
if stat, ok := info.Sys().(*syscall.Stat_t); ok { if stat, ok := info.Sys().(*syscall.Stat_t); ok {
return fmt.Sprintf("nlink=%d", stat.Nlink), nil return KeyVal(fmt.Sprintf("nlink=%d", stat.Nlink)), nil
} }
return "", nil return emptyKV, nil
} }
xattrKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (string, error) { xattrKeywordFunc = func(path string, info os.FileInfo, r io.Reader) (KeyVal, error) {
if hdr, ok := info.Sys().(*tar.Header); ok { if hdr, ok := info.Sys().(*tar.Header); ok {
if len(hdr.Xattrs) == 0 { if len(hdr.Xattrs) == 0 {
return "", nil return emptyKV, nil
} }
klist := []string{} klist := []KeyVal{}
for k, v := range hdr.Xattrs { for k, v := range hdr.Xattrs {
klist = append(klist, fmt.Sprintf("xattr.%s=%s", k, base64.StdEncoding.EncodeToString([]byte(v)))) klist = append(klist, KeyVal(fmt.Sprintf("xattr.%s=%s", k, base64.StdEncoding.EncodeToString([]byte(v)))))
} }
return strings.Join(klist, " "), nil return KeyVal(strings.Join(KeyValToString(klist), " ")), nil
} }
if !info.Mode().IsRegular() && !info.Mode().IsDir() { if !info.Mode().IsRegular() && !info.Mode().IsDir() {
return "", nil return emptyKV, nil
} }
xlist, err := xattr.List(path) xlist, err := xattr.List(path)
if err != nil { if err != nil {
return "", err return emptyKV, err
} }
klist := make([]string, len(xlist)) klist := make([]KeyVal, len(xlist))
for i := range xlist { for i := range xlist {
data, err := xattr.Get(path, xlist[i]) data, err := xattr.Get(path, xlist[i])
if err != nil { if err != nil {
return "", err return emptyKV, err
} }
klist[i] = fmt.Sprintf("xattr.%s=%s", xlist[i], base64.StdEncoding.EncodeToString(data)) klist[i] = KeyVal(fmt.Sprintf("xattr.%s=%s", xlist[i], base64.StdEncoding.EncodeToString(data)))
} }
return strings.Join(klist, " "), nil return KeyVal(strings.Join(KeyValToString(klist), " ")), nil
} }
) )

View file

@ -54,7 +54,7 @@ func TestKeywordsTimeNano(t *testing.T) {
{857125628319, 0}, {857125628319, 0},
} { } {
mtime := time.Unix(test.sec, test.nsec) mtime := time.Unix(test.sec, test.nsec)
expected := fmt.Sprintf("time=%d.%9.9d", test.sec, test.nsec) expected := KeyVal(fmt.Sprintf("time=%d.%9.9d", test.sec, test.nsec))
got, err := timeKeywordFunc("", fakeFileInfo{ got, err := timeKeywordFunc("", fakeFileInfo{
mtime: mtime, mtime: mtime,
}, nil) }, nil)
@ -81,7 +81,7 @@ func TestKeywordsTimeTar(t *testing.T) {
{857125628319, 0}, {857125628319, 0},
} { } {
mtime := time.Unix(test.sec, test.nsec) mtime := time.Unix(test.sec, test.nsec)
expected := fmt.Sprintf("tar_time=%d.%9.9d", test.sec, 0) expected := KeyVal(fmt.Sprintf("tar_time=%d.%9.9d", test.sec, 0))
got, err := tartimeKeywordFunc("", fakeFileInfo{ got, err := tartimeKeywordFunc("", fakeFileInfo{
mtime: mtime, mtime: mtime,
}, nil) }, nil)
@ -93,3 +93,34 @@ func TestKeywordsTimeTar(t *testing.T) {
} }
} }
} }
func TestKeywordSynonym(t *testing.T) {
checklist := []struct {
give string
expect Keyword
}{
{give: "time", expect: "time"},
{give: "md5", expect: "md5digest"},
{give: "md5digest", expect: "md5digest"},
{give: "rmd160", expect: "ripemd160digest"},
{give: "rmd160digest", expect: "ripemd160digest"},
{give: "ripemd160digest", expect: "ripemd160digest"},
{give: "sha1", expect: "sha1digest"},
{give: "sha1digest", expect: "sha1digest"},
{give: "sha256", expect: "sha256digest"},
{give: "sha256digest", expect: "sha256digest"},
{give: "sha384", expect: "sha384digest"},
{give: "sha384digest", expect: "sha384digest"},
{give: "sha512", expect: "sha512digest"},
{give: "sha512digest", expect: "sha512digest"},
{give: "xattr", expect: "xattr"},
{give: "xattrs", expect: "xattr"},
}
for i, check := range checklist {
got := KeywordSynonym(check.give)
if got != check.expect {
t.Errorf("%d: expected %q; got %q", i, check.expect, got)
}
}
}

View file

@ -48,7 +48,7 @@ func ParseSpec(r io.Reader) (*DirectoryHierarchy, error) {
// parse the options // parse the options
f := strings.Fields(str) f := strings.Fields(str)
e.Name = f[0] e.Name = f[0]
e.Keywords = f[1:] e.Keywords = StringToKeyVals(f[1:])
if e.Name == "/set" { if e.Name == "/set" {
creator.curSet = &e creator.curSet = &e
} else if e.Name == "/unset" { } else if e.Name == "/unset" {
@ -80,7 +80,7 @@ func ParseSpec(r io.Reader) (*DirectoryHierarchy, error) {
} else { } else {
e.Type = RelativeType e.Type = RelativeType
} }
e.Keywords = f[1:] e.Keywords = StringToKeyVals(f[1:])
// TODO: gather keywords if using tar stream // TODO: gather keywords if using tar stream
e.Parent = creator.curDir e.Parent = creator.curDir
for i := range e.Keywords { for i := range e.Keywords {

50
tar.go
View file

@ -17,11 +17,15 @@ type Streamer interface {
Hierarchy() (*DirectoryHierarchy, error) Hierarchy() (*DirectoryHierarchy, error)
} }
var tarDefaultSetKeywords = []string{"type=file", "flags=none", "mode=0664"} var tarDefaultSetKeywords = []KeyVal{
"type=file",
"flags=none",
"mode=0664",
}
// NewTarStreamer streams a tar archive and creates a file hierarchy based off // NewTarStreamer streams a tar archive and creates a file hierarchy based off
// of the tar metadata headers // of the tar metadata headers
func NewTarStreamer(r io.Reader, keywords []string) Streamer { func NewTarStreamer(r io.Reader, keywords []Keyword) Streamer {
pR, pW := io.Pipe() pR, pW := io.Pipe()
ts := &tarStream{ ts := &tarStream{
pipeReader: pR, pipeReader: pR,
@ -45,17 +49,17 @@ type tarStream struct {
pipeWriter *io.PipeWriter pipeWriter *io.PipeWriter
teeReader io.Reader teeReader io.Reader
tarReader *tar.Reader tarReader *tar.Reader
keywords []string keywords []Keyword
err error err error
} }
func (ts *tarStream) readHeaders() { func (ts *tarStream) readHeaders() {
// remove "time" keyword // remove "time" keyword
notimekws := []string{} notimekws := []Keyword{}
for _, kw := range ts.keywords { for _, kw := range ts.keywords {
if !inSlice(kw, notimekws) { if !InKeywordSlice(kw, notimekws) {
if kw == "time" { if kw == "time" {
if !inSlice("tar_time", ts.keywords) { if !InKeywordSlice("tar_time", ts.keywords) {
notimekws = append(notimekws, "tar_time") notimekws = append(notimekws, "tar_time")
} }
} else { } else {
@ -74,7 +78,7 @@ func (ts *tarStream) readHeaders() {
Type: CommentType, Type: CommentType,
}, },
Set: nil, Set: nil,
Keywords: []string{"type=dir"}, Keywords: []KeyVal{"type=dir"},
} }
metadataEntries := signatureEntries("<user specified tar archive>") metadataEntries := signatureEntries("<user specified tar archive>")
for _, e := range metadataEntries { for _, e := range metadataEntries {
@ -242,7 +246,7 @@ func populateTree(root, e *Entry, hdr *tar.Header) error {
Name: encoded, Name: encoded,
Type: RelativeType, Type: RelativeType,
Parent: parent, Parent: parent,
Keywords: []string{"type=dir"}, // temp data Keywords: []KeyVal{"type=dir"}, // temp data
Set: nil, // temp data Set: nil, // temp data
} }
pathname, err := newEntry.Path() pathname, err := newEntry.Path()
@ -276,7 +280,7 @@ func populateTree(root, e *Entry, hdr *tar.Header) error {
// root: the "head" of the sub-tree to flatten // root: the "head" of the sub-tree to flatten
// creator: a dhCreator that helps with the '/set' keyword // creator: a dhCreator that helps with the '/set' keyword
// keywords: keywords specified by the user that should be evaluated // keywords: keywords specified by the user that should be evaluated
func flatten(root *Entry, creator *dhCreator, keywords []string) { func flatten(root *Entry, creator *dhCreator, keywords []Keyword) {
if root == nil || creator == nil { if root == nil || creator == nil {
return return
} }
@ -292,18 +296,19 @@ func flatten(root *Entry, creator *dhCreator, keywords []string) {
if root.Set != nil { if root.Set != nil {
// Check if we need a new set // Check if we need a new set
consolidatedKeys := keyvalSelector(append(tarDefaultSetKeywords, root.Set.Keywords...), keywords)
if creator.curSet == nil { if creator.curSet == nil {
creator.curSet = &Entry{ creator.curSet = &Entry{
Type: SpecialType, Type: SpecialType,
Name: "/set", Name: "/set",
Keywords: keywordSelector(append(tarDefaultSetKeywords, root.Set.Keywords...), keywords), Keywords: consolidatedKeys,
Pos: len(creator.DH.Entries), Pos: len(creator.DH.Entries),
} }
creator.DH.Entries = append(creator.DH.Entries, *creator.curSet) creator.DH.Entries = append(creator.DH.Entries, *creator.curSet)
} else { } else {
needNewSet := false needNewSet := false
for _, k := range root.Set.Keywords { for _, k := range root.Set.Keywords {
if !inSlice(k, creator.curSet.Keywords) { if !inKeyValSlice(k, creator.curSet.Keywords) {
needNewSet = true needNewSet = true
break break
} }
@ -313,7 +318,7 @@ func flatten(root *Entry, creator *dhCreator, keywords []string) {
Name: "/set", Name: "/set",
Type: SpecialType, Type: SpecialType,
Pos: len(creator.DH.Entries), Pos: len(creator.DH.Entries),
Keywords: keywordSelector(append(tarDefaultSetKeywords, root.Set.Keywords...), keywords), Keywords: consolidatedKeys,
} }
creator.DH.Entries = append(creator.DH.Entries, *creator.curSet) creator.DH.Entries = append(creator.DH.Entries, *creator.curSet)
} }
@ -331,7 +336,7 @@ func flatten(root *Entry, creator *dhCreator, keywords []string) {
} }
root.Set = creator.curSet root.Set = creator.curSet
if creator.curSet != nil { if creator.curSet != nil {
root.Keywords = setDifference(root.Keywords, creator.curSet.Keywords) root.Keywords = keyValDifference(root.Keywords, creator.curSet.Keywords)
} }
root.Pos = len(creator.DH.Entries) root.Pos = len(creator.DH.Entries)
creator.DH.Entries = append(creator.DH.Entries, *root) creator.DH.Entries = append(creator.DH.Entries, *root)
@ -376,11 +381,11 @@ func resolveHardlinks(root *Entry, hardlinks map[string][]string, countlinks boo
} }
linkfile.Keywords = basefile.Keywords linkfile.Keywords = basefile.Keywords
if countlinks { if countlinks {
linkfile.Keywords = append(linkfile.Keywords, fmt.Sprintf("nlink=%d", len(links)+1)) linkfile.Keywords = append(linkfile.Keywords, KeyVal(fmt.Sprintf("nlink=%d", len(links)+1)))
} }
} }
if countlinks { if countlinks {
basefile.Keywords = append(basefile.Keywords, fmt.Sprintf("nlink=%d", len(links)+1)) basefile.Keywords = append(basefile.Keywords, KeyVal(fmt.Sprintf("nlink=%d", len(links)+1)))
} }
} }
} }
@ -410,19 +415,6 @@ func filter(root *Entry, p func(*Entry) bool) []Entry {
return nil return nil
} }
func setDifference(this, that []string) []string {
if len(this) == 0 {
return that
}
diff := []string{}
for _, kv := range this {
if !inSlice(kv, that) {
diff = append(diff, kv)
}
}
return diff
}
func (ts *tarStream) setErr(err error) { func (ts *tarStream) setErr(err error) {
ts.err = err ts.err = err
} }
@ -444,7 +436,7 @@ func (ts *tarStream) Hierarchy() (*DirectoryHierarchy, error) {
if ts.root == nil { if ts.root == nil {
return nil, fmt.Errorf("root Entry not found, nothing to flatten") return nil, fmt.Errorf("root Entry not found, nothing to flatten")
} }
resolveHardlinks(ts.root, ts.hardlinks, inSlice("nlink", ts.keywords)) resolveHardlinks(ts.root, ts.hardlinks, InKeywordSlice(Keyword("nlink"), ts.keywords))
flatten(ts.root, &ts.creator, ts.keywords) flatten(ts.root, &ts.creator, ts.keywords)
return ts.creator.DH, nil return ts.creator.DH, nil
} }

View file

@ -128,7 +128,7 @@ func TestArchiveCreation(t *testing.T) {
if err != nil { if err != nil {
t.Fatal(err) t.Fatal(err)
} }
str := NewTarStreamer(fh, []string{"sha1"}) str := NewTarStreamer(fh, []Keyword{"sha1"})
if _, err := io.Copy(ioutil.Discard, str); err != nil && err != io.EOF { if _, err := io.Copy(ioutil.Discard, str); err != nil && err != io.EOF {
t.Fatal(err) t.Fatal(err)
@ -145,7 +145,7 @@ func TestArchiveCreation(t *testing.T) {
} }
// Test the tar manifest against the actual directory // Test the tar manifest against the actual directory
res, err := Check("./testdata/collection", tdh, []string{"sha1"}) res, err := Check("./testdata/collection", tdh, []Keyword{"sha1"})
if err != nil { if err != nil {
t.Fatal(err) t.Fatal(err)
} }
@ -158,7 +158,7 @@ func TestArchiveCreation(t *testing.T) {
} }
// Test the tar manifest against itself // Test the tar manifest against itself
res, err = TarCheck(tdh, tdh, []string{"sha1"}) res, err = TarCheck(tdh, tdh, []Keyword{"sha1"})
if err != nil { if err != nil {
t.Fatal(err) t.Fatal(err)
} }
@ -170,11 +170,11 @@ func TestArchiveCreation(t *testing.T) {
} }
// Validate the directory manifest against the archive // Validate the directory manifest against the archive
dh, err := Walk("./testdata/collection", nil, []string{"sha1"}) dh, err := Walk("./testdata/collection", nil, []Keyword{"sha1"})
if err != nil { if err != nil {
t.Fatal(err) t.Fatal(err)
} }
res, err = TarCheck(tdh, dh, []string{"sha1"}) res, err = TarCheck(tdh, dh, []Keyword{"sha1"})
if err != nil { if err != nil {
t.Fatal(err) t.Fatal(err)
} }
@ -212,7 +212,7 @@ func TestTreeTraversal(t *testing.T) {
t.Fatal(err) t.Fatal(err)
} }
res, err := TarCheck(tdh, tdh, []string{"sha1"}) res, err := TarCheck(tdh, tdh, []Keyword{"sha1"})
if err != nil { if err != nil {
t.Fatal(err) t.Fatal(err)
} }
@ -224,7 +224,7 @@ func TestTreeTraversal(t *testing.T) {
} }
// top-level "." directory will contain contents of traversal.tar // top-level "." directory will contain contents of traversal.tar
res, err = Check("./testdata/.", tdh, []string{"sha1"}) res, err = Check("./testdata/.", tdh, []Keyword{"sha1"})
if err != nil { if err != nil {
t.Fatal(err) t.Fatal(err)
} }
@ -262,7 +262,7 @@ func TestTreeTraversal(t *testing.T) {
} }
// Implied top-level "." directory will contain the contents of singlefile.tar // Implied top-level "." directory will contain the contents of singlefile.tar
res, err = Check("./testdata/.", tdh, []string{"sha1"}) res, err = Check("./testdata/.", tdh, []Keyword{"sha1"})
if err != nil { if err != nil {
t.Fatal(err) t.Fatal(err)
} }

24
test/cli/0004.sh Normal file
View file

@ -0,0 +1,24 @@
#!/bin/bash
set -e
#set -x
name=$(basename $0)
root=$1
gomtree=$(readlink -f ${root}/gomtree)
t=$(mktemp -d /tmp/go-mtree.XXXXXX)
echo "[${name}] Running in ${t}"
pushd ${root}
git archive --format=tar HEAD^{tree} . > ${t}/${name}.tar
mkdir -p ${t}/extract
tar -C ${t}/extract/ -xf ${t}/${name}.tar
## This is a checking that keyword synonyms are respected
${gomtree} -k sha1digest -c -p ${t}/extract/ > ${t}/${name}.mtree
${gomtree} -k sha1 -f ${t}/${name}.mtree -p ${t}/extract/
${gomtree} -k sha1 -c -p ${t}/extract/ > ${t}/${name}.mtree
${gomtree} -k sha1digest -f ${t}/${name}.mtree -p ${t}/extract/
popd
rm -rf ${t}

19
test/cli/0005.sh Normal file
View file

@ -0,0 +1,19 @@
#!/bin/bash
set -e
name=$(basename $0)
root=$1
gomtree=$(readlink -f ${root}/gomtree)
t=$(mktemp -d /tmp/go-mtree.XXXXXX)
echo "[${name}] Running in ${t}"
pushd ${root}
mkdir -p ${t}/extract
git archive --format=tar HEAD^{tree} . | tar -C ${t}/extract/ -x
${gomtree} -k sha1digest -c -p ${t}/extract/ > ${t}/${name}.mtree
${gomtree} -f ${t}/${name}.mtree -k md5digest -p ${t}/extract/
popd
rm -rf ${t}

View file

@ -2,6 +2,11 @@ package mtree
import "fmt" import "fmt"
const (
// AppName is the name ... of this library/application
AppName = "gomtree"
)
const ( const (
// VersionMajor is for an API incompatible changes // VersionMajor is for an API incompatible changes
VersionMajor = 0 VersionMajor = 0

25
walk.go
View file

@ -15,12 +15,12 @@ import (
// returns true, then the path is not included in the spec. // returns true, then the path is not included in the spec.
type ExcludeFunc func(path string, info os.FileInfo) bool type ExcludeFunc func(path string, info os.FileInfo) bool
var defaultSetKeywords = []string{"type=file", "nlink=1", "flags=none", "mode=0664"} var defaultSetKeywords = []KeyVal{"type=file", "nlink=1", "flags=none", "mode=0664"}
// Walk from root directory and assemble the DirectoryHierarchy. excludes // Walk from root directory and assemble the DirectoryHierarchy. excludes
// provided are used to skip paths. keywords are the set to collect from the // provided are used to skip paths. keywords are the set to collect from the
// walked paths. The recommended default list is DefaultKeywords. // walked paths. The recommended default list is DefaultKeywords.
func Walk(root string, exlcudes []ExcludeFunc, keywords []string) (*DirectoryHierarchy, error) { func Walk(root string, excludes []ExcludeFunc, keywords []Keyword) (*DirectoryHierarchy, error) {
creator := dhCreator{DH: &DirectoryHierarchy{}} creator := dhCreator{DH: &DirectoryHierarchy{}}
// insert signature and metadata comments first (user, machine, tree, date) // insert signature and metadata comments first (user, machine, tree, date)
metadataEntries := signatureEntries(root) metadataEntries := signatureEntries(root)
@ -32,7 +32,7 @@ func Walk(root string, exlcudes []ExcludeFunc, keywords []string) (*DirectoryHie
if err != nil { if err != nil {
return err return err
} }
for _, ex := range exlcudes { for _, ex := range excludes {
if ex(path, info) { if ex(path, info) {
return nil return nil
} }
@ -71,7 +71,7 @@ func Walk(root string, exlcudes []ExcludeFunc, keywords []string) (*DirectoryHie
Name: "/set", Name: "/set",
Type: SpecialType, Type: SpecialType,
Pos: len(creator.DH.Entries), Pos: len(creator.DH.Entries),
Keywords: keywordSelector(defaultSetKeywords, keywords), Keywords: keyvalSelector(defaultSetKeywords, keywords),
} }
for _, keyword := range SetKeywords { for _, keyword := range SetKeywords {
err := func() error { err := func() error {
@ -103,7 +103,7 @@ func Walk(root string, exlcudes []ExcludeFunc, keywords []string) (*DirectoryHie
creator.DH.Entries = append(creator.DH.Entries, e) creator.DH.Entries = append(creator.DH.Entries, e)
} else if creator.curSet != nil { } else if creator.curSet != nil {
// check the attributes of the /set keywords and re-set if changed // check the attributes of the /set keywords and re-set if changed
klist := []string{} klist := []KeyVal{}
for _, keyword := range SetKeywords { for _, keyword := range SetKeywords {
err := func() error { err := func() error {
var r io.Reader var r io.Reader
@ -135,7 +135,7 @@ func Walk(root string, exlcudes []ExcludeFunc, keywords []string) (*DirectoryHie
needNewSet := false needNewSet := false
for _, k := range klist { for _, k := range klist {
if !inSlice(k, creator.curSet.Keywords) { if !inKeyValSlice(k, creator.curSet.Keywords) {
needNewSet = true needNewSet = true
} }
} }
@ -144,7 +144,7 @@ func Walk(root string, exlcudes []ExcludeFunc, keywords []string) (*DirectoryHie
Name: "/set", Name: "/set",
Type: SpecialType, Type: SpecialType,
Pos: len(creator.DH.Entries), Pos: len(creator.DH.Entries),
Keywords: keywordSelector(append(defaultSetKeywords, klist...), keywords), Keywords: keyvalSelector(append(defaultSetKeywords, klist...), keywords),
} }
creator.curSet = &e creator.curSet = &e
creator.DH.Entries = append(creator.DH.Entries, e) creator.DH.Entries = append(creator.DH.Entries, e)
@ -181,7 +181,7 @@ func Walk(root string, exlcudes []ExcludeFunc, keywords []string) (*DirectoryHie
if err != nil { if err != nil {
return err return err
} }
if str != "" && !inSlice(str, creator.curSet.Keywords) { if str != "" && !inKeyValSlice(str, creator.curSet.Keywords) {
e.Keywords = append(e.Keywords, str) e.Keywords = append(e.Keywords, str)
} }
return nil return nil
@ -209,15 +209,6 @@ func Walk(root string, exlcudes []ExcludeFunc, keywords []string) (*DirectoryHie
return creator.DH, err return creator.DH, err
} }
func inSlice(a string, list []string) bool {
for _, b := range list {
if b == a {
return true
}
}
return false
}
// startWalk walks the file tree rooted at root, calling walkFn for each file or // startWalk walks the file tree rooted at root, calling walkFn for each file or
// directory in the tree, including root. All errors that arise visiting files // directory in the tree, including root. All errors that arise visiting files
// and directories are filtered by walkFn. The files are walked in lexical // and directories are filtered by walkFn. The files are walked in lexical