| // Copyright 2015 Google Inc. All rights reserved |
| // |
| // Licensed under the Apache License, Version 2.0 (the "License"); |
| // you may not use this file except in compliance with the License. |
| // You may obtain a copy of the License at |
| // |
| // http://www.apache.org/licenses/LICENSE-2.0 |
| // |
| // Unless required by applicable law or agreed to in writing, software |
| // distributed under the License is distributed on an "AS IS" BASIS, |
| // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| // See the License for the specific language governing permissions and |
| // limitations under the License. |
| |
| package kati |
| |
| import ( |
| "bytes" |
| "crypto/sha1" |
| "encoding/binary" |
| "encoding/gob" |
| "encoding/json" |
| "fmt" |
| "io" |
| "io/ioutil" |
| "net/url" |
| "os" |
| "sort" |
| "strconv" |
| "time" |
| ) |
| |
| const ( |
| valueTypeRecursive = 'R' |
| valueTypeSimple = 'S' |
| valueTypeTSV = 'T' |
| valueTypeUndefined = 'U' |
| valueTypeAssign = 'a' |
| valueTypeExpr = 'e' |
| valueTypeFunc = 'f' |
| valueTypeLiteral = 'l' |
| valueTypeNop = 'n' |
| valueTypeParamref = 'p' |
| valueTypeVarref = 'r' |
| valueTypeVarsubst = 's' |
| valueTypeTmpval = 't' |
| ) |
| |
| var JSON LoadSaver |
| var GOB LoadSaver |
| |
| func init() { |
| JSON = jsonLoadSaver{} |
| GOB = gobLoadSaver{} |
| } |
| |
| type jsonLoadSaver struct{} |
| type gobLoadSaver struct{} |
| |
| func dumpInt(w io.Writer, i int) { |
| v := int32(i) |
| err := binary.Write(w, binary.LittleEndian, &v) |
| if err != nil { |
| panic(err) |
| } |
| } |
| |
| func dumpString(w io.Writer, s string) { |
| dumpInt(w, len(s)) |
| _, err := io.WriteString(w, s) |
| if err != nil { |
| panic(err) |
| } |
| } |
| |
| func dumpBytes(w io.Writer, b []byte) { |
| dumpInt(w, len(b)) |
| _, err := w.Write(b) |
| if err != nil { |
| panic(err) |
| } |
| } |
| |
| func dumpByte(w io.Writer, b byte) { |
| err := writeByte(w, b) |
| if err != nil { |
| panic(err) |
| } |
| } |
| |
| type serializableVar struct { |
| Type string |
| V string |
| Origin string |
| Children []serializableVar |
| } |
| |
| type serializableDepNode struct { |
| Output int |
| Cmds []string |
| Deps []int |
| Parents []int |
| HasRule bool |
| IsOrderOnly bool |
| IsPhony bool |
| ActualInputs []int |
| TargetSpecificVars []int |
| Filename string |
| Lineno int |
| } |
| |
| type serializableTargetSpecificVar struct { |
| Name string |
| Value serializableVar |
| } |
| |
| type serializableGraph struct { |
| Nodes []*serializableDepNode |
| Vars map[string]serializableVar |
| Tsvs []serializableTargetSpecificVar |
| Targets []string |
| Roots []string |
| AccessedMks []*accessedMakefile |
| Exports map[string]bool |
| } |
| |
| func encGob(v interface{}) string { |
| var buf bytes.Buffer |
| e := gob.NewEncoder(&buf) |
| err := e.Encode(v) |
| if err != nil { |
| panic(err) |
| } |
| return buf.String() |
| } |
| |
| func encVar(k string, v Var) string { |
| var buf bytes.Buffer |
| dumpString(&buf, k) |
| v.dump(&buf) |
| return buf.String() |
| } |
| |
| type depNodesSerializer struct { |
| nodes []*serializableDepNode |
| tsvs []serializableTargetSpecificVar |
| tsvMap map[string]int |
| targets []string |
| targetMap map[string]int |
| done map[string]bool |
| } |
| |
| func newDepNodesSerializer() *depNodesSerializer { |
| return &depNodesSerializer{ |
| tsvMap: make(map[string]int), |
| targetMap: make(map[string]int), |
| done: make(map[string]bool), |
| } |
| } |
| |
| func (ns *depNodesSerializer) serializeTarget(t string) int { |
| id, present := ns.targetMap[t] |
| if present { |
| return id |
| } |
| id = len(ns.targets) |
| ns.targetMap[t] = id |
| ns.targets = append(ns.targets, t) |
| return id |
| } |
| |
| func (ns *depNodesSerializer) serializeDepNodes(nodes []*DepNode) { |
| for _, n := range nodes { |
| if ns.done[n.Output] { |
| continue |
| } |
| ns.done[n.Output] = true |
| |
| var deps []int |
| for _, d := range n.Deps { |
| deps = append(deps, ns.serializeTarget(d.Output)) |
| } |
| var parents []int |
| for _, d := range n.Parents { |
| parents = append(parents, ns.serializeTarget(d.Output)) |
| } |
| var actualInputs []int |
| for _, i := range n.ActualInputs { |
| actualInputs = append(actualInputs, ns.serializeTarget(i)) |
| } |
| |
| // Sort keys for consistent serialization. |
| var tsvKeys []string |
| for k := range n.TargetSpecificVars { |
| tsvKeys = append(tsvKeys, k) |
| } |
| sort.Strings(tsvKeys) |
| |
| var vars []int |
| for _, k := range tsvKeys { |
| v := n.TargetSpecificVars[k] |
| sv := serializableTargetSpecificVar{Name: k, Value: v.serialize()} |
| //gob := encGob(sv) |
| gob := encVar(k, v) |
| id, present := ns.tsvMap[gob] |
| if !present { |
| id = len(ns.tsvs) |
| ns.tsvMap[gob] = id |
| ns.tsvs = append(ns.tsvs, sv) |
| } |
| vars = append(vars, id) |
| } |
| |
| ns.nodes = append(ns.nodes, &serializableDepNode{ |
| Output: ns.serializeTarget(n.Output), |
| Cmds: n.Cmds, |
| Deps: deps, |
| Parents: parents, |
| HasRule: n.HasRule, |
| IsOrderOnly: n.IsOrderOnly, |
| IsPhony: n.IsPhony, |
| ActualInputs: actualInputs, |
| TargetSpecificVars: vars, |
| Filename: n.Filename, |
| Lineno: n.Lineno, |
| }) |
| ns.serializeDepNodes(n.Deps) |
| } |
| } |
| |
| func makeSerializableVars(vars Vars) (r map[string]serializableVar) { |
| r = make(map[string]serializableVar) |
| for k, v := range vars { |
| r[k] = v.serialize() |
| } |
| return r |
| } |
| |
| func makeSerializableGraph(g *DepGraph, roots []string) serializableGraph { |
| ns := newDepNodesSerializer() |
| ns.serializeDepNodes(g.nodes) |
| v := makeSerializableVars(g.vars) |
| return serializableGraph{ |
| Nodes: ns.nodes, |
| Vars: v, |
| Tsvs: ns.tsvs, |
| Targets: ns.targets, |
| Roots: roots, |
| AccessedMks: g.accessedMks, |
| Exports: g.exports, |
| } |
| } |
| |
| func (jsonLoadSaver) Save(g *DepGraph, filename string, roots []string) error { |
| startTime := time.Now() |
| sg := makeSerializableGraph(g, roots) |
| o, err := json.MarshalIndent(sg, " ", " ") |
| if err != nil { |
| return err |
| } |
| f, err := os.Create(filename) |
| if err != nil { |
| return err |
| } |
| _, err = f.Write(o) |
| if err != nil { |
| f.Close() |
| return err |
| } |
| err = f.Close() |
| if err != nil { |
| return err |
| } |
| logStats("json serialize time: %q", time.Since(startTime)) |
| return nil |
| } |
| |
| func (gobLoadSaver) Save(g *DepGraph, filename string, roots []string) error { |
| startTime := time.Now() |
| f, err := os.Create(filename) |
| if err != nil { |
| return err |
| } |
| e := gob.NewEncoder(f) |
| var sg serializableGraph |
| { |
| startTime := time.Now() |
| sg = makeSerializableGraph(g, roots) |
| logStats("gob serialize prepare time: %q", time.Since(startTime)) |
| } |
| { |
| startTime := time.Now() |
| e.Encode(sg) |
| logStats("gob serialize output time: %q", time.Since(startTime)) |
| } |
| err = f.Close() |
| if err != nil { |
| return err |
| } |
| logStats("gob serialize time: %q", time.Since(startTime)) |
| return nil |
| } |
| |
| func cacheFilename(mk string, roots []string) string { |
| filename := ".kati_cache." + mk |
| for _, r := range roots { |
| filename += "." + r |
| } |
| return url.QueryEscape(filename) |
| } |
| |
| func saveCache(g *DepGraph, roots []string) { |
| if len(g.accessedMks) == 0 { |
| panic("No Makefile is read") |
| } |
| cacheFile := cacheFilename(g.accessedMks[0].Filename, roots) |
| for _, mk := range g.accessedMks { |
| // Inconsistent, do not dump this result. |
| if mk.State == fileInconsistent { |
| if exists(cacheFile) { |
| os.Remove(cacheFile) |
| } |
| return |
| } |
| } |
| GOB.Save(g, cacheFile, roots) |
| } |
| |
| func deserializeSingleChild(sv serializableVar) Value { |
| if len(sv.Children) != 1 { |
| panic(fmt.Sprintf("unexpected number of children: %q", sv)) |
| } |
| return deserializeVar(sv.Children[0]) |
| } |
| |
| func deserializeVar(sv serializableVar) (r Value) { |
| switch sv.Type { |
| case "literal": |
| return literal(sv.V) |
| case "tmpval": |
| return tmpval([]byte(sv.V)) |
| case "expr": |
| var e expr |
| for _, v := range sv.Children { |
| e = append(e, deserializeVar(v)) |
| } |
| return e |
| case "varref": |
| return &varref{varname: deserializeSingleChild(sv)} |
| case "paramref": |
| v, err := strconv.Atoi(sv.V) |
| if err != nil { |
| panic(err) |
| } |
| return paramref(v) |
| case "varsubst": |
| return varsubst{ |
| varname: deserializeVar(sv.Children[0]), |
| pat: deserializeVar(sv.Children[1]), |
| subst: deserializeVar(sv.Children[2]), |
| } |
| |
| case "func": |
| name := deserializeVar(sv.Children[0]).(literal) |
| f := funcMap[string(name[1:])]() |
| f.AddArg(name) |
| for _, a := range sv.Children[1:] { |
| f.AddArg(deserializeVar(a)) |
| } |
| return f |
| case "funcEvalAssign": |
| return &funcEvalAssign{ |
| lhs: sv.Children[0].V, |
| op: sv.Children[1].V, |
| rhs: deserializeVar(sv.Children[2]), |
| } |
| case "funcNop": |
| return &funcNop{expr: sv.V} |
| |
| case "simple": |
| return &simpleVar{ |
| value: sv.V, |
| origin: sv.Origin, |
| } |
| case "recursive": |
| return &recursiveVar{ |
| expr: deserializeSingleChild(sv), |
| origin: sv.Origin, |
| } |
| |
| case ":=", "=", "+=", "?=": |
| return &targetSpecificVar{ |
| v: deserializeSingleChild(sv).(Var), |
| op: sv.Type, |
| } |
| |
| default: |
| panic(fmt.Sprintf("unknown serialized variable type: %q", sv)) |
| } |
| } |
| |
| func deserializeVars(vars map[string]serializableVar) Vars { |
| r := make(Vars) |
| for k, v := range vars { |
| r[k] = deserializeVar(v).(Var) |
| } |
| return r |
| } |
| |
| func deserializeNodes(g serializableGraph) (r []*DepNode) { |
| nodes := g.Nodes |
| tsvs := g.Tsvs |
| targets := g.Targets |
| // Deserialize all TSVs first so that multiple rules can share memory. |
| var tsvValues []Var |
| for _, sv := range tsvs { |
| tsvValues = append(tsvValues, deserializeVar(sv.Value).(Var)) |
| } |
| |
| nodeMap := make(map[string]*DepNode) |
| for _, n := range nodes { |
| var actualInputs []string |
| for _, i := range n.ActualInputs { |
| actualInputs = append(actualInputs, targets[i]) |
| } |
| |
| d := &DepNode{ |
| Output: targets[n.Output], |
| Cmds: n.Cmds, |
| HasRule: n.HasRule, |
| IsOrderOnly: n.IsOrderOnly, |
| IsPhony: n.IsPhony, |
| ActualInputs: actualInputs, |
| Filename: n.Filename, |
| Lineno: n.Lineno, |
| TargetSpecificVars: make(Vars), |
| } |
| |
| for _, id := range n.TargetSpecificVars { |
| sv := tsvs[id] |
| d.TargetSpecificVars[sv.Name] = tsvValues[id] |
| } |
| |
| nodeMap[targets[n.Output]] = d |
| r = append(r, d) |
| } |
| |
| for _, n := range nodes { |
| d := nodeMap[targets[n.Output]] |
| for _, o := range n.Deps { |
| c, present := nodeMap[targets[o]] |
| if !present { |
| panic(fmt.Sprintf("unknown target: %d (%s)", o, targets[o])) |
| } |
| d.Deps = append(d.Deps, c) |
| } |
| for _, o := range n.Parents { |
| c, present := nodeMap[targets[o]] |
| if !present { |
| panic(fmt.Sprintf("unknown target: %d (%s)", o, targets[o])) |
| } |
| d.Parents = append(d.Parents, c) |
| } |
| } |
| |
| return r |
| } |
| |
| func human(n int) string { |
| if n >= 10*1000*1000*1000 { |
| return fmt.Sprintf("%.2fGB", float32(n)/1000/1000/1000) |
| } |
| if n >= 10*1000*1000 { |
| return fmt.Sprintf("%.2fMB", float32(n)/1000/1000) |
| } |
| if n >= 10*1000 { |
| return fmt.Sprintf("%.2fkB", float32(n)/1000) |
| } |
| return fmt.Sprintf("%dB", n) |
| } |
| |
| func showSerializedNodesStats(nodes []*serializableDepNode) { |
| outputSize := 0 |
| cmdSize := 0 |
| depsSize := 0 |
| actualInputSize := 0 |
| tsvSize := 0 |
| filenameSize := 0 |
| linenoSize := 0 |
| for _, n := range nodes { |
| outputSize += 4 |
| for _, c := range n.Cmds { |
| cmdSize += len(c) |
| } |
| for _ = range n.Deps { |
| depsSize += 4 |
| } |
| for _ = range n.ActualInputs { |
| actualInputSize += 4 |
| } |
| for _ = range n.TargetSpecificVars { |
| tsvSize += 4 |
| } |
| filenameSize += len(n.Filename) |
| linenoSize += 4 |
| } |
| size := outputSize + cmdSize + depsSize + actualInputSize + tsvSize + filenameSize + linenoSize |
| logStats("%d nodes %s", len(nodes), human(size)) |
| logStats(" output %s", human(outputSize)) |
| logStats(" command %s", human(cmdSize)) |
| logStats(" deps %s", human(depsSize)) |
| logStats(" inputs %s", human(actualInputSize)) |
| logStats(" tsv %s", human(tsvSize)) |
| logStats(" filename %s", human(filenameSize)) |
| logStats(" lineno %s", human(linenoSize)) |
| } |
| |
| func (v serializableVar) size() int { |
| size := 0 |
| size += len(v.Type) |
| size += len(v.V) |
| size += len(v.Origin) |
| for _, c := range v.Children { |
| size += c.size() |
| } |
| return size |
| } |
| |
| func showSerializedVarsStats(vars map[string]serializableVar) { |
| nameSize := 0 |
| valueSize := 0 |
| for k, v := range vars { |
| nameSize += len(k) |
| valueSize += v.size() |
| } |
| size := nameSize + valueSize |
| logStats("%d vars %s", len(vars), human(size)) |
| logStats(" name %s", human(nameSize)) |
| logStats(" value %s", human(valueSize)) |
| } |
| |
| func showSerializedTsvsStats(vars []serializableTargetSpecificVar) { |
| nameSize := 0 |
| valueSize := 0 |
| for _, v := range vars { |
| nameSize += len(v.Name) |
| valueSize += v.Value.size() |
| } |
| size := nameSize + valueSize |
| logStats("%d tsvs %s", len(vars), human(size)) |
| logStats(" name %s", human(nameSize)) |
| logStats(" value %s", human(valueSize)) |
| } |
| |
| func showSerializedTargetsStats(targets []string) { |
| size := 0 |
| for _, t := range targets { |
| size += len(t) |
| } |
| logStats("%d targets %s", len(targets), human(size)) |
| } |
| |
| func showSerializedAccessedMksStats(accessedMks []*accessedMakefile) { |
| size := 0 |
| for _, rm := range accessedMks { |
| size += len(rm.Filename) + len(rm.Hash) + 4 |
| } |
| logStats("%d makefiles %s", len(accessedMks), human(size)) |
| } |
| |
| func showSerializedGraphStats(g serializableGraph) { |
| showSerializedNodesStats(g.Nodes) |
| showSerializedVarsStats(g.Vars) |
| showSerializedTsvsStats(g.Tsvs) |
| showSerializedTargetsStats(g.Targets) |
| showSerializedAccessedMksStats(g.AccessedMks) |
| } |
| |
| func deserializeGraph(g serializableGraph) *DepGraph { |
| if LogFlag || StatsFlag { |
| showSerializedGraphStats(g) |
| } |
| nodes := deserializeNodes(g) |
| vars := deserializeVars(g.Vars) |
| return &DepGraph{ |
| nodes: nodes, |
| vars: vars, |
| accessedMks: g.AccessedMks, |
| exports: g.Exports, |
| } |
| } |
| |
| func (jsonLoadSaver) Load(filename string) (*DepGraph, error) { |
| startTime := time.Now() |
| f, err := os.Open(filename) |
| if err != nil { |
| return nil, err |
| } |
| defer f.Close() |
| |
| d := json.NewDecoder(f) |
| g := serializableGraph{Vars: make(map[string]serializableVar)} |
| err = d.Decode(&g) |
| if err != nil { |
| return nil, err |
| } |
| dg := deserializeGraph(g) |
| logStats("gob deserialize time: %q", time.Since(startTime)) |
| return dg, nil |
| } |
| |
| func (gobLoadSaver) Load(filename string) (*DepGraph, error) { |
| startTime := time.Now() |
| f, err := os.Open(filename) |
| if err != nil { |
| return nil, err |
| } |
| defer f.Close() |
| |
| d := gob.NewDecoder(f) |
| g := serializableGraph{Vars: make(map[string]serializableVar)} |
| err = d.Decode(&g) |
| if err != nil { |
| return nil, err |
| } |
| dg := deserializeGraph(g) |
| logStats("json deserialize time: %q", time.Since(startTime)) |
| return dg, nil |
| } |
| |
| func loadCache(makefile string, roots []string) *DepGraph { |
| startTime := time.Now() |
| defer func() { |
| logStats("Cache lookup time: %q", time.Since(startTime)) |
| }() |
| |
| filename := cacheFilename(makefile, roots) |
| if !exists(filename) { |
| logAlways("Cache not found") |
| return nil |
| } |
| |
| g, err := GOB.Load(filename) |
| if err != nil { |
| logAlways("Cache load error: %v", err) |
| return nil |
| } |
| for _, mk := range g.accessedMks { |
| if mk.State != fileExists && mk.State != fileNotExists { |
| panic(fmt.Sprintf("Internal error: broken state: %d", mk.State)) |
| } |
| if mk.State == fileNotExists { |
| if exists(mk.Filename) { |
| logAlways("Cache expired: %s", mk.Filename) |
| return nil |
| } |
| } else { |
| c, err := ioutil.ReadFile(mk.Filename) |
| if err != nil { |
| logAlways("Cache expired: %s", mk.Filename) |
| return nil |
| } |
| h := sha1.Sum(c) |
| if !bytes.Equal(h[:], mk.Hash[:]) { |
| logAlways("Cache expired: %s", mk.Filename) |
| return nil |
| } |
| } |
| } |
| g.isCached = true |
| logAlways("Cache found!") |
| return g |
| } |