| package main |
| |
| import ( |
| "bytes" |
| "crypto/sha1" |
| "encoding/binary" |
| "encoding/gob" |
| "encoding/json" |
| "fmt" |
| "io" |
| "io/ioutil" |
| "net/url" |
| "os" |
| "sort" |
| "strconv" |
| "time" |
| ) |
| |
| const ( |
| VALUE_TYPE_RECURSIVE = 'R' |
| VALUE_TYPE_SIMPLE = 'S' |
| VALUE_TYPE_TSV = 'T' |
| VALUE_TYPE_UNDEFINED = 'U' |
| VALUE_TYPE_ASSIGN = 'a' |
| VALUE_TYPE_EXPR = 'e' |
| VALUE_TYPE_FUNC = 'f' |
| VALUE_TYPE_LITERAL = 'l' |
| VALUE_TYPE_NOP = 'n' |
| VALUE_TYPE_PARAMREF = 'p' |
| VALUE_TYPE_VARREF = 'r' |
| VALUE_TYPE_VARSUBST = 's' |
| VALUE_TYPE_TMPVAL = 't' |
| ) |
| |
| func dumpData(w io.Writer, data []byte) { |
| for len(data) != 0 { |
| written, err := w.Write(data) |
| if err == io.EOF { |
| return |
| } |
| if err != nil { |
| panic(err) |
| } |
| data = data[written:] |
| } |
| } |
| |
| func dumpInt(w io.Writer, i int) { |
| v := int32(i) |
| binary.Write(w, binary.LittleEndian, &v) |
| } |
| |
| func dumpString(w io.Writer, s string) { |
| dumpInt(w, len(s)) |
| dumpData(w, []byte(s)) |
| } |
| |
| func dumpBytes(w io.Writer, b []byte) { |
| dumpInt(w, len(b)) |
| dumpData(w, b) |
| } |
| |
| func dumpByte(w io.Writer, b byte) { |
| w.Write([]byte{b}) |
| } |
| |
| type SerializableVar struct { |
| Type string |
| V string |
| Origin string |
| Children []SerializableVar |
| } |
| |
| type SerializableDepNode struct { |
| Output int |
| Cmds []string |
| Deps []int |
| Parents []int |
| HasRule bool |
| IsOrderOnly bool |
| IsPhony bool |
| ActualInputs []int |
| TargetSpecificVars []int |
| Filename string |
| Lineno int |
| } |
| |
| type SerializableTargetSpecificVar struct { |
| Name string |
| Value SerializableVar |
| } |
| |
| type SerializableGraph struct { |
| Nodes []*SerializableDepNode |
| Vars map[string]SerializableVar |
| Tsvs []SerializableTargetSpecificVar |
| Targets []string |
| Roots []string |
| ReadMks []*ReadMakefile |
| Exports map[string]bool |
| } |
| |
| func encGob(v interface{}) string { |
| var buf bytes.Buffer |
| e := gob.NewEncoder(&buf) |
| err := e.Encode(v) |
| if err != nil { |
| panic(err) |
| } |
| return buf.String() |
| } |
| |
| func encVar(k string, v Var) string { |
| var buf bytes.Buffer |
| dumpString(&buf, k) |
| v.Dump(&buf) |
| return buf.String() |
| } |
| |
| type DepNodesSerializer struct { |
| nodes []*SerializableDepNode |
| tsvs []SerializableTargetSpecificVar |
| tsvMap map[string]int |
| targets []string |
| targetMap map[string]int |
| done map[string]bool |
| } |
| |
| func NewDepNodesSerializer() *DepNodesSerializer { |
| return &DepNodesSerializer{ |
| tsvMap: make(map[string]int), |
| targetMap: make(map[string]int), |
| done: make(map[string]bool), |
| } |
| } |
| |
| func (ns *DepNodesSerializer) SerializeTarget(t string) int { |
| id, present := ns.targetMap[t] |
| if present { |
| return id |
| } |
| id = len(ns.targets) |
| ns.targetMap[t] = id |
| ns.targets = append(ns.targets, t) |
| return id |
| } |
| |
| func (ns *DepNodesSerializer) SerializeDepNodes(nodes []*DepNode) { |
| for _, n := range nodes { |
| if ns.done[n.Output] { |
| continue |
| } |
| ns.done[n.Output] = true |
| |
| var deps []int |
| for _, d := range n.Deps { |
| deps = append(deps, ns.SerializeTarget(d.Output)) |
| } |
| var parents []int |
| for _, d := range n.Parents { |
| parents = append(parents, ns.SerializeTarget(d.Output)) |
| } |
| var actualInputs []int |
| for _, i := range n.ActualInputs { |
| actualInputs = append(actualInputs, ns.SerializeTarget(i)) |
| } |
| |
| // Sort keys for consistent serialization. |
| var tsvKeys []string |
| for k := range n.TargetSpecificVars { |
| tsvKeys = append(tsvKeys, k) |
| } |
| sort.Strings(tsvKeys) |
| |
| var vars []int |
| for _, k := range tsvKeys { |
| v := n.TargetSpecificVars[k] |
| sv := SerializableTargetSpecificVar{Name: k, Value: v.Serialize()} |
| //gob := encGob(sv) |
| gob := encVar(k, v) |
| id, present := ns.tsvMap[gob] |
| if !present { |
| id = len(ns.tsvs) |
| ns.tsvMap[gob] = id |
| ns.tsvs = append(ns.tsvs, sv) |
| } |
| vars = append(vars, id) |
| } |
| |
| ns.nodes = append(ns.nodes, &SerializableDepNode{ |
| Output: ns.SerializeTarget(n.Output), |
| Cmds: n.Cmds, |
| Deps: deps, |
| Parents: parents, |
| HasRule: n.HasRule, |
| IsOrderOnly: n.IsOrderOnly, |
| IsPhony: n.IsPhony, |
| ActualInputs: actualInputs, |
| TargetSpecificVars: vars, |
| Filename: n.Filename, |
| Lineno: n.Lineno, |
| }) |
| ns.SerializeDepNodes(n.Deps) |
| } |
| } |
| |
| func MakeSerializableVars(vars Vars) (r map[string]SerializableVar) { |
| r = make(map[string]SerializableVar) |
| for k, v := range vars { |
| r[k] = v.Serialize() |
| } |
| return r |
| } |
| |
| func MakeSerializableGraph(g *DepGraph, roots []string) SerializableGraph { |
| ns := NewDepNodesSerializer() |
| ns.SerializeDepNodes(g.nodes) |
| v := MakeSerializableVars(g.vars) |
| return SerializableGraph{ |
| Nodes: ns.nodes, |
| Vars: v, |
| Tsvs: ns.tsvs, |
| Targets: ns.targets, |
| Roots: roots, |
| ReadMks: g.readMks, |
| Exports: g.exports, |
| } |
| } |
| |
| func DumpDepGraphAsJson(g *DepGraph, filename string, roots []string) { |
| sg := MakeSerializableGraph(g, roots) |
| o, err := json.MarshalIndent(sg, " ", " ") |
| if err != nil { |
| panic(err) |
| } |
| f, err2 := os.Create(filename) |
| if err2 != nil { |
| panic(err2) |
| } |
| f.Write(o) |
| err = f.Close() |
| if err != nil { |
| panic(err) |
| } |
| } |
| |
| func DumpDepGraph(g *DepGraph, filename string, roots []string) { |
| f, err := os.Create(filename) |
| if err != nil { |
| panic(err) |
| } |
| e := gob.NewEncoder(f) |
| startTime := time.Now() |
| sg := MakeSerializableGraph(g, roots) |
| LogStats("serialize prepare time: %q", time.Now().Sub(startTime)) |
| startTime = time.Now() |
| e.Encode(sg) |
| LogStats("serialize output time: %q", time.Now().Sub(startTime)) |
| err = f.Close() |
| if err != nil { |
| panic(err) |
| } |
| } |
| |
| func GetCacheFilename(mk string, roots []string) string { |
| filename := ".kati_cache." + mk |
| for _, r := range roots { |
| filename += "." + r |
| } |
| return url.QueryEscape(filename) |
| } |
| |
| func DumpDepGraphCache(g *DepGraph, roots []string) { |
| if len(g.readMks) == 0 { |
| panic("No Makefile is read") |
| } |
| cacheFile := GetCacheFilename(g.readMks[0].Filename, roots) |
| for _, mk := range g.readMks { |
| // Inconsistent, do not dump this result. |
| if mk.State == 2 { |
| if exists(cacheFile) { |
| os.Remove(cacheFile) |
| } |
| return |
| } |
| } |
| DumpDepGraph(g, cacheFile, roots) |
| } |
| |
| func DeserializeSingleChild(sv SerializableVar) Value { |
| if len(sv.Children) != 1 { |
| panic(fmt.Sprintf("unexpected number of children: %q", sv)) |
| } |
| return DeserializeVar(sv.Children[0]) |
| } |
| |
| func DeserializeVar(sv SerializableVar) (r Value) { |
| switch sv.Type { |
| case "literal": |
| return literal(sv.V) |
| case "tmpval": |
| return tmpval([]byte(sv.V)) |
| case "expr": |
| var e Expr |
| for _, v := range sv.Children { |
| e = append(e, DeserializeVar(v)) |
| } |
| return e |
| case "varref": |
| return varref{varname: DeserializeSingleChild(sv)} |
| case "paramref": |
| v, err := strconv.Atoi(sv.V) |
| if err != nil { |
| panic(err) |
| } |
| return paramref(v) |
| case "varsubst": |
| return varsubst{ |
| varname: DeserializeVar(sv.Children[0]), |
| pat: DeserializeVar(sv.Children[1]), |
| subst: DeserializeVar(sv.Children[2]), |
| } |
| |
| case "func": |
| name := DeserializeVar(sv.Children[0]).(literal) |
| f := funcMap[string(name[1:])]() |
| f.AddArg(name) |
| for _, a := range sv.Children[1:] { |
| f.AddArg(DeserializeVar(a)) |
| } |
| return f |
| case "funcEvalAssign": |
| return &funcEvalAssign{ |
| lhs: sv.Children[0].V, |
| op: sv.Children[1].V, |
| rhs: DeserializeVar(sv.Children[2]), |
| } |
| case "funcNop": |
| return &funcNop{expr: sv.V} |
| |
| case "simple": |
| return SimpleVar{ |
| value: []byte(sv.V), |
| origin: sv.Origin, |
| } |
| case "recursive": |
| return RecursiveVar{ |
| expr: DeserializeSingleChild(sv), |
| origin: sv.Origin, |
| } |
| |
| case ":=", "=", "+=", "?=": |
| return TargetSpecificVar{ |
| v: DeserializeSingleChild(sv).(Var), |
| op: sv.Type, |
| } |
| |
| default: |
| panic(fmt.Sprintf("unknown serialized variable type: %q", sv)) |
| } |
| return UndefinedVar{} |
| } |
| |
| func DeserializeVars(vars map[string]SerializableVar) Vars { |
| r := make(Vars) |
| for k, v := range vars { |
| r[k] = DeserializeVar(v).(Var) |
| } |
| return r |
| } |
| |
| func DeserializeNodes(g SerializableGraph) (r []*DepNode) { |
| nodes := g.Nodes |
| tsvs := g.Tsvs |
| targets := g.Targets |
| // Deserialize all TSVs first so that multiple rules can share memory. |
| var tsvValues []Var |
| for _, sv := range tsvs { |
| tsvValues = append(tsvValues, DeserializeVar(sv.Value).(Var)) |
| } |
| |
| nodeMap := make(map[string]*DepNode) |
| for _, n := range nodes { |
| var actualInputs []string |
| for _, i := range n.ActualInputs { |
| actualInputs = append(actualInputs, targets[i]) |
| } |
| |
| d := &DepNode{ |
| Output: targets[n.Output], |
| Cmds: n.Cmds, |
| HasRule: n.HasRule, |
| IsOrderOnly: n.IsOrderOnly, |
| IsPhony: n.IsPhony, |
| ActualInputs: actualInputs, |
| Filename: n.Filename, |
| Lineno: n.Lineno, |
| TargetSpecificVars: make(Vars), |
| } |
| |
| for _, id := range n.TargetSpecificVars { |
| sv := tsvs[id] |
| d.TargetSpecificVars[sv.Name] = tsvValues[id] |
| } |
| |
| nodeMap[targets[n.Output]] = d |
| r = append(r, d) |
| } |
| |
| for _, n := range nodes { |
| d := nodeMap[targets[n.Output]] |
| for _, o := range n.Deps { |
| c, present := nodeMap[targets[o]] |
| if !present { |
| panic(fmt.Sprintf("unknown target: %s", o)) |
| } |
| d.Deps = append(d.Deps, c) |
| } |
| for _, o := range n.Parents { |
| c, present := nodeMap[targets[o]] |
| if !present { |
| panic(fmt.Sprintf("unknown target: %s", o)) |
| } |
| d.Parents = append(d.Parents, c) |
| } |
| } |
| |
| return r |
| } |
| |
| func human(n int) string { |
| if n >= 10*1000*1000*1000 { |
| return fmt.Sprintf("%.2fGB", float32(n)/1000/1000/1000) |
| } else if n >= 10*1000*1000 { |
| return fmt.Sprintf("%.2fMB", float32(n)/1000/1000) |
| } else if n >= 10*1000 { |
| return fmt.Sprintf("%.2fkB", float32(n)/1000) |
| } else { |
| return fmt.Sprintf("%dB", n) |
| } |
| } |
| |
| func showSerializedNodesStats(nodes []*SerializableDepNode) { |
| outputSize := 0 |
| cmdSize := 0 |
| depsSize := 0 |
| actualInputSize := 0 |
| tsvSize := 0 |
| filenameSize := 0 |
| linenoSize := 0 |
| for _, n := range nodes { |
| outputSize += 4 |
| for _, c := range n.Cmds { |
| cmdSize += len(c) |
| } |
| for _ = range n.Deps { |
| depsSize += 4 |
| } |
| for _ = range n.ActualInputs { |
| actualInputSize += 4 |
| } |
| for _ = range n.TargetSpecificVars { |
| tsvSize += 4 |
| } |
| filenameSize += len(n.Filename) |
| linenoSize += 4 |
| } |
| size := outputSize + cmdSize + depsSize + actualInputSize + tsvSize + filenameSize + linenoSize |
| LogStats("%d nodes %s", len(nodes), human(size)) |
| LogStats(" output %s", human(outputSize)) |
| LogStats(" command %s", human(cmdSize)) |
| LogStats(" deps %s", human(depsSize)) |
| LogStats(" inputs %s", human(actualInputSize)) |
| LogStats(" tsv %s", human(tsvSize)) |
| LogStats(" filename %s", human(filenameSize)) |
| LogStats(" lineno %s", human(linenoSize)) |
| } |
| |
| func (v SerializableVar) size() int { |
| size := 0 |
| size += len(v.Type) |
| size += len(v.V) |
| size += len(v.Origin) |
| for _, c := range v.Children { |
| size += c.size() |
| } |
| return size |
| } |
| |
| func showSerializedVarsStats(vars map[string]SerializableVar) { |
| nameSize := 0 |
| valueSize := 0 |
| for k, v := range vars { |
| nameSize += len(k) |
| valueSize += v.size() |
| } |
| size := nameSize + valueSize |
| LogStats("%d vars %s", len(vars), human(size)) |
| LogStats(" name %s", human(nameSize)) |
| LogStats(" value %s", human(valueSize)) |
| } |
| |
| func showSerializedTsvsStats(vars []SerializableTargetSpecificVar) { |
| nameSize := 0 |
| valueSize := 0 |
| for _, v := range vars { |
| nameSize += len(v.Name) |
| valueSize += v.Value.size() |
| } |
| size := nameSize + valueSize |
| LogStats("%d tsvs %s", len(vars), human(size)) |
| LogStats(" name %s", human(nameSize)) |
| LogStats(" value %s", human(valueSize)) |
| } |
| |
| func showSerializedTargetsStats(targets []string) { |
| size := 0 |
| for _, t := range targets { |
| size += len(t) |
| } |
| LogStats("%d targets %s", len(targets), human(size)) |
| } |
| |
| func showSerializedReadMksStats(readMks []*ReadMakefile) { |
| size := 0 |
| for _, rm := range readMks { |
| size += len(rm.Filename) + len(rm.Hash) + 4 |
| } |
| LogStats("%d makefiles %s", len(readMks), human(size)) |
| } |
| |
| func showSerializedGraphStats(g SerializableGraph) { |
| showSerializedNodesStats(g.Nodes) |
| showSerializedVarsStats(g.Vars) |
| showSerializedTsvsStats(g.Tsvs) |
| showSerializedTargetsStats(g.Targets) |
| showSerializedReadMksStats(g.ReadMks) |
| } |
| |
| func DeserializeGraph(g SerializableGraph) *DepGraph { |
| if katiLogFlag || katiStatsFlag { |
| showSerializedGraphStats(g) |
| } |
| nodes := DeserializeNodes(g) |
| vars := DeserializeVars(g.Vars) |
| return &DepGraph{ |
| nodes: nodes, |
| vars: vars, |
| readMks: g.ReadMks, |
| exports: g.Exports, |
| } |
| } |
| |
| func LoadDepGraphFromJson(filename string) *DepGraph { |
| f, err := os.Open(filename) |
| if err != nil { |
| panic(err) |
| } |
| defer f.Close() |
| |
| d := json.NewDecoder(f) |
| g := SerializableGraph{Vars: make(map[string]SerializableVar)} |
| err = d.Decode(&g) |
| if err != nil { |
| panic(err) |
| } |
| return DeserializeGraph(g) |
| } |
| |
| func LoadDepGraph(filename string) *DepGraph { |
| f, err := os.Open(filename) |
| if err != nil { |
| panic(err) |
| } |
| defer f.Close() |
| |
| d := gob.NewDecoder(f) |
| g := SerializableGraph{Vars: make(map[string]SerializableVar)} |
| err = d.Decode(&g) |
| if err != nil { |
| panic(err) |
| } |
| return DeserializeGraph(g) |
| } |
| |
| func LoadDepGraphCache(makefile string, roots []string) *DepGraph { |
| startTime := time.Now() |
| defer func() { |
| LogStats("Cache lookup time: %q", time.Now().Sub(startTime)) |
| }() |
| |
| filename := GetCacheFilename(makefile, roots) |
| if !exists(filename) { |
| LogAlways("Cache not found") |
| return nil |
| } |
| |
| g := LoadDepGraph(filename) |
| for _, mk := range g.readMks { |
| if mk.State != FILE_EXISTS && mk.State != FILE_NOT_EXISTS { |
| panic(fmt.Sprintf("Internal error: broken state: %d", mk.State)) |
| } |
| if mk.State == FILE_NOT_EXISTS { |
| if exists(mk.Filename) { |
| LogAlways("Cache expired: %s", mk.Filename) |
| return nil |
| } |
| } else { |
| c, err := ioutil.ReadFile(mk.Filename) |
| if err != nil { |
| LogAlways("Cache expired: %s", mk.Filename) |
| return nil |
| } |
| h := sha1.Sum(c) |
| if !bytes.Equal(h[:], mk.Hash[:]) { |
| LogAlways("Cache expired: %s", mk.Filename) |
| return nil |
| } |
| } |
| } |
| g.isCached = true |
| LogAlways("Cache found!") |
| return g |
| } |