| // Copyright 2016 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| package main |
| |
| /* |
| Generate the tasks.json file. |
| */ |
| |
| import ( |
| "encoding/json" |
| "flag" |
| "fmt" |
| "io/ioutil" |
| "os" |
| "path" |
| "path/filepath" |
| "regexp" |
| "runtime" |
| "sort" |
| "strconv" |
| "strings" |
| "time" |
| |
| "github.com/skia-dev/glog" |
| "go.skia.org/infra/go/sklog" |
| "go.skia.org/infra/go/util" |
| "go.skia.org/infra/task_scheduler/go/specs" |
| ) |
| |
| const ( |
| BUNDLE_RECIPES_NAME = "Housekeeper-PerCommit-BundleRecipes" |
| ISOLATE_SKIMAGE_NAME = "Housekeeper-PerCommit-IsolateSkImage" |
| ISOLATE_SKP_NAME = "Housekeeper-PerCommit-IsolateSKP" |
| ISOLATE_SVG_NAME = "Housekeeper-PerCommit-IsolateSVG" |
| ISOLATE_NDK_LINUX_NAME = "Housekeeper-PerCommit-IsolateAndroidNDKLinux" |
| ISOLATE_WIN_TOOLCHAIN_NAME = "Housekeeper-PerCommit-IsolateWinToolchain" |
| ISOLATE_WIN_VULKAN_SDK_NAME = "Housekeeper-PerCommit-IsolateWinVulkanSDK" |
| |
| DEFAULT_OS_DEBIAN = "Debian-9.1" |
| DEFAULT_OS_LINUX_GCE = "Debian-9.2" |
| DEFAULT_OS_MAC = "Mac-10.13.1" |
| DEFAULT_OS_UBUNTU = "Ubuntu-14.04" |
| DEFAULT_OS_WIN = "Windows-2016Server-14393" |
| |
| // Name prefix for upload jobs. |
| PREFIX_UPLOAD = "Upload" |
| ) |
| |
| var ( |
| // "Constants" |
| |
| // Top-level list of all jobs to run at each commit; loaded from |
| // jobs.json. |
| JOBS []string |
| |
| // General configuration information. |
| CONFIG struct { |
| GsBucketCoverage string `json:"gs_bucket_coverage"` |
| GsBucketGm string `json:"gs_bucket_gm"` |
| GsBucketNano string `json:"gs_bucket_nano"` |
| GsBucketCalm string `json:"gs_bucket_calm"` |
| NoUpload []string `json:"no_upload"` |
| Pool string `json:"pool"` |
| } |
| |
| // alternateSwarmDimensions can be set in an init function to override the default swarming bot |
| // dimensions for the given task. |
| alternateSwarmDimensions func(parts map[string]string) []string |
| |
| // internalHardwareLabelFn can be set in an init function to provide an |
| // internal_hardware_label variable to the recipe. |
| internalHardwareLabelFn func(parts map[string]string) *int |
| |
| // Defines the structure of job names. |
| jobNameSchema *JobNameSchema |
| |
| // Git 2.13. |
| cipdGit1 = &specs.CipdPackage{ |
| Name: fmt.Sprintf("infra/git/${platform}"), |
| Path: "git", |
| Version: fmt.Sprintf("version:2.13.0.chromium9"), |
| } |
| cipdGit2 = &specs.CipdPackage{ |
| Name: fmt.Sprintf("infra/tools/git/${platform}"), |
| Path: "git", |
| Version: fmt.Sprintf("git_revision:a78b5f3658c0578a017db48df97d20ac09822bcd"), |
| } |
| |
| // Flags. |
| builderNameSchemaFile = flag.String("builder_name_schema", "", "Path to the builder_name_schema.json file. If not specified, uses infra/bots/recipe_modules/builder_name_schema/builder_name_schema.json from this repo.") |
| assetsDir = flag.String("assets_dir", "", "Directory containing assets.") |
| cfgFile = flag.String("cfg_file", "", "JSON file containing general configuration information.") |
| jobsFile = flag.String("jobs", "", "JSON file containing jobs to run.") |
| ) |
| |
| // internalHardwareLabel returns the internal ID for the bot, if any. |
| func internalHardwareLabel(parts map[string]string) *int { |
| if internalHardwareLabelFn != nil { |
| return internalHardwareLabelFn(parts) |
| } |
| return nil |
| } |
| |
| // linuxGceDimensions are the Swarming dimensions for Linux GCE |
| // instances. |
| func linuxGceDimensions() []string { |
| return []string{ |
| // Specify CPU to avoid running builds on bots with a more unique CPU. |
| "cpu:x86-64-Haswell_GCE", |
| "gpu:none", |
| fmt.Sprintf("os:%s", DEFAULT_OS_LINUX_GCE), |
| fmt.Sprintf("pool:%s", CONFIG.Pool), |
| } |
| } |
| |
| // deriveCompileTaskName returns the name of a compile task based on the given |
| // job name. |
| func deriveCompileTaskName(jobName string, parts map[string]string) string { |
| if jobName == "Housekeeper-Nightly-Bookmaker" { |
| return "Build-Debian9-GCC-x86_64-Release" |
| } else if parts["role"] == "Housekeeper" { |
| return "Build-Debian9-GCC-x86_64-Release-Shared" |
| } else if parts["role"] == "Test" || parts["role"] == "Perf" { |
| task_os := parts["os"] |
| ec := []string{} |
| if val := parts["extra_config"]; val != "" { |
| ec = strings.Split(val, "_") |
| ignore := []string{"Skpbench", "AbandonGpuContext", "PreAbandonGpuContext", "Valgrind", "ReleaseAndAbandonGpuContext", "CCPR", "FSAA", "FAAA", "FDAA", "NativeFonts", "GDI", "NoGPUThreads"} |
| keep := make([]string, 0, len(ec)) |
| for _, part := range ec { |
| if !util.In(part, ignore) { |
| keep = append(keep, part) |
| } |
| } |
| ec = keep |
| } |
| if task_os == "Android" { |
| if !util.In("Android", ec) { |
| ec = append([]string{"Android"}, ec...) |
| } |
| task_os = "Debian9" |
| } else if task_os == "Chromecast" { |
| task_os = "Debian9" |
| ec = append([]string{"Chromecast"}, ec...) |
| } else if strings.Contains(task_os, "ChromeOS") { |
| ec = append([]string{"Chromebook", "GLES"}, ec...) |
| task_os = "Debian9" |
| } else if task_os == "iOS" { |
| ec = append([]string{task_os}, ec...) |
| task_os = "Mac" |
| } else if strings.Contains(task_os, "Win") { |
| task_os = "Win" |
| } else if strings.Contains(task_os, "Ubuntu") || strings.Contains(task_os, "Debian") { |
| task_os = "Debian9" |
| } |
| jobNameMap := map[string]string{ |
| "role": "Build", |
| "os": task_os, |
| "compiler": parts["compiler"], |
| "target_arch": parts["arch"], |
| "configuration": parts["configuration"], |
| } |
| if len(ec) > 0 { |
| jobNameMap["extra_config"] = strings.Join(ec, "_") |
| } |
| name, err := jobNameSchema.MakeJobName(jobNameMap) |
| if err != nil { |
| glog.Fatal(err) |
| } |
| return name |
| } else { |
| return jobName |
| } |
| } |
| |
| // swarmDimensions generates swarming bot dimensions for the given task. |
| func swarmDimensions(parts map[string]string) []string { |
| if alternateSwarmDimensions != nil { |
| return alternateSwarmDimensions(parts) |
| } |
| return defaultSwarmDimensions(parts) |
| } |
| |
| // defaultSwarmDimensions generates default swarming bot dimensions for the given task. |
| func defaultSwarmDimensions(parts map[string]string) []string { |
| d := map[string]string{ |
| "pool": CONFIG.Pool, |
| } |
| if os, ok := parts["os"]; ok { |
| d["os"], ok = map[string]string{ |
| "Android": "Android", |
| "Chromecast": "Android", |
| "ChromeOS": "ChromeOS", |
| "Debian9": DEFAULT_OS_DEBIAN, |
| "Mac": DEFAULT_OS_MAC, |
| "Ubuntu14": DEFAULT_OS_UBUNTU, |
| "Ubuntu16": "Ubuntu-16.10", |
| "Ubuntu17": "Ubuntu-17.04", |
| "Win": DEFAULT_OS_WIN, |
| "Win10": "Windows-10-15063", |
| "Win2k8": "Windows-2008ServerR2-SP1", |
| "Win2016": DEFAULT_OS_WIN, |
| "Win7": "Windows-7-SP1", |
| "Win8": "Windows-8.1-SP0", |
| "iOS": "iOS-10.3.1", |
| }[os] |
| if !ok { |
| glog.Fatalf("Entry %q not found in OS mapping.", os) |
| } |
| } else { |
| d["os"] = DEFAULT_OS_DEBIAN |
| } |
| if parts["role"] == "Test" || parts["role"] == "Perf" || parts["role"] == "Calmbench" { |
| if strings.Contains(parts["os"], "Android") || strings.Contains(parts["os"], "Chromecast") { |
| // For Android, the device type is a better dimension |
| // than CPU or GPU. |
| deviceInfo, ok := map[string][]string{ |
| "AndroidOne": {"sprout", "MOB30Q"}, |
| "Chorizo": {"chorizo", "1.24_82923"}, |
| "GalaxyS6": {"zerofltetmo", "NRD90M_G920TUVU5FQK1"}, |
| "GalaxyS7_G930A": {"heroqlteatt", "NRD90M_G930AUCS4BQC2"}, |
| "GalaxyS7_G930FD": {"herolte", "NRD90M_G930FXXU1DQAS"}, |
| "MotoG4": {"athene", "NPJ25.93-14"}, |
| "NVIDIA_Shield": {"foster", "NRD90M_1915764_848"}, |
| "Nexus5": {"hammerhead", "M4B30Z_3437181"}, |
| "Nexus5x": {"bullhead", "OPR6.170623.023"}, |
| "Nexus7": {"grouper", "LMY47V_1836172"}, // 2012 Nexus 7 |
| "NexusPlayer": {"fugu", "OPR6.170623.021"}, |
| "Pixel": {"sailfish", "OPR3.170623.008"}, |
| "Pixel2XL": {"taimen", "OPD1.170816.023"}, |
| "PixelC": {"dragon", "OPR1.170623.034"}, |
| "PixelXL": {"marlin", "OPR3.170623.008"}, |
| }[parts["model"]] |
| if !ok { |
| glog.Fatalf("Entry %q not found in Android mapping.", parts["model"]) |
| } |
| d["device_type"] = deviceInfo[0] |
| d["device_os"] = deviceInfo[1] |
| // TODO(kjlubick): Remove the python dimension after we have removed the |
| // Nexus5x devices from the local lab (on Monday, Dec 11, 2017 should be fine). |
| d["python"] = "2.7.9" // This indicates a RPI, e.g. in Skolo. Golo is 2.7.12 |
| if parts["model"] == "Nexus5x" { |
| d["python"] = "2.7.12" |
| } |
| } else if strings.Contains(parts["os"], "iOS") { |
| device, ok := map[string]string{ |
| "iPadMini4": "iPad5,1", |
| "iPhone6": "iPhone7,2", |
| "iPhone7": "iPhone9,1", |
| "iPadPro": "iPad6,3", |
| }[parts["model"]] |
| if !ok { |
| glog.Fatalf("Entry %q not found in iOS mapping.", parts["model"]) |
| } |
| d["device"] = device |
| } else if parts["cpu_or_gpu"] == "CPU" { |
| modelMapping, ok := map[string]map[string]string{ |
| "AVX": { |
| "MacMini7.1": "x86-64-E5-2697_v2", |
| "Golo": "x86-64-E5-2670", |
| }, |
| "AVX2": { |
| "GCE": "x86-64-Haswell_GCE", |
| "NUC5i7RYH": "x86-64-i7-5557U", |
| }, |
| "AVX512": { |
| "GCE": "x86-64-Skylake_GCE", |
| }, |
| }[parts["cpu_or_gpu_value"]] |
| if !ok { |
| glog.Fatalf("Entry %q not found in CPU mapping.", parts["cpu_or_gpu_value"]) |
| } |
| cpu, ok := modelMapping[parts["model"]] |
| if !ok { |
| glog.Fatalf("Entry %q not found in %q model mapping.", parts["model"], parts["cpu_or_gpu_value"]) |
| } |
| d["cpu"] = cpu |
| if parts["model"] == "GCE" && d["os"] == DEFAULT_OS_DEBIAN { |
| d["os"] = DEFAULT_OS_LINUX_GCE |
| } |
| if parts["model"] == "GCE" && d["os"] == DEFAULT_OS_WIN { |
| // Use normal-size machines for Test and Perf tasks on Win GCE. |
| d["machine_type"] = "n1-standard-16" |
| } |
| } else { |
| if strings.Contains(parts["os"], "Win") { |
| gpu, ok := map[string]string{ |
| "GT610": "10de:104a-22.21.13.8205", |
| "GTX1070": "10de:1ba1-23.21.13.8813", |
| "GTX660": "10de:11c0-23.21.13.8813", |
| "GTX960": "10de:1401-23.21.13.8813", |
| "IntelHD530": "8086:1912-21.20.16.4590", |
| "IntelHD4400": "8086:0a16-20.19.15.4703", |
| "IntelHD4600": "8086:0412-20.19.15.4703", |
| "IntelIris540": "8086:1926-21.20.16.4590", |
| "IntelIris6100": "8086:162b-20.19.15.4703", |
| "RadeonHD7770": "1002:683d-22.19.165.512", |
| "RadeonR9M470X": "1002:6646-22.19.165.512", |
| "QuadroP400": "10de:1cb3-22.21.13.8205", |
| }[parts["cpu_or_gpu_value"]] |
| if !ok { |
| glog.Fatalf("Entry %q not found in Win GPU mapping.", parts["cpu_or_gpu_value"]) |
| } |
| d["gpu"] = gpu |
| |
| // Specify cpu dimension for NUCs and ShuttleCs. We temporarily have two |
| // types of machines with a GTX960. |
| cpu, ok := map[string]string{ |
| "NUC6i7KYK": "x86-64-i7-6770HQ", |
| "ShuttleC": "x86-64-i7-6700K", |
| }[parts["model"]] |
| if ok { |
| d["cpu"] = cpu |
| } |
| } else if strings.Contains(parts["os"], "Ubuntu") || strings.Contains(parts["os"], "Debian") { |
| gpu, ok := map[string]string{ |
| // Intel drivers come from CIPD, so no need to specify the version here. |
| "IntelBayTrail": "8086:0f31", |
| "IntelHD2000": "8086:0102", |
| "IntelHD405": "8086:22b1", |
| "IntelIris640": "8086:5926", |
| "QuadroP400": "10de:1cb3-384.59", |
| }[parts["cpu_or_gpu_value"]] |
| if !ok { |
| glog.Fatalf("Entry %q not found in Ubuntu GPU mapping.", parts["cpu_or_gpu_value"]) |
| } |
| d["gpu"] = gpu |
| } else if strings.Contains(parts["os"], "Mac") { |
| gpu, ok := map[string]string{ |
| "IntelIris5100": "8086:0a2e", |
| }[parts["cpu_or_gpu_value"]] |
| if !ok { |
| glog.Fatalf("Entry %q not found in Mac GPU mapping.", parts["cpu_or_gpu_value"]) |
| } |
| d["gpu"] = gpu |
| } else if strings.Contains(parts["os"], "ChromeOS") { |
| version, ok := map[string]string{ |
| "MaliT604": "9901.12.0", |
| "MaliT764": "10172.0.0", |
| "MaliT860": "10172.0.0", |
| "PowerVRGX6250": "10176.5.0", |
| "TegraK1": "10172.0.0", |
| "IntelHDGraphics615": "10032.17.0", |
| }[parts["cpu_or_gpu_value"]] |
| if !ok { |
| glog.Fatalf("Entry %q not found in ChromeOS GPU mapping.", parts["cpu_or_gpu_value"]) |
| } |
| d["gpu"] = parts["cpu_or_gpu_value"] |
| d["release_version"] = version |
| } else { |
| glog.Fatalf("Unknown GPU mapping for OS %q.", parts["os"]) |
| } |
| } |
| } else { |
| d["gpu"] = "none" |
| if d["os"] == DEFAULT_OS_DEBIAN { |
| return linuxGceDimensions() |
| } else if d["os"] == DEFAULT_OS_WIN { |
| // Windows CPU bots. |
| d["cpu"] = "x86-64-Haswell_GCE" |
| // Use many-core machines for Build tasks on Win GCE, except for Goma. |
| if strings.Contains(parts["extra_config"], "Goma") { |
| d["machine_type"] = "n1-standard-16" |
| } else { |
| d["machine_type"] = "n1-highcpu-64" |
| } |
| } else if d["os"] == DEFAULT_OS_MAC { |
| // Mac CPU bots. |
| d["cpu"] = "x86-64-E5-2697_v2" |
| // skia:7408 |
| d["cert"] = "1" |
| } |
| } |
| |
| rv := make([]string, 0, len(d)) |
| for k, v := range d { |
| rv = append(rv, fmt.Sprintf("%s:%s", k, v)) |
| } |
| sort.Strings(rv) |
| return rv |
| } |
| |
| // relpath returns the relative path to the given file from the config file. |
| func relpath(f string) string { |
| _, filename, _, _ := runtime.Caller(0) |
| dir := path.Dir(filename) |
| rel := dir |
| if *cfgFile != "" { |
| rel = path.Dir(*cfgFile) |
| } |
| rv, err := filepath.Rel(rel, path.Join(dir, f)) |
| if err != nil { |
| sklog.Fatal(err) |
| } |
| return rv |
| } |
| |
| // bundleRecipes generates the task to bundle and isolate the recipes. |
| func bundleRecipes(b *specs.TasksCfgBuilder) string { |
| b.MustAddTask(BUNDLE_RECIPES_NAME, &specs.TaskSpec{ |
| CipdPackages: []*specs.CipdPackage{cipdGit1, cipdGit2}, |
| Dimensions: linuxGceDimensions(), |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "bundle_recipes", |
| fmt.Sprintf("buildername=%s", BUNDLE_RECIPES_NAME), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| }, |
| Isolate: relpath("bundle_recipes.isolate"), |
| Priority: 0.7, |
| }) |
| return BUNDLE_RECIPES_NAME |
| } |
| |
| // useBundledRecipes returns true iff the given bot should use bundled recipes |
| // instead of syncing recipe DEPS itself. |
| func useBundledRecipes(parts map[string]string) bool { |
| // Use bundled recipes for all test/perf tasks. |
| return true |
| } |
| |
| type isolateAssetCfg struct { |
| isolateFile string |
| cipdPkg string |
| } |
| |
| var ISOLATE_ASSET_MAPPING = map[string]isolateAssetCfg{ |
| ISOLATE_SKIMAGE_NAME: { |
| isolateFile: "isolate_skimage.isolate", |
| cipdPkg: "skimage", |
| }, |
| ISOLATE_SKP_NAME: { |
| isolateFile: "isolate_skp.isolate", |
| cipdPkg: "skp", |
| }, |
| ISOLATE_SVG_NAME: { |
| isolateFile: "isolate_svg.isolate", |
| cipdPkg: "svg", |
| }, |
| ISOLATE_NDK_LINUX_NAME: { |
| isolateFile: "isolate_ndk_linux.isolate", |
| cipdPkg: "android_ndk_linux", |
| }, |
| ISOLATE_WIN_TOOLCHAIN_NAME: { |
| isolateFile: "isolate_win_toolchain.isolate", |
| cipdPkg: "win_toolchain", |
| }, |
| ISOLATE_WIN_VULKAN_SDK_NAME: { |
| isolateFile: "isolate_win_vulkan_sdk.isolate", |
| cipdPkg: "win_vulkan_sdk", |
| }, |
| } |
| |
| // bundleRecipes generates the task to bundle and isolate the recipes. |
| func isolateCIPDAsset(b *specs.TasksCfgBuilder, name string) string { |
| b.MustAddTask(name, &specs.TaskSpec{ |
| CipdPackages: []*specs.CipdPackage{ |
| b.MustGetCipdPackageFromAsset(ISOLATE_ASSET_MAPPING[name].cipdPkg), |
| }, |
| Dimensions: linuxGceDimensions(), |
| Isolate: relpath(ISOLATE_ASSET_MAPPING[name].isolateFile), |
| Priority: 0.7, |
| }) |
| return name |
| } |
| |
| // getIsolatedCIPDDeps returns the slice of Isolate_* tasks a given task needs. |
| // This allows us to save time on I/O bound bots, like the RPIs. |
| func getIsolatedCIPDDeps(parts map[string]string) []string { |
| deps := []string{} |
| // Only do this on the RPIs for now. Other, faster machines shouldn't see much |
| // benefit and we don't need the extra complexity, for now |
| rpiOS := []string{"Android", "ChromeOS", "iOS"} |
| |
| if o := parts["os"]; strings.Contains(o, "Chromecast") { |
| // Chromecasts don't have enough disk space to fit all of the content, |
| // so we do a subset of the skps. |
| deps = append(deps, ISOLATE_SKP_NAME) |
| } else if e := parts["extra_config"]; strings.Contains(e, "Skpbench") { |
| // Skpbench only needs skps |
| deps = append(deps, ISOLATE_SKP_NAME) |
| } else if util.In(o, rpiOS) { |
| deps = append(deps, ISOLATE_SKP_NAME) |
| deps = append(deps, ISOLATE_SVG_NAME) |
| deps = append(deps, ISOLATE_SKIMAGE_NAME) |
| } |
| |
| return deps |
| } |
| |
| // compile generates a compile task. Returns the name of the last task in the |
| // generated chain of tasks, which the Job should add as a dependency. |
| func compile(b *specs.TasksCfgBuilder, name string, parts map[string]string) string { |
| // Collect the necessary CIPD packages. |
| pkgs := []*specs.CipdPackage{} |
| deps := []string{} |
| |
| // Android bots require a toolchain. |
| if strings.Contains(name, "Android") { |
| if strings.Contains(name, "Mac") { |
| pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("android_ndk_darwin")) |
| } else if strings.Contains(name, "Win") { |
| pkg := b.MustGetCipdPackageFromAsset("android_ndk_windows") |
| pkg.Path = "n" |
| pkgs = append(pkgs, pkg) |
| } else { |
| deps = append(deps, isolateCIPDAsset(b, ISOLATE_NDK_LINUX_NAME)) |
| } |
| } else if strings.Contains(name, "Chromecast") { |
| pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("cast_toolchain")) |
| pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("chromebook_arm_gles")) |
| } else if strings.Contains(name, "Chromebook") { |
| pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("clang_linux")) |
| if parts["target_arch"] == "x86_64" { |
| pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("chromebook_x86_64_gles")) |
| } else if parts["target_arch"] == "arm" { |
| pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("armhf_sysroot")) |
| pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("chromebook_arm_gles")) |
| } |
| } else if strings.Contains(name, "Debian") { |
| if strings.Contains(name, "Clang") { |
| pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("clang_linux")) |
| } |
| if strings.Contains(name, "Vulkan") { |
| pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("linux_vulkan_sdk")) |
| } |
| if strings.Contains(name, "EMCC") { |
| pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("emscripten_sdk")) |
| } |
| } else if strings.Contains(name, "Win") { |
| deps = append(deps, isolateCIPDAsset(b, ISOLATE_WIN_TOOLCHAIN_NAME)) |
| if strings.Contains(name, "Clang") { |
| pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("clang_win")) |
| } |
| if strings.Contains(name, "Vulkan") { |
| deps = append(deps, isolateCIPDAsset(b, ISOLATE_WIN_VULKAN_SDK_NAME)) |
| } |
| } |
| |
| dimensions := swarmDimensions(parts) |
| |
| // Add the task. |
| b.MustAddTask(name, &specs.TaskSpec{ |
| CipdPackages: pkgs, |
| Dimensions: dimensions, |
| Dependencies: deps, |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "compile", |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildername=%s", name), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| }, |
| Isolate: relpath("compile_skia.isolate"), |
| Priority: 0.8, |
| }) |
| // All compile tasks are runnable as their own Job. Assert that the Job |
| // is listed in JOBS. |
| if !util.In(name, JOBS) { |
| glog.Fatalf("Job %q is missing from the JOBS list!", name) |
| } |
| |
| // Upload the skiaserve binary only for Linux Android compile bots. |
| // See skbug.com/7399 for context. |
| if parts["configuration"] == "Release" && |
| parts["extra_config"] == "Android" && |
| !strings.Contains(parts["os"], "Win") && |
| !strings.Contains(parts["os"], "Mac") { |
| uploadName := fmt.Sprintf("%s%s%s", PREFIX_UPLOAD, jobNameSchema.Sep, name) |
| b.MustAddTask(uploadName, &specs.TaskSpec{ |
| Dependencies: []string{name}, |
| Dimensions: linuxGceDimensions(), |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "upload_skiaserve", |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildername=%s", name), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| }, |
| // We're using the same isolate as upload_dm_results |
| Isolate: relpath("upload_dm_results.isolate"), |
| Priority: 0.8, |
| }) |
| return uploadName |
| } |
| |
| return name |
| } |
| |
| // recreateSKPs generates a RecreateSKPs task. Returns the name of the last |
| // task in the generated chain of tasks, which the Job should add as a |
| // dependency. |
| func recreateSKPs(b *specs.TasksCfgBuilder, name string) string { |
| b.MustAddTask(name, &specs.TaskSpec{ |
| CipdPackages: []*specs.CipdPackage{b.MustGetCipdPackageFromAsset("go")}, |
| Dimensions: linuxGceDimensions(), |
| ExecutionTimeout: 4 * time.Hour, |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "recreate_skps", |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildername=%s", name), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| }, |
| IoTimeout: 40 * time.Minute, |
| Isolate: relpath("compile_skia.isolate"), |
| Priority: 0.8, |
| }) |
| return name |
| } |
| |
| // updateMetaConfig generates a UpdateMetaConfig task. Returns the name of the |
| // last task in the generated chain of tasks, which the Job should add as a |
| // dependency. |
| func updateMetaConfig(b *specs.TasksCfgBuilder, name string) string { |
| b.MustAddTask(name, &specs.TaskSpec{ |
| CipdPackages: []*specs.CipdPackage{}, |
| Dimensions: linuxGceDimensions(), |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "update_meta_config", |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildername=%s", name), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| }, |
| Isolate: relpath("meta_config.isolate"), |
| Priority: 0.8, |
| }) |
| return name |
| } |
| |
| // ctSKPs generates a CT SKPs task. Returns the name of the last task in the |
| // generated chain of tasks, which the Job should add as a dependency. |
| func ctSKPs(b *specs.TasksCfgBuilder, name string) string { |
| b.MustAddTask(name, &specs.TaskSpec{ |
| CipdPackages: []*specs.CipdPackage{}, |
| Dimensions: []string{ |
| "pool:SkiaCT", |
| fmt.Sprintf("os:%s", DEFAULT_OS_LINUX_GCE), |
| }, |
| ExecutionTimeout: 24 * time.Hour, |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "ct_skps", |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildername=%s", name), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| }, |
| IoTimeout: time.Hour, |
| Isolate: relpath("ct_skps_skia.isolate"), |
| Priority: 0.8, |
| }) |
| return name |
| } |
| |
| // checkGeneratedFiles verifies that no generated SKSL files have been edited |
| // by hand. |
| func checkGeneratedFiles(b *specs.TasksCfgBuilder, name string) string { |
| b.MustAddTask(name, &specs.TaskSpec{ |
| CipdPackages: []*specs.CipdPackage{}, |
| Dimensions: linuxGceDimensions(), |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "check_generated_files", |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildername=%s", name), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| }, |
| Isolate: relpath("compile_skia.isolate"), |
| Priority: 0.8, |
| }) |
| return name |
| } |
| |
| // housekeeper generates a Housekeeper task. Returns the name of the last task |
| // in the generated chain of tasks, which the Job should add as a dependency. |
| func housekeeper(b *specs.TasksCfgBuilder, name, compileTaskName string) string { |
| b.MustAddTask(name, &specs.TaskSpec{ |
| CipdPackages: []*specs.CipdPackage{b.MustGetCipdPackageFromAsset("go")}, |
| Dependencies: []string{compileTaskName}, |
| Dimensions: linuxGceDimensions(), |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "housekeeper", |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildername=%s", name), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| }, |
| Isolate: relpath("housekeeper_skia.isolate"), |
| Priority: 0.8, |
| }) |
| return name |
| } |
| |
| // bookmaker generates a Bookmaker task. Returns the name of the last task |
| // in the generated chain of tasks, which the Job should add as a dependency. |
| func bookmaker(b *specs.TasksCfgBuilder, name, compileTaskName string) string { |
| b.MustAddTask(name, &specs.TaskSpec{ |
| CipdPackages: []*specs.CipdPackage{b.MustGetCipdPackageFromAsset("go")}, |
| Dependencies: []string{compileTaskName}, |
| Dimensions: linuxGceDimensions(), |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "bookmaker", |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildername=%s", name), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| }, |
| Isolate: relpath("compile_skia.isolate"), |
| Priority: 0.8, |
| ExecutionTimeout: 2 * time.Hour, |
| IoTimeout: 2 * time.Hour, |
| }) |
| return name |
| } |
| |
| // infra generates an infra_tests task. Returns the name of the last task in the |
| // generated chain of tasks, which the Job should add as a dependency. |
| func infra(b *specs.TasksCfgBuilder, name string) string { |
| b.MustAddTask(name, &specs.TaskSpec{ |
| CipdPackages: []*specs.CipdPackage{b.MustGetCipdPackageFromAsset("go")}, |
| Dimensions: linuxGceDimensions(), |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "infra", |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildername=%s", name), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| }, |
| Isolate: relpath("infra_skia.isolate"), |
| Priority: 0.8, |
| }) |
| return name |
| } |
| |
| // calmbench generates a calmbench task. Returns the name of the last task in the |
| // generated chain of tasks, which the Job should add as a dependency. |
| func calmbench(b *specs.TasksCfgBuilder, name string, parts map[string]string) string { |
| s := &specs.TaskSpec{ |
| CipdPackages: []*specs.CipdPackage{b.MustGetCipdPackageFromAsset("clang_linux")}, |
| Dimensions: swarmDimensions(parts), |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "calmbench", |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildername=%s", name), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| }, |
| Isolate: relpath("infra_skia.isolate"), |
| Priority: 0.8, |
| } |
| |
| s.Dependencies = append(s.Dependencies, ISOLATE_SKP_NAME, ISOLATE_SVG_NAME) |
| |
| b.MustAddTask(name, s) |
| |
| // Upload results if necessary. |
| if strings.Contains(name, "Release") && doUpload(name) { |
| uploadName := fmt.Sprintf("%s%s%s", PREFIX_UPLOAD, jobNameSchema.Sep, name) |
| b.MustAddTask(uploadName, &specs.TaskSpec{ |
| Dependencies: []string{name}, |
| Dimensions: linuxGceDimensions(), |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "upload_calmbench_results", |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildername=%s", name), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| fmt.Sprintf("gs_bucket=%s", CONFIG.GsBucketCalm), |
| }, |
| // We're using the same isolate as upload_nano_results |
| Isolate: relpath("upload_nano_results.isolate"), |
| Priority: 0.8, |
| }) |
| return uploadName |
| } |
| |
| return name |
| } |
| |
| // doUpload indicates whether the given Job should upload its results. |
| func doUpload(name string) bool { |
| for _, s := range CONFIG.NoUpload { |
| m, err := regexp.MatchString(s, name) |
| if err != nil { |
| glog.Fatal(err) |
| } |
| if m { |
| return false |
| } |
| } |
| return true |
| } |
| |
| // test generates a Test task. Returns the name of the last task in the |
| // generated chain of tasks, which the Job should add as a dependency. |
| func test(b *specs.TasksCfgBuilder, name string, parts map[string]string, compileTaskName string, pkgs []*specs.CipdPackage) string { |
| s := &specs.TaskSpec{ |
| CipdPackages: pkgs, |
| Dependencies: []string{compileTaskName}, |
| Dimensions: swarmDimensions(parts), |
| ExecutionTimeout: 4 * time.Hour, |
| Expiration: 20 * time.Hour, |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "test", |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildbucket_build_id=%s", specs.PLACEHOLDER_BUILDBUCKET_BUILD_ID), |
| fmt.Sprintf("buildername=%s", name), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| }, |
| IoTimeout: 40 * time.Minute, |
| Isolate: relpath("test_skia.isolate"), |
| MaxAttempts: 1, |
| Priority: 0.8, |
| } |
| if useBundledRecipes(parts) { |
| s.Dependencies = append(s.Dependencies, BUNDLE_RECIPES_NAME) |
| if strings.Contains(parts["os"], "Win") { |
| s.Isolate = relpath("test_skia_bundled_win.isolate") |
| } else { |
| s.Isolate = relpath("test_skia_bundled_unix.isolate") |
| } |
| } |
| if deps := getIsolatedCIPDDeps(parts); len(deps) > 0 { |
| s.Dependencies = append(s.Dependencies, deps...) |
| } |
| if strings.Contains(parts["extra_config"], "Valgrind") { |
| s.ExecutionTimeout = 9 * time.Hour |
| s.Expiration = 48 * time.Hour |
| s.IoTimeout = time.Hour |
| s.CipdPackages = append(s.CipdPackages, b.MustGetCipdPackageFromAsset("valgrind")) |
| s.Dimensions = append(s.Dimensions, "valgrind:1") |
| } else if strings.Contains(parts["extra_config"], "MSAN") { |
| s.ExecutionTimeout = 9 * time.Hour |
| } else if parts["arch"] == "x86" && parts["configuration"] == "Debug" { |
| // skia:6737 |
| s.ExecutionTimeout = 6 * time.Hour |
| } |
| iid := internalHardwareLabel(parts) |
| if iid != nil { |
| s.ExtraArgs = append(s.ExtraArgs, fmt.Sprintf("internal_hardware_label=%d", *iid)) |
| } |
| b.MustAddTask(name, s) |
| |
| // Upload results if necessary. TODO(kjlubick): If we do coverage analysis at the same |
| // time as normal tests (which would be nice), cfg.json needs to have Coverage removed. |
| if doUpload(name) { |
| uploadName := fmt.Sprintf("%s%s%s", PREFIX_UPLOAD, jobNameSchema.Sep, name) |
| b.MustAddTask(uploadName, &specs.TaskSpec{ |
| Dependencies: []string{name}, |
| Dimensions: linuxGceDimensions(), |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "upload_dm_results", |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildername=%s", name), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| fmt.Sprintf("gs_bucket=%s", CONFIG.GsBucketGm), |
| }, |
| Isolate: relpath("upload_dm_results.isolate"), |
| Priority: 0.8, |
| }) |
| return uploadName |
| } |
| |
| return name |
| } |
| |
| func coverage(b *specs.TasksCfgBuilder, name string, parts map[string]string, compileTaskName string, pkgs []*specs.CipdPackage) string { |
| shards := 1 |
| deps := []string{} |
| |
| tf := parts["test_filter"] |
| if strings.Contains(tf, "Shard") { |
| // Expected Shard_NN |
| shardstr := strings.Split(tf, "_")[1] |
| var err error |
| shards, err = strconv.Atoi(shardstr) |
| if err != nil { |
| glog.Fatalf("Expected int for number of shards %q in %s: %s", shardstr, name, err) |
| } |
| } |
| for i := 0; i < shards; i++ { |
| n := strings.Replace(name, tf, fmt.Sprintf("shard_%02d_%02d", i, shards), 1) |
| s := &specs.TaskSpec{ |
| CipdPackages: pkgs, |
| Dependencies: []string{compileTaskName}, |
| Dimensions: swarmDimensions(parts), |
| ExecutionTimeout: 4 * time.Hour, |
| Expiration: 20 * time.Hour, |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "test", |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildername=%s", n), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| }, |
| IoTimeout: 40 * time.Minute, |
| Isolate: relpath("test_skia.isolate"), |
| MaxAttempts: 1, |
| Priority: 0.8, |
| } |
| if useBundledRecipes(parts) { |
| s.Dependencies = append(s.Dependencies, BUNDLE_RECIPES_NAME) |
| if strings.Contains(parts["os"], "Win") { |
| s.Isolate = relpath("test_skia_bundled_win.isolate") |
| } else { |
| s.Isolate = relpath("test_skia_bundled_unix.isolate") |
| } |
| } |
| if deps := getIsolatedCIPDDeps(parts); len(deps) > 0 { |
| s.Dependencies = append(s.Dependencies, deps...) |
| } |
| b.MustAddTask(n, s) |
| deps = append(deps, n) |
| } |
| |
| uploadName := fmt.Sprintf("%s%s%s", "Upload", jobNameSchema.Sep, name) |
| // We need clang_linux to get access to the llvm-profdata and llvm-cov binaries |
| // which are used to deal with the raw coverage data output by the Test step. |
| pkgs = append([]*specs.CipdPackage{}, b.MustGetCipdPackageFromAsset("clang_linux")) |
| deps = append(deps, compileTaskName) |
| |
| b.MustAddTask(uploadName, &specs.TaskSpec{ |
| // A dependency on compileTaskName makes the TaskScheduler link the |
| // isolated output of the compile step to the input of the upload step, |
| // which gives us access to the instrumented binary. The binary is |
| // needed to figure out symbol names and line numbers. |
| Dependencies: deps, |
| Dimensions: linuxGceDimensions(), |
| CipdPackages: pkgs, |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "upload_coverage_results", |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildername=%s", name), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| fmt.Sprintf("gs_bucket=%s", CONFIG.GsBucketCoverage), |
| }, |
| Isolate: relpath("upload_coverage_results.isolate"), |
| Priority: 0.8, |
| }) |
| return uploadName |
| } |
| |
| // perf generates a Perf task. Returns the name of the last task in the |
| // generated chain of tasks, which the Job should add as a dependency. |
| func perf(b *specs.TasksCfgBuilder, name string, parts map[string]string, compileTaskName string, pkgs []*specs.CipdPackage) string { |
| recipe := "perf" |
| isolate := relpath("perf_skia.isolate") |
| if strings.Contains(parts["extra_config"], "Skpbench") { |
| recipe = "skpbench" |
| isolate = relpath("skpbench_skia.isolate") |
| if useBundledRecipes(parts) { |
| if strings.Contains(parts["os"], "Win") { |
| isolate = relpath("skpbench_skia_bundled_win.isolate") |
| } else { |
| isolate = relpath("skpbench_skia_bundled_unix.isolate") |
| } |
| } |
| } else if useBundledRecipes(parts) { |
| if strings.Contains(parts["os"], "Win") { |
| isolate = relpath("perf_skia_bundled_win.isolate") |
| } else { |
| isolate = relpath("perf_skia_bundled_unix.isolate") |
| } |
| } |
| s := &specs.TaskSpec{ |
| CipdPackages: pkgs, |
| Dependencies: []string{compileTaskName}, |
| Dimensions: swarmDimensions(parts), |
| ExecutionTimeout: 4 * time.Hour, |
| Expiration: 20 * time.Hour, |
| ExtraArgs: []string{ |
| "--workdir", "../../..", recipe, |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildername=%s", name), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| }, |
| IoTimeout: 40 * time.Minute, |
| Isolate: isolate, |
| MaxAttempts: 1, |
| Priority: 0.8, |
| } |
| if useBundledRecipes(parts) { |
| s.Dependencies = append(s.Dependencies, BUNDLE_RECIPES_NAME) |
| } |
| if deps := getIsolatedCIPDDeps(parts); len(deps) > 0 { |
| s.Dependencies = append(s.Dependencies, deps...) |
| } |
| |
| if strings.Contains(parts["extra_config"], "Valgrind") { |
| s.ExecutionTimeout = 9 * time.Hour |
| s.Expiration = 48 * time.Hour |
| s.IoTimeout = time.Hour |
| s.CipdPackages = append(s.CipdPackages, b.MustGetCipdPackageFromAsset("valgrind")) |
| s.Dimensions = append(s.Dimensions, "valgrind:1") |
| } else if strings.Contains(parts["extra_config"], "MSAN") { |
| s.ExecutionTimeout = 9 * time.Hour |
| } else if parts["arch"] == "x86" && parts["configuration"] == "Debug" { |
| // skia:6737 |
| s.ExecutionTimeout = 6 * time.Hour |
| } |
| iid := internalHardwareLabel(parts) |
| if iid != nil { |
| s.ExtraArgs = append(s.ExtraArgs, fmt.Sprintf("internal_hardware_label=%d", *iid)) |
| } |
| b.MustAddTask(name, s) |
| |
| // Upload results if necessary. |
| if strings.Contains(name, "Release") && doUpload(name) { |
| uploadName := fmt.Sprintf("%s%s%s", PREFIX_UPLOAD, jobNameSchema.Sep, name) |
| b.MustAddTask(uploadName, &specs.TaskSpec{ |
| Dependencies: []string{name}, |
| Dimensions: linuxGceDimensions(), |
| ExtraArgs: []string{ |
| "--workdir", "../../..", "upload_nano_results", |
| fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), |
| fmt.Sprintf("buildername=%s", name), |
| fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), |
| fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), |
| fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), |
| fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), |
| fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), |
| fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), |
| fmt.Sprintf("gs_bucket=%s", CONFIG.GsBucketNano), |
| }, |
| Isolate: relpath("upload_nano_results.isolate"), |
| Priority: 0.8, |
| }) |
| return uploadName |
| } |
| return name |
| } |
| |
| // process generates tasks and jobs for the given job name. |
| func process(b *specs.TasksCfgBuilder, name string) { |
| deps := []string{} |
| |
| // Bundle Recipes. |
| if name == BUNDLE_RECIPES_NAME { |
| deps = append(deps, bundleRecipes(b)) |
| } |
| |
| // Isolate CIPD assets. |
| if _, ok := ISOLATE_ASSET_MAPPING[name]; ok { |
| deps = append(deps, isolateCIPDAsset(b, name)) |
| } |
| |
| parts, err := jobNameSchema.ParseJobName(name) |
| if err != nil { |
| glog.Fatal(err) |
| } |
| |
| // RecreateSKPs. |
| if strings.Contains(name, "RecreateSKPs") { |
| deps = append(deps, recreateSKPs(b, name)) |
| } |
| |
| // UpdateMetaConfig bot. |
| if strings.Contains(name, "UpdateMetaConfig") { |
| deps = append(deps, updateMetaConfig(b, name)) |
| } |
| |
| // CT bots. |
| if strings.Contains(name, "-CT_") { |
| deps = append(deps, ctSKPs(b, name)) |
| } |
| |
| // Infra tests. |
| if name == "Housekeeper-PerCommit-InfraTests" { |
| deps = append(deps, infra(b, name)) |
| } |
| |
| // Compile bots. |
| if parts["role"] == "Build" { |
| deps = append(deps, compile(b, name, parts)) |
| } |
| |
| // Calmbench bots. |
| if parts["role"] == "Calmbench" { |
| deps = append(deps, calmbench(b, name, parts)) |
| } |
| |
| // Most remaining bots need a compile task. |
| compileTaskName := deriveCompileTaskName(name, parts) |
| compileTaskParts, err := jobNameSchema.ParseJobName(compileTaskName) |
| if err != nil { |
| glog.Fatal(err) |
| } |
| // These bots do not need a compile task. |
| if parts["role"] != "Build" && parts["role"] != "Calmbench" && |
| name != "Housekeeper-PerCommit-BundleRecipes" && |
| name != "Housekeeper-PerCommit-InfraTests" && |
| name != "Housekeeper-PerCommit-CheckGeneratedFiles" && |
| !strings.Contains(name, "RecreateSKPs") && |
| !strings.Contains(name, "UpdateMetaConfig") && |
| !strings.Contains(name, "-CT_") && |
| !strings.Contains(name, "Housekeeper-PerCommit-Isolate") { |
| compile(b, compileTaskName, compileTaskParts) |
| } |
| |
| // Housekeepers. |
| if name == "Housekeeper-PerCommit" { |
| deps = append(deps, housekeeper(b, name, compileTaskName)) |
| } |
| if name == "Housekeeper-PerCommit-CheckGeneratedFiles" { |
| deps = append(deps, checkGeneratedFiles(b, name)) |
| } |
| if name == "Housekeeper-Nightly-Bookmaker" { |
| deps = append(deps, bookmaker(b, name, compileTaskName)) |
| } |
| |
| // Common assets needed by the remaining bots. |
| |
| pkgs := []*specs.CipdPackage{} |
| |
| if deps := getIsolatedCIPDDeps(parts); len(deps) == 0 { |
| pkgs = []*specs.CipdPackage{ |
| b.MustGetCipdPackageFromAsset("skimage"), |
| b.MustGetCipdPackageFromAsset("skp"), |
| b.MustGetCipdPackageFromAsset("svg"), |
| } |
| } |
| |
| if strings.Contains(name, "Ubuntu") || strings.Contains(name, "Debian") { |
| if strings.Contains(name, "SAN") { |
| pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("clang_linux")) |
| } |
| if strings.Contains(name, "Vulkan") { |
| pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("linux_vulkan_sdk")) |
| } |
| if strings.Contains(name, "Intel") && strings.Contains(name, "GPU") { |
| if strings.Contains(name, "Release") { |
| pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("linux_vulkan_intel_driver_release")) |
| } else { |
| pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("linux_vulkan_intel_driver_debug")) |
| } |
| } |
| } |
| |
| // Test bots. |
| |
| if parts["role"] == "Test" { |
| if strings.Contains(parts["extra_config"], "Coverage") { |
| deps = append(deps, coverage(b, name, parts, compileTaskName, pkgs)) |
| } else if !strings.Contains(name, "-CT_") { |
| deps = append(deps, test(b, name, parts, compileTaskName, pkgs)) |
| } |
| |
| } |
| |
| // Perf bots. |
| if parts["role"] == "Perf" && !strings.Contains(name, "-CT_") { |
| deps = append(deps, perf(b, name, parts, compileTaskName, pkgs)) |
| } |
| |
| // Add the Job spec. |
| j := &specs.JobSpec{ |
| Priority: 0.8, |
| TaskSpecs: deps, |
| Trigger: specs.TRIGGER_ANY_BRANCH, |
| } |
| if strings.Contains(name, "-Nightly-") { |
| j.Trigger = specs.TRIGGER_NIGHTLY |
| } else if strings.Contains(name, "-Weekly-") || strings.Contains(name, "CT_DM_1m_SKPs") { |
| j.Trigger = specs.TRIGGER_WEEKLY |
| } else if strings.Contains(name, "Flutter") || strings.Contains(name, "PDFium") || strings.Contains(name, "CommandBuffer") { |
| j.Trigger = specs.TRIGGER_MASTER_ONLY |
| } |
| b.MustAddJob(name, j) |
| } |
| |
| func loadJson(flag *string, defaultFlag string, val interface{}) { |
| if *flag == "" { |
| *flag = defaultFlag |
| } |
| b, err := ioutil.ReadFile(*flag) |
| if err != nil { |
| glog.Fatal(err) |
| } |
| if err := json.Unmarshal(b, val); err != nil { |
| glog.Fatal(err) |
| } |
| } |
| |
| // Regenerate the tasks.json file. |
| func main() { |
| b := specs.MustNewTasksCfgBuilder() |
| b.SetAssetsDir(*assetsDir) |
| infraBots := path.Join(b.CheckoutRoot(), "infra", "bots") |
| |
| // Load the jobs from a JSON file. |
| loadJson(jobsFile, path.Join(infraBots, "jobs.json"), &JOBS) |
| |
| // Load general config information from a JSON file. |
| loadJson(cfgFile, path.Join(infraBots, "cfg.json"), &CONFIG) |
| |
| // Create the JobNameSchema. |
| if *builderNameSchemaFile == "" { |
| *builderNameSchemaFile = path.Join(b.CheckoutRoot(), "infra", "bots", "recipe_modules", "builder_name_schema", "builder_name_schema.json") |
| } |
| schema, err := NewJobNameSchema(*builderNameSchemaFile) |
| if err != nil { |
| glog.Fatal(err) |
| } |
| jobNameSchema = schema |
| |
| // Create Tasks and Jobs. |
| for _, name := range JOBS { |
| process(b, name) |
| } |
| |
| b.MustFinish() |
| } |
| |
| // TODO(borenet): The below really belongs in its own file, probably next to the |
| // builder_name_schema.json file. |
| |
| // JobNameSchema is a struct used for (de)constructing Job names in a |
| // predictable format. |
| type JobNameSchema struct { |
| Schema map[string][]string `json:"builder_name_schema"` |
| Sep string `json:"builder_name_sep"` |
| } |
| |
| // NewJobNameSchema returns a JobNameSchema instance based on the given JSON |
| // file. |
| func NewJobNameSchema(jsonFile string) (*JobNameSchema, error) { |
| var rv JobNameSchema |
| f, err := os.Open(jsonFile) |
| if err != nil { |
| return nil, err |
| } |
| defer util.Close(f) |
| if err := json.NewDecoder(f).Decode(&rv); err != nil { |
| return nil, err |
| } |
| return &rv, nil |
| } |
| |
| // ParseJobName splits the given Job name into its component parts, according |
| // to the schema. |
| func (s *JobNameSchema) ParseJobName(n string) (map[string]string, error) { |
| split := strings.Split(n, s.Sep) |
| if len(split) < 2 { |
| return nil, fmt.Errorf("Invalid job name: %q", n) |
| } |
| role := split[0] |
| split = split[1:] |
| keys, ok := s.Schema[role] |
| if !ok { |
| return nil, fmt.Errorf("Invalid job name; %q is not a valid role.", role) |
| } |
| extraConfig := "" |
| if len(split) == len(keys)+1 { |
| extraConfig = split[len(split)-1] |
| split = split[:len(split)-1] |
| } |
| if len(split) != len(keys) { |
| return nil, fmt.Errorf("Invalid job name; %q has incorrect number of parts.", n) |
| } |
| rv := make(map[string]string, len(keys)+2) |
| rv["role"] = role |
| if extraConfig != "" { |
| rv["extra_config"] = extraConfig |
| } |
| for i, k := range keys { |
| rv[k] = split[i] |
| } |
| return rv, nil |
| } |
| |
| // MakeJobName assembles the given parts of a Job name, according to the schema. |
| func (s *JobNameSchema) MakeJobName(parts map[string]string) (string, error) { |
| role, ok := parts["role"] |
| if !ok { |
| return "", fmt.Errorf("Invalid job parts; jobs must have a role.") |
| } |
| keys, ok := s.Schema[role] |
| if !ok { |
| return "", fmt.Errorf("Invalid job parts; unknown role %q", role) |
| } |
| rvParts := make([]string, 0, len(parts)) |
| rvParts = append(rvParts, role) |
| for _, k := range keys { |
| v, ok := parts[k] |
| if !ok { |
| return "", fmt.Errorf("Invalid job parts; missing %q", k) |
| } |
| rvParts = append(rvParts, v) |
| } |
| if _, ok := parts["extra_config"]; ok { |
| rvParts = append(rvParts, parts["extra_config"]) |
| } |
| return strings.Join(rvParts, s.Sep), nil |
| } |