blob: 23672ddbea43db5e3a4b9d0e60e98a6cbedb6cda [file] [log] [blame]
csmartdalton008b9d82017-02-22 12:00:42 -07001/*
2 * Copyright 2017 Google Inc.
3 *
4 * Use of this source code is governed by a BSD-style license that can be
5 * found in the LICENSE file.
6 */
7
Chris Dalton040238b2017-12-18 14:22:34 -07008#ifndef SK_COMMON_FLAGS_GPU_H
9#define SK_COMMON_FLAGS_GPU_H
csmartdalton008b9d82017-02-22 12:00:42 -070010
Brian Osman195c05b2017-08-30 15:14:04 -040011#include "GrTypesPriv.h"
csmartdalton008b9d82017-02-22 12:00:42 -070012#include "SkCommandLineFlags.h"
13#include "SkTypes.h"
14
Chris Dalton040238b2017-12-18 14:22:34 -070015DECLARE_int32(gpuThreads);
16DECLARE_bool(cachePathMasks);
17DECLARE_bool(noGS);
csmartdalton008b9d82017-02-22 12:00:42 -070018DECLARE_string(pr);
Robert Phillipscb2275b2018-08-07 08:34:38 -040019DECLARE_bool(disableExplicitAlloc);
Robert Phillipsb7a98ef2018-10-10 09:26:00 -040020DECLARE_bool(reduceOpListSplitting);
csmartdalton008b9d82017-02-22 12:00:42 -070021
Brian Osman195c05b2017-08-30 15:14:04 -040022inline GpuPathRenderers get_named_pathrenderers_flags(const char* name) {
Chris Dalton9acfc6c2018-07-26 12:34:49 -060023 if (!strcmp(name, "none")) {
24 return GpuPathRenderers::kNone;
csmartdalton008b9d82017-02-22 12:00:42 -070025 } else if (!strcmp(name, "dashline")) {
26 return GpuPathRenderers::kDashLine;
27 } else if (!strcmp(name, "nvpr")) {
28 return GpuPathRenderers::kStencilAndCover;
Chris Dalton9acfc6c2018-07-26 12:34:49 -060029 } else if (!strcmp(name, "ccpr")) {
30 return GpuPathRenderers::kCoverageCounting;
31 } else if (!strcmp(name, "aahairline")) {
32 return GpuPathRenderers::kAAHairline;
csmartdalton008b9d82017-02-22 12:00:42 -070033 } else if (!strcmp(name, "aaconvex")) {
34 return GpuPathRenderers::kAAConvex;
35 } else if (!strcmp(name, "aalinearizing")) {
36 return GpuPathRenderers::kAALinearizing;
Jim Van Verth83010462017-03-16 08:45:39 -040037 } else if (!strcmp(name, "small")) {
38 return GpuPathRenderers::kSmall;
csmartdalton008b9d82017-02-22 12:00:42 -070039 } else if (!strcmp(name, "tess")) {
Brian Osman8a9de3d2017-03-01 14:59:05 -050040 return GpuPathRenderers::kTessellating;
Chris Dalton9acfc6c2018-07-26 12:34:49 -060041 } else if (!strcmp(name, "all")) {
42 return GpuPathRenderers::kAll;
csmartdalton008b9d82017-02-22 12:00:42 -070043 }
44 SK_ABORT(SkStringPrintf("error: unknown named path renderer \"%s\"\n", name).c_str());
45 return GpuPathRenderers::kNone;
46}
47
Brian Osman195c05b2017-08-30 15:14:04 -040048inline GpuPathRenderers CollectGpuPathRenderersFromFlags() {
csmartdalton008b9d82017-02-22 12:00:42 -070049 if (FLAGS_pr.isEmpty()) {
Chris Dalton9acfc6c2018-07-26 12:34:49 -060050 return GpuPathRenderers::kAll;
csmartdalton008b9d82017-02-22 12:00:42 -070051 }
Chris Dalton9acfc6c2018-07-26 12:34:49 -060052 GpuPathRenderers gpuPathRenderers = '~' == FLAGS_pr[0][0]
53 ? GpuPathRenderers::kAll : GpuPathRenderers::kNone;
csmartdalton008b9d82017-02-22 12:00:42 -070054 for (int i = 0; i < FLAGS_pr.count(); ++i) {
55 const char* name = FLAGS_pr[i];
56 if (name[0] == '~') {
57 gpuPathRenderers &= ~get_named_pathrenderers_flags(&name[1]);
58 } else {
59 gpuPathRenderers |= get_named_pathrenderers_flags(name);
60 }
61 }
62 return gpuPathRenderers;
63}
64
Chris Dalton040238b2017-12-18 14:22:34 -070065/**
66 * Helper to set GrContextOptions from common GPU flags.
67 */
68void SetCtxOptionsFromCommonFlags(struct GrContextOptions*);
69
csmartdalton008b9d82017-02-22 12:00:42 -070070#endif