DEFINE_int32 + DEFINE_uint32 -> DEFINE_int
There's really no big benefit to distinguishing these.
Change-Id: Ib329d32b1fc43e98ba40658cfea37261203ecdb9
Reviewed-on: https://skia-review.googlesource.com/c/skia/+/202801
Commit-Queue: Mike Klein <mtklein@google.com>
Reviewed-by: Brian Osman <brianosman@google.com>
diff --git a/tests/ProcessorTest.cpp b/tests/ProcessorTest.cpp
index 1f9f01e..f5c37e4 100644
--- a/tests/ProcessorTest.cpp
+++ b/tests/ProcessorTest.cpp
@@ -233,8 +233,8 @@
#include "CommandLineFlags.h"
static DEFINE_bool(randomProcessorTest, false,
"Use non-deterministic seed for random processor tests?");
-static DEFINE_uint32(processorSeed, 0,
- "Use specific seed for processor tests. Overridden by --randomProcessorTest.");
+static DEFINE_int(processorSeed, 0,
+ "Use specific seed for processor tests. Overridden by --randomProcessorTest.");
#if GR_TEST_UTILS