blob: d610b3b9d22d1631dad14ae1422e3bf0c1bea3f7 [file] [log] [blame]
/*
* Copyright 2016 Google Inc.
*
* Use of this source code is governed by a BSD-style license that can be
* found in the LICENSE file.
*/
#include "src/sksl/SkSLCompiler.h"
#include <memory>
#include <unordered_set>
#include "src/sksl/SkSLAnalysis.h"
#include "src/sksl/SkSLByteCodeGenerator.h"
#include "src/sksl/SkSLCFGGenerator.h"
#include "src/sksl/SkSLCPPCodeGenerator.h"
#include "src/sksl/SkSLGLSLCodeGenerator.h"
#include "src/sksl/SkSLHCodeGenerator.h"
#include "src/sksl/SkSLIRGenerator.h"
#include "src/sksl/SkSLMetalCodeGenerator.h"
#include "src/sksl/SkSLPipelineStageCodeGenerator.h"
#include "src/sksl/SkSLRehydrator.h"
#include "src/sksl/SkSLSPIRVCodeGenerator.h"
#include "src/sksl/SkSLSPIRVtoHLSL.h"
#include "src/sksl/ir/SkSLEnum.h"
#include "src/sksl/ir/SkSLExpression.h"
#include "src/sksl/ir/SkSLExpressionStatement.h"
#include "src/sksl/ir/SkSLFunctionCall.h"
#include "src/sksl/ir/SkSLIntLiteral.h"
#include "src/sksl/ir/SkSLModifiersDeclaration.h"
#include "src/sksl/ir/SkSLNop.h"
#include "src/sksl/ir/SkSLSymbolTable.h"
#include "src/sksl/ir/SkSLTernaryExpression.h"
#include "src/sksl/ir/SkSLUnresolvedFunction.h"
#include "src/sksl/ir/SkSLVarDeclarations.h"
#include "src/utils/SkBitSet.h"
#include <fstream>
#if !defined(SKSL_STANDALONE) & SK_SUPPORT_GPU
#include "include/gpu/GrContextOptions.h"
#include "src/gpu/GrShaderCaps.h"
#endif
#ifdef SK_ENABLE_SPIRV_VALIDATION
#include "spirv-tools/libspirv.hpp"
#endif
#if defined(SKSL_STANDALONE)
// In standalone mode, we load the textual sksl source files. GN generates or copies these files
// to the skslc executable directory. The "data" in this mode is just the filename.
#define MODULE_DATA(name) MakeModulePath("sksl_" #name ".sksl")
#else
// At runtime, we load the dehydrated sksl data files. The data is a (pointer, size) pair.
#include "src/sksl/generated/sksl_fp.dehydrated.sksl"
#include "src/sksl/generated/sksl_frag.dehydrated.sksl"
#include "src/sksl/generated/sksl_geom.dehydrated.sksl"
#include "src/sksl/generated/sksl_gpu.dehydrated.sksl"
#include "src/sksl/generated/sksl_interp.dehydrated.sksl"
#include "src/sksl/generated/sksl_pipeline.dehydrated.sksl"
#include "src/sksl/generated/sksl_public.dehydrated.sksl"
#include "src/sksl/generated/sksl_vert.dehydrated.sksl"
#define MODULE_DATA(name) MakeModuleData(SKSL_INCLUDE_sksl_##name,\
SKSL_INCLUDE_sksl_##name##_LENGTH)
#endif
namespace SkSL {
class AutoSource {
public:
AutoSource(Compiler* compiler, const String* source)
: fCompiler(compiler), fOldSource(fCompiler->fSource) {
fCompiler->fSource = source;
}
~AutoSource() { fCompiler->fSource = fOldSource; }
Compiler* fCompiler;
const String* fOldSource;
};
Compiler::Compiler(const ShaderCapsClass* caps, Flags flags)
: fContext(std::make_shared<Context>())
, fCaps(caps)
, fInliner(fContext.get(), fCaps)
, fFlags(flags)
, fErrorCount(0) {
SkASSERT(fCaps);
fRootSymbolTable = std::make_shared<SymbolTable>(this, /*builtin=*/true);
fPrivateSymbolTable = std::make_shared<SymbolTable>(fRootSymbolTable, /*builtin=*/true);
fIRGenerator = std::make_unique<IRGenerator>(fContext.get(), fCaps, *this);
#define TYPE(t) fContext->f##t##_Type.get()
const SkSL::Symbol* rootTypes[] = {
TYPE(Void),
TYPE( Float), TYPE( Float2), TYPE( Float3), TYPE( Float4),
TYPE( Half), TYPE( Half2), TYPE( Half3), TYPE( Half4),
TYPE( Int), TYPE( Int2), TYPE( Int3), TYPE( Int4),
TYPE( UInt), TYPE( UInt2), TYPE( UInt3), TYPE( UInt4),
TYPE( Short), TYPE( Short2), TYPE( Short3), TYPE( Short4),
TYPE(UShort), TYPE(UShort2), TYPE(UShort3), TYPE(UShort4),
TYPE( Byte), TYPE( Byte2), TYPE( Byte3), TYPE( Byte4),
TYPE( UByte), TYPE( UByte2), TYPE( UByte3), TYPE( UByte4),
TYPE( Bool), TYPE( Bool2), TYPE( Bool3), TYPE( Bool4),
TYPE(Float2x2), TYPE(Float2x3), TYPE(Float2x4),
TYPE(Float3x2), TYPE(Float3x3), TYPE(Float3x4),
TYPE(Float4x2), TYPE(Float4x3), TYPE(Float4x4),
TYPE(Half2x2), TYPE(Half2x3), TYPE(Half2x4),
TYPE(Half3x2), TYPE(Half3x3), TYPE(Half3x4),
TYPE(Half4x2), TYPE(Half4x3), TYPE(Half4x4),
TYPE(GenType), TYPE(GenHType), TYPE(GenIType), TYPE(GenUType), TYPE(GenBType),
TYPE(Mat), TYPE(MatH), TYPE(Vec),
TYPE(GVec), TYPE(GVec2), TYPE(GVec3), TYPE(GVec4),
TYPE(HVec), TYPE(IVec), TYPE(UVec), TYPE(SVec), TYPE(USVec),
TYPE(ByteVec), TYPE(UByteVec), TYPE(BVec),
TYPE(FragmentProcessor),
};
const SkSL::Symbol* privateTypes[] = {
TYPE(Sampler1D), TYPE(Sampler2D), TYPE(Sampler3D),
TYPE(SamplerExternalOES),
TYPE(SamplerCube),
TYPE(Sampler2DRect),
TYPE(Sampler1DArray), TYPE(Sampler2DArray), TYPE(SamplerCubeArray),
TYPE(SamplerBuffer),
TYPE(Sampler2DMS), TYPE(Sampler2DMSArray),
TYPE(ISampler2D),
TYPE(Image2D), TYPE(IImage2D),
TYPE(SubpassInput), TYPE(SubpassInputMS),
TYPE(GSampler1D), TYPE(GSampler2D), TYPE(GSampler3D),
TYPE(GSamplerCube),
TYPE(GSampler2DRect),
TYPE(GSampler1DArray), TYPE(GSampler2DArray), TYPE(GSamplerCubeArray),
TYPE(GSamplerBuffer),
TYPE(GSampler2DMS), TYPE(GSampler2DMSArray),
TYPE(Sampler1DShadow), TYPE(Sampler2DShadow), TYPE(SamplerCubeShadow),
TYPE(Sampler2DRectShadow),
TYPE(Sampler1DArrayShadow), TYPE(Sampler2DArrayShadow), TYPE(SamplerCubeArrayShadow),
TYPE(GSampler2DArrayShadow), TYPE(GSamplerCubeArrayShadow),
TYPE(Sampler),
TYPE(Texture2D),
};
for (const SkSL::Symbol* type : rootTypes) {
fRootSymbolTable->addWithoutOwnership(type);
}
for (const SkSL::Symbol* type : privateTypes) {
fPrivateSymbolTable->addWithoutOwnership(type);
}
#undef TYPE
// sk_Caps is "builtin", but all references to it are resolved to Settings, so we don't need to
// treat it as builtin (ie, no need to clone it into the Program).
fPrivateSymbolTable->add(
std::make_unique<Variable>(/*offset=*/-1,
fIRGenerator->fModifiers->addToPool(Modifiers()),
"sk_Caps",
fContext->fSkCaps_Type.get(),
/*builtin=*/false,
Variable::Storage::kGlobal));
fRootModule = {fRootSymbolTable, /*fIntrinsics=*/nullptr};
fPrivateModule = {fPrivateSymbolTable, /*fIntrinsics=*/nullptr};
}
Compiler::~Compiler() {}
const ParsedModule& Compiler::loadGPUModule() {
if (!fGPUModule.fSymbols) {
fGPUModule = this->parseModule(Program::kFragment_Kind, MODULE_DATA(gpu), fPrivateModule);
}
return fGPUModule;
}
const ParsedModule& Compiler::loadFragmentModule() {
if (!fFragmentModule.fSymbols) {
fFragmentModule = this->parseModule(Program::kFragment_Kind, MODULE_DATA(frag),
this->loadGPUModule());
}
return fFragmentModule;
}
const ParsedModule& Compiler::loadVertexModule() {
if (!fVertexModule.fSymbols) {
fVertexModule = this->parseModule(Program::kVertex_Kind, MODULE_DATA(vert),
this->loadGPUModule());
}
return fVertexModule;
}
const ParsedModule& Compiler::loadGeometryModule() {
if (!fGeometryModule.fSymbols) {
fGeometryModule = this->parseModule(Program::kGeometry_Kind, MODULE_DATA(geom),
this->loadGPUModule());
}
return fGeometryModule;
}
const ParsedModule& Compiler::loadFPModule() {
if (!fFPModule.fSymbols) {
fFPModule = this->parseModule(Program::kFragmentProcessor_Kind, MODULE_DATA(fp),
this->loadGPUModule());
}
return fFPModule;
}
const ParsedModule& Compiler::loadPublicModule() {
if (!fPublicModule.fSymbols) {
fPublicModule = this->parseModule(Program::kGeneric_Kind, MODULE_DATA(public), fRootModule);
}
return fPublicModule;
}
const ParsedModule& Compiler::loadPipelineModule() {
if (!fPipelineModule.fSymbols) {
fPipelineModule = this->parseModule(Program::kPipelineStage_Kind, MODULE_DATA(pipeline),
this->loadPublicModule());
// Add some aliases to the pipeline module so that it's friendlier, and more like GLSL
fPipelineModule.fSymbols->addAlias("shader", fContext->fFragmentProcessor_Type.get());
fPipelineModule.fSymbols->addAlias("vec2", fContext->fFloat2_Type.get());
fPipelineModule.fSymbols->addAlias("vec3", fContext->fFloat3_Type.get());
fPipelineModule.fSymbols->addAlias("vec4", fContext->fFloat4_Type.get());
fPipelineModule.fSymbols->addAlias("bvec2", fContext->fBool2_Type.get());
fPipelineModule.fSymbols->addAlias("bvec3", fContext->fBool3_Type.get());
fPipelineModule.fSymbols->addAlias("bvec4", fContext->fBool4_Type.get());
fPipelineModule.fSymbols->addAlias("mat2", fContext->fFloat2x2_Type.get());
fPipelineModule.fSymbols->addAlias("mat3", fContext->fFloat3x3_Type.get());
fPipelineModule.fSymbols->addAlias("mat4", fContext->fFloat4x4_Type.get());
fPipelineModule.fSymbols->addAlias("mat2x2", fContext->fFloat2x2_Type.get());
fPipelineModule.fSymbols->addAlias("mat2x3", fContext->fFloat2x3_Type.get());
fPipelineModule.fSymbols->addAlias("mat2x4", fContext->fFloat2x4_Type.get());
fPipelineModule.fSymbols->addAlias("mat3x2", fContext->fFloat3x2_Type.get());
fPipelineModule.fSymbols->addAlias("mat3x3", fContext->fFloat3x3_Type.get());
fPipelineModule.fSymbols->addAlias("mat3x4", fContext->fFloat3x4_Type.get());
fPipelineModule.fSymbols->addAlias("mat4x2", fContext->fFloat4x2_Type.get());
fPipelineModule.fSymbols->addAlias("mat4x3", fContext->fFloat4x3_Type.get());
fPipelineModule.fSymbols->addAlias("mat4x4", fContext->fFloat4x4_Type.get());
}
return fPipelineModule;
}
const ParsedModule& Compiler::loadInterpreterModule() {
if (!fInterpreterModule.fSymbols) {
fInterpreterModule = this->parseModule(Program::kGeneric_Kind, MODULE_DATA(interp),
this->loadPublicModule());
}
return fInterpreterModule;
}
const ParsedModule& Compiler::moduleForProgramKind(Program::Kind kind) {
switch (kind) {
case Program::kVertex_Kind: return this->loadVertexModule(); break;
case Program::kFragment_Kind: return this->loadFragmentModule(); break;
case Program::kGeometry_Kind: return this->loadGeometryModule(); break;
case Program::kFragmentProcessor_Kind: return this->loadFPModule(); break;
case Program::kPipelineStage_Kind: return this->loadPipelineModule(); break;
case Program::kGeneric_Kind: return this->loadInterpreterModule(); break;
}
SkUNREACHABLE;
}
LoadedModule Compiler::loadModule(Program::Kind kind,
ModuleData data,
std::shared_ptr<SymbolTable> base) {
if (!base) {
// NOTE: This is a workaround. The only time 'base' is null is when dehydrating includes.
// In that case, skslc doesn't know which module it's preparing, nor what the correct base
// module is. We can't use 'Root', because many GPU intrinsics reference private types,
// like samplers or textures. Today, 'Private' does contain the union of all known types,
// so this is safe. If we ever have types that only exist in 'Public' (for example), this
// logic needs to be smarter (by choosing the correct base for the module we're compiling).
base = fPrivateSymbolTable;
}
#if defined(SKSL_STANDALONE)
SkASSERT(data.fPath);
std::ifstream in(data.fPath);
std::unique_ptr<String> text = std::make_unique<String>(std::istreambuf_iterator<char>(in),
std::istreambuf_iterator<char>());
if (in.rdstate()) {
printf("error reading %s\n", data.fPath);
abort();
}
const String* source = fRootSymbolTable->takeOwnershipOfString(std::move(text));
AutoSource as(this, source);
Program::Settings settings;
SkASSERT(fIRGenerator->fCanInline);
fIRGenerator->fCanInline = false;
ParsedModule baseModule = {base, /*fIntrinsics=*/nullptr};
IRGenerator::IRBundle ir =
fIRGenerator->convertProgram(kind, &settings, baseModule,
/*isBuiltinCode=*/true, source->c_str(), source->length(),
/*externalValues=*/nullptr);
SkASSERT(ir.fSharedElements.empty());
LoadedModule module = { kind, std::move(ir.fSymbolTable), std::move(ir.fElements) };
fIRGenerator->fCanInline = true;
if (this->fErrorCount) {
printf("Unexpected errors: %s\n", this->fErrorText.c_str());
SkDEBUGFAILF("%s %s\n", data.fPath, this->fErrorText.c_str());
}
fModifiers.push_back(std::move(ir.fModifiers));
#else
SkASSERT(data.fData && (data.fSize != 0));
Rehydrator rehydrator(fContext.get(), fIRGenerator->fModifiers.get(), base, this,
data.fData, data.fSize);
LoadedModule module = { kind, rehydrator.symbolTable(), rehydrator.elements() };
fModifiers.push_back(fIRGenerator->releaseModifiers());
#endif
return module;
}
ParsedModule Compiler::parseModule(Program::Kind kind, ModuleData data, const ParsedModule& base) {
LoadedModule module = this->loadModule(kind, data, base.fSymbols);
this->optimize(module);
// For modules that just declare (but don't define) intrinsic functions, there will be no new
// program elements. In that case, we can share our parent's intrinsic map:
if (module.fElements.empty()) {
return {module.fSymbols, base.fIntrinsics};
}
auto intrinsics = std::make_shared<IRIntrinsicMap>(base.fIntrinsics.get());
// Now, transfer all of the program elements to an intrinsic map. This maps certain types of
// global objects to the declaring ProgramElement.
for (std::unique_ptr<ProgramElement>& element : module.fElements) {
switch (element->kind()) {
case ProgramElement::Kind::kFunction: {
const FunctionDefinition& f = element->as<FunctionDefinition>();
SkASSERT(f.declaration().isBuiltin());
intrinsics->insertOrDie(f.declaration().description(), std::move(element));
break;
}
case ProgramElement::Kind::kFunctionPrototype: {
// These are already in the symbol table.
break;
}
case ProgramElement::Kind::kEnum: {
const Enum& e = element->as<Enum>();
SkASSERT(e.isBuiltin());
intrinsics->insertOrDie(e.typeName(), std::move(element));
break;
}
case ProgramElement::Kind::kGlobalVar: {
const GlobalVarDeclaration& global = element->as<GlobalVarDeclaration>();
const Variable& var = global.declaration()->as<VarDeclaration>().var();
SkASSERT(var.isBuiltin());
intrinsics->insertOrDie(var.name(), std::move(element));
break;
}
case ProgramElement::Kind::kInterfaceBlock: {
const Variable& var = element->as<InterfaceBlock>().variable();
SkASSERT(var.isBuiltin());
intrinsics->insertOrDie(var.name(), std::move(element));
break;
}
default:
printf("Unsupported element: %s\n", element->description().c_str());
SkASSERT(false);
break;
}
}
return {module.fSymbols, std::move(intrinsics)};
}
// add the definition created by assigning to the lvalue to the definition set
void Compiler::addDefinition(const Expression* lvalue, std::unique_ptr<Expression>* expr,
DefinitionMap* definitions) {
switch (lvalue->kind()) {
case Expression::Kind::kVariableReference: {
const Variable& var = *lvalue->as<VariableReference>().variable();
if (var.storage() == Variable::Storage::kLocal) {
definitions->set(&var, expr);
}
break;
}
case Expression::Kind::kSwizzle:
// We consider the variable written to as long as at least some of its components have
// been written to. This will lead to some false negatives (we won't catch it if you
// write to foo.x and then read foo.y), but being stricter could lead to false positives
// (we write to foo.x, and then pass foo to a function which happens to only read foo.x,
// but since we pass foo as a whole it is flagged as an error) unless we perform a much
// more complicated whole-program analysis. This is probably good enough.
this->addDefinition(lvalue->as<Swizzle>().base().get(),
(std::unique_ptr<Expression>*) &fContext->fDefined_Expression,
definitions);
break;
case Expression::Kind::kIndex:
// see comments in Swizzle
this->addDefinition(lvalue->as<IndexExpression>().base().get(),
(std::unique_ptr<Expression>*) &fContext->fDefined_Expression,
definitions);
break;
case Expression::Kind::kFieldAccess:
// see comments in Swizzle
this->addDefinition(lvalue->as<FieldAccess>().base().get(),
(std::unique_ptr<Expression>*) &fContext->fDefined_Expression,
definitions);
break;
case Expression::Kind::kTernary:
// To simplify analysis, we just pretend that we write to both sides of the ternary.
// This allows for false positives (meaning we fail to detect that a variable might not
// have been assigned), but is preferable to false negatives.
this->addDefinition(lvalue->as<TernaryExpression>().ifTrue().get(),
(std::unique_ptr<Expression>*) &fContext->fDefined_Expression,
definitions);
this->addDefinition(lvalue->as<TernaryExpression>().ifFalse().get(),
(std::unique_ptr<Expression>*) &fContext->fDefined_Expression,
definitions);
break;
case Expression::Kind::kExternalValue:
break;
default:
// not an lvalue, can't happen
SkASSERT(false);
}
}
// add local variables defined by this node to the set
void Compiler::addDefinitions(const BasicBlock::Node& node, DefinitionMap* definitions) {
if (node.isExpression()) {
Expression* expr = node.expression()->get();
switch (expr->kind()) {
case Expression::Kind::kBinary: {
BinaryExpression* b = &expr->as<BinaryExpression>();
if (b->getOperator() == Token::Kind::TK_EQ) {
this->addDefinition(b->left().get(), &b->right(), definitions);
} else if (Compiler::IsAssignment(b->getOperator())) {
this->addDefinition(
b->left().get(),
(std::unique_ptr<Expression>*) &fContext->fDefined_Expression,
definitions);
}
break;
}
case Expression::Kind::kFunctionCall: {
const FunctionCall& c = expr->as<FunctionCall>();
const std::vector<const Variable*>& parameters = c.function().parameters();
for (size_t i = 0; i < parameters.size(); ++i) {
if (parameters[i]->modifiers().fFlags & Modifiers::kOut_Flag) {
this->addDefinition(
c.arguments()[i].get(),
(std::unique_ptr<Expression>*) &fContext->fDefined_Expression,
definitions);
}
}
break;
}
case Expression::Kind::kPrefix: {
const PrefixExpression* p = &expr->as<PrefixExpression>();
if (p->getOperator() == Token::Kind::TK_MINUSMINUS ||
p->getOperator() == Token::Kind::TK_PLUSPLUS) {
this->addDefinition(
p->operand().get(),
(std::unique_ptr<Expression>*) &fContext->fDefined_Expression,
definitions);
}
break;
}
case Expression::Kind::kPostfix: {
const PostfixExpression* p = &expr->as<PostfixExpression>();
if (p->getOperator() == Token::Kind::TK_MINUSMINUS ||
p->getOperator() == Token::Kind::TK_PLUSPLUS) {
this->addDefinition(
p->operand().get(),
(std::unique_ptr<Expression>*) &fContext->fDefined_Expression,
definitions);
}
break;
}
case Expression::Kind::kVariableReference: {
const VariableReference* v = &expr->as<VariableReference>();
if (v->refKind() != VariableReference::RefKind::kRead) {
this->addDefinition(
v,
(std::unique_ptr<Expression>*) &fContext->fDefined_Expression,
definitions);
}
break;
}
default:
break;
}
} else if (node.isStatement()) {
Statement* stmt = node.statement()->get();
if (stmt->is<VarDeclaration>()) {
VarDeclaration& vd = stmt->as<VarDeclaration>();
if (vd.value()) {
definitions->set(&vd.var(), &vd.value());
}
}
}
}
void Compiler::scanCFG(CFG* cfg, BlockId blockId, SkBitSet* processedSet) {
BasicBlock& block = cfg->fBlocks[blockId];
// compute definitions after this block
DefinitionMap after = block.fBefore;
for (const BasicBlock::Node& n : block.fNodes) {
this->addDefinitions(n, &after);
}
// propagate definitions to exits
for (BlockId exitId : block.fExits) {
if (exitId == blockId) {
continue;
}
BasicBlock& exit = cfg->fBlocks[exitId];
after.foreach([&](const Variable* var, std::unique_ptr<Expression>** e1Ptr) {
std::unique_ptr<Expression>* e1 = *e1Ptr;
std::unique_ptr<Expression>** exitDef = exit.fBefore.find(var);
if (!exitDef) {
// exit has no definition for it, just copy it and reprocess exit block
processedSet->reset(exitId);
exit.fBefore[var] = e1;
} else {
// exit has a (possibly different) value already defined
std::unique_ptr<Expression>* e2 = *exitDef;
if (e1 != e2) {
// definition has changed, merge and reprocess the exit block
processedSet->reset(exitId);
if (e1 && e2) {
*exitDef = (std::unique_ptr<Expression>*)&fContext->fDefined_Expression;
} else {
*exitDef = nullptr;
}
}
}
});
}
}
// returns a map which maps all local variables in the function to null, indicating that their value
// is initially unknown
static DefinitionMap compute_start_state(const CFG& cfg) {
DefinitionMap result;
for (const auto& block : cfg.fBlocks) {
for (const auto& node : block.fNodes) {
if (node.isStatement()) {
const Statement* s = node.statement()->get();
if (s->is<VarDeclaration>()) {
result[&s->as<VarDeclaration>().var()] = nullptr;
}
}
}
}
return result;
}
/**
* Returns true if assigning to this lvalue has no effect.
*/
static bool is_dead(const Expression& lvalue, ProgramUsage* usage) {
switch (lvalue.kind()) {
case Expression::Kind::kVariableReference:
return usage->isDead(*lvalue.as<VariableReference>().variable());
case Expression::Kind::kSwizzle:
return is_dead(*lvalue.as<Swizzle>().base(), usage);
case Expression::Kind::kFieldAccess:
return is_dead(*lvalue.as<FieldAccess>().base(), usage);
case Expression::Kind::kIndex: {
const IndexExpression& idx = lvalue.as<IndexExpression>();
return is_dead(*idx.base(), usage) &&
!idx.index()->hasProperty(Expression::Property::kSideEffects);
}
case Expression::Kind::kTernary: {
const TernaryExpression& t = lvalue.as<TernaryExpression>();
return !t.test()->hasSideEffects() &&
is_dead(*t.ifTrue(), usage) &&
is_dead(*t.ifFalse(), usage);
}
case Expression::Kind::kExternalValue:
return false;
default:
#ifdef SK_DEBUG
ABORT("invalid lvalue: %s\n", lvalue.description().c_str());
#endif
return false;
}
}
/**
* Returns true if this is an assignment which can be collapsed down to just the right hand side due
* to a dead target and lack of side effects on the left hand side.
*/
static bool dead_assignment(const BinaryExpression& b, ProgramUsage* usage) {
if (!Compiler::IsAssignment(b.getOperator())) {
return false;
}
return is_dead(*b.left(), usage);
}
void Compiler::computeDataFlow(CFG* cfg) {
cfg->fBlocks[cfg->fStart].fBefore = compute_start_state(*cfg);
// We set bits in the "processed" set after a block has been scanned.
SkBitSet processedSet(cfg->fBlocks.size());
while (SkBitSet::OptionalIndex blockId = processedSet.findFirstUnset()) {
processedSet.set(*blockId);
this->scanCFG(cfg, *blockId, &processedSet);
}
}
/**
* Attempts to replace the expression pointed to by iter with a new one (in both the CFG and the
* IR). If the expression can be cleanly removed, returns true and updates the iterator to point to
* the newly-inserted element. Otherwise updates only the IR and returns false (and the CFG will
* need to be regenerated).
*/
static bool try_replace_expression(BasicBlock* b,
std::vector<BasicBlock::Node>::iterator* iter,
std::unique_ptr<Expression>* newExpression) {
std::unique_ptr<Expression>* target = (*iter)->expression();
if (!b->tryRemoveExpression(iter)) {
*target = std::move(*newExpression);
return false;
}
*target = std::move(*newExpression);
return b->tryInsertExpression(iter, target);
}
/**
* Returns true if the expression is a constant numeric literal with the specified value, or a
* constant vector with all elements equal to the specified value.
*/
template <typename T = SKSL_FLOAT>
static bool is_constant(const Expression& expr, T value) {
switch (expr.kind()) {
case Expression::Kind::kIntLiteral:
return expr.as<IntLiteral>().value() == value;
case Expression::Kind::kFloatLiteral:
return expr.as<FloatLiteral>().value() == value;
case Expression::Kind::kConstructor: {
const Constructor& constructor = expr.as<Constructor>();
if (constructor.isCompileTimeConstant()) {
const Type& constructorType = constructor.type();
switch (constructorType.typeKind()) {
case Type::TypeKind::kVector:
if (constructor.componentType().isFloat()) {
for (int i = 0; i < constructorType.columns(); ++i) {
if (constructor.getFVecComponent(i) != value) {
return false;
}
}
return true;
} else if (constructor.componentType().isInteger()) {
for (int i = 0; i < constructorType.columns(); ++i) {
if (constructor.getIVecComponent(i) != value) {
return false;
}
}
return true;
}
// Other types (e.g. boolean) might occur, but aren't supported here.
return false;
case Type::TypeKind::kScalar:
SkASSERT(constructor.arguments().size() == 1);
return is_constant<T>(*constructor.arguments()[0], value);
default:
return false;
}
}
return false;
}
default:
return false;
}
}
/**
* Collapses the binary expression pointed to by iter down to just the right side (in both the IR
* and CFG structures).
*/
static void delete_left(BasicBlock* b,
std::vector<BasicBlock::Node>::iterator* iter,
Compiler::OptimizationContext* optimizationContext) {
optimizationContext->fUpdated = true;
std::unique_ptr<Expression>* target = (*iter)->expression();
BinaryExpression& bin = (*target)->as<BinaryExpression>();
Expression& left = *bin.left();
std::unique_ptr<Expression>& rightPointer = bin.right();
SkASSERT(!left.hasSideEffects());
bool result;
if (bin.getOperator() == Token::Kind::TK_EQ) {
result = b->tryRemoveLValueBefore(iter, &left);
} else {
result = b->tryRemoveExpressionBefore(iter, &left);
}
// Remove references within LHS.
optimizationContext->fUsage->remove(&left);
*target = std::move(rightPointer);
if (!result) {
optimizationContext->fNeedsRescan = true;
return;
}
if (*iter == b->fNodes.begin()) {
optimizationContext->fNeedsRescan = true;
return;
}
--(*iter);
if (!(*iter)->isExpression() || (*iter)->expression() != &rightPointer) {
optimizationContext->fNeedsRescan = true;
return;
}
*iter = b->fNodes.erase(*iter);
SkASSERT((*iter)->expression() == target);
}
/**
* Collapses the binary expression pointed to by iter down to just the left side (in both the IR and
* CFG structures).
*/
static void delete_right(BasicBlock* b,
std::vector<BasicBlock::Node>::iterator* iter,
Compiler::OptimizationContext* optimizationContext) {
optimizationContext->fUpdated = true;
std::unique_ptr<Expression>* target = (*iter)->expression();
BinaryExpression& bin = (*target)->as<BinaryExpression>();
std::unique_ptr<Expression>& leftPointer = bin.left();
Expression& right = *bin.right();
SkASSERT(!right.hasSideEffects());
// Remove references within RHS.
optimizationContext->fUsage->remove(&right);
if (!b->tryRemoveExpressionBefore(iter, &right)) {
*target = std::move(leftPointer);
optimizationContext->fNeedsRescan = true;
return;
}
*target = std::move(leftPointer);
if (*iter == b->fNodes.begin()) {
optimizationContext->fNeedsRescan = true;
return;
}
--(*iter);
if ((!(*iter)->isExpression() || (*iter)->expression() != &leftPointer)) {
optimizationContext->fNeedsRescan = true;
return;
}
*iter = b->fNodes.erase(*iter);
SkASSERT((*iter)->expression() == target);
}
/**
* Constructs the specified type using a single argument.
*/
static std::unique_ptr<Expression> construct(const Type* type, std::unique_ptr<Expression> v) {
ExpressionArray args;
args.push_back(std::move(v));
std::unique_ptr<Expression> result = std::make_unique<Constructor>(-1, type, std::move(args));
return result;
}
/**
* Used in the implementations of vectorize_left and vectorize_right. Given a vector type and an
* expression x, deletes the expression pointed to by iter and replaces it with <type>(x).
*/
static void vectorize(BasicBlock* b,
std::vector<BasicBlock::Node>::iterator* iter,
const Type& type,
std::unique_ptr<Expression>* otherExpression,
Compiler::OptimizationContext* optimizationContext) {
SkASSERT((*(*iter)->expression())->kind() == Expression::Kind::kBinary);
SkASSERT(type.isVector());
SkASSERT((*otherExpression)->type().isScalar());
optimizationContext->fUpdated = true;
std::unique_ptr<Expression>* target = (*iter)->expression();
if (!b->tryRemoveExpression(iter)) {
*target = construct(&type, std::move(*otherExpression));
optimizationContext->fNeedsRescan = true;
} else {
*target = construct(&type, std::move(*otherExpression));
if (!b->tryInsertExpression(iter, target)) {
optimizationContext->fNeedsRescan = true;
}
}
}
/**
* Given a binary expression of the form x <op> vec<n>(y), deletes the right side and vectorizes the
* left to yield vec<n>(x).
*/
static void vectorize_left(BasicBlock* b,
std::vector<BasicBlock::Node>::iterator* iter,
Compiler::OptimizationContext* optimizationContext) {
BinaryExpression& bin = (*(*iter)->expression())->as<BinaryExpression>();
// Remove references within RHS. Vectorization of LHS doesn't change reference counts.
optimizationContext->fUsage->remove(bin.right().get());
vectorize(b, iter, bin.right()->type(), &bin.left(), optimizationContext);
}
/**
* Given a binary expression of the form vec<n>(x) <op> y, deletes the left side and vectorizes the
* right to yield vec<n>(y).
*/
static void vectorize_right(BasicBlock* b,
std::vector<BasicBlock::Node>::iterator* iter,
Compiler::OptimizationContext* optimizationContext) {
BinaryExpression& bin = (*(*iter)->expression())->as<BinaryExpression>();
// Remove references within LHS. Vectorization of RHS doesn't change reference counts.
optimizationContext->fUsage->remove(bin.left().get());
vectorize(b, iter, bin.left()->type(), &bin.right(), optimizationContext);
}
// Mark that an expression which we were writing to is no longer being written to
static void clear_write(Expression& expr) {
switch (expr.kind()) {
case Expression::Kind::kVariableReference: {
expr.as<VariableReference>().setRefKind(VariableReference::RefKind::kRead);
break;
}
case Expression::Kind::kFieldAccess:
clear_write(*expr.as<FieldAccess>().base());
break;
case Expression::Kind::kSwizzle:
clear_write(*expr.as<Swizzle>().base());
break;
case Expression::Kind::kIndex:
clear_write(*expr.as<IndexExpression>().base());
break;
default:
ABORT("shouldn't be writing to this kind of expression\n");
break;
}
}
void Compiler::simplifyExpression(DefinitionMap& definitions,
BasicBlock& b,
std::vector<BasicBlock::Node>::iterator* iter,
OptimizationContext* optimizationContext) {
Expression* expr = (*iter)->expression()->get();
SkASSERT(expr);
if ((*iter)->fConstantPropagation) {
std::unique_ptr<Expression> optimized = expr->constantPropagate(*fIRGenerator,
definitions);
if (optimized) {
optimizationContext->fUpdated = true;
optimized = fIRGenerator->coerce(std::move(optimized), expr->type());
SkASSERT(optimized);
// Remove references within 'expr', add references within 'optimized'
optimizationContext->fUsage->replace(expr, optimized.get());
if (!try_replace_expression(&b, iter, &optimized)) {
optimizationContext->fNeedsRescan = true;
return;
}
SkASSERT((*iter)->isExpression());
expr = (*iter)->expression()->get();
}
}
switch (expr->kind()) {
case Expression::Kind::kVariableReference: {
const VariableReference& ref = expr->as<VariableReference>();
const Variable* var = ref.variable();
if (ref.refKind() != VariableReference::RefKind::kWrite &&
ref.refKind() != VariableReference::RefKind::kPointer &&
var->storage() == Variable::Storage::kLocal && !definitions[var] &&
optimizationContext->fSilences.find(var) == optimizationContext->fSilences.end()) {
optimizationContext->fSilences.insert(var);
this->error(expr->fOffset,
"'" + var->name() + "' has not been assigned");
}
break;
}
case Expression::Kind::kTernary: {
TernaryExpression* t = &expr->as<TernaryExpression>();
if (t->test()->is<BoolLiteral>()) {
// ternary has a constant test, replace it with either the true or
// false branch
if (t->test()->as<BoolLiteral>().value()) {
(*iter)->setExpression(std::move(t->ifTrue()), optimizationContext->fUsage);
} else {
(*iter)->setExpression(std::move(t->ifFalse()), optimizationContext->fUsage);
}
optimizationContext->fUpdated = true;
optimizationContext->fNeedsRescan = true;
}
break;
}
case Expression::Kind::kBinary: {
BinaryExpression* bin = &expr->as<BinaryExpression>();
if (dead_assignment(*bin, optimizationContext->fUsage)) {
delete_left(&b, iter, optimizationContext);
break;
}
Expression& left = *bin->left();
Expression& right = *bin->right();
const Type& leftType = left.type();
const Type& rightType = right.type();
// collapse useless expressions like x * 1 or x + 0
if ((!leftType.isScalar() && !leftType.isVector()) ||
(!rightType.isScalar() && !rightType.isVector())) {
break;
}
switch (bin->getOperator()) {
case Token::Kind::TK_STAR:
if (is_constant(left, 1)) {
if (leftType.isVector() && rightType.isScalar()) {
// float4(1) * x -> float4(x)
vectorize_right(&b, iter, optimizationContext);
} else {
// 1 * x -> x
// 1 * float4(x) -> float4(x)
// float4(1) * float4(x) -> float4(x)
delete_left(&b, iter, optimizationContext);
}
}
else if (is_constant(left, 0)) {
if (leftType.isScalar() && rightType.isVector() &&
!right.hasSideEffects()) {
// 0 * float4(x) -> float4(0)
vectorize_left(&b, iter, optimizationContext);
} else {
// 0 * x -> 0
// float4(0) * x -> float4(0)
// float4(0) * float4(x) -> float4(0)
if (!right.hasSideEffects()) {
delete_right(&b, iter, optimizationContext);
}
}
}
else if (is_constant(right, 1)) {
if (leftType.isScalar() && rightType.isVector()) {
// x * float4(1) -> float4(x)
vectorize_left(&b, iter, optimizationContext);
} else {
// x * 1 -> x
// float4(x) * 1 -> float4(x)
// float4(x) * float4(1) -> float4(x)
delete_right(&b, iter, optimizationContext);
}
}
else if (is_constant(right, 0)) {
if (leftType.isVector() && rightType.isScalar() && !left.hasSideEffects()) {
// float4(x) * 0 -> float4(0)
vectorize_right(&b, iter, optimizationContext);
} else {
// x * 0 -> 0
// x * float4(0) -> float4(0)
// float4(x) * float4(0) -> float4(0)
if (!left.hasSideEffects()) {
delete_left(&b, iter, optimizationContext);
}
}
}
break;
case Token::Kind::TK_PLUS:
if (is_constant(left, 0)) {
if (leftType.isVector() && rightType.isScalar()) {
// float4(0) + x -> float4(x)
vectorize_right(&b, iter, optimizationContext);
} else {
// 0 + x -> x
// 0 + float4(x) -> float4(x)
// float4(0) + float4(x) -> float4(x)
delete_left(&b, iter, optimizationContext);
}
} else if (is_constant(right, 0)) {
if (leftType.isScalar() && rightType.isVector()) {
// x + float4(0) -> float4(x)
vectorize_left(&b, iter, optimizationContext);
} else {
// x + 0 -> x
// float4(x) + 0 -> float4(x)
// float4(x) + float4(0) -> float4(x)
delete_right(&b, iter, optimizationContext);
}
}
break;
case Token::Kind::TK_MINUS:
if (is_constant(right, 0)) {
if (leftType.isScalar() && rightType.isVector()) {
// x - float4(0) -> float4(x)
vectorize_left(&b, iter, optimizationContext);
} else {
// x - 0 -> x
// float4(x) - 0 -> float4(x)
// float4(x) - float4(0) -> float4(x)
delete_right(&b, iter, optimizationContext);
}
}
break;
case Token::Kind::TK_SLASH:
if (is_constant(right, 1)) {
if (leftType.isScalar() && rightType.isVector()) {
// x / float4(1) -> float4(x)
vectorize_left(&b, iter, optimizationContext);
} else {
// x / 1 -> x
// float4(x) / 1 -> float4(x)
// float4(x) / float4(1) -> float4(x)
delete_right(&b, iter, optimizationContext);
}
} else if (is_constant(left, 0)) {
if (leftType.isScalar() && rightType.isVector() &&
!right.hasSideEffects()) {
// 0 / float4(x) -> float4(0)
vectorize_left(&b, iter, optimizationContext);
} else {
// 0 / x -> 0
// float4(0) / x -> float4(0)
// float4(0) / float4(x) -> float4(0)
if (!right.hasSideEffects()) {
delete_right(&b, iter, optimizationContext);
}
}
}
break;
case Token::Kind::TK_PLUSEQ:
if (is_constant(right, 0)) {
clear_write(left);
delete_right(&b, iter, optimizationContext);
}
break;
case Token::Kind::TK_MINUSEQ:
if (is_constant(right, 0)) {
clear_write(left);
delete_right(&b, iter, optimizationContext);
}
break;
case Token::Kind::TK_STAREQ:
if (is_constant(right, 1)) {
clear_write(left);
delete_right(&b, iter, optimizationContext);
}
break;
case Token::Kind::TK_SLASHEQ:
if (is_constant(right, 1)) {
clear_write(left);
delete_right(&b, iter, optimizationContext);
}
break;
default:
break;
}
break;
}
case Expression::Kind::kConstructor: {
// Find constructors embedded inside constructors and flatten them out where possible.
// - float4(float2(1, 2), 3, 4) --> float4(1, 2, 3, 4)
// - float4(w, float3(sin(x), cos(y), tan(z))) --> float4(w, sin(x), cos(y), tan(z))
// Leave single-argument constructors alone, though. These might be casts or splats.
Constructor& c = expr->as<Constructor>();
if (c.type().columns() > 1) {
// Inspect each constructor argument to see if it's a candidate for flattening.
// Remember matched arguments in a bitfield, "argsToOptimize".
int argsToOptimize = 0;
int currBit = 1;
for (const std::unique_ptr<Expression>& arg : c.arguments()) {
if (arg->is<Constructor>()) {
Constructor& inner = arg->as<Constructor>();
if (inner.arguments().size() > 1 &&
inner.type().componentType() == c.type().componentType()) {
argsToOptimize |= currBit;
}
}
currBit <<= 1;
}
if (argsToOptimize) {
// We found at least one argument that could be flattened out. Re-walk the
// constructor args and flatten the candidates we found during our initial pass.
ExpressionArray flattened;
flattened.reserve_back(c.type().columns());
currBit = 1;
for (const std::unique_ptr<Expression>& arg : c.arguments()) {
if (argsToOptimize & currBit) {
Constructor& inner = arg->as<Constructor>();
for (const std::unique_ptr<Expression>& innerArg : inner.arguments()) {
flattened.push_back(innerArg->clone());
}
} else {
flattened.push_back(arg->clone());
}
currBit <<= 1;
}
auto optimized = std::unique_ptr<Expression>(
new Constructor(c.fOffset, &c.type(), std::move(flattened)));
// No fUsage change; no references have been added or removed anywhere.
optimizationContext->fUpdated = true;
if (!try_replace_expression(&b, iter, &optimized)) {
optimizationContext->fNeedsRescan = true;
return;
}
SkASSERT((*iter)->isExpression());
break;
}
}
break;
}
case Expression::Kind::kSwizzle: {
Swizzle& s = expr->as<Swizzle>();
// Detect identity swizzles like `foo.rgba`.
if ((int) s.components().size() == s.base()->type().columns()) {
bool identity = true;
for (int i = 0; i < (int) s.components().size(); ++i) {
if (s.components()[i] != i) {
identity = false;
break;
}
}
if (identity) {
optimizationContext->fUpdated = true;
// No fUsage change: foo.rgba and foo have equivalent reference counts
if (!try_replace_expression(&b, iter, &s.base())) {
optimizationContext->fNeedsRescan = true;
return;
}
SkASSERT((*iter)->isExpression());
break;
}
}
// Detect swizzles of swizzles, e.g. replace `foo.argb.r000` with `foo.a000`.
if (s.base()->is<Swizzle>()) {
Swizzle& base = s.base()->as<Swizzle>();
ComponentArray final;
for (int c : s.components()) {
final.push_back(base.components()[c]);
}
optimizationContext->fUpdated = true;
std::unique_ptr<Expression> replacement(new Swizzle(*fContext, base.base()->clone(),
final));
// No fUsage change: `foo.gbr.gbr` and `foo.brg` have equivalent reference counts
if (!try_replace_expression(&b, iter, &replacement)) {
optimizationContext->fNeedsRescan = true;
return;
}
SkASSERT((*iter)->isExpression());
break;
}
// Optimize swizzles of constructors.
if (s.base()->is<Constructor>()) {
Constructor& base = s.base()->as<Constructor>();
std::unique_ptr<Expression> replacement;
const Type& componentType = base.type().componentType();
int swizzleSize = s.components().size();
// The IR generator has already converted any zero/one swizzle components into
// constructors containing zero/one args. Confirm that this is true by checking that
// our swizzle components are all `xyzw` (values 0 through 3).
SkASSERT(std::all_of(s.components().begin(), s.components().end(),
[](int8_t c) { return c >= 0 && c <= 3; }));
if (base.arguments().size() == 1 && base.arguments().front()->type().isScalar()) {
// `half4(scalar).zyy` can be optimized to `half3(scalar)`. The swizzle
// components don't actually matter since all fields are the same.
ExpressionArray newArgs;
newArgs.push_back(base.arguments().front()->clone());
replacement = std::make_unique<Constructor>(
base.fOffset,
&componentType.toCompound(*fContext, swizzleSize, /*rows=*/1),
std::move(newArgs));
// No fUsage change: `half4(foo).xy` and `half2(foo)` have equivalent reference
// counts.
optimizationContext->fUpdated = true;
if (!try_replace_expression(&b, iter, &replacement)) {
optimizationContext->fNeedsRescan = true;
return;
}
SkASSERT((*iter)->isExpression());
break;
}
// Swizzles can duplicate some elements and discard others, e.g.
// `half4(1, 2, 3, 4).xxz` --> `half3(1, 1, 3)`. However, there are constraints:
// - Expressions with side effects need to occur exactly once, even if they
// would otherwise be swizzle-eliminated
// - Non-trivial expressions should not be repeated, but elimination is OK.
//
// Look up the argument for the constructor at each index. This is typically simple
// but for weird cases like `half4(bar.yz, half2(foo))`, it can be harder than it
// seems. This example would result in:
// argMap[0] = {.fArgIndex = 0, .fComponent = 0} (bar.yz .x)
// argMap[1] = {.fArgIndex = 0, .fComponent = 1} (bar.yz .y)
// argMap[2] = {.fArgIndex = 1, .fComponent = 0} (half2(foo) .x)
// argMap[3] = {.fArgIndex = 1, .fComponent = 1} (half2(foo) .y)
struct ConstructorArgMap {
int8_t fArgIndex;
int8_t fComponent;
};
int numConstructorArgs = base.type().columns();
ConstructorArgMap argMap[4] = {};
int writeIdx = 0;
for (int argIdx = 0; argIdx < (int) base.arguments().size(); ++argIdx) {
const Expression& expr = *base.arguments()[argIdx];
int argWidth = expr.type().columns();
for (int componentIdx = 0; componentIdx < argWidth; ++componentIdx) {
argMap[writeIdx].fArgIndex = argIdx;
argMap[writeIdx].fComponent = componentIdx;
++writeIdx;
}
}
SkASSERT(writeIdx == numConstructorArgs);
// Count up the number of times each constructor argument is used by the
// swizzle.
// `half4(bar.yz, half2(foo)).xwxy` -> { 3, 1 }
// - bar.yz is referenced 3 times, by `.x_xy`
// - half(foo) is referenced 1 time, by `._w__`
int8_t exprUsed[4] = {};
for (int c : s.components()) {
exprUsed[argMap[c].fArgIndex]++;
}
bool safeToOptimize = true;
for (int index = 0; index < numConstructorArgs; ++index) {
int8_t constructorArgIndex = argMap[index].fArgIndex;
const Expression& baseArg = *base.arguments()[constructorArgIndex];
// Check that non-trivial expressions are not swizzled in more than once.
if (exprUsed[constructorArgIndex] > 1 &&
!Analysis::IsTrivialExpression(baseArg)) {
safeToOptimize = false;
break;
}
// Check that side-effect-bearing expressions are swizzled in exactly once.
if (exprUsed[constructorArgIndex] != 1 && baseArg.hasSideEffects()) {
safeToOptimize = false;
break;
}
}
if (safeToOptimize) {
struct ReorderedArgument {
int8_t fArgIndex;
ComponentArray fComponents;
};
SkSTArray<4, ReorderedArgument> reorderedArgs;
for (int c : s.components()) {
const ConstructorArgMap& argument = argMap[c];
const Expression& baseArg = *base.arguments()[argument.fArgIndex];
if (baseArg.type().isScalar()) {
// This argument is a scalar; add it to the list as-is.
SkASSERT(argument.fComponent == 0);
reorderedArgs.push_back({argument.fArgIndex,
ComponentArray{}});
} else {
// This argument is a component from a vector.
SkASSERT(argument.fComponent < baseArg.type().columns());
if (reorderedArgs.empty() ||
reorderedArgs.back().fArgIndex != argument.fArgIndex) {
// This can't be combined with the previous argument. Add a new one.
reorderedArgs.push_back({argument.fArgIndex,
ComponentArray{argument.fComponent}});
} else {
// Since we know this argument uses components, it should already
// have at least one component set.
SkASSERT(!reorderedArgs.back().fComponents.empty());
// Build up the current argument with one more component.
reorderedArgs.back().fComponents.push_back(argument.fComponent);
}
}
}
// Convert our reordered argument list to an actual array of expressions, with
// the new order and any new inner swizzles that need to be applied. Note that
// we expect followup passes to clean up the inner swizzles.
ExpressionArray newArgs;
newArgs.reserve_back(swizzleSize);
for (const ReorderedArgument& reorderedArg : reorderedArgs) {
const Expression& baseArg = *base.arguments()[reorderedArg.fArgIndex];
if (reorderedArg.fComponents.empty()) {
newArgs.push_back(baseArg.clone());
} else {
newArgs.push_back(std::make_unique<Swizzle>(*fContext, baseArg.clone(),
reorderedArg.fComponents));
}
}
// Create a new constructor.
replacement = std::make_unique<Constructor>(
base.fOffset,
&componentType.toCompound(*fContext, swizzleSize, /*rows=*/1),
std::move(newArgs));
// Remove references within 'expr', add references within 'optimized'
optimizationContext->fUpdated = true;
optimizationContext->fUsage->replace(expr, replacement.get());
if (!try_replace_expression(&b, iter, &replacement)) {
optimizationContext->fNeedsRescan = true;
return;
}
SkASSERT((*iter)->isExpression());
}
break;
}
break;
}
default:
break;
}
}
// Returns true if this statement could potentially execute a break at the current level. We ignore
// nested loops and switches, since any breaks inside of them will merely break the loop / switch.
static bool contains_conditional_break(Statement& stmt) {
class ContainsConditionalBreak : public ProgramVisitor {
public:
bool visitStatement(const Statement& stmt) override {
switch (stmt.kind()) {
case Statement::Kind::kBlock:
return this->INHERITED::visitStatement(stmt);
case Statement::Kind::kBreak:
return fInConditional > 0;
case Statement::Kind::kIf: {
++fInConditional;
bool result = this->INHERITED::visitStatement(stmt);
--fInConditional;
return result;
}
default:
return false;
}
}
int fInConditional = 0;
using INHERITED = ProgramVisitor;
};
return ContainsConditionalBreak{}.visitStatement(stmt);
}
// returns true if this statement definitely executes a break at the current level (we ignore
// nested loops and switches, since any breaks inside of them will merely break the loop / switch)
static bool contains_unconditional_break(Statement& stmt) {
class ContainsUnconditionalBreak : public ProgramVisitor {
public:
bool visitStatement(const Statement& stmt) override {
switch (stmt.kind()) {
case Statement::Kind::kBlock:
return this->INHERITED::visitStatement(stmt);
case Statement::Kind::kBreak:
return true;
default:
return false;
}
}
using INHERITED = ProgramVisitor;
};
return ContainsUnconditionalBreak{}.visitStatement(stmt);
}
static void move_all_but_break(std::unique_ptr<Statement>& stmt, StatementArray* target) {
switch (stmt->kind()) {
case Statement::Kind::kBlock: {
// Recurse into the block.
Block& block = static_cast<Block&>(*stmt);
StatementArray blockStmts;
blockStmts.reserve_back(block.children().size());
for (std::unique_ptr<Statement>& stmt : block.children()) {
move_all_but_break(stmt, &blockStmts);
}
target->push_back(std::make_unique<Block>(block.fOffset, std::move(blockStmts),
block.symbolTable(), block.isScope()));
break;
}
case Statement::Kind::kBreak:
// Do not append a break to the target.
break;
default:
// Append normal statements to the target.
target->push_back(std::move(stmt));
break;
}
}
// Returns a block containing all of the statements that will be run if the given case matches
// (which, owing to the statements being owned by unique_ptrs, means the switch itself will be
// broken by this call and must then be discarded).
// Returns null (and leaves the switch unmodified) if no such simple reduction is possible, such as
// when break statements appear inside conditionals.
static std::unique_ptr<Statement> block_for_case(SwitchStatement* switchStatement,
SwitchCase* caseToCapture) {
// We have to be careful to not move any of the pointers until after we're sure we're going to
// succeed, so before we make any changes at all, we check the switch-cases to decide on a plan
// of action. First, find the switch-case we are interested in.
auto iter = switchStatement->cases().begin();
for (; iter != switchStatement->cases().end(); ++iter) {
if (iter->get() == caseToCapture) {
break;
}
}
// Next, walk forward through the rest of the switch. If we find a conditional break, we're
// stuck and can't simplify at all. If we find an unconditional break, we have a range of
// statements that we can use for simplification.
auto startIter = iter;
Statement* unconditionalBreakStmt = nullptr;
for (; iter != switchStatement->cases().end(); ++iter) {
for (std::unique_ptr<Statement>& stmt : (*iter)->statements()) {
if (contains_conditional_break(*stmt)) {
// We can't reduce switch-cases to a block when they have conditional breaks.
return nullptr;
}
if (contains_unconditional_break(*stmt)) {
// We found an unconditional break. We can use this block, but we need to strip
// out the break statement.
unconditionalBreakStmt = stmt.get();
break;
}
}
if (unconditionalBreakStmt != nullptr) {
break;
}
}
// We fell off the bottom of the switch or encountered a break. We know the range of statements
// that we need to move over, and we know it's safe to do so.
StatementArray caseStmts;
// We can move over most of the statements as-is.
while (startIter != iter) {
for (std::unique_ptr<Statement>& stmt : (*startIter)->statements()) {
caseStmts.push_back(std::move(stmt));
}
++startIter;
}
// If we found an unconditional break at the end, we need to move what we can while avoiding
// that break.
if (unconditionalBreakStmt != nullptr) {
for (std::unique_ptr<Statement>& stmt : (*startIter)->statements()) {
if (stmt.get() == unconditionalBreakStmt) {
move_all_but_break(stmt, &caseStmts);
unconditionalBreakStmt = nullptr;
break;
}
caseStmts.push_back(std::move(stmt));
}
}
SkASSERT(unconditionalBreakStmt == nullptr); // Verify that we fixed the unconditional break.
// Return our newly-synthesized block.
return std::make_unique<Block>(/*offset=*/-1, std::move(caseStmts), switchStatement->symbols());
}
void Compiler::simplifyStatement(DefinitionMap& definitions,
BasicBlock& b,
std::vector<BasicBlock::Node>::iterator* iter,
OptimizationContext* optimizationContext) {
ProgramUsage* usage = optimizationContext->fUsage;
Statement* stmt = (*iter)->statement()->get();
switch (stmt->kind()) {
case Statement::Kind::kVarDeclaration: {
const auto& varDecl = stmt->as<VarDeclaration>();
if (usage->isDead(varDecl.var()) &&
(!varDecl.value() ||
!varDecl.value()->hasSideEffects())) {
if (varDecl.value()) {
SkASSERT((*iter)->statement()->get() == stmt);
if (!b.tryRemoveExpressionBefore(iter, varDecl.value().get())) {
optimizationContext->fNeedsRescan = true;
}
}
(*iter)->setStatement(std::make_unique<Nop>(), usage);
optimizationContext->fUpdated = true;
}
break;
}
case Statement::Kind::kIf: {
IfStatement& i = stmt->as<IfStatement>();
if (i.test()->kind() == Expression::Kind::kBoolLiteral) {
// constant if, collapse down to a single branch
if (i.test()->as<BoolLiteral>().value()) {
SkASSERT(i.ifTrue());
(*iter)->setStatement(std::move(i.ifTrue()), usage);
} else {
if (i.ifFalse()) {
(*iter)->setStatement(std::move(i.ifFalse()), usage);
} else {
(*iter)->setStatement(std::make_unique<Nop>(), usage);
}
}
optimizationContext->fUpdated = true;
optimizationContext->fNeedsRescan = true;
break;
}
if (i.ifFalse() && i.ifFalse()->isEmpty()) {
// else block doesn't do anything, remove it
i.ifFalse().reset();
optimizationContext->fUpdated = true;
optimizationContext->fNeedsRescan = true;
}
if (!i.ifFalse() && i.ifTrue()->isEmpty()) {
// if block doesn't do anything, no else block
if (i.test()->hasSideEffects()) {
// test has side effects, keep it
(*iter)->setStatement(
std::make_unique<ExpressionStatement>(std::move(i.test())), usage);
} else {
// no if, no else, no test side effects, kill the whole if
// statement
(*iter)->setStatement(std::make_unique<Nop>(), usage);
}
optimizationContext->fUpdated = true;
optimizationContext->fNeedsRescan = true;
}
break;
}
case Statement::Kind::kSwitch: {
SwitchStatement& s = stmt->as<SwitchStatement>();
int64_t switchValue;
if (fIRGenerator->getConstantInt(*s.value(), &switchValue)) {
// switch is constant, replace it with the case that matches
bool found = false;
SwitchCase* defaultCase = nullptr;
for (const std::unique_ptr<SwitchCase>& c : s.cases()) {
if (!c->value()) {
defaultCase = c.get();
continue;
}
int64_t caseValue;
SkAssertResult(fIRGenerator->getConstantInt(*c->value(), &caseValue));
if (caseValue == switchValue) {
std::unique_ptr<Statement> newBlock = block_for_case(&s, c.get());
if (newBlock) {
(*iter)->setStatement(std::move(newBlock), usage);
found = true;
break;
} else {
if (s.isStatic() && !(fFlags & kPermitInvalidStaticTests_Flag) &&
optimizationContext->fSilences.find(&s) ==
optimizationContext->fSilences.end()) {
this->error(s.fOffset,
"static switch contains non-static conditional break");
optimizationContext->fSilences.insert(&s);
}
return; // can't simplify
}
}
}
if (!found) {
// no matching case. use default if it exists, or kill the whole thing
if (defaultCase) {
std::unique_ptr<Statement> newBlock = block_for_case(&s, defaultCase);
if (newBlock) {
(*iter)->setStatement(std::move(newBlock), usage);
} else {
if (s.isStatic() && !(fFlags & kPermitInvalidStaticTests_Flag) &&
optimizationContext->fSilences.find(&s) ==
optimizationContext->fSilences.end()) {
this->error(s.fOffset,
"static switch contains non-static conditional break");
optimizationContext->fSilences.insert(&s);
}
return; // can't simplify
}
} else {
(*iter)->setStatement(std::make_unique<Nop>(), usage);
}
}
optimizationContext->fUpdated = true;
optimizationContext->fNeedsRescan = true;
}
break;
}
case Statement::Kind::kExpression: {
ExpressionStatement& e = stmt->as<ExpressionStatement>();
SkASSERT((*iter)->statement()->get() == &e);
if (!e.expression()->hasSideEffects()) {
// Expression statement with no side effects, kill it
if (!b.tryRemoveExpressionBefore(iter, e.expression().get())) {
optimizationContext->fNeedsRescan = true;
}
SkASSERT((*iter)->statement()->get() == stmt);
(*iter)->setStatement(std::make_unique<Nop>(), usage);
optimizationContext->fUpdated = true;
}
break;
}
default:
break;
}
}
bool Compiler::scanCFG(FunctionDefinition& f, ProgramUsage* usage) {
bool madeChanges = false;
CFG cfg = CFGGenerator().getCFG(f);
this->computeDataFlow(&cfg);
// check for unreachable code
for (size_t i = 0; i < cfg.fBlocks.size(); i++) {
const BasicBlock& block = cfg.fBlocks[i];
if (i != cfg.fStart && !block.fIsReachable && block.fNodes.size()) {
int offset;
const BasicBlock::Node& node = block.fNodes[0];
if (node.isStatement()) {
offset = (*node.statement())->fOffset;
} else {
offset = (*node.expression())->fOffset;
if ((*node.expression())->is<BoolLiteral>()) {
// Function inlining can generate do { ... } while(false) loops which always
// break, so the boolean condition is considered unreachable. Since not being
// able to reach a literal is a non-issue in the first place, we don't report an
// error in this case.
continue;
}
}
this->error(offset, String("unreachable"));
}
}
if (fErrorCount) {
return madeChanges;
}
// check for dead code & undefined variables, perform constant propagation
OptimizationContext optimizationContext;
optimizationContext.fUsage = usage;
SkBitSet eliminatedBlockIds(cfg.fBlocks.size());
do {
if (optimizationContext.fNeedsRescan) {
cfg = CFGGenerator().getCFG(f);
this->computeDataFlow(&cfg);
optimizationContext.fNeedsRescan = false;
}
eliminatedBlockIds.reset();
optimizationContext.fUpdated = false;
for (BlockId blockId = 0; blockId < cfg.fBlocks.size(); ++blockId) {
if (eliminatedBlockIds.test(blockId)) {
// We reached a block ID that might have been eliminated. Be cautious and rescan.
optimizationContext.fUpdated = true;
optimizationContext.fNeedsRescan = true;
break;
}
BasicBlock& b = cfg.fBlocks[blockId];
if (blockId > 0 && !b.fIsReachable) {
// Block was reachable before optimization, but has since become unreachable. In
// addition to being dead code, it's broken - since control flow can't reach it, no
// prior variable definitions can reach it, and therefore variables might look to
// have not been properly assigned. Kill it by replacing all statements with Nops.
for (BasicBlock::Node& node : b.fNodes) {
if (node.isStatement() && !(*node.statement())->is<Nop>()) {
// Eliminating a node runs the risk of eliminating that node's exits as
// well. Keep track of this and do a rescan if we are about to access one
// of these.
for (BlockId id : b.fExits) {
eliminatedBlockIds.set(id);
}
node.setStatement(std::make_unique<Nop>(), usage);
madeChanges = true;
}
}
continue;
}
DefinitionMap definitions = b.fBefore;
for (auto iter = b.fNodes.begin(); iter != b.fNodes.end() &&
!optimizationContext.fNeedsRescan; ++iter) {
if (iter->isExpression()) {
this->simplifyExpression(definitions, b, &iter, &optimizationContext);
} else {
this->simplifyStatement(definitions, b, &iter, &optimizationContext);
}
if (optimizationContext.fNeedsRescan) {
break;
}
this->addDefinitions(*iter, &definitions);
}
if (optimizationContext.fNeedsRescan) {
break;
}
}
madeChanges |= optimizationContext.fUpdated;
} while (optimizationContext.fUpdated);
SkASSERT(!optimizationContext.fNeedsRescan);
// verify static ifs & switches, clean up dead variable decls
for (BasicBlock& b : cfg.fBlocks) {
for (auto iter = b.fNodes.begin(); iter != b.fNodes.end() &&
!optimizationContext.fNeedsRescan;) {
if (iter->isStatement()) {
const Statement& s = **iter->statement();
switch (s.kind()) {
case Statement::Kind::kIf:
if (s.as<IfStatement>().isStatic() &&
!(fFlags & kPermitInvalidStaticTests_Flag)) {
this->error(s.fOffset, "static if has non-static test");
}
++iter;
break;
case Statement::Kind::kSwitch:
if (s.as<SwitchStatement>().isStatic() &&
!(fFlags & kPermitInvalidStaticTests_Flag) &&
optimizationContext.fSilences.find(&s) ==
optimizationContext.fSilences.end()) {
this->error(s.fOffset, "static switch has non-static test");
}
++iter;
break;
default:
++iter;
break;
}
} else {
++iter;
}
}
}
// check for missing return
if (f.declaration().returnType() != *fContext->fVoid_Type) {
if (cfg.fBlocks[cfg.fExit].fIsReachable) {
this->error(f.fOffset, String("function '" + String(f.declaration().name()) +
"' can exit without returning a value"));
}
}
return madeChanges;
}
std::unique_ptr<Program> Compiler::convertProgram(
Program::Kind kind,
String text,
const Program::Settings& settings,
const std::vector<std::unique_ptr<ExternalValue>>* externalValues) {
SkASSERT(!externalValues || (kind == Program::kGeneric_Kind));
// Loading and optimizing our base module might reset the inliner, so do that first,
// *then* configure the inliner with the settings for this program.
const ParsedModule& baseModule = this->moduleForProgramKind(kind);
fErrorText = "";
fErrorCount = 0;
fInliner.reset(fIRGenerator->fModifiers.get(), &settings);
// Not using AutoSource, because caller is likely to call errorText() if we fail to compile
std::unique_ptr<String> textPtr(new String(std::move(text)));
fSource = textPtr.get();
// Enable node pooling while converting and optimizing the program for a performance boost.
// The Program will take ownership of the pool.
std::unique_ptr<Pool> pool = Pool::Create();
pool->attachToThread();
IRGenerator::IRBundle ir =
fIRGenerator->convertProgram(kind, &settings, baseModule, /*isBuiltinCode=*/false,
textPtr->c_str(), textPtr->size(), externalValues);
auto program = std::make_unique<Program>(kind,
std::move(textPtr),
settings,
fCaps,
fContext,
std::move(ir.fElements),
std::move(ir.fSharedElements),
std::move(ir.fModifiers),
std::move(ir.fSymbolTable),
std::move(pool),
ir.fInputs);
bool success = false;
if (fErrorCount) {
// Do not return programs that failed to compile.
} else if (settings.fOptimize && !this->optimize(*program)) {
// Do not return programs that failed to optimize.
} else {
// We have a successful program!
success = true;
}
program->fPool->detachFromThread();
return success ? std::move(program) : nullptr;
}
bool Compiler::optimize(LoadedModule& module) {
SkASSERT(!fErrorCount);
const Program::Settings* oldSettings = fIRGenerator->fSettings;
Program::Settings settings;
fIRGenerator->fKind = module.fKind;
fIRGenerator->fSettings = &settings;
std::unique_ptr<ProgramUsage> usage = Analysis::GetUsage(module);
fInliner.reset(fModifiers.back().get(), &settings);
while (fErrorCount == 0) {
bool madeChanges = false;
// Scan and optimize based on the control-flow graph for each function.
for (const auto& element : module.fElements) {
if (element->is<FunctionDefinition>()) {
madeChanges |= this->scanCFG(element->as<FunctionDefinition>(), usage.get());
}
}
// Perform inline-candidate analysis and inline any functions deemed suitable.
madeChanges |= fInliner.analyze(module.fElements, module.fSymbols.get(), usage.get());
if (!madeChanges) {
break;
}
}
fIRGenerator->fSettings = oldSettings;
return fErrorCount == 0;
}
bool Compiler::optimize(Program& program) {
SkASSERT(!fErrorCount);
fIRGenerator->fKind = program.fKind;
fIRGenerator->fSettings = &program.fSettings;
ProgramUsage* usage = program.fUsage.get();
while (fErrorCount == 0) {
bool madeChanges = false;
// Scan and optimize based on the control-flow graph for each function.
for (const auto& element : program.ownedElements()) {
if (element->is<FunctionDefinition>()) {
madeChanges |= this->scanCFG(element->as<FunctionDefinition>(), usage);
}
}
// Perform inline-candidate analysis and inline any functions deemed suitable.
madeChanges |= fInliner.analyze(program.ownedElements(), program.fSymbols.get(), usage);
// Remove dead functions. We wait until after analysis so that we still report errors,
// even in unused code.
if (program.fSettings.fRemoveDeadFunctions) {
auto isDeadFunction = [&](const ProgramElement* element) {
if (!element->is<FunctionDefinition>()) {
return false;
}
const FunctionDefinition& fn = element->as<FunctionDefinition>();
if (fn.declaration().name() != "main" && usage->get(fn.declaration()) == 0) {
usage->remove(*element);
madeChanges = true;
return true;
}
return false;
};
program.fElements.erase(
std::remove_if(program.fElements.begin(), program.fElements.end(),
[&](const std::unique_ptr<ProgramElement>& element) {
return isDeadFunction(element.get());
}),
program.fElements.end());
program.fSharedElements.erase(
std::remove_if(program.fSharedElements.begin(), program.fSharedElements.end(),
isDeadFunction),
program.fSharedElements.end());
}
if (program.fKind != Program::kFragmentProcessor_Kind) {
// Remove declarations of dead global variables
auto isDeadVariable = [&](const ProgramElement* element) {
if (!element->is<GlobalVarDeclaration>()) {
return false;
}
const GlobalVarDeclaration& global = element->as<GlobalVarDeclaration>();
const VarDeclaration& varDecl = global.declaration()->as<VarDeclaration>();
if (usage->isDead(varDecl.var())) {
madeChanges = true;
return true;
}
return false;
};
program.fElements.erase(
std::remove_if(program.fElements.begin(), program.fElements.end(),
[&](const std::unique_ptr<ProgramElement>& element) {
return isDeadVariable(element.get());
}),
program.fElements.end());
program.fSharedElements.erase(
std::remove_if(program.fSharedElements.begin(), program.fSharedElements.end(),
isDeadVariable),
program.fSharedElements.end());
}
if (!madeChanges) {
break;
}
}
return fErrorCount == 0;
}
#if defined(SKSL_STANDALONE) || SK_SUPPORT_GPU
bool Compiler::toSPIRV(Program& program, OutputStream& out) {
#ifdef SK_ENABLE_SPIRV_VALIDATION
StringStream buffer;
AutoSource as(this, program.fSource.get());
SPIRVCodeGenerator cg(fContext.get(), &program, this, &buffer);
bool result = cg.generateCode();
if (result && program.fSettings.fValidateSPIRV) {
spvtools::SpirvTools tools(SPV_ENV_VULKAN_1_0);
const String& data = buffer.str();
SkASSERT(0 == data.size() % 4);
String errors;
auto dumpmsg = [&errors](spv_message_level_t, const char*, const spv_position_t&,
const char* m) {
errors.appendf("SPIR-V validation error: %s\n", m);
};
tools.SetMessageConsumer(dumpmsg);
// Verify that the SPIR-V we produced is valid. At runtime, we will abort() with a message
// explaining the error. In standalone mode (skslc), we will send the message, plus the
// entire disassembled SPIR-V (for easier context & debugging) as *our* error message.
result = tools.Validate((const uint32_t*) data.c_str(), data.size() / 4);
if (!result) {
#if defined(SKSL_STANDALONE)
// Convert the string-stream to a SPIR-V disassembly.
std::string disassembly;
if (tools.Disassemble((const uint32_t*)data.data(), data.size() / 4, &disassembly)) {
errors.append(disassembly);
}
this->error(-1, errors);
#else
SkDEBUGFAILF("%s", errors.c_str());
#endif
}
out.write(data.c_str(), data.size());
}
#else
AutoSource as(this, program.fSource.get());
SPIRVCodeGenerator cg(fContext.get(), &program, this, &out);
bool result = cg.generateCode();
#endif
return result;
}
bool Compiler::toSPIRV(Program& program, String* out) {
StringStream buffer;
bool result = this->toSPIRV(program, buffer);
if (result) {
*out = buffer.str();
}
return result;
}
bool Compiler::toGLSL(Program& program, OutputStream& out) {
AutoSource as(this, program.fSource.get());
GLSLCodeGenerator cg(fContext.get(), &program, this, &out);
bool result = cg.generateCode();
return result;
}
bool Compiler::toGLSL(Program& program, String* out) {
StringStream buffer;
bool result = this->toGLSL(program, buffer);
if (result) {
*out = buffer.str();
}
return result;
}
bool Compiler::toHLSL(Program& program, String* out) {
String spirv;
if (!this->toSPIRV(program, &spirv)) {
return false;
}
return SPIRVtoHLSL(spirv, out);
}
bool Compiler::toMetal(Program& program, OutputStream& out) {
MetalCodeGenerator cg(fContext.get(), &program, this, &out);
bool result = cg.generateCode();
return result;
}
bool Compiler::toMetal(Program& program, String* out) {
StringStream buffer;
bool result = this->toMetal(program, buffer);
if (result) {
*out = buffer.str();
}
return result;
}
#if defined(SKSL_STANDALONE) || GR_TEST_UTILS
bool Compiler::toCPP(Program& program, String name, OutputStream& out) {
AutoSource as(this, program.fSource.get());
CPPCodeGenerator cg(fContext.get(), &program, this, name, &out);
bool result = cg.generateCode();
return result;
}
bool Compiler::toH(Program& program, String name, OutputStream& out) {
AutoSource as(this, program.fSource.get());
HCodeGenerator cg(fContext.get(), &program, this, name, &out);
bool result = cg.generateCode();
return result;
}
#endif // defined(SKSL_STANDALONE) || GR_TEST_UTILS
#endif // defined(SKSL_STANDALONE) || SK_SUPPORT_GPU
#if !defined(SKSL_STANDALONE) && SK_SUPPORT_GPU
bool Compiler::toPipelineStage(Program& program, PipelineStageArgs* outArgs) {
AutoSource as(this, program.fSource.get());
StringStream buffer;
PipelineStageCodeGenerator cg(fContext.get(), &program, this, &buffer, outArgs);
bool result = cg.generateCode();
if (result) {
outArgs->fCode = buffer.str();
}
return result;
}
#endif
std::unique_ptr<ByteCode> Compiler::toByteCode(Program& program) {
AutoSource as(this, program.fSource.get());
std::unique_ptr<ByteCode> result(new ByteCode());
ByteCodeGenerator cg(fContext.get(), &program, this, result.get());
bool success = cg.generateCode();
if (success) {
return result;
}
return nullptr;
}
const char* Compiler::OperatorName(Token::Kind op) {
switch (op) {
case Token::Kind::TK_PLUS: return "+";
case Token::Kind::TK_MINUS: return "-";
case Token::Kind::TK_STAR: return "*";
case Token::Kind::TK_SLASH: return "/";
case Token::Kind::TK_PERCENT: return "%";
case Token::Kind::TK_SHL: return "<<";
case Token::Kind::TK_SHR: return ">>";
case Token::Kind::TK_LOGICALNOT: return "!";
case Token::Kind::TK_LOGICALAND: return "&&";
case Token::Kind::TK_LOGICALOR: return "||";
case Token::Kind::TK_LOGICALXOR: return "^^";
case Token::Kind::TK_BITWISENOT: return "~";
case Token::Kind::TK_BITWISEAND: return "&";
case Token::Kind::TK_BITWISEOR: return "|";
case Token::Kind::TK_BITWISEXOR: return "^";
case Token::Kind::TK_EQ: return "=";
case Token::Kind::TK_EQEQ: return "==";
case Token::Kind::TK_NEQ: return "!=";
case Token::Kind::TK_LT: return "<";
case Token::Kind::TK_GT: return ">";
case Token::Kind::TK_LTEQ: return "<=";
case Token::Kind::TK_GTEQ: return ">=";
case Token::Kind::TK_PLUSEQ: return "+=";
case Token::Kind::TK_MINUSEQ: return "-=";
case Token::Kind::TK_STAREQ: return "*=";
case Token::Kind::TK_SLASHEQ: return "/=";
case Token::Kind::TK_PERCENTEQ: return "%=";
case Token::Kind::TK_SHLEQ: return "<<=";
case Token::Kind::TK_SHREQ: return ">>=";
case Token::Kind::TK_BITWISEANDEQ: return "&=";
case Token::Kind::TK_BITWISEOREQ: return "|=";
case Token::Kind::TK_BITWISEXOREQ: return "^=";
case Token::Kind::TK_PLUSPLUS: return "++";
case Token::Kind::TK_MINUSMINUS: return "--";
case Token::Kind::TK_COMMA: return ",";
default:
ABORT("unsupported operator: %d\n", (int) op);
}
}
bool Compiler::IsAssignment(Token::Kind op) {
switch (op) {
case Token::Kind::TK_EQ: // fall through
case Token::Kind::TK_PLUSEQ: // fall through
case Token::Kind::TK_MINUSEQ: // fall through
case Token::Kind::TK_STAREQ: // fall through
case Token::Kind::TK_SLASHEQ: // fall through
case Token::Kind::TK_PERCENTEQ: // fall through
case Token::Kind::TK_SHLEQ: // fall through
case Token::Kind::TK_SHREQ: // fall through
case Token::Kind::TK_BITWISEOREQ: // fall through
case Token::Kind::TK_BITWISEXOREQ: // fall through
case Token::Kind::TK_BITWISEANDEQ:
return true;
default:
return false;
}
}
Token::Kind Compiler::RemoveAssignment(Token::Kind op) {
switch (op) {
case Token::Kind::TK_PLUSEQ: return Token::Kind::TK_PLUS;
case Token::Kind::TK_MINUSEQ: return Token::Kind::TK_MINUS;
case Token::Kind::TK_STAREQ: return Token::Kind::TK_STAR;
case Token::Kind::TK_SLASHEQ: return Token::Kind::TK_SLASH;
case Token::Kind::TK_PERCENTEQ: return Token::Kind::TK_PERCENT;
case Token::Kind::TK_SHLEQ: return Token::Kind::TK_SHL;
case Token::Kind::TK_SHREQ: return Token::Kind::TK_SHR;
case Token::Kind::TK_BITWISEOREQ: return Token::Kind::TK_BITWISEOR;
case Token::Kind::TK_BITWISEXOREQ: return Token::Kind::TK_BITWISEXOR;
case Token::Kind::TK_BITWISEANDEQ: return Token::Kind::TK_BITWISEAND;
default: return op;
}
}
Position Compiler::position(int offset) {
if (fSource && offset >= 0) {
int line = 1;
int column = 1;
for (int i = 0; i < offset; i++) {
if ((*fSource)[i] == '\n') {
++line;
column = 1;
}
else {
++column;
}
}
return Position(line, column);
} else {
return Position(-1, -1);
}
}
void Compiler::error(int offset, String msg) {
fErrorCount++;
Position pos = this->position(offset);
fErrorText += "error: " + (pos.fLine >= 1 ? to_string(pos.fLine) + ": " : "") + msg + "\n";
}
String Compiler::errorText(bool showCount) {
if (showCount) {
this->writeErrorCount();
}
fErrorCount = 0;
String result = fErrorText;
fErrorText = "";
return result;
}
void Compiler::writeErrorCount() {
if (fErrorCount) {
fErrorText += to_string(fErrorCount) + " error";
if (fErrorCount > 1) {
fErrorText += "s";
}
fErrorText += "\n";
}
}
} // namespace SkSL