| /* |
| * Copyright (c) 1999, 2017, Oracle and/or its affiliates. All rights reserved. |
| * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
| * |
| * This code is free software; you can redistribute it and/or modify it |
| * under the terms of the GNU General Public License version 2 only, as |
| * published by the Free Software Foundation. |
| * |
| * This code is distributed in the hope that it will be useful, but WITHOUT |
| * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
| * version 2 for more details (a copy is included in the LICENSE file that |
| * accompanied this code). |
| * |
| * You should have received a copy of the GNU General Public License version |
| * 2 along with this work; if not, write to the Free Software Foundation, |
| * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
| * |
| * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
| * or visit www.oracle.com if you need additional information or have any |
| * questions. |
| * |
| */ |
| |
| #include "precompiled.hpp" |
| #include "c1/c1_Compilation.hpp" |
| #include "c1/c1_Compiler.hpp" |
| #include "c1/c1_FrameMap.hpp" |
| #include "c1/c1_GraphBuilder.hpp" |
| #include "c1/c1_LinearScan.hpp" |
| #include "c1/c1_MacroAssembler.hpp" |
| #include "c1/c1_Runtime1.hpp" |
| #include "c1/c1_ValueType.hpp" |
| #include "compiler/compileBroker.hpp" |
| #include "interpreter/linkResolver.hpp" |
| #include "memory/allocation.hpp" |
| #include "memory/allocation.inline.hpp" |
| #include "memory/resourceArea.hpp" |
| #include "prims/nativeLookup.hpp" |
| #include "runtime/arguments.hpp" |
| #include "runtime/interfaceSupport.hpp" |
| #include "runtime/sharedRuntime.hpp" |
| #include "utilities/bitMap.inline.hpp" |
| |
| |
| Compiler::Compiler() : AbstractCompiler(compiler_c1) { |
| } |
| |
| void Compiler::init_c1_runtime() { |
| BufferBlob* buffer_blob = CompilerThread::current()->get_buffer_blob(); |
| Arena* arena = new (mtCompiler) Arena(mtCompiler); |
| Runtime1::initialize(buffer_blob); |
| FrameMap::initialize(); |
| // initialize data structures |
| ValueType::initialize(arena); |
| GraphBuilder::initialize(); |
| // note: to use more than one instance of LinearScan at a time this function call has to |
| // be moved somewhere outside of this constructor: |
| Interval::initialize(arena); |
| } |
| |
| |
| void Compiler::initialize() { |
| // Buffer blob must be allocated per C1 compiler thread at startup |
| BufferBlob* buffer_blob = init_buffer_blob(); |
| |
| if (should_perform_init()) { |
| if (buffer_blob == NULL) { |
| // When we come here we are in state 'initializing'; entire C1 compilation |
| // can be shut down. |
| set_state(failed); |
| } else { |
| init_c1_runtime(); |
| set_state(initialized); |
| } |
| } |
| } |
| |
| int Compiler::code_buffer_size() { |
| assert(SegmentedCodeCache, "Should be only used with a segmented code cache"); |
| return Compilation::desired_max_code_buffer_size() + Compilation::desired_max_constant_size(); |
| } |
| |
| BufferBlob* Compiler::init_buffer_blob() { |
| // Allocate buffer blob once at startup since allocation for each |
| // compilation seems to be too expensive (at least on Intel win32). |
| assert (CompilerThread::current()->get_buffer_blob() == NULL, "Should initialize only once"); |
| |
| // setup CodeBuffer. Preallocate a BufferBlob of size |
| // NMethodSizeLimit plus some extra space for constants. |
| int code_buffer_size = Compilation::desired_max_code_buffer_size() + |
| Compilation::desired_max_constant_size(); |
| |
| BufferBlob* buffer_blob = BufferBlob::create("C1 temporary CodeBuffer", code_buffer_size); |
| if (buffer_blob != NULL) { |
| CompilerThread::current()->set_buffer_blob(buffer_blob); |
| } |
| |
| return buffer_blob; |
| } |
| |
| bool Compiler::is_intrinsic_supported(const methodHandle& method) { |
| vmIntrinsics::ID id = method->intrinsic_id(); |
| assert(id != vmIntrinsics::_none, "must be a VM intrinsic"); |
| |
| if (method->is_synchronized()) { |
| // C1 does not support intrinsification of synchronized methods. |
| return false; |
| } |
| |
| switch (id) { |
| case vmIntrinsics::_compareAndSetLong: |
| if (!VM_Version::supports_cx8()) return false; |
| break; |
| case vmIntrinsics::_getAndAddInt: |
| if (!VM_Version::supports_atomic_getadd4()) return false; |
| break; |
| case vmIntrinsics::_getAndAddLong: |
| if (!VM_Version::supports_atomic_getadd8()) return false; |
| break; |
| case vmIntrinsics::_getAndSetInt: |
| if (!VM_Version::supports_atomic_getset4()) return false; |
| break; |
| case vmIntrinsics::_getAndSetLong: |
| if (!VM_Version::supports_atomic_getset8()) return false; |
| break; |
| case vmIntrinsics::_getAndSetObject: |
| #ifdef _LP64 |
| if (!UseCompressedOops && !VM_Version::supports_atomic_getset8()) return false; |
| if (UseCompressedOops && !VM_Version::supports_atomic_getset4()) return false; |
| #else |
| if (!VM_Version::supports_atomic_getset4()) return false; |
| #endif |
| break; |
| case vmIntrinsics::_onSpinWait: |
| if (!VM_Version::supports_on_spin_wait()) return false; |
| break; |
| case vmIntrinsics::_arraycopy: |
| case vmIntrinsics::_currentTimeMillis: |
| case vmIntrinsics::_nanoTime: |
| case vmIntrinsics::_Reference_get: |
| // Use the intrinsic version of Reference.get() so that the value in |
| // the referent field can be registered by the G1 pre-barrier code. |
| // Also to prevent commoning reads from this field across safepoint |
| // since GC can change its value. |
| case vmIntrinsics::_loadFence: |
| case vmIntrinsics::_storeFence: |
| case vmIntrinsics::_fullFence: |
| case vmIntrinsics::_floatToRawIntBits: |
| case vmIntrinsics::_intBitsToFloat: |
| case vmIntrinsics::_doubleToRawLongBits: |
| case vmIntrinsics::_longBitsToDouble: |
| case vmIntrinsics::_getClass: |
| case vmIntrinsics::_isInstance: |
| case vmIntrinsics::_isPrimitive: |
| case vmIntrinsics::_currentThread: |
| case vmIntrinsics::_dabs: |
| case vmIntrinsics::_dsqrt: |
| case vmIntrinsics::_dsin: |
| case vmIntrinsics::_dcos: |
| case vmIntrinsics::_dtan: |
| case vmIntrinsics::_dlog: |
| case vmIntrinsics::_dlog10: |
| case vmIntrinsics::_dexp: |
| case vmIntrinsics::_dpow: |
| case vmIntrinsics::_fmaD: |
| case vmIntrinsics::_fmaF: |
| case vmIntrinsics::_getObject: |
| case vmIntrinsics::_getBoolean: |
| case vmIntrinsics::_getByte: |
| case vmIntrinsics::_getShort: |
| case vmIntrinsics::_getChar: |
| case vmIntrinsics::_getInt: |
| case vmIntrinsics::_getLong: |
| case vmIntrinsics::_getFloat: |
| case vmIntrinsics::_getDouble: |
| case vmIntrinsics::_putObject: |
| case vmIntrinsics::_putBoolean: |
| case vmIntrinsics::_putByte: |
| case vmIntrinsics::_putShort: |
| case vmIntrinsics::_putChar: |
| case vmIntrinsics::_putInt: |
| case vmIntrinsics::_putLong: |
| case vmIntrinsics::_putFloat: |
| case vmIntrinsics::_putDouble: |
| case vmIntrinsics::_getObjectVolatile: |
| case vmIntrinsics::_getBooleanVolatile: |
| case vmIntrinsics::_getByteVolatile: |
| case vmIntrinsics::_getShortVolatile: |
| case vmIntrinsics::_getCharVolatile: |
| case vmIntrinsics::_getIntVolatile: |
| case vmIntrinsics::_getLongVolatile: |
| case vmIntrinsics::_getFloatVolatile: |
| case vmIntrinsics::_getDoubleVolatile: |
| case vmIntrinsics::_putObjectVolatile: |
| case vmIntrinsics::_putBooleanVolatile: |
| case vmIntrinsics::_putByteVolatile: |
| case vmIntrinsics::_putShortVolatile: |
| case vmIntrinsics::_putCharVolatile: |
| case vmIntrinsics::_putIntVolatile: |
| case vmIntrinsics::_putLongVolatile: |
| case vmIntrinsics::_putFloatVolatile: |
| case vmIntrinsics::_putDoubleVolatile: |
| case vmIntrinsics::_getShortUnaligned: |
| case vmIntrinsics::_getCharUnaligned: |
| case vmIntrinsics::_getIntUnaligned: |
| case vmIntrinsics::_getLongUnaligned: |
| case vmIntrinsics::_putShortUnaligned: |
| case vmIntrinsics::_putCharUnaligned: |
| case vmIntrinsics::_putIntUnaligned: |
| case vmIntrinsics::_putLongUnaligned: |
| case vmIntrinsics::_checkIndex: |
| case vmIntrinsics::_updateCRC32: |
| case vmIntrinsics::_updateBytesCRC32: |
| case vmIntrinsics::_updateByteBufferCRC32: |
| #if defined(SPARC) || defined(S390) || defined(PPC64) |
| case vmIntrinsics::_updateBytesCRC32C: |
| case vmIntrinsics::_updateDirectByteBufferCRC32C: |
| #endif |
| case vmIntrinsics::_vectorizedMismatch: |
| case vmIntrinsics::_compareAndSetInt: |
| case vmIntrinsics::_compareAndSetObject: |
| case vmIntrinsics::_getCharStringU: |
| case vmIntrinsics::_putCharStringU: |
| #ifdef TRACE_HAVE_INTRINSICS |
| case vmIntrinsics::_counterTime: |
| case vmIntrinsics::_getBufferWriter: |
| #if defined(_LP64) || !defined(TRACE_ID_CLASS_SHIFT) |
| case vmIntrinsics::_getClassId: |
| #endif |
| #endif |
| break; |
| default: |
| return false; // Intrinsics not on the previous list are not available. |
| } |
| |
| return true; |
| } |
| |
| void Compiler::compile_method(ciEnv* env, ciMethod* method, int entry_bci, DirectiveSet* directive) { |
| BufferBlob* buffer_blob = CompilerThread::current()->get_buffer_blob(); |
| assert(buffer_blob != NULL, "Must exist"); |
| // invoke compilation |
| { |
| // We are nested here because we need for the destructor |
| // of Compilation to occur before we release the any |
| // competing compiler thread |
| ResourceMark rm; |
| Compilation c(this, env, method, entry_bci, buffer_blob, directive); |
| } |
| } |
| |
| |
| void Compiler::print_timers() { |
| Compilation::print_timers(); |
| } |