| /* |
| * Copyright (c) 1997, 2017, Oracle and/or its affiliates. All rights reserved. |
| * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
| * |
| * This code is free software; you can redistribute it and/or modify it |
| * under the terms of the GNU General Public License version 2 only, as |
| * published by the Free Software Foundation. |
| * |
| * This code is distributed in the hope that it will be useful, but WITHOUT |
| * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
| * version 2 for more details (a copy is included in the LICENSE file that |
| * accompanied this code). |
| * |
| * You should have received a copy of the GNU General Public License version |
| * 2 along with this work; if not, write to the Free Software Foundation, |
| * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
| * |
| * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
| * or visit www.oracle.com if you need additional information or have any |
| * questions. |
| * |
| */ |
| |
| #ifndef SHARE_VM_RUNTIME_GLOBALS_HPP |
| #define SHARE_VM_RUNTIME_GLOBALS_HPP |
| |
| #include "utilities/align.hpp" |
| #include "utilities/globalDefinitions.hpp" |
| #include "utilities/macros.hpp" |
| |
| #include <float.h> // for DBL_MAX |
| |
| // The larger HeapWordSize for 64bit requires larger heaps |
| // for the same application running in 64bit. See bug 4967770. |
| // The minimum alignment to a heap word size is done. Other |
| // parts of the memory system may require additional alignment |
| // and are responsible for those alignments. |
| #ifdef _LP64 |
| #define ScaleForWordSize(x) align_down_((x) * 13 / 10, HeapWordSize) |
| #else |
| #define ScaleForWordSize(x) (x) |
| #endif |
| |
| // use this for flags that are true per default in the tiered build |
| // but false in non-tiered builds, and vice versa |
| #ifdef TIERED |
| #define trueInTiered true |
| #define falseInTiered false |
| #else |
| #define trueInTiered false |
| #define falseInTiered true |
| #endif |
| |
| #include CPU_HEADER(globals) |
| #include OS_HEADER(globals) |
| #include OS_CPU_HEADER(globals) |
| #ifdef COMPILER1 |
| #include CPU_HEADER(c1_globals) |
| #include OS_HEADER(c1_globals) |
| #endif |
| #ifdef COMPILER2 |
| #include CPU_HEADER(c2_globals) |
| #include OS_HEADER(c2_globals) |
| #endif |
| #ifdef SHARK |
| #ifdef ZERO |
| # include "shark_globals_zero.hpp" |
| #endif |
| #endif |
| |
| #if !defined(COMPILER1) && !defined(COMPILER2) && !defined(SHARK) && !INCLUDE_JVMCI |
| define_pd_global(bool, BackgroundCompilation, false); |
| define_pd_global(bool, UseTLAB, false); |
| define_pd_global(bool, CICompileOSR, false); |
| define_pd_global(bool, UseTypeProfile, false); |
| define_pd_global(bool, UseOnStackReplacement, false); |
| define_pd_global(bool, InlineIntrinsics, false); |
| define_pd_global(bool, PreferInterpreterNativeStubs, true); |
| define_pd_global(bool, ProfileInterpreter, false); |
| define_pd_global(bool, ProfileTraps, false); |
| define_pd_global(bool, TieredCompilation, false); |
| |
| define_pd_global(intx, CompileThreshold, 0); |
| |
| define_pd_global(intx, OnStackReplacePercentage, 0); |
| define_pd_global(bool, ResizeTLAB, false); |
| define_pd_global(intx, FreqInlineSize, 0); |
| define_pd_global(size_t, NewSizeThreadIncrease, 4*K); |
| define_pd_global(intx, InlineClassNatives, true); |
| define_pd_global(intx, InlineUnsafeOps, true); |
| define_pd_global(intx, InitialCodeCacheSize, 160*K); |
| define_pd_global(intx, ReservedCodeCacheSize, 32*M); |
| define_pd_global(intx, NonProfiledCodeHeapSize, 0); |
| define_pd_global(intx, ProfiledCodeHeapSize, 0); |
| define_pd_global(intx, NonNMethodCodeHeapSize, 32*M); |
| |
| define_pd_global(intx, CodeCacheExpansionSize, 32*K); |
| define_pd_global(intx, CodeCacheMinBlockLength, 1); |
| define_pd_global(intx, CodeCacheMinimumUseSpace, 200*K); |
| define_pd_global(size_t, MetaspaceSize, ScaleForWordSize(4*M)); |
| define_pd_global(bool, NeverActAsServerClassMachine, true); |
| define_pd_global(uint64_t,MaxRAM, 1ULL*G); |
| #define CI_COMPILER_COUNT 0 |
| #else |
| |
| #if defined(COMPILER2) || INCLUDE_JVMCI |
| #define CI_COMPILER_COUNT 2 |
| #else |
| #define CI_COMPILER_COUNT 1 |
| #endif // COMPILER2 || INCLUDE_JVMCI |
| |
| #endif // no compilers |
| |
| // string type aliases used only in this file |
| typedef const char* ccstr; |
| typedef const char* ccstrlist; // represents string arguments which accumulate |
| |
| // function type that will construct default range string |
| typedef const char* (*RangeStrFunc)(void); |
| |
| struct Flag { |
| enum Flags { |
| // latest value origin |
| DEFAULT = 0, |
| COMMAND_LINE = 1, |
| ENVIRON_VAR = 2, |
| CONFIG_FILE = 3, |
| MANAGEMENT = 4, |
| ERGONOMIC = 5, |
| ATTACH_ON_DEMAND = 6, |
| INTERNAL = 7, |
| |
| LAST_VALUE_ORIGIN = INTERNAL, |
| VALUE_ORIGIN_BITS = 4, |
| VALUE_ORIGIN_MASK = right_n_bits(VALUE_ORIGIN_BITS), |
| |
| // flag kind |
| KIND_PRODUCT = 1 << 4, |
| KIND_MANAGEABLE = 1 << 5, |
| KIND_DIAGNOSTIC = 1 << 6, |
| KIND_EXPERIMENTAL = 1 << 7, |
| KIND_NOT_PRODUCT = 1 << 8, |
| KIND_DEVELOP = 1 << 9, |
| KIND_PLATFORM_DEPENDENT = 1 << 10, |
| KIND_READ_WRITE = 1 << 11, |
| KIND_C1 = 1 << 12, |
| KIND_C2 = 1 << 13, |
| KIND_ARCH = 1 << 14, |
| KIND_SHARK = 1 << 15, |
| KIND_LP64_PRODUCT = 1 << 16, |
| KIND_COMMERCIAL = 1 << 17, |
| KIND_JVMCI = 1 << 18, |
| |
| // set this bit if the flag was set on the command line |
| ORIG_COMMAND_LINE = 1 << 19, |
| |
| KIND_MASK = ~(VALUE_ORIGIN_MASK | ORIG_COMMAND_LINE) |
| }; |
| |
| enum Error { |
| // no error |
| SUCCESS = 0, |
| // flag name is missing |
| MISSING_NAME, |
| // flag value is missing |
| MISSING_VALUE, |
| // error parsing the textual form of the value |
| WRONG_FORMAT, |
| // flag is not writable |
| NON_WRITABLE, |
| // flag value is outside of its bounds |
| OUT_OF_BOUNDS, |
| // flag value violates its constraint |
| VIOLATES_CONSTRAINT, |
| // there is no flag with the given name |
| INVALID_FLAG, |
| // the flag can only be set only on command line during invocation of the VM |
| COMMAND_LINE_ONLY, |
| // the flag may only be set once |
| SET_ONLY_ONCE, |
| // the flag is not writable in this combination of product/debug build |
| CONSTANT, |
| // other, unspecified error related to setting the flag |
| ERR_OTHER |
| }; |
| |
| enum MsgType { |
| NONE = 0, |
| DIAGNOSTIC_FLAG_BUT_LOCKED, |
| EXPERIMENTAL_FLAG_BUT_LOCKED, |
| DEVELOPER_FLAG_BUT_PRODUCT_BUILD, |
| NOTPRODUCT_FLAG_BUT_PRODUCT_BUILD |
| }; |
| |
| const char* _type; |
| const char* _name; |
| void* _addr; |
| NOT_PRODUCT(const char* _doc;) |
| Flags _flags; |
| size_t _name_len; |
| |
| // points to all Flags static array |
| static Flag* flags; |
| |
| // number of flags |
| static size_t numFlags; |
| |
| static Flag* find_flag(const char* name) { return find_flag(name, strlen(name), true, true); }; |
| static Flag* find_flag(const char* name, size_t length, bool allow_locked = false, bool return_flag = false); |
| static Flag* fuzzy_match(const char* name, size_t length, bool allow_locked = false); |
| |
| static const char* get_int_default_range_str(); |
| static const char* get_uint_default_range_str(); |
| static const char* get_intx_default_range_str(); |
| static const char* get_uintx_default_range_str(); |
| static const char* get_uint64_t_default_range_str(); |
| static const char* get_size_t_default_range_str(); |
| static const char* get_double_default_range_str(); |
| |
| Flag::Error check_writable(bool changed); |
| |
| bool is_bool() const; |
| bool get_bool() const; |
| Flag::Error set_bool(bool value); |
| |
| bool is_int() const; |
| int get_int() const; |
| Flag::Error set_int(int value); |
| |
| bool is_uint() const; |
| uint get_uint() const; |
| Flag::Error set_uint(uint value); |
| |
| bool is_intx() const; |
| intx get_intx() const; |
| Flag::Error set_intx(intx value); |
| |
| bool is_uintx() const; |
| uintx get_uintx() const; |
| Flag::Error set_uintx(uintx value); |
| |
| bool is_uint64_t() const; |
| uint64_t get_uint64_t() const; |
| Flag::Error set_uint64_t(uint64_t value); |
| |
| bool is_size_t() const; |
| size_t get_size_t() const; |
| Flag::Error set_size_t(size_t value); |
| |
| bool is_double() const; |
| double get_double() const; |
| Flag::Error set_double(double value); |
| |
| bool is_ccstr() const; |
| bool ccstr_accumulates() const; |
| ccstr get_ccstr() const; |
| Flag::Error set_ccstr(ccstr value); |
| |
| Flags get_origin(); |
| void set_origin(Flags origin); |
| |
| size_t get_name_length(); |
| |
| bool is_default(); |
| bool is_ergonomic(); |
| bool is_command_line(); |
| void set_command_line(); |
| |
| bool is_product() const; |
| bool is_manageable() const; |
| bool is_diagnostic() const; |
| bool is_experimental() const; |
| bool is_notproduct() const; |
| bool is_develop() const; |
| bool is_read_write() const; |
| bool is_commercial() const; |
| |
| bool is_constant_in_binary() const; |
| |
| bool is_unlocker() const; |
| bool is_unlocked() const; |
| bool is_writeable() const; |
| bool is_external() const; |
| |
| bool is_unlocker_ext() const; |
| bool is_unlocked_ext() const; |
| bool is_writeable_ext() const; |
| bool is_external_ext() const; |
| |
| void clear_diagnostic(); |
| |
| Flag::MsgType get_locked_message(char*, int) const; |
| void get_locked_message_ext(char*, int) const; |
| |
| // printRanges will print out flags type, name and range values as expected by -XX:+PrintFlagsRanges |
| void print_on(outputStream* st, bool withComments = false, bool printRanges = false); |
| void print_kind_and_origin(outputStream* st); |
| void print_as_flag(outputStream* st); |
| |
| static const char* flag_error_str(Flag::Error error); |
| }; |
| |
| // debug flags control various aspects of the VM and are global accessible |
| |
| // use FlagSetting to temporarily change some debug flag |
| // e.g. FlagSetting fs(DebugThisAndThat, true); |
| // restored to previous value upon leaving scope |
| class FlagSetting { |
| bool val; |
| bool* flag; |
| public: |
| FlagSetting(bool& fl, bool newValue) { flag = &fl; val = fl; fl = newValue; } |
| ~FlagSetting() { *flag = val; } |
| }; |
| |
| |
| class CounterSetting { |
| intx* counter; |
| public: |
| CounterSetting(intx* cnt) { counter = cnt; (*counter)++; } |
| ~CounterSetting() { (*counter)--; } |
| }; |
| |
| class IntFlagSetting { |
| int val; |
| int* flag; |
| public: |
| IntFlagSetting(int& fl, int newValue) { flag = &fl; val = fl; fl = newValue; } |
| ~IntFlagSetting() { *flag = val; } |
| }; |
| |
| class UIntFlagSetting { |
| uint val; |
| uint* flag; |
| public: |
| UIntFlagSetting(uint& fl, uint newValue) { flag = &fl; val = fl; fl = newValue; } |
| ~UIntFlagSetting() { *flag = val; } |
| }; |
| |
| class UIntXFlagSetting { |
| uintx val; |
| uintx* flag; |
| public: |
| UIntXFlagSetting(uintx& fl, uintx newValue) { flag = &fl; val = fl; fl = newValue; } |
| ~UIntXFlagSetting() { *flag = val; } |
| }; |
| |
| class DoubleFlagSetting { |
| double val; |
| double* flag; |
| public: |
| DoubleFlagSetting(double& fl, double newValue) { flag = &fl; val = fl; fl = newValue; } |
| ~DoubleFlagSetting() { *flag = val; } |
| }; |
| |
| class SizeTFlagSetting { |
| size_t val; |
| size_t* flag; |
| public: |
| SizeTFlagSetting(size_t& fl, size_t newValue) { flag = &fl; val = fl; fl = newValue; } |
| ~SizeTFlagSetting() { *flag = val; } |
| }; |
| |
| // Helper class for temporarily saving the value of a flag during a scope. |
| template <size_t SIZE> |
| class FlagGuard { |
| unsigned char _value[SIZE]; |
| void* const _addr; |
| |
| // Hide operator new, this class should only be allocated on the stack. |
| // NOTE: Cannot include memory/allocation.hpp here due to circular |
| // dependencies. |
| void* operator new(size_t size) throw(); |
| void* operator new [](size_t size) throw(); |
| |
| public: |
| FlagGuard(void* flag_addr) : _addr(flag_addr) { |
| memcpy(_value, _addr, SIZE); |
| } |
| |
| ~FlagGuard() { |
| memcpy(_addr, _value, SIZE); |
| } |
| }; |
| |
| #define FLAG_GUARD(f) FlagGuard<sizeof(f)> f ## _guard(&f) |
| |
| class CommandLineFlags { |
| public: |
| static Flag::Error boolAt(const char* name, size_t len, bool* value, bool allow_locked = false, bool return_flag = false); |
| static Flag::Error boolAt(const char* name, bool* value, bool allow_locked = false, bool return_flag = false) { return boolAt(name, strlen(name), value, allow_locked, return_flag); } |
| static Flag::Error boolAtPut(Flag* flag, bool* value, Flag::Flags origin); |
| static Flag::Error boolAtPut(const char* name, size_t len, bool* value, Flag::Flags origin); |
| static Flag::Error boolAtPut(const char* name, bool* value, Flag::Flags origin) { return boolAtPut(name, strlen(name), value, origin); } |
| |
| static Flag::Error intAt(const char* name, size_t len, int* value, bool allow_locked = false, bool return_flag = false); |
| static Flag::Error intAt(const char* name, int* value, bool allow_locked = false, bool return_flag = false) { return intAt(name, strlen(name), value, allow_locked, return_flag); } |
| static Flag::Error intAtPut(Flag* flag, int* value, Flag::Flags origin); |
| static Flag::Error intAtPut(const char* name, size_t len, int* value, Flag::Flags origin); |
| static Flag::Error intAtPut(const char* name, int* value, Flag::Flags origin) { return intAtPut(name, strlen(name), value, origin); } |
| |
| static Flag::Error uintAt(const char* name, size_t len, uint* value, bool allow_locked = false, bool return_flag = false); |
| static Flag::Error uintAt(const char* name, uint* value, bool allow_locked = false, bool return_flag = false) { return uintAt(name, strlen(name), value, allow_locked, return_flag); } |
| static Flag::Error uintAtPut(Flag* flag, uint* value, Flag::Flags origin); |
| static Flag::Error uintAtPut(const char* name, size_t len, uint* value, Flag::Flags origin); |
| static Flag::Error uintAtPut(const char* name, uint* value, Flag::Flags origin) { return uintAtPut(name, strlen(name), value, origin); } |
| |
| static Flag::Error intxAt(const char* name, size_t len, intx* value, bool allow_locked = false, bool return_flag = false); |
| static Flag::Error intxAt(const char* name, intx* value, bool allow_locked = false, bool return_flag = false) { return intxAt(name, strlen(name), value, allow_locked, return_flag); } |
| static Flag::Error intxAtPut(Flag* flag, intx* value, Flag::Flags origin); |
| static Flag::Error intxAtPut(const char* name, size_t len, intx* value, Flag::Flags origin); |
| static Flag::Error intxAtPut(const char* name, intx* value, Flag::Flags origin) { return intxAtPut(name, strlen(name), value, origin); } |
| |
| static Flag::Error uintxAt(const char* name, size_t len, uintx* value, bool allow_locked = false, bool return_flag = false); |
| static Flag::Error uintxAt(const char* name, uintx* value, bool allow_locked = false, bool return_flag = false) { return uintxAt(name, strlen(name), value, allow_locked, return_flag); } |
| static Flag::Error uintxAtPut(Flag* flag, uintx* value, Flag::Flags origin); |
| static Flag::Error uintxAtPut(const char* name, size_t len, uintx* value, Flag::Flags origin); |
| static Flag::Error uintxAtPut(const char* name, uintx* value, Flag::Flags origin) { return uintxAtPut(name, strlen(name), value, origin); } |
| |
| static Flag::Error size_tAt(const char* name, size_t len, size_t* value, bool allow_locked = false, bool return_flag = false); |
| static Flag::Error size_tAt(const char* name, size_t* value, bool allow_locked = false, bool return_flag = false) { return size_tAt(name, strlen(name), value, allow_locked, return_flag); } |
| static Flag::Error size_tAtPut(Flag* flag, size_t* value, Flag::Flags origin); |
| static Flag::Error size_tAtPut(const char* name, size_t len, size_t* value, Flag::Flags origin); |
| static Flag::Error size_tAtPut(const char* name, size_t* value, Flag::Flags origin) { return size_tAtPut(name, strlen(name), value, origin); } |
| |
| static Flag::Error uint64_tAt(const char* name, size_t len, uint64_t* value, bool allow_locked = false, bool return_flag = false); |
| static Flag::Error uint64_tAt(const char* name, uint64_t* value, bool allow_locked = false, bool return_flag = false) { return uint64_tAt(name, strlen(name), value, allow_locked, return_flag); } |
| static Flag::Error uint64_tAtPut(Flag* flag, uint64_t* value, Flag::Flags origin); |
| static Flag::Error uint64_tAtPut(const char* name, size_t len, uint64_t* value, Flag::Flags origin); |
| static Flag::Error uint64_tAtPut(const char* name, uint64_t* value, Flag::Flags origin) { return uint64_tAtPut(name, strlen(name), value, origin); } |
| |
| static Flag::Error doubleAt(const char* name, size_t len, double* value, bool allow_locked = false, bool return_flag = false); |
| static Flag::Error doubleAt(const char* name, double* value, bool allow_locked = false, bool return_flag = false) { return doubleAt(name, strlen(name), value, allow_locked, return_flag); } |
| static Flag::Error doubleAtPut(Flag* flag, double* value, Flag::Flags origin); |
| static Flag::Error doubleAtPut(const char* name, size_t len, double* value, Flag::Flags origin); |
| static Flag::Error doubleAtPut(const char* name, double* value, Flag::Flags origin) { return doubleAtPut(name, strlen(name), value, origin); } |
| |
| static Flag::Error ccstrAt(const char* name, size_t len, ccstr* value, bool allow_locked = false, bool return_flag = false); |
| static Flag::Error ccstrAt(const char* name, ccstr* value, bool allow_locked = false, bool return_flag = false) { return ccstrAt(name, strlen(name), value, allow_locked, return_flag); } |
| // Contract: Flag will make private copy of the incoming value. |
| // Outgoing value is always malloc-ed, and caller MUST call free. |
| static Flag::Error ccstrAtPut(const char* name, size_t len, ccstr* value, Flag::Flags origin); |
| static Flag::Error ccstrAtPut(const char* name, ccstr* value, Flag::Flags origin) { return ccstrAtPut(name, strlen(name), value, origin); } |
| |
| // Returns false if name is not a command line flag. |
| static bool wasSetOnCmdline(const char* name, bool* value); |
| static void printSetFlags(outputStream* out); |
| |
| // printRanges will print out flags type, name and range values as expected by -XX:+PrintFlagsRanges |
| static void printFlags(outputStream* out, bool withComments, bool printRanges = false); |
| |
| static void verify() PRODUCT_RETURN; |
| }; |
| |
| // use this for flags that are true by default in the debug version but |
| // false in the optimized version, and vice versa |
| #ifdef ASSERT |
| #define trueInDebug true |
| #define falseInDebug false |
| #else |
| #define trueInDebug false |
| #define falseInDebug true |
| #endif |
| |
| // use this for flags that are true per default in the product build |
| // but false in development builds, and vice versa |
| #ifdef PRODUCT |
| #define trueInProduct true |
| #define falseInProduct false |
| #else |
| #define trueInProduct false |
| #define falseInProduct true |
| #endif |
| |
| // develop flags are settable / visible only during development and are constant in the PRODUCT version |
| // product flags are always settable / visible |
| // notproduct flags are settable / visible only during development and are not declared in the PRODUCT version |
| |
| // A flag must be declared with one of the following types: |
| // bool, int, uint, intx, uintx, size_t, ccstr, double, or uint64_t. |
| // The type "ccstr" is an alias for "const char*" and is used |
| // only in this file, because the macrology requires single-token type names. |
| |
| // Note: Diagnostic options not meant for VM tuning or for product modes. |
| // They are to be used for VM quality assurance or field diagnosis |
| // of VM bugs. They are hidden so that users will not be encouraged to |
| // try them as if they were VM ordinary execution options. However, they |
| // are available in the product version of the VM. Under instruction |
| // from support engineers, VM customers can turn them on to collect |
| // diagnostic information about VM problems. To use a VM diagnostic |
| // option, you must first specify +UnlockDiagnosticVMOptions. |
| // (This master switch also affects the behavior of -Xprintflags.) |
| // |
| // experimental flags are in support of features that are not |
| // part of the officially supported product, but are available |
| // for experimenting with. They could, for example, be performance |
| // features that may not have undergone full or rigorous QA, but which may |
| // help performance in some cases and released for experimentation |
| // by the community of users and developers. This flag also allows one to |
| // be able to build a fully supported product that nonetheless also |
| // ships with some unsupported, lightly tested, experimental features. |
| // Like the UnlockDiagnosticVMOptions flag above, there is a corresponding |
| // UnlockExperimentalVMOptions flag, which allows the control and |
| // modification of the experimental flags. |
| // |
| // Nota bene: neither diagnostic nor experimental options should be used casually, |
| // and they are not supported on production loads, except under explicit |
| // direction from support engineers. |
| // |
| // manageable flags are writeable external product flags. |
| // They are dynamically writeable through the JDK management interface |
| // (com.sun.management.HotSpotDiagnosticMXBean API) and also through JConsole. |
| // These flags are external exported interface (see CCC). The list of |
| // manageable flags can be queried programmatically through the management |
| // interface. |
| // |
| // A flag can be made as "manageable" only if |
| // - the flag is defined in a CCC as an external exported interface. |
| // - the VM implementation supports dynamic setting of the flag. |
| // This implies that the VM must *always* query the flag variable |
| // and not reuse state related to the flag state at any given time. |
| // - you want the flag to be queried programmatically by the customers. |
| // |
| // product_rw flags are writeable internal product flags. |
| // They are like "manageable" flags but for internal/private use. |
| // The list of product_rw flags are internal/private flags which |
| // may be changed/removed in a future release. It can be set |
| // through the management interface to get/set value |
| // when the name of flag is supplied. |
| // |
| // A flag can be made as "product_rw" only if |
| // - the VM implementation supports dynamic setting of the flag. |
| // This implies that the VM must *always* query the flag variable |
| // and not reuse state related to the flag state at any given time. |
| // |
| // Note that when there is a need to support develop flags to be writeable, |
| // it can be done in the same way as product_rw. |
| // |
| // range is a macro that will expand to min and max arguments for range |
| // checking code if provided - see commandLineFlagRangeList.hpp |
| // |
| // constraint is a macro that will expand to custom function call |
| // for constraint checking if provided - see commandLineFlagConstraintList.hpp |
| // |
| // writeable is a macro that controls if and how the value can change during the runtime |
| // |
| // writeable(Always) is optional and allows the flag to have its value changed |
| // without any limitations at any time |
| // |
| // writeable(Once) flag value's can be only set once during the lifetime of VM |
| // |
| // writeable(CommandLineOnly) flag value's can be only set from command line |
| // (multiple times allowed) |
| // |
| |
| |
| #define RUNTIME_FLAGS(develop, \ |
| develop_pd, \ |
| product, \ |
| product_pd, \ |
| diagnostic, \ |
| diagnostic_pd, \ |
| experimental, \ |
| notproduct, \ |
| manageable, \ |
| product_rw, \ |
| lp64_product, \ |
| range, \ |
| constraint, \ |
| writeable) \ |
| \ |
| lp64_product(bool, UseCompressedOops, false, \ |
| "Use 32-bit object references in 64-bit VM. " \ |
| "lp64_product means flag is always constant in 32 bit VM") \ |
| \ |
| lp64_product(bool, UseCompressedClassPointers, false, \ |
| "Use 32-bit class pointers in 64-bit VM. " \ |
| "lp64_product means flag is always constant in 32 bit VM") \ |
| \ |
| notproduct(bool, CheckCompressedOops, true, \ |
| "Generate checks in encoding/decoding code in debug VM") \ |
| \ |
| product(uintx, HeapSearchSteps, 3 PPC64_ONLY(+17), \ |
| "Heap allocation steps through preferred address regions to find" \ |
| " where it can allocate the heap. Number of steps to take per " \ |
| "region.") \ |
| range(1, max_uintx) \ |
| \ |
| lp64_product(intx, ObjectAlignmentInBytes, 8, \ |
| "Default object alignment in bytes, 8 is minimum") \ |
| range(8, 256) \ |
| constraint(ObjectAlignmentInBytesConstraintFunc,AtParse) \ |
| \ |
| product(bool, AssumeMP, false, \ |
| "Instruct the VM to assume multiple processors are available") \ |
| \ |
| /* UseMembar is theoretically a temp flag used for memory barrier */ \ |
| /* removal testing. It was supposed to be removed before FCS but has */ \ |
| /* been re-added (see 6401008) */ \ |
| product_pd(bool, UseMembar, \ |
| "(Unstable) Issues membars on thread state transitions") \ |
| \ |
| develop(bool, CleanChunkPoolAsync, true, \ |
| "Clean the chunk pool asynchronously") \ |
| \ |
| experimental(bool, AlwaysSafeConstructors, false, \ |
| "Force safe construction, as if all fields are final.") \ |
| \ |
| diagnostic(bool, UnlockDiagnosticVMOptions, trueInDebug, \ |
| "Enable normal processing of flags relating to field diagnostics")\ |
| \ |
| experimental(bool, UnlockExperimentalVMOptions, false, \ |
| "Enable normal processing of flags relating to experimental " \ |
| "features") \ |
| \ |
| product(bool, JavaMonitorsInStackTrace, true, \ |
| "Print information about Java monitor locks when the stacks are" \ |
| "dumped") \ |
| \ |
| product_pd(bool, UseLargePages, \ |
| "Use large page memory") \ |
| \ |
| product_pd(bool, UseLargePagesIndividualAllocation, \ |
| "Allocate large pages individually for better affinity") \ |
| \ |
| develop(bool, LargePagesIndividualAllocationInjectError, false, \ |
| "Fail large pages individual allocation") \ |
| \ |
| product(bool, UseLargePagesInMetaspace, false, \ |
| "Use large page memory in metaspace. " \ |
| "Only used if UseLargePages is enabled.") \ |
| \ |
| product(bool, UseNUMA, false, \ |
| "Use NUMA if available") \ |
| \ |
| product(bool, UseNUMAInterleaving, false, \ |
| "Interleave memory across NUMA nodes if available") \ |
| \ |
| product(size_t, NUMAInterleaveGranularity, 2*M, \ |
| "Granularity to use for NUMA interleaving on Windows OS") \ |
| range(os::vm_allocation_granularity(), NOT_LP64(2*G) LP64_ONLY(8192*G)) \ |
| \ |
| product(bool, ForceNUMA, false, \ |
| "Force NUMA optimizations on single-node/UMA systems") \ |
| \ |
| product(uintx, NUMAChunkResizeWeight, 20, \ |
| "Percentage (0-100) used to weight the current sample when " \ |
| "computing exponentially decaying average for " \ |
| "AdaptiveNUMAChunkSizing") \ |
| range(0, 100) \ |
| \ |
| product(size_t, NUMASpaceResizeRate, 1*G, \ |
| "Do not reallocate more than this amount per collection") \ |
| range(0, max_uintx) \ |
| \ |
| product(bool, UseAdaptiveNUMAChunkSizing, true, \ |
| "Enable adaptive chunk sizing for NUMA") \ |
| \ |
| product(bool, NUMAStats, false, \ |
| "Print NUMA stats in detailed heap information") \ |
| \ |
| product(uintx, NUMAPageScanRate, 256, \ |
| "Maximum number of pages to include in the page scan procedure") \ |
| range(0, max_uintx) \ |
| \ |
| product_pd(bool, NeedsDeoptSuspend, \ |
| "True for register window machines (sparc/ia64)") \ |
| \ |
| product(intx, UseSSE, 99, \ |
| "Highest supported SSE instructions set on x86/x64") \ |
| range(0, 99) \ |
| \ |
| product(bool, UseAES, false, \ |
| "Control whether AES instructions are used when available") \ |
| \ |
| product(bool, UseFMA, false, \ |
| "Control whether FMA instructions are used when available") \ |
| \ |
| product(bool, UseSHA, false, \ |
| "Control whether SHA instructions are used when available") \ |
| \ |
| diagnostic(bool, UseGHASHIntrinsics, false, \ |
| "Use intrinsics for GHASH versions of crypto") \ |
| \ |
| product(size_t, LargePageSizeInBytes, 0, \ |
| "Large page size (0 to let VM choose the page size)") \ |
| range(0, max_uintx) \ |
| \ |
| product(size_t, LargePageHeapSizeThreshold, 128*M, \ |
| "Use large pages if maximum heap is at least this big") \ |
| range(0, max_uintx) \ |
| \ |
| product(bool, ForceTimeHighResolution, false, \ |
| "Using high time resolution (for Win32 only)") \ |
| \ |
| develop(bool, TracePcPatching, false, \ |
| "Trace usage of frame::patch_pc") \ |
| \ |
| develop(bool, TraceRelocator, false, \ |
| "Trace the bytecode relocator") \ |
| \ |
| develop(bool, TraceLongCompiles, false, \ |
| "Print out every time compilation is longer than " \ |
| "a given threshold") \ |
| \ |
| develop(bool, SafepointALot, false, \ |
| "Generate a lot of safepoints. This works with " \ |
| "GuaranteedSafepointInterval") \ |
| \ |
| product_pd(bool, BackgroundCompilation, \ |
| "A thread requesting compilation is not blocked during " \ |
| "compilation") \ |
| \ |
| product(bool, PrintVMQWaitTime, false, \ |
| "Print out the waiting time in VM operation queue") \ |
| \ |
| product(bool, MethodFlushing, true, \ |
| "Reclamation of zombie and not-entrant methods") \ |
| \ |
| develop(bool, VerifyStack, false, \ |
| "Verify stack of each thread when it is entering a runtime call") \ |
| \ |
| diagnostic(bool, ForceUnreachable, false, \ |
| "Make all non code cache addresses to be unreachable by " \ |
| "forcing use of 64bit literal fixups") \ |
| \ |
| notproduct(bool, StressDerivedPointers, false, \ |
| "Force scavenge when a derived pointer is detected on stack " \ |
| "after rtm call") \ |
| \ |
| develop(bool, TraceDerivedPointers, false, \ |
| "Trace traversal of derived pointers on stack") \ |
| \ |
| notproduct(bool, TraceCodeBlobStacks, false, \ |
| "Trace stack-walk of codeblobs") \ |
| \ |
| product(bool, PrintJNIResolving, false, \ |
| "Used to implement -v:jni") \ |
| \ |
| notproduct(bool, PrintRewrites, false, \ |
| "Print methods that are being rewritten") \ |
| \ |
| product(bool, UseInlineCaches, true, \ |
| "Use Inline Caches for virtual calls ") \ |
| \ |
| diagnostic(bool, InlineArrayCopy, true, \ |
| "Inline arraycopy native that is known to be part of " \ |
| "base library DLL") \ |
| \ |
| diagnostic(bool, InlineObjectHash, true, \ |
| "Inline Object::hashCode() native that is known to be part " \ |
| "of base library DLL") \ |
| \ |
| diagnostic(bool, InlineNatives, true, \ |
| "Inline natives that are known to be part of base library DLL") \ |
| \ |
| diagnostic(bool, InlineMathNatives, true, \ |
| "Inline SinD, CosD, etc.") \ |
| \ |
| diagnostic(bool, InlineClassNatives, true, \ |
| "Inline Class.isInstance, etc") \ |
| \ |
| diagnostic(bool, InlineThreadNatives, true, \ |
| "Inline Thread.currentThread, etc") \ |
| \ |
| diagnostic(bool, InlineUnsafeOps, true, \ |
| "Inline memory ops (native methods) from Unsafe") \ |
| \ |
| product(bool, CriticalJNINatives, true, \ |
| "Check for critical JNI entry points") \ |
| \ |
| notproduct(bool, StressCriticalJNINatives, false, \ |
| "Exercise register saving code in critical natives") \ |
| \ |
| diagnostic(bool, UseAESIntrinsics, false, \ |
| "Use intrinsics for AES versions of crypto") \ |
| \ |
| diagnostic(bool, UseAESCTRIntrinsics, false, \ |
| "Use intrinsics for the paralleled version of AES/CTR crypto") \ |
| \ |
| diagnostic(bool, UseSHA1Intrinsics, false, \ |
| "Use intrinsics for SHA-1 crypto hash function. " \ |
| "Requires that UseSHA is enabled.") \ |
| \ |
| diagnostic(bool, UseSHA256Intrinsics, false, \ |
| "Use intrinsics for SHA-224 and SHA-256 crypto hash functions. " \ |
| "Requires that UseSHA is enabled.") \ |
| \ |
| diagnostic(bool, UseSHA512Intrinsics, false, \ |
| "Use intrinsics for SHA-384 and SHA-512 crypto hash functions. " \ |
| "Requires that UseSHA is enabled.") \ |
| \ |
| diagnostic(bool, UseCRC32Intrinsics, false, \ |
| "use intrinsics for java.util.zip.CRC32") \ |
| \ |
| diagnostic(bool, UseCRC32CIntrinsics, false, \ |
| "use intrinsics for java.util.zip.CRC32C") \ |
| \ |
| diagnostic(bool, UseAdler32Intrinsics, false, \ |
| "use intrinsics for java.util.zip.Adler32") \ |
| \ |
| diagnostic(bool, UseVectorizedMismatchIntrinsic, false, \ |
| "Enables intrinsification of ArraysSupport.vectorizedMismatch()") \ |
| \ |
| diagnostic(ccstrlist, DisableIntrinsic, "", \ |
| "do not expand intrinsics whose (internal) names appear here") \ |
| \ |
| develop(bool, TraceCallFixup, false, \ |
| "Trace all call fixups") \ |
| \ |
| develop(bool, DeoptimizeALot, false, \ |
| "Deoptimize at every exit from the runtime system") \ |
| \ |
| notproduct(ccstrlist, DeoptimizeOnlyAt, "", \ |
| "A comma separated list of bcis to deoptimize at") \ |
| \ |
| product(bool, DeoptimizeRandom, false, \ |
| "Deoptimize random frames on random exit from the runtime system")\ |
| \ |
| notproduct(bool, ZombieALot, false, \ |
| "Create zombies (non-entrant) at exit from the runtime system") \ |
| \ |
| product(bool, UnlinkSymbolsALot, false, \ |
| "Unlink unreferenced symbols from the symbol table at safepoints")\ |
| \ |
| notproduct(bool, WalkStackALot, false, \ |
| "Trace stack (no print) at every exit from the runtime system") \ |
| \ |
| product(bool, Debugging, false, \ |
| "Set when executing debug methods in debug.cpp " \ |
| "(to prevent triggering assertions)") \ |
| \ |
| notproduct(bool, StrictSafepointChecks, trueInDebug, \ |
| "Enable strict checks that safepoints cannot happen for threads " \ |
| "that use NoSafepointVerifier") \ |
| \ |
| notproduct(bool, VerifyLastFrame, false, \ |
| "Verify oops on last frame on entry to VM") \ |
| \ |
| product(bool, FailOverToOldVerifier, true, \ |
| "Fail over to old verifier when split verifier fails") \ |
| \ |
| develop(bool, ShowSafepointMsgs, false, \ |
| "Show message about safepoint synchronization") \ |
| \ |
| product(bool, SafepointTimeout, false, \ |
| "Time out and warn or fail after SafepointTimeoutDelay " \ |
| "milliseconds if failed to reach safepoint") \ |
| \ |
| develop(bool, DieOnSafepointTimeout, false, \ |
| "Die upon failure to reach safepoint (see SafepointTimeout)") \ |
| \ |
| /* 50 retries * (5 * current_retry_count) millis = ~6.375 seconds */ \ |
| /* typically, at most a few retries are needed */ \ |
| product(intx, SuspendRetryCount, 50, \ |
| "Maximum retry count for an external suspend request") \ |
| range(0, max_intx) \ |
| \ |
| product(intx, SuspendRetryDelay, 5, \ |
| "Milliseconds to delay per retry (* current_retry_count)") \ |
| range(0, max_intx) \ |
| \ |
| product(bool, AssertOnSuspendWaitFailure, false, \ |
| "Assert/Guarantee on external suspend wait failure") \ |
| \ |
| product(bool, TraceSuspendWaitFailures, false, \ |
| "Trace external suspend wait failures") \ |
| \ |
| product(bool, MaxFDLimit, true, \ |
| "Bump the number of file descriptors to maximum in Solaris") \ |
| \ |
| diagnostic(bool, LogEvents, true, \ |
| "Enable the various ring buffer event logs") \ |
| \ |
| diagnostic(uintx, LogEventsBufferEntries, 10, \ |
| "Number of ring buffer event logs") \ |
| range(1, NOT_LP64(1*K) LP64_ONLY(1*M)) \ |
| \ |
| product(bool, BytecodeVerificationRemote, true, \ |
| "Enable the Java bytecode verifier for remote classes") \ |
| \ |
| product(bool, BytecodeVerificationLocal, false, \ |
| "Enable the Java bytecode verifier for local classes") \ |
| \ |
| develop(bool, ForceFloatExceptions, trueInDebug, \ |
| "Force exceptions on FP stack under/overflow") \ |
| \ |
| develop(bool, VerifyStackAtCalls, false, \ |
| "Verify that the stack pointer is unchanged after calls") \ |
| \ |
| develop(bool, TraceJavaAssertions, false, \ |
| "Trace java language assertions") \ |
| \ |
| notproduct(bool, CheckAssertionStatusDirectives, false, \ |
| "Temporary - see javaClasses.cpp") \ |
| \ |
| notproduct(bool, PrintMallocFree, false, \ |
| "Trace calls to C heap malloc/free allocation") \ |
| \ |
| notproduct(bool, VerifyCodeCache, false, \ |
| "Verify code cache on memory allocation/deallocation") \ |
| \ |
| develop(bool, UseMallocOnly, false, \ |
| "Use only malloc/free for allocation (no resource area/arena)") \ |
| \ |
| develop(bool, PrintMalloc, false, \ |
| "Print all malloc/free calls") \ |
| \ |
| develop(bool, PrintMallocStatistics, false, \ |
| "Print malloc/free statistics") \ |
| \ |
| develop(bool, ZapResourceArea, trueInDebug, \ |
| "Zap freed resource/arena space with 0xABABABAB") \ |
| \ |
| notproduct(bool, ZapVMHandleArea, trueInDebug, \ |
| "Zap freed VM handle space with 0xBCBCBCBC") \ |
| \ |
| develop(bool, ZapJNIHandleArea, trueInDebug, \ |
| "Zap freed JNI handle space with 0xFEFEFEFE") \ |
| \ |
| notproduct(bool, ZapStackSegments, trueInDebug, \ |
| "Zap allocated/freed stack segments with 0xFADFADED") \ |
| \ |
| develop(bool, ZapUnusedHeapArea, trueInDebug, \ |
| "Zap unused heap space with 0xBAADBABE") \ |
| \ |
| develop(bool, CheckZapUnusedHeapArea, false, \ |
| "Check zapping of unused heap space") \ |
| \ |
| develop(bool, ZapFillerObjects, trueInDebug, \ |
| "Zap filler objects with 0xDEAFBABE") \ |
| \ |
| develop(bool, PrintVMMessages, true, \ |
| "Print VM messages on console") \ |
| \ |
| notproduct(uintx, ErrorHandlerTest, 0, \ |
| "If > 0, provokes an error after VM initialization; the value " \ |
| "determines which error to provoke. See test_error_handler() " \ |
| "in vmError.cpp.") \ |
| \ |
| notproduct(uintx, TestCrashInErrorHandler, 0, \ |
| "If > 0, provokes an error inside VM error handler (a secondary " \ |
| "crash). see test_error_handler() in vmError.cpp") \ |
| \ |
| notproduct(bool, TestSafeFetchInErrorHandler, false, \ |
| "If true, tests SafeFetch inside error handler.") \ |
| \ |
| notproduct(bool, TestUnresponsiveErrorHandler, false, \ |
| "If true, simulates an unresponsive error handler.") \ |
| \ |
| develop(bool, Verbose, false, \ |
| "Print additional debugging information from other modes") \ |
| \ |
| develop(bool, PrintMiscellaneous, false, \ |
| "Print uncategorized debugging information (requires +Verbose)") \ |
| \ |
| develop(bool, WizardMode, false, \ |
| "Print much more debugging information") \ |
| \ |
| product(bool, ShowMessageBoxOnError, false, \ |
| "Keep process alive on VM fatal error") \ |
| \ |
| product(bool, CreateCoredumpOnCrash, true, \ |
| "Create core/mini dump on VM fatal error") \ |
| \ |
| product(uint64_t, ErrorLogTimeout, 2 * 60, \ |
| "Timeout, in seconds, to limit the time spent on writing an " \ |
| "error log in case of a crash.") \ |
| range(0, (uint64_t)max_jlong/1000) \ |
| \ |
| product_pd(bool, UseOSErrorReporting, \ |
| "Let VM fatal error propagate to the OS (ie. WER on Windows)") \ |
| \ |
| product(bool, SuppressFatalErrorMessage, false, \ |
| "Report NO fatal error message (avoid deadlock)") \ |
| \ |
| product(ccstrlist, OnError, "", \ |
| "Run user-defined commands on fatal error; see VMError.cpp " \ |
| "for examples") \ |
| \ |
| product(ccstrlist, OnOutOfMemoryError, "", \ |
| "Run user-defined commands on first java.lang.OutOfMemoryError") \ |
| \ |
| manageable(bool, HeapDumpBeforeFullGC, false, \ |
| "Dump heap to file before any major stop-the-world GC") \ |
| \ |
| manageable(bool, HeapDumpAfterFullGC, false, \ |
| "Dump heap to file after any major stop-the-world GC") \ |
| \ |
| manageable(bool, HeapDumpOnOutOfMemoryError, false, \ |
| "Dump heap to file when java.lang.OutOfMemoryError is thrown") \ |
| \ |
| manageable(ccstr, HeapDumpPath, NULL, \ |
| "When HeapDumpOnOutOfMemoryError is on, the path (filename or " \ |
| "directory) of the dump file (defaults to java_pid<pid>.hprof " \ |
| "in the working directory)") \ |
| \ |
| develop(bool, BreakAtWarning, false, \ |
| "Execute breakpoint upon encountering VM warning") \ |
| \ |
| develop(bool, UseFakeTimers, false, \ |
| "Tell whether the VM should use system time or a fake timer") \ |
| \ |
| product(ccstr, NativeMemoryTracking, "off", \ |
| "Native memory tracking options") \ |
| \ |
| diagnostic(bool, PrintNMTStatistics, false, \ |
| "Print native memory tracking summary data if it is on") \ |
| \ |
| diagnostic(bool, LogCompilation, false, \ |
| "Log compilation activity in detail to LogFile") \ |
| \ |
| product(bool, PrintCompilation, false, \ |
| "Print compilations") \ |
| \ |
| diagnostic(bool, TraceNMethodInstalls, false, \ |
| "Trace nmethod installation") \ |
| \ |
| diagnostic(intx, ScavengeRootsInCode, 2, \ |
| "0: do not allow scavengable oops in the code cache; " \ |
| "1: allow scavenging from the code cache; " \ |
| "2: emit as many constants as the compiler can see") \ |
| range(0, 2) \ |
| \ |
| product(bool, AlwaysRestoreFPU, false, \ |
| "Restore the FPU control word after every JNI call (expensive)") \ |
| \ |
| diagnostic(bool, PrintCompilation2, false, \ |
| "Print additional statistics per compilation") \ |
| \ |
| diagnostic(bool, PrintAdapterHandlers, false, \ |
| "Print code generated for i2c/c2i adapters") \ |
| \ |
| diagnostic(bool, VerifyAdapterCalls, trueInDebug, \ |
| "Verify that i2c/c2i adapters are called properly") \ |
| \ |
| develop(bool, VerifyAdapterSharing, false, \ |
| "Verify that the code for shared adapters is the equivalent") \ |
| \ |
| diagnostic(bool, PrintAssembly, false, \ |
| "Print assembly code (using external disassembler.so)") \ |
| \ |
| diagnostic(ccstr, PrintAssemblyOptions, NULL, \ |
| "Print options string passed to disassembler.so") \ |
| \ |
| notproduct(bool, PrintNMethodStatistics, false, \ |
| "Print a summary statistic for the generated nmethods") \ |
| \ |
| diagnostic(bool, PrintNMethods, false, \ |
| "Print assembly code for nmethods when generated") \ |
| \ |
| diagnostic(bool, PrintNativeNMethods, false, \ |
| "Print assembly code for native nmethods when generated") \ |
| \ |
| develop(bool, PrintDebugInfo, false, \ |
| "Print debug information for all nmethods when generated") \ |
| \ |
| develop(bool, PrintRelocations, false, \ |
| "Print relocation information for all nmethods when generated") \ |
| \ |
| develop(bool, PrintDependencies, false, \ |
| "Print dependency information for all nmethods when generated") \ |
| \ |
| develop(bool, PrintExceptionHandlers, false, \ |
| "Print exception handler tables for all nmethods when generated") \ |
| \ |
| develop(bool, StressCompiledExceptionHandlers, false, \ |
| "Exercise compiled exception handlers") \ |
| \ |
| develop(bool, InterceptOSException, false, \ |
| "Start debugger when an implicit OS (e.g. NULL) " \ |
| "exception happens") \ |
| \ |
| product(bool, PrintCodeCache, false, \ |
| "Print the code cache memory usage when exiting") \ |
| \ |
| develop(bool, PrintCodeCache2, false, \ |
| "Print detailed usage information on the code cache when exiting")\ |
| \ |
| product(bool, PrintCodeCacheOnCompilation, false, \ |
| "Print the code cache memory usage each time a method is " \ |
| "compiled") \ |
| \ |
| diagnostic(bool, PrintStubCode, false, \ |
| "Print generated stub code") \ |
| \ |
| product(bool, StackTraceInThrowable, true, \ |
| "Collect backtrace in throwable when exception happens") \ |
| \ |
| product(bool, OmitStackTraceInFastThrow, true, \ |
| "Omit backtraces for some 'hot' exceptions in optimized code") \ |
| \ |
| product(bool, ProfilerPrintByteCodeStatistics, false, \ |
| "Print bytecode statistics when dumping profiler output") \ |
| \ |
| product(bool, ProfilerRecordPC, false, \ |
| "Collect ticks for each 16 byte interval of compiled code") \ |
| \ |
| product(bool, ProfileVM, false, \ |
| "Profile ticks that fall within VM (either in the VM Thread " \ |
| "or VM code called through stubs)") \ |
| \ |
| product(bool, ProfileIntervals, false, \ |
| "Print profiles for each interval (see ProfileIntervalsTicks)") \ |
| \ |
| notproduct(bool, ProfilerCheckIntervals, false, \ |
| "Collect and print information on spacing of profiler ticks") \ |
| \ |
| product(bool, PrintWarnings, true, \ |
| "Print JVM warnings to output stream") \ |
| \ |
| notproduct(uintx, WarnOnStalledSpinLock, 0, \ |
| "Print warnings for stalled SpinLocks") \ |
| \ |
| product(bool, RegisterFinalizersAtInit, true, \ |
| "Register finalizable objects at end of Object.<init> or " \ |
| "after allocation") \ |
| \ |
| develop(bool, RegisterReferences, true, \ |
| "Tell whether the VM should register soft/weak/final/phantom " \ |
| "references") \ |
| \ |
| develop(bool, IgnoreRewrites, false, \ |
| "Suppress rewrites of bytecodes in the oopmap generator. " \ |
| "This is unsafe!") \ |
| \ |
| develop(bool, PrintCodeCacheExtension, false, \ |
| "Print extension of code cache") \ |
| \ |
| develop(bool, UsePrivilegedStack, true, \ |
| "Enable the security JVM functions") \ |
| \ |
| develop(bool, ProtectionDomainVerification, true, \ |
| "Verify protection domain before resolution in system dictionary")\ |
| \ |
| product(bool, ClassUnloading, true, \ |
| "Do unloading of classes") \ |
| \ |
| product(bool, ClassUnloadingWithConcurrentMark, true, \ |
| "Do unloading of classes with a concurrent marking cycle") \ |
| \ |
| develop(bool, DisableStartThread, false, \ |
| "Disable starting of additional Java threads " \ |
| "(for debugging only)") \ |
| \ |
| develop(bool, MemProfiling, false, \ |
| "Write memory usage profiling to log file") \ |
| \ |
| notproduct(bool, PrintSystemDictionaryAtExit, false, \ |
| "Print the system dictionary at exit") \ |
| \ |
| experimental(intx, PredictedLoadedClassCount, 0, \ |
| "Experimental: Tune loaded class cache starting size") \ |
| \ |
| diagnostic(bool, UnsyncloadClass, false, \ |
| "Unstable: VM calls loadClass unsynchronized. Custom " \ |
| "class loader must call VM synchronized for findClass " \ |
| "and defineClass.") \ |
| \ |
| product(bool, AlwaysLockClassLoader, false, \ |
| "Require the VM to acquire the class loader lock before calling " \ |
| "loadClass() even for class loaders registering " \ |
| "as parallel capable") \ |
| \ |
| product(bool, AllowParallelDefineClass, false, \ |
| "Allow parallel defineClass requests for class loaders " \ |
| "registering as parallel capable") \ |
| \ |
| product(bool, MustCallLoadClassInternal, false, \ |
| "Call loadClassInternal() rather than loadClass()") \ |
| \ |
| product_pd(bool, DontYieldALot, \ |
| "Throw away obvious excess yield calls") \ |
| \ |
| develop(bool, UseDetachedThreads, true, \ |
| "Use detached threads that are recycled upon termination " \ |
| "(for Solaris only)") \ |
| \ |
| product(bool, UseLWPSynchronization, true, \ |
| "Use LWP-based instead of libthread-based synchronization " \ |
| "(SPARC only)") \ |
| \ |
| experimental(ccstr, SyncKnobs, NULL, \ |
| "(Unstable) Various monitor synchronization tunables") \ |
| \ |
| experimental(intx, EmitSync, 0, \ |
| "(Unsafe, Unstable) " \ |
| "Control emission of inline sync fast-path code") \ |
| \ |
| product(intx, MonitorBound, 0, "Bound Monitor population") \ |
| range(0, max_jint) \ |
| \ |
| product(bool, MonitorInUseLists, true, "Track Monitors for Deflation") \ |
| \ |
| experimental(intx, MonitorUsedDeflationThreshold, 90, \ |
| "Percentage of used monitors before triggering cleanup " \ |
| "safepoint which deflates monitors (0 is off). " \ |
| "The check is performed on GuaranteedSafepointInterval.") \ |
| range(0, 100) \ |
| \ |
| experimental(intx, SyncFlags, 0, "(Unsafe, Unstable) " \ |
| "Experimental Sync flags") \ |
| \ |
| experimental(intx, SyncVerbose, 0, "(Unstable)") \ |
| \ |
| diagnostic(bool, InlineNotify, true, "intrinsify subset of notify") \ |
| \ |
| experimental(intx, hashCode, 5, \ |
| "(Unstable) select hashCode generation algorithm") \ |
| \ |
| product(bool, FilterSpuriousWakeups, true, \ |
| "When true prevents OS-level spurious, or premature, wakeups " \ |
| "from Object.wait (Ignored for Windows)") \ |
| \ |
| experimental(intx, NativeMonitorTimeout, -1, "(Unstable)") \ |
| \ |
| experimental(intx, NativeMonitorFlags, 0, "(Unstable)") \ |
| \ |
| experimental(intx, NativeMonitorSpinLimit, 20, "(Unstable)") \ |
| \ |
| develop(bool, UsePthreads, false, \ |
| "Use pthread-based instead of libthread-based synchronization " \ |
| "(SPARC only)") \ |
| \ |
| product(bool, ReduceSignalUsage, false, \ |
| "Reduce the use of OS signals in Java and/or the VM") \ |
| \ |
| develop_pd(bool, ShareVtableStubs, \ |
| "Share vtable stubs (smaller code but worse branch prediction") \ |
| \ |
| develop(bool, LoadLineNumberTables, true, \ |
| "Tell whether the class file parser loads line number tables") \ |
| \ |
| develop(bool, LoadLocalVariableTables, true, \ |
| "Tell whether the class file parser loads local variable tables") \ |
| \ |
| develop(bool, LoadLocalVariableTypeTables, true, \ |
| "Tell whether the class file parser loads local variable type" \ |
| "tables") \ |
| \ |
| product(bool, AllowUserSignalHandlers, false, \ |
| "Do not complain if the application installs signal handlers " \ |
| "(Solaris & Linux only)") \ |
| \ |
| product(bool, UseSignalChaining, true, \ |
| "Use signal-chaining to invoke signal handlers installed " \ |
| "by the application (Solaris & Linux only)") \ |
| \ |
| product(bool, AllowJNIEnvProxy, false, \ |
| "Allow JNIEnv proxies for jdbx") \ |
| \ |
| product(bool, RestoreMXCSROnJNICalls, false, \ |
| "Restore MXCSR when returning from JNI calls") \ |
| \ |
| product(bool, CheckJNICalls, false, \ |
| "Verify all arguments to JNI calls") \ |
| \ |
| product(bool, CheckEndorsedAndExtDirs, false, \ |
| "Verify the endorsed and extension directories are not used") \ |
| \ |
| product(bool, UseFastJNIAccessors, true, \ |
| "Use optimized versions of Get<Primitive>Field") \ |
| \ |
| product(intx, MaxJNILocalCapacity, 65536, \ |
| "Maximum allowable local JNI handle capacity to " \ |
| "EnsureLocalCapacity() and PushLocalFrame(), " \ |
| "where <= 0 is unlimited, default: 65536") \ |
| range(min_intx, max_intx) \ |
| \ |
| product(bool, EagerXrunInit, false, \ |
| "Eagerly initialize -Xrun libraries; allows startup profiling, " \ |
| "but not all -Xrun libraries may support the state of the VM " \ |
| "at this time") \ |
| \ |
| product(bool, PreserveAllAnnotations, false, \ |
| "Preserve RuntimeInvisibleAnnotations as well " \ |
| "as RuntimeVisibleAnnotations") \ |
| \ |
| develop(uintx, PreallocatedOutOfMemoryErrorCount, 4, \ |
| "Number of OutOfMemoryErrors preallocated with backtrace") \ |
| \ |
| product(bool, UseXMMForArrayCopy, false, \ |
| "Use SSE2 MOVQ instruction for Arraycopy") \ |
| \ |
| product(intx, FieldsAllocationStyle, 1, \ |
| "0 - type based with oops first, " \ |
| "1 - with oops last, " \ |
| "2 - oops in super and sub classes are together") \ |
| range(0, 2) \ |
| \ |
| product(bool, CompactFields, true, \ |
| "Allocate nonstatic fields in gaps between previous fields") \ |
| \ |
| notproduct(bool, PrintFieldLayout, false, \ |
| "Print field layout for each class") \ |
| \ |
| /* Need to limit the extent of the padding to reasonable size. */\ |
| /* 8K is well beyond the reasonable HW cache line size, even with */\ |
| /* aggressive prefetching, while still leaving the room for segregating */\ |
| /* among the distinct pages. */\ |
| product(intx, ContendedPaddingWidth, 128, \ |
| "How many bytes to pad the fields/classes marked @Contended with")\ |
| range(0, 8192) \ |
| constraint(ContendedPaddingWidthConstraintFunc,AfterErgo) \ |
| \ |
| product(bool, EnableContended, true, \ |
| "Enable @Contended annotation support") \ |
| \ |
| product(bool, RestrictContended, true, \ |
| "Restrict @Contended to trusted classes") \ |
| \ |
| product(bool, UseBiasedLocking, true, \ |
| "Enable biased locking in JVM") \ |
| \ |
| product(intx, BiasedLockingStartupDelay, 0, \ |
| "Number of milliseconds to wait before enabling biased locking") \ |
| range(0, (intx)(max_jint-(max_jint%PeriodicTask::interval_gran))) \ |
| constraint(BiasedLockingStartupDelayFunc,AfterErgo) \ |
| \ |
| diagnostic(bool, PrintBiasedLockingStatistics, false, \ |
| "Print statistics of biased locking in JVM") \ |
| \ |
| product(intx, BiasedLockingBulkRebiasThreshold, 20, \ |
| "Threshold of number of revocations per type to try to " \ |
| "rebias all objects in the heap of that type") \ |
| range(0, max_intx) \ |
| constraint(BiasedLockingBulkRebiasThresholdFunc,AfterErgo) \ |
| \ |
| product(intx, BiasedLockingBulkRevokeThreshold, 40, \ |
| "Threshold of number of revocations per type to permanently " \ |
| "revoke biases of all objects in the heap of that type") \ |
| range(0, max_intx) \ |
| constraint(BiasedLockingBulkRevokeThresholdFunc,AfterErgo) \ |
| \ |
| product(intx, BiasedLockingDecayTime, 25000, \ |
| "Decay time (in milliseconds) to re-enable bulk rebiasing of a " \ |
| "type after previous bulk rebias") \ |
| range(500, max_intx) \ |
| constraint(BiasedLockingDecayTimeFunc,AfterErgo) \ |
| \ |
| product(bool, ExitOnOutOfMemoryError, false, \ |
| "JVM exits on the first occurrence of an out-of-memory error") \ |
| \ |
| product(bool, CrashOnOutOfMemoryError, false, \ |
| "JVM aborts, producing an error log and core/mini dump, on the " \ |
| "first occurrence of an out-of-memory error") \ |
| \ |
| /* tracing */ \ |
| \ |
| develop(bool, StressRewriter, false, \ |
| "Stress linktime bytecode rewriting") \ |
| \ |
| product(ccstr, TraceJVMTI, NULL, \ |
| "Trace flags for JVMTI functions and events") \ |
| \ |
| /* This option can change an EMCP method into an obsolete method. */ \ |
| /* This can affect tests that except specific methods to be EMCP. */ \ |
| /* This option should be used with caution. */ \ |
| product(bool, StressLdcRewrite, false, \ |
| "Force ldc -> ldc_w rewrite during RedefineClasses") \ |
| \ |
| /* change to false by default sometime after Mustang */ \ |
| product(bool, VerifyMergedCPBytecodes, true, \ |
| "Verify bytecodes after RedefineClasses constant pool merging") \ |
| \ |
| develop(bool, TraceBytecodes, false, \ |
| "Trace bytecode execution") \ |
| \ |
| develop(bool, TraceICs, false, \ |
| "Trace inline cache changes") \ |
| \ |
| notproduct(bool, TraceInvocationCounterOverflow, false, \ |
| "Trace method invocation counter overflow") \ |
| \ |
| develop(bool, TraceInlineCacheClearing, false, \ |
| "Trace clearing of inline caches in nmethods") \ |
| \ |
| develop(bool, TraceDependencies, false, \ |
| "Trace dependencies") \ |
| \ |
| develop(bool, VerifyDependencies, trueInDebug, \ |
| "Exercise and verify the compilation dependency mechanism") \ |
| \ |
| develop(bool, TraceNewOopMapGeneration, false, \ |
| "Trace OopMapGeneration") \ |
| \ |
| develop(bool, TraceNewOopMapGenerationDetailed, false, \ |
| "Trace OopMapGeneration: print detailed cell states") \ |
| \ |
| develop(bool, TimeOopMap, false, \ |
| "Time calls to GenerateOopMap::compute_map() in sum") \ |
| \ |
| develop(bool, TimeOopMap2, false, \ |
| "Time calls to GenerateOopMap::compute_map() individually") \ |
| \ |
| develop(bool, TraceOopMapRewrites, false, \ |
| "Trace rewriting of method oops during oop map generation") \ |
| \ |
| develop(bool, TraceICBuffer, false, \ |
| "Trace usage of IC buffer") \ |
| \ |
| develop(bool, TraceCompiledIC, false, \ |
| "Trace changes of compiled IC") \ |
| \ |
| /* gc */ \ |
| \ |
| product(bool, UseSerialGC, false, \ |
| "Use the Serial garbage collector") \ |
| \ |
| product(bool, UseG1GC, false, \ |
| "Use the Garbage-First garbage collector") \ |
| \ |
| product(bool, UseParallelGC, false, \ |
| "Use the Parallel Scavenge garbage collector") \ |
| \ |
| product(bool, UseParallelOldGC, false, \ |
| "Use the Parallel Old garbage collector") \ |
| \ |
| product(uintx, HeapMaximumCompactionInterval, 20, \ |
| "How often should we maximally compact the heap (not allowing " \ |
| "any dead space)") \ |
| range(0, max_uintx) \ |
| \ |
| product(uintx, HeapFirstMaximumCompactionCount, 3, \ |
| "The collection count for the first maximum compaction") \ |
| range(0, max_uintx) \ |
| \ |
| product(bool, UseMaximumCompactionOnSystemGC, true, \ |
| "Use maximum compaction in the Parallel Old garbage collector " \ |
| "for a system GC") \ |
| \ |
| product(uintx, ParallelOldDeadWoodLimiterMean, 50, \ |
| "The mean used by the parallel compact dead wood " \ |
| "limiter (a number between 0-100)") \ |
| range(0, 100) \ |
| \ |
| product(uintx, ParallelOldDeadWoodLimiterStdDev, 80, \ |
| "The standard deviation used by the parallel compact dead wood " \ |
| "limiter (a number between 0-100)") \ |
| range(0, 100) \ |
| \ |
| product(uint, ParallelGCThreads, 0, \ |
| "Number of parallel threads parallel gc will use") \ |
| constraint(ParallelGCThreadsConstraintFunc,AfterErgo) \ |
| \ |
| diagnostic(bool, UseSemaphoreGCThreadsSynchronization, true, \ |
| "Use semaphore synchronization for the GC Threads, " \ |
| "instead of synchronization based on mutexes") \ |
| \ |
| product(bool, UseDynamicNumberOfGCThreads, false, \ |
| "Dynamically choose the number of parallel threads " \ |
| "parallel gc will use") \ |
| \ |
| diagnostic(bool, InjectGCWorkerCreationFailure, false, \ |
| "Inject thread creation failures for " \ |
| "UseDynamicNumberOfGCThreads") \ |
| \ |
| diagnostic(bool, ForceDynamicNumberOfGCThreads, false, \ |
| "Force dynamic selection of the number of " \ |
| "parallel threads parallel gc will use to aid debugging") \ |
| \ |
| product(size_t, HeapSizePerGCThread, ScaleForWordSize(64*M), \ |
| "Size of heap (bytes) per GC thread used in calculating the " \ |
| "number of GC threads") \ |
| range((size_t)os::vm_page_size(), (size_t)max_uintx) \ |
| \ |
| product(uint, ConcGCThreads, 0, \ |
| "Number of threads concurrent gc will use") \ |
| constraint(ConcGCThreadsConstraintFunc,AfterErgo) \ |
| \ |
| product(uint, GCTaskTimeStampEntries, 200, \ |
| "Number of time stamp entries per gc worker thread") \ |
| range(1, max_jint) \ |
| \ |
| product(bool, AlwaysTenure, false, \ |
| "Always tenure objects in eden (ParallelGC only)") \ |
| \ |
| product(bool, NeverTenure, false, \ |
| "Never tenure objects in eden, may tenure on overflow " \ |
| "(ParallelGC only)") \ |
| \ |
| product(bool, ScavengeBeforeFullGC, true, \ |
| "Scavenge youngest generation before each full GC.") \ |
| \ |
| product(bool, UseConcMarkSweepGC, false, \ |
| "Use Concurrent Mark-Sweep GC in the old generation") \ |
| \ |
| product(bool, ExplicitGCInvokesConcurrent, false, \ |
| "A System.gc() request invokes a concurrent collection; " \ |
| "(effective only when using concurrent collectors)") \ |
| \ |
| product(bool, GCLockerInvokesConcurrent, false, \ |
| "The exit of a JNI critical section necessitating a scavenge, " \ |
| "also kicks off a background concurrent collection") \ |
| \ |
| product(uintx, GCLockerEdenExpansionPercent, 5, \ |
| "How much the GC can expand the eden by while the GC locker " \ |
| "is active (as a percentage)") \ |
| range(0, 100) \ |
| \ |
| diagnostic(uintx, GCLockerRetryAllocationCount, 2, \ |
| "Number of times to retry allocations when " \ |
| "blocked by the GC locker") \ |
| range(0, max_uintx) \ |
| \ |
| product(bool, UseCMSBestFit, true, \ |
| "Use CMS best fit allocation strategy") \ |
| \ |
| product(uintx, ParallelGCBufferWastePct, 10, \ |
| "Wasted fraction of parallel allocation buffer") \ |
| range(0, 100) \ |
| \ |
| product(uintx, TargetPLABWastePct, 10, \ |
| "Target wasted space in last buffer as percent of overall " \ |
| "allocation") \ |
| range(1, 100) \ |
| \ |
| product(uintx, PLABWeight, 75, \ |
| "Percentage (0-100) used to weight the current sample when " \ |
| "computing exponentially decaying average for ResizePLAB") \ |
| range(0, 100) \ |
| \ |
| product(bool, ResizePLAB, true, \ |
| "Dynamically resize (survivor space) promotion LAB's") \ |
| \ |
| product(int, ParGCArrayScanChunk, 50, \ |
| "Scan a subset of object array and push remainder, if array is " \ |
| "bigger than this") \ |
| range(1, max_jint/3) \ |
| \ |
| product(bool, ParGCUseLocalOverflow, false, \ |
| "Instead of a global overflow list, use local overflow stacks") \ |
| \ |
| product(bool, ParGCTrimOverflow, true, \ |
| "Eagerly trim the local overflow lists " \ |
| "(when ParGCUseLocalOverflow)") \ |
| \ |
| notproduct(bool, ParGCWorkQueueOverflowALot, false, \ |
| "Simulate work queue overflow in ParNew") \ |
| \ |
| notproduct(uintx, ParGCWorkQueueOverflowInterval, 1000, \ |
| "An `interval' counter that determines how frequently " \ |
| "we simulate overflow; a smaller number increases frequency") \ |
| \ |
| product(uintx, ParGCDesiredObjsFromOverflowList, 20, \ |
| "The desired number of objects to claim from the overflow list") \ |
| range(0, max_uintx) \ |
| \ |
| diagnostic(uintx, ParGCStridesPerThread, 2, \ |
| "The number of strides per worker thread that we divide up the " \ |
| "card table scanning work into") \ |
| range(1, max_uintx) \ |
| constraint(ParGCStridesPerThreadConstraintFunc,AfterErgo) \ |
| \ |
| diagnostic(intx, ParGCCardsPerStrideChunk, 256, \ |
| "The number of cards in each chunk of the parallel chunks used " \ |
| "during card table scanning") \ |
| range(1, max_intx) \ |
| constraint(ParGCCardsPerStrideChunkConstraintFunc,AfterMemoryInit)\ |
| \ |
| product(uintx, OldPLABWeight, 50, \ |
| "Percentage (0-100) used to weight the current sample when " \ |
| "computing exponentially decaying average for resizing " \ |
| "OldPLABSize") \ |
| range(0, 100) \ |
| \ |
| product(bool, ResizeOldPLAB, true, \ |
| "Dynamically resize (old gen) promotion LAB's") \ |
| \ |
| product(size_t, CMSOldPLABMax, 1024, \ |
| "Maximum size of CMS gen promotion LAB caches per worker " \ |
| "per block size") \ |
| range(1, max_uintx) \ |
| constraint(CMSOldPLABMaxConstraintFunc,AfterMemoryInit) \ |
| \ |
| product(size_t, CMSOldPLABMin, 16, \ |
| "Minimum size of CMS gen promotion LAB caches per worker " \ |
| "per block size") \ |
| range(1, max_uintx) \ |
| constraint(CMSOldPLABMinConstraintFunc,AfterMemoryInit) \ |
| \ |
| product(uintx, CMSOldPLABNumRefills, 4, \ |
| "Nominal number of refills of CMS gen promotion LAB cache " \ |
| "per worker per block size") \ |
| range(1, max_uintx) \ |
| \ |
| product(bool, CMSOldPLABResizeQuicker, false, \ |
| "React on-the-fly during a scavenge to a sudden " \ |
| "change in block demand rate") \ |
| \ |
| product(uintx, CMSOldPLABToleranceFactor, 4, \ |
| "The tolerance of the phase-change detector for on-the-fly " \ |
| "PLAB resizing during a scavenge") \ |
| range(1, max_uintx) \ |
| \ |
| product(uintx, CMSOldPLABReactivityFactor, 2, \ |
| "The gain in the feedback loop for on-the-fly PLAB resizing " \ |
| "during a scavenge") \ |
| range(1, max_uintx) \ |
| \ |
| product(bool, AlwaysPreTouch, false, \ |
| "Force all freshly committed pages to be pre-touched") \ |
| \ |
| product(size_t, PreTouchParallelChunkSize, 1 * G, \ |
| "Per-thread chunk size for parallel memory pre-touch.") \ |
| range(1, SIZE_MAX / 2) \ |
| \ |
| product_pd(size_t, CMSYoungGenPerWorker, \ |
| "The maximum size of young gen chosen by default per GC worker " \ |
| "thread available") \ |
| range(1, max_uintx) \ |
| \ |
| product(uintx, CMSIncrementalSafetyFactor, 10, \ |
| "Percentage (0-100) used to add conservatism when computing the " \ |
| "duty cycle") \ |
| range(0, 100) \ |
| \ |
| product(uintx, CMSExpAvgFactor, 50, \ |
| "Percentage (0-100) used to weight the current sample when " \ |
| "computing exponential averages for CMS statistics") \ |
| range(0, 100) \ |
| \ |
| product(uintx, CMS_FLSWeight, 75, \ |
| "Percentage (0-100) used to weight the current sample when " \ |
| "computing exponentially decaying averages for CMS FLS " \ |
| "statistics") \ |
| range(0, 100) \ |
| \ |
| product(uintx, CMS_FLSPadding, 1, \ |
| "The multiple of deviation from mean to use for buffering " \ |
| "against volatility in free list demand") \ |
| range(0, max_juint) \ |
| \ |
| product(uintx, FLSCoalescePolicy, 2, \ |
| "CMS: aggressiveness level for coalescing, increasing " \ |
| "from 0 to 4") \ |
| range(0, 4) \ |
| \ |
| product(bool, FLSAlwaysCoalesceLarge, false, \ |
| "CMS: larger free blocks are always available for coalescing") \ |
| \ |
| product(double, FLSLargestBlockCoalesceProximity, 0.99, \ |
| "CMS: the smaller the percentage the greater the coalescing " \ |
| "force") \ |
| range(0.0, 1.0) \ |
| \ |
| product(double, CMSSmallCoalSurplusPercent, 1.05, \ |
| "CMS: the factor by which to inflate estimated demand of small " \ |
| "block sizes to prevent coalescing with an adjoining block") \ |
| range(0.0, DBL_MAX) \ |
| \ |
| product(double, CMSLargeCoalSurplusPercent, 0.95, \ |
| "CMS: the factor by which to inflate estimated demand of large " \ |
| "block sizes to prevent coalescing with an adjoining block") \ |
| range(0.0, DBL_MAX) \ |
| \ |
| product(double, CMSSmallSplitSurplusPercent, 1.10, \ |
| "CMS: the factor by which to inflate estimated demand of small " \ |
| "block sizes to prevent splitting to supply demand for smaller " \ |
| "blocks") \ |
| range(0.0, DBL_MAX) \ |
| \ |
| product(double, CMSLargeSplitSurplusPercent, 1.00, \ |
| "CMS: the factor by which to inflate estimated demand of large " \ |
| "block sizes to prevent splitting to supply demand for smaller " \ |
| "blocks") \ |
| range(0.0, DBL_MAX) \ |
| \ |
| product(bool, CMSExtrapolateSweep, false, \ |
| "CMS: cushion for block demand during sweep") \ |
| \ |
| product(uintx, CMS_SweepWeight, 75, \ |
| "Percentage (0-100) used to weight the current sample when " \ |
| "computing exponentially decaying average for inter-sweep " \ |
| "duration") \ |
| range(0, 100) \ |
| \ |
| product(uintx, CMS_SweepPadding, 1, \ |
| "The multiple of deviation from mean to use for buffering " \ |
| "against volatility in inter-sweep duration") \ |
| range(0, max_juint) \ |
| \ |
| product(uintx, CMS_SweepTimerThresholdMillis, 10, \ |
| "Skip block flux-rate sampling for an epoch unless inter-sweep " \ |
| "duration exceeds this threshold in milliseconds") \ |
| range(0, max_uintx) \ |
| \ |
| product(bool, CMSClassUnloadingEnabled, true, \ |
| "Whether class unloading enabled when using CMS GC") \ |
| \ |
| product(uintx, CMSClassUnloadingMaxInterval, 0, \ |
| "When CMS class unloading is enabled, the maximum CMS cycle " \ |
| "count for which classes may not be unloaded") \ |
| range(0, max_uintx) \ |
| \ |
| product(uintx, CMSIndexedFreeListReplenish, 4, \ |
| "Replenish an indexed free list with this number of chunks") \ |
| range(1, max_uintx) \ |
| \ |
| product(bool, CMSReplenishIntermediate, true, \ |
| "Replenish all intermediate free-list caches") \ |
| \ |
| product(bool, CMSSplitIndexedFreeListBlocks, true, \ |
| "When satisfying batched demand, split blocks from the " \ |
| "IndexedFreeList whose size is a multiple of requested size") \ |
| \ |
| product(bool, CMSLoopWarn, false, \ |
| "Warn in case of excessive CMS looping") \ |
| \ |
| /* where does the range max value of (max_jint - 1) come from? */ \ |
| product(size_t, MarkStackSizeMax, NOT_LP64(4*M) LP64_ONLY(512*M), \ |
| "Maximum size of marking stack") \ |
| range(1, (max_jint - 1)) \ |
| \ |
| product(size_t, MarkStackSize, NOT_LP64(32*K) LP64_ONLY(4*M), \ |
| "Size of marking stack") \ |
| constraint(MarkStackSizeConstraintFunc,AfterErgo) \ |
| \ |
| notproduct(bool, CMSMarkStackOverflowALot, false, \ |
| "Simulate frequent marking stack / work queue overflow") \ |
| \ |
| notproduct(uintx, CMSMarkStackOverflowInterval, 1000, \ |
| "An \"interval\" counter that determines how frequently " \ |
| "to simulate overflow; a smaller number increases frequency") \ |
| \ |
| product(uintx, CMSMaxAbortablePrecleanLoops, 0, \ |
| "Maximum number of abortable preclean iterations, if > 0") \ |
| range(0, max_uintx) \ |
| \ |
| product(intx, CMSMaxAbortablePrecleanTime, 5000, \ |
| "Maximum time in abortable preclean (in milliseconds)") \ |
| range(0, max_intx) \ |
| \ |
| product(uintx, CMSAbortablePrecleanMinWorkPerIteration, 100, \ |
| "Nominal minimum work per abortable preclean iteration") \ |
| range(0, max_uintx) \ |
| \ |
| manageable(intx, CMSAbortablePrecleanWaitMillis, 100, \ |
| "Time that we sleep between iterations when not given " \ |
| "enough work per iteration") \ |
| range(0, max_intx) \ |
| \ |
| /* 4096 = CardTableModRefBS::card_size_in_words * BitsPerWord */ \ |
| product(size_t, CMSRescanMultiple, 32, \ |
| "Size (in cards) of CMS parallel rescan task") \ |
| range(1, SIZE_MAX / 4096) \ |
| constraint(CMSRescanMultipleConstraintFunc,AfterMemoryInit) \ |
| \ |
| /* 4096 = CardTableModRefBS::card_size_in_words * BitsPerWord */ \ |
| product(size_t, CMSConcMarkMultiple, 32, \ |
| "Size (in cards) of CMS concurrent MT marking task") \ |
| range(1, SIZE_MAX / 4096) \ |
| constraint(CMSConcMarkMultipleConstraintFunc,AfterMemoryInit) \ |
| \ |
| product(bool, CMSAbortSemantics, false, \ |
| "Whether abort-on-overflow semantics is implemented") \ |
| \ |
| product(bool, CMSParallelInitialMarkEnabled, true, \ |
| "Use the parallel initial mark.") \ |
| \ |
| product(bool, CMSParallelRemarkEnabled, true, \ |
| "Whether parallel remark enabled (only if ParNewGC)") \ |
| \ |
| product(bool, CMSParallelSurvivorRemarkEnabled, true, \ |
| "Whether parallel remark of survivor space " \ |
| "enabled (effective only if CMSParallelRemarkEnabled)") \ |
| \ |
| product(bool, CMSPLABRecordAlways, true, \ |
| "Always record survivor space PLAB boundaries (effective only " \ |
| "if CMSParallelSurvivorRemarkEnabled)") \ |
| \ |
| product(bool, CMSEdenChunksRecordAlways, true, \ |
| "Always record eden chunks used for the parallel initial mark " \ |
| "or remark of eden") \ |
| \ |
| product(bool, CMSConcurrentMTEnabled, true, \ |
| "Whether multi-threaded concurrent work enabled " \ |
| "(effective only if ParNewGC)") \ |
| \ |
| product(bool, CMSPrecleaningEnabled, true, \ |
| "Whether concurrent precleaning enabled") \ |
| \ |
| product(uintx, CMSPrecleanIter, 3, \ |
| "Maximum number of precleaning iteration passes") \ |
| range(0, 9) \ |
| \ |
| product(uintx, CMSPrecleanDenominator, 3, \ |
| "CMSPrecleanNumerator:CMSPrecleanDenominator yields convergence " \ |
| "ratio") \ |
| range(1, max_uintx) \ |
| constraint(CMSPrecleanDenominatorConstraintFunc,AfterErgo) \ |
| \ |
| product(uintx, CMSPrecleanNumerator, 2, \ |
| "CMSPrecleanNumerator:CMSPrecleanDenominator yields convergence " \ |
| "ratio") \ |
| range(0, max_uintx-1) \ |
| constraint(CMSPrecleanNumeratorConstraintFunc,AfterErgo) \ |
| \ |
| product(bool, CMSPrecleanRefLists1, true, \ |
| "Preclean ref lists during (initial) preclean phase") \ |
| \ |
| product(bool, CMSPrecleanRefLists2, false, \ |
| "Preclean ref lists during abortable preclean phase") \ |
| \ |
| product(bool, CMSPrecleanSurvivors1, false, \ |
| "Preclean survivors during (initial) preclean phase") \ |
| \ |
| product(bool, CMSPrecleanSurvivors2, true, \ |
| "Preclean survivors during abortable preclean phase") \ |
| \ |
| product(uintx, CMSPrecleanThreshold, 1000, \ |
| "Do not iterate again if number of dirty cards is less than this")\ |
| range(100, max_uintx) \ |
| \ |
| product(bool, CMSCleanOnEnter, true, \ |
| "Clean-on-enter optimization for reducing number of dirty cards") \ |
| \ |
| product(uintx, CMSRemarkVerifyVariant, 1, \ |
| "Choose variant (1,2) of verification following remark") \ |
| range(1, 2) \ |
| \ |
| product(size_t, CMSScheduleRemarkEdenSizeThreshold, 2*M, \ |
| "If Eden size is below this, do not try to schedule remark") \ |
| range(0, max_uintx) \ |
| \ |
| product(uintx, CMSScheduleRemarkEdenPenetration, 50, \ |
| "The Eden occupancy percentage (0-100) at which " \ |
| "to try and schedule remark pause") \ |
| range(0, 100) \ |
| \ |
| product(uintx, CMSScheduleRemarkSamplingRatio, 5, \ |
| "Start sampling eden top at least before young gen " \ |
| "occupancy reaches 1/<ratio> of the size at which " \ |
| "we plan to schedule remark") \ |
| range(1, max_uintx) \ |
| \ |
| product(uintx, CMSSamplingGrain, 16*K, \ |
| "The minimum distance between eden samples for CMS (see above)") \ |
| range(ObjectAlignmentInBytes, max_uintx) \ |
| constraint(CMSSamplingGrainConstraintFunc,AfterMemoryInit) \ |
| \ |
| product(bool, CMSScavengeBeforeRemark, false, \ |
| "Attempt scavenge before the CMS remark step") \ |
| \ |
| product(uintx, CMSWorkQueueDrainThreshold, 10, \ |
| "Don't drain below this size per parallel worker/thief") \ |
| range(1, max_juint) \ |
| constraint(CMSWorkQueueDrainThresholdConstraintFunc,AfterErgo) \ |
| \ |
| manageable(intx, CMSWaitDuration, 2000, \ |
| "Time in milliseconds that CMS thread waits for young GC") \ |
| range(min_jint, max_jint) \ |
| \ |
| develop(uintx, CMSCheckInterval, 1000, \ |
| "Interval in milliseconds that CMS thread checks if it " \ |
| "should start a collection cycle") \ |
| \ |
| product(bool, CMSYield, true, \ |
| "Yield between steps of CMS") \ |
| \ |
| product(size_t, CMSBitMapYieldQuantum, 10*M, \ |
| "Bitmap operations should process at most this many bits " \ |
| "between yields") \ |
| range(1, max_uintx) \ |
| constraint(CMSBitMapYieldQuantumConstraintFunc,AfterMemoryInit) \ |
| \ |
| product(bool, CMSPrintChunksInDump, false, \ |
| "If logging for the \"gc\" and \"promotion\" tags is enabled on" \ |
| "trace level include more detailed information about the" \ |
| "free chunks") \ |
| \ |
| product(bool, CMSPrintObjectsInDump, false, \ |
| "If logging for the \"gc\" and \"promotion\" tags is enabled on" \ |
| "trace level include more detailed information about the" \ |
| "allocated objects") \ |
| \ |
| diagnostic(bool, FLSVerifyAllHeapReferences, false, \ |
| "Verify that all references across the FLS boundary " \ |
| "are to valid objects") \ |
| \ |
| diagnostic(bool, FLSVerifyLists, false, \ |
| "Do lots of (expensive) FreeListSpace verification") \ |
| \ |
| diagnostic(bool, FLSVerifyIndexTable, false, \ |
| "Do lots of (expensive) FLS index table verification") \ |
| \ |
| develop(bool, FLSVerifyDictionary, false, \ |
| "Do lots of (expensive) FLS dictionary verification") \ |
| \ |
| develop(bool, VerifyBlockOffsetArray, false, \ |
| "Do (expensive) block offset array verification") \ |
| \ |
| diagnostic(bool, BlockOffsetArrayUseUnallocatedBlock, false, \ |
| "Maintain _unallocated_block in BlockOffsetArray " \ |
| "(currently applicable only to CMS collector)") \ |
| \ |
| product(intx, RefDiscoveryPolicy, 0, \ |
| "Select type of reference discovery policy: " \ |
| "reference-based(0) or referent-based(1)") \ |
| range(ReferenceProcessor::DiscoveryPolicyMin, \ |
| ReferenceProcessor::DiscoveryPolicyMax) \ |
| \ |
| product(bool, ParallelRefProcEnabled, false, \ |
| "Enable parallel reference processing whenever possible") \ |
| \ |
| product(bool, ParallelRefProcBalancingEnabled, true, \ |
| "Enable balancing of reference processing queues") \ |
| \ |
| product(uintx, CMSTriggerRatio, 80, \ |
| "Percentage of MinHeapFreeRatio in CMS generation that is " \ |
| "allocated before a CMS collection cycle commences") \ |
| range(0, 100) \ |
| \ |
| product(uintx, CMSBootstrapOccupancy, 50, \ |
| "Percentage CMS generation occupancy at which to " \ |
| "initiate CMS collection for bootstrapping collection stats") \ |
| range(0, 100) \ |
| \ |
| product(intx, CMSInitiatingOccupancyFraction, -1, \ |
| "Percentage CMS generation occupancy to start a CMS collection " \ |
| "cycle. A negative value means that CMSTriggerRatio is used") \ |
| range(min_intx, 100) \ |
| \ |
| product(uintx, InitiatingHeapOccupancyPercent, 45, \ |
| "The percent occupancy (IHOP) of the current old generation " \ |
| "capacity above which a concurrent mark cycle will be initiated " \ |
| "Its value may change over time if adaptive IHOP is enabled, " \ |
| "otherwise the value remains constant. " \ |
| "In the latter case a value of 0 will result as frequent as " \ |
| "possible concurrent marking cycles. A value of 100 disables " \ |
| "concurrent marking. " \ |
| "Fragmentation waste in the old generation is not considered " \ |
| "free space in this calculation. (G1 collector only)") \ |
| range(0, 100) \ |
| \ |
| manageable(intx, CMSTriggerInterval, -1, \ |
| "Commence a CMS collection cycle (at least) every so many " \ |
| "milliseconds (0 permanently, -1 disabled)") \ |
| range(-1, max_intx) \ |
| \ |
| product(bool, UseCMSInitiatingOccupancyOnly, false, \ |
| "Only use occupancy as a criterion for starting a CMS collection")\ |
| \ |
| product(uintx, CMSIsTooFullPercentage, 98, \ |
| "An absolute ceiling above which CMS will always consider the " \ |
| "unloading of classes when class unloading is enabled") \ |
| range(0, 100) \ |
| \ |
| develop(bool, CMSTestInFreeList, false, \ |
| "Check if the coalesced range is already in the " \ |
| "free lists as claimed") \ |
| \ |
| notproduct(bool, CMSVerifyReturnedBytes, false, \ |
| "Check that all the garbage collected was returned to the " \ |
| "free lists") \ |
| \ |
| notproduct(bool, ScavengeALot, false, \ |
| "Force scavenge at every Nth exit from the runtime system " \ |
| "(N=ScavengeALotInterval)") \ |
| \ |
| develop(bool, FullGCALot, false, \ |
| "Force full gc at every Nth exit from the runtime system " \ |
| "(N=FullGCALotInterval)") \ |
| \ |
| notproduct(bool, GCALotAtAllSafepoints, false, \ |
| "Enforce ScavengeALot/GCALot at all potential safepoints") \ |
| \ |
| notproduct(bool, PromotionFailureALot, false, \ |
| "Use promotion failure handling on every youngest generation " \ |
| "collection") \ |
| \ |
| develop(uintx, PromotionFailureALotCount, 1000, \ |
| "Number of promotion failures occurring at PLAB " \ |
| "refill attempts (ParNew) or promotion attempts " \ |
| "(other young collectors)") \ |
| \ |
| develop(uintx, PromotionFailureALotInterval, 5, \ |
| "Total collections between promotion failures a lot") \ |
| \ |
| experimental(uintx, WorkStealingSleepMillis, 1, \ |
| "Sleep time when sleep is used for yields") \ |
| \ |
| experimental(uintx, WorkStealingYieldsBeforeSleep, 5000, \ |
| "Number of yields before a sleep is done during work stealing") \ |
| \ |
| experimental(uintx, WorkStealingHardSpins, 4096, \ |
| "Number of iterations in a spin loop between checks on " \ |
| "time out of hard spin") \ |
| \ |
| experimental(uintx, WorkStealingSpinToYieldRatio, 10, \ |
| "Ratio of hard spins to calls to yield") \ |
| \ |
| develop(uintx, ObjArrayMarkingStride, 2048, \ |
| "Number of object array elements to push onto the marking stack " \ |
| "before pushing a continuation entry") \ |
| \ |
| develop(bool, MetadataAllocationFailALot, false, \ |
| "Fail metadata allocations at intervals controlled by " \ |
| "MetadataAllocationFailALotInterval") \ |
| \ |
| develop(uintx, MetadataAllocationFailALotInterval, 1000, \ |
| "Metadata allocation failure a lot interval") \ |
| \ |
| notproduct(bool, ExecuteInternalVMTests, false, \ |
| "Enable execution of internal VM tests") \ |
| \ |
| notproduct(bool, VerboseInternalVMTests, false, \ |
| "Turn on logging for internal VM tests.") \ |
| \ |
| product(bool, ExecutingUnitTests, false, \ |
| "Whether the JVM is running unit tests or not") \ |
| \ |
| product_pd(bool, UseTLAB, "Use thread-local object allocation") \ |
| \ |
| product_pd(bool, ResizeTLAB, \ |
| "Dynamically resize TLAB size for threads") \ |
| \ |
| product(bool, ZeroTLAB, false, \ |
| "Zero out the newly created TLAB") \ |
| \ |
| product(bool, FastTLABRefill, true, \ |
| "Use fast TLAB refill code") \ |
| \ |
| product(bool, TLABStats, true, \ |
| "Provide more detailed and expensive TLAB statistics.") \ |
| \ |
| product_pd(bool, NeverActAsServerClassMachine, \ |
| "Never act like a server-class machine") \ |
| \ |
| product(bool, AlwaysActAsServerClassMachine, false, \ |
| "Always act like a server-class machine") \ |
| \ |
| product_pd(uint64_t, MaxRAM, \ |
| "Real memory size (in bytes) used to set maximum heap size") \ |
| range(0, 0XFFFFFFFFFFFFFFFF) \ |
| \ |
| product(size_t, ErgoHeapSizeLimit, 0, \ |
| "Maximum ergonomically set heap size (in bytes); zero means use " \ |
| "MaxRAM * MaxRAMPercentage / 100") \ |
| range(0, max_uintx) \ |
| \ |
| experimental(bool, UseCGroupMemoryLimitForHeap, false, \ |
| "Use CGroup memory limit as physical memory limit for heap " \ |
| "sizing") \ |
| \ |
| product(uintx, MaxRAMFraction, 4, \ |
| "Maximum fraction (1/n) of real memory used for maximum heap " \ |
| "size. " \ |
| "Deprecated, use MaxRAMPercentage instead") \ |
| range(1, max_uintx) \ |
| \ |
| product(uintx, MinRAMFraction, 2, \ |
| "Minimum fraction (1/n) of real memory used for maximum heap " \ |
| "size on systems with small physical memory size. " \ |
| "Deprecated, use MinRAMPercentage instead") \ |
| range(1, max_uintx) \ |
| \ |
| product(uintx, InitialRAMFraction, 64, \ |
| "Fraction (1/n) of real memory used for initial heap size. " \ |
| "Deprecated, use InitialRAMPercentage instead") \ |
| range(1, max_uintx) \ |
| \ |
| product(double, MaxRAMPercentage, 25.0, \ |
| "Maximum percentage of real memory used for maximum heap size") \ |
| range(0.0, 100.0) \ |
| \ |
| product(double, MinRAMPercentage, 50.0, \ |
| "Minimum percentage of real memory used for maximum heap" \ |
| "size on systems with small physical memory size") \ |
| range(0.0, 100.0) \ |
| \ |
| product(double, InitialRAMPercentage, 1.5625, \ |
| "Percentage of real memory used for initial heap size") \ |
| range(0.0, 100.0) \ |
| \ |
| develop(uintx, MaxVirtMemFraction, 2, \ |
| "Maximum fraction (1/n) of virtual memory used for ergonomically "\ |
| "determining maximum heap size") \ |
| \ |
| product(bool, UseAdaptiveSizePolicy, true, \ |
| "Use adaptive generation sizing policies") \ |
| \ |
| product(bool, UsePSAdaptiveSurvivorSizePolicy, true, \ |
| "Use adaptive survivor sizing policies") \ |
| \ |
| product(bool, UseAdaptiveGenerationSizePolicyAtMinorCollection, true, \ |
| "Use adaptive young-old sizing policies at minor collections") \ |
| \ |
| product(bool, UseAdaptiveGenerationSizePolicyAtMajorCollection, true, \ |
| "Use adaptive young-old sizing policies at major collections") \ |
| \ |
| product(bool, UseAdaptiveSizePolicyWithSystemGC, false, \ |
| "Include statistics from System.gc() for adaptive size policy") \ |
| \ |
| product(bool, UseAdaptiveGCBoundary, false, \ |
| "Allow young-old boundary to move") \ |
| \ |
| develop(intx, PSAdaptiveSizePolicyResizeVirtualSpaceAlot, -1, \ |
| "Resize the virtual spaces of the young or old generations") \ |
| range(-1, 1) \ |
| \ |
| product(uintx, AdaptiveSizeThroughPutPolicy, 0, \ |
| "Policy for changing generation size for throughput goals") \ |
| range(0, 1) \ |
| \ |
| product(uintx, AdaptiveSizePolicyInitializingSteps, 20, \ |
| "Number of steps where heuristics is used before data is used") \ |
| range(0, max_uintx) \ |
| \ |
| develop(uintx, AdaptiveSizePolicyReadyThreshold, 5, \ |
| "Number of collections before the adaptive sizing is started") \ |
| \ |
| product(uintx, AdaptiveSizePolicyOutputInterval, 0, \ |
| "Collection interval for printing information; zero means never") \ |
| range(0, max_uintx) \ |
| \ |
| product(bool, UseAdaptiveSizePolicyFootprintGoal, true, \ |
| "Use adaptive minimum footprint as a goal") \ |
| \ |
| product(uintx, AdaptiveSizePolicyWeight, 10, \ |
| "Weight given to exponential resizing, between 0 and 100") \ |
| range(0, 100) \ |
| \ |
| product(uintx, AdaptiveTimeWeight, 25, \ |
| "Weight given to time in adaptive policy, between 0 and 100") \ |
| range(0, 100) \ |
| \ |
| product(uintx, PausePadding, 1, \ |
| "How much buffer to keep for pause time") \ |
| range(0, max_juint) \ |
| \ |
| product(uintx, PromotedPadding, 3, \ |
| "How much buffer to keep for promotion failure") \ |
| range(0, max_juint) \ |
| \ |
| product(uintx, SurvivorPadding, 3, \ |
| "How much buffer to keep for survivor overflow") \ |
| range(0, max_juint) \ |
| \ |
| product(uintx, ThresholdTolerance, 10, \ |
| "Allowed collection cost difference between generations") \ |
| range(0, 100) \ |
| \ |
| product(uintx, AdaptiveSizePolicyCollectionCostMargin, 50, \ |
| "If collection costs are within margin, reduce both by full " \ |
| "delta") \ |
| range(0, 100) \ |
| \ |
| product(uintx, YoungGenerationSizeIncrement, 20, \ |
| "Adaptive size percentage change in young generation") \ |
| range(0, 100) \ |
| \ |
| product(uintx, YoungGenerationSizeSupplement, 80, \ |
| "Supplement to YoungedGenerationSizeIncrement used at startup") \ |
| range(0, 100) \ |
| \ |
| product(uintx, YoungGenerationSizeSupplementDecay, 8, \ |
| "Decay factor to YoungedGenerationSizeSupplement") \ |
| range(1, max_uintx) \ |
| \ |
| product(uintx, TenuredGenerationSizeIncrement, 20, \ |
| "Adaptive size percentage change in tenured generation") \ |
| range(0, 100) \ |
| \ |
| product(uintx, TenuredGenerationSizeSupplement, 80, \ |
| "Supplement to TenuredGenerationSizeIncrement used at startup") \ |
| range(0, 100) \ |
| \ |
| product(uintx, TenuredGenerationSizeSupplementDecay, 2, \ |
| "Decay factor to TenuredGenerationSizeIncrement") \ |
| range(1, max_uintx) \ |
| \ |
| product(uintx, MaxGCPauseMillis, max_uintx - 1, \ |
| "Adaptive size policy maximum GC pause time goal in millisecond, "\ |
| "or (G1 Only) the maximum GC time per MMU time slice") \ |
| range(1, max_uintx - 1) \ |
| constraint(MaxGCPauseMillisConstraintFunc,AfterErgo) \ |
| \ |
| product(uintx, GCPauseIntervalMillis, 0, \ |
| "Time slice for MMU specification") \ |
| constraint(GCPauseIntervalMillisConstraintFunc,AfterErgo) \ |
| \ |
| product(uintx, MaxGCMinorPauseMillis, max_uintx, \ |
| "Adaptive size policy maximum GC minor pause time goal " \ |
| "in millisecond") \ |
| range(0, max_uintx) \ |
| \ |
| product(uintx, GCTimeRatio, 99, \ |
| "Adaptive size policy application time to GC time ratio") \ |
| range(0, max_juint) \ |
| \ |
| product(uintx, AdaptiveSizeDecrementScaleFactor, 4, \ |
| "Adaptive size scale down factor for shrinking") \ |
| range(1, max_uintx) \ |
| \ |
| product(bool, UseAdaptiveSizeDecayMajorGCCost, true, \ |
| "Adaptive size decays the major cost for long major intervals") \ |
| \ |
| product(uintx, AdaptiveSizeMajorGCDecayTimeScale, 10, \ |
| "Time scale over which major costs decay") \ |
| range(0, max_uintx) \ |
| \ |
| product(uintx, MinSurvivorRatio, 3, \ |
| "Minimum ratio of young generation/survivor space size") \ |
| range(3, max_uintx) \ |
| \ |
| product(uintx, InitialSurvivorRatio, 8, \ |
| "Initial ratio of young generation/survivor space size") \ |
| range(0, max_uintx) \ |
| \ |
| product(size_t, BaseFootPrintEstimate, 256*M, \ |
| "Estimate of footprint other than Java Heap") \ |
| range(0, max_uintx) \ |
| \ |
| product(bool, UseGCOverheadLimit, true, \ |
| "Use policy to limit of proportion of time spent in GC " \ |
| "before an OutOfMemory error is thrown") \ |
| \ |
| product(uintx, GCTimeLimit, 98, \ |
| "Limit of the proportion of time spent in GC before " \ |
| "an OutOfMemoryError is thrown (used with GCHeapFreeLimit)") \ |
| range(0, 100) \ |
| \ |
| product(uintx, GCHeapFreeLimit, 2, \ |
| "Minimum percentage of free space after a full GC before an " \ |
| "OutOfMemoryError is thrown (used with GCTimeLimit)") \ |
| range(0, 100) \ |
| \ |
| develop(uintx, AdaptiveSizePolicyGCTimeLimitThreshold, 5, \ |
| "Number of consecutive collections before gc time limit fires") \ |
| range(1, max_uintx) \ |
| \ |
| product(intx, PrefetchCopyIntervalInBytes, -1, \ |
| "How far ahead to prefetch destination area (<= 0 means off)") \ |
| range(-1, max_jint) \ |
| \ |
| product(intx, PrefetchScanIntervalInBytes, -1, \ |
| "How far ahead to prefetch scan area (<= 0 means off)") \ |
| range(-1, max_jint) \ |
| \ |
| product(intx, PrefetchFieldsAhead, -1, \ |
| "How many fields ahead to prefetch in oop scan (<= 0 means off)") \ |
| range(-1, max_jint) \ |
| \ |
| diagnostic(bool, VerifyDuringStartup, false, \ |
| "Verify memory system before executing any Java code " \ |
| "during VM initialization") \ |
| \ |
| diagnostic(bool, VerifyBeforeExit, trueInDebug, \ |
| "Verify system before exiting") \ |
| \ |
| diagnostic(bool, VerifyBeforeGC, false, \ |
| "Verify memory system before GC") \ |
| \ |
| diagnostic(bool, VerifyAfterGC, false, \ |
| "Verify memory system after GC") \ |
| \ |
| diagnostic(bool, VerifyDuringGC, false, \ |
| "Verify memory system during GC (between phases)") \ |
| \ |
| diagnostic(ccstrlist, VerifySubSet, "", \ |
| "Memory sub-systems to verify when Verify*GC flag(s) " \ |
| "are enabled. One or more sub-systems can be specified " \ |
| "in a comma separated string. Sub-systems are: " \ |
| "threads, heap, symbol_table, string_table, codecache, " \ |
| "dictionary, classloader_data_graph, metaspace, jni_handles, " \ |
| "codecache_oops") \ |
| \ |
| diagnostic(bool, GCParallelVerificationEnabled, true, \ |
| "Enable parallel memory system verification") \ |
| \ |
| diagnostic(bool, DeferInitialCardMark, false, \ |
| "When +ReduceInitialCardMarks, explicitly defer any that " \ |
| "may arise from new_pre_store_barrier") \ |
| \ |
| product(bool, UseCondCardMark, false, \ |
| "Check for already marked card before updating card table") \ |
| \ |
| diagnostic(bool, VerifyRememberedSets, false, \ |
| "Verify GC remembered sets") \ |
| \ |
| diagnostic(bool, VerifyObjectStartArray, true, \ |
| "Verify GC object start array if verify before/after") \ |
| \ |
| product(bool, DisableExplicitGC, false, \ |
| "Ignore calls to System.gc()") \ |
| \ |
| notproduct(bool, CheckMemoryInitialization, false, \ |
| "Check memory initialization") \ |
| \ |
| diagnostic(bool, BindCMSThreadToCPU, false, \ |
| "Bind CMS Thread to CPU if possible") \ |
| \ |
| diagnostic(uintx, CPUForCMSThread, 0, \ |
| "When BindCMSThreadToCPU is true, the CPU to bind CMS thread to") \ |
| range(0, max_juint) \ |
| \ |
| product(bool, BindGCTaskThreadsToCPUs, false, \ |
| "Bind GCTaskThreads to CPUs if possible") \ |
| \ |
| product(bool, UseGCTaskAffinity, false, \ |
| "Use worker affinity when asking for GCTasks") \ |
| \ |
| product(uintx, ProcessDistributionStride, 4, \ |
| "Stride through processors when distributing processes") \ |
| range(0, max_juint) \ |
| \ |
| product(uintx, CMSCoordinatorYieldSleepCount, 10, \ |
| "Number of times the coordinator GC thread will sleep while " \ |
| "yielding before giving up and resuming GC") \ |
| range(0, max_juint) \ |
| \ |
| product(uintx, CMSYieldSleepCount, 0, \ |
| "Number of times a GC thread (minus the coordinator) " \ |
| "will sleep while yielding before giving up and resuming GC") \ |
| range(0, max_juint) \ |
| \ |
| product(bool, PrintGC, false, \ |
| "Print message at garbage collection. " \ |
| "Deprecated, use -Xlog:gc instead.") \ |
| \ |
| product(bool, PrintGCDetails, false, \ |
| "Print more details at garbage collection. " \ |
| "Deprecated, use -Xlog:gc* instead.") \ |
| \ |
| develop(intx, ConcGCYieldTimeout, 0, \ |
| "If non-zero, assert that GC threads yield within this " \ |
| "number of milliseconds") \ |
| range(0, max_intx) \ |
| \ |
| develop(bool, TraceFinalizerRegistration, false, \ |
| "Trace registration of final references") \ |
| \ |
| notproduct(bool, TraceScavenge, false, \ |
| "Trace scavenge") \ |
| \ |
| product(bool, IgnoreEmptyClassPaths, false, \ |
| "Ignore empty path elements in -classpath") \ |
| \ |
| product(size_t, InitialBootClassLoaderMetaspaceSize, \ |
| NOT_LP64(2200*K) LP64_ONLY(4*M), \ |
| "Initial size of the boot class loader data metaspace") \ |
| range(30*K, max_uintx/BytesPerWord) \ |
| constraint(InitialBootClassLoaderMetaspaceSizeConstraintFunc, AfterErgo)\ |
| \ |
| product(bool, TraceYoungGenTime, false, \ |
| "Trace accumulated time for young collection") \ |
| \ |
| product(bool, TraceOldGenTime, false, \ |
| "Trace accumulated time for old collection") \ |
| \ |
| product(bool, PrintHeapAtSIGBREAK, true, \ |
| "Print heap layout in response to SIGBREAK") \ |
| \ |
| manageable(bool, PrintClassHistogram, false, \ |
| "Print a histogram of class instances") \ |
| \ |
| develop(bool, TraceGCTaskManager, false, \ |
| "Trace actions of the GC task manager") \ |
| \ |
| develop(bool, TraceGCTaskQueue, false, \ |
| "Trace actions of the GC task queues") \ |
| \ |
| develop(bool, TraceParallelOldGCMarkingPhase, false, \ |
| "Trace marking phase in ParallelOldGC") \ |
| \ |
| develop(bool, TraceParallelOldGCDensePrefix, false, \ |
| "Trace dense prefix computation for ParallelOldGC") \ |
| \ |
| develop(bool, IgnoreLibthreadGPFault, false, \ |
| "Suppress workaround for libthread GP fault") \ |
| \ |
| experimental(double, ObjectCountCutOffPercent, 0.5, \ |
| "The percentage of the used heap that the instances of a class " \ |
| "must occupy for the class to generate a trace event") \ |
| range(0.0, 100.0) \ |
| \ |
| /* JVMTI heap profiling */ \ |
| \ |
| diagnostic(bool, TraceJVMTIObjectTagging, false, \ |
| "Trace JVMTI object tagging calls") \ |
| \ |
| diagnostic(bool, VerifyBeforeIteration, false, \ |
| "Verify memory system before JVMTI iteration") \ |
| \ |
| /* compiler interface */ \ |
| \ |
| develop(bool, CIPrintCompilerName, false, \ |
| "when CIPrint is active, print the name of the active compiler") \ |
| \ |
| diagnostic(bool, CIPrintCompileQueue, false, \ |
| "display the contents of the compile queue whenever a " \ |
| "compilation is enqueued") \ |
| \ |
| develop(bool, CIPrintRequests, false, \ |
| "display every request for compilation") \ |
| \ |
| product(bool, CITime, false, \ |
| "collect timing information for compilation") \ |
| \ |
| develop(bool, CITimeVerbose, false, \ |
| "be more verbose in compilation timings") \ |
| \ |
| develop(bool, CITimeEach, false, \ |
| "display timing information after each successful compilation") \ |
| \ |
| develop(bool, CICountOSR, false, \ |
| "use a separate counter when assigning ids to osr compilations") \ |
| \ |
| develop(bool, CICompileNatives, true, \ |
| "compile native methods if supported by the compiler") \ |
| \ |
| develop_pd(bool, CICompileOSR, \ |
| "compile on stack replacement methods if supported by the " \ |
| "compiler") \ |
| \ |
| develop(bool, CIPrintMethodCodes, false, \ |
| "print method bytecodes of the compiled code") \ |
| \ |
| develop(bool, CIPrintTypeFlow, false, \ |
| "print the results of ciTypeFlow analysis") \ |
| \ |
| develop(bool, CITraceTypeFlow, false, \ |
| "detailed per-bytecode tracing of ciTypeFlow analysis") \ |
| \ |
| develop(intx, OSROnlyBCI, -1, \ |
| "OSR only at this bci. Negative values mean exclude that bci") \ |
| \ |
| /* compiler */ \ |
| \ |
| /* notice: the max range value here is max_jint, not max_intx */ \ |
| /* because of overflow issue */ \ |
| product(intx, CICompilerCount, CI_COMPILER_COUNT, \ |
| "Number of compiler threads to run") \ |
| range(0, max_jint) \ |
| constraint(CICompilerCountConstraintFunc, AfterErgo) \ |
| \ |
| product(intx, CompilationPolicyChoice, 0, \ |
| "which compilation policy (0-3)") \ |
| range(0, 3) \ |
| \ |
| develop(bool, UseStackBanging, true, \ |
| "use stack banging for stack overflow checks (required for " \ |
| "proper StackOverflow handling; disable only to measure cost " \ |
| "of stackbanging)") \ |
| \ |
| develop(bool, UseStrictFP, true, \ |
| "use strict fp if modifier strictfp is set") \ |
| \ |
| develop(bool, GenerateSynchronizationCode, true, \ |
| "generate locking/unlocking code for synchronized methods and " \ |
| "monitors") \ |
| \ |
| develop(bool, GenerateRangeChecks, true, \ |
| "Generate range checks for array accesses") \ |
| \ |
| diagnostic_pd(bool, ImplicitNullChecks, \ |
| "Generate code for implicit null checks") \ |
| \ |
| product_pd(bool, TrapBasedNullChecks, \ |
| "Generate code for null checks that uses a cmp and trap " \ |
| "instruction raising SIGTRAP. This is only used if an access to" \ |
| "null (+offset) will not raise a SIGSEGV, i.e.," \ |
| "ImplicitNullChecks don't work (PPC64).") \ |
| \ |
| product(bool, PrintSafepointStatistics, false, \ |
| "Print statistics about safepoint synchronization") \ |
| \ |
| product(intx, PrintSafepointStatisticsCount, 300, \ |
| "Total number of safepoint statistics collected " \ |
| "before printing them out") \ |
| range(1, max_intx) \ |
| \ |
| product(intx, PrintSafepointStatisticsTimeout, -1, \ |
| "Print safepoint statistics only when safepoint takes " \ |
| "more than PrintSafepointSatisticsTimeout in millis") \ |
| LP64_ONLY(range(-1, max_intx/MICROUNITS)) \ |
| NOT_LP64(range(-1, max_intx)) \ |
| \ |
| product(bool, Inline, true, \ |
| "Enable inlining") \ |
| \ |
| product(bool, ClipInlining, true, \ |
| "Clip inlining if aggregate method exceeds DesiredMethodLimit") \ |
| \ |
| develop(bool, UseCHA, true, \ |
| "Enable CHA") \ |
| \ |
| product(bool, UseTypeProfile, true, \ |
| "Check interpreter profile for historically monomorphic calls") \ |
| \ |
| diagnostic(bool, PrintInlining, false, \ |
| "Print inlining optimizations") \ |
| \ |
| product(bool, UsePopCountInstruction, false, \ |
| "Use population count instruction") \ |
| \ |
| develop(bool, EagerInitialization, false, \ |
| "Eagerly initialize classes if possible") \ |
| \ |
| diagnostic(bool, LogTouchedMethods, false, \ |
| "Log methods which have been ever touched in runtime") \ |
| \ |
| diagnostic(bool, PrintTouchedMethodsAtExit, false, \ |
| "Print all methods that have been ever touched in runtime") \ |
| \ |
| develop(bool, TraceMethodReplacement, false, \ |
| "Print when methods are replaced do to recompilation") \ |
| \ |
| develop(bool, PrintMethodFlushing, false, \ |
| "Print the nmethods being flushed") \ |
| \ |
| diagnostic(bool, PrintMethodFlushingStatistics, false, \ |
| "print statistics about method flushing") \ |
| \ |
| diagnostic(intx, HotMethodDetectionLimit, 100000, \ |
| "Number of compiled code invocations after which " \ |
| "the method is considered as hot by the flusher") \ |
| range(1, max_jint) \ |
| \ |
| diagnostic(intx, MinPassesBeforeFlush, 10, \ |
| "Minimum number of sweeper passes before an nmethod " \ |
| "can be flushed") \ |
| range(0, max_intx) \ |
| \ |
| product(bool, UseCodeAging, true, \ |
| "Insert counter to detect warm methods") \ |
| \ |
| diagnostic(bool, StressCodeAging, false, \ |
| "Start with counters compiled in") \ |
| \ |
| develop(bool, StressCodeBuffers, false, \ |
| "Exercise code buffer expansion and other rare state changes") \ |
| \ |
| diagnostic(bool, DebugNonSafepoints, trueInDebug, \ |
| "Generate extra debugging information for non-safepoints in " \ |
| "nmethods") \ |
| \ |
| product(bool, PrintVMOptions, false, \ |
| "Print flags that appeared on the command line") \ |
| \ |
| product(bool, IgnoreUnrecognizedVMOptions, false, \ |
| "Ignore unrecognized VM options") \ |
| \ |
| product(bool, PrintCommandLineFlags, false, \ |
| "Print flags specified on command line or set by ergonomics") \ |
| \ |
| product(bool, PrintFlagsInitial, false, \ |
| "Print all VM flags before argument processing and exit VM") \ |
| \ |
| product(bool, PrintFlagsFinal, false, \ |
| "Print all VM flags after argument and ergonomic processing") \ |
| \ |
| notproduct(bool, PrintFlagsWithComments, false, \ |
| "Print all VM flags with default values and descriptions and " \ |
| "exit") \ |
| \ |
| product(bool, PrintFlagsRanges, false, \ |
| "Print VM flags and their ranges and exit VM") \ |
| \ |
| diagnostic(bool, SerializeVMOutput, true, \ |
| "Use a mutex to serialize output to tty and LogFile") \ |
| \ |
| diagnostic(bool, DisplayVMOutput, true, \ |
| "Display all VM output on the tty, independently of LogVMOutput") \ |
| \ |
| diagnostic(bool, LogVMOutput, false, \ |
| "Save VM output to LogFile") \ |
| \ |
| diagnostic(ccstr, LogFile, NULL, \ |
| "If LogVMOutput or LogCompilation is on, save VM output to " \ |
| "this file [default: ./hotspot_pid%p.log] (%p replaced with pid)")\ |
| \ |
| product(ccstr, ErrorFile, NULL, \ |
| "If an error occurs, save the error data to this file " \ |
| "[default: ./hs_err_pid%p.log] (%p replaced with pid)") \ |
| \ |
| product(bool, DisplayVMOutputToStderr, false, \ |
| "If DisplayVMOutput is true, display all VM output to stderr") \ |
| \ |
| product(bool, DisplayVMOutputToStdout, false, \ |
| "If DisplayVMOutput is true, display all VM output to stdout") \ |
| \ |
| product(bool, UseHeavyMonitors, false, \ |
| "use heavyweight instead of lightweight Java monitors") \ |
| \ |
| product(bool, PrintStringTableStatistics, false, \ |
| "print statistics about the StringTable and SymbolTable") \ |
| \ |
| diagnostic(bool, VerifyStringTableAtExit, false, \ |
| "verify StringTable contents at exit") \ |
| \ |
| notproduct(bool, PrintSymbolTableSizeHistogram, false, \ |
| "print histogram of the symbol table") \ |
| \ |
| notproduct(bool, ExitVMOnVerifyError, false, \ |
| "standard exit from VM if bytecode verify error " \ |
| "(only in debug mode)") \ |
| \ |
| diagnostic(ccstr, AbortVMOnException, NULL, \ |
| "Call fatal if this exception is thrown. Example: " \ |
| "java -XX:AbortVMOnException=java.lang.NullPointerException Foo") \ |
| \ |
| diagnostic(ccstr, AbortVMOnExceptionMessage, NULL, \ |
| "Call fatal if the exception pointed by AbortVMOnException " \ |
| "has this message") \ |
| \ |
| develop(bool, DebugVtables, false, \ |
| "add debugging code to vtable dispatch") \ |
| \ |
| notproduct(bool, PrintVtableStats, false, \ |
| "print vtables stats at end of run") \ |
| \ |
| develop(bool, TraceCreateZombies, false, \ |
| "trace creation of zombie nmethods") \ |
| \ |
| notproduct(bool, IgnoreLockingAssertions, false, \ |
| "disable locking assertions (for speed)") \ |
| \ |
| product(bool, RangeCheckElimination, true, \ |
| "Eliminate range checks") \ |
| \ |
| develop_pd(bool, UncommonNullCast, \ |
| "track occurrences of null in casts; adjust compiler tactics") \ |
| \ |
| develop(bool, TypeProfileCasts, true, \ |
| "treat casts like calls for purposes of type profiling") \ |
| \ |
| develop(bool, DelayCompilationDuringStartup, true, \ |
| "Delay invoking the compiler until main application class is " \ |
| "loaded") \ |
| \ |
| develop(bool, CompileTheWorld, false, \ |
| "Compile all methods in all classes in bootstrap class path " \ |
| "(stress test)") \ |
| \ |
| develop(bool, CompileTheWorldPreloadClasses, true, \ |
| "Preload all classes used by a class before start loading") \ |
| \ |
| notproduct(intx, CompileTheWorldSafepointInterval, 100, \ |
| "Force a safepoint every n compiles so sweeper can keep up") \ |
| \ |
| develop(bool, FillDelaySlots, true, \ |
| "Fill delay slots (on SPARC only)") \ |
| \ |
| develop(bool, TimeLivenessAnalysis, false, \ |
| "Time computation of bytecode liveness analysis") \ |
| \ |
| develop(bool, TraceLivenessGen, false, \ |
| "Trace the generation of liveness analysis information") \ |
| \ |
| notproduct(bool, TraceLivenessQuery, false, \ |
| "Trace queries of liveness analysis information") \ |
| \ |
| notproduct(bool, CollectIndexSetStatistics, false, \ |
| "Collect information about IndexSets") \ |
| \ |
| develop(bool, UseLoopSafepoints, true, \ |
| "Generate Safepoint nodes in every loop") \ |
| \ |
| develop(intx, FastAllocateSizeLimit, 128*K, \ |
| /* Note: This value is zero mod 1<<13 for a cheap sparc set. */ \ |
| "Inline allocations larger than this in doublewords must go slow")\ |
| \ |
| product(bool, AggressiveOpts, false, \ |
| "Enable aggressive optimizations - see arguments.cpp") \ |
| \ |
| product_pd(bool, CompactStrings, \ |
| "Enable Strings to use single byte chars in backing store") \ |
| \ |
| product_pd(uintx, TypeProfileLevel, \ |
| "=XYZ, with Z: Type profiling of arguments at call; " \ |
| "Y: Type profiling of return value at call; " \ |
| "X: Type profiling of parameters to methods; " \ |
| "X, Y and Z in 0=off ; 1=jsr292 only; 2=all methods") \ |
| constraint(TypeProfileLevelConstraintFunc, AfterErgo) \ |
| \ |
| product(intx, TypeProfileArgsLimit, 2, \ |
| "max number of call arguments to consider for type profiling") \ |
| range(0, 16) \ |
| \ |
| product(intx, TypeProfileParmsLimit, 2, \ |
| "max number of incoming parameters to consider for type profiling"\ |
| ", -1 for all") \ |
| range(-1, 64) \ |
| \ |
| /* statistics */ \ |
| develop(bool, CountCompiledCalls, false, \ |
| "Count method invocations") \ |
| \ |
| notproduct(bool, CountRuntimeCalls, false, \ |
| "Count VM runtime calls") \ |
| \ |
| develop(bool, CountJNICalls, false, \ |
| "Count jni method invocations") \ |
| \ |
| notproduct(bool, CountJVMCalls, false, \ |
| "Count jvm method invocations") \ |
| \ |
| notproduct(bool, CountRemovableExceptions, false, \ |
| "Count exceptions that could be replaced by branches due to " \ |
| "inlining") \ |
| \ |
| notproduct(bool, ICMissHistogram, false, \ |
| "Produce histogram of IC misses") \ |
| \ |
| /* interpreter */ \ |
| product_pd(bool, RewriteBytecodes, \ |
| "Allow rewriting of bytecodes (bytecodes are not immutable)") \ |
| \ |
| product_pd(bool, RewriteFrequentPairs, \ |
| "Rewrite frequently used bytecode pairs into a single bytecode") \ |
| \ |
| diagnostic(bool, PrintInterpreter, false, \ |
| "Print the generated interpreter code") \ |
| \ |
| product(bool, UseInterpreter, true, \ |
| "Use interpreter for non-compiled methods") \ |
| \ |
| develop(bool, UseFastSignatureHandlers, true, \ |
| "Use fast signature handlers for native calls") \ |
| \ |
| product(bool, UseLoopCounter, true, \ |
| "Increment invocation counter on backward branch") \ |
| \ |
| product_pd(bool, UseOnStackReplacement, \ |
| "Use on stack replacement, calls runtime if invoc. counter " \ |
| "overflows in loop") \ |
| \ |
| notproduct(bool, TraceOnStackReplacement, false, \ |
| "Trace on stack replacement") \ |
| \ |
| product_pd(bool, PreferInterpreterNativeStubs, \ |
| "Use always interpreter stubs for native methods invoked via " \ |
| "interpreter") \ |
| \ |
| develop(bool, CountBytecodes, false, \ |
| "Count number of bytecodes executed") \ |
| \ |
| develop(bool, PrintBytecodeHistogram, false, \ |
| "Print histogram of the executed bytecodes") \ |
| \ |
| develop(bool, PrintBytecodePairHistogram, false, \ |
| "Print histogram of the executed bytecode pairs") \ |
| \ |
| diagnostic(bool, PrintSignatureHandlers, false, \ |
| "Print code generated for native method signature handlers") \ |
| \ |
| develop(bool, VerifyOops, false, \ |
| "Do plausibility checks for oops") \ |
| \ |
| develop(bool, CheckUnhandledOops, false, \ |
| "Check for unhandled oops in VM code") \ |
| \ |
| develop(bool, VerifyJNIFields, trueInDebug, \ |
| "Verify jfieldIDs for instance fields") \ |
| \ |
| notproduct(bool, VerifyJNIEnvThread, false, \ |
| "Verify JNIEnv.thread == Thread::current() when entering VM " \ |
| "from JNI") \ |
| \ |
| develop(bool, VerifyFPU, false, \ |
| "Verify FPU state (check for NaN's, etc.)") \ |
| \ |
| develop(bool, VerifyThread, false, \ |
| "Watch the thread register for corruption (SPARC only)") \ |
| \ |
| develop(bool, VerifyActivationFrameSize, false, \ |
| "Verify that activation frame didn't become smaller than its " \ |
| "minimal size") \ |
| \ |
| develop(bool, TraceFrequencyInlining, false, \ |
| "Trace frequency based inlining") \ |
| \ |
| develop_pd(bool, InlineIntrinsics, \ |
| "Inline intrinsics that can be statically resolved") \ |
| \ |
| product_pd(bool, ProfileInterpreter, \ |
| "Profile at the bytecode level during interpretation") \ |
| \ |
| develop(bool, TraceProfileInterpreter, false, \ |
| "Trace profiling at the bytecode level during interpretation. " \ |
| "This outputs the profiling information collected to improve " \ |
| "jit compilation.") \ |
| \ |
| develop_pd(bool, ProfileTraps, \ |
| "Profile deoptimization traps at the bytecode level") \ |
| \ |
| product(intx, ProfileMaturityPercentage, 20, \ |
| "number of method invocations/branches (expressed as % of " \ |
| "CompileThreshold) before using the method's profile") \ |
| range(0, 100) \ |
| \ |
| diagnostic(bool, PrintMethodData, false, \ |
| "Print the results of +ProfileInterpreter at end of run") \ |
| \ |
| develop(bool, VerifyDataPointer, trueInDebug, \ |
| "Verify the method data pointer during interpreter profiling") \ |
| \ |
| develop(bool, VerifyCompiledCode, false, \ |
| "Include miscellaneous runtime verifications in nmethod code; " \ |
| "default off because it disturbs nmethod size heuristics") \ |
| \ |
| notproduct(bool, CrashGCForDumpingJavaThread, false, \ |
| "Manually make GC thread crash then dump java stack trace; " \ |
| "Test only") \ |
| \ |
| /* compilation */ \ |
| product(bool, UseCompiler, true, \ |
| "Use Just-In-Time compilation") \ |
| \ |
| develop(bool, TraceCompilationPolicy, false, \ |
| "Trace compilation policy") \ |
| \ |
| develop(bool, TimeCompilationPolicy, false, \ |
| "Time the compilation policy") \ |
| \ |
| product(bool, UseCounterDecay, true, \ |
| "Adjust recompilation counters") \ |
| \ |
| develop(intx, CounterHalfLifeTime, 30, \ |
| "Half-life time of invocation counters (in seconds)") \ |
| \ |
| develop(intx, CounterDecayMinIntervalLength, 500, \ |
| "The minimum interval (in milliseconds) between invocation of " \ |
| "CounterDecay") \ |
| \ |
| product(bool, AlwaysCompileLoopMethods, false, \ |
| "When using recompilation, never interpret methods " \ |
| "containing loops") \ |
| \ |
| product(bool, DontCompileHugeMethods, true, \ |
| "Do not compile methods > HugeMethodLimit") \ |
| \ |
| /* Bytecode escape analysis estimation. */ \ |
| product(bool, EstimateArgEscape, true, \ |
| "Analyze bytecodes to estimate escape state of arguments") \ |
| \ |
| product(intx, BCEATraceLevel, 0, \ |
| "How much tracing to do of bytecode escape analysis estimates " \ |
| "(0-3)") \ |
| range(0, 3) \ |
| \ |
| product(intx, MaxBCEAEstimateLevel, 5, \ |
| "Maximum number of nested calls that are analyzed by BC EA") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, MaxBCEAEstimateSize, 150, \ |
| "Maximum bytecode size of a method to be analyzed by BC EA") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, AllocatePrefetchStyle, 1, \ |
| "0 = no prefetch, " \ |
| "1 = generate prefetch instructions for each allocation, " \ |
| "2 = use TLAB watermark to gate allocation prefetch, " \ |
| "3 = generate one prefetch instruction per cache line") \ |
| range(0, 3) \ |
| \ |
| product(intx, AllocatePrefetchDistance, -1, \ |
| "Distance to prefetch ahead of allocation pointer. " \ |
| "-1: use system-specific value (automatically determined") \ |
| constraint(AllocatePrefetchDistanceConstraintFunc, AfterMemoryInit)\ |
| \ |
| product(intx, AllocatePrefetchLines, 3, \ |
| "Number of lines to prefetch ahead of array allocation pointer") \ |
| range(1, 64) \ |
| \ |
| product(intx, AllocateInstancePrefetchLines, 1, \ |
| "Number of lines to prefetch ahead of instance allocation " \ |
| "pointer") \ |
| range(1, 64) \ |
| \ |
| product(intx, AllocatePrefetchStepSize, 16, \ |
| "Step size in bytes of sequential prefetch instructions") \ |
| range(1, 512) \ |
| constraint(AllocatePrefetchStepSizeConstraintFunc,AfterMemoryInit)\ |
| \ |
| product(intx, AllocatePrefetchInstr, 0, \ |
| "Select instruction to prefetch ahead of allocation pointer") \ |
| constraint(AllocatePrefetchInstrConstraintFunc, AfterMemoryInit) \ |
| \ |
| /* deoptimization */ \ |
| develop(bool, TraceDeoptimization, false, \ |
| "Trace deoptimization") \ |
| \ |
| develop(bool, PrintDeoptimizationDetails, false, \ |
| "Print more information about deoptimization") \ |
| \ |
| develop(bool, DebugDeoptimization, false, \ |
| "Tracing various information while debugging deoptimization") \ |
| \ |
| product(intx, SelfDestructTimer, 0, \ |
| "Will cause VM to terminate after a given time (in minutes) " \ |
| "(0 means off)") \ |
| range(0, max_intx) \ |
| \ |
| product(intx, MaxJavaStackTraceDepth, 1024, \ |
| "The maximum number of lines in the stack trace for Java " \ |
| "exceptions (0 means all)") \ |
| range(0, max_jint/2) \ |
| \ |
| /* notice: the max range value here is max_jint, not max_intx */ \ |
| /* because of overflow issue */ \ |
| diagnostic(intx, GuaranteedSafepointInterval, 1000, \ |
| "Guarantee a safepoint (at least) every so many milliseconds " \ |
| "(0 means none)") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, SafepointTimeoutDelay, 10000, \ |
| "Delay in milliseconds for option SafepointTimeout") \ |
| LP64_ONLY(range(0, max_intx/MICROUNITS)) \ |
| NOT_LP64(range(0, max_intx)) \ |
| \ |
| product(intx, NmethodSweepActivity, 10, \ |
| "Removes cold nmethods from code cache if > 0. Higher values " \ |
| "result in more aggressive sweeping") \ |
| range(0, 2000) \ |
| \ |
| notproduct(bool, LogSweeper, false, \ |
| "Keep a ring buffer of sweeper activity") \ |
| \ |
| notproduct(intx, SweeperLogEntries, 1024, \ |
| "Number of records in the ring buffer of sweeper activity") \ |
| \ |
| notproduct(intx, MemProfilingInterval, 500, \ |
| "Time between each invocation of the MemProfiler") \ |
| \ |
| develop(intx, MallocCatchPtr, -1, \ |
| "Hit breakpoint when mallocing/freeing this pointer") \ |
| \ |
| notproduct(ccstrlist, SuppressErrorAt, "", \ |
| "List of assertions (file:line) to muzzle") \ |
| \ |
| develop(intx, StackPrintLimit, 100, \ |
| "number of stack frames to print in VM-level stack dump") \ |
| \ |
| notproduct(intx, MaxElementPrintSize, 256, \ |
| "maximum number of elements to print") \ |
| \ |
| notproduct(intx, MaxSubklassPrintSize, 4, \ |
| "maximum number of subklasses to print when printing klass") \ |
| \ |
| product(intx, MaxInlineLevel, 9, \ |
| "maximum number of nested calls that are inlined") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, MaxRecursiveInlineLevel, 1, \ |
| "maximum number of nested recursive calls that are inlined") \ |
| range(0, max_jint) \ |
| \ |
| develop(intx, MaxForceInlineLevel, 100, \ |
| "maximum number of nested calls that are forced for inlining " \ |
| "(using CompileCommand or marked w/ @ForceInline)") \ |
| range(0, max_jint) \ |
| \ |
| product_pd(intx, InlineSmallCode, \ |
| "Only inline already compiled methods if their code size is " \ |
| "less than this") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, MaxInlineSize, 35, \ |
| "The maximum bytecode size of a method to be inlined") \ |
| range(0, max_jint) \ |
| \ |
| product_pd(intx, FreqInlineSize, \ |
| "The maximum bytecode size of a frequent method to be inlined") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, MaxTrivialSize, 6, \ |
| "The maximum bytecode size of a trivial method to be inlined") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, MinInliningThreshold, 250, \ |
| "The minimum invocation count a method needs to have to be " \ |
| "inlined") \ |
| range(0, max_jint) \ |
| \ |
| develop(intx, MethodHistogramCutoff, 100, \ |
| "The cutoff value for method invocation histogram (+CountCalls)") \ |
| \ |
| diagnostic(intx, ProfilerNumberOfInterpretedMethods, 25, \ |
| "Number of interpreted methods to show in profile") \ |
| \ |
| diagnostic(intx, ProfilerNumberOfCompiledMethods, 25, \ |
| "Number of compiled methods to show in profile") \ |
| \ |
| diagnostic(intx, ProfilerNumberOfStubMethods, 25, \ |
| "Number of stub methods to show in profile") \ |
| \ |
| diagnostic(intx, ProfilerNumberOfRuntimeStubNodes, 25, \ |
| "Number of runtime stub nodes to show in profile") \ |
| \ |
| product(intx, ProfileIntervalsTicks, 100, \ |
| "Number of ticks between printing of interval profile " \ |
| "(+ProfileIntervals)") \ |
| range(0, max_intx) \ |
| \ |
| notproduct(intx, ScavengeALotInterval, 1, \ |
| "Interval between which scavenge will occur with +ScavengeALot") \ |
| \ |
| notproduct(intx, FullGCALotInterval, 1, \ |
| "Interval between which full gc will occur with +FullGCALot") \ |
| \ |
| notproduct(intx, FullGCALotStart, 0, \ |
| "For which invocation to start FullGCAlot") \ |
| \ |
| notproduct(intx, FullGCALotDummies, 32*K, \ |
| "Dummy object allocated with +FullGCALot, forcing all objects " \ |
| "to move") \ |
| \ |
| develop(intx, DontYieldALotInterval, 10, \ |
| "Interval between which yields will be dropped (milliseconds)") \ |
| \ |
| develop(intx, ProfilerPCTickThreshold, 15, \ |
| "Number of ticks in a PC buckets to be a hotspot") \ |
| \ |
| notproduct(intx, DeoptimizeALotInterval, 5, \ |
| "Number of exits until DeoptimizeALot kicks in") \ |
| \ |
| notproduct(intx, ZombieALotInterval, 5, \ |
| "Number of exits until ZombieALot kicks in") \ |
| \ |
| diagnostic(uintx, MallocMaxTestWords, 0, \ |
| "If non-zero, maximum number of words that malloc/realloc can " \ |
| "allocate (for testing only)") \ |
| range(0, max_uintx) \ |
| \ |
| product(intx, TypeProfileWidth, 2, \ |
| "Number of receiver types to record in call/cast profile") \ |
| range(0, 8) \ |
| \ |
| develop(intx, BciProfileWidth, 2, \ |
| "Number of return bci's to record in ret profile") \ |
| \ |
| product(intx, PerMethodRecompilationCutoff, 400, \ |
| "After recompiling N times, stay in the interpreter (-1=>'Inf')") \ |
| range(-1, max_intx) \ |
| \ |
| product(intx, PerBytecodeRecompilationCutoff, 200, \ |
| "Per-BCI limit on repeated recompilation (-1=>'Inf')") \ |
| range(-1, max_intx) \ |
| \ |
| product(intx, PerMethodTrapLimit, 100, \ |
| "Limit on traps (of one kind) in a method (includes inlines)") \ |
| range(0, max_jint) \ |
| \ |
| experimental(intx, PerMethodSpecTrapLimit, 5000, \ |
| "Limit on speculative traps (of one kind) in a method " \ |
| "(includes inlines)") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, PerBytecodeTrapLimit, 4, \ |
| "Limit on traps (of one kind) at a particular BCI") \ |
| range(0, max_jint) \ |
| \ |
| experimental(intx, SpecTrapLimitExtraEntries, 3, \ |
| "Extra method data trap entries for speculation") \ |
| \ |
| develop(intx, InlineFrequencyRatio, 20, \ |
| "Ratio of call site execution to caller method invocation") \ |
| range(0, max_jint) \ |
| \ |
| diagnostic_pd(intx, InlineFrequencyCount, \ |
| "Count of call site execution necessary to trigger frequent " \ |
| "inlining") \ |
| range(0, max_jint) \ |
| \ |
| develop(intx, InlineThrowCount, 50, \ |
| "Force inlining of interpreted methods that throw this often") \ |
| range(0, max_jint) \ |
| \ |
| develop(intx, InlineThrowMaxSize, 200, \ |
| "Force inlining of throwing methods smaller than this") \ |
| range(0, max_jint) \ |
| \ |
| develop(intx, ProfilerNodeSize, 1024, \ |
| "Size in K to allocate for the Profile Nodes of each thread") \ |
| range(0, 1024) \ |
| \ |
| /* gc parameters */ \ |
| product(size_t, InitialHeapSize, 0, \ |
| "Initial heap size (in bytes); zero means use ergonomics") \ |
| constraint(InitialHeapSizeConstraintFunc,AfterErgo) \ |
| \ |
| product(size_t, MaxHeapSize, ScaleForWordSize(96*M), \ |
| "Maximum heap size (in bytes)") \ |
| constraint(MaxHeapSizeConstraintFunc,AfterErgo) \ |
| \ |
| product(size_t, OldSize, ScaleForWordSize(4*M), \ |
| "Initial tenured generation size (in bytes)") \ |
| range(0, max_uintx) \ |
| \ |
| product(size_t, NewSize, ScaleForWordSize(1*M), \ |
| "Initial new generation size (in bytes)") \ |
| constraint(NewSizeConstraintFunc,AfterErgo) \ |
| \ |
| product(size_t, MaxNewSize, max_uintx, \ |
| "Maximum new generation size (in bytes), max_uintx means set " \ |
| "ergonomically") \ |
| range(0, max_uintx) \ |
| \ |
| product_pd(size_t, HeapBaseMinAddress, \ |
| "OS specific low limit for heap base address") \ |
| constraint(HeapBaseMinAddressConstraintFunc,AfterErgo) \ |
| \ |
| product(size_t, PretenureSizeThreshold, 0, \ |
| "Maximum size in bytes of objects allocated in DefNew " \ |
| "generation; zero means no maximum") \ |
| range(0, max_uintx) \ |
| \ |
| product(size_t, MinTLABSize, 2*K, \ |
| "Minimum allowed TLAB size (in bytes)") \ |
| range(1, max_uintx/2) \ |
| constraint(MinTLABSizeConstraintFunc,AfterMemoryInit) \ |
| \ |
| product(size_t, TLABSize, 0, \ |
| "Starting TLAB size (in bytes); zero means set ergonomically") \ |
| constraint(TLABSizeConstraintFunc,AfterMemoryInit) \ |
| \ |
| product(size_t, YoungPLABSize, 4096, \ |
| "Size of young gen promotion LAB's (in HeapWords)") \ |
| constraint(YoungPLABSizeConstraintFunc,AfterMemoryInit) \ |
| \ |
| product(size_t, OldPLABSize, 1024, \ |
| "Size of old gen promotion LAB's (in HeapWords), or Number " \ |
| "of blocks to attempt to claim when refilling CMS LAB's") \ |
| constraint(OldPLABSizeConstraintFunc,AfterMemoryInit) \ |
| \ |
| product(uintx, TLABAllocationWeight, 35, \ |
| "Allocation averaging weight") \ |
| range(0, 100) \ |
| \ |
| /* Limit the lower bound of this flag to 1 as it is used */ \ |
| /* in a division expression. */ \ |
| product(uintx, TLABWasteTargetPercent, 1, \ |
| "Percentage of Eden that can be wasted") \ |
| range(1, 100) \ |
| \ |
| product(uintx, TLABRefillWasteFraction, 64, \ |
| "Maximum TLAB waste at a refill (internal fragmentation)") \ |
| range(1, max_juint) \ |
| \ |
| product(uintx, TLABWasteIncrement, 4, \ |
| "Increment allowed waste at slow allocation") \ |
| range(0, max_jint) \ |
| constraint(TLABWasteIncrementConstraintFunc,AfterMemoryInit) \ |
| \ |
| product(uintx, SurvivorRatio, 8, \ |
| "Ratio of eden/survivor space size") \ |
| range(1, max_uintx-2) \ |
| constraint(SurvivorRatioConstraintFunc,AfterMemoryInit) \ |
| \ |
| product(uintx, NewRatio, 2, \ |
| "Ratio of old/new generation sizes") \ |
| range(0, max_uintx-1) \ |
| \ |
| product_pd(size_t, NewSizeThreadIncrease, \ |
| "Additional size added to desired new generation size per " \ |
| "non-daemon thread (in bytes)") \ |
| range(0, max_uintx) \ |
| \ |
| product_pd(size_t, MetaspaceSize, \ |
| "Initial threshold (in bytes) at which a garbage collection " \ |
| "is done to reduce Metaspace usage") \ |
| constraint(MetaspaceSizeConstraintFunc,AfterErgo) \ |
| \ |
| product(size_t, MaxMetaspaceSize, max_uintx, \ |
| "Maximum size of Metaspaces (in bytes)") \ |
| constraint(MaxMetaspaceSizeConstraintFunc,AfterErgo) \ |
| \ |
| product(size_t, CompressedClassSpaceSize, 1*G, \ |
| "Maximum size of class area in Metaspace when compressed " \ |
| "class pointers are used") \ |
| range(1*M, 3*G) \ |
| \ |
| manageable(uintx, MinHeapFreeRatio, 40, \ |
| "The minimum percentage of heap free after GC to avoid expansion."\ |
| " For most GCs this applies to the old generation. In G1 and" \ |
| " ParallelGC it applies to the whole heap.") \ |
| range(0, 100) \ |
| constraint(MinHeapFreeRatioConstraintFunc,AfterErgo) \ |
| \ |
| manageable(uintx, MaxHeapFreeRatio, 70, \ |
| "The maximum percentage of heap free after GC to avoid shrinking."\ |
| " For most GCs this applies to the old generation. In G1 and" \ |
| " ParallelGC it applies to the whole heap.") \ |
| range(0, 100) \ |
| constraint(MaxHeapFreeRatioConstraintFunc,AfterErgo) \ |
| \ |
| product(bool, ShrinkHeapInSteps, true, \ |
| "When disabled, informs the GC to shrink the java heap directly" \ |
| " to the target size at the next full GC rather than requiring" \ |
| " smaller steps during multiple full GCs.") \ |
| \ |
| product(intx, SoftRefLRUPolicyMSPerMB, 1000, \ |
| "Number of milliseconds per MB of free space in the heap") \ |
| range(0, max_intx) \ |
| constraint(SoftRefLRUPolicyMSPerMBConstraintFunc,AfterMemoryInit) \ |
| \ |
| product(size_t, MinHeapDeltaBytes, ScaleForWordSize(128*K), \ |
| "The minimum change in heap space due to GC (in bytes)") \ |
| range(0, max_uintx) \ |
| \ |
| product(size_t, MinMetaspaceExpansion, ScaleForWordSize(256*K), \ |
| "The minimum expansion of Metaspace (in bytes)") \ |
| range(0, max_uintx) \ |
| \ |
| product(uintx, MaxMetaspaceFreeRatio, 70, \ |
| "The maximum percentage of Metaspace free after GC to avoid " \ |
| "shrinking") \ |
| range(0, 100) \ |
| constraint(MaxMetaspaceFreeRatioConstraintFunc,AfterErgo) \ |
| \ |
| product(uintx, MinMetaspaceFreeRatio, 40, \ |
| "The minimum percentage of Metaspace free after GC to avoid " \ |
| "expansion") \ |
| range(0, 99) \ |
| constraint(MinMetaspaceFreeRatioConstraintFunc,AfterErgo) \ |
| \ |
| product(size_t, MaxMetaspaceExpansion, ScaleForWordSize(4*M), \ |
| "The maximum expansion of Metaspace without full GC (in bytes)") \ |
| range(0, max_uintx) \ |
| \ |
| product(uintx, QueuedAllocationWarningCount, 0, \ |
| "Number of times an allocation that queues behind a GC " \ |
| "will retry before printing a warning") \ |
| range(0, max_uintx) \ |
| \ |
| diagnostic(uintx, VerifyGCStartAt, 0, \ |
| "GC invoke count where +VerifyBefore/AfterGC kicks in") \ |
| range(0, max_uintx) \ |
| \ |
| diagnostic(intx, VerifyGCLevel, 0, \ |
| "Generation level at which to start +VerifyBefore/AfterGC") \ |
| range(0, 1) \ |
| \ |
| product(uintx, MaxTenuringThreshold, 15, \ |
| "Maximum value for tenuring threshold") \ |
| range(0, markOopDesc::max_age + 1) \ |
| constraint(MaxTenuringThresholdConstraintFunc,AfterErgo) \ |
| \ |
| product(uintx, InitialTenuringThreshold, 7, \ |
| "Initial value for tenuring threshold") \ |
| range(0, markOopDesc::max_age + 1) \ |
| constraint(InitialTenuringThresholdConstraintFunc,AfterErgo) \ |
| \ |
| product(uintx, TargetSurvivorRatio, 50, \ |
| "Desired percentage of survivor space used after scavenge") \ |
| range(0, 100) \ |
| \ |
| product(uintx, MarkSweepDeadRatio, 5, \ |
| "Percentage (0-100) of the old gen allowed as dead wood. " \ |
| "Serial mark sweep treats this as both the minimum and maximum " \ |
| "value. " \ |
| "CMS uses this value only if it falls back to mark sweep. " \ |
| "Par compact uses a variable scale based on the density of the " \ |
| "generation and treats this as the maximum value when the heap " \ |
| "is either completely full or completely empty. Par compact " \ |
| "also has a smaller default value; see arguments.cpp.") \ |
| range(0, 100) \ |
| \ |
| product(uint, MarkSweepAlwaysCompactCount, 4, \ |
| "How often should we fully compact the heap (ignoring the dead " \ |
| "space parameters)") \ |
| range(1, max_juint) \ |
| \ |
| develop(uintx, GCExpandToAllocateDelayMillis, 0, \ |
| "Delay between expansion and allocation (in milliseconds)") \ |
| \ |
| develop(uintx, GCWorkerDelayMillis, 0, \ |
| "Delay in scheduling GC workers (in milliseconds)") \ |
| \ |
| product(intx, DeferThrSuspendLoopCount, 4000, \ |
| "(Unstable) Number of times to iterate in safepoint loop " \ |
| "before blocking VM threads ") \ |
| range(-1, max_jint-1) \ |
| \ |
| product(intx, DeferPollingPageLoopCount, -1, \ |
| "(Unsafe,Unstable) Number of iterations in safepoint loop " \ |
| "before changing safepoint polling page to RO ") \ |
| range(-1, max_jint-1) \ |
| \ |
| product(intx, SafepointSpinBeforeYield, 2000, "(Unstable)") \ |
| range(0, max_intx) \ |
| \ |
| product(bool, PSChunkLargeArrays, true, \ |
| "Process large arrays in chunks") \ |
| \ |
| product(uintx, GCDrainStackTargetSize, 64, \ |
| "Number of entries we will try to leave on the stack " \ |
| "during parallel gc") \ |
| range(0, max_juint) \ |
| \ |
| /* stack parameters */ \ |
| product_pd(intx, StackYellowPages, \ |
| "Number of yellow zone (recoverable overflows) pages of size " \ |
| "4KB. If pages are bigger yellow zone is aligned up.") \ |
| range(MIN_STACK_YELLOW_PAGES, (DEFAULT_STACK_YELLOW_PAGES+5)) \ |
| \ |
| product_pd(intx, StackRedPages, \ |
| "Number of red zone (unrecoverable overflows) pages of size " \ |
| "4KB. If pages are bigger red zone is aligned up.") \ |
| range(MIN_STACK_RED_PAGES, (DEFAULT_STACK_RED_PAGES+2)) \ |
| \ |
| product_pd(intx, StackReservedPages, \ |
| "Number of reserved zone (reserved to annotated methods) pages" \ |
| " of size 4KB. If pages are bigger reserved zone is aligned up.") \ |
| range(MIN_STACK_RESERVED_PAGES, (DEFAULT_STACK_RESERVED_PAGES+10))\ |
| \ |
| product(bool, RestrictReservedStack, true, \ |
| "Restrict @ReservedStackAccess to trusted classes") \ |
| \ |
| /* greater stack shadow pages can't generate instruction to bang stack */ \ |
| product_pd(intx, StackShadowPages, \ |
| "Number of shadow zone (for overflow checking) pages of size " \ |
| "4KB. If pages are bigger shadow zone is aligned up. " \ |
| "This should exceed the depth of the VM and native call stack.") \ |
| range(MIN_STACK_SHADOW_PAGES, (DEFAULT_STACK_SHADOW_PAGES+30)) \ |
| \ |
| product_pd(intx, ThreadStackSize, \ |
| "Thread Stack Size (in Kbytes)") \ |
| range(0, 1 * M) \ |
| \ |
| product_pd(intx, VMThreadStackSize, \ |
| "Non-Java Thread Stack Size (in Kbytes)") \ |
| range(0, max_intx/(1 * K)) \ |
| \ |
| product_pd(intx, CompilerThreadStackSize, \ |
| "Compiler Thread Stack Size (in Kbytes)") \ |
| range(0, max_intx/(1 * K)) \ |
| \ |
| develop_pd(size_t, JVMInvokeMethodSlack, \ |
| "Stack space (bytes) required for JVM_InvokeMethod to complete") \ |
| \ |
| /* code cache parameters */ \ |
| develop_pd(uintx, CodeCacheSegmentSize, \ |
| "Code cache segment size (in bytes) - smallest unit of " \ |
| "allocation") \ |
| range(1, 1024) \ |
| constraint(CodeCacheSegmentSizeConstraintFunc, AfterErgo) \ |
| \ |
| develop_pd(intx, CodeEntryAlignment, \ |
| "Code entry alignment for generated code (in bytes)") \ |
| constraint(CodeEntryAlignmentConstraintFunc, AfterErgo) \ |
| \ |
| product_pd(intx, OptoLoopAlignment, \ |
| "Align inner loops to zero relative to this modulus") \ |
| range(1, 16) \ |
| constraint(OptoLoopAlignmentConstraintFunc, AfterErgo) \ |
| \ |
| product_pd(uintx, InitialCodeCacheSize, \ |
| "Initial code cache size (in bytes)") \ |
| range(0, max_uintx) \ |
| \ |
| develop_pd(uintx, CodeCacheMinimumUseSpace, \ |
| "Minimum code cache size (in bytes) required to start VM.") \ |
| range(0, max_uintx) \ |
| \ |
| product(bool, SegmentedCodeCache, false, \ |
| "Use a segmented code cache") \ |
| \ |
| product_pd(uintx, ReservedCodeCacheSize, \ |
| "Reserved code cache size (in bytes) - maximum code cache size") \ |
| range(0, max_uintx) \ |
| \ |
| product_pd(uintx, NonProfiledCodeHeapSize, \ |
| "Size of code heap with non-profiled methods (in bytes)") \ |
| range(0, max_uintx) \ |
| \ |
| product_pd(uintx, ProfiledCodeHeapSize, \ |
| "Size of code heap with profiled methods (in bytes)") \ |
| range(0, max_uintx) \ |
| \ |
| product_pd(uintx, NonNMethodCodeHeapSize, \ |
| "Size of code heap with non-nmethods (in bytes)") \ |
| range(0, max_uintx) \ |
| \ |
| product_pd(uintx, CodeCacheExpansionSize, \ |
| "Code cache expansion size (in bytes)") \ |
| range(0, max_uintx) \ |
| \ |
| diagnostic_pd(uintx, CodeCacheMinBlockLength, \ |
| "Minimum number of segments in a code cache block") \ |
| range(1, 100) \ |
| \ |
| notproduct(bool, ExitOnFullCodeCache, false, \ |
| "Exit the VM if we fill the code cache") \ |
| \ |
| product(bool, UseCodeCacheFlushing, true, \ |
| "Remove cold/old nmethods from the code cache") \ |
| \ |
| product(uintx, StartAggressiveSweepingAt, 10, \ |
| "Start aggressive sweeping if X[%] of the code cache is free." \ |
| "Segmented code cache: X[%] of the non-profiled heap." \ |
| "Non-segmented code cache: X[%] of the total code cache") \ |
| range(0, 100) \ |
| \ |
| /* AOT parameters */ \ |
| product(bool, UseAOT, AOT_ONLY(true) NOT_AOT(false), \ |
| "Use AOT compiled files") \ |
| \ |
| product(ccstrlist, AOTLibrary, NULL, \ |
| "AOT library") \ |
| \ |
| product(bool, PrintAOT, false, \ |
| "Print used AOT klasses and methods") \ |
| \ |
| notproduct(bool, PrintAOTStatistics, false, \ |
| "Print AOT statistics") \ |
| \ |
| diagnostic(bool, UseAOTStrictLoading, false, \ |
| "Exit the VM if any of the AOT libraries has invalid config") \ |
| \ |
| product(bool, CalculateClassFingerprint, false, \ |
| "Calculate class fingerprint") \ |
| \ |
| /* interpreter debugging */ \ |
| develop(intx, BinarySwitchThreshold, 5, \ |
| "Minimal number of lookupswitch entries for rewriting to binary " \ |
| "switch") \ |
| \ |
| develop(intx, StopInterpreterAt, 0, \ |
| "Stop interpreter execution at specified bytecode number") \ |
| \ |
| develop(intx, TraceBytecodesAt, 0, \ |
| "Trace bytecodes starting with specified bytecode number") \ |
| \ |
| /* compiler interface */ \ |
| develop(intx, CIStart, 0, \ |
| "The id of the first compilation to permit") \ |
| \ |
| develop(intx, CIStop, max_jint, \ |
| "The id of the last compilation to permit") \ |
| \ |
| develop(intx, CIStartOSR, 0, \ |
| "The id of the first osr compilation to permit " \ |
| "(CICountOSR must be on)") \ |
| \ |
| develop(intx, CIStopOSR, max_jint, \ |
| "The id of the last osr compilation to permit " \ |
| "(CICountOSR must be on)") \ |
| \ |
| develop(intx, CIBreakAtOSR, -1, \ |
| "The id of osr compilation to break at") \ |
| \ |
| develop(intx, CIBreakAt, -1, \ |
| "The id of compilation to break at") \ |
| \ |
| product(ccstrlist, CompileOnly, "", \ |
| "List of methods (pkg/class.name) to restrict compilation to") \ |
| \ |
| product(ccstr, CompileCommandFile, NULL, \ |
| "Read compiler commands from this file [.hotspot_compiler]") \ |
| \ |
| diagnostic(ccstr, CompilerDirectivesFile, NULL, \ |
| "Read compiler directives from this file") \ |
| \ |
| product(ccstrlist, CompileCommand, "", \ |
| "Prepend to .hotspot_compiler; e.g. log,java/lang/String.<init>") \ |
| \ |
| develop(bool, ReplayCompiles, false, \ |
| "Enable replay of compilations from ReplayDataFile") \ |
| \ |
| product(ccstr, ReplayDataFile, NULL, \ |
| "File containing compilation replay information" \ |
| "[default: ./replay_pid%p.log] (%p replaced with pid)") \ |
| \ |
| product(ccstr, InlineDataFile, NULL, \ |
| "File containing inlining replay information" \ |
| "[default: ./inline_pid%p.log] (%p replaced with pid)") \ |
| \ |
| develop(intx, ReplaySuppressInitializers, 2, \ |
| "Control handling of class initialization during replay: " \ |
| "0 - don't do anything special; " \ |
| "1 - treat all class initializers as empty; " \ |
| "2 - treat class initializers for application classes as empty; " \ |
| "3 - allow all class initializers to run during bootstrap but " \ |
| " pretend they are empty after starting replay") \ |
| range(0, 3) \ |
| \ |
| develop(bool, ReplayIgnoreInitErrors, false, \ |
| "Ignore exceptions thrown during initialization for replay") \ |
| \ |
| product(bool, DumpReplayDataOnError, true, \ |
| "Record replay data for crashing compiler threads") \ |
| \ |
| product(bool, CICompilerCountPerCPU, false, \ |
| "1 compiler thread for log(N CPUs)") \ |
| \ |
| develop(intx, CIFireOOMAt, -1, \ |
| "Fire OutOfMemoryErrors throughout CI for testing the compiler " \ |
| "(non-negative value throws OOM after this many CI accesses " \ |
| "in each compile)") \ |
| notproduct(intx, CICrashAt, -1, \ |
| "id of compilation to trigger assert in compiler thread for " \ |
| "the purpose of testing, e.g. generation of replay data") \ |
| notproduct(bool, CIObjectFactoryVerify, false, \ |
| "enable potentially expensive verification in ciObjectFactory") \ |
| \ |
| /* Priorities */ \ |
| product_pd(bool, UseThreadPriorities, "Use native thread priorities") \ |
| \ |
| product(intx, ThreadPriorityPolicy, 0, \ |
| "0 : Normal. "\ |
| " VM chooses priorities that are appropriate for normal "\ |
| " applications. On Solaris NORM_PRIORITY and above are mapped "\ |
| " to normal native priority. Java priorities below " \ |
| " NORM_PRIORITY map to lower native priority values. On "\ |
| " Windows applications are allowed to use higher native "\ |
| " priorities. However, with ThreadPriorityPolicy=0, VM will "\ |
| " not use the highest possible native priority, "\ |
| " THREAD_PRIORITY_TIME_CRITICAL, as it may interfere with "\ |
| " system threads. On Linux thread priorities are ignored "\ |
| " because the OS does not support static priority in "\ |
| " SCHED_OTHER scheduling class which is the only choice for "\ |
| " non-root, non-realtime applications. "\ |
| "1 : Aggressive. "\ |
| " Java thread priorities map over to the entire range of "\ |
| " native thread priorities. Higher Java thread priorities map "\ |
| " to higher native thread priorities. This policy should be "\ |
| " used with care, as sometimes it can cause performance "\ |
| " degradation in the application and/or the entire system. On "\ |
| " Linux this policy requires root privilege.") \ |
| range(0, 1) \ |
| \ |
| product(bool, ThreadPriorityVerbose, false, \ |
| "Print priority changes") \ |
| \ |
| product(intx, CompilerThreadPriority, -1, \ |
| "The native priority at which compiler threads should run " \ |
| "(-1 means no change)") \ |
| range(min_jint, max_jint) \ |
| constraint(CompilerThreadPriorityConstraintFunc, AfterErgo) \ |
| \ |
| product(intx, VMThreadPriority, -1, \ |
| "The native priority at which the VM thread should run " \ |
| "(-1 means no change)") \ |
| range(-1, 127) \ |
| \ |
| product(bool, CompilerThreadHintNoPreempt, true, \ |
| "(Solaris only) Give compiler threads an extra quanta") \ |
| \ |
| product(bool, VMThreadHintNoPreempt, false, \ |
| "(Solaris only) Give VM thread an extra quanta") \ |
| \ |
| product(intx, JavaPriority1_To_OSPriority, -1, \ |
| "Map Java priorities to OS priorities") \ |
| range(-1, 127) \ |
| \ |
| product(intx, JavaPriority2_To_OSPriority, -1, \ |
| "Map Java priorities to OS priorities") \ |
| range(-1, 127) \ |
| \ |
| product(intx, JavaPriority3_To_OSPriority, -1, \ |
| "Map Java priorities to OS priorities") \ |
| range(-1, 127) \ |
| \ |
| product(intx, JavaPriority4_To_OSPriority, -1, \ |
| "Map Java priorities to OS priorities") \ |
| range(-1, 127) \ |
| \ |
| product(intx, JavaPriority5_To_OSPriority, -1, \ |
| "Map Java priorities to OS priorities") \ |
| range(-1, 127) \ |
| \ |
| product(intx, JavaPriority6_To_OSPriority, -1, \ |
| "Map Java priorities to OS priorities") \ |
| range(-1, 127) \ |
| \ |
| product(intx, JavaPriority7_To_OSPriority, -1, \ |
| "Map Java priorities to OS priorities") \ |
| range(-1, 127) \ |
| \ |
| product(intx, JavaPriority8_To_OSPriority, -1, \ |
| "Map Java priorities to OS priorities") \ |
| range(-1, 127) \ |
| \ |
| product(intx, JavaPriority9_To_OSPriority, -1, \ |
| "Map Java priorities to OS priorities") \ |
| range(-1, 127) \ |
| \ |
| product(intx, JavaPriority10_To_OSPriority,-1, \ |
| "Map Java priorities to OS priorities") \ |
| range(-1, 127) \ |
| \ |
| experimental(bool, UseCriticalJavaThreadPriority, false, \ |
| "Java thread priority 10 maps to critical scheduling priority") \ |
| \ |
| experimental(bool, UseCriticalCompilerThreadPriority, false, \ |
| "Compiler thread(s) run at critical scheduling priority") \ |
| \ |
| experimental(bool, UseCriticalCMSThreadPriority, false, \ |
| "ConcurrentMarkSweep thread runs at critical scheduling priority")\ |
| \ |
| /* compiler debugging */ \ |
| notproduct(intx, CompileTheWorldStartAt, 1, \ |
| "First class to consider when using +CompileTheWorld") \ |
| \ |
| notproduct(intx, CompileTheWorldStopAt, max_jint, \ |
| "Last class to consider when using +CompileTheWorld") \ |
| \ |
| develop(intx, NewCodeParameter, 0, \ |
| "Testing Only: Create a dedicated integer parameter before " \ |
| "putback") \ |
| \ |
| /* new oopmap storage allocation */ \ |
| develop(intx, MinOopMapAllocation, 8, \ |
| "Minimum number of OopMap entries in an OopMapSet") \ |
| \ |
| /* Background Compilation */ \ |
| develop(intx, LongCompileThreshold, 50, \ |
| "Used with +TraceLongCompiles") \ |
| \ |
| /* recompilation */ \ |
| product_pd(intx, CompileThreshold, \ |
| "number of interpreted method invocations before (re-)compiling") \ |
| constraint(CompileThresholdConstraintFunc, AfterErgo) \ |
| \ |
| product(double, CompileThresholdScaling, 1.0, \ |
| "Factor to control when first compilation happens " \ |
| "(both with and without tiered compilation): " \ |
| "values greater than 1.0 delay counter overflow, " \ |
| "values between 0 and 1.0 rush counter overflow, " \ |
| "value of 1.0 leaves compilation thresholds unchanged " \ |
| "value of 0.0 is equivalent to -Xint. " \ |
| "" \ |
| "Flag can be set as per-method option. " \ |
| "If a value is specified for a method, compilation thresholds " \ |
| "for that method are scaled by both the value of the global flag "\ |
| "and the value of the per-method flag.") \ |
| range(0.0, DBL_MAX) \ |
| \ |
| product(intx, Tier0InvokeNotifyFreqLog, 7, \ |
| "Interpreter (tier 0) invocation notification frequency") \ |
| range(0, 30) \ |
| \ |
| product(intx, Tier2InvokeNotifyFreqLog, 11, \ |
| "C1 without MDO (tier 2) invocation notification frequency") \ |
| range(0, 30) \ |
| \ |
| product(intx, Tier3InvokeNotifyFreqLog, 10, \ |
| "C1 with MDO profiling (tier 3) invocation notification " \ |
| "frequency") \ |
| range(0, 30) \ |
| \ |
| product(intx, Tier23InlineeNotifyFreqLog, 20, \ |
| "Inlinee invocation (tiers 2 and 3) notification frequency") \ |
| range(0, 30) \ |
| \ |
| product(intx, Tier0BackedgeNotifyFreqLog, 10, \ |
| "Interpreter (tier 0) invocation notification frequency") \ |
| range(0, 30) \ |
| \ |
| product(intx, Tier2BackedgeNotifyFreqLog, 14, \ |
| "C1 without MDO (tier 2) invocation notification frequency") \ |
| range(0, 30) \ |
| \ |
| product(intx, Tier3BackedgeNotifyFreqLog, 13, \ |
| "C1 with MDO profiling (tier 3) invocation notification " \ |
| "frequency") \ |
| range(0, 30) \ |
| \ |
| product(intx, Tier2CompileThreshold, 0, \ |
| "threshold at which tier 2 compilation is invoked") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier2BackEdgeThreshold, 0, \ |
| "Back edge threshold at which tier 2 compilation is invoked") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier3InvocationThreshold, 200, \ |
| "Compile if number of method invocations crosses this " \ |
| "threshold") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier3MinInvocationThreshold, 100, \ |
| "Minimum invocation to compile at tier 3") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier3CompileThreshold, 2000, \ |
| "Threshold at which tier 3 compilation is invoked (invocation " \ |
| "minimum must be satisfied)") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier3BackEdgeThreshold, 60000, \ |
| "Back edge threshold at which tier 3 OSR compilation is invoked") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier3AOTInvocationThreshold, 10000, \ |
| "Compile if number of method invocations crosses this " \ |
| "threshold if coming from AOT") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier3AOTMinInvocationThreshold, 1000, \ |
| "Minimum invocation to compile at tier 3 if coming from AOT") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier3AOTCompileThreshold, 15000, \ |
| "Threshold at which tier 3 compilation is invoked (invocation " \ |
| "minimum must be satisfied) if coming from AOT") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier3AOTBackEdgeThreshold, 120000, \ |
| "Back edge threshold at which tier 3 OSR compilation is invoked " \ |
| "if coming from AOT") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier4InvocationThreshold, 5000, \ |
| "Compile if number of method invocations crosses this " \ |
| "threshold") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier4MinInvocationThreshold, 600, \ |
| "Minimum invocation to compile at tier 4") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier4CompileThreshold, 15000, \ |
| "Threshold at which tier 4 compilation is invoked (invocation " \ |
| "minimum must be satisfied") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier4BackEdgeThreshold, 40000, \ |
| "Back edge threshold at which tier 4 OSR compilation is invoked") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier3DelayOn, 5, \ |
| "If C2 queue size grows over this amount per compiler thread " \ |
| "stop compiling at tier 3 and start compiling at tier 2") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier3DelayOff, 2, \ |
| "If C2 queue size is less than this amount per compiler thread " \ |
| "allow methods compiled at tier 2 transition to tier 3") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier3LoadFeedback, 5, \ |
| "Tier 3 thresholds will increase twofold when C1 queue size " \ |
| "reaches this amount per compiler thread") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, Tier4LoadFeedback, 3, \ |
| "Tier 4 thresholds will increase twofold when C2 queue size " \ |
| "reaches this amount per compiler thread") \ |
| range(0, max_jint) \ |
| \ |
| product(intx, TieredCompileTaskTimeout, 50, \ |
| "Kill compile task if method was not used within " \ |
| "given timeout in milliseconds") \ |
| range(0, max_intx) \ |
| \ |
| product(intx, TieredStopAtLevel, 4, \ |
| "Stop at given compilation level") \ |
| range(0, 4) \ |
| \ |
| product(intx, Tier0ProfilingStartPercentage, 200, \ |
| "Start profiling in interpreter if the counters exceed tier 3 " \ |
| "thresholds by the specified percentage") \ |
| range(0, max_jint) \ |
| \ |
| product(uintx, IncreaseFirstTierCompileThresholdAt, 50, \ |
| "Increase the compile threshold for C1 compilation if the code " \ |
| "cache is filled by the specified percentage") \ |
| range(0, 99) \ |
| \ |
| product(intx, TieredRateUpdateMinTime, 1, \ |
| "Minimum rate sampling interval (in milliseconds)") \ |
| range(0, max_intx) \ |
| \ |
| product(intx, TieredRateUpdateMaxTime, 25, \ |
| "Maximum rate sampling interval (in milliseconds)") \ |
| range(0, max_intx) \ |
| \ |
| product_pd(bool, TieredCompilation, \ |
| "Enable tiered compilation") \ |
| \ |
| product(bool, PrintTieredEvents, false, \ |
| "Print tiered events notifications") \ |
| \ |
| product_pd(intx, OnStackReplacePercentage, \ |
| "NON_TIERED number of method invocations/branches (expressed as " \ |
| "% of CompileThreshold) before (re-)compiling OSR code") \ |
| constraint(OnStackReplacePercentageConstraintFunc, AfterErgo) \ |
| \ |
| product(intx, InterpreterProfilePercentage, 33, \ |
| "NON_TIERED number of method invocations/branches (expressed as " \ |
| "% of CompileThreshold) before profiling in the interpreter") \ |
| range(0, 100) \ |
| \ |
| develop(intx, MaxRecompilationSearchLength, 10, \ |
| "The maximum number of frames to inspect when searching for " \ |
| "recompilee") \ |
| \ |
| develop(intx, MaxInterpretedSearchLength, 3, \ |
| "The maximum number of interpreted frames to skip when searching "\ |
| "for recompilee") \ |
| \ |
| develop(intx, DesiredMethodLimit, 8000, \ |
| "The desired maximum method size (in bytecodes) after inlining") \ |
| \ |
| develop(intx, HugeMethodLimit, 8000, \ |
| "Don't compile methods larger than this if " \ |
| "+DontCompileHugeMethods") \ |
| \ |
| /* New JDK 1.4 reflection implementation */ \ |
| \ |
| develop(intx, FastSuperclassLimit, 8, \ |
| "Depth of hardwired instanceof accelerator array") \ |
| \ |
| /* Properties for Java libraries */ \ |
| \ |
| product(uint64_t, MaxDirectMemorySize, 0, \ |
| "Maximum total size of NIO direct-buffer allocations") \ |
| range(0, max_jlong) \ |
| \ |
| /* Flags used for temporary code during development */ \ |
| \ |
| diagnostic(bool, UseNewCode, false, \ |
| "Testing Only: Use the new version while testing") \ |
| \ |
| diagnostic(bool, UseNewCode2, false, \ |
| "Testing Only: Use the new version while testing") \ |
| \ |
| diagnostic(bool, UseNewCode3, false, \ |
| "Testing Only: Use the new version while testing") \ |
| \ |
| /* flags for performance data collection */ \ |
| \ |
| product(bool, UsePerfData, true, \ |
| "Flag to disable jvmstat instrumentation for performance testing "\ |
| "and problem isolation purposes") \ |
| \ |
| product(bool, PerfDataSaveToFile, false, \ |
| "Save PerfData memory to hsperfdata_<pid> file on exit") \ |
| \ |
| product(ccstr, PerfDataSaveFile, NULL, \ |
| "Save PerfData memory to the specified absolute pathname. " \ |
| "The string %p in the file name (if present) " \ |
| "will be replaced by pid") \ |
| \ |
| product(intx, PerfDataSamplingInterval, 50, \ |
| "Data sampling interval (in milliseconds)") \ |
| range(PeriodicTask::min_interval, max_jint) \ |
| constraint(PerfDataSamplingIntervalFunc, AfterErgo) \ |
| \ |
| product(bool, PerfDisableSharedMem, false, \ |
| "Store performance data in standard memory") \ |
| \ |
| product(intx, PerfDataMemorySize, 32*K, \ |
| "Size of performance data memory region. Will be rounded " \ |
| "up to a multiple of the native os page size.") \ |
| range(128, 32*64*K) \ |
| \ |
| product(intx, PerfMaxStringConstLength, 1024, \ |
| "Maximum PerfStringConstant string length before truncation") \ |
| range(32, 32*K) \ |
| \ |
| product(bool, PerfAllowAtExitRegistration, false, \ |
| "Allow registration of atexit() methods") \ |
| \ |
| product(bool, PerfBypassFileSystemCheck, false, \ |
| "Bypass Win32 file system criteria checks (Windows Only)") \ |
| \ |
| product(intx, UnguardOnExecutionViolation, 0, \ |
| "Unguard page and retry on no-execute fault (Win32 only) " \ |
| "0=off, 1=conservative, 2=aggressive") \ |
| range(0, 2) \ |
| \ |
| /* Serviceability Support */ \ |
| \ |
| product(bool, ManagementServer, false, \ |
| "Create JMX Management Server") \ |
| \ |
| product(bool, DisableAttachMechanism, false, \ |
| "Disable mechanism that allows tools to attach to this VM") \ |
| \ |
| product(bool, StartAttachListener, false, \ |
| "Always start Attach Listener at VM startup") \ |
| \ |
| product(bool, EnableDynamicAgentLoading, true, \ |
| "Allow tools to load agents with the attach mechanism") \ |
| \ |
| manageable(bool, PrintConcurrentLocks, false, \ |
| "Print java.util.concurrent locks in thread dump") \ |
| \ |
| product(bool, TransmitErrorReport, false, \ |
| "Enable error report transmission on erroneous termination") \ |
| \ |
| product(ccstr, ErrorReportServer, NULL, \ |
| "Override built-in error report server address") \ |
| \ |
| /* Shared spaces */ \ |
| \ |
| product(bool, UseSharedSpaces, true, \ |
| "Use shared spaces for metadata") \ |
| \ |
| product(bool, VerifySharedSpaces, false, \ |
| "Verify shared spaces (false for default archive, true for " \ |
| "archive specified by -XX:SharedArchiveFile)") \ |
| \ |
| product(bool, RequireSharedSpaces, false, \ |
| "Require shared spaces for metadata") \ |
| \ |
| product(bool, DumpSharedSpaces, false, \ |
| "Special mode: JVM reads a class list, loads classes, builds " \ |
| "shared spaces, and dumps the shared spaces to a file to be " \ |
| "used in future JVM runs") \ |
| \ |
| product(bool, PrintSharedArchiveAndExit, false, \ |
| "Print shared archive file contents") \ |
| \ |
| product(bool, PrintSharedDictionary, false, \ |
| "If PrintSharedArchiveAndExit is true, also print the shared " \ |
| "dictionary") \ |
| \ |
| product(size_t, SharedReadWriteSize, 0, \ |
| "Deprecated") \ |
| \ |
| product(size_t, SharedReadOnlySize, 0, \ |
| "Deprecated") \ |
| \ |
| product(size_t, SharedMiscDataSize, 0, \ |
| "Deprecated") \ |
| \ |
| product(size_t, SharedMiscCodeSize, 0, \ |
| "Deprecated") \ |
| \ |
| product(size_t, SharedBaseAddress, LP64_ONLY(32*G) \ |
| NOT_LP64(LINUX_ONLY(2*G) NOT_LINUX(0)), \ |
| "Address to allocate shared memory region for class data") \ |
| range(0, SIZE_MAX) \ |
| \ |
| product(uintx, SharedSymbolTableBucketSize, 4, \ |
| "Average number of symbols per bucket in shared table") \ |
| range(2, 246) \ |
| \ |
| diagnostic(bool, IgnoreUnverifiableClassesDuringDump, false, \ |
| "Do not quit -Xshare:dump even if we encounter unverifiable " \ |
| "classes. Just exclude them from the shared dictionary.") \ |
| \ |
| diagnostic(bool, PrintMethodHandleStubs, false, \ |
| "Print generated stub code for method handles") \ |
| \ |
| develop(bool, TraceMethodHandles, false, \ |
| "trace internal method handle operations") \ |
| \ |
| diagnostic(bool, VerifyMethodHandles, trueInDebug, \ |
| "perform extra checks when constructing method handles") \ |
| \ |
| diagnostic(bool, ShowHiddenFrames, false, \ |
| "show method handle implementation frames (usually hidden)") \ |
| \ |
| experimental(bool, TrustFinalNonStaticFields, false, \ |
| "trust final non-static declarations for constant folding") \ |
| \ |
| diagnostic(bool, FoldStableValues, true, \ |
| "Optimize loads from stable fields (marked w/ @Stable)") \ |
| \ |
| develop(bool, TraceInvokeDynamic, false, \ |
| "trace internal invoke dynamic operations") \ |
| \ |
| diagnostic(bool, PauseAtStartup, false, \ |
| "Causes the VM to pause at startup time and wait for the pause " \ |
| "file to be removed (default: ./vm.paused.<pid>)") \ |
| \ |
| diagnostic(ccstr, PauseAtStartupFile, NULL, \ |
| "The file to create and for whose removal to await when pausing " \ |
| "at startup. (default: ./vm.paused.<pid>)") \ |
| \ |
| diagnostic(bool, PauseAtExit, false, \ |
| "Pause and wait for keypress on exit if a debugger is attached") \ |
| \ |
| product(bool, ExtendedDTraceProbes, false, \ |
| "Enable performance-impacting dtrace probes") \ |
| \ |
| product(bool, DTraceMethodProbes, false, \ |
| "Enable dtrace probes for method-entry and method-exit") \ |
| \ |
| product(bool, DTraceAllocProbes, false, \ |
| "Enable dtrace probes for object allocation") \ |
| \ |
| product(bool, DTraceMonitorProbes, false, \ |
| "Enable dtrace probes for monitor events") \ |
| \ |
| product(bool, RelaxAccessControlCheck, false, \ |
| "Relax the access control checks in the verifier") \ |
| \ |
| product(uintx, StringTableSize, defaultStringTableSize, \ |
| "Number of buckets in the interned String table") \ |
| range(minimumStringTableSize, 111*defaultStringTableSize) \ |
| \ |
| experimental(uintx, SymbolTableSize, defaultSymbolTableSize, \ |
| "Number of buckets in the JVM internal Symbol table") \ |
| range(minimumSymbolTableSize, 111*defaultSymbolTableSize) \ |
| \ |
| product(bool, UseStringDeduplication, false, \ |
| "Use string deduplication") \ |
| \ |
| product(uintx, StringDeduplicationAgeThreshold, 3, \ |
| "A string must reach this age (or be promoted to an old region) " \ |
| "to be considered for deduplication") \ |
| range(1, markOopDesc::max_age) \ |
| \ |
| diagnostic(bool, StringDeduplicationResizeALot, false, \ |
| "Force table resize every time the table is scanned") \ |
| \ |
| diagnostic(bool, StringDeduplicationRehashALot, false, \ |
| "Force table rehash every time the table is scanned") \ |
| \ |
| diagnostic(bool, WhiteBoxAPI, false, \ |
| "Enable internal testing APIs") \ |
| \ |
| experimental(intx, SurvivorAlignmentInBytes, 0, \ |
| "Default survivor space alignment in bytes") \ |
| constraint(SurvivorAlignmentInBytesConstraintFunc,AfterErgo) \ |
| \ |
| product(bool , AllowNonVirtualCalls, false, \ |
| "Obey the ACC_SUPER flag and allow invokenonvirtual calls") \ |
| \ |
| product(ccstr, DumpLoadedClassList, NULL, \ |
| "Dump the names all loaded classes, that could be stored into " \ |
| "the CDS archive, in the specified file") \ |
| \ |
| product(ccstr, SharedClassListFile, NULL, \ |
| "Override the default CDS class list") \ |
| \ |
| diagnostic(ccstr, SharedArchiveFile, NULL, \ |
| "Override the default location of the CDS archive file") \ |
| \ |
| product(ccstr, ExtraSharedClassListFile, NULL, \ |
| "Extra classlist for building the CDS archive file") \ |
| \ |
| experimental(size_t, ArrayAllocatorMallocLimit, \ |
| SOLARIS_ONLY(64*K) NOT_SOLARIS((size_t)-1), \ |
| "Allocation less than this value will be allocated " \ |
| "using malloc. Larger allocations will use mmap.") \ |
| \ |
| experimental(bool, AlwaysAtomicAccesses, false, \ |
| "Accesses to all variables should always be atomic") \ |
| \ |
| product(bool, EnableTracing, false, \ |
| "Enable event-based tracing") \ |
| \ |
| product(bool, UseLockedTracing, false, \ |
| "Use locked-tracing when doing event-based tracing") \ |
| \ |
| diagnostic(bool, UseUnalignedAccesses, false, \ |
| "Use unaligned memory accesses in Unsafe") \ |
| \ |
| product_pd(bool, PreserveFramePointer, \ |
| "Use the FP register for holding the frame pointer " \ |
| "and not as a general purpose register.") \ |
| \ |
| diagnostic(bool, CheckIntrinsics, true, \ |
| "When a class C is loaded, check that " \ |
| "(1) all intrinsics defined by the VM for class C are present "\ |
| "in the loaded class file and are marked with the " \ |
| "@HotSpotIntrinsicCandidate annotation, that " \ |
| "(2) there is an intrinsic registered for all loaded methods " \ |
| "that are annotated with the @HotSpotIntrinsicCandidate " \ |
| "annotation, and that " \ |
| "(3) no orphan methods exist for class C (i.e., methods for " \ |
| "which the VM declares an intrinsic but that are not declared "\ |
| "in the loaded class C. " \ |
| "Check (3) is available only in debug builds.") \ |
| \ |
| diagnostic_pd(intx, InitArrayShortSize, \ |
| "Threshold small size (in bytes) for clearing arrays. " \ |
| "Anything this size or smaller may get converted to discrete " \ |
| "scalar stores.") \ |
| range(0, max_intx) \ |
| constraint(InitArrayShortSizeConstraintFunc, AfterErgo) \ |
| \ |
| diagnostic(bool, CompilerDirectivesIgnoreCompileCommands, false, \ |
| "Disable backwards compatibility for compile commands.") \ |
| \ |
| diagnostic(bool, CompilerDirectivesPrint, false, \ |
| "Print compiler directives on installation.") \ |
| diagnostic(int, CompilerDirectivesLimit, 50, \ |
| "Limit on number of compiler directives.") |
| |
| |
| /* |
| * Macros for factoring of globals |
| */ |
| |
| // Interface macros |
| #define DECLARE_PRODUCT_FLAG(type, name, value, doc) extern "C" type name; |
| #define DECLARE_PD_PRODUCT_FLAG(type, name, doc) extern "C" type name; |
| #define DECLARE_DIAGNOSTIC_FLAG(type, name, value, doc) extern "C" type name; |
| #define DECLARE_PD_DIAGNOSTIC_FLAG(type, name, doc) extern "C" type name; |
| #define DECLARE_EXPERIMENTAL_FLAG(type, name, value, doc) extern "C" type name; |
| #define DECLARE_MANAGEABLE_FLAG(type, name, value, doc) extern "C" type name; |
| #define DECLARE_PRODUCT_RW_FLAG(type, name, value, doc) extern "C" type name; |
| #ifdef PRODUCT |
| #define DECLARE_DEVELOPER_FLAG(type, name, value, doc) const type name = value; |
| #define DECLARE_PD_DEVELOPER_FLAG(type, name, doc) const type name = pd_##name; |
| #define DECLARE_NOTPRODUCT_FLAG(type, name, value, doc) const type name = value; |
| #else |
| #define DECLARE_DEVELOPER_FLAG(type, name, value, doc) extern "C" type name; |
| #define DECLARE_PD_DEVELOPER_FLAG(type, name, doc) extern "C" type name; |
| #define DECLARE_NOTPRODUCT_FLAG(type, name, value, doc) extern "C" type name; |
| #endif // PRODUCT |
| // Special LP64 flags, product only needed for now. |
| #ifdef _LP64 |
| #define DECLARE_LP64_PRODUCT_FLAG(type, name, value, doc) extern "C" type name; |
| #else |
| #define DECLARE_LP64_PRODUCT_FLAG(type, name, value, doc) const type name = value; |
| #endif // _LP64 |
| |
| // Implementation macros |
| #define MATERIALIZE_PRODUCT_FLAG(type, name, value, doc) type name = value; |
| #define MATERIALIZE_PD_PRODUCT_FLAG(type, name, doc) type name = pd_##name; |
| #define MATERIALIZE_DIAGNOSTIC_FLAG(type, name, value, doc) type name = value; |
| #define MATERIALIZE_PD_DIAGNOSTIC_FLAG(type, name, doc) type name = pd_##name; |
| #define MATERIALIZE_EXPERIMENTAL_FLAG(type, name, value, doc) type name = value; |
| #define MATERIALIZE_MANAGEABLE_FLAG(type, name, value, doc) type name = value; |
| #define MATERIALIZE_PRODUCT_RW_FLAG(type, name, value, doc) type name = value; |
| #ifdef PRODUCT |
| #define MATERIALIZE_DEVELOPER_FLAG(type, name, value, doc) |
| #define MATERIALIZE_PD_DEVELOPER_FLAG(type, name, doc) |
| #define MATERIALIZE_NOTPRODUCT_FLAG(type, name, value, doc) |
| #else |
| #define MATERIALIZE_DEVELOPER_FLAG(type, name, value, doc) type name = value; |
| #define MATERIALIZE_PD_DEVELOPER_FLAG(type, name, doc) type name = pd_##name; |
| #define MATERIALIZE_NOTPRODUCT_FLAG(type, name, value, doc) type name = value; |
| #endif // PRODUCT |
| #ifdef _LP64 |
| #define MATERIALIZE_LP64_PRODUCT_FLAG(type, name, value, doc) type name = value; |
| #else |
| #define MATERIALIZE_LP64_PRODUCT_FLAG(type, name, value, doc) /* flag is constant */ |
| #endif // _LP64 |
| |
| // Only materialize src code for range checking when required, ignore otherwise |
| #define IGNORE_RANGE(a, b) |
| // Only materialize src code for contraint checking when required, ignore otherwise |
| #define IGNORE_CONSTRAINT(func,type) |
| |
| #define IGNORE_WRITEABLE(type) |
| |
| RUNTIME_FLAGS(DECLARE_DEVELOPER_FLAG, \ |
| DECLARE_PD_DEVELOPER_FLAG, \ |
| DECLARE_PRODUCT_FLAG, \ |
| DECLARE_PD_PRODUCT_FLAG, \ |
| DECLARE_DIAGNOSTIC_FLAG, \ |
| DECLARE_PD_DIAGNOSTIC_FLAG, \ |
| DECLARE_EXPERIMENTAL_FLAG, \ |
| DECLARE_NOTPRODUCT_FLAG, \ |
| DECLARE_MANAGEABLE_FLAG, \ |
| DECLARE_PRODUCT_RW_FLAG, \ |
| DECLARE_LP64_PRODUCT_FLAG, \ |
| IGNORE_RANGE, \ |
| IGNORE_CONSTRAINT, \ |
| IGNORE_WRITEABLE) |
| |
| RUNTIME_OS_FLAGS(DECLARE_DEVELOPER_FLAG, \ |
| DECLARE_PD_DEVELOPER_FLAG, \ |
| DECLARE_PRODUCT_FLAG, \ |
| DECLARE_PD_PRODUCT_FLAG, \ |
| DECLARE_DIAGNOSTIC_FLAG, \ |
| DECLARE_PD_DIAGNOSTIC_FLAG, \ |
| DECLARE_NOTPRODUCT_FLAG, \ |
| IGNORE_RANGE, \ |
| IGNORE_CONSTRAINT, \ |
| IGNORE_WRITEABLE) |
| |
| ARCH_FLAGS(DECLARE_DEVELOPER_FLAG, \ |
| DECLARE_PRODUCT_FLAG, \ |
| DECLARE_DIAGNOSTIC_FLAG, \ |
| DECLARE_EXPERIMENTAL_FLAG, \ |
| DECLARE_NOTPRODUCT_FLAG, \ |
| IGNORE_RANGE, \ |
| IGNORE_CONSTRAINT, \ |
| IGNORE_WRITEABLE) |
| |
| // Extensions |
| |
| #include "runtime/globals_ext.hpp" |
| |
| #endif // SHARE_VM_RUNTIME_GLOBALS_HPP |