blob: b897c86d24981fb0c83081341debdc2d9473db3b [file] [log] [blame]
/*
* Copyright (c) 2020, 2022, Oracle and/or its affiliates. All rights reserved.
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
*
* This code is free software; you can redistribute it and/or modify it
* under the terms of the GNU General Public License version 2 only, as
* published by the Free Software Foundation.
*
* This code is distributed in the hope that it will be useful, but WITHOUT
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
* version 2 for more details (a copy is included in the LICENSE file that
* accompanied this code).
*
* You should have received a copy of the GNU General Public License version
* 2 along with this work; if not, write to the Free Software Foundation,
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
*
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
* or visit www.oracle.com if you need additional information or have any
* questions.
*
*/
#ifndef SHARE_RUNTIME_STACKOVERFLOW_HPP
#define SHARE_RUNTIME_STACKOVERFLOW_HPP
#include "utilities/align.hpp"
#include "utilities/debug.hpp"
class JavaThread;
// StackOverflow handling is encapsulated in this class. This class contains state variables
// for each JavaThread that are used to detect stack overflow though explicit checks or through
// checks in the signal handler when stack banging into guard pages causes a trap.
// The state variables also record whether guard pages are enabled or disabled.
class StackOverflow {
friend class JVMCIVMStructs;
friend class JavaThread;
public:
// State of the stack guard pages for the containing thread.
enum StackGuardState {
stack_guard_unused, // not needed
stack_guard_reserved_disabled,
stack_guard_yellow_reserved_disabled,// disabled (temporarily) after stack overflow
stack_guard_enabled // enabled
};
StackOverflow() :
_stack_guard_state(stack_guard_unused),
_stack_overflow_limit(nullptr),
_reserved_stack_activation(nullptr), // stack base not known yet
_shadow_zone_safe_limit(nullptr),
_shadow_zone_growth_watermark(nullptr),
_stack_base(nullptr), _stack_end(nullptr) {}
// Initialization after thread is started.
void initialize(address base, address end) {
_stack_base = base;
_stack_end = end;
set_stack_overflow_limit();
set_shadow_zone_limits();
set_reserved_stack_activation(base);
}
private:
StackGuardState _stack_guard_state;
// Precompute the limit of the stack as used in stack overflow checks.
// We load it from here to simplify the stack overflow check in assembly.
address _stack_overflow_limit;
address _reserved_stack_activation;
address _shadow_zone_safe_limit;
address _shadow_zone_growth_watermark;
// Support for stack overflow handling, copied down from thread.
address _stack_base;
address _stack_end;
address stack_end() const { return _stack_end; }
address stack_base() const { assert(_stack_base != nullptr, "Sanity check"); return _stack_base; }
// Stack overflow support
// --------------------------------------------------------------------------------
//
// The Java thread stack is structured as follows:
//
// (low addresses)
//
// -- <-- stack_end() ---
// | |
// | red zone |
// | |
// -- <-- stack_red_zone_base() |
// | |
// | guard
// | yellow zone zone
// | |
// | |
// -- <-- stack_yellow_zone_base() |
// | |
// | |
// | reserved zone |
// | |
// -- <-- stack_reserved_zone_base() --- ---
// ^
// | <-- stack_overflow_limit() [somewhere in here]
// | shadow
// | zone
// | size
// v
// --- <-- shadow_zone_safe_limit()
// (Here and below: not yet touched stack)
//
//
// (Here and below: touched at least once) ---
// ^
// | shadow
// | zone
// | size
// v
// --- <-- shadow_zone_growth_watermark()
//
//
// --
// |
// | shadow zone
// |
// --
// x frame n
// --
// x frame n-1
// x
// --
// ...
//
// --
// x frame 0
// -- <-- stack_base()
//
// (high addresses)
//
//
// The stack overflow mechanism detects overflows by touching ("banging") the stack
// ahead of current stack pointer (SP). The entirety of guard zone is memory protected,
// therefore such access would trap when touching the guard zone, and one of the following
// things would happen.
//
// Access in the red zone: unrecoverable stack overflow. Crash the VM, generate a report,
// crash dump, and other diagnostics.
//
// Access in the yellow zone: recoverable, reportable stack overflow. Create and throw
// a StackOverflowError, remove the protection of yellow zone temporarily to let exception
// handlers run. If exception handlers themselves run out of stack, they will crash VM due
// to access to red zone.
//
// Access in the reserved zone: recoverable, reportable, transparent for privileged methods
// stack overflow. Perform a stack walk to check if there's a method annotated with
// @ReservedStackAccess on the call stack. If such method is found, remove the protection of
// reserved zone temporarily, and let the method run. If not, handle the access like a yellow
// zone trap.
//
// The banging itself happens within the "shadow zone" that extends from the current SP.
//
// The goals for properly implemented shadow zone banging are:
//
// a) Allow native/VM methods to run without stack overflow checks within some reasonable
// headroom. Default shadow zone size should accommodate the largest normally expected
// native/VM stack use.
// b) Guarantee the stack overflow checks work even if SP is dangerously close to guard zone.
// If SP is very low, banging at the edge of shadow zone (SP+shadow-zone-size) can slip
// into adjacent thread stack, or even into other readable memory. This would potentially
// pass the check by accident.
// c) Allow for incremental stack growth on some OSes. This is enabled by handling traps
// from not yet committed thread stacks, even outside the guard zone. The banging should
// not allow uncommitted "gaps" on thread stack. See for example the uses of
// os::map_stack_shadow_pages().
// d) Make sure the stack overflow trap happens in the code that is known to runtime, so
// the traps can be reasonably handled: handling a spurious trap from executing Java code
// is hard, while properly handling the trap from VM/native code is nearly impossible.
//
// The simplest code that satisfies all these requirements is banging the shadow zone
// page by page at every Java/native method entry.
//
// While that code is sufficient, it comes with the large performance cost. This performance
// cost can be reduced by several *optional* techniques:
//
// 1. Guarantee that stack would not take another page. If so, the current bang was
// enough to verify we are not near the guard zone. This kind of insight is usually only
// available for compilers that can know the size of the frame exactly.
//
// Examples: PhaseOutput::need_stack_bang.
//
// 2. Check the current SP in relation to shadow zone safe limit.
//
// Define "safe limit" as the highest SP where banging would not touch the guard zone.
// Then, do the page-by-page bang only if current SP is above that safe limit, OR some
// OS-es need it to get the stack mapped.
//
// Examples: AbstractAssembler::generate_stack_overflow_check, JavaCalls::call_helper,
// os::stack_shadow_pages_available, os::map_stack_shadow_pages and their uses.
//
// 3. Check the current SP in relation to the shadow zone growth watermark.
//
// Define "shadow zone growth watermark" as the highest SP where we banged already.
// Invariant: growth watermark is always above the safe limit, which allows testing
// for watermark and safe limit at the same time in the most frequent case.
//
// Easy and overwhelmingly frequent case: SP is above the growth watermark, and
// by extension above the safe limit. In this case, we know that the guard zone is far away
// (safe limit), and that the stack was banged before for stack growth (growth watermark).
// Therefore, we can skip the banging altogether.
//
// Harder cases: SP is below the growth watermark. In might be due to two things:
// we have not banged the stack for growth (below growth watermark only), or we are
// close to guard zone (also below safe limit). Do the full banging. Once done, we
// can adjust the growth watermark, thus recording the bang for stack growth had
// happened.
//
// Examples: TemplateInterpreterGenerator::bang_stack_shadow_pages on x86 and others.
private:
// These values are derived from flags StackRedPages, StackYellowPages,
// StackReservedPages and StackShadowPages.
static size_t _stack_red_zone_size;
static size_t _stack_yellow_zone_size;
static size_t _stack_reserved_zone_size;
static size_t _stack_shadow_zone_size;
public:
static void initialize_stack_zone_sizes();
static size_t stack_red_zone_size() {
assert(_stack_red_zone_size > 0, "Don't call this before the field is initialized.");
return _stack_red_zone_size;
}
// Returns base of red zone (one-beyond the highest red zone address, so
// itself outside red zone and the highest address of the yellow zone).
address stack_red_zone_base() const {
return (address)(stack_end() + stack_red_zone_size());
}
// Returns true if address points into the red zone.
bool in_stack_red_zone(address a) const {
return a < stack_red_zone_base() && a >= stack_end();
}
static size_t stack_yellow_zone_size() {
assert(_stack_yellow_zone_size > 0, "Don't call this before the field is initialized.");
return _stack_yellow_zone_size;
}
static size_t stack_reserved_zone_size() {
// _stack_reserved_zone_size may be 0. This indicates the feature is off.
return _stack_reserved_zone_size;
}
// Returns base of the reserved zone (one-beyond the highest reserved zone address).
address stack_reserved_zone_base() const {
return (address)(stack_end() +
(stack_red_zone_size() + stack_yellow_zone_size() + stack_reserved_zone_size()));
}
// Returns true if address points into the reserved zone.
bool in_stack_reserved_zone(address a) const {
return (a < stack_reserved_zone_base()) &&
(a >= (address)((intptr_t)stack_reserved_zone_base() - stack_reserved_zone_size()));
}
static size_t stack_yellow_reserved_zone_size() {
return _stack_yellow_zone_size + _stack_reserved_zone_size;
}
// Returns true if a points into either yellow or reserved zone.
bool in_stack_yellow_reserved_zone(address a) const {
return (a < stack_reserved_zone_base()) && (a >= stack_red_zone_base());
}
// Size of red + yellow + reserved zones.
static size_t stack_guard_zone_size() {
return stack_red_zone_size() + stack_yellow_reserved_zone_size();
}
static size_t stack_shadow_zone_size() {
assert(_stack_shadow_zone_size > 0, "Don't call this before the field is initialized.");
return _stack_shadow_zone_size;
}
address shadow_zone_safe_limit() const {
assert(_shadow_zone_safe_limit != nullptr, "Don't call this before the field is initialized.");
return _shadow_zone_safe_limit;
}
void create_stack_guard_pages();
void remove_stack_guard_pages();
void enable_stack_reserved_zone(bool check_if_disabled = false);
void disable_stack_reserved_zone();
void enable_stack_yellow_reserved_zone();
void disable_stack_yellow_reserved_zone();
void disable_stack_red_zone();
bool stack_guard_zone_unused() const { return _stack_guard_state == stack_guard_unused; }
bool stack_yellow_reserved_zone_disabled() const {
return _stack_guard_state == stack_guard_yellow_reserved_disabled;
}
size_t stack_available(address cur_sp) const {
// This code assumes java stacks grow down
address low_addr; // Limit on the address for deepest stack depth
if (_stack_guard_state == stack_guard_unused) {
low_addr = stack_end();
} else {
low_addr = stack_reserved_zone_base();
}
return cur_sp > low_addr ? cur_sp - low_addr : 0;
}
bool stack_guards_enabled() const;
address reserved_stack_activation() const { return _reserved_stack_activation; }
void set_reserved_stack_activation(address addr) {
assert(_reserved_stack_activation == stack_base()
|| _reserved_stack_activation == nullptr
|| addr == stack_base(), "Must not be set twice");
_reserved_stack_activation = addr;
}
// Attempt to reguard the stack after a stack overflow may have occurred.
// Returns true if (a) guard pages are not needed on this thread, (b) the
// pages are already guarded, or (c) the pages were successfully reguarded.
// Returns false if there is not enough stack space to reguard the pages, in
// which case the caller should unwind a frame and try again. The argument
// should be the caller's (approximate) sp.
bool reguard_stack(address cur_sp);
// Similar to above but see if current stackpoint is out of the guard area
// and reguard if possible.
bool reguard_stack(void);
bool reguard_stack_if_needed(void);
void set_stack_overflow_limit() {
_stack_overflow_limit =
stack_end() + MAX2(stack_guard_zone_size(), stack_shadow_zone_size());
}
void set_shadow_zone_limits() {
_shadow_zone_safe_limit =
stack_end() + stack_guard_zone_size() + stack_shadow_zone_size();
_shadow_zone_growth_watermark =
stack_base();
}
address stack_overflow_limit() const { return _stack_overflow_limit; }
};
#endif // SHARE_RUNTIME_STACKOVERFLOW_HPP