mirror of
https://github.com/ziglang/zig.git
synced 2024-11-30 17:12:31 +00:00
363 lines
11 KiB
C++
Vendored
363 lines
11 KiB
C++
Vendored
//===-- sanitizer_stoptheworld_netbsd_libcdep.cpp -------------------------===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
// See sanitizer_stoptheworld.h for details.
|
|
// This implementation was inspired by Markus Gutschke's linuxthreads.cc.
|
|
//
|
|
// This is a NetBSD variation of Linux stoptheworld implementation
|
|
// See sanitizer_stoptheworld_linux_libcdep.cpp for code comments.
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "sanitizer_platform.h"
|
|
|
|
#if SANITIZER_NETBSD
|
|
|
|
#include "sanitizer_stoptheworld.h"
|
|
|
|
#include "sanitizer_atomic.h"
|
|
#include "sanitizer_platform_limits_posix.h"
|
|
|
|
#include <sys/types.h>
|
|
|
|
#include <sys/ptrace.h>
|
|
#include <sys/uio.h>
|
|
#include <sys/wait.h>
|
|
|
|
#include <machine/reg.h>
|
|
|
|
#include <elf.h>
|
|
#include <errno.h>
|
|
#include <sched.h>
|
|
#include <signal.h>
|
|
#include <stddef.h>
|
|
|
|
#define internal_sigaction_norestorer internal_sigaction
|
|
|
|
#include "sanitizer_common.h"
|
|
#include "sanitizer_flags.h"
|
|
#include "sanitizer_libc.h"
|
|
#include "sanitizer_linux.h"
|
|
#include "sanitizer_mutex.h"
|
|
#include "sanitizer_placement_new.h"
|
|
|
|
namespace __sanitizer {
|
|
|
|
class SuspendedThreadsListNetBSD final : public SuspendedThreadsList {
|
|
public:
|
|
SuspendedThreadsListNetBSD() { thread_ids_.reserve(1024); }
|
|
|
|
tid_t GetThreadID(uptr index) const;
|
|
uptr ThreadCount() const;
|
|
bool ContainsTid(tid_t thread_id) const;
|
|
void Append(tid_t tid);
|
|
|
|
PtraceRegistersStatus GetRegistersAndSP(uptr index,
|
|
InternalMmapVector<uptr> *buffer,
|
|
uptr *sp) const;
|
|
|
|
private:
|
|
InternalMmapVector<tid_t> thread_ids_;
|
|
};
|
|
|
|
struct TracerThreadArgument {
|
|
StopTheWorldCallback callback;
|
|
void *callback_argument;
|
|
BlockingMutex mutex;
|
|
atomic_uintptr_t done;
|
|
uptr parent_pid;
|
|
};
|
|
|
|
class ThreadSuspender {
|
|
public:
|
|
explicit ThreadSuspender(pid_t pid, TracerThreadArgument *arg)
|
|
: arg(arg), pid_(pid) {
|
|
CHECK_GE(pid, 0);
|
|
}
|
|
bool SuspendAllThreads();
|
|
void ResumeAllThreads();
|
|
void KillAllThreads();
|
|
SuspendedThreadsListNetBSD &suspended_threads_list() {
|
|
return suspended_threads_list_;
|
|
}
|
|
TracerThreadArgument *arg;
|
|
|
|
private:
|
|
SuspendedThreadsListNetBSD suspended_threads_list_;
|
|
pid_t pid_;
|
|
};
|
|
|
|
void ThreadSuspender::ResumeAllThreads() {
|
|
int pterrno;
|
|
if (!internal_iserror(internal_ptrace(PT_DETACH, pid_, (void *)(uptr)1, 0),
|
|
&pterrno)) {
|
|
VReport(2, "Detached from process %d.\n", pid_);
|
|
} else {
|
|
VReport(1, "Could not detach from process %d (errno %d).\n", pid_, pterrno);
|
|
}
|
|
}
|
|
|
|
void ThreadSuspender::KillAllThreads() {
|
|
internal_ptrace(PT_KILL, pid_, nullptr, 0);
|
|
}
|
|
|
|
bool ThreadSuspender::SuspendAllThreads() {
|
|
int pterrno;
|
|
if (internal_iserror(internal_ptrace(PT_ATTACH, pid_, nullptr, 0),
|
|
&pterrno)) {
|
|
Printf("Could not attach to process %d (errno %d).\n", pid_, pterrno);
|
|
return false;
|
|
}
|
|
|
|
int status;
|
|
uptr waitpid_status;
|
|
HANDLE_EINTR(waitpid_status, internal_waitpid(pid_, &status, 0));
|
|
|
|
VReport(2, "Attached to process %d.\n", pid_);
|
|
|
|
#ifdef PT_LWPNEXT
|
|
struct ptrace_lwpstatus pl;
|
|
int op = PT_LWPNEXT;
|
|
#else
|
|
struct ptrace_lwpinfo pl;
|
|
int op = PT_LWPINFO;
|
|
#endif
|
|
|
|
pl.pl_lwpid = 0;
|
|
|
|
int val;
|
|
while ((val = internal_ptrace(op, pid_, (void *)&pl, sizeof(pl))) != -1 &&
|
|
pl.pl_lwpid != 0) {
|
|
suspended_threads_list_.Append(pl.pl_lwpid);
|
|
VReport(2, "Appended thread %d in process %d.\n", pl.pl_lwpid, pid_);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
// Pointer to the ThreadSuspender instance for use in signal handler.
|
|
static ThreadSuspender *thread_suspender_instance = nullptr;
|
|
|
|
// Synchronous signals that should not be blocked.
|
|
static const int kSyncSignals[] = {SIGABRT, SIGILL, SIGFPE, SIGSEGV,
|
|
SIGBUS, SIGXCPU, SIGXFSZ};
|
|
|
|
static void TracerThreadDieCallback() {
|
|
ThreadSuspender *inst = thread_suspender_instance;
|
|
if (inst && stoptheworld_tracer_pid == internal_getpid()) {
|
|
inst->KillAllThreads();
|
|
thread_suspender_instance = nullptr;
|
|
}
|
|
}
|
|
|
|
// Signal handler to wake up suspended threads when the tracer thread dies.
|
|
static void TracerThreadSignalHandler(int signum, __sanitizer_siginfo *siginfo,
|
|
void *uctx) {
|
|
SignalContext ctx(siginfo, uctx);
|
|
Printf("Tracer caught signal %d: addr=0x%zx pc=0x%zx sp=0x%zx\n", signum,
|
|
ctx.addr, ctx.pc, ctx.sp);
|
|
ThreadSuspender *inst = thread_suspender_instance;
|
|
if (inst) {
|
|
if (signum == SIGABRT)
|
|
inst->KillAllThreads();
|
|
else
|
|
inst->ResumeAllThreads();
|
|
RAW_CHECK(RemoveDieCallback(TracerThreadDieCallback));
|
|
thread_suspender_instance = nullptr;
|
|
atomic_store(&inst->arg->done, 1, memory_order_relaxed);
|
|
}
|
|
internal__exit((signum == SIGABRT) ? 1 : 2);
|
|
}
|
|
|
|
// Size of alternative stack for signal handlers in the tracer thread.
|
|
static const int kHandlerStackSize = 8192;
|
|
|
|
// This function will be run as a cloned task.
|
|
static int TracerThread(void *argument) {
|
|
TracerThreadArgument *tracer_thread_argument =
|
|
(TracerThreadArgument *)argument;
|
|
|
|
// Check if parent is already dead.
|
|
if (internal_getppid() != tracer_thread_argument->parent_pid)
|
|
internal__exit(4);
|
|
|
|
// Wait for the parent thread to finish preparations.
|
|
tracer_thread_argument->mutex.Lock();
|
|
tracer_thread_argument->mutex.Unlock();
|
|
|
|
RAW_CHECK(AddDieCallback(TracerThreadDieCallback));
|
|
|
|
ThreadSuspender thread_suspender(internal_getppid(), tracer_thread_argument);
|
|
// Global pointer for the signal handler.
|
|
thread_suspender_instance = &thread_suspender;
|
|
|
|
// Alternate stack for signal handling.
|
|
InternalMmapVector<char> handler_stack_memory(kHandlerStackSize);
|
|
stack_t handler_stack;
|
|
internal_memset(&handler_stack, 0, sizeof(handler_stack));
|
|
handler_stack.ss_sp = handler_stack_memory.data();
|
|
handler_stack.ss_size = kHandlerStackSize;
|
|
internal_sigaltstack(&handler_stack, nullptr);
|
|
|
|
// Install our handler for synchronous signals. Other signals should be
|
|
// blocked by the mask we inherited from the parent thread.
|
|
for (uptr i = 0; i < ARRAY_SIZE(kSyncSignals); i++) {
|
|
__sanitizer_sigaction act;
|
|
internal_memset(&act, 0, sizeof(act));
|
|
act.sigaction = TracerThreadSignalHandler;
|
|
act.sa_flags = SA_ONSTACK | SA_SIGINFO;
|
|
internal_sigaction_norestorer(kSyncSignals[i], &act, 0);
|
|
}
|
|
|
|
int exit_code = 0;
|
|
if (!thread_suspender.SuspendAllThreads()) {
|
|
VReport(1, "Failed suspending threads.\n");
|
|
exit_code = 3;
|
|
} else {
|
|
tracer_thread_argument->callback(thread_suspender.suspended_threads_list(),
|
|
tracer_thread_argument->callback_argument);
|
|
thread_suspender.ResumeAllThreads();
|
|
exit_code = 0;
|
|
}
|
|
RAW_CHECK(RemoveDieCallback(TracerThreadDieCallback));
|
|
thread_suspender_instance = nullptr;
|
|
atomic_store(&tracer_thread_argument->done, 1, memory_order_relaxed);
|
|
return exit_code;
|
|
}
|
|
|
|
class ScopedStackSpaceWithGuard {
|
|
public:
|
|
explicit ScopedStackSpaceWithGuard(uptr stack_size) {
|
|
stack_size_ = stack_size;
|
|
guard_size_ = GetPageSizeCached();
|
|
// FIXME: Omitting MAP_STACK here works in current kernels but might break
|
|
// in the future.
|
|
guard_start_ =
|
|
(uptr)MmapOrDie(stack_size_ + guard_size_, "ScopedStackWithGuard");
|
|
CHECK(MprotectNoAccess((uptr)guard_start_, guard_size_));
|
|
}
|
|
~ScopedStackSpaceWithGuard() {
|
|
UnmapOrDie((void *)guard_start_, stack_size_ + guard_size_);
|
|
}
|
|
void *Bottom() const {
|
|
return (void *)(guard_start_ + stack_size_ + guard_size_);
|
|
}
|
|
|
|
private:
|
|
uptr stack_size_;
|
|
uptr guard_size_;
|
|
uptr guard_start_;
|
|
};
|
|
|
|
static __sanitizer_sigset_t blocked_sigset;
|
|
static __sanitizer_sigset_t old_sigset;
|
|
|
|
struct ScopedSetTracerPID {
|
|
explicit ScopedSetTracerPID(uptr tracer_pid) {
|
|
stoptheworld_tracer_pid = tracer_pid;
|
|
stoptheworld_tracer_ppid = internal_getpid();
|
|
}
|
|
~ScopedSetTracerPID() {
|
|
stoptheworld_tracer_pid = 0;
|
|
stoptheworld_tracer_ppid = 0;
|
|
}
|
|
};
|
|
|
|
void StopTheWorld(StopTheWorldCallback callback, void *argument) {
|
|
// Prepare the arguments for TracerThread.
|
|
struct TracerThreadArgument tracer_thread_argument;
|
|
tracer_thread_argument.callback = callback;
|
|
tracer_thread_argument.callback_argument = argument;
|
|
tracer_thread_argument.parent_pid = internal_getpid();
|
|
atomic_store(&tracer_thread_argument.done, 0, memory_order_relaxed);
|
|
const uptr kTracerStackSize = 2 * 1024 * 1024;
|
|
ScopedStackSpaceWithGuard tracer_stack(kTracerStackSize);
|
|
|
|
tracer_thread_argument.mutex.Lock();
|
|
|
|
internal_sigfillset(&blocked_sigset);
|
|
for (uptr i = 0; i < ARRAY_SIZE(kSyncSignals); i++)
|
|
internal_sigdelset(&blocked_sigset, kSyncSignals[i]);
|
|
int rv = internal_sigprocmask(SIG_BLOCK, &blocked_sigset, &old_sigset);
|
|
CHECK_EQ(rv, 0);
|
|
uptr tracer_pid = internal_clone(TracerThread, tracer_stack.Bottom(),
|
|
CLONE_VM | CLONE_FS | CLONE_FILES,
|
|
&tracer_thread_argument);
|
|
internal_sigprocmask(SIG_SETMASK, &old_sigset, 0);
|
|
int local_errno = 0;
|
|
if (internal_iserror(tracer_pid, &local_errno)) {
|
|
VReport(1, "Failed spawning a tracer thread (errno %d).\n", local_errno);
|
|
tracer_thread_argument.mutex.Unlock();
|
|
} else {
|
|
ScopedSetTracerPID scoped_set_tracer_pid(tracer_pid);
|
|
|
|
tracer_thread_argument.mutex.Unlock();
|
|
|
|
while (atomic_load(&tracer_thread_argument.done, memory_order_relaxed) == 0)
|
|
sched_yield();
|
|
|
|
for (;;) {
|
|
uptr waitpid_status = internal_waitpid(tracer_pid, nullptr, __WALL);
|
|
if (!internal_iserror(waitpid_status, &local_errno))
|
|
break;
|
|
if (local_errno == EINTR)
|
|
continue;
|
|
VReport(1, "Waiting on the tracer thread failed (errno %d).\n",
|
|
local_errno);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
tid_t SuspendedThreadsListNetBSD::GetThreadID(uptr index) const {
|
|
CHECK_LT(index, thread_ids_.size());
|
|
return thread_ids_[index];
|
|
}
|
|
|
|
uptr SuspendedThreadsListNetBSD::ThreadCount() const {
|
|
return thread_ids_.size();
|
|
}
|
|
|
|
bool SuspendedThreadsListNetBSD::ContainsTid(tid_t thread_id) const {
|
|
for (uptr i = 0; i < thread_ids_.size(); i++) {
|
|
if (thread_ids_[i] == thread_id)
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
void SuspendedThreadsListNetBSD::Append(tid_t tid) {
|
|
thread_ids_.push_back(tid);
|
|
}
|
|
|
|
PtraceRegistersStatus SuspendedThreadsListNetBSD::GetRegistersAndSP(
|
|
uptr index, InternalMmapVector<uptr> *buffer, uptr *sp) const {
|
|
lwpid_t tid = GetThreadID(index);
|
|
pid_t ppid = internal_getppid();
|
|
struct reg regs;
|
|
int pterrno;
|
|
bool isErr =
|
|
internal_iserror(internal_ptrace(PT_GETREGS, ppid, ®s, tid), &pterrno);
|
|
if (isErr) {
|
|
VReport(1,
|
|
"Could not get registers from process %d thread %d (errno %d).\n",
|
|
ppid, tid, pterrno);
|
|
return pterrno == ESRCH ? REGISTERS_UNAVAILABLE_FATAL
|
|
: REGISTERS_UNAVAILABLE;
|
|
}
|
|
|
|
*sp = PTRACE_REG_SP(®s);
|
|
buffer->resize(RoundUpTo(sizeof(regs), sizeof(uptr)) / sizeof(uptr));
|
|
internal_memcpy(buffer->data(), ®s, sizeof(regs));
|
|
|
|
return REGISTERS_AVAILABLE;
|
|
}
|
|
|
|
} // namespace __sanitizer
|
|
|
|
#endif
|