| 1 | //===-- sanitizer_stoptheworld_netbsd_libcdep.cpp -------------------------===// |
| 2 | // |
| 3 | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
| 4 | // See https://llvm.org/LICENSE.txt for license information. |
| 5 | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
| 6 | // |
| 7 | //===----------------------------------------------------------------------===// |
| 8 | // |
| 9 | // See sanitizer_stoptheworld.h for details. |
| 10 | // This implementation was inspired by Markus Gutschke's linuxthreads.cc. |
| 11 | // |
| 12 | // This is a NetBSD variation of Linux stoptheworld implementation |
| 13 | // See sanitizer_stoptheworld_linux_libcdep.cpp for code comments. |
| 14 | // |
| 15 | //===----------------------------------------------------------------------===// |
| 16 | |
| 17 | #include "sanitizer_platform.h" |
| 18 | |
| 19 | #if SANITIZER_NETBSD |
| 20 | |
| 21 | #include "sanitizer_stoptheworld.h" |
| 22 | |
| 23 | #include "sanitizer_atomic.h" |
| 24 | #include "sanitizer_platform_limits_posix.h" |
| 25 | |
| 26 | #include <sys/types.h> |
| 27 | |
| 28 | #include <sys/ptrace.h> |
| 29 | #include <sys/uio.h> |
| 30 | #include <sys/wait.h> |
| 31 | |
| 32 | #include <machine/reg.h> |
| 33 | |
| 34 | #include <elf.h> |
| 35 | #include <errno.h> |
| 36 | #include <sched.h> |
| 37 | #include <signal.h> |
| 38 | #include <stddef.h> |
| 39 | |
| 40 | #define internal_sigaction_norestorer internal_sigaction |
| 41 | |
| 42 | #include "sanitizer_common.h" |
| 43 | #include "sanitizer_flags.h" |
| 44 | #include "sanitizer_libc.h" |
| 45 | #include "sanitizer_linux.h" |
| 46 | #include "sanitizer_mutex.h" |
| 47 | #include "sanitizer_placement_new.h" |
| 48 | |
| 49 | namespace __sanitizer { |
| 50 | |
| 51 | class SuspendedThreadsListNetBSD final : public SuspendedThreadsList { |
| 52 | public: |
| 53 | SuspendedThreadsListNetBSD() { thread_ids_.reserve(1024); } |
| 54 | |
| 55 | tid_t GetThreadID(uptr index) const; |
| 56 | uptr ThreadCount() const; |
| 57 | bool ContainsTid(tid_t thread_id) const; |
| 58 | void Append(tid_t tid); |
| 59 | |
| 60 | PtraceRegistersStatus GetRegistersAndSP(uptr index, |
| 61 | InternalMmapVector<uptr> *buffer, |
| 62 | uptr *sp) const; |
| 63 | |
| 64 | private: |
| 65 | InternalMmapVector<tid_t> thread_ids_; |
| 66 | }; |
| 67 | |
| 68 | struct TracerThreadArgument { |
| 69 | StopTheWorldCallback callback; |
| 70 | void *callback_argument; |
| 71 | Mutex mutex; |
| 72 | atomic_uintptr_t done; |
| 73 | uptr parent_pid; |
| 74 | }; |
| 75 | |
| 76 | class ThreadSuspender { |
| 77 | public: |
| 78 | explicit ThreadSuspender(pid_t pid, TracerThreadArgument *arg) |
| 79 | : arg(arg), pid_(pid) { |
| 80 | CHECK_GE(pid, 0); |
| 81 | } |
| 82 | bool SuspendAllThreads(); |
| 83 | void ResumeAllThreads(); |
| 84 | void KillAllThreads(); |
| 85 | SuspendedThreadsListNetBSD &suspended_threads_list() { |
| 86 | return suspended_threads_list_; |
| 87 | } |
| 88 | TracerThreadArgument *arg; |
| 89 | |
| 90 | private: |
| 91 | SuspendedThreadsListNetBSD suspended_threads_list_; |
| 92 | pid_t pid_; |
| 93 | }; |
| 94 | |
| 95 | void ThreadSuspender::ResumeAllThreads() { |
| 96 | int pterrno; |
| 97 | if (!internal_iserror(internal_ptrace(PT_DETACH, pid_, (void *)(uptr)1, 0), |
| 98 | &pterrno)) { |
| 99 | VReport(2, "Detached from process %d.\n" , pid_); |
| 100 | } else { |
| 101 | VReport(1, "Could not detach from process %d (errno %d).\n" , pid_, pterrno); |
| 102 | } |
| 103 | } |
| 104 | |
| 105 | void ThreadSuspender::KillAllThreads() { |
| 106 | internal_ptrace(PT_KILL, pid_, nullptr, 0); |
| 107 | } |
| 108 | |
| 109 | bool ThreadSuspender::SuspendAllThreads() { |
| 110 | int pterrno; |
| 111 | if (internal_iserror(internal_ptrace(PT_ATTACH, pid_, nullptr, 0), |
| 112 | &pterrno)) { |
| 113 | Printf("Could not attach to process %d (errno %d).\n" , pid_, pterrno); |
| 114 | return false; |
| 115 | } |
| 116 | |
| 117 | int status; |
| 118 | uptr waitpid_status; |
| 119 | HANDLE_EINTR(waitpid_status, internal_waitpid(pid_, &status, 0)); |
| 120 | |
| 121 | VReport(2, "Attached to process %d.\n" , pid_); |
| 122 | |
| 123 | #ifdef PT_LWPNEXT |
| 124 | struct ptrace_lwpstatus pl; |
| 125 | int op = PT_LWPNEXT; |
| 126 | #else |
| 127 | struct ptrace_lwpinfo pl; |
| 128 | int op = PT_LWPINFO; |
| 129 | #endif |
| 130 | |
| 131 | pl.pl_lwpid = 0; |
| 132 | |
| 133 | int val; |
| 134 | while ((val = internal_ptrace(op, pid_, (void *)&pl, sizeof(pl))) != -1 && |
| 135 | pl.pl_lwpid != 0) { |
| 136 | suspended_threads_list_.Append(pl.pl_lwpid); |
| 137 | VReport(2, "Appended thread %d in process %d.\n" , pl.pl_lwpid, pid_); |
| 138 | } |
| 139 | return true; |
| 140 | } |
| 141 | |
| 142 | // Pointer to the ThreadSuspender instance for use in signal handler. |
| 143 | static ThreadSuspender *thread_suspender_instance = nullptr; |
| 144 | |
| 145 | // Synchronous signals that should not be blocked. |
| 146 | static const int kSyncSignals[] = {SIGABRT, SIGILL, SIGFPE, SIGSEGV, |
| 147 | SIGBUS, SIGXCPU, SIGXFSZ}; |
| 148 | |
| 149 | static void TracerThreadDieCallback() { |
| 150 | ThreadSuspender *inst = thread_suspender_instance; |
| 151 | if (inst && stoptheworld_tracer_pid == internal_getpid()) { |
| 152 | inst->KillAllThreads(); |
| 153 | thread_suspender_instance = nullptr; |
| 154 | } |
| 155 | } |
| 156 | |
| 157 | // Signal handler to wake up suspended threads when the tracer thread dies. |
| 158 | static void TracerThreadSignalHandler(int signum, __sanitizer_siginfo *siginfo, |
| 159 | void *uctx) { |
| 160 | SignalContext ctx(siginfo, uctx); |
| 161 | Printf("Tracer caught signal %d: addr=%p pc=%p sp=%p\n" , signum, |
| 162 | (void *)ctx.addr, (void *)ctx.pc, (void *)ctx.sp); |
| 163 | ThreadSuspender *inst = thread_suspender_instance; |
| 164 | if (inst) { |
| 165 | if (signum == SIGABRT) |
| 166 | inst->KillAllThreads(); |
| 167 | else |
| 168 | inst->ResumeAllThreads(); |
| 169 | RAW_CHECK(RemoveDieCallback(TracerThreadDieCallback)); |
| 170 | thread_suspender_instance = nullptr; |
| 171 | atomic_store(&inst->arg->done, 1, memory_order_relaxed); |
| 172 | } |
| 173 | internal__exit((signum == SIGABRT) ? 1 : 2); |
| 174 | } |
| 175 | |
| 176 | // Size of alternative stack for signal handlers in the tracer thread. |
| 177 | static const int kHandlerStackSize = 8192; |
| 178 | |
| 179 | // This function will be run as a cloned task. |
| 180 | static int TracerThread(void *argument) { |
| 181 | TracerThreadArgument *tracer_thread_argument = |
| 182 | (TracerThreadArgument *)argument; |
| 183 | |
| 184 | // Check if parent is already dead. |
| 185 | if (internal_getppid() != tracer_thread_argument->parent_pid) |
| 186 | internal__exit(4); |
| 187 | |
| 188 | // Wait for the parent thread to finish preparations. |
| 189 | tracer_thread_argument->mutex.Lock(); |
| 190 | tracer_thread_argument->mutex.Unlock(); |
| 191 | |
| 192 | RAW_CHECK(AddDieCallback(TracerThreadDieCallback)); |
| 193 | |
| 194 | ThreadSuspender thread_suspender(internal_getppid(), tracer_thread_argument); |
| 195 | // Global pointer for the signal handler. |
| 196 | thread_suspender_instance = &thread_suspender; |
| 197 | |
| 198 | // Alternate stack for signal handling. |
| 199 | InternalMmapVector<char> handler_stack_memory(kHandlerStackSize); |
| 200 | stack_t handler_stack; |
| 201 | internal_memset(&handler_stack, 0, sizeof(handler_stack)); |
| 202 | handler_stack.ss_sp = handler_stack_memory.data(); |
| 203 | handler_stack.ss_size = kHandlerStackSize; |
| 204 | internal_sigaltstack(&handler_stack, nullptr); |
| 205 | |
| 206 | // Install our handler for synchronous signals. Other signals should be |
| 207 | // blocked by the mask we inherited from the parent thread. |
| 208 | for (uptr i = 0; i < ARRAY_SIZE(kSyncSignals); i++) { |
| 209 | __sanitizer_sigaction act; |
| 210 | internal_memset(&act, 0, sizeof(act)); |
| 211 | act.sigaction = TracerThreadSignalHandler; |
| 212 | act.sa_flags = SA_ONSTACK | SA_SIGINFO; |
| 213 | internal_sigaction_norestorer(kSyncSignals[i], &act, 0); |
| 214 | } |
| 215 | |
| 216 | int exit_code = 0; |
| 217 | if (!thread_suspender.SuspendAllThreads()) { |
| 218 | VReport(1, "Failed suspending threads.\n" ); |
| 219 | exit_code = 3; |
| 220 | } else { |
| 221 | tracer_thread_argument->callback(thread_suspender.suspended_threads_list(), |
| 222 | tracer_thread_argument->callback_argument); |
| 223 | thread_suspender.ResumeAllThreads(); |
| 224 | exit_code = 0; |
| 225 | } |
| 226 | RAW_CHECK(RemoveDieCallback(TracerThreadDieCallback)); |
| 227 | thread_suspender_instance = nullptr; |
| 228 | atomic_store(&tracer_thread_argument->done, 1, memory_order_relaxed); |
| 229 | return exit_code; |
| 230 | } |
| 231 | |
| 232 | class ScopedStackSpaceWithGuard { |
| 233 | public: |
| 234 | explicit ScopedStackSpaceWithGuard(uptr stack_size) { |
| 235 | stack_size_ = stack_size; |
| 236 | guard_size_ = GetPageSizeCached(); |
| 237 | // FIXME: Omitting MAP_STACK here works in current kernels but might break |
| 238 | // in the future. |
| 239 | guard_start_ = |
| 240 | (uptr)MmapOrDie(stack_size_ + guard_size_, "ScopedStackWithGuard" ); |
| 241 | CHECK(MprotectNoAccess((uptr)guard_start_, guard_size_)); |
| 242 | } |
| 243 | ~ScopedStackSpaceWithGuard() { |
| 244 | UnmapOrDie((void *)guard_start_, stack_size_ + guard_size_); |
| 245 | } |
| 246 | void *Bottom() const { |
| 247 | return (void *)(guard_start_ + stack_size_ + guard_size_); |
| 248 | } |
| 249 | |
| 250 | private: |
| 251 | uptr stack_size_; |
| 252 | uptr guard_size_; |
| 253 | uptr guard_start_; |
| 254 | }; |
| 255 | |
| 256 | static __sanitizer_sigset_t blocked_sigset; |
| 257 | static __sanitizer_sigset_t old_sigset; |
| 258 | |
| 259 | struct ScopedSetTracerPID { |
| 260 | explicit ScopedSetTracerPID(uptr tracer_pid) { |
| 261 | stoptheworld_tracer_pid = tracer_pid; |
| 262 | stoptheworld_tracer_ppid = internal_getpid(); |
| 263 | } |
| 264 | ~ScopedSetTracerPID() { |
| 265 | stoptheworld_tracer_pid = 0; |
| 266 | stoptheworld_tracer_ppid = 0; |
| 267 | } |
| 268 | }; |
| 269 | |
| 270 | void StopTheWorld(StopTheWorldCallback callback, void *argument) { |
| 271 | // Prepare the arguments for TracerThread. |
| 272 | struct TracerThreadArgument tracer_thread_argument; |
| 273 | tracer_thread_argument.callback = callback; |
| 274 | tracer_thread_argument.callback_argument = argument; |
| 275 | tracer_thread_argument.parent_pid = internal_getpid(); |
| 276 | atomic_store(&tracer_thread_argument.done, 0, memory_order_relaxed); |
| 277 | const uptr kTracerStackSize = 2 * 1024 * 1024; |
| 278 | ScopedStackSpaceWithGuard tracer_stack(kTracerStackSize); |
| 279 | |
| 280 | tracer_thread_argument.mutex.Lock(); |
| 281 | |
| 282 | internal_sigfillset(&blocked_sigset); |
| 283 | for (uptr i = 0; i < ARRAY_SIZE(kSyncSignals); i++) |
| 284 | internal_sigdelset(&blocked_sigset, kSyncSignals[i]); |
| 285 | int rv = internal_sigprocmask(SIG_BLOCK, &blocked_sigset, &old_sigset); |
| 286 | CHECK_EQ(rv, 0); |
| 287 | uptr tracer_pid = internal_clone(TracerThread, tracer_stack.Bottom(), |
| 288 | CLONE_VM | CLONE_FS | CLONE_FILES, |
| 289 | &tracer_thread_argument); |
| 290 | internal_sigprocmask(SIG_SETMASK, &old_sigset, 0); |
| 291 | int local_errno = 0; |
| 292 | if (internal_iserror(tracer_pid, &local_errno)) { |
| 293 | VReport(1, "Failed spawning a tracer thread (errno %d).\n" , local_errno); |
| 294 | tracer_thread_argument.mutex.Unlock(); |
| 295 | } else { |
| 296 | ScopedSetTracerPID scoped_set_tracer_pid(tracer_pid); |
| 297 | |
| 298 | tracer_thread_argument.mutex.Unlock(); |
| 299 | |
| 300 | while (atomic_load(&tracer_thread_argument.done, memory_order_relaxed) == 0) |
| 301 | sched_yield(); |
| 302 | |
| 303 | for (;;) { |
| 304 | uptr waitpid_status = internal_waitpid(tracer_pid, nullptr, __WALL); |
| 305 | if (!internal_iserror(waitpid_status, &local_errno)) |
| 306 | break; |
| 307 | if (local_errno == EINTR) |
| 308 | continue; |
| 309 | VReport(1, "Waiting on the tracer thread failed (errno %d).\n" , |
| 310 | local_errno); |
| 311 | break; |
| 312 | } |
| 313 | } |
| 314 | } |
| 315 | |
| 316 | tid_t SuspendedThreadsListNetBSD::GetThreadID(uptr index) const { |
| 317 | CHECK_LT(index, thread_ids_.size()); |
| 318 | return thread_ids_[index]; |
| 319 | } |
| 320 | |
| 321 | uptr SuspendedThreadsListNetBSD::ThreadCount() const { |
| 322 | return thread_ids_.size(); |
| 323 | } |
| 324 | |
| 325 | bool SuspendedThreadsListNetBSD::ContainsTid(tid_t thread_id) const { |
| 326 | for (uptr i = 0; i < thread_ids_.size(); i++) { |
| 327 | if (thread_ids_[i] == thread_id) |
| 328 | return true; |
| 329 | } |
| 330 | return false; |
| 331 | } |
| 332 | |
| 333 | void SuspendedThreadsListNetBSD::Append(tid_t tid) { |
| 334 | thread_ids_.push_back(tid); |
| 335 | } |
| 336 | |
| 337 | PtraceRegistersStatus SuspendedThreadsListNetBSD::GetRegistersAndSP( |
| 338 | uptr index, InternalMmapVector<uptr> *buffer, uptr *sp) const { |
| 339 | lwpid_t tid = GetThreadID(index); |
| 340 | pid_t ppid = internal_getppid(); |
| 341 | struct reg regs; |
| 342 | int pterrno; |
| 343 | bool isErr = |
| 344 | internal_iserror(internal_ptrace(PT_GETREGS, ppid, ®s, tid), &pterrno); |
| 345 | if (isErr) { |
| 346 | VReport(1, |
| 347 | "Could not get registers from process %d thread %d (errno %d).\n" , |
| 348 | ppid, tid, pterrno); |
| 349 | return pterrno == ESRCH ? REGISTERS_UNAVAILABLE_FATAL |
| 350 | : REGISTERS_UNAVAILABLE; |
| 351 | } |
| 352 | |
| 353 | *sp = PTRACE_REG_SP(®s); |
| 354 | buffer->resize(RoundUpTo(sizeof(regs), sizeof(uptr)) / sizeof(uptr)); |
| 355 | internal_memcpy(buffer->data(), ®s, sizeof(regs)); |
| 356 | |
| 357 | return REGISTERS_AVAILABLE; |
| 358 | } |
| 359 | |
| 360 | } // namespace __sanitizer |
| 361 | |
| 362 | #endif |
| 363 | |