1 #include "private/pthread_support.h"
3 #if defined(GC_PTHREADS) && !defined(GC_SOLARIS_THREADS) \
4 && !defined(GC_IRIX_THREADS) && !defined(GC_WIN32_THREADS) \
5 && !defined(GC_DARWIN_THREADS) && !defined(GC_AIX_THREADS) \
6 && !defined(GC_OPENBSD_THREADS)
14 /* work around a dlopen issue (bug #75390), undefs to avoid warnings with redefinitions */
15 #undef PACKAGE_BUGREPORT
18 #undef PACKAGE_TARNAME
19 #undef PACKAGE_VERSION
20 #include "mono/utils/mono-compiler.h"
22 #ifdef MONO_DEBUGGER_SUPPORTED
23 #include "include/libgc-mono-debugger.h"
27 volatile int __nacl_thread_suspension_needed = 0;
28 pthread_t nacl_thread_parker = -1;
30 volatile int nacl_thread_parked[MAX_NACL_GC_THREADS];
31 volatile int nacl_thread_used[MAX_NACL_GC_THREADS];
32 volatile int nacl_thread_parking_inited = 0;
33 volatile int nacl_num_gc_threads = 0;
34 pthread_mutex_t nacl_thread_alloc_lock = PTHREAD_MUTEX_INITIALIZER;
35 __thread int nacl_thread_idx = -1;
36 __thread GC_thread nacl_gc_thread_self = NULL;
43 # define NSIG (MAXSIG+1)
46 # elif defined(__SIGRTMAX)
47 # define NSIG (__SIGRTMAX+1)
54 void GC_print_sig_mask()
59 if (pthread_sigmask(SIG_BLOCK, NULL, &blocked) != 0)
60 ABORT("pthread_sigmask");
61 GC_printf0("Blocked: ");
62 for (i = 1; i < NSIG; i++) {
63 if (sigismember(&blocked, i)) { GC_printf1("%ld ",(long) i); }
70 /* Remove the signals that we want to allow in thread stopping */
71 /* handler from a set. */
72 void GC_remove_allowed_signals(sigset_t *set)
75 if (sigdelset(set, SIGINT) != 0
76 || sigdelset(set, SIGQUIT) != 0
77 || sigdelset(set, SIGABRT) != 0
78 || sigdelset(set, SIGTERM) != 0) {
79 ABORT("sigdelset() failed");
84 /* Handlers write to the thread structure, which is in the heap, */
85 /* and hence can trigger a protection fault. */
86 if (sigdelset(set, SIGSEGV) != 0
88 || sigdelset(set, SIGBUS) != 0
91 ABORT("sigdelset() failed");
96 static sigset_t suspend_handler_mask;
98 word GC_stop_count; /* Incremented at the beginning of GC_stop_world. */
100 #ifdef GC_OSF1_THREADS
101 GC_bool GC_retry_signals = TRUE;
103 GC_bool GC_retry_signals = FALSE;
107 * We use signals to stop threads during GC.
109 * Suspended threads wait in signal handler for SIG_THR_RESTART.
110 * That's more portable than semaphores or condition variables.
111 * (We do use sem_post from a signal handler, but that should be portable.)
113 * The thread suspension signal SIG_SUSPEND is now defined in gc_priv.h.
114 * Note that we can't just stop a thread; we need it to save its stack
115 * pointer(s) and acknowledge.
118 #ifndef SIG_THR_RESTART
119 # if defined(GC_HPUX_THREADS) || defined(GC_OSF1_THREADS)
121 # define SIG_THR_RESTART _SIGRTMIN + 5
123 # define SIG_THR_RESTART SIGRTMIN + 5
126 # define SIG_THR_RESTART SIGXCPU
130 sem_t GC_suspend_ack_sem;
132 static void _GC_suspend_handler(int sig)
136 pthread_t my_thread = pthread_self();
138 # ifdef PARALLEL_MARK
139 word my_mark_no = GC_mark_no;
140 /* Marker can't proceed until we acknowledge. Thus this is */
141 /* guaranteed to be the mark_no correspending to our */
142 /* suspension, i.e. the marker can't have incremented it yet. */
144 word my_stop_count = GC_stop_count;
146 if (sig != SIG_SUSPEND) ABORT("Bad signal in suspend_handler");
149 GC_printf1("Suspending 0x%lx\n", my_thread);
152 me = GC_lookup_thread(my_thread);
153 /* The lookup here is safe, since I'm doing this on behalf */
154 /* of a thread which holds the allocation lock in order */
155 /* to stop the world. Thus concurrent modification of the */
156 /* data structure is impossible. */
157 if (me -> stop_info.last_stop_count == my_stop_count) {
158 /* Duplicate signal. OK if we are retrying. */
159 if (!GC_retry_signals) {
160 WARN("Duplicate suspend signal in thread %lx\n",
166 me -> stop_info.stack_ptr = (ptr_t)GC_save_regs_in_stack();
168 me -> stop_info.stack_ptr = (ptr_t)(&dummy);
171 me -> backing_store_ptr = (ptr_t)GC_save_regs_in_stack();
174 /* Tell the thread that wants to stop the world that this */
175 /* thread has been stopped. Note that sem_post() is */
176 /* the only async-signal-safe primitive in LinuxThreads. */
177 sem_post(&GC_suspend_ack_sem);
178 me -> stop_info.last_stop_count = my_stop_count;
180 /* Wait until that thread tells us to restart by sending */
181 /* this thread a SIG_THR_RESTART signal. */
182 /* SIG_THR_RESTART should be masked at this point. Thus there */
185 me->stop_info.signal = 0;
186 sigsuspend(&suspend_handler_mask); /* Wait for signal */
187 } while (me->stop_info.signal != SIG_THR_RESTART);
188 /* If the RESTART signal gets lost, we can still lose. That should be */
189 /* less likely than losing the SUSPEND signal, since we don't do much */
190 /* between the sem_post and sigsuspend. */
191 /* We'd need more handshaking to work around that, since we don't want */
192 /* to accidentally leave a RESTART signal pending, thus causing us to */
193 /* continue prematurely in a future round. */
195 /* Tell the thread that wants to start the world that this */
196 /* thread has been started. Note that sem_post() is */
197 /* the only async-signal-safe primitive in LinuxThreads. */
198 sem_post(&GC_suspend_ack_sem);
202 GC_printf1("Continuing 0x%lx\n", my_thread);
208 void GC_suspend_handler(int sig)
210 int old_errno = errno;
211 _GC_suspend_handler(sig);
215 static void _GC_restart_handler(int sig)
217 pthread_t my_thread = pthread_self();
220 if (sig != SIG_THR_RESTART) ABORT("Bad signal in suspend_handler");
222 /* Let the GC_suspend_handler() know that we got a SIG_THR_RESTART. */
223 /* The lookup here is safe, since I'm doing this on behalf */
224 /* of a thread which holds the allocation lock in order */
225 /* to stop the world. Thus concurrent modification of the */
226 /* data structure is impossible. */
227 me = GC_lookup_thread(my_thread);
228 me->stop_info.signal = SIG_THR_RESTART;
231 ** Note: even if we didn't do anything useful here,
232 ** it would still be necessary to have a signal handler,
233 ** rather than ignoring the signals, otherwise
234 ** the signals will not be delivered at all, and
235 ** will thus not interrupt the sigsuspend() above.
239 GC_printf1("In GC_restart_handler for 0x%lx\n", pthread_self());
244 # define IF_IA64(x) x
248 /* We hold allocation lock. Should do exactly the right thing if the */
249 /* world is stopped. Should not fail if it isn't. */
250 static void pthread_push_all_stacks()
252 GC_bool found_me = FALSE;
256 /* On IA64, we also need to scan the register backing store. */
257 IF_IA64(ptr_t bs_lo; ptr_t bs_hi;)
258 pthread_t me = pthread_self();
260 if (!GC_thr_initialized) GC_thr_init();
262 GC_printf1("Pushing stacks from thread 0x%lx\n", (unsigned long) me);
264 for (i = 0; i < THREAD_TABLE_SZ; i++) {
265 for (p = GC_threads[i]; p != 0; p = p -> next) {
266 if (p -> flags & FINISHED) continue;
267 if (pthread_equal(p -> id, me)) {
269 lo = (ptr_t)GC_save_regs_in_stack();
274 IF_IA64(bs_hi = (ptr_t)GC_save_regs_in_stack();)
276 lo = p -> stop_info.stack_ptr;
277 IF_IA64(bs_hi = p -> backing_store_ptr;)
279 if ((p -> flags & MAIN_THREAD) == 0) {
281 IF_IA64(bs_lo = p -> backing_store_end);
283 /* The original stack. */
285 IF_IA64(bs_lo = BACKING_STORE_BASE;)
288 GC_printf3("Stack for thread 0x%lx = [%lx,%lx)\n",
289 (unsigned long) p -> id,
290 (unsigned long) lo, (unsigned long) hi);
292 if (0 == lo) ABORT("GC_push_all_stacks: sp not set!\n");
293 if (p->altstack && lo >= p->altstack && lo <= p->altstack + p->altstack_size)
294 hi = p->altstack + p->altstack_size;
295 /* FIXME: Need to scan the normal stack too, but how ? */
297 # ifdef STACK_GROWS_UP
298 /* We got them backwards! */
299 GC_push_all_stack(hi, lo);
301 GC_push_all_stack(lo, hi);
304 /* Push reg_storage as roots, this will cover the reg context */
305 GC_push_all_stack(p -> stop_info.reg_storage, p -> stop_info.reg_storage + NACL_GC_REG_STORAGE_SIZE);
309 GC_printf3("Reg stack for thread 0x%lx = [%lx,%lx)\n",
310 (unsigned long) p -> id,
311 (unsigned long) bs_lo, (unsigned long) bs_hi);
313 if (pthread_equal(p -> id, me)) {
314 GC_push_all_eager(bs_lo, bs_hi);
316 GC_push_all_stack(bs_lo, bs_hi);
321 if (!found_me && !GC_in_thread_creation)
322 ABORT("Collecting from unknown thread.");
325 void GC_restart_handler(int sig)
327 int old_errno = errno;
328 _GC_restart_handler (sig);
332 /* We hold allocation lock. Should do exactly the right thing if the */
333 /* world is stopped. Should not fail if it isn't. */
334 void GC_push_all_stacks()
336 pthread_push_all_stacks();
339 /* There seems to be a very rare thread stopping problem. To help us */
340 /* debug that, we save the ids of the stopping thread. */
341 pthread_t GC_stopping_thread;
344 #ifdef PLATFORM_ANDROID
346 int android_thread_kill(pid_t tid, int sig)
349 int old_errno = errno;
351 ret = tkill(tid, sig);
361 /* We hold the allocation lock. Suspend all threads that might */
362 /* still be running. Return the number of suspend signals that */
367 int n_live_threads = 0;
371 pthread_t my_thread = pthread_self();
373 GC_stopping_thread = my_thread; /* debugging only. */
374 GC_stopping_pid = getpid(); /* debugging only. */
375 for (i = 0; i < THREAD_TABLE_SZ; i++) {
376 for (p = GC_threads[i]; p != 0; p = p -> next) {
377 if (p -> id != my_thread) {
378 if (p -> flags & FINISHED) continue;
379 if (p -> stop_info.last_stop_count == GC_stop_count) continue;
380 if (p -> thread_blocked) /* Will wait */ continue;
383 GC_printf1("Sending suspend signal to 0x%lx\n", p -> id);
386 #ifndef PLATFORM_ANDROID
387 result = pthread_kill(p -> id, SIG_SUSPEND);
389 result = android_thread_kill(p -> kernel_id, SIG_SUSPEND);
393 /* Not really there anymore. Possible? */
399 ABORT("pthread_kill failed");
404 return n_live_threads;
410 /* Caller holds allocation lock. */
411 static void pthread_stop_world()
419 GC_printf1("Stopping the world from 0x%lx\n", pthread_self());
422 n_live_threads = GC_suspend_all();
424 if (GC_retry_signals) {
425 unsigned long wait_usecs = 0; /* Total wait since retry. */
426 # define WAIT_UNIT 3000
427 # define RETRY_INTERVAL 100000
431 sem_getvalue(&GC_suspend_ack_sem, &ack_count);
432 if (ack_count == n_live_threads) break;
433 if (wait_usecs > RETRY_INTERVAL) {
434 int newly_sent = GC_suspend_all();
437 if (GC_print_stats) {
438 GC_printf1("Resent %ld signals after timeout\n",
442 sem_getvalue(&GC_suspend_ack_sem, &ack_count);
443 if (newly_sent < n_live_threads - ack_count) {
444 WARN("Lost some threads during GC_stop_world?!\n",0);
445 n_live_threads = ack_count + newly_sent;
450 wait_usecs += WAIT_UNIT;
453 for (i = 0; i < n_live_threads; i++) {
454 while (0 != (code = sem_wait(&GC_suspend_ack_sem))) {
455 if (errno != EINTR) {
456 GC_err_printf1("Sem_wait returned %ld\n", (unsigned long)code);
457 ABORT("sem_wait for handler failed");
462 GC_printf1("World stopped from 0x%lx\n", pthread_self());
464 GC_stopping_thread = 0; /* debugging only */
471 GC_printf1("pthread_stop_world: num_threads %d\n", nacl_num_gc_threads - 1);
473 nacl_thread_parker = pthread_self();
474 __nacl_thread_suspension_needed = 1;
477 #define NACL_PARK_WAIT_NANOSECONDS 100000
478 #define NANOS_PER_SECOND 1000000000
479 int num_threads_parked = 0;
482 /* Check the 'parked' flag for each thread the GC knows about */
483 for (i = 0; i < MAX_NACL_GC_THREADS && num_used < nacl_num_gc_threads; i++) {
484 if (nacl_thread_used[i] == 1) {
486 if (nacl_thread_parked[i] == 1) {
487 num_threads_parked++;
491 /* -1 for the current thread */
492 if (num_threads_parked >= nacl_num_gc_threads - 1)
495 ts.tv_nsec = NACL_PARK_WAIT_NANOSECONDS;
497 GC_printf1("sleeping waiting for %d threads to park...\n", nacl_num_gc_threads - num_threads_parked - 1);
500 if (++num_sleeps > NANOS_PER_SECOND / NACL_PARK_WAIT_NANOSECONDS) {
501 GC_printf1("GC appears stalled waiting for %d threads to park...\n", nacl_num_gc_threads - num_threads_parked - 1);
514 #define NACL_STORE_REGS() \
516 __asm__ __volatile__ ("push %rbx");\
517 __asm__ __volatile__ ("push %rbp");\
518 __asm__ __volatile__ ("push %r12");\
519 __asm__ __volatile__ ("push %r13");\
520 __asm__ __volatile__ ("push %r14");\
521 __asm__ __volatile__ ("push %r15");\
522 __asm__ __volatile__ ("mov %%esp, %0" : "=m" (nacl_gc_thread_self->stop_info.stack_ptr));\
523 memcpy(nacl_gc_thread_self->stop_info.reg_storage, nacl_gc_thread_self->stop_info.stack_ptr, NACL_GC_REG_STORAGE_SIZE * sizeof(ptr_t));\
524 __asm__ __volatile__ ("naclasp $48, %r15");\
529 #define NACL_STORE_REGS() \
531 __asm__ __volatile__ ("push %ebx");\
532 __asm__ __volatile__ ("push %ebp");\
533 __asm__ __volatile__ ("push %esi");\
534 __asm__ __volatile__ ("push %edi");\
535 __asm__ __volatile__ ("mov %%esp, %0" : "=m" (nacl_gc_thread_self->stop_info.stack_ptr));\
536 memcpy(nacl_gc_thread_self->stop_info.reg_storage, nacl_gc_thread_self->stop_info.stack_ptr, NACL_GC_REG_STORAGE_SIZE * sizeof(ptr_t));\
537 __asm__ __volatile__ ("add $16, %esp");\
542 #define NACL_STORE_REGS() \
544 __asm__ __volatile__ ("push {r4-r12,lr}");\
545 __asm__ __volatile__ ("mov r0, %0" : : "r" (&nacl_gc_thread_self->stop_info.stack_ptr)); \
546 __asm__ __volatile__ ("bic r0, r0, #0xc0000000");\
547 __asm__ __volatile__ ("str sp, [r0]");\
548 memcpy(nacl_gc_thread_self->stop_info.reg_storage, nacl_gc_thread_self->stop_info.stack_ptr, NACL_GC_REG_STORAGE_SIZE * sizeof(ptr_t));\
549 __asm__ __volatile__ ("add sp, sp, #40");\
550 __asm__ __volatile__ ("bic sp, sp, #0xc0000000");\
554 #error "Please port NACL_STORE_REGS"
558 void nacl_pre_syscall_hook()
561 if (nacl_thread_idx != -1) {
563 nacl_gc_thread_self->stop_info.stack_ptr = (ptr_t)(&local_dummy);
564 nacl_thread_parked[nacl_thread_idx] = 1;
568 void __nacl_suspend_thread_if_needed();
570 void nacl_post_syscall_hook()
572 /* Calling __nacl_suspend_thread_if_needed() right away should guarantee we don't mutate the GC set. */
573 __nacl_suspend_thread_if_needed();
574 if (nacl_thread_idx != -1) {
575 nacl_thread_parked[nacl_thread_idx] = 0;
579 void __nacl_suspend_thread_if_needed() {
580 if (__nacl_thread_suspension_needed) {
581 pthread_t self = pthread_self();
583 /* Don't try to park the thread parker. */
584 if (nacl_thread_parker == self)
587 /* This can happen when a thread is created */
588 /* outside of the GC system (wthread mostly). */
589 if (nacl_thread_idx < 0)
592 /* If it was already 'parked', we're returning from a syscall, */
593 /* so don't bother storing registers again, the GC has a set. */
594 if (!nacl_thread_parked[nacl_thread_idx]) {
596 nacl_gc_thread_self->stop_info.stack_ptr = (ptr_t)(&local_dummy);
598 nacl_thread_parked[nacl_thread_idx] = 1;
599 while (__nacl_thread_suspension_needed)
601 nacl_thread_parked[nacl_thread_idx] = 0;
603 /* Clear out the reg storage for next suspend. */
604 memset(nacl_gc_thread_self->stop_info.reg_storage, 0, NACL_GC_REG_STORAGE_SIZE * sizeof(ptr_t));
610 /* Caller holds allocation lock. */
614 GC_notify_event (GC_EVENT_PRE_STOP_WORLD);
615 /* Make sure all free list construction has stopped before we start. */
616 /* No new construction can start, since free list construction is */
617 /* required to acquire and release the GC lock before it starts, */
618 /* and we have the lock. */
619 # ifdef PARALLEL_MARK
620 GC_acquire_mark_lock();
621 GC_ASSERT(GC_fl_builder_count == 0);
622 /* We should have previously waited for it to become zero. */
623 # endif /* PARALLEL_MARK */
625 #ifdef MONO_DEBUGGER_SUPPORTED
626 if (gc_thread_vtable && gc_thread_vtable->stop_world)
627 gc_thread_vtable->stop_world ();
630 pthread_stop_world ();
631 # ifdef PARALLEL_MARK
632 GC_release_mark_lock();
635 GC_notify_event (GC_EVENT_POST_STOP_WORLD);
638 /* Caller holds allocation lock, and has held it continuously since */
639 /* the world stopped. */
640 static void pthread_start_world()
643 pthread_t my_thread = pthread_self();
645 register GC_thread p;
646 register int n_live_threads = 0;
651 GC_printf0("World starting\n");
654 GC_notify_event (GC_EVENT_PRE_START_WORLD);
656 for (i = 0; i < THREAD_TABLE_SZ; i++) {
657 for (p = GC_threads[i]; p != 0; p = p -> next) {
658 if (p -> id != my_thread) {
659 if (p -> flags & FINISHED) continue;
660 if (p -> thread_blocked) continue;
663 GC_printf1("Sending restart signal to 0x%lx\n", p -> id);
666 #ifndef PLATFORM_ANDROID
667 result = pthread_kill(p -> id, SIG_THR_RESTART);
669 result = android_thread_kill(p -> kernel_id, SIG_THR_RESTART);
673 /* Not really there anymore. Possible? */
679 ABORT("pthread_kill failed");
686 GC_printf0 ("All threads signaled");
689 for (i = 0; i < n_live_threads; i++) {
690 while (0 != (code = sem_wait(&GC_suspend_ack_sem))) {
691 if (errno != EINTR) {
692 GC_err_printf1("Sem_wait returned %ld\n", (unsigned long)code);
693 ABORT("sem_wait for handler failed");
699 GC_notify_event (GC_EVENT_POST_START_WORLD);
701 GC_printf0("World started\n");
705 GC_notify_event (GC_EVENT_PRE_START_WORLD);
707 GC_printf0("World starting\n");
709 __nacl_thread_suspension_needed = 0;
711 GC_notify_event (GC_EVENT_POST_START_WORLD);
715 void GC_start_world()
717 #ifdef MONO_DEBUGGER_SUPPORTED
718 if (gc_thread_vtable && gc_thread_vtable->start_world)
719 gc_thread_vtable->start_world();
722 pthread_start_world ();
725 static void pthread_stop_init() {
727 struct sigaction act;
729 if (sem_init(&GC_suspend_ack_sem, 0, 0) != 0)
730 ABORT("sem_init failed");
732 act.sa_flags = SA_RESTART;
733 if (sigfillset(&act.sa_mask) != 0) {
734 ABORT("sigfillset() failed");
736 GC_remove_allowed_signals(&act.sa_mask);
737 /* SIG_THR_RESTART is set in the resulting mask. */
738 /* It is unmasked by the handler when necessary. */
739 act.sa_handler = GC_suspend_handler;
740 if (sigaction(SIG_SUSPEND, &act, NULL) != 0) {
741 ABORT("Cannot set SIG_SUSPEND handler");
744 act.sa_handler = GC_restart_handler;
745 if (sigaction(SIG_THR_RESTART, &act, NULL) != 0) {
746 ABORT("Cannot set SIG_THR_RESTART handler");
749 /* Inititialize suspend_handler_mask. It excludes SIG_THR_RESTART. */
750 if (sigfillset(&suspend_handler_mask) != 0) ABORT("sigfillset() failed");
751 GC_remove_allowed_signals(&suspend_handler_mask);
752 if (sigdelset(&suspend_handler_mask, SIG_THR_RESTART) != 0)
753 ABORT("sigdelset() failed");
755 /* Check for GC_RETRY_SIGNALS. */
756 if (0 != GETENV("GC_RETRY_SIGNALS")) {
757 GC_retry_signals = TRUE;
759 if (0 != GETENV("GC_NO_RETRY_SIGNALS")) {
760 GC_retry_signals = FALSE;
763 if (GC_print_stats && GC_retry_signals) {
764 GC_printf0("Will retry suspend signal if necessary.\n");
770 /* We hold the allocation lock. */
773 #ifdef MONO_DEBUGGER_SUPPORTED
774 if (gc_thread_vtable && gc_thread_vtable->initialize)
775 gc_thread_vtable->initialize ();
778 pthread_stop_init ();
781 #ifdef MONO_DEBUGGER_SUPPORTED
783 GCThreadFunctions *gc_thread_vtable = NULL;
786 GC_mono_debugger_get_stack_ptr (void)
790 me = GC_lookup_thread (pthread_self ());
791 return &me->stop_info.stack_ptr;