X-Git-Url: http://wien.tomnetworks.com/gitweb/?a=blobdiff_plain;f=mono%2Fmetadata%2Fthreadpool.c;h=61f69437e3fec21439d40698091adfaeff9c2d0a;hb=83c3ff04e80cbffc33b087e084dda74b911c4a61;hp=cf421dacb56f29859147685d5d26d8d18669e7c3;hpb=fcd84c6ca49020f733101af6bb490018262b88c5;p=mono.git diff --git a/mono/metadata/threadpool.c b/mono/metadata/threadpool.c old mode 100644 new mode 100755 index cf421dacb56..61f69437e3f --- a/mono/metadata/threadpool.c +++ b/mono/metadata/threadpool.c @@ -6,15 +6,17 @@ * Gonzalo Paniagua Javier (gonzalo@ximian.com) * * Copyright 2001-2003 Ximian, Inc (http://www.ximian.com) - * Copyright 2004-2009 Novell, Inc (http://www.novell.com) + * Copyright 2004-2010 Novell, Inc (http://www.novell.com) */ #include #include -#define THREADS_PER_CPU 10 /* 8 + THREADS_PER_CPU * number of CPUs = max threads */ -#define THREAD_EXIT_TIMEOUT 1000 -#define INITIAL_QUEUE_LENGTH 128 +#ifdef MONO_SMALL_CONFIG +#define QUEUE_LENGTH 16 /* Must be 2^N */ +#else +#define QUEUE_LENGTH 64 /* Must be 2^N */ +#endif #include #include @@ -28,10 +30,12 @@ #include #include #include +#include #include #include #include #include +#include #include #ifdef HAVE_SYS_TIME_H #include @@ -59,8 +63,29 @@ #define THREAD_WANTS_A_BREAK(t) ((t->state & (ThreadState_StopRequested | \ ThreadState_SuspendRequested)) != 0) -#undef EPOLL_DEBUG -// +#define SPIN_TRYLOCK(i) (InterlockedCompareExchange (&(i), 1, 0) == 0) +#define SPIN_LOCK(i) do { \ + if (SPIN_TRYLOCK (i)) \ + break; \ + } while (1) + +#define SPIN_UNLOCK(i) i = 0 + +#define EPOLL_DEBUG(...) +#define EPOLL_DEBUG_STMT(...) +#define TP_DEBUG(...) +#define TP_DEBUG_STMT(...) + +/* DEBUG: prints tp data every 2s */ +#undef DEBUG + +/* +#define EPOLL_DEBUG(...) g_message(__VA_ARGS__) +#define EPOLL_DEBUG_STMT(...) do { __VA_ARGS__ } while (0) +#define TP_DEBUG(...) g_message(__VA_ARGS__) +#define TP_DEBUG_STMT(...) do { __VA_ARGS__ } while (0) +*/ + /* map of CounterSample.cs */ struct _MonoCounterSample { gint64 rawValue; @@ -74,13 +99,11 @@ struct _MonoCounterSample { }; /* mono_thread_pool_init called */ -static int tp_inited; - -static int pending_io_items; +static volatile int tp_inited; typedef struct { CRITICAL_SECTION io_lock; /* access to sock_to_state */ - int inited; + int inited; // 0 -> not initialized , 1->initializing, 2->initialized, 3->cleaned up int pipe [2]; MonoGHashTable *sock_to_state; @@ -103,49 +126,51 @@ typedef struct { MonoObject *state; MonoObject *res; MonoArray *out_args; - /* This is a HANDLE, we use guint64 so the managed object layout remains constant */ - /* THIS FIELD IS NOT USED ANY MORE. Remove it when we feel like breaking corlib compatibility with 2.6 */ - guint64 wait_event; } ASyncCall; typedef struct { - CRITICAL_SECTION lock; - MonoArray *array; - int first_elem; - int next_elem; + MonoSemType lock; + MonoMList *first; /* GC root */ + MonoMList *last; + MonoMList *unused; /* Up to 20 chunks. GC root */ + gint head; + gint tail; + MonoSemType new_job; + volatile gint waiting; /* threads waiting for a work item */ /**/ - GQueue *idle_threads; - int idle_started; /* Have we started the idle threads? Interlocked */ + volatile gint pool_status; /* 0 -> not initialized, 1 -> initialized, 2 -> cleaning up */ /* min, max, n and busy -> Interlocked */ - int min_threads; - int max_threads; - int nthreads; - int busy_threads; + volatile gint min_threads; + volatile gint max_threads; + volatile gint nthreads; + volatile gint busy_threads; void (*async_invoke) (gpointer data); void *pc_nitems; /* Performance counter for total number of items in added */ - /* We don't need the rate here since we can compute the different ourselves */ - /* void *perfc_rate; */ - MonoCounterSample last_sample; - + void *pc_nthreads; /* Performance counter for total number of active threads */ + /**/ + volatile gint destroy_thread; + volatile gint ignore_times; /* Used when there's a thread being created or destroyed */ + volatile gint sp_lock; /* spin lock used to protect ignore_times */ + volatile gint64 last_check; + volatile gint64 time_sum; + volatile gint n_sum; + gint64 averages [2]; + /**/ + //TP_DEBUG_ONLY (gint nodes_created); + //TP_DEBUG_ONLY (gint nodes_reused); + gboolean is_io; } ThreadPool; static ThreadPool async_tp; static ThreadPool async_io_tp; -typedef struct { - HANDLE wait_handle; - gpointer data; - gint timeout; - gboolean die; -} IdleThreadData; - static void async_invoke_thread (gpointer data); -static void mono_async_invoke (MonoAsyncResult *ares); +static MonoObject *mono_async_invoke (ThreadPool *tp, MonoAsyncResult *ares); static void threadpool_free_queue (ThreadPool *tp); static void threadpool_append_job (ThreadPool *tp, MonoObject *ar); -static void *threadpool_queue_idle_thread (ThreadPool *tp, IdleThreadData *it); +static void threadpool_append_jobs (ThreadPool *tp, MonoObject **jobs, gint njobs); static void threadpool_init (ThreadPool *tp, int min_threads, int max_threads, void (*async_invoke) (gpointer)); static void threadpool_start_idle_threads (ThreadPool *tp); static void threadpool_kill_idle_threads (ThreadPool *tp); @@ -154,6 +179,17 @@ static MonoClass *async_call_klass; static MonoClass *socket_async_call_klass; static MonoClass *process_async_call_klass; +static GPtrArray *wsqs; +CRITICAL_SECTION wsqs_lock; + +/* Hooks */ +static MonoThreadPoolFunc tp_start_func; +static MonoThreadPoolFunc tp_finish_func; +static gpointer tp_hooks_user_data; +static MonoThreadPoolItemFunc tp_item_begin_func; +static MonoThreadPoolItemFunc tp_item_end_func; +static gpointer tp_item_user_data; + #define INIT_POLLFD(a, b, c) {(a)->fd = b; (a)->events = c; (a)->revents = 0;} enum { AIO_OP_FIRST, @@ -170,22 +206,42 @@ enum { }; #ifdef DISABLE_SOCKETS + #define socket_io_cleanup(x) + +static int +get_event_from_state (MonoSocketAsyncResult *state) +{ + g_assert_not_reached (); + return -1; +} + +static int +get_events_from_list (MonoMList *list) +{ + return 0; +} + #else + static void socket_io_cleanup (SocketIOData *data) { - if (data->inited == 0) + EnterCriticalSection (&data->io_lock); + if (data->inited != 2) { + LeaveCriticalSection (&data->io_lock); return; + } + data->inited = 3; - EnterCriticalSection (&data->io_lock); - data->inited = 0; -#ifdef PLATFORM_WIN32 +#ifdef HOST_WIN32 closesocket (data->pipe [0]); closesocket (data->pipe [1]); #else - close (data->pipe [0]); - close (data->pipe [1]); + if (data->pipe [0] > -1) + close (data->pipe [0]); + if (data->pipe [1] > -1) + close (data->pipe [1]); #endif data->pipe [0] = -1; data->pipe [1] = -1; @@ -194,10 +250,6 @@ socket_io_cleanup (SocketIOData *data) data->new_sem = NULL; mono_g_hash_table_destroy (data->sock_to_state); data->sock_to_state = NULL; - EnterCriticalSection (&async_io_tp.lock); - threadpool_free_queue (&async_io_tp); - threadpool_kill_idle_threads (&async_io_tp); - LeaveCriticalSection (&async_io_tp.lock); g_free (data->newpfd); data->newpfd = NULL; #ifdef HAVE_EPOLL @@ -223,7 +275,7 @@ get_event_from_state (MonoSocketAsyncResult *state) case AIO_OP_CONNECT: return MONO_POLLOUT; default: /* Should never happen */ - g_print ("get_event_from_state: unknown value in switch!!!\n"); + g_message ("get_event_from_state: unknown value in switch!!!"); return 0; } } @@ -271,99 +323,38 @@ threadpool_jobs_dec (MonoObject *obj) return FALSE; } -#ifndef DISABLE_SOCKETS -static void -async_invoke_io_thread (gpointer data) +#ifdef HAVE_EPOLL +static MonoObject * +get_io_event (MonoMList **list, gint event) { - MonoDomain *domain; - MonoInternalThread *thread; - const gchar *version; - IdleThreadData idle_data = {0}; - - idle_data.timeout = INFINITE; - idle_data.wait_handle = CreateEvent (NULL, FALSE, FALSE, NULL); - - thread = mono_thread_internal_current (); - - version = mono_get_runtime_info ()->framework_version; - for (;;) { - MonoSocketAsyncResult *state; - MonoAsyncResult *ar; - - state = (MonoSocketAsyncResult *) data; - if (state) { - InterlockedDecrement (&pending_io_items); - ar = state->ares; - switch (state->operation) { - case AIO_OP_RECEIVE: - state->total = ICALL_RECV (state); - break; - case AIO_OP_SEND: - state->total = ICALL_SEND (state); - break; - } - - /* worker threads invokes methods in different domains, - * so we need to set the right domain here */ - domain = ((MonoObject *)ar)->vtable->domain; - - g_assert (domain); + MonoObject *state; + MonoMList *current; + MonoMList *prev; - if (domain->state == MONO_APPDOMAIN_UNLOADED || domain->state == MONO_APPDOMAIN_UNLOADING) { - threadpool_jobs_dec ((MonoObject *)ar); - data = NULL; - } else { - mono_thread_push_appdomain_ref (domain); - if (threadpool_jobs_dec ((MonoObject *)ar)) { - data = NULL; - mono_thread_pop_appdomain_ref (); - continue; - } - if (mono_domain_set (domain, FALSE)) { - ASyncCall *ac; - - mono_async_invoke (ar); - ac = (ASyncCall *) ar->object_data; - /* - if (ac->msg->exc != NULL) - mono_unhandled_exception (ac->msg->exc); - */ - mono_domain_set (mono_get_root_domain (), TRUE); - } - mono_thread_pop_appdomain_ref (); - InterlockedDecrement (&async_io_tp.busy_threads); - /* If the callee changes the background status, set it back to TRUE */ - if (*version != '1' && !mono_thread_test_state (thread , ThreadState_Background)) - ves_icall_System_Threading_Thread_SetState (thread, ThreadState_Background); - } - } + current = *list; + prev = NULL; + state = NULL; + while (current) { + state = mono_mlist_get_data (current); + if (get_event_from_state ((MonoSocketAsyncResult *) state) == event) + break; - data = threadpool_queue_idle_thread (&async_io_tp, &idle_data); - while (!idle_data.die && !data) { - guint32 wr; - wr = WaitForSingleObjectEx (idle_data.wait_handle, idle_data.timeout, TRUE); - if (THREAD_WANTS_A_BREAK (thread)) - mono_thread_interruption_checkpoint (); - - if (wr != WAIT_TIMEOUT && wr != WAIT_IO_COMPLETION) { - data = idle_data.data; - idle_data.data = NULL; - break; /* We have to exit */ - } - } + state = NULL; + prev = current; + current = mono_mlist_next (current); + } - if (!data) { - InterlockedDecrement (&async_io_tp.nthreads); - CloseHandle (idle_data.wait_handle); - idle_data.wait_handle = NULL; - return; + if (current) { + if (prev) { + mono_mlist_set_next (prev, mono_mlist_next (current)); + } else { + *list = mono_mlist_next (*list); } - - InterlockedIncrement (&async_io_tp.busy_threads); } - g_assert_not_reached (); + return state; } +#endif static MonoMList * process_io_event (MonoMList *list, int event) @@ -383,10 +374,7 @@ process_io_event (MonoMList *list, int event) if (list != NULL) { oldlist = mono_mlist_remove_item (oldlist, list); -#ifdef EPOLL_DEBUG - g_print ("Dispatching event %d on socket %p\n", event, state->handle); -#endif - InterlockedIncrement (&pending_io_items); + EPOLL_DEBUG ("Dispatching event %d on socket %p", event, state->handle); threadpool_append_job (&async_io_tp, (MonoObject *) state); } @@ -420,7 +408,11 @@ mark_bad_fds (mono_pollfd *pfds, int nfds) static void socket_io_poll_main (gpointer p) { +#if MONO_SMALL_CONFIG +#define INITIAL_POLLFD_SIZE 128 +#else #define INITIAL_POLLFD_SIZE 1024 +#endif #define POLL_ERRORS (MONO_POLLERR | MONO_POLLHUP | MONO_POLLNVAL) SocketIOData *data = p; mono_pollfd *pfds; @@ -484,7 +476,8 @@ socket_io_poll_main (gpointer p) for (i = 1; i < allocated; i++) { pfd = &pfds [i]; - if (pfd->fd == -1 || pfd->fd == data->newpfd->fd) + if (pfd->fd == -1 || data->newpfd == NULL || + pfd->fd == data->newpfd->fd) break; } @@ -499,7 +492,7 @@ socket_io_poll_main (gpointer p) for (; i < allocated; i++) INIT_POLLFD (&pfds [i], -1, 0); } -#ifndef PLATFORM_WIN32 +#ifndef HOST_WIN32 nread = read (data->pipe [0], one, 1); #else nread = recv ((SOCKET) data->pipe [0], one, 1, 0); @@ -520,7 +513,7 @@ socket_io_poll_main (gpointer p) continue; EnterCriticalSection (&data->io_lock); - if (data->inited == 0) { + if (data->inited == 3) { g_free (pfds); LeaveCriticalSection (&data->io_lock); return; /* cleanup called */ @@ -557,6 +550,7 @@ socket_io_poll_main (gpointer p) #ifdef HAVE_EPOLL #define EPOLL_ERRORS (EPOLLERR | EPOLLHUP) +#define EPOLL_NEVENTS 128 static void socket_io_epoll_main (gpointer p) { @@ -564,13 +558,14 @@ socket_io_epoll_main (gpointer p) int epollfd; MonoInternalThread *thread; struct epoll_event *events, *evt; - const int nevents = 512; int ready = 0, i; + gpointer async_results [EPOLL_NEVENTS * 2]; // * 2 because each loop can add up to 2 results here + gint nresults; data = p; epollfd = data->epollfd; thread = mono_thread_internal_current (); - events = g_new0 (struct epoll_event, nevents); + events = g_new0 (struct epoll_event, EPOLL_NEVENTS); while (1) { do { @@ -578,83 +573,80 @@ socket_io_epoll_main (gpointer p) if (THREAD_WANTS_A_BREAK (thread)) mono_thread_interruption_checkpoint (); } -#ifdef EPOLL_DEBUG - g_print ("epoll_wait init\n"); -#endif - ready = epoll_wait (epollfd, events, nevents, -1); -#ifdef EPOLL_DEBUG - { - int err = errno; - g_print ("epoll_wait end with %d ready sockets (%d %s).\n", ready, err, (err) ? g_strerror (err) : ""); - errno = err; - } -#endif + EPOLL_DEBUG ("epoll_wait init"); + ready = epoll_wait (epollfd, events, EPOLL_NEVENTS, -1); + EPOLL_DEBUG_STMT( + int err = errno; + EPOLL_DEBUG ("epoll_wait end with %d ready sockets (%d %s).", ready, err, (ready == -1) ? g_strerror (err) : ""); + errno = err; + ); } while (ready == -1 && errno == EINTR); if (ready == -1) { int err = errno; g_free (events); if (err != EBADF) - g_warning ("epoll_wait: %d %s\n", err, g_strerror (err)); + g_warning ("epoll_wait: %d %s", err, g_strerror (err)); close (epollfd); return; } EnterCriticalSection (&data->io_lock); - if (data->inited == 0) { -#ifdef EPOLL_DEBUG - g_print ("data->inited == 0\n"); -#endif + if (data->inited == 3) { g_free (events); close (epollfd); + LeaveCriticalSection (&data->io_lock); return; /* cleanup called */ } + nresults = 0; for (i = 0; i < ready; i++) { int fd; MonoMList *list; + MonoObject *ares; evt = &events [i]; fd = evt->data.fd; list = mono_g_hash_table_lookup (data->sock_to_state, GINT_TO_POINTER (fd)); -#ifdef EPOLL_DEBUG - g_print ("Event %d on %d list length: %d\n", evt->events, fd, mono_mlist_length (list)); -#endif + EPOLL_DEBUG ("Event %d on %d list length: %d", evt->events, fd, mono_mlist_length (list)); if (list != NULL && (evt->events & (EPOLLIN | EPOLL_ERRORS)) != 0) { - list = process_io_event (list, MONO_POLLIN); + ares = get_io_event (&list, MONO_POLLIN); + if (ares != NULL) + async_results [nresults++] = ares; } if (list != NULL && (evt->events & (EPOLLOUT | EPOLL_ERRORS)) != 0) { - list = process_io_event (list, MONO_POLLOUT); + ares = get_io_event (&list, MONO_POLLOUT); + if (ares != NULL) + async_results [nresults++] = ares; } if (list != NULL) { mono_g_hash_table_replace (data->sock_to_state, GINT_TO_POINTER (fd), list); evt->events = get_events_from_list (list); -#ifdef EPOLL_DEBUG - g_print ("MOD %d to %d\n", fd, evt->events); -#endif + EPOLL_DEBUG ("MOD %d to %d", fd, evt->events); if (epoll_ctl (epollfd, EPOLL_CTL_MOD, fd, evt)) { if (epoll_ctl (epollfd, EPOLL_CTL_ADD, fd, evt) == -1) { -#ifdef EPOLL_DEBUG - int err = errno; - g_message ("epoll_ctl(MOD): %d %s fd: %d events: %d", err, g_strerror (err), fd, evt->events); - errno = err; -#endif + EPOLL_DEBUG_STMT ( + int err = errno; + EPOLL_DEBUG ("epoll_ctl(MOD): %d %s fd: %d events: %d", err, g_strerror (err), fd, evt->events); + errno = err; + ); } } } else { mono_g_hash_table_remove (data->sock_to_state, GINT_TO_POINTER (fd)); -#ifdef EPOLL_DEBUG - g_print ("DEL %d\n", fd); -#endif + EPOLL_DEBUG ("DEL %d", fd); epoll_ctl (epollfd, EPOLL_CTL_DEL, fd, evt); } } LeaveCriticalSection (&data->io_lock); + threadpool_append_jobs (&async_io_tp, (MonoObject **) async_results, nresults); + memset (async_results, 0, sizeof (gpointer) * nresults); } } +#undef EPOLL_NEVENTS #endif /* @@ -667,14 +659,17 @@ mono_thread_pool_remove_socket (int sock) MonoMList *list, *next; MonoSocketAsyncResult *state; - if (socket_io_data.inited == FALSE) + if (socket_io_data.inited == 0) return; EnterCriticalSection (&socket_io_data.io_lock); + if (socket_io_data.sock_to_state == NULL) { + LeaveCriticalSection (&socket_io_data.io_lock); + return; + } list = mono_g_hash_table_lookup (socket_io_data.sock_to_state, GINT_TO_POINTER (sock)); - if (list) { + if (list) mono_g_hash_table_remove (socket_io_data.sock_to_state, GINT_TO_POINTER (sock)); - } LeaveCriticalSection (&socket_io_data.io_lock); while (list) { @@ -693,7 +688,7 @@ mono_thread_pool_remove_socket (int sock) } } -#ifdef PLATFORM_WIN32 +#ifdef HOST_WIN32 static void connect_hack (gpointer x) { @@ -703,7 +698,7 @@ connect_hack (gpointer x) while (connect ((SOCKET) socket_io_data.pipe [1], (SOCKADDR *) addr, sizeof (struct sockaddr_in))) { Sleep (500); if (++count > 3) { - g_warning ("Error initializing async. sockets %d.\n", WSAGetLastError ()); + g_warning ("Error initializing async. sockets %d.", WSAGetLastError ()); g_assert (WSAGetLastError ()); } } @@ -713,25 +708,29 @@ connect_hack (gpointer x) static void socket_io_init (SocketIOData *data) { -#ifdef PLATFORM_WIN32 +#ifdef HOST_WIN32 struct sockaddr_in server; struct sockaddr_in client; SOCKET srv; int len; #endif int inited; + guint32 stack_size; - inited = InterlockedCompareExchange (&data->inited, -1, -1); - if (inited == 1) + if (data->inited >= 2) // 2 -> initialized, 3-> cleaned up return; - EnterCriticalSection (&data->io_lock); - inited = InterlockedCompareExchange (&data->inited, -1, -1); - if (inited == 1) { - LeaveCriticalSection (&data->io_lock); - return; + inited = InterlockedCompareExchange (&data->inited, 1, 0); + if (inited >= 1) { + while (TRUE) { + if (data->inited >= 2) + return; + SleepEx (1, FALSE); + } } + EnterCriticalSection (&data->io_lock); + #ifdef HAVE_EPOLL data->epoll_disabled = (g_getenv ("MONO_DISABLE_AIO") != NULL); if (FALSE == data->epoll_disabled) { @@ -746,7 +745,7 @@ socket_io_init (SocketIOData *data) data->epoll_disabled = TRUE; #endif -#ifndef PLATFORM_WIN32 +#ifndef HOST_WIN32 if (data->epoll_disabled) { if (pipe (data->pipe) != 0) { int err = errno; @@ -781,12 +780,14 @@ socket_io_init (SocketIOData *data) closesocket (srv); #endif data->sock_to_state = mono_g_hash_table_new_type (g_direct_hash, g_direct_equal, MONO_HASH_VALUE_GC); - mono_thread_create_internal (mono_get_root_domain (), threadpool_start_idle_threads, &async_io_tp, TRUE); if (data->epoll_disabled) { data->new_sem = CreateSemaphore (NULL, 1, 1, NULL); g_assert (data->new_sem != NULL); } + + stack_size = mono_threads_get_default_stacksize (); + mono_threads_set_default_stacksize (128 * 1024); if (data->epoll_disabled) { mono_thread_create_internal (mono_get_root_domain (), socket_io_poll_main, data, TRUE); } @@ -795,8 +796,9 @@ socket_io_init (SocketIOData *data) mono_thread_create_internal (mono_get_root_domain (), socket_io_epoll_main, data, TRUE); } #endif - InterlockedCompareExchange (&data->inited, 1, 0); + mono_threads_set_default_stacksize (stack_size); LeaveCriticalSection (&data->io_lock); + data->inited = 2; } static void @@ -808,11 +810,15 @@ socket_io_add_poll (MonoSocketAsyncResult *state) SocketIOData *data = &socket_io_data; int w; -#if defined(PLATFORM_MACOSX) || defined(PLATFORM_BSD) || defined(PLATFORM_WIN32) || defined(PLATFORM_SOLARIS) + if (mono_runtime_is_shutting_down () || data->inited == 3 || data->sock_to_state == NULL) + return; + +#if defined(PLATFORM_MACOSX) || defined(PLATFORM_BSD) || defined(HOST_WIN32) || defined(PLATFORM_SOLARIS) /* select() for connect() does not work well on the Mac. Bug #75436. */ /* Bug #77637 for the BSD 6 case */ /* Bug #78888 for the Windows case */ if (state->operation == AIO_OP_CONNECT && state->blocking == TRUE) { + //FIXME: increment number of threads while this one is waiting? threadpool_append_job (&async_io_tp, (MonoObject *) state); return; } @@ -822,6 +828,11 @@ socket_io_add_poll (MonoSocketAsyncResult *state) data->newpfd = g_new0 (mono_pollfd, 1); EnterCriticalSection (&data->io_lock); + if (data->sock_to_state == NULL) { + LeaveCriticalSection (&data->io_lock); + return; + } + /* FIXME: 64 bit issue: handle can be a pointer on windows? */ list = mono_g_hash_table_lookup (data->sock_to_state, GINT_TO_POINTER (state->handle)); if (list == NULL) { @@ -835,7 +846,7 @@ socket_io_add_poll (MonoSocketAsyncResult *state) mono_g_hash_table_replace (data->sock_to_state, GINT_TO_POINTER (state->handle), list); LeaveCriticalSection (&data->io_lock); *msg = (char) state->operation; -#ifndef PLATFORM_WIN32 +#ifndef HOST_WIN32 w = write (data->pipe [1], msg, 1); w = w; #else @@ -853,9 +864,16 @@ socket_io_add_epoll (MonoSocketAsyncResult *state) int epoll_op, ievt; int fd; + if (mono_runtime_is_shutting_down () || data->inited == 3 || data->sock_to_state == NULL) + return TRUE; + memset (&event, 0, sizeof (struct epoll_event)); fd = GPOINTER_TO_INT (state->handle); EnterCriticalSection (&data->io_lock); + if (data->sock_to_state == NULL) { + LeaveCriticalSection (&data->io_lock); + return TRUE; + } list = mono_g_hash_table_lookup (data->sock_to_state, GINT_TO_POINTER (fd)); if (list == NULL) { list = mono_mlist_alloc ((MonoObject*)state); @@ -873,20 +891,18 @@ socket_io_add_epoll (MonoSocketAsyncResult *state) mono_g_hash_table_replace (data->sock_to_state, state->handle, list); event.data.fd = fd; -#ifdef EPOLL_DEBUG - g_print ("%s %d with %d\n", epoll_op == EPOLL_CTL_ADD ? "ADD" : "MOD", fd, event.events); -#endif + EPOLL_DEBUG ("%s %d with %d", epoll_op == EPOLL_CTL_ADD ? "ADD" : "MOD", fd, event.events); if (epoll_ctl (data->epollfd, epoll_op, fd, &event) == -1) { int err = errno; if (epoll_op == EPOLL_CTL_ADD && err == EEXIST) { epoll_op = EPOLL_CTL_MOD; if (epoll_ctl (data->epollfd, epoll_op, fd, &event) == -1) { - g_message ("epoll_ctl(MOD): %d %s\n", err, g_strerror (err)); + g_message ("epoll_ctl(MOD): %d %s", err, g_strerror (err)); } } } - LeaveCriticalSection (&data->io_lock); + return TRUE; } #endif @@ -894,7 +910,11 @@ socket_io_add_epoll (MonoSocketAsyncResult *state) static void socket_io_add (MonoAsyncResult *ares, MonoSocketAsyncResult *state) { + if (async_tp.pool_status == 2 || mono_runtime_is_shutting_down ()) + return; + socket_io_init (&socket_io_data); + MONO_OBJECT_SETREF (state, ares, ares); #ifdef HAVE_EPOLL if (socket_io_data.epoll_disabled == FALSE) { @@ -905,6 +925,7 @@ socket_io_add (MonoAsyncResult *ares, MonoSocketAsyncResult *state) socket_io_add_poll (state); } +#ifndef DISABLE_SOCKETS static gboolean socket_io_filter (MonoObject *target, MonoObject *state) { @@ -950,8 +971,9 @@ socket_io_filter (MonoObject *target, MonoObject *state) } #endif /* !DISABLE_SOCKETS */ -static void -mono_async_invoke (MonoAsyncResult *ares) +/* Returns the exception thrown when invoking, if any */ +static MonoObject * +mono_async_invoke (ThreadPool *tp, MonoAsyncResult *ares) { ASyncCall *ac = (ASyncCall *)ares->object_data; MonoObject *res, *exc = NULL; @@ -966,30 +988,36 @@ mono_async_invoke (MonoAsyncResult *ares) ares->original_context = NULL; } - ac->msg->exc = NULL; - res = mono_message_invoke (ares->async_delegate, ac->msg, &exc, &out_args); - MONO_OBJECT_SETREF (ac, res, res); - MONO_OBJECT_SETREF (ac, msg->exc, exc); - MONO_OBJECT_SETREF (ac, out_args, out_args); - - mono_monitor_enter ((MonoObject *) ares); - ares->completed = 1; - if (ares->handle != NULL) - wait_event = mono_wait_handle_get_handle ((MonoWaitHandle *) ares->handle); - mono_monitor_exit ((MonoObject *) ares); - /* notify listeners */ - if (wait_event != NULL) - SetEvent (wait_event); - - /* call async callback if cb_method != null*/ - if (ac->cb_method) { - MonoObject *exc = NULL; - void *pa = &ares; - mono_runtime_invoke (ac->cb_method, ac->cb_target, pa, &exc); - /* 'exc' will be the previous ac->msg->exc if not NULL and not - * catched. If catched, this will be set to NULL and the - * exception will not be printed. */ - MONO_OBJECT_SETREF (ac->msg, exc, exc); + if (ac == NULL) { + /* Fast path from ThreadPool.*QueueUserWorkItem */ + void *pa = ares->async_state; + mono_runtime_delegate_invoke (ares->async_delegate, &pa, &exc); + } else { + ac->msg->exc = NULL; + res = mono_message_invoke (ares->async_delegate, ac->msg, &exc, &out_args); + MONO_OBJECT_SETREF (ac, res, res); + MONO_OBJECT_SETREF (ac, msg->exc, exc); + MONO_OBJECT_SETREF (ac, out_args, out_args); + + mono_monitor_enter ((MonoObject *) ares); + ares->completed = 1; + if (ares->handle != NULL) + wait_event = mono_wait_handle_get_handle ((MonoWaitHandle *) ares->handle); + mono_monitor_exit ((MonoObject *) ares); + /* notify listeners */ + if (wait_event != NULL) + SetEvent (wait_event); + + /* call async callback if cb_method != null*/ + if (ac != NULL && ac->cb_method) { + MonoObject *exc = NULL; + void *pa = &ares; + mono_runtime_invoke (ac->cb_method, ac->cb_target, pa, &exc); + /* 'exc' will be the previous ac->msg->exc if not NULL and not + * catched. If catched, this will be set to NULL and the + * exception will not be printed. */ + MONO_OBJECT_SETREF (ac->msg, exc, exc); + } } /* restore original thread execution context if flow isn't suppressed, i.e. non null */ @@ -997,23 +1025,25 @@ mono_async_invoke (MonoAsyncResult *ares) mono_thread_set_execution_context (ares->original_context); ares->original_context = NULL; } - + return exc; } static void threadpool_start_idle_threads (ThreadPool *tp) { - int needed; - int existing; + int n; - needed = (int) InterlockedCompareExchange (&tp->min_threads, 0, -1); do { - existing = (int) InterlockedCompareExchange (&tp->nthreads, 0, -1); - if (existing >= needed) - break; - InterlockedIncrement (&tp->nthreads); - mono_thread_create_internal (mono_get_root_domain (), tp->async_invoke, NULL, TRUE); - SleepEx (250, TRUE); + while (1) { + n = tp->nthreads; + if (n >= tp->min_threads) + return; + if (InterlockedCompareExchange (&tp->nthreads, n + 1, n) == n) + break; + } + mono_perfcounter_update_value (tp->pc_nthreads, TRUE, 1); + mono_thread_create_internal (mono_get_root_domain (), tp->async_invoke, tp, TRUE); + SleepEx (100, TRUE); } while (1); } @@ -1021,11 +1051,11 @@ static void threadpool_init (ThreadPool *tp, int min_threads, int max_threads, void (*async_invoke) (gpointer)) { memset (tp, 0, sizeof (ThreadPool)); - InitializeCriticalSection (&tp->lock); + MONO_SEM_INIT (&tp->lock, 1); tp->min_threads = min_threads; tp->max_threads = max_threads; tp->async_invoke = async_invoke; - tp->idle_threads = g_queue_new (); + MONO_SEM_INIT (&tp->new_job, 0); } static void * @@ -1047,41 +1077,116 @@ init_perf_counter (const char *category, const char *counter) return mono_perfcounter_get_impl (category_str, counter_str, NULL, machine, &type, &custom); } +#ifdef DEBUG +static void +print_pool_info (ThreadPool *tp) +{ + +// if (tp->tail - tp->head == 0) +// return; + + g_print ("Pool status? %d\n", InterlockedCompareExchange (&tp->pool_status, 0, 0)); + g_print ("Min. threads: %d\n", InterlockedCompareExchange (&tp->min_threads, 0, 0)); + g_print ("Max. threads: %d\n", InterlockedCompareExchange (&tp->max_threads, 0, 0)); + g_print ("nthreads: %d\n", InterlockedCompareExchange (&tp->nthreads, 0, 0)); + g_print ("busy threads: %d\n", InterlockedCompareExchange (&tp->busy_threads, 0, 0)); + g_print ("Waiting: %d\n", InterlockedCompareExchange (&tp->waiting, 0, 0)); + g_print ("Queued: %d\n", (tp->tail - tp->head)); + if (tp == &async_tp) { + int i; + EnterCriticalSection (&wsqs_lock); + for (i = 0; i < wsqs->len; i++) { + g_print ("\tWSQ %d: %d\n", i, mono_wsq_count (g_ptr_array_index (wsqs, i))); + } + LeaveCriticalSection (&wsqs_lock); + } else { + g_print ("\tSockets: %d\n", mono_g_hash_table_size (socket_io_data.sock_to_state)); + } + g_print ("-------------\n"); +} + +static void +signal_handler (int signo) +{ + ThreadPool *tp; + + tp = &async_tp; + MONO_SEM_WAIT (&tp->lock); + g_print ("\n-----Non-IO-----\n"); + print_pool_info (tp); + MONO_SEM_POST (&tp->lock); + tp = &async_io_tp; + MONO_SEM_WAIT (&tp->lock); + g_print ("\n-----IO-----\n"); + print_pool_info (tp); + MONO_SEM_POST (&tp->lock); + alarm (2); +} +#endif + void mono_thread_pool_init () { - int threads_per_cpu = THREADS_PER_CPU; - int cpu_count; - int n; + gint threads_per_cpu = 1; + gint thread_count; + gint cpu_count = mono_cpu_count (); + int result; - if ((int) InterlockedCompareExchange (&tp_inited, 1, 0) == 1) + if (tp_inited == 2) return; + result = InterlockedCompareExchange (&tp_inited, 1, 0); + if (result == 1) { + while (1) { + SleepEx (1, FALSE); + if (tp_inited == 2) + return; + } + } + + MONO_GC_REGISTER_ROOT (async_tp.first); + MONO_GC_REGISTER_ROOT (async_tp.last); + MONO_GC_REGISTER_ROOT (async_tp.unused); + MONO_GC_REGISTER_ROOT (async_io_tp.first); + MONO_GC_REGISTER_ROOT (async_io_tp.unused); + MONO_GC_REGISTER_ROOT (async_io_tp.last); + MONO_GC_REGISTER_ROOT (socket_io_data.sock_to_state); InitializeCriticalSection (&socket_io_data.io_lock); if (g_getenv ("MONO_THREADS_PER_CPU") != NULL) { threads_per_cpu = atoi (g_getenv ("MONO_THREADS_PER_CPU")); - if (threads_per_cpu < THREADS_PER_CPU) - threads_per_cpu = THREADS_PER_CPU; + if (threads_per_cpu < 1) + threads_per_cpu = 1; } - cpu_count = mono_cpu_count (); - n = 8 + 2 * cpu_count; /* 8 is minFreeThreads for ASP.NET */ - threadpool_init (&async_tp, n, n + threads_per_cpu * cpu_count, async_invoke_thread); -#ifndef DISABLE_SOCKETS - threadpool_init (&async_io_tp, 2 * cpu_count, 8 * cpu_count, async_invoke_io_thread); -#endif + thread_count = MIN (cpu_count * threads_per_cpu, 100 * cpu_count); + threadpool_init (&async_tp, thread_count, MAX (100 * cpu_count, thread_count), async_invoke_thread); + threadpool_init (&async_io_tp, cpu_count * 2, cpu_count * 4, async_invoke_thread); + async_io_tp.is_io = TRUE; async_call_klass = mono_class_from_name (mono_defaults.corlib, "System", "MonoAsyncCall"); g_assert (async_call_klass); + InitializeCriticalSection (&wsqs_lock); + wsqs = g_ptr_array_sized_new (MAX (100 * cpu_count, thread_count)); + mono_wsq_init (); + async_tp.pc_nitems = init_perf_counter ("Mono Threadpool", "Work Items Added"); g_assert (async_tp.pc_nitems); - mono_perfcounter_get_sample (async_tp.pc_nitems, FALSE, &async_tp.last_sample); async_io_tp.pc_nitems = init_perf_counter ("Mono Threadpool", "IO Work Items Added"); g_assert (async_io_tp.pc_nitems); - mono_perfcounter_get_sample (async_io_tp.pc_nitems, FALSE, &async_io_tp.last_sample); + + async_tp.pc_nthreads = init_perf_counter ("Mono Threadpool", "# of Threads"); + g_assert (async_tp.pc_nthreads); + + async_io_tp.pc_nthreads = init_perf_counter ("Mono Threadpool", "# of IO Threads"); + g_assert (async_io_tp.pc_nthreads); + tp_inited = 2; +#ifdef DEBUG + signal (SIGALRM, signal_handler); + alarm (2); +#endif } MonoAsyncResult * @@ -1092,7 +1197,7 @@ mono_thread_pool_add (MonoObject *target, MonoMethodMessage *msg, MonoDelegate * MonoAsyncResult *ares; ASyncCall *ac; - ac = (ASyncCall*)mono_object_new (mono_domain_get (), async_call_klass); + ac = (ASyncCall*)mono_object_new (domain, async_call_klass); MONO_OBJECT_SETREF (ac, msg, msg); MONO_OBJECT_SETREF (ac, state, state); @@ -1110,9 +1215,6 @@ mono_thread_pool_add (MonoObject *target, MonoMethodMessage *msg, MonoDelegate * return ares; } #endif - if (InterlockedCompareExchange (&async_tp.idle_started, 1, 0) == 0) - mono_thread_create_internal (mono_get_root_domain (), threadpool_start_idle_threads, &async_tp, TRUE); - threadpool_append_job (&async_tp, (MonoObject *) ares); return ares; } @@ -1130,8 +1232,7 @@ mono_thread_pool_finish (MonoAsyncResult *ares, MonoArray **out_args, MonoObject mono_monitor_enter ((MonoObject *) ares); if (ares->endinvoke_called) { - *exc = (MonoObject *)mono_exception_from_name (mono_defaults.corlib, "System", - "InvalidOperationException"); + *exc = (MonoObject *) mono_get_exception_invalid_operation (NULL); mono_monitor_exit ((MonoObject *) ares); return NULL; } @@ -1163,43 +1264,44 @@ mono_thread_pool_finish (MonoAsyncResult *ares, MonoArray **out_args, MonoObject static void threadpool_kill_idle_threads (ThreadPool *tp) { - IdleThreadData *it; - - if (!tp || !tp->idle_threads) - return; + gint n; - while ((it = g_queue_pop_head (tp->idle_threads)) != NULL) { - it->data = NULL; - it->die = TRUE; - SetEvent (it->wait_handle); + n = (gint) InterlockedCompareExchange (&tp->max_threads, 0, -1); + while (n) { + n--; + MONO_SEM_POST (&tp->new_job); } - g_queue_free (tp->idle_threads); - tp->idle_threads = NULL; } void mono_thread_pool_cleanup (void) { - EnterCriticalSection (&async_tp.lock); + if (async_tp.pool_status == 0 || async_tp.pool_status == 2) + return; + + if (async_tp.pool_status == 1 && InterlockedCompareExchange (&async_tp.pool_status, 2, 1) == 2) + return; + + InterlockedExchange (&async_io_tp.pool_status, 2); + MONO_SEM_WAIT (&async_tp.lock); threadpool_free_queue (&async_tp); threadpool_kill_idle_threads (&async_tp); - LeaveCriticalSection (&async_tp.lock); - socket_io_cleanup (&socket_io_data); /* Empty when DISABLE_SOCKETS is defined */ - /* Do we want/need these? - DeleteCriticalSection (&async_tp.lock); - DeleteCriticalSection (&async_tp.table_lock); - DeleteCriticalSection (&socket_io_data.io_lock); - */ -} + MONO_SEM_POST (&async_tp.lock); -static void -null_array (MonoArray *a, int first, int last) -{ - /* We must null the old array because it might - contain cross-appdomain references, which - will crash the GC when the domains are - unloaded. */ - memset (mono_array_addr (a, MonoObject*, first), 0, sizeof (MonoObject*) * (last - first)); + socket_io_cleanup (&socket_io_data); /* Empty when DISABLE_SOCKETS is defined */ + MONO_SEM_WAIT (&async_io_tp.lock); + threadpool_free_queue (&async_io_tp); + threadpool_kill_idle_threads (&async_io_tp); + MONO_SEM_POST (&async_io_tp.lock); + MONO_SEM_DESTROY (&async_io_tp.new_job); + + EnterCriticalSection (&wsqs_lock); + mono_wsq_cleanup (); + if (wsqs) + g_ptr_array_free (wsqs, TRUE); + wsqs = NULL; + LeaveCriticalSection (&wsqs_lock); + MONO_SEM_DESTROY (&async_tp.new_job); } /* Caller must enter &tp->lock */ @@ -1207,208 +1309,198 @@ static MonoObject* dequeue_job_nolock (ThreadPool *tp) { MonoObject *ar; - int count; + MonoArray *array; + MonoMList *list; - if (!tp->array || tp->first_elem == tp->next_elem) - return NULL; - ar = mono_array_get (tp->array, MonoObject*, tp->first_elem); - mono_array_set (tp->array, MonoObject*, tp->first_elem, NULL); - tp->first_elem++; - count = tp->next_elem - tp->first_elem; - /* reduce the size of the array if it's mostly empty */ - if (mono_array_length (tp->array) > INITIAL_QUEUE_LENGTH && count < (mono_array_length (tp->array) / 3)) { - MonoArray *newa = mono_array_new_cached (mono_get_root_domain (), mono_defaults.object_class, mono_array_length (tp->array) / 2); - mono_array_memcpy_refs (newa, 0, tp->array, tp->first_elem, count); - null_array (tp->array, tp->first_elem, tp->next_elem); - tp->array = newa; - tp->first_elem = 0; - tp->next_elem = count; - } + list = tp->first; + do { + if (mono_runtime_is_shutting_down ()) + return NULL; + if (!list || tp->head == tp->tail) + return NULL; + + array = (MonoArray *) mono_mlist_get_data (list); + ar = mono_array_get (array, MonoObject *, tp->head % QUEUE_LENGTH); + mono_array_set (array, MonoObject *, tp->head % QUEUE_LENGTH, NULL); + tp->head++; + if ((tp->head % QUEUE_LENGTH) == 0) { + list = tp->first; + tp->first = mono_mlist_next (list); + if (tp->first == NULL) + tp->last = NULL; + if (mono_mlist_length (tp->unused) < 20) { + /* reuse this chunk */ + tp->unused = mono_mlist_set_next (list, tp->unused); + } + tp->head -= QUEUE_LENGTH; + tp->tail -= QUEUE_LENGTH; + } + list = tp->first; + } while (ar == NULL); return ar; } -/* Call after entering &tp->lock */ -static int -signal_idle_threads (ThreadPool *tp) +static gboolean +threadpool_start_thread (ThreadPool *tp) { - IdleThreadData *it; - int result = 0; - int njobs; - - njobs = tp->next_elem - tp->first_elem; - while (njobs > 0 && (it = g_queue_pop_head (tp->idle_threads)) != NULL) { - it->data = dequeue_job_nolock (tp); - if (it->data == NULL) - break; /* Should never happen */ - result++; - njobs--; - it->timeout = INFINITE; - SetEvent (it->wait_handle); + gint n; + + while (!mono_runtime_is_shutting_down () && (n = tp->nthreads) < tp->max_threads) { + if (InterlockedCompareExchange (&tp->nthreads, n + 1, n) == n) { + mono_perfcounter_update_value (tp->pc_nthreads, TRUE, 1); + mono_thread_create_internal (mono_get_root_domain (), tp->async_invoke, tp, TRUE); + return TRUE; + } } - return njobs; + + return FALSE; } -/* Call after entering &tp->lock */ -static gboolean -threadpool_start_thread (ThreadPool *tp, gpointer arg) +static void +pulse_on_new_job (ThreadPool *tp) { - gint max; - gint n; + if (tp->waiting) + MONO_SEM_POST (&tp->new_job); +} - max = (gint) InterlockedCompareExchange (&tp->max_threads, 0, -1); - n = (gint) InterlockedCompareExchange (&tp->nthreads, 0, -1); - if (max <= n) - return FALSE; - InterlockedIncrement (&tp->nthreads); - mono_thread_create_internal (mono_get_root_domain (), tp->async_invoke, arg, TRUE); - return TRUE; +void +icall_append_job (MonoObject *ar) +{ + threadpool_append_job (&async_tp, ar); } -/* -static const char * -get_queue_name (ThreadPool *tp) +static void +threadpool_append_job (ThreadPool *tp, MonoObject *ar) { - if (tp == &async_tp) - return "TP"; - if (tp == &async_io_tp) - return "IO"; - return "(Unknown)"; + threadpool_append_jobs (tp, &ar, 1); } -*/ -static gpointer -threadpool_queue_idle_thread (ThreadPool *tp, IdleThreadData *it) +static MonoMList * +create_or_reuse_list (ThreadPool *tp) { - /* - MonoCounterSample sample; - float rate; - */ - gpointer result = NULL; - CRITICAL_SECTION *cs = &tp->lock; - - EnterCriticalSection (cs); - if (tp->idle_threads == NULL) { - it->die = TRUE; - LeaveCriticalSection (cs); - return NULL; /* We are shutting down */ - } - /* - if (mono_100ns_ticks () - tp->last_sample.timeStamp > 10000 * 1000) { - float elapsed_ticks; - mono_perfcounter_get_sample (tp->pc_nitems, FALSE, &sample); - - elapsed_ticks = (float) (sample.timeStamp - tp->last_sample.timeStamp); - rate = ((float) (sample.rawValue - tp->last_sample.rawValue)) / elapsed_ticks * 10000000; - printf ("Queue: %s NThreads: %d Rate: %.2f Total items: %lld Time(ms): %.2f\n", get_queue_name (tp), - InterlockedCompareExchange (&tp->nthreads, 0, -1), rate, - sample.rawValue - tp->last_sample.rawValue, elapsed_ticks / 10000); - memcpy (&tp->last_sample, &sample, sizeof (sample)); - } - */ + MonoMList *list; + MonoArray *array; - it->data = result = dequeue_job_nolock (tp); - if (result != NULL) { - signal_idle_threads (tp); + list = NULL; + if (tp->unused) { + list = tp->unused; + tp->unused = mono_mlist_next (list); + mono_mlist_set_next (list, NULL); + //TP_DEBUG (tp->nodes_reused++); } else { - int min, n; - min = (gint) InterlockedCompareExchange (&tp->min_threads, 0, -1); - n = (gint) InterlockedCompareExchange (&tp->nthreads, 0, -1); - if (n <= min) { - g_queue_push_tail (tp->idle_threads, it); - } else { - /* TODO: figure out when threads should be told to die */ - /* it->die = TRUE; */ - g_queue_push_tail (tp->idle_threads, it); - } + array = mono_array_new_cached (mono_get_root_domain (), mono_defaults.object_class, QUEUE_LENGTH); + list = mono_mlist_alloc ((MonoObject *) array); + //TP_DEBUG (tp->nodes_created++); } - LeaveCriticalSection (cs); - return result; + return list; } static void -threadpool_append_job (ThreadPool *tp, MonoObject *ar) +threadpool_append_jobs (ThreadPool *tp, MonoObject **jobs, gint njobs) { - CRITICAL_SECTION *cs; - - cs = &tp->lock; - threadpool_jobs_inc (ar); - EnterCriticalSection (cs); - if (tp->idle_threads == NULL) { - LeaveCriticalSection (cs); - return; /* We are shutting down */ - } - if (ar->vtable->domain->state == MONO_APPDOMAIN_UNLOADING || - ar->vtable->domain->state == MONO_APPDOMAIN_UNLOADED) { - LeaveCriticalSection (cs); + static int job_counter; + MonoArray *array; + MonoMList *list; + MonoObject *ar; + gint i; + gboolean lock_taken = FALSE; /* We won't take the lock when the local queue is used */ + + if (mono_runtime_is_shutting_down ()) return; - } - mono_perfcounter_update_value (tp->pc_nitems, TRUE, 1); - if (tp->array && (tp->next_elem < mono_array_length (tp->array))) { - mono_array_setref (tp->array, tp->next_elem, ar); - tp->next_elem++; - if (signal_idle_threads (tp) > 0 && threadpool_start_thread (tp, ar)) { - tp->next_elem--; - mono_array_setref (tp->array, tp->next_elem, NULL); + if (tp->pool_status == 0 && InterlockedCompareExchange (&tp->pool_status, 1, 0) == 0) + mono_thread_create_internal (mono_get_root_domain (), threadpool_start_idle_threads, tp, TRUE); + + for (i = 0; i < njobs; i++) { + ar = jobs [i]; + if (ar == NULL || mono_domain_is_unloading (ar->vtable->domain)) + continue; /* Might happen when cleaning domain jobs */ + if (!tp->is_io && (InterlockedIncrement (&job_counter) % 10) == 0) { + MonoAsyncResult *o = (MonoAsyncResult *) ar; + o->add_time = mono_100ns_ticks (); } - LeaveCriticalSection (cs); - return; - } + threadpool_jobs_inc (ar); + mono_perfcounter_update_value (tp->pc_nitems, TRUE, 1); + if (!tp->is_io && mono_wsq_local_push (ar)) + continue; - if (!tp->array) { - MONO_GC_REGISTER_ROOT (tp->array); - tp->array = mono_array_new_cached (mono_get_root_domain (), mono_defaults.object_class, INITIAL_QUEUE_LENGTH); - } else { - int count = tp->next_elem - tp->first_elem; - /* slide the array or create a larger one if it's full */ - if (tp->first_elem) { - mono_array_memcpy_refs (tp->array, 0, tp->array, tp->first_elem, count); - null_array (tp->array, count, tp->next_elem); - } else { - MonoArray *newa = mono_array_new_cached (mono_get_root_domain (), mono_defaults.object_class, mono_array_length (tp->array) * 2); - mono_array_memcpy_refs (newa, 0, tp->array, tp->first_elem, count); - null_array (tp->array, count, tp->next_elem); - tp->array = newa; + if (!lock_taken) { + MONO_SEM_WAIT (&tp->lock); + lock_taken = TRUE; + } + if ((tp->tail % QUEUE_LENGTH) == 0) { + list = create_or_reuse_list (tp); + if (tp->last != NULL) + mono_mlist_set_next (tp->last, list); + tp->last = list; + if (tp->first == NULL) + tp->first = tp->last; } - tp->first_elem = 0; - tp->next_elem = count; + + array = (MonoArray *) mono_mlist_get_data (tp->last); + mono_array_setref (array, tp->tail % QUEUE_LENGTH, ar); + tp->tail++; } - mono_array_setref (tp->array, tp->next_elem, ar); - tp->next_elem++; - if (signal_idle_threads (tp) > 0 && threadpool_start_thread (tp, ar)) { - tp->next_elem--; - mono_array_setref (tp->array, tp->next_elem, NULL); + if (lock_taken) + MONO_SEM_POST (&tp->lock); + + if (!tp->is_io && tp->waiting == 0) { + gint64 ticks = mono_100ns_ticks (); + + if (tp->last_check == 0 || (ticks - tp->last_check) > 5000000) { + SPIN_LOCK (tp->sp_lock); + tp->last_check = ticks; + SPIN_UNLOCK (tp->sp_lock); + threadpool_start_thread (tp); + } } - LeaveCriticalSection (cs); + for (i = 0; i < MIN(njobs, tp->max_threads); i++) + pulse_on_new_job (tp); } - static void threadpool_clear_queue (ThreadPool *tp, MonoDomain *domain) { - int i, count = 0; - EnterCriticalSection (&tp->lock); - /*remove*/ - for (i = tp->first_elem; i < tp->next_elem; ++i) { - MonoObject *obj = mono_array_get (tp->array, MonoObject*, i); - if (obj->vtable->domain == domain) { - mono_array_set (tp->array, MonoObject*, i, NULL); - InterlockedDecrement (&domain->threadpool_jobs); - ++count; + MonoMList *current; + MonoArray *array; + MonoObject *obj; + int domain_count; + int i; + + domain_count = 0; + MONO_SEM_WAIT (&tp->lock); + current = tp->first; + while (current) { + array = (MonoArray *) mono_mlist_get_data (current); + for (i = 0; i < QUEUE_LENGTH; i++) { + obj = mono_array_get (array, MonoObject*, i); + if (obj != NULL && obj->vtable->domain == domain) { + domain_count++; + mono_array_setref (array, i, NULL); + threadpool_jobs_dec (obj); + } } + current = mono_mlist_next (current); } - /*compact*/ - if (count) { - int idx = 0; - for (i = tp->first_elem; i < tp->next_elem; ++i) { - MonoObject *obj = mono_array_get (tp->array, MonoObject*, i); - if (obj) - mono_array_set (tp->array, MonoObject*, idx++, obj); - } - tp->first_elem = 0; - tp->next_elem = count; + + if (!domain_count) { + MONO_SEM_POST (&tp->lock); + return; + } + + current = tp->first; + tp->first = NULL; + tp->last = NULL; + tp->head = 0; + tp->tail = 0; + MONO_SEM_POST (&tp->lock); + /* Re-add everything but the nullified elements */ + while (current) { + array = (MonoArray *) mono_mlist_get_data (current); + threadpool_append_jobs (tp, mono_array_addr (array, MonoObject *, 0), QUEUE_LENGTH); + memset (mono_array_addr (array, MonoObject *, 0), 0, sizeof (MonoObject *) * QUEUE_LENGTH); + current = mono_mlist_next (current); } - LeaveCriticalSection (&tp->lock); } /* @@ -1433,7 +1525,7 @@ mono_thread_pool_remove_domain_jobs (MonoDomain *domain, int timeout) * We avoid that by setting up a semaphore to be pulsed by the thread that reaches zero. */ sem_handle = CreateSemaphore (NULL, 0, 1, NULL); - + domain->cleanup_semaphore = sem_handle; /* * The memory barrier here is required to have global ordering between assigning to cleanup_semaphone @@ -1460,16 +1552,187 @@ mono_thread_pool_remove_domain_jobs (MonoDomain *domain, int timeout) static void threadpool_free_queue (ThreadPool *tp) { - if (tp->array) - null_array (tp->array, tp->first_elem, tp->next_elem); - tp->array = NULL; - tp->first_elem = tp->next_elem = 0; + tp->head = tp->tail = 0; + tp->first = NULL; + tp->unused = NULL; } gboolean mono_thread_pool_is_queue_array (MonoArray *o) { - return o == async_tp.array || o == async_io_tp.array; + gpointer obj = o; + + // FIXME: need some fix in sgen code. + // There are roots at: async*tp.unused (MonoMList) and wsqs [n]->queue (MonoArray) + return obj == async_tp.first || obj == async_io_tp.first; +} + +static void +add_wsq (MonoWSQ *wsq) +{ + int i; + + if (wsq == NULL) + return; + + EnterCriticalSection (&wsqs_lock); + if (wsqs == NULL) { + LeaveCriticalSection (&wsqs_lock); + return; + } + for (i = 0; i < wsqs->len; i++) { + if (g_ptr_array_index (wsqs, i) == NULL) { + wsqs->pdata [i] = wsq; + LeaveCriticalSection (&wsqs_lock); + return; + } + } + g_ptr_array_add (wsqs, wsq); + LeaveCriticalSection (&wsqs_lock); +} + +static void +remove_wsq (MonoWSQ *wsq) +{ + if (wsq == NULL) + return; + + EnterCriticalSection (&wsqs_lock); + if (wsqs == NULL) { + LeaveCriticalSection (&wsqs_lock); + return; + } + g_ptr_array_remove_fast (wsqs, wsq); + LeaveCriticalSection (&wsqs_lock); +} + +static void +try_steal (gpointer *data, gboolean retry) +{ + int i; + int ms; + + if (wsqs == NULL || data == NULL || *data != NULL) + return; + + ms = 0; + do { + if (mono_runtime_is_shutting_down ()) + return; + for (i = 0; wsqs != NULL && i < wsqs->len; i++) { + if (mono_runtime_is_shutting_down ()) { + return; + } + mono_wsq_try_steal (wsqs->pdata [i], data, ms); + if (*data != NULL) { + return; + } + } + ms += 10; + } while (retry && ms < 11); +} + +static gboolean +dequeue_or_steal (ThreadPool *tp, gpointer *data) +{ + if (mono_runtime_is_shutting_down ()) + return FALSE; + TP_DEBUG ("Dequeue"); + MONO_SEM_WAIT (&tp->lock); + *data = dequeue_job_nolock (tp); + MONO_SEM_POST (&tp->lock); + if (!tp->is_io && !*data) + try_steal (data, FALSE); + return (*data != NULL); +} + +static void +process_idle_times (ThreadPool *tp, gint64 t) +{ + gint64 ticks; + gint64 avg; + gboolean compute_avg; + gint new_threads; + gint64 per1; + + if (tp->ignore_times || t <= 0) + return; + + compute_avg = FALSE; + ticks = mono_100ns_ticks (); + t = ticks - t; + SPIN_LOCK (tp->sp_lock); + if (tp->ignore_times) { + SPIN_UNLOCK (tp->sp_lock); + return; + } + tp->time_sum += t; + tp->n_sum++; + if (tp->last_check == 0) + tp->last_check = ticks; + else if (tp->last_check > 0 && (ticks - tp->last_check) > 5000000) { + tp->ignore_times = 1; + compute_avg = TRUE; + } + SPIN_UNLOCK (tp->sp_lock); + + if (!compute_avg) + return; + + //printf ("Items: %d Time elapsed: %.3fs\n", tp->n_sum, (ticks - tp->last_check) / 10000.0); + tp->last_check = ticks; + new_threads = 0; + avg = tp->time_sum / tp->n_sum; + if (tp->averages [1] == 0) { + tp->averages [1] = avg; + } else { + per1 = ((100 * (ABS (avg - tp->averages [1]))) / tp->averages [1]); + if (per1 > 5) { + if (avg > tp->averages [1]) { + if (tp->averages [1] < tp->averages [0]) { + new_threads = -1; + } else { + new_threads = 1; + } + } else if (avg < tp->averages [1] && tp->averages [1] < tp->averages [0]) { + new_threads = 1; + } + } else { + int min, n; + min = tp->min_threads; + n = tp->nthreads; + if ((n - min) < min && tp->busy_threads == n) + new_threads = 1; + } + /* + if (new_threads != 0) { + printf ("n: %d per1: %lld avg=%lld avg1=%lld avg0=%lld\n", new_threads, per1, avg, tp->averages [1], tp->averages [0]); + } + */ + } + + tp->time_sum = 0; + tp->n_sum = 0; + + tp->averages [0] = tp->averages [1]; + tp->averages [1] = avg; + tp->ignore_times = 0; + + if (tp->waiting == 0 && new_threads == 1) { + threadpool_start_thread (tp); + } else if (new_threads == -1) { + if (tp->destroy_thread == 0 && InterlockedCompareExchange (&tp->destroy_thread, 1, 0) == 0) + pulse_on_new_job (tp); + } +} + +static gboolean +should_i_die (ThreadPool *tp) +{ + gboolean result = FALSE; + if (tp->destroy_thread == 1 && InterlockedCompareExchange (&tp->destroy_thread, 0, 1) == 1) + result = (tp->nthreads > tp->min_threads); + return result; } static void @@ -1477,76 +1740,158 @@ async_invoke_thread (gpointer data) { MonoDomain *domain; MonoInternalThread *thread; - const gchar *version; - IdleThreadData idle_data = {0}; + MonoWSQ *wsq; + ThreadPool *tp; + gboolean must_die; - idle_data.timeout = INFINITE; - idle_data.wait_handle = CreateEvent (NULL, FALSE, FALSE, NULL); - + tp = data; + wsq = NULL; + if (!tp->is_io) { + wsq = mono_wsq_create (); + add_wsq (wsq); + } + thread = mono_thread_internal_current (); - version = mono_get_runtime_info ()->framework_version; + if (tp_start_func) + tp_start_func (tp_hooks_user_data); + data = NULL; for (;;) { MonoAsyncResult *ar; + gboolean is_io_task; + int n_naps = 0; + is_io_task = FALSE; ar = (MonoAsyncResult *) data; if (ar) { + InterlockedIncrement (&tp->busy_threads); +#ifndef DISABLE_SOCKETS + is_io_task = (strcmp (((MonoObject *) data)->vtable->klass->name, "AsyncResult")); + if (is_io_task) { + MonoSocketAsyncResult *state = (MonoSocketAsyncResult *) data; + ar = state->ares; + switch (state->operation) { + case AIO_OP_RECEIVE: + state->total = ICALL_RECV (state); + break; + case AIO_OP_SEND: + state->total = ICALL_SEND (state); + break; + } + } +#endif /* worker threads invokes methods in different domains, * so we need to set the right domain here */ domain = ((MonoObject *)ar)->vtable->domain; - g_assert (domain); - if (domain->state == MONO_APPDOMAIN_UNLOADED || domain->state == MONO_APPDOMAIN_UNLOADING) { + if (mono_domain_is_unloading (domain) || mono_runtime_is_shutting_down ()) { threadpool_jobs_dec ((MonoObject *)ar); data = NULL; + ar = NULL; + InterlockedDecrement (&tp->busy_threads); } else { mono_thread_push_appdomain_ref (domain); if (threadpool_jobs_dec ((MonoObject *)ar)) { data = NULL; + ar = NULL; mono_thread_pop_appdomain_ref (); + InterlockedDecrement (&tp->busy_threads); continue; } if (mono_domain_set (domain, FALSE)) { - ASyncCall *ac; + /* ASyncCall *ac; */ - mono_async_invoke (ar); - ac = (ASyncCall *) ar->object_data; + if (tp_item_begin_func) + tp_item_begin_func (tp_item_user_data); + + if (!is_io_task && ar->add_time > 0) + process_idle_times (tp, ar->add_time); + /*FIXME: Do something with the exception returned? */ + mono_async_invoke (tp, ar); + if (tp_item_end_func) + tp_item_end_func (tp_item_user_data); /* + ac = (ASyncCall *) ar->object_data; if (ac->msg->exc != NULL) mono_unhandled_exception (ac->msg->exc); */ mono_domain_set (mono_get_root_domain (), TRUE); } mono_thread_pop_appdomain_ref (); - InterlockedDecrement (&async_tp.busy_threads); + InterlockedDecrement (&tp->busy_threads); /* If the callee changes the background status, set it back to TRUE */ - if (*version != '1' && !mono_thread_test_state (thread , ThreadState_Background)) + if (!mono_thread_test_state (thread , ThreadState_Background)) ves_icall_System_Threading_Thread_SetState (thread, ThreadState_Background); } } - data = threadpool_queue_idle_thread (&async_tp, &idle_data); - while (!idle_data.die && !data) { - guint32 wr; - wr = WaitForSingleObjectEx (idle_data.wait_handle, idle_data.timeout, TRUE); - if (THREAD_WANTS_A_BREAK (thread)) - mono_thread_interruption_checkpoint (); - - if (wr != WAIT_TIMEOUT && wr != WAIT_IO_COMPLETION) { - data = idle_data.data; - break; /* We have to exit */ + + ar = NULL; + data = NULL; + must_die = should_i_die (tp); + TP_DEBUG ("Trying to get a job"); + if (!must_die && (tp->is_io || !mono_wsq_local_pop (&data))) + dequeue_or_steal (tp, &data); + TP_DEBUG ("Done trying to get a job %p", data); + + n_naps = 0; + while (!must_die && !data && n_naps < 4) { + gboolean res; + + TP_DEBUG ("Waiting"); + InterlockedIncrement (&tp->waiting); +#if defined(__OpenBSD__) + while ((res = mono_sem_wait (&tp->new_job, TRUE)) == -1) {// && errno == EINTR) { +#else + while ((res = mono_sem_timedwait (&tp->new_job, 2000, TRUE)) == -1) {// && errno == EINTR) { +#endif + if (mono_runtime_is_shutting_down ()) + break; + if (THREAD_WANTS_A_BREAK (thread)) + mono_thread_interruption_checkpoint (); + } + TP_DEBUG ("Done waiting"); + InterlockedDecrement (&tp->waiting); + if (mono_runtime_is_shutting_down ()) + break; + must_die = should_i_die (tp); + dequeue_or_steal (tp, &data); + n_naps++; + } + + if (!data && tp->is_io && !mono_runtime_is_shutting_down ()) { + mono_wsq_local_pop (&data); + if (data && must_die) { + InterlockedCompareExchange (&tp->destroy_thread, 1, 0); + pulse_on_new_job (tp); } } - idle_data.data = NULL; if (!data) { - InterlockedDecrement (&async_tp.nthreads); - CloseHandle (idle_data.wait_handle); - idle_data.wait_handle = NULL; - return; + gint nt; + gboolean down; + while (1) { + nt = tp->nthreads; + down = mono_runtime_is_shutting_down (); + if (!down && nt <= tp->min_threads) + break; + if (down || InterlockedCompareExchange (&tp->nthreads, nt - 1, nt) == nt) { + mono_perfcounter_update_value (tp->pc_nthreads, TRUE, -1); + TP_DEBUG ("DIE"); + if (!tp->is_io) { + remove_wsq (wsq); + while (mono_wsq_local_pop (&data)) { + threadpool_jobs_dec (data); + data = NULL; + } + mono_wsq_destroy (wsq); + } + if (tp_finish_func) + tp_finish_func (tp_hooks_user_data); + return; + } + } } - - InterlockedIncrement (&async_tp.busy_threads); } g_assert_not_reached (); @@ -1555,72 +1900,62 @@ async_invoke_thread (gpointer data) void ves_icall_System_Threading_ThreadPool_GetAvailableThreads (gint *workerThreads, gint *completionPortThreads) { - gint busy, busy_io; - - MONO_ARCH_SAVE_REGS; - - busy = (gint) InterlockedCompareExchange (&async_tp.busy_threads, 0, -1); - busy_io = (gint) InterlockedCompareExchange (&async_io_tp.busy_threads, 0, -1); - *workerThreads = async_tp.max_threads - busy; - *completionPortThreads = async_io_tp.max_threads - busy_io; + *workerThreads = async_tp.max_threads - async_tp.busy_threads; + *completionPortThreads = async_io_tp.max_threads - async_io_tp.busy_threads; } void ves_icall_System_Threading_ThreadPool_GetMaxThreads (gint *workerThreads, gint *completionPortThreads) { - MONO_ARCH_SAVE_REGS; - - *workerThreads = (gint) InterlockedCompareExchange (&async_tp.max_threads, 0, -1); - *completionPortThreads = (gint) InterlockedCompareExchange (&async_io_tp.max_threads, 0, -1); + *workerThreads = async_tp.max_threads; + *completionPortThreads = async_io_tp.max_threads; } void ves_icall_System_Threading_ThreadPool_GetMinThreads (gint *workerThreads, gint *completionPortThreads) { - gint workers, workers_io; - - MONO_ARCH_SAVE_REGS; - - workers = (gint) InterlockedCompareExchange (&async_tp.min_threads, 0, -1); - workers_io = (gint) InterlockedCompareExchange (&async_io_tp.min_threads, 0, -1); - - *workerThreads = workers; - *completionPortThreads = workers_io; -} - -static void -start_idle_threads (void) -{ - threadpool_start_idle_threads (&async_tp); + *workerThreads = async_tp.min_threads; + *completionPortThreads = async_io_tp.min_threads; } MonoBoolean ves_icall_System_Threading_ThreadPool_SetMinThreads (gint workerThreads, gint completionPortThreads) { - MONO_ARCH_SAVE_REGS; + gint max_threads; + gint max_io_threads; - if (workerThreads < 0 || workerThreads > async_tp.max_threads) + max_threads = async_tp.max_threads; + if (workerThreads <= 0 || workerThreads > max_threads) return FALSE; - if (completionPortThreads < 0 || completionPortThreads > async_io_tp.max_threads) + max_io_threads = async_io_tp.max_threads; + if (completionPortThreads <= 0 || completionPortThreads > max_io_threads) return FALSE; InterlockedExchange (&async_tp.min_threads, workerThreads); InterlockedExchange (&async_io_tp.min_threads, completionPortThreads); - mono_thread_create_internal (mono_get_root_domain (), start_idle_threads, NULL, TRUE); + if (workerThreads > async_tp.nthreads) + mono_thread_create_internal (mono_get_root_domain (), threadpool_start_idle_threads, &async_tp, TRUE); + if (completionPortThreads > async_io_tp.nthreads) + mono_thread_create_internal (mono_get_root_domain (), threadpool_start_idle_threads, &async_io_tp, TRUE); return TRUE; } MonoBoolean ves_icall_System_Threading_ThreadPool_SetMaxThreads (gint workerThreads, gint completionPortThreads) { - MONO_ARCH_SAVE_REGS; + gint min_threads; + gint min_io_threads; + gint cpu_count; - if (workerThreads < async_tp.max_threads) + cpu_count = mono_cpu_count (); + min_threads = async_tp.min_threads; + if (workerThreads < min_threads || workerThreads < cpu_count) return FALSE; /* We don't really have the concept of completion ports. Do we care here? */ - if (completionPortThreads < async_io_tp.max_threads) + min_io_threads = async_io_tp.min_threads; + if (completionPortThreads < min_io_threads || completionPortThreads < cpu_count) return FALSE; InterlockedExchange (&async_tp.max_threads, workerThreads); @@ -1628,3 +1963,38 @@ ves_icall_System_Threading_ThreadPool_SetMaxThreads (gint workerThreads, gint co return TRUE; } +/** + * mono_install_threadpool_thread_hooks + * @start_func: the function to be called right after a new threadpool thread is created. Can be NULL. + * @finish_func: the function to be called right before a thredpool thread is exiting. Can be NULL. + * @user_data: argument passed to @start_func and @finish_func. + * + * @start_fun will be called right after a threadpool thread is created and @finish_func right before a threadpool thread exits. + * The calls will be made from the thread itself. + */ +void +mono_install_threadpool_thread_hooks (MonoThreadPoolFunc start_func, MonoThreadPoolFunc finish_func, gpointer user_data) +{ + tp_start_func = start_func; + tp_finish_func = finish_func; + tp_hooks_user_data = user_data; +} + +/** + * mono_install_threadpool_item_hooks + * @begin_func: the function to be called before a threadpool work item processing starts. + * @end_func: the function to be called after a threadpool work item is finished. + * @user_data: argument passed to @begin_func and @end_func. + * + * The calls will be made from the thread itself and from the same AppDomain + * where the work item was executed. + * + */ +void +mono_install_threadpool_item_hooks (MonoThreadPoolItemFunc begin_func, MonoThreadPoolItemFunc end_func, gpointer user_data) +{ + tp_item_begin_func = begin_func; + tp_item_end_func = end_func; + tp_item_user_data = user_data; +} +