[arm64] Add sgen support.
[mono.git] / mono / metadata / sgen-pinning.c
index cf4778bf0f06adcc8e0fe760714df3035df8e545..d0a38ef426eb94b8b90c82bcedd4490e3fc35ce2 100644 (file)
-#define PIN_STAGING_AREA_SIZE  1024
+/*
+ * sgen-pinning.c: The pin queue.
+ *
+ * Copyright 2001-2003 Ximian, Inc
+ * Copyright 2003-2010 Novell, Inc.
+ * Copyright (C) 2012 Xamarin Inc
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Library General Public
+ * License 2.0 as published by the Free Software Foundation;
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+ * Library General Public License for more details.
+ *
+ * You should have received a copy of the GNU Library General Public
+ * License 2.0 along with this library; if not, write to the Free
+ * Software Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
+ */
 
-static void* pin_staging_area [PIN_STAGING_AREA_SIZE];
-static int pin_staging_area_index;
+#include "config.h"
+#ifdef HAVE_SGEN_GC
 
-static void
-init_pinning (void)
+#include "metadata/sgen-gc.h"
+#include "metadata/sgen-pinning.h"
+#include "metadata/sgen-protocol.h"
+
+static void** pin_queue;
+static int pin_queue_size = 0;
+static int next_pin_slot = 0;
+static int last_num_pinned = 0;
+
+#define PIN_HASH_SIZE 1024
+static void *pin_hash_filter [PIN_HASH_SIZE];
+
+void
+sgen_init_pinning (void)
+{
+       memset (pin_hash_filter, 0, sizeof (pin_hash_filter));
+}
+
+void
+sgen_finish_pinning (void)
 {
-       pin_staging_area_index = 0;
+       last_num_pinned = next_pin_slot;
+       next_pin_slot = 0;
 }
 
 static void
-evacuate_pin_staging_area (void)
+realloc_pin_queue (void)
+{
+       int new_size = pin_queue_size? pin_queue_size + pin_queue_size/2: 1024;
+       void **new_pin = sgen_alloc_internal_dynamic (sizeof (void*) * new_size, INTERNAL_MEM_PIN_QUEUE, TRUE);
+       memcpy (new_pin, pin_queue, sizeof (void*) * next_pin_slot);
+       sgen_free_internal_dynamic (pin_queue, sizeof (void*) * pin_queue_size, INTERNAL_MEM_PIN_QUEUE);
+       pin_queue = new_pin;
+       pin_queue_size = new_size;
+       SGEN_LOG (4, "Reallocated pin queue to size: %d", new_size);
+}
+
+void
+sgen_pin_stage_ptr (void *ptr)
+{
+       /*very simple multiplicative hash function, tons better than simple and'ng */ 
+       int hash_idx = ((mword)ptr * 1737350767) & (PIN_HASH_SIZE - 1);
+       if (pin_hash_filter [hash_idx] == ptr)
+               return;
+
+       pin_hash_filter [hash_idx] = ptr;
+
+       if (next_pin_slot >= pin_queue_size)
+               realloc_pin_queue ();
+
+       pin_queue [next_pin_slot++] = ptr;
+}
+
+static int
+optimized_pin_queue_search (void *addr)
+{
+       int first = 0, last = next_pin_slot;
+       while (first < last) {
+               int middle = first + ((last - first) >> 1);
+               if (addr <= pin_queue [middle])
+                       last = middle;
+               else
+                       first = middle + 1;
+       }
+       g_assert (first == last);
+       return first;
+}
+
+void**
+sgen_find_optimized_pin_queue_area (void *start, void *end, int *num)
+{
+       int first, last;
+       first = optimized_pin_queue_search (start);
+       last = optimized_pin_queue_search (end);
+       *num = last - first;
+       if (first == last)
+               return NULL;
+       return pin_queue + first;
+}
+
+void
+sgen_find_section_pin_queue_start_end (GCMemSection *section)
+{
+       SGEN_LOG (6, "Pinning from section %p (%p-%p)", section, section->data, section->end_data);
+       section->pin_queue_start = sgen_find_optimized_pin_queue_area (section->data, section->end_data, &section->pin_queue_num_entries);
+       SGEN_LOG (6, "Found %d pinning addresses in section %p", section->pin_queue_num_entries, section);
+}
+
+/*This will setup the given section for the while pin queue. */
+void
+sgen_pinning_setup_section (GCMemSection *section)
+{
+       section->pin_queue_start = pin_queue;
+       section->pin_queue_num_entries = next_pin_slot;
+}
+
+void
+sgen_pinning_trim_queue_to_section (GCMemSection *section)
+{
+       next_pin_slot = section->pin_queue_num_entries;
+}
+
+void
+sgen_pin_queue_clear_discarded_entries (GCMemSection *section, int max_pin_slot)
+{
+       void **start = section->pin_queue_start + section->pin_queue_num_entries;
+       void **end = pin_queue + max_pin_slot;
+       void *addr;
+
+       if (!start)
+               return;
+
+       for (; start < end; ++start) {
+               addr = *start;
+               if ((char*)addr < section->data || (char*)addr > section->end_data)
+                       break;
+               *start = NULL;
+       }
+}
+
+/* reduce the info in the pin queue, removing duplicate pointers and sorting them */
+void
+sgen_optimize_pin_queue (int start_slot)
+{
+       void **start, **cur, **end;
+       /* sort and uniq pin_queue: we just sort and we let the rest discard multiple values */
+       /* it may be better to keep ranges of pinned memory instead of individually pinning objects */
+       SGEN_LOG (5, "Sorting pin queue, size: %d", next_pin_slot);
+       if ((next_pin_slot - start_slot) > 1)
+               sgen_sort_addresses (pin_queue + start_slot, next_pin_slot - start_slot);
+       start = cur = pin_queue + start_slot;
+       end = pin_queue + next_pin_slot;
+       while (cur < end) {
+               *start = *cur++;
+               while (*start == *cur && cur < end)
+                       cur++;
+               start++;
+       };
+       next_pin_slot = start - pin_queue;
+       SGEN_LOG (5, "Pin queue reduced to size: %d", next_pin_slot);
+}
+
+int
+sgen_get_pinned_count (void)
+{
+       return next_pin_slot;
+}
+
+void
+sgen_dump_pin_queue (void)
+{
+       int i;
+
+       for (i = 0; i < last_num_pinned; ++i) {
+               SGEN_LOG (3, "Bastard pinning obj %p (%s), size: %d", pin_queue [i], sgen_safe_name (pin_queue [i]), sgen_safe_object_get_size (pin_queue [i]));
+       }       
+}
+
+typedef struct _CementHashEntry CementHashEntry;
+struct _CementHashEntry {
+       char *obj;
+       unsigned int count;
+};
+
+static CementHashEntry cement_hash [SGEN_CEMENT_HASH_SIZE];
+static CementHashEntry cement_hash_concurrent [SGEN_CEMENT_HASH_SIZE];
+
+static gboolean cement_enabled = TRUE;
+static gboolean cement_concurrent = FALSE;
+
+void
+sgen_cement_init (gboolean enabled)
+{
+       cement_enabled = enabled;
+}
+
+void
+sgen_cement_reset (void)
+{
+       SGEN_ASSERT (1, !cement_concurrent, "Concurrent cementing cannot simply be reset");
+
+       memset (cement_hash, 0, sizeof (cement_hash));
+       binary_protocol_cement_reset ();
+}
+
+/*
+ * The reason we cannot simply reset cementing at the start of a
+ * concurrent collection is that the nursery collections running
+ * concurrently must keep pinning the cemented objects, because we
+ * don't have the global remsets that point to them anymore - if the
+ * nursery collector moved the cemented objects, we'd have invalid
+ * pointers in the major heap.
+ *
+ * What we do instead is to reset cementing at the start of concurrent
+ * collections in such a way that nursery collections happening during
+ * the major collection still pin the formerly cemented objects.  We
+ * have a shadow cementing table for that purpose.  The nursery
+ * collections still work with the old cementing table, while the
+ * major collector builds up a new cementing table, adding global
+ * remsets whenever needed like usual.  When the major collector
+ * finishes, the old cementing table is replaced by the new one.
+ */
+
+void
+sgen_cement_concurrent_start (void)
+{
+       SGEN_ASSERT (1, !cement_concurrent, "Concurrent cementing has already been started");
+       cement_concurrent = TRUE;
+
+       memset (cement_hash_concurrent, 0, sizeof (cement_hash));
+}
+
+void
+sgen_cement_concurrent_finish (void)
+{
+       SGEN_ASSERT (1, cement_concurrent, "Concurrent cementing hasn't been started");
+       cement_concurrent = FALSE;
+
+       memcpy (cement_hash, cement_hash_concurrent, sizeof (cement_hash));
+}
+
+gboolean
+sgen_cement_lookup (char *obj)
+{
+       int i = mono_aligned_addr_hash (obj) % SGEN_CEMENT_HASH_SIZE;
+
+       SGEN_ASSERT (5, sgen_ptr_in_nursery (obj), "Looking up cementing for non-nursery objects makes no sense");
+
+       if (!cement_enabled)
+               return FALSE;
+
+       if (!cement_hash [i].obj)
+               return FALSE;
+       if (cement_hash [i].obj != obj)
+               return FALSE;
+
+       return cement_hash [i].count >= SGEN_CEMENT_THRESHOLD;
+}
+
+gboolean
+sgen_cement_lookup_or_register (char *obj)
 {
        int i;
+       CementHashEntry *hash;
+       gboolean concurrent_cementing = sgen_concurrent_collection_in_progress ();
+
+       if (!cement_enabled)
+               return FALSE;
+
+       if (concurrent_cementing)
+               SGEN_ASSERT (5, cement_concurrent, "Cementing wasn't inited with concurrent flag");
 
-       g_assert (pin_staging_area_index >= 0 && pin_staging_area_index <= PIN_STAGING_AREA_SIZE);
+       if (concurrent_cementing)
+               hash = cement_hash_concurrent;
+       else
+               hash = cement_hash;
 
        /*
-        * The pinning addresses might come from undefined memory, this is normal. Since they
-        * are used in lots of functions, we make the memory defined here instead of having
-        * to add a supression for those functions.
+        * We use modulus hashing, which is fine with constants as gcc
+        * can optimize them to multiplication, but with variable
+        * values it would be a bad idea given armv7 has no hardware
+        * for division, making it 20x slower than a multiplication.
+        *
+        * This code path can be quite hot, depending on the workload,
+        * so if we make the hash size user-adjustable we should
+        * figure out something not involving division.
         */
-       VALGRIND_MAKE_MEM_DEFINED (pin_staging_area, pin_staging_area_index * sizeof (void*));
+       i = mono_aligned_addr_hash (obj) % SGEN_CEMENT_HASH_SIZE;
 
-       sort_addresses (pin_staging_area, pin_staging_area_index);
+       SGEN_ASSERT (5, sgen_ptr_in_nursery (obj), "Can only cement pointers to nursery objects");
 
-       while (next_pin_slot + pin_staging_area_index > pin_queue_size)
-               realloc_pin_queue ();
-
-       pin_queue [next_pin_slot++] = pin_staging_area [0];
-       for (i = 1; i < pin_staging_area_index; ++i) {
-               void *p = pin_staging_area [i];
-               if (p != pin_queue [next_pin_slot - 1])
-                       pin_queue [next_pin_slot++] = p;
+       if (!hash [i].obj) {
+               SGEN_ASSERT (5, !hash [i].count, "Cementing hash inconsistent");
+               hash [i].obj = obj;
+       } else if (hash [i].obj != obj) {
+               return FALSE;
        }
 
-       g_assert (next_pin_slot <= pin_queue_size);
+       if (hash [i].count >= SGEN_CEMENT_THRESHOLD)
+               return TRUE;
 
-       pin_staging_area_index = 0;
+       ++hash [i].count;
+       if (hash [i].count == SGEN_CEMENT_THRESHOLD) {
+               if (G_UNLIKELY (MONO_GC_OBJ_CEMENTED_ENABLED())) {
+                       MonoVTable *vt G_GNUC_UNUSED = (MonoVTable*)SGEN_LOAD_VTABLE (obj);
+                       MONO_GC_OBJ_CEMENTED ((mword)obj, sgen_safe_object_get_size ((MonoObject*)obj),
+                                       vt->klass->name_space, vt->klass->name);
+               }
+               binary_protocol_cement (obj, (gpointer)SGEN_LOAD_VTABLE (obj),
+                               sgen_safe_object_get_size ((MonoObject*)obj));
+       }
+
+       return FALSE;
 }
 
-static void
-pin_stage_ptr (void *ptr)
+void
+sgen_cement_iterate (IterateObjectCallbackFunc callback, void *callback_data)
 {
-       if (pin_staging_area_index >= PIN_STAGING_AREA_SIZE)
-               evacuate_pin_staging_area ();
+       int i;
+       for (i = 0; i < SGEN_CEMENT_HASH_SIZE; ++i) {
+               if (!cement_hash [i].count)
+                       continue;
 
-       pin_staging_area [pin_staging_area_index++] = ptr;
+               SGEN_ASSERT (5, cement_hash [i].count >= SGEN_CEMENT_THRESHOLD, "Cementing hash inconsistent");
+
+               callback (cement_hash [i].obj, 0, callback_data);
+       }
 }
+
+void
+sgen_cement_clear_below_threshold (void)
+{
+       int i;
+       for (i = 0; i < SGEN_CEMENT_HASH_SIZE; ++i) {
+               if (cement_hash [i].count < SGEN_CEMENT_THRESHOLD) {
+                       cement_hash [i].obj = NULL;
+                       cement_hash [i].count = 0;
+               }
+       }
+}
+
+#endif /* HAVE_SGEN_GC */