X-Git-Url: http://wien.tomnetworks.com/gitweb/?p=hs-boehmgc.git;a=blobdiff_plain;f=gc-7.2%2Ftests%2Ftest.c;fp=gc-7.2%2Ftests%2Ftest.c;h=7912b06ad246019cd87a6034795706c2f96b11cd;hp=0000000000000000000000000000000000000000;hb=324587ba93dc77f37406d41fd2a20d0e0d94fb1d;hpb=2a4ea609491b225a1ceb06da70396e93916f137a diff --git a/gc-7.2/tests/test.c b/gc-7.2/tests/test.c new file mode 100644 index 0000000..7912b06 --- /dev/null +++ b/gc-7.2/tests/test.c @@ -0,0 +1,1776 @@ +/* + * Copyright 1988, 1989 Hans-J. Boehm, Alan J. Demers + * Copyright (c) 1991-1994 by Xerox Corporation. All rights reserved. + * Copyright (c) 1996 by Silicon Graphics. All rights reserved. + * + * THIS MATERIAL IS PROVIDED AS IS, WITH ABSOLUTELY NO WARRANTY EXPRESSED + * OR IMPLIED. ANY USE IS AT YOUR OWN RISK. + * + * Permission is hereby granted to use or copy this program + * for any purpose, provided the above notices are retained on all copies. + * Permission to modify the code and to distribute modified code is granted, + * provided the above notices are retained, and a notice that the code was + * modified is included with the above copyright notice. + */ +/* An incomplete test for the garbage collector. */ +/* Some more obscure entry points are not tested at all. */ +/* This must be compiled with the same flags used to build the */ +/* GC. It uses GC internals to allow more precise results */ +/* checking for some of the tests. */ + +# ifdef HAVE_CONFIG_H +# include "private/config.h" +# endif + +# undef GC_BUILD + +#if (defined(DBG_HDRS_ALL) || defined(MAKE_BACK_GRAPH)) && !defined(GC_DEBUG) +# define GC_DEBUG +#endif + +#include "gc.h" + +#ifndef NTHREADS /* Number of additional threads to fork. */ +# define NTHREADS 5 /* excludes main thread, which also runs a test. */ + /* Not respected by PCR test. */ +#endif + +# if defined(mips) && defined(SYSTYPE_BSD43) + /* MIPS RISCOS 4 */ +# else +# include +# endif +# include +# if defined(_WIN32_WCE) && !defined(__GNUC__) +# include +/* # define assert ASSERT */ +# else +# include /* Not normally used, but handy for debugging. */ +# endif + +# include "gc_typed.h" +# include "private/gc_priv.h" /* For output, locking, MIN_WORDS, */ + /* some statistics and gcconfig.h. */ + +# if defined(MSWIN32) || defined(MSWINCE) +# include +# endif + +#ifdef GC_PRINT_VERBOSE_STATS +# define print_stats VERBOSE +# define INIT_PRINT_STATS /* empty */ +#else + /* Use own variable as GC_print_stats might not be exported. */ + static int print_stats = 0; +# ifdef GC_READ_ENV_FILE + /* GETENV uses GC internal function in this case. */ +# define INIT_PRINT_STATS /* empty */ +# else +# define INIT_PRINT_STATS \ + { \ + if (0 != GETENV("GC_PRINT_VERBOSE_STATS")) \ + print_stats = VERBOSE; \ + else if (0 != GETENV("GC_PRINT_STATS")) \ + print_stats = 1; \ + } +# endif +#endif /* !GC_PRINT_VERBOSE_STATS */ + +# ifdef PCR +# include "th/PCR_ThCrSec.h" +# include "th/PCR_Th.h" +# define GC_printf printf +# endif + +# if defined(GC_PTHREADS) +# include +# endif + +# if (!defined(THREADS) || !defined(HANDLE_FORK) \ + || (defined(DARWIN) && defined(MPROTECT_VDB) \ + && !defined(NO_INCREMENTAL) && !defined(MAKE_BACK_GRAPH))) \ + && !defined(NO_TEST_HANDLE_FORK) +# define NO_TEST_HANDLE_FORK +# endif + +# ifndef NO_TEST_HANDLE_FORK +# include +# define INIT_FORK_SUPPORT GC_set_handle_fork(1) +# else +# define INIT_FORK_SUPPORT /* empty */ +# endif + +# if defined(GC_WIN32_THREADS) && !defined(GC_PTHREADS) + static CRITICAL_SECTION incr_cs; +# endif + +# include + +#ifndef GC_ALPHA_VERSION +# define GC_ALPHA_VERSION GC_TMP_ALPHA_VERSION +#endif + +#define CHECH_GCLIB_VERSION \ + if (GC_get_version() != ((GC_VERSION_MAJOR<<16) \ + | (GC_VERSION_MINOR<<8) \ + | GC_ALPHA_VERSION)) { \ + GC_printf("libgc version mismatch\n"); \ + exit(1); \ + } + +/* Call GC_INIT only on platforms on which we think we really need it, */ +/* so that we can test automatic initialization on the rest. */ +#if defined(CYGWIN32) || defined (AIX) || defined(DARWIN) \ + || defined(THREAD_LOCAL_ALLOC) \ + || (defined(MSWINCE) && !defined(GC_WINMAIN_REDIRECT)) +# define GC_OPT_INIT GC_INIT() +#else +# define GC_OPT_INIT /* empty */ +#endif + +#define GC_COND_INIT() \ + INIT_FORK_SUPPORT; GC_OPT_INIT; CHECH_GCLIB_VERSION; INIT_PRINT_STATS + +#define CHECK_OUT_OF_MEMORY(p) \ + if ((p) == NULL) { \ + GC_printf("Out of memory\n"); \ + exit(1); \ + } + +/* Allocation Statistics. Incremented without synchronization. */ +/* FIXME: We should be using synchronization. */ +int stubborn_count = 0; +int uncollectable_count = 0; +int collectable_count = 0; +int atomic_count = 0; +int realloc_count = 0; + +#if defined(GC_AMIGA_FASTALLOC) && defined(AMIGA) + + void GC_amiga_free_all_mem(void); + void Amiga_Fail(void){GC_amiga_free_all_mem();abort();} +# define FAIL (void)Amiga_Fail() + void *GC_amiga_gctest_malloc_explicitly_typed(size_t lb, GC_descr d){ + void *ret=GC_malloc_explicitly_typed(lb,d); + if(ret==NULL){ + if(!GC_dont_gc){ + GC_gcollect(); + ret=GC_malloc_explicitly_typed(lb,d); + } + if(ret==NULL){ + GC_printf("Out of memory, (typed allocations are not directly " + "supported with the GC_AMIGA_FASTALLOC option.)\n"); + FAIL; + } + } + return ret; + } + void *GC_amiga_gctest_calloc_explicitly_typed(size_t a,size_t lb, GC_descr d){ + void *ret=GC_calloc_explicitly_typed(a,lb,d); + if(ret==NULL){ + if(!GC_dont_gc){ + GC_gcollect(); + ret=GC_calloc_explicitly_typed(a,lb,d); + } + if(ret==NULL){ + GC_printf("Out of memory, (typed allocations are not directly " + "supported with the GC_AMIGA_FASTALLOC option.)\n"); + FAIL; + } + } + return ret; + } +# define GC_malloc_explicitly_typed(a,b) GC_amiga_gctest_malloc_explicitly_typed(a,b) +# define GC_calloc_explicitly_typed(a,b,c) GC_amiga_gctest_calloc_explicitly_typed(a,b,c) + +#else /* !AMIGA_FASTALLOC */ + +# if defined(PCR) || defined(LINT2) +# define FAIL (void)abort() +# else +# define FAIL ABORT("Test failed") +# endif + +#endif /* !AMIGA_FASTALLOC */ + +/* AT_END may be defined to exercise the interior pointer test */ +/* if the collector is configured with ALL_INTERIOR_POINTERS. */ +/* As it stands, this test should succeed with either */ +/* configuration. In the FIND_LEAK configuration, it should */ +/* find lots of leaks, since we free almost nothing. */ + +struct SEXPR { + struct SEXPR * sexpr_car; + struct SEXPR * sexpr_cdr; +}; + + +typedef struct SEXPR * sexpr; + +# define INT_TO_SEXPR(x) ((sexpr)(GC_word)(x)) +# define SEXPR_TO_INT(x) ((int)(GC_word)(x)) + +# undef nil +# define nil (INT_TO_SEXPR(0)) +# define car(x) ((x) -> sexpr_car) +# define cdr(x) ((x) -> sexpr_cdr) +# define is_nil(x) ((x) == nil) + + +int extra_count = 0; /* Amount of space wasted in cons node */ + +/* Silly implementation of Lisp cons. Intentionally wastes lots of space */ +/* to test collector. */ +# ifdef VERY_SMALL_CONFIG +# define cons small_cons +# else +sexpr cons (sexpr x, sexpr y) +{ + sexpr r; + int *p; + int my_extra = extra_count; + + stubborn_count++; + r = (sexpr) GC_MALLOC_STUBBORN(sizeof(struct SEXPR) + my_extra); + CHECK_OUT_OF_MEMORY(r); + for (p = (int *)r; + ((char *)p) < ((char *)r) + my_extra + sizeof(struct SEXPR); p++) { + if (*p) { + GC_printf("Found nonzero at %p - allocator is broken\n", p); + FAIL; + } + *p = (int)((13 << 12) + ((p - (int *)r) & 0xfff)); + } +# ifdef AT_END + r = (sexpr)((char *)r + (my_extra & ~7)); +# endif + r -> sexpr_car = x; + r -> sexpr_cdr = y; + my_extra++; + if ( my_extra >= 5000 ) { + extra_count = 0; + } else { + extra_count = my_extra; + } + GC_END_STUBBORN_CHANGE((char *)r); + return(r); +} +# endif + +#ifdef GC_GCJ_SUPPORT + +#include "gc_mark.h" +#include "gc_gcj.h" + +/* The following struct emulates the vtable in gcj. */ +/* This assumes the default value of MARK_DESCR_OFFSET. */ +struct fake_vtable { + void * dummy; /* class pointer in real gcj. */ + GC_word descr; +}; + +struct fake_vtable gcj_class_struct1 = { 0, sizeof(struct SEXPR) + + sizeof(struct fake_vtable *) }; + /* length based descriptor. */ +struct fake_vtable gcj_class_struct2 = + { 0, ((GC_word)3 << (CPP_WORDSZ - 3)) | GC_DS_BITMAP}; + /* Bitmap based descriptor. */ + +struct GC_ms_entry * fake_gcj_mark_proc(word * addr, + struct GC_ms_entry *mark_stack_ptr, + struct GC_ms_entry *mark_stack_limit, + word env ) +{ + sexpr x; + if (1 == env) { + /* Object allocated with debug allocator. */ + addr = (word *)GC_USR_PTR_FROM_BASE(addr); + } + x = (sexpr)(addr + 1); /* Skip the vtable pointer. */ + mark_stack_ptr = GC_MARK_AND_PUSH( + (void *)(x -> sexpr_cdr), mark_stack_ptr, + mark_stack_limit, (void * *)&(x -> sexpr_cdr)); + mark_stack_ptr = GC_MARK_AND_PUSH( + (void *)(x -> sexpr_car), mark_stack_ptr, + mark_stack_limit, (void * *)&(x -> sexpr_car)); + return(mark_stack_ptr); +} + +#endif /* GC_GCJ_SUPPORT */ + + +sexpr small_cons (sexpr x, sexpr y) +{ + sexpr r; + + collectable_count++; + r = (sexpr) GC_MALLOC(sizeof(struct SEXPR)); + CHECK_OUT_OF_MEMORY(r); + r -> sexpr_car = x; + r -> sexpr_cdr = y; + return(r); +} + +sexpr small_cons_uncollectable (sexpr x, sexpr y) +{ + sexpr r; + + uncollectable_count++; + r = (sexpr) GC_MALLOC_UNCOLLECTABLE(sizeof(struct SEXPR)); + CHECK_OUT_OF_MEMORY(r); + r -> sexpr_car = x; + r -> sexpr_cdr = (sexpr)(~(GC_word)y); + return(r); +} + +#ifdef GC_GCJ_SUPPORT + + +sexpr gcj_cons(sexpr x, sexpr y) +{ + GC_word * r; + sexpr result; + + r = (GC_word *) GC_GCJ_MALLOC(sizeof(struct SEXPR) + + sizeof(struct fake_vtable*), + &gcj_class_struct2); + CHECK_OUT_OF_MEMORY(r); + result = (sexpr)(r + 1); + result -> sexpr_car = x; + result -> sexpr_cdr = y; + return(result); +} +#endif + +/* Return reverse(x) concatenated with y */ +sexpr reverse1(sexpr x, sexpr y) +{ + if (is_nil(x)) { + return(y); + } else { + return( reverse1(cdr(x), cons(car(x), y)) ); + } +} + +sexpr reverse(sexpr x) +{ +# ifdef TEST_WITH_SYSTEM_MALLOC + malloc(100000); +# endif + return( reverse1(x, nil) ); +} + +sexpr ints(int low, int up) +{ + if (low > up) { + return(nil); + } else { + return(small_cons(small_cons(INT_TO_SEXPR(low), nil), ints(low+1, up))); + } +} + +#ifdef GC_GCJ_SUPPORT +/* Return reverse(x) concatenated with y */ +sexpr gcj_reverse1(sexpr x, sexpr y) +{ + if (is_nil(x)) { + return(y); + } else { + return( gcj_reverse1(cdr(x), gcj_cons(car(x), y)) ); + } +} + +sexpr gcj_reverse(sexpr x) +{ + return( gcj_reverse1(x, nil) ); +} + +sexpr gcj_ints(int low, int up) +{ + if (low > up) { + return(nil); + } else { + return(gcj_cons(gcj_cons(INT_TO_SEXPR(low), nil), gcj_ints(low+1, up))); + } +} +#endif /* GC_GCJ_SUPPORT */ + +/* To check uncollectable allocation we build lists with disguised cdr */ +/* pointers, and make sure they don't go away. */ +sexpr uncollectable_ints(int low, int up) +{ + if (low > up) { + return(nil); + } else { + return(small_cons_uncollectable(small_cons(INT_TO_SEXPR(low), nil), + uncollectable_ints(low+1, up))); + } +} + +void check_ints(sexpr list, int low, int up) +{ + if (SEXPR_TO_INT(car(car(list))) != low) { + GC_printf( + "List reversal produced incorrect list - collector is broken\n"); + FAIL; + } + if (low == up) { + if (cdr(list) != nil) { + GC_printf("List too long - collector is broken\n"); + FAIL; + } + } else { + check_ints(cdr(list), low+1, up); + } +} + +# define UNCOLLECTABLE_CDR(x) (sexpr)(~(GC_word)(cdr(x))) + +void check_uncollectable_ints(sexpr list, int low, int up) +{ + if (SEXPR_TO_INT(car(car(list))) != low) { + GC_printf("Uncollectable list corrupted - collector is broken\n"); + FAIL; + } + if (low == up) { + if (UNCOLLECTABLE_CDR(list) != nil) { + GC_printf("Uncollectable list too long - collector is broken\n"); + FAIL; + } + } else { + check_uncollectable_ints(UNCOLLECTABLE_CDR(list), low+1, up); + } +} + +/* Not used, but useful for debugging: */ +void print_int_list(sexpr x) +{ + if (is_nil(x)) { + GC_printf("NIL\n"); + } else { + GC_printf("(%d)", SEXPR_TO_INT(car(car(x)))); + if (!is_nil(cdr(x))) { + GC_printf(", "); + print_int_list(cdr(x)); + } else { + GC_printf("\n"); + } + } +} + +/* ditto: */ +void check_marks_int_list(sexpr x) +{ + if (!GC_is_marked((ptr_t)x)) + GC_printf("[unm:%p]", x); + else + GC_printf("[mkd:%p]", x); + if (is_nil(x)) { + GC_printf("NIL\n"); + } else { + if (!GC_is_marked((ptr_t)car(x))) + GC_printf("[unm car:%p]", car(x)); + GC_printf("(%d)", SEXPR_TO_INT(car(car(x)))); + if (!is_nil(cdr(x))) { + GC_printf(", "); + check_marks_int_list(cdr(x)); + } else { + GC_printf("\n"); + } + } +} + +/* + * A tiny list reversal test to check thread creation. + */ +#ifdef THREADS + +# ifdef VERY_SMALL_CONFIG +# define TINY_REVERSE_UPPER_VALUE 4 +# else +# define TINY_REVERSE_UPPER_VALUE 10 +# endif + +# if defined(GC_WIN32_THREADS) && !defined(GC_PTHREADS) + DWORD __stdcall tiny_reverse_test(void * arg) +# else + void * tiny_reverse_test(void * arg) +# endif +{ + int i; + for (i = 0; i < 5; ++i) { + check_ints(reverse(reverse(ints(1, TINY_REVERSE_UPPER_VALUE))), + 1, TINY_REVERSE_UPPER_VALUE); + } + return 0; +} + +# if defined(GC_PTHREADS) + void fork_a_thread(void) + { + pthread_t t; + int code; + if ((code = pthread_create(&t, 0, tiny_reverse_test, 0)) != 0) { + GC_printf("Small thread creation failed %d\n", code); + FAIL; + } + if ((code = pthread_join(t, 0)) != 0) { + GC_printf("Small thread join failed %d\n", code); + FAIL; + } + } + +# elif defined(GC_WIN32_THREADS) + void fork_a_thread(void) + { + DWORD thread_id; + HANDLE h; + h = GC_CreateThread(NULL, 0, tiny_reverse_test, 0, 0, &thread_id); + if (h == (HANDLE)NULL) { + GC_printf("Small thread creation failed %d\n", + (int)GetLastError()); + FAIL; + } + if (WaitForSingleObject(h, INFINITE) != WAIT_OBJECT_0) { + GC_printf("Small thread wait failed %d\n", + (int)GetLastError()); + FAIL; + } + } + +# endif + +#endif + +/* Try to force a to be strangely aligned */ +struct { + char dummy; + sexpr aa; +} A; +#define a A.aa + +/* + * Repeatedly reverse lists built out of very different sized cons cells. + * Check that we didn't lose anything. + */ +void *GC_CALLBACK reverse_test_inner(void *data) +{ + int i; + sexpr b; + sexpr c; + sexpr d; + sexpr e; + sexpr *f, *g, *h; + + if (data == 0) { + /* This stack frame is not guaranteed to be scanned. */ + return GC_call_with_gc_active(reverse_test_inner, (void*)(word)1); + } + +# if /*defined(MSWIN32) ||*/ defined(MACOS) + /* Win32S only allows 128K stacks */ +# define BIG 1000 +# elif defined(PCR) + /* PCR default stack is 100K. Stack frames are up to 120 bytes. */ +# define BIG 700 +# elif defined(MSWINCE) || defined(RTEMS) + /* WinCE only allows 64K stacks */ +# define BIG 500 +# elif defined(OSF1) + /* OSF has limited stack space by default, and large frames. */ +# define BIG 200 +# elif defined(__MACH__) && defined(__ppc64__) +# define BIG 2500 +# else +# define BIG 4500 +# endif + + A.dummy = 17; + a = ints(1, 49); + b = ints(1, 50); + c = ints(1, BIG); + d = uncollectable_ints(1, 100); + e = uncollectable_ints(1, 1); + /* Check that realloc updates object descriptors correctly */ + collectable_count++; + f = (sexpr *)GC_MALLOC(4 * sizeof(sexpr)); + realloc_count++; + f = (sexpr *)GC_REALLOC((void *)f, 6 * sizeof(sexpr)); + CHECK_OUT_OF_MEMORY(f); + f[5] = ints(1,17); + collectable_count++; + g = (sexpr *)GC_MALLOC(513 * sizeof(sexpr)); + realloc_count++; + g = (sexpr *)GC_REALLOC((void *)g, 800 * sizeof(sexpr)); + CHECK_OUT_OF_MEMORY(g); + g[799] = ints(1,18); + collectable_count++; + h = (sexpr *)GC_MALLOC(1025 * sizeof(sexpr)); + realloc_count++; + h = (sexpr *)GC_REALLOC((void *)h, 2000 * sizeof(sexpr)); + CHECK_OUT_OF_MEMORY(h); +# ifdef GC_GCJ_SUPPORT + h[1999] = gcj_ints(1,200); + for (i = 0; i < 51; ++i) + h[1999] = gcj_reverse(h[1999]); + /* Leave it as the reveresed list for now. */ +# else + h[1999] = ints(1,200); +# endif + /* Try to force some collections and reuse of small list elements */ + for (i = 0; i < 10; i++) { + (void)ints(1, BIG); + } + /* Superficially test interior pointer recognition on stack */ + c = (sexpr)((char *)c + sizeof(char *)); + d = (sexpr)((char *)d + sizeof(char *)); + + GC_FREE((void *)e); + + check_ints(b,1,50); + check_ints(a,1,49); + for (i = 0; i < 50; i++) { + check_ints(b,1,50); + b = reverse(reverse(b)); + } + check_ints(b,1,50); + check_ints(a,1,49); + for (i = 0; i < 60; i++) { +# if defined(GC_PTHREADS) || defined(GC_WIN32_THREADS) + if (i % 10 == 0) fork_a_thread(); +# endif + /* This maintains the invariant that a always points to a list of */ + /* 49 integers. Thus this is thread safe without locks, */ + /* assuming atomic pointer assignments. */ + a = reverse(reverse(a)); +# if !defined(AT_END) && !defined(THREADS) + /* This is not thread safe, since realloc explicitly deallocates */ + if (i & 1) { + a = (sexpr)GC_REALLOC((void *)a, 500); + } else { + a = (sexpr)GC_REALLOC((void *)a, 8200); + } +# endif + } + check_ints(a,1,49); + check_ints(b,1,50); + + /* Restore c and d values. */ + c = (sexpr)((char *)c - sizeof(char *)); + d = (sexpr)((char *)d - sizeof(char *)); + + check_ints(c,1,BIG); + check_uncollectable_ints(d, 1, 100); + check_ints(f[5], 1,17); + check_ints(g[799], 1,18); +# ifdef GC_GCJ_SUPPORT + h[1999] = gcj_reverse(h[1999]); +# endif + check_ints(h[1999], 1,200); +# ifndef THREADS + a = 0; +# endif + *(sexpr volatile *)&b = 0; + *(sexpr volatile *)&c = 0; + return 0; +} + +void reverse_test(void) +{ + /* Test GC_do_blocking/GC_call_with_gc_active. */ + (void)GC_do_blocking(reverse_test_inner, 0); +} + +#undef a + +/* + * The rest of this builds balanced binary trees, checks that they don't + * disappear, and tests finalization. + */ +typedef struct treenode { + int level; + struct treenode * lchild; + struct treenode * rchild; +} tn; + +int finalizable_count = 0; +int finalized_count = 0; +volatile int dropped_something = 0; + +void GC_CALLBACK finalizer(void * obj, void * client_data) +{ + tn * t = (tn *)obj; + +# ifdef PCR + PCR_ThCrSec_EnterSys(); +# endif +# if defined(GC_PTHREADS) + static pthread_mutex_t incr_lock = PTHREAD_MUTEX_INITIALIZER; + pthread_mutex_lock(&incr_lock); +# elif defined(GC_WIN32_THREADS) + EnterCriticalSection(&incr_cs); +# endif + if ((int)(GC_word)client_data != t -> level) { + GC_printf("Wrong finalization data - collector is broken\n"); + FAIL; + } + finalized_count++; + t -> level = -1; /* detect duplicate finalization immediately */ +# ifdef PCR + PCR_ThCrSec_ExitSys(); +# endif +# if defined(GC_PTHREADS) + pthread_mutex_unlock(&incr_lock); +# elif defined(GC_WIN32_THREADS) + LeaveCriticalSection(&incr_cs); +# endif +} + +size_t counter = 0; + +# define MAX_FINALIZED (NTHREADS*4000) + +# if !defined(MACOS) + GC_FAR GC_word live_indicators[MAX_FINALIZED] = {0}; +#else + /* Too big for THINK_C. have to allocate it dynamically. */ + GC_word *live_indicators = 0; +#endif + +int live_indicators_count = 0; + +tn * mktree(int n) +{ + tn * result = (tn *)GC_MALLOC(sizeof(tn)); + + collectable_count++; +# if defined(MACOS) + /* get around static data limitations. */ + if (!live_indicators) { + live_indicators = + (GC_word*)NewPtrClear(MAX_FINALIZED * sizeof(GC_word)); + CHECK_OUT_OF_MEMORY(live_indicators); + } +# endif + if (n == 0) return(0); + CHECK_OUT_OF_MEMORY(result); + result -> level = n; + result -> lchild = mktree(n-1); + result -> rchild = mktree(n-1); + if (counter++ % 17 == 0 && n >= 2) { + tn * tmp; + + CHECK_OUT_OF_MEMORY(result->lchild); + tmp = result -> lchild -> rchild; + CHECK_OUT_OF_MEMORY(result->rchild); + result -> lchild -> rchild = result -> rchild -> lchild; + result -> rchild -> lchild = tmp; + } + if (counter++ % 119 == 0) { + int my_index; + + { +# ifdef PCR + PCR_ThCrSec_EnterSys(); +# endif +# if defined(GC_PTHREADS) + static pthread_mutex_t incr_lock = PTHREAD_MUTEX_INITIALIZER; + pthread_mutex_lock(&incr_lock); +# elif defined(GC_WIN32_THREADS) + EnterCriticalSection(&incr_cs); +# endif + /* Losing a count here causes erroneous report of failure. */ + finalizable_count++; + my_index = live_indicators_count++; +# ifdef PCR + PCR_ThCrSec_ExitSys(); +# endif +# if defined(GC_PTHREADS) + pthread_mutex_unlock(&incr_lock); +# elif defined(GC_WIN32_THREADS) + LeaveCriticalSection(&incr_cs); +# endif + } + + GC_REGISTER_FINALIZER((void *)result, finalizer, (void *)(GC_word)n, + (GC_finalization_proc *)0, (void * *)0); + if (my_index >= MAX_FINALIZED) { + GC_printf("live_indicators overflowed\n"); + FAIL; + } + live_indicators[my_index] = 13; + if (GC_GENERAL_REGISTER_DISAPPEARING_LINK( + (void * *)(&(live_indicators[my_index])), + (void *)result) != 0) { + GC_printf("GC_general_register_disappearing_link failed\n"); + FAIL; + } + if (GC_unregister_disappearing_link( + (void * *) + (&(live_indicators[my_index]))) == 0) { + GC_printf("GC_unregister_disappearing_link failed\n"); + FAIL; + } + if (GC_GENERAL_REGISTER_DISAPPEARING_LINK( + (void * *)(&(live_indicators[my_index])), + (void *)result) != 0) { + GC_printf("GC_general_register_disappearing_link failed 2\n"); + FAIL; + } + GC_reachable_here(result); + } + return(result); +} + +void chktree(tn *t, int n) +{ + if (n == 0 && t != 0) { + GC_printf("Clobbered a leaf - collector is broken\n"); + FAIL; + } + if (n == 0) return; + if (t -> level != n) { + GC_printf("Lost a node at level %d - collector is broken\n", n); + FAIL; + } + if (counter++ % 373 == 0) { + collectable_count++; + (void) GC_MALLOC(counter%5001); + } + chktree(t -> lchild, n-1); + if (counter++ % 73 == 0) { + collectable_count++; + (void) GC_MALLOC(counter%373); + } + chktree(t -> rchild, n-1); +} + + +#if defined(GC_PTHREADS) +pthread_key_t fl_key; + +void * alloc8bytes(void) +{ +# if defined(SMALL_CONFIG) || defined(GC_DEBUG) + collectable_count++; + return(GC_MALLOC(8)); +# else + void ** my_free_list_ptr; + void * my_free_list; + + my_free_list_ptr = (void **)pthread_getspecific(fl_key); + if (my_free_list_ptr == 0) { + uncollectable_count++; + my_free_list_ptr = GC_NEW_UNCOLLECTABLE(void *); + CHECK_OUT_OF_MEMORY(my_free_list_ptr); + if (pthread_setspecific(fl_key, my_free_list_ptr) != 0) { + GC_printf("pthread_setspecific failed\n"); + FAIL; + } + } + my_free_list = *my_free_list_ptr; + if (my_free_list == 0) { + my_free_list = GC_malloc_many(8); + CHECK_OUT_OF_MEMORY(my_free_list); + } + *my_free_list_ptr = GC_NEXT(my_free_list); + GC_NEXT(my_free_list) = 0; + collectable_count++; + return(my_free_list); +# endif +} + +#else +# define alloc8bytes() GC_MALLOC_ATOMIC(8) +#endif + +void alloc_small(int n) +{ + int i; + + for (i = 0; i < n; i += 8) { + atomic_count++; + if (alloc8bytes() == 0) { + GC_printf("Out of memory\n"); + FAIL; + } + } +} + +# if defined(THREADS) && defined(GC_DEBUG) +# ifdef VERY_SMALL_CONFIG +# define TREE_HEIGHT 12 +# else +# define TREE_HEIGHT 15 +# endif +# else +# ifdef VERY_SMALL_CONFIG +# define TREE_HEIGHT 13 +# else +# define TREE_HEIGHT 16 +# endif +# endif +void tree_test(void) +{ + tn * root; + int i; + + root = mktree(TREE_HEIGHT); +# ifndef VERY_SMALL_CONFIG + alloc_small(5000000); +# endif + chktree(root, TREE_HEIGHT); + if (finalized_count && ! dropped_something) { + GC_printf("Premature finalization - collector is broken\n"); + FAIL; + } + dropped_something = 1; + GC_noop1((word)root); /* Root needs to remain live until */ + /* dropped_something is set. */ + root = mktree(TREE_HEIGHT); + chktree(root, TREE_HEIGHT); + for (i = TREE_HEIGHT; i >= 0; i--) { + root = mktree(i); + chktree(root, i); + } +# ifndef VERY_SMALL_CONFIG + alloc_small(5000000); +# endif +} + +unsigned n_tests = 0; + +GC_word bm_huge[10] = { + 0xffffffff, + 0xffffffff, + 0xffffffff, + 0xffffffff, + 0xffffffff, + 0xffffffff, + 0xffffffff, + 0xffffffff, + 0xffffffff, + 0x00ffffff, +}; + +/* A very simple test of explicitly typed allocation */ +void typed_test(void) +{ + GC_word * old, * new; + GC_word bm3 = 0x3; + GC_word bm2 = 0x2; + GC_word bm_large = 0xf7ff7fff; + GC_descr d1 = GC_make_descriptor(&bm3, 2); + GC_descr d2 = GC_make_descriptor(&bm2, 2); + GC_descr d3 = GC_make_descriptor(&bm_large, 32); + GC_descr d4 = GC_make_descriptor(bm_huge, 320); + GC_word * x = (GC_word *)GC_malloc_explicitly_typed(2000, d4); + int i; + +# ifndef LINT + (void)GC_make_descriptor(&bm_large, 32); +# endif + collectable_count++; + old = 0; + for (i = 0; i < 4000; i++) { + collectable_count++; + new = (GC_word *) GC_malloc_explicitly_typed(4 * sizeof(GC_word), d1); + CHECK_OUT_OF_MEMORY(new); + if (0 != new[0] || 0 != new[1]) { + GC_printf("Bad initialization by GC_malloc_explicitly_typed\n"); + FAIL; + } + new[0] = 17; + new[1] = (GC_word)old; + old = new; + collectable_count++; + new = (GC_word *) GC_malloc_explicitly_typed(4 * sizeof(GC_word), d2); + CHECK_OUT_OF_MEMORY(new); + new[0] = 17; + new[1] = (GC_word)old; + old = new; + collectable_count++; + new = (GC_word *) GC_malloc_explicitly_typed(33 * sizeof(GC_word), d3); + CHECK_OUT_OF_MEMORY(new); + new[0] = 17; + new[1] = (GC_word)old; + old = new; + collectable_count++; + new = (GC_word *) GC_calloc_explicitly_typed(4, 2 * sizeof(GC_word), + d1); + CHECK_OUT_OF_MEMORY(new); + new[0] = 17; + new[1] = (GC_word)old; + old = new; + collectable_count++; + if (i & 0xff) { + new = (GC_word *) GC_calloc_explicitly_typed(7, 3 * sizeof(GC_word), + d2); + } else { + new = (GC_word *) GC_calloc_explicitly_typed(1001, + 3 * sizeof(GC_word), + d2); + if (new && (0 != new[0] || 0 != new[1])) { + GC_printf("Bad initialization by GC_malloc_explicitly_typed\n"); + FAIL; + } + } + CHECK_OUT_OF_MEMORY(new); + new[0] = 17; + new[1] = (GC_word)old; + old = new; + } + for (i = 0; i < 20000; i++) { + if (new[0] != 17) { + GC_printf("typed alloc failed at %lu\n", (unsigned long)i); + FAIL; + } + new[0] = 0; + old = new; + new = (GC_word *)(old[1]); + } + GC_gcollect(); + GC_noop1((word)x); +} + +int fail_count = 0; + +/*ARGSUSED*/ +void GC_CALLBACK fail_proc1(void * x) +{ + fail_count++; +} + +static void uniq(void *p, ...) { + va_list a; + void *q[100]; + int n = 0, i, j; + q[n++] = p; + va_start(a,p); + for (;(q[n] = va_arg(a,void *)) != NULL;n++) ; + va_end(a); + for (i=0; i= (n)) +#endif + +void * GC_CALLBACK inc_int_counter(void *pcounter) +{ + ++(*(int *)pcounter); + return NULL; +} + +void run_one_test(void) +{ +# ifndef DBG_HDRS_ALL + char *x; + char **z; +# ifdef LINT + char *y = 0; +# else + char *y = (char *)(GC_word)fail_proc1; +# endif + CLOCK_TYPE typed_time; +# endif + CLOCK_TYPE start_time; + CLOCK_TYPE reverse_time; + CLOCK_TYPE tree_time; + unsigned long time_diff; + +# ifdef FIND_LEAK + GC_printf( + "This test program is not designed for leak detection mode\n"); + GC_printf("Expect lots of problems\n"); +# endif + GC_FREE(0); +# ifndef DBG_HDRS_ALL + collectable_count += 3; + if ((GC_size(GC_malloc(7)) != 8 && + GC_size(GC_malloc(7)) != MIN_WORDS * sizeof(GC_word)) + || GC_size(GC_malloc(15)) != 16) { + GC_printf("GC_size produced unexpected results\n"); + FAIL; + } + collectable_count += 1; + if (GC_size(GC_malloc(0)) != MIN_WORDS * sizeof(GC_word)) { + GC_printf("GC_malloc(0) failed: GC_size returns %ld\n", + (unsigned long)GC_size(GC_malloc(0))); + FAIL; + } + collectable_count += 1; + if (GC_size(GC_malloc_uncollectable(0)) != MIN_WORDS * sizeof(GC_word)) { + GC_printf("GC_malloc_uncollectable(0) failed\n"); + FAIL; + } + GC_is_valid_displacement_print_proc = fail_proc1; + GC_is_visible_print_proc = fail_proc1; + collectable_count += 1; + x = GC_malloc(16); + if (GC_base(GC_PTR_ADD(x, 13)) != x) { + GC_printf("GC_base(heap ptr) produced incorrect result\n"); + FAIL; + } + (void)GC_PRE_INCR(x, 0); + (void)GC_POST_INCR(x); + (void)GC_POST_DECR(x); + if (GC_base(x) != x) { + GC_printf("Bad INCR/DECR result\n"); + FAIL; + } +# ifndef PCR + if (GC_base(y) != 0) { + GC_printf("GC_base(fn_ptr) produced incorrect result\n"); + FAIL; + } +# endif + if (GC_same_obj(x+5, x) != x + 5) { + GC_printf("GC_same_obj produced incorrect result\n"); + FAIL; + } + if (GC_is_visible(y) != y || GC_is_visible(x) != x) { + GC_printf("GC_is_visible produced incorrect result\n"); + FAIL; + } + z = GC_malloc(8); + GC_PTR_STORE(z, x); + if (*z != x) { + GC_printf("GC_PTR_STORE failed: %p != %p\n", *z, x); + FAIL; + } + if (!TEST_FAIL_COUNT(1)) { +# if!(defined(POWERPC) || defined(IA64)) || defined(M68K) + /* On POWERPCs function pointers point to a descriptor in the */ + /* data segment, so there should have been no failures. */ + /* The same applies to IA64. Something similar seems to */ + /* be going on with NetBSD/M68K. */ + GC_printf("GC_is_visible produced wrong failure indication\n"); + FAIL; +# endif + } + if (GC_is_valid_displacement(y) != y + || GC_is_valid_displacement(x) != x + || GC_is_valid_displacement(x + 3) != x + 3) { + GC_printf("GC_is_valid_displacement produced incorrect result\n"); + FAIL; + } + { + size_t i; + + GC_malloc(17); + for (i = sizeof(GC_word); i < 512; i *= 2) { + GC_word result = (GC_word) GC_memalign(i, 17); + if (result % i != 0 || result == 0 || *(int *)result != 0) FAIL; + } + } +# ifndef ALL_INTERIOR_POINTERS +# if defined(RS6000) || defined(POWERPC) + if (!TEST_FAIL_COUNT(1)) +# else + if (!TEST_FAIL_COUNT(GC_get_all_interior_pointers() ? 1 : 2)) +# endif + { + GC_printf( + "GC_is_valid_displacement produced wrong failure indication\n"); + FAIL; + } +# endif +# endif /* DBG_HDRS_ALL */ + /* Test floating point alignment */ + collectable_count += 2; + { + double *dp = GC_MALLOC(sizeof(double)); + CHECK_OUT_OF_MEMORY(dp); + *dp = 1.0; + dp = GC_MALLOC(sizeof(double)); + CHECK_OUT_OF_MEMORY(dp); + *dp = 1.0; + } + /* Test size 0 allocation a bit more */ + { + size_t i; + for (i = 0; i < 10000; ++i) { + GC_MALLOC(0); + GC_FREE(GC_MALLOC(0)); + GC_MALLOC_ATOMIC(0); + GC_FREE(GC_MALLOC_ATOMIC(0)); + } + } +# ifdef GC_GCJ_SUPPORT + GC_REGISTER_DISPLACEMENT(sizeof(struct fake_vtable *)); + GC_init_gcj_malloc(0, (void *)(GC_word)fake_gcj_mark_proc); +# endif + /* Make sure that fn arguments are visible to the collector. */ + uniq( + GC_malloc(12), GC_malloc(12), GC_malloc(12), + (GC_gcollect(),GC_malloc(12)), + GC_malloc(12), GC_malloc(12), GC_malloc(12), + (GC_gcollect(),GC_malloc(12)), + GC_malloc(12), GC_malloc(12), GC_malloc(12), + (GC_gcollect(),GC_malloc(12)), + GC_malloc(12), GC_malloc(12), GC_malloc(12), + (GC_gcollect(),GC_malloc(12)), + GC_malloc(12), GC_malloc(12), GC_malloc(12), + (GC_gcollect(),GC_malloc(12)), + (void *)0); + /* GC_malloc(0) must return NULL or something we can deallocate. */ + GC_free(GC_malloc(0)); + GC_free(GC_malloc_atomic(0)); + GC_free(GC_malloc(0)); + GC_free(GC_malloc_atomic(0)); + /* Repeated list reversal test. */ + GET_TIME(start_time); + reverse_test(); + if (print_stats) { + GET_TIME(reverse_time); + time_diff = MS_TIME_DIFF(reverse_time, start_time); + GC_log_printf("-------------Finished reverse_test at time %u (%p)\n", + (unsigned) time_diff, &start_time); + } +# ifndef DBG_HDRS_ALL + typed_test(); + if (print_stats) { + GET_TIME(typed_time); + time_diff = MS_TIME_DIFF(typed_time, start_time); + GC_log_printf("-------------Finished typed_test at time %u (%p)\n", + (unsigned) time_diff, &start_time); + } +# endif /* DBG_HDRS_ALL */ + tree_test(); + if (print_stats) { + GET_TIME(tree_time); + time_diff = MS_TIME_DIFF(tree_time, start_time); + GC_log_printf("-------------Finished tree_test at time %u (%p)\n", + (unsigned) time_diff, &start_time); + } + /* Run reverse_test a second time, so we hopefully notice corruption. */ + reverse_test(); + if (print_stats) { + GET_TIME(reverse_time); + time_diff = MS_TIME_DIFF(reverse_time, start_time); + GC_log_printf( + "-------------Finished second reverse_test at time %u (%p)\n", + (unsigned)time_diff, &start_time); + } + /* GC_allocate_ml and GC_need_to_lock are no longer exported, and */ + /* AO_fetch_and_add1() may be unavailable to update a counter. */ + (void)GC_call_with_alloc_lock(inc_int_counter, &n_tests); +# ifndef NO_TEST_HANDLE_FORK + if (fork() == 0) { + GC_gcollect(); + tiny_reverse_test(0); + GC_gcollect(); + if (print_stats) + GC_log_printf("Finished a child process\n"); + exit(0); + } +# endif + if (print_stats) + GC_log_printf("Finished %p\n", &start_time); +} + +#define NUMBER_ROUND_UP(v, bound) ((((v) + (bound) - 1) / (bound)) * (bound)) + +void check_heap_stats(void) +{ + size_t max_heap_sz; + int i; + int still_live; +# ifdef FINALIZE_ON_DEMAND + int late_finalize_count = 0; +# endif + +# ifdef VERY_SMALL_CONFIG + /* The upper bounds are a guess, which has been empirically */ + /* adjusted. On low end uniprocessors with incremental GC */ + /* these may be particularly dubious, since empirically the */ + /* heap tends to grow largely as a result of the GC not */ + /* getting enough cycles. */ +# if CPP_WORDSZ == 64 + max_heap_sz = 4500000; +# else + max_heap_sz = 2800000; +# endif +# else +# if CPP_WORDSZ == 64 + max_heap_sz = 23000000; +# else + max_heap_sz = 16000000; +# endif +# endif +# ifdef GC_DEBUG + max_heap_sz *= 2; +# ifdef SAVE_CALL_CHAIN + max_heap_sz *= 3; +# ifdef SAVE_CALL_COUNT + max_heap_sz += max_heap_sz * SAVE_CALL_COUNT/4; +# endif +# endif +# endif + max_heap_sz *= n_tests; +# if defined(USE_MMAP) || defined(MSWIN32) + max_heap_sz = NUMBER_ROUND_UP(max_heap_sz, 4 * 1024 * 1024); +# endif + /* Garbage collect repeatedly so that all inaccessible objects */ + /* can be finalized. */ + while (GC_collect_a_little()) { } + for (i = 0; i < 16; i++) { + GC_gcollect(); +# ifdef FINALIZE_ON_DEMAND + late_finalize_count += +# endif + GC_invoke_finalizers(); + } + if (print_stats) { + GC_log_printf("Primordial thread stack bottom: %p\n", + GC_stackbottom); + } + GC_printf("Completed %u tests\n", n_tests); + GC_printf("Allocated %d collectable objects\n", collectable_count); + GC_printf("Allocated %d uncollectable objects\n", + uncollectable_count); + GC_printf("Allocated %d atomic objects\n", atomic_count); + GC_printf("Allocated %d stubborn objects\n", stubborn_count); + GC_printf("Finalized %d/%d objects - ", + finalized_count, finalizable_count); +# ifdef FINALIZE_ON_DEMAND + if (finalized_count != late_finalize_count) { + GC_printf("Demand finalization error\n"); + FAIL; + } +# endif + if (finalized_count > finalizable_count + || finalized_count < finalizable_count/2) { + GC_printf("finalization is probably broken\n"); + FAIL; + } else { + GC_printf("finalization is probably ok\n"); + } + still_live = 0; + for (i = 0; i < MAX_FINALIZED; i++) { + if (live_indicators[i] != 0) { + still_live++; + } + } + i = finalizable_count - finalized_count - still_live; + if (0 != i) { + GC_printf("%d disappearing links remain and %d more objects " + "were not finalized\n", still_live, i); + if (i > 10) { + GC_printf("\tVery suspicious!\n"); + } else { + GC_printf("\tSlightly suspicious, but probably OK\n"); + } + } + GC_printf("Total number of bytes allocated is %lu\n", + (unsigned long)GC_get_total_bytes()); + GC_printf("Final heap size is %lu bytes\n", + (unsigned long)GC_get_heap_size()); + if (GC_get_total_bytes() < n_tests * +# ifdef VERY_SMALL_CONFIG + 2700000 +# else + 33500000 +# endif + ) { + GC_printf("Incorrect execution - missed some allocations\n"); + FAIL; + } + if (GC_get_heap_size() + GC_get_unmapped_bytes() > max_heap_sz) { + GC_printf("Unexpected heap growth - collector may be broken" + " (heapsize: %lu, expected: %lu)\n", + (unsigned long)(GC_get_heap_size() + GC_get_unmapped_bytes()), + (unsigned long)max_heap_sz); + FAIL; + } +# ifdef THREADS + GC_unregister_my_thread(); /* just to check it works (for main) */ +# endif + GC_printf("Collector appears to work\n"); +} + +#if defined(MACOS) +void SetMinimumStack(long minSize) +{ + long newApplLimit; + + if (minSize > LMGetDefltStack()) + { + newApplLimit = (long) GetApplLimit() + - (minSize - LMGetDefltStack()); + SetApplLimit((Ptr) newApplLimit); + MaxApplZone(); + } +} + +#define cMinStackSpace (512L * 1024L) + +#endif + +void GC_CALLBACK warn_proc(char *msg, GC_word p) +{ + GC_printf(msg, (unsigned long)p); + /*FAIL;*/ +} + +#if defined(MSWINCE) && defined(UNDER_CE) +# define WINMAIN_LPTSTR LPWSTR +#else +# define WINMAIN_LPTSTR LPSTR +#endif + +#if !defined(PCR) && !defined(GC_WIN32_THREADS) && !defined(GC_PTHREADS) \ + || defined(LINT) +#if defined(MSWIN32) && !defined(__MINGW32__) || defined(MSWINCE) + int APIENTRY WinMain(HINSTANCE instance, HINSTANCE prev, + WINMAIN_LPTSTR cmd, int n) +#elif defined(RTEMS) +# include +# define CONFIGURE_APPLICATION_NEEDS_CLOCK_DRIVER +# define CONFIGURE_APPLICATION_NEEDS_CONSOLE_DRIVER +# define CONFIGURE_RTEMS_INIT_TASKS_TABLE +# define CONFIGURE_MAXIMUM_TASKS 1 +# define CONFIGURE_INIT +# define CONFIGURE_INIT_TASK_STACK_SIZE (64*1024) +# include + rtems_task Init(rtems_task_argument ignord) +#else + int main(void) +#endif +{ + n_tests = 0; +# if defined(MACOS) + /* Make sure we have lots and lots of stack space. */ + SetMinimumStack(cMinStackSpace); + /* Cheat and let stdio initialize toolbox for us. */ + printf("Testing GC Macintosh port\n"); +# endif + GC_COND_INIT(); + GC_set_warn_proc(warn_proc); +# if (defined(MPROTECT_VDB) || defined(PROC_VDB) || defined(GWW_VDB)) \ + && !defined(MAKE_BACK_GRAPH) && !defined(NO_INCREMENTAL) + GC_enable_incremental(); + GC_printf("Switched to incremental mode\n"); +# if defined(MPROTECT_VDB) + GC_printf("Emulating dirty bits with mprotect/signals\n"); +# else +# ifdef PROC_VDB + GC_printf("Reading dirty bits from /proc\n"); +# elif defined(GWW_VDB) + GC_printf("Using GetWriteWatch-based implementation\n"); +# else + GC_printf("Using DEFAULT_VDB dirty bit implementation\n"); +# endif +# endif +# endif + run_one_test(); + check_heap_stats(); +# ifndef MSWINCE + fflush(stdout); +# endif +# ifdef LINT + /* Entry points we should be testing, but aren't. */ + /* Some can be tested by defining GC_DEBUG at the top of this file */ + /* This is a bit SunOS4 specific. */ + GC_noop(GC_expand_hp, GC_add_roots, GC_clear_roots, + GC_register_disappearing_link, + GC_register_finalizer_ignore_self, + GC_debug_register_displacement, GC_debug_change_stubborn, + GC_debug_end_stubborn_change, GC_debug_malloc_uncollectable, + GC_debug_free, GC_debug_realloc, + GC_generic_malloc_words_small, GC_init, + GC_malloc_ignore_off_page, GC_malloc_atomic_ignore_off_page, + GC_set_max_heap_size, GC_get_bytes_since_gc, + GC_get_total_bytes, GC_pre_incr, GC_post_incr); +# endif +# ifdef MSWIN32 + GC_win32_free_heap(); +# endif +# ifdef RTEMS + exit(0); +# else + return(0); +# endif +} +# endif + +#if defined(GC_WIN32_THREADS) && !defined(GC_PTHREADS) + +DWORD __stdcall thr_run_one_test(void *arg) +{ + run_one_test(); + return 0; +} + +#ifdef MSWINCE +HANDLE win_created_h; +HWND win_handle; + +LRESULT CALLBACK window_proc(HWND hwnd, UINT uMsg, WPARAM wParam, LPARAM lParam) +{ + LRESULT ret = 0; + switch (uMsg) { + case WM_HIBERNATE: + GC_printf("Received WM_HIBERNATE, calling GC_gcollect\n"); + /* Force "unmap as much memory as possible" mode. */ + GC_gcollect_and_unmap(); + break; + case WM_CLOSE: + GC_printf("Received WM_CLOSE, closing window\n"); + DestroyWindow(hwnd); + break; + case WM_DESTROY: + PostQuitMessage(0); + break; + default: + ret = DefWindowProc(hwnd, uMsg, wParam, lParam); + break; + } + return ret; +} + +DWORD __stdcall thr_window(void *arg) +{ + WNDCLASS win_class = { + CS_NOCLOSE, + window_proc, + 0, + 0, + GetModuleHandle(NULL), + NULL, + NULL, + (HBRUSH)(COLOR_APPWORKSPACE+1), + NULL, + TEXT("GCtestWindow") + }; + MSG msg; + + if (!RegisterClass(&win_class)) + FAIL; + + win_handle = CreateWindowEx( + 0, + TEXT("GCtestWindow"), + TEXT("GCtest"), + 0, + CW_USEDEFAULT, CW_USEDEFAULT, CW_USEDEFAULT, CW_USEDEFAULT, + NULL, + NULL, + GetModuleHandle(NULL), + NULL); + + if (win_handle == NULL) + FAIL; + + SetEvent(win_created_h); + + ShowWindow(win_handle, SW_SHOW); + UpdateWindow(win_handle); + + while (GetMessage(&msg, NULL, 0, 0)) { + TranslateMessage(&msg); + DispatchMessage(&msg); + } + + return 0; +} +#endif + +int APIENTRY WinMain(HINSTANCE instance, HINSTANCE prev, + WINMAIN_LPTSTR cmd, int n) +{ +# if NTHREADS > 0 + HANDLE h[NTHREADS]; + int i; +# endif +# ifdef MSWINCE + HANDLE win_thr_h; +# endif + DWORD thread_id; +# if defined(GC_DLL) && !defined(GC_NO_THREADS_DISCOVERY) \ + && !defined(MSWINCE) && !defined(THREAD_LOCAL_ALLOC) \ + && !defined(PARALLEL_MARK) + GC_use_threads_discovery(); + /* Test with implicit thread registration if possible. */ + GC_printf("Using DllMain to track threads\n"); +# endif + GC_COND_INIT(); +# if !defined(MAKE_BACK_GRAPH) && !defined(NO_INCREMENTAL) + GC_enable_incremental(); +# endif + InitializeCriticalSection(&incr_cs); + GC_set_warn_proc(warn_proc); +# ifdef MSWINCE + win_created_h = CreateEvent(NULL, FALSE, FALSE, NULL); + if (win_created_h == (HANDLE)NULL) { + GC_printf("Event creation failed %d\n", (int)GetLastError()); + FAIL; + } + win_thr_h = GC_CreateThread(NULL, 0, thr_window, 0, 0, &thread_id); + if (win_thr_h == (HANDLE)NULL) { + GC_printf("Thread creation failed %d\n", (int)GetLastError()); + FAIL; + } + if (WaitForSingleObject(win_created_h, INFINITE) != WAIT_OBJECT_0) + FAIL; + CloseHandle(win_created_h); +# endif +# if NTHREADS > 0 + for (i = 0; i < NTHREADS; i++) { + h[i] = GC_CreateThread(NULL, 0, thr_run_one_test, 0, 0, &thread_id); + if (h[i] == (HANDLE)NULL) { + GC_printf("Thread creation failed %d\n", (int)GetLastError()); + FAIL; + } + } +# endif /* NTHREADS > 0 */ + run_one_test(); +# if NTHREADS > 0 + for (i = 0; i < NTHREADS; i++) { + if (WaitForSingleObject(h[i], INFINITE) != WAIT_OBJECT_0) { + GC_printf("Thread wait failed %d\n", (int)GetLastError()); + FAIL; + } + } +# endif /* NTHREADS > 0 */ +# ifdef MSWINCE + PostMessage(win_handle, WM_CLOSE, 0, 0); + if (WaitForSingleObject(win_thr_h, INFINITE) != WAIT_OBJECT_0) + FAIL; +# endif + check_heap_stats(); + return(0); +} + +#endif /* GC_WIN32_THREADS */ + + +#ifdef PCR +int test(void) +{ + PCR_Th_T * th1; + PCR_Th_T * th2; + int code; + + n_tests = 0; + /* GC_enable_incremental(); */ + GC_set_warn_proc(warn_proc); + th1 = PCR_Th_Fork(run_one_test, 0); + th2 = PCR_Th_Fork(run_one_test, 0); + run_one_test(); + if (PCR_Th_T_Join(th1, &code, NIL, PCR_allSigsBlocked, PCR_waitForever) + != PCR_ERes_okay || code != 0) { + GC_printf("Thread 1 failed\n"); + } + if (PCR_Th_T_Join(th2, &code, NIL, PCR_allSigsBlocked, PCR_waitForever) + != PCR_ERes_okay || code != 0) { + GC_printf("Thread 2 failed\n"); + } + check_heap_stats(); + return(0); +} +#endif + +#if defined(GC_PTHREADS) +void * thr_run_one_test(void * arg) +{ + run_one_test(); + return(0); +} + +#ifdef GC_DEBUG +# define GC_free GC_debug_free +#endif + +int main(void) +{ + pthread_t th[NTHREADS]; + pthread_attr_t attr; + int code; + int i; +# ifdef GC_IRIX_THREADS + /* Force a larger stack to be preallocated */ + /* Since the initial can't always grow later. */ + *((volatile char *)&code - 1024*1024) = 0; /* Require 1 MB */ +# endif /* GC_IRIX_THREADS */ +# if defined(GC_HPUX_THREADS) + /* Default stack size is too small, especially with the 64 bit ABI */ + /* Increase it. */ + if (pthread_default_stacksize_np(1024*1024, 0) != 0) { + GC_printf("pthread_default_stacksize_np failed\n"); + } +# endif /* GC_HPUX_THREADS */ +# ifdef PTW32_STATIC_LIB + pthread_win32_process_attach_np (); + pthread_win32_thread_attach_np (); +# endif +# if defined(GC_DARWIN_THREADS) && !defined(GC_NO_THREADS_DISCOVERY) \ + && !defined(DARWIN_DONT_PARSE_STACK) && !defined(THREAD_LOCAL_ALLOC) + /* Test with the Darwin implicit thread registration. */ + GC_use_threads_discovery(); + GC_printf("Using Darwin task-threads-based world stop and push\n"); +# endif + GC_COND_INIT(); + + pthread_attr_init(&attr); +# if defined(GC_IRIX_THREADS) || defined(GC_FREEBSD_THREADS) \ + || defined(GC_DARWIN_THREADS) || defined(GC_AIX_THREADS) \ + || defined(GC_OPENBSD_THREADS) + pthread_attr_setstacksize(&attr, 1000000); +# endif + n_tests = 0; +# if (defined(MPROTECT_VDB)) && !defined(REDIRECT_MALLOC) \ + && !defined(MAKE_BACK_GRAPH) && !defined(USE_PROC_FOR_LIBRARIES) \ + && !defined(NO_INCREMENTAL) + GC_enable_incremental(); + GC_printf("Switched to incremental mode\n"); +# if defined(MPROTECT_VDB) + GC_printf("Emulating dirty bits with mprotect/signals\n"); +# else +# ifdef PROC_VDB + GC_printf("Reading dirty bits from /proc\n"); +# else + GC_printf("Using DEFAULT_VDB dirty bit implementation\n"); +# endif +# endif +# endif + GC_set_warn_proc(warn_proc); + if ((code = pthread_key_create(&fl_key, 0)) != 0) { + GC_printf("Key creation failed %d\n", code); + FAIL; + } + for (i = 0; i < NTHREADS; ++i) { + if ((code = pthread_create(th+i, &attr, thr_run_one_test, 0)) != 0) { + GC_printf("Thread %d creation failed %d\n", i, code); + FAIL; + } + } + run_one_test(); + for (i = 0; i < NTHREADS; ++i) { + if ((code = pthread_join(th[i], 0)) != 0) { + GC_printf("Thread %d failed %d\n", i, code); + FAIL; + } + } + check_heap_stats(); + (void)fflush(stdout); + pthread_attr_destroy(&attr); + GC_printf("Completed %u collections\n", (unsigned)GC_get_gc_no()); +# ifdef PTW32_STATIC_LIB + pthread_win32_thread_detach_np (); + pthread_win32_process_detach_np (); +# endif + return(0); +} +#endif /* GC_PTHREADS */