summaryrefslogtreecommitdiff
path: root/libjava/boehm.cc
diff options
context:
space:
mode:
authorupstream source tree <ports@midipix.org>2015-03-15 20:14:05 -0400
committerupstream source tree <ports@midipix.org>2015-03-15 20:14:05 -0400
commit554fd8c5195424bdbcabf5de30fdc183aba391bd (patch)
tree976dc5ab7fddf506dadce60ae936f43f58787092 /libjava/boehm.cc
downloadcbb-gcc-4.6.4-554fd8c5195424bdbcabf5de30fdc183aba391bd.tar.bz2
cbb-gcc-4.6.4-554fd8c5195424bdbcabf5de30fdc183aba391bd.tar.xz
obtained gcc-4.6.4.tar.bz2 from upstream website;upstream
verified gcc-4.6.4.tar.bz2.sig; imported gcc-4.6.4 source tree from verified upstream tarball. downloading a git-generated archive based on the 'upstream' tag should provide you with a source tree that is binary identical to the one extracted from the above tarball. if you have obtained the source via the command 'git clone', however, do note that line-endings of files in your working directory might differ from line-endings of the respective files in the upstream repository.
Diffstat (limited to 'libjava/boehm.cc')
-rw-r--r--libjava/boehm.cc761
1 files changed, 761 insertions, 0 deletions
diff --git a/libjava/boehm.cc b/libjava/boehm.cc
new file mode 100644
index 000000000..855d23cb0
--- /dev/null
+++ b/libjava/boehm.cc
@@ -0,0 +1,761 @@
+// boehm.cc - interface between libjava and Boehm GC.
+
+/* Copyright (C) 1998, 1999, 2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007
+ Free Software Foundation
+
+ This file is part of libgcj.
+
+This software is copyrighted work licensed under the terms of the
+Libgcj License. Please consult the file "LIBGCJ_LICENSE" for
+details. */
+
+#include <config.h>
+
+#include <stdio.h>
+#include <limits.h>
+
+#include <jvm.h>
+#include <gcj/cni.h>
+
+#include <java/lang/Class.h>
+#include <java/lang/reflect/Modifier.h>
+#include <java-interp.h>
+
+// More nastiness: the GC wants to define TRUE and FALSE. We don't
+// need the Java definitions (themselves a hack), so we undefine them.
+#undef TRUE
+#undef FALSE
+
+// We include two autoconf headers. Avoid multiple definition warnings.
+#undef PACKAGE_NAME
+#undef PACKAGE_STRING
+#undef PACKAGE_TARNAME
+#undef PACKAGE_VERSION
+
+#ifdef HAVE_DLFCN_H
+#undef _GNU_SOURCE
+#define _GNU_SOURCE
+#include <dlfcn.h>
+#endif
+
+extern "C"
+{
+#include <gc_config.h>
+
+// Set GC_DEBUG before including gc.h!
+#ifdef LIBGCJ_GC_DEBUG
+# define GC_DEBUG
+#endif
+
+#include <gc_mark.h>
+#include <gc_gcj.h>
+#include <javaxfc.h> // GC_finalize_all declaration.
+
+#ifdef THREAD_LOCAL_ALLOC
+# define GC_REDIRECT_TO_LOCAL
+# include <gc_local_alloc.h>
+#endif
+
+ // From boehm's misc.c
+ void GC_enable();
+ void GC_disable();
+};
+
+#define MAYBE_MARK(Obj, Top, Limit, Source) \
+ Top=GC_MARK_AND_PUSH((GC_PTR) Obj, Top, Limit, (GC_PTR *) Source)
+
+// `kind' index used when allocating Java arrays.
+static int array_kind_x;
+
+// Freelist used for Java arrays.
+static void **array_free_list;
+
+static int _Jv_GC_has_static_roots (const char *filename, void *, size_t);
+
+
+
+// This is called by the GC during the mark phase. It marks a Java
+// object. We use `void *' arguments and return, and not what the
+// Boehm GC wants, to avoid pollution in our headers.
+void *
+_Jv_MarkObj (void *addr, void *msp, void *msl, void *env)
+{
+ struct GC_ms_entry *mark_stack_ptr = (struct GC_ms_entry *)msp;
+ struct GC_ms_entry *mark_stack_limit = (struct GC_ms_entry *)msl;
+
+ if (env == (void *)1) /* Object allocated with debug allocator. */
+ addr = (GC_PTR)GC_USR_PTR_FROM_BASE(addr);
+ jobject obj = (jobject) addr;
+
+ _Jv_VTable *dt = *(_Jv_VTable **) addr;
+ // The object might not yet have its vtable set, or it might
+ // really be an object on the freelist. In either case, the vtable slot
+ // will either be 0, or it will point to a cleared object.
+ // This assumes Java objects have size at least 3 words,
+ // including the header. But this should remain true, since this
+ // should only be used with debugging allocation or with large objects.
+ if (__builtin_expect (! dt || !(dt -> get_finalizer()), false))
+ return mark_stack_ptr;
+ jclass klass = dt->clas;
+ GC_PTR p;
+
+ p = (GC_PTR) dt;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, obj);
+
+# ifndef JV_HASH_SYNCHRONIZATION
+ // Every object has a sync_info pointer.
+ p = (GC_PTR) obj->sync_info;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, obj);
+# endif
+
+ if (__builtin_expect (klass == &java::lang::Class::class$, false))
+ {
+ // Currently we allocate some of the memory referenced from class objects
+ // as pointerfree memory, and then mark it more intelligently here.
+ // We ensure that the ClassClass mark descriptor forces invocation of
+ // this procedure.
+ // Correctness of this is subtle, but it looks OK to me for now. For the incremental
+ // collector, we need to make sure that the class object is written whenever
+ // any of the subobjects are altered and may need rescanning. This may be tricky
+ // during construction, and this may not be the right way to do this with
+ // incremental collection.
+ // If we overflow the mark stack, we will rescan the class object, so we should
+ // be OK. The same applies if we redo the mark phase because win32 unmapped part
+ // of our root set. - HB
+ jclass c = (jclass) addr;
+
+ p = (GC_PTR) c->name;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+ p = (GC_PTR) c->superclass;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+
+ p = (GC_PTR) c->constants.tags;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+ p = (GC_PTR) c->constants.data;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+
+ // If the class is an array, then the methods field holds a
+ // pointer to the element class. If the class is primitive,
+ // then the methods field holds a pointer to the array class.
+ p = (GC_PTR) c->methods;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+
+ p = (GC_PTR) c->fields;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+
+ // The vtable might be allocated even for compiled code.
+ p = (GC_PTR) c->vtable;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+
+ p = (GC_PTR) c->interfaces;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+ p = (GC_PTR) c->loader;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+
+ // The dispatch tables can be allocated at runtime.
+ p = (GC_PTR) c->ancestors;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+
+ p = (GC_PTR) c->idt;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+
+ p = (GC_PTR) c->arrayclass;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+ p = (GC_PTR) c->protectionDomain;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+ p = (GC_PTR) c->hack_signers;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+ p = (GC_PTR) c->aux_info;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+
+ p = (GC_PTR) c->reflection_data;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+
+ // The class chain must be marked for runtime-allocated Classes
+ // loaded by the bootstrap ClassLoader.
+ p = (GC_PTR) c->next_or_version;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, c);
+ }
+ else
+ {
+ // NOTE: each class only holds information about the class
+ // itself. So we must do the marking for the entire inheritance
+ // tree in order to mark all fields. FIXME: what about
+ // interfaces? We skip Object here, because Object only has a
+ // sync_info, and we handled that earlier.
+ // Note: occasionally `klass' can be null. For instance, this
+ // can happen if a GC occurs between the point where an object
+ // is allocated and where the vtbl slot is set.
+ while (klass && klass != &java::lang::Object::class$)
+ {
+ jfieldID field = JvGetFirstInstanceField (klass);
+ jint max = JvNumInstanceFields (klass);
+
+ for (int i = 0; i < max; ++i)
+ {
+ if (JvFieldIsRef (field))
+ {
+ jobject val = JvGetObjectField (obj, field);
+ p = (GC_PTR) val;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, obj);
+ }
+ field = field->getNextField ();
+ }
+ klass = klass->getSuperclass();
+ }
+ }
+
+ return mark_stack_ptr;
+}
+
+// This is called by the GC during the mark phase. It marks a Java
+// array (of objects). We use `void *' arguments and return, and not
+// what the Boehm GC wants, to avoid pollution in our headers.
+void *
+_Jv_MarkArray (void *addr, void *msp, void *msl, void *env)
+{
+ struct GC_ms_entry *mark_stack_ptr = (struct GC_ms_entry *)msp;
+ struct GC_ms_entry *mark_stack_limit = (struct GC_ms_entry *)msl;
+
+ if (env == (void *)1) /* Object allocated with debug allocator. */
+ addr = (void *)GC_USR_PTR_FROM_BASE(addr);
+ jobjectArray array = (jobjectArray) addr;
+
+ _Jv_VTable *dt = *(_Jv_VTable **) addr;
+ // Assumes size >= 3 words. That's currently true since arrays have
+ // a vtable, sync pointer, and size. If the sync pointer goes away,
+ // we may need to round up the size.
+ if (__builtin_expect (! dt || !(dt -> get_finalizer()), false))
+ return mark_stack_ptr;
+ GC_PTR p;
+
+ p = (GC_PTR) dt;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, array);
+
+# ifndef JV_HASH_SYNCHRONIZATION
+ // Every object has a sync_info pointer.
+ p = (GC_PTR) array->sync_info;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, array);
+# endif
+
+ for (int i = 0; i < JvGetArrayLength (array); ++i)
+ {
+ jobject obj = elements (array)[i];
+ p = (GC_PTR) obj;
+ MAYBE_MARK (p, mark_stack_ptr, mark_stack_limit, array);
+ }
+
+ return mark_stack_ptr;
+}
+
+// Generate a GC marking descriptor for a class.
+//
+// We assume that the gcj mark proc has index 0. This is a dubious assumption,
+// since another one could be registered first. But the compiler also
+// knows this, so in that case everything else will break, too.
+#define GCJ_DEFAULT_DESCR GC_MAKE_PROC(GC_GCJ_RESERVED_MARK_PROC_INDEX,0)
+
+void *
+_Jv_BuildGCDescr(jclass self)
+{
+ jlong desc = 0;
+ jint bits_per_word = CHAR_BIT * sizeof (void *);
+
+ // Note: for now we only consider a bitmap mark descriptor. We
+ // could also handle the case where the first N fields of a type are
+ // references. However, this is not very likely to be used by many
+ // classes, and it is easier to compute things this way.
+
+ // The vtable pointer.
+ desc |= 1ULL << (bits_per_word - 1);
+#ifndef JV_HASH_SYNCHRONIZATION
+ // The sync_info field.
+ desc |= 1ULL << (bits_per_word - 2);
+#endif
+
+ for (jclass klass = self; klass != NULL; klass = klass->getSuperclass())
+ {
+ jfieldID field = JvGetFirstInstanceField(klass);
+ int count = JvNumInstanceFields(klass);
+
+ for (int i = 0; i < count; ++i)
+ {
+ if (field->isRef())
+ {
+ unsigned int off = field->getOffset();
+ // If we run into a weird situation, we bail.
+ if (off % sizeof (void *) != 0)
+ return (void *) (GCJ_DEFAULT_DESCR);
+ off /= sizeof (void *);
+ // If we find a field outside the range of our bitmap,
+ // fall back to procedure marker. The bottom 2 bits are
+ // reserved.
+ if (off >= (unsigned) bits_per_word - 2)
+ return (void *) (GCJ_DEFAULT_DESCR);
+ desc |= 1ULL << (bits_per_word - off - 1);
+ }
+
+ field = field->getNextField();
+ }
+ }
+
+ // For bitmap mark type, bottom bits are 01.
+ desc |= 1;
+ // Bogus warning avoidance (on many platforms).
+ return (void *) (unsigned long) desc;
+}
+
+// Allocate some space that is known to be pointer-free.
+void *
+_Jv_AllocBytes (jsize size)
+{
+ void *r = GC_MALLOC_ATOMIC (size);
+ // We have to explicitly zero memory here, as the GC doesn't
+ // guarantee that PTRFREE allocations are zeroed. Note that we
+ // don't have to do this for other allocation types because we set
+ // the `ok_init' flag in the type descriptor.
+ memset (r, 0, size);
+ return r;
+}
+
+#ifdef LIBGCJ_GC_DEBUG
+
+void *
+_Jv_AllocObj (jsize size, jclass klass)
+{
+ return GC_GCJ_MALLOC (size, klass->vtable);
+}
+
+void *
+_Jv_AllocPtrFreeObj (jsize size, jclass klass)
+{
+#ifdef JV_HASH_SYNCHRONIZATION
+ void * obj = GC_MALLOC_ATOMIC(size);
+ *((_Jv_VTable **) obj) = klass->vtable;
+#else
+ void * obj = GC_GCJ_MALLOC(size, klass->vtable);
+#endif
+ return obj;
+}
+
+#endif /* LIBGCJ_GC_DEBUG */
+// In the non-debug case, the above two functions are defined
+// as inline functions in boehm-gc.h. In the debug case we
+// really want to take advantage of the definitions in gc_gcj.h.
+
+// Allocate space for a new Java array.
+// Used only for arrays of objects.
+void *
+_Jv_AllocArray (jsize size, jclass klass)
+{
+ void *obj;
+
+#ifdef LIBGCJ_GC_DEBUG
+ // There isn't much to lose by scanning this conservatively.
+ // If we didn't, the mark proc would have to understand that
+ // it needed to skip the header.
+ obj = GC_MALLOC(size);
+#else
+ const jsize min_heap_addr = 16*1024;
+ // A heuristic. If size is less than this value, the size
+ // stored in the array can't possibly be misinterpreted as
+ // a pointer. Thus we lose nothing by scanning the object
+ // completely conservatively, since no misidentification can
+ // take place.
+
+ if (size < min_heap_addr)
+ obj = GC_MALLOC(size);
+ else
+ obj = GC_generic_malloc (size, array_kind_x);
+#endif
+ *((_Jv_VTable **) obj) = klass->vtable;
+ return obj;
+}
+
+/* Allocate space for a new non-Java object, which does not have the usual
+ Java object header but may contain pointers to other GC'ed objects. */
+void *
+_Jv_AllocRawObj (jsize size)
+{
+ return (void *) GC_MALLOC (size ? size : 1);
+}
+
+#ifdef INTERPRETER
+typedef _Jv_ClosureList *closure_list_pointer;
+
+/* Release closures in a _Jv_ClosureList. */
+static void
+finalize_closure_list (GC_PTR obj, GC_PTR)
+{
+ _Jv_ClosureList **clpp = (_Jv_ClosureList **)obj;
+ _Jv_ClosureList::releaseClosures (clpp);
+}
+
+/* Allocate a double-indirect pointer to a _Jv_ClosureList that will
+ get garbage-collected after this double-indirect pointer becomes
+ unreachable by any other objects, including finalizable ones. */
+_Jv_ClosureList **
+_Jv_ClosureListFinalizer ()
+{
+ _Jv_ClosureList **clpp;
+ clpp = (_Jv_ClosureList **)_Jv_AllocBytes (sizeof (*clpp));
+ GC_REGISTER_FINALIZER_UNREACHABLE (clpp, finalize_closure_list,
+ NULL, NULL, NULL);
+ return clpp;
+}
+#endif // INTERPRETER
+
+static void
+call_finalizer (GC_PTR obj, GC_PTR client_data)
+{
+ _Jv_FinalizerFunc *fn = (_Jv_FinalizerFunc *) client_data;
+ jobject jobj = (jobject) obj;
+
+ (*fn) (jobj);
+}
+
+void
+_Jv_RegisterFinalizer (void *object, _Jv_FinalizerFunc *meth)
+{
+ GC_REGISTER_FINALIZER_NO_ORDER (object, call_finalizer, (GC_PTR) meth,
+ NULL, NULL);
+}
+
+void
+_Jv_RunFinalizers (void)
+{
+ GC_invoke_finalizers ();
+}
+
+void
+_Jv_RunAllFinalizers (void)
+{
+ GC_finalize_all ();
+}
+
+void
+_Jv_RunGC (void)
+{
+ GC_gcollect ();
+}
+
+long
+_Jv_GCTotalMemory (void)
+{
+ return GC_get_heap_size ();
+}
+
+long
+_Jv_GCFreeMemory (void)
+{
+ return GC_get_free_bytes ();
+}
+
+void
+_Jv_GCSetInitialHeapSize (size_t size)
+{
+ size_t current = GC_get_heap_size ();
+ if (size > current)
+ GC_expand_hp (size - current);
+}
+
+void
+_Jv_GCSetMaximumHeapSize (size_t size)
+{
+ GC_set_max_heap_size ((GC_word) size);
+}
+
+int
+_Jv_SetGCFreeSpaceDivisor (int div)
+{
+ return (int)GC_set_free_space_divisor ((GC_word)div);
+}
+
+void
+_Jv_DisableGC (void)
+{
+ GC_disable();
+}
+
+void
+_Jv_EnableGC (void)
+{
+ GC_enable();
+}
+
+static void * handle_out_of_memory(size_t)
+{
+ _Jv_ThrowNoMemory();
+}
+
+static void
+gcj_describe_type_fn(void *obj, char *out_buf)
+{
+ _Jv_VTable *dt = *(_Jv_VTable **) obj;
+
+ if (! dt /* Shouldn't happen */)
+ {
+ strcpy(out_buf, "GCJ (bad)");
+ return;
+ }
+ jclass klass = dt->clas;
+ if (!klass /* shouldn't happen */)
+ {
+ strcpy(out_buf, "GCJ (bad)");
+ return;
+ }
+ jstring name = klass -> getName();
+ size_t len = name -> length();
+ if (len >= GC_TYPE_DESCR_LEN) len = GC_TYPE_DESCR_LEN - 1;
+ JvGetStringUTFRegion (name, 0, len, out_buf);
+ out_buf[len] = '\0';
+}
+
+void
+_Jv_InitGC (void)
+{
+ int proc;
+ static bool gc_initialized;
+
+ if (gc_initialized)
+ return;
+
+ gc_initialized = 1;
+
+ // Ignore pointers that do not point to the start of an object.
+ GC_all_interior_pointers = 0;
+
+#if defined (HAVE_DLFCN_H) && defined (HAVE_DLADDR)
+ // Tell the collector to ask us before scanning DSOs.
+ GC_register_has_static_roots_callback (_Jv_GC_has_static_roots);
+#endif
+
+ // Configure the collector to use the bitmap marking descriptors that we
+ // stash in the class vtable.
+ // We always use mark proc descriptor 0, since the compiler knows
+ // about it.
+ GC_init_gcj_malloc (0, (void *) _Jv_MarkObj);
+
+ // Cause an out of memory error to be thrown from the allocators,
+ // instead of returning 0. This is cheaper than checking on allocation.
+ GC_oom_fn = handle_out_of_memory;
+
+ GC_java_finalization = 1;
+
+ // We use a different mark procedure for object arrays. This code
+ // configures a different object `kind' for object array allocation and
+ // marking.
+ array_free_list = GC_new_free_list();
+ proc = GC_new_proc((GC_mark_proc)_Jv_MarkArray);
+ array_kind_x = GC_new_kind(array_free_list, GC_MAKE_PROC (proc, 0), 0, 1);
+
+ // Arrange to have the GC print Java class names in backtraces, etc.
+ GC_register_describe_type_fn(GC_gcj_kind, gcj_describe_type_fn);
+ GC_register_describe_type_fn(GC_gcj_debug_kind, gcj_describe_type_fn);
+}
+
+#ifdef JV_HASH_SYNCHRONIZATION
+// Allocate an object with a fake vtable pointer, which causes only
+// the first field (beyond the fake vtable pointer) to be traced.
+// Eventually this should probably be generalized.
+
+static _Jv_VTable trace_one_vtable = {
+ 0, // class pointer
+ (void *)(2 * sizeof(void *)),
+ // descriptor; scan 2 words incl. vtable ptr.
+ // Least significant bits must be zero to
+ // identify this as a length descriptor
+ {0} // First method
+};
+
+void *
+_Jv_AllocTraceOne (jsize size /* includes vtable slot */)
+{
+ return GC_GCJ_MALLOC (size, &trace_one_vtable);
+}
+
+// Ditto for two words.
+// the first field (beyond the fake vtable pointer) to be traced.
+// Eventually this should probably be generalized.
+
+static _Jv_VTable trace_two_vtable =
+{
+ 0, // class pointer
+ (void *)(3 * sizeof(void *)),
+ // descriptor; scan 3 words incl. vtable ptr.
+ {0} // First method
+};
+
+void *
+_Jv_AllocTraceTwo (jsize size /* includes vtable slot */)
+{
+ return GC_GCJ_MALLOC (size, &trace_two_vtable);
+}
+
+#endif /* JV_HASH_SYNCHRONIZATION */
+
+void
+_Jv_GCInitializeFinalizers (void (*notifier) (void))
+{
+ GC_finalize_on_demand = 1;
+ GC_finalizer_notifier = notifier;
+}
+
+void
+_Jv_GCRegisterDisappearingLink (jobject *objp)
+{
+ // This test helps to ensure that we meet a precondition of
+ // GC_general_register_disappearing_link, viz. "Obj must be a
+ // pointer to the first word of an object we allocated."
+ if (GC_base(*objp))
+ GC_general_register_disappearing_link ((GC_PTR *) objp, (GC_PTR) *objp);
+}
+
+jboolean
+_Jv_GCCanReclaimSoftReference (jobject)
+{
+ // For now, always reclaim soft references. FIXME.
+ return true;
+}
+
+
+
+#if defined (HAVE_DLFCN_H) && defined (HAVE_DLADDR)
+
+// We keep a store of the filenames of DSOs that need to be
+// conservatively scanned by the garbage collector. During collection
+// the gc calls _Jv_GC_has_static_roots() to see if the data segment
+// of a DSO should be scanned.
+typedef struct filename_node
+{
+ char *name;
+ struct filename_node *link;
+} filename_node;
+
+#define FILENAME_STORE_SIZE 17
+static filename_node *filename_store[FILENAME_STORE_SIZE];
+
+// Find a filename in filename_store.
+static filename_node **
+find_file (const char *filename)
+{
+ int index = strlen (filename) % FILENAME_STORE_SIZE;
+ filename_node **node = &filename_store[index];
+
+ while (*node)
+ {
+ if (strcmp ((*node)->name, filename) == 0)
+ return node;
+ node = &(*node)->link;
+ }
+
+ return node;
+}
+
+// Print the store of filenames of DSOs that need collection.
+void
+_Jv_print_gc_store (void)
+{
+ for (int i = 0; i < FILENAME_STORE_SIZE; i++)
+ {
+ filename_node *node = filename_store[i];
+ while (node)
+ {
+ fprintf (stderr, "%s\n", node->name);
+ node = node->link;
+ }
+ }
+}
+
+// Create a new node in the store of libraries to collect.
+static filename_node *
+new_node (const char *filename)
+{
+ filename_node *node = (filename_node*)_Jv_Malloc (sizeof (filename_node));
+ node->name = (char *)_Jv_Malloc (strlen (filename) + 1);
+ node->link = NULL;
+ strcpy (node->name, filename);
+
+ return node;
+}
+
+// Nonzero if the gc should scan this lib.
+static int
+_Jv_GC_has_static_roots (const char *filename, void *, size_t)
+{
+ if (filename == NULL || strlen (filename) == 0)
+ // No filename; better safe than sorry.
+ return 1;
+
+ filename_node **node = find_file (filename);
+ if (*node)
+ return 1;
+
+ return 0;
+}
+
+#endif
+
+// Register the DSO that contains p for collection.
+void
+_Jv_RegisterLibForGc (const void *p __attribute__ ((__unused__)))
+{
+#if defined (HAVE_DLFCN_H) && defined (HAVE_DLADDR)
+ Dl_info info;
+
+ if (dladdr (const_cast<void *>(p), &info) != 0)
+ {
+ filename_node **node = find_file (info.dli_fname);
+ if (! *node)
+ *node = new_node (info.dli_fname);
+ }
+#endif
+}
+
+void
+_Jv_SuspendThread (_Jv_Thread_t *thread)
+{
+#if defined(GC_PTHREADS) && !defined(GC_SOLARIS_THREADS) \
+ && !defined(GC_WIN32_THREADS) && !defined(GC_DARWIN_THREADS)
+ GC_suspend_thread (_Jv_GetPlatformThreadID (thread));
+#endif
+}
+
+void
+_Jv_ResumeThread (_Jv_Thread_t *thread)
+{
+#if defined(GC_PTHREADS) && !defined(GC_SOLARIS_THREADS) \
+ && !defined(GC_WIN32_THREADS) && !defined(GC_DARWIN_THREADS)
+ GC_resume_thread (_Jv_GetPlatformThreadID (thread));
+#endif
+}
+
+int
+_Jv_IsThreadSuspended (_Jv_Thread_t *thread)
+{
+#if defined(GC_PTHREADS) && !defined(GC_SOLARIS_THREADS) \
+ && !defined(GC_WIN32_THREADS) && !defined(GC_DARWIN_THREADS)
+ return GC_is_thread_suspended (_Jv_GetPlatformThreadID (thread));
+#else
+ return 0;
+#endif
+}
+
+void
+_Jv_GCAttachThread ()
+{
+ // The registration interface is only defined on posixy systems and
+ // only actually works if pthread_getattr_np is defined.
+ // FIXME: until gc7 it is simpler to disable this on solaris.
+#if defined(HAVE_PTHREAD_GETATTR_NP) && !defined(GC_SOLARIS_THREADS)
+ GC_register_my_thread ();
+#endif
+}
+
+void
+_Jv_GCDetachThread ()
+{
+#if defined(HAVE_PTHREAD_GETATTR_NP) && !defined(GC_SOLARIS_THREADS)
+ GC_unregister_my_thread ();
+#endif
+}