| 1 | /*
|
|---|
| 2 | * Copyright 1988, 1989 Hans-J. Boehm, Alan J. Demers
|
|---|
| 3 | * Copyright (c) 1991-1994 by Xerox Corporation. All rights reserved.
|
|---|
| 4 | * Copyright (c) 1996 by Silicon Graphics. All rights reserved.
|
|---|
| 5 | * Copyright (c) 2000 by Hewlett-Packard Company. All rights reserved.
|
|---|
| 6 | *
|
|---|
| 7 | * THIS MATERIAL IS PROVIDED AS IS, WITH ABSOLUTELY NO WARRANTY EXPRESSED
|
|---|
| 8 | * OR IMPLIED. ANY USE IS AT YOUR OWN RISK.
|
|---|
| 9 | *
|
|---|
| 10 | * Permission is hereby granted to use or copy this program
|
|---|
| 11 | * for any purpose, provided the above notices are retained on all copies.
|
|---|
| 12 | * Permission to modify the code and to distribute modified code is granted,
|
|---|
| 13 | * provided the above notices are retained, and a notice that the code was
|
|---|
| 14 | * modified is included with the above copyright notice.
|
|---|
| 15 | */
|
|---|
| 16 |
|
|---|
| 17 | /*
|
|---|
| 18 | * These are extra allocation routines which are likely to be less
|
|---|
| 19 | * frequently used than those in malloc.c. They are separate in the
|
|---|
| 20 | * hope that the .o file will be excluded from statically linked
|
|---|
| 21 | * executables. We should probably break this up further.
|
|---|
| 22 | */
|
|---|
| 23 |
|
|---|
| 24 | #include <stdio.h>
|
|---|
| 25 | #include "private/gc_priv.h"
|
|---|
| 26 |
|
|---|
| 27 | extern ptr_t GC_clear_stack(); /* in misc.c, behaves like identity */
|
|---|
| 28 | void GC_extend_size_map(); /* in misc.c. */
|
|---|
| 29 | GC_bool GC_alloc_reclaim_list(); /* in malloc.c */
|
|---|
| 30 |
|
|---|
| 31 | /* Some externally visible but unadvertised variables to allow access to */
|
|---|
| 32 | /* free lists from inlined allocators without including gc_priv.h */
|
|---|
| 33 | /* or introducing dependencies on internal data structure layouts. */
|
|---|
| 34 | ptr_t * GC_CONST GC_objfreelist_ptr = GC_objfreelist;
|
|---|
| 35 | ptr_t * GC_CONST GC_aobjfreelist_ptr = GC_aobjfreelist;
|
|---|
| 36 | ptr_t * GC_CONST GC_uobjfreelist_ptr = GC_uobjfreelist;
|
|---|
| 37 | # ifdef ATOMIC_UNCOLLECTABLE
|
|---|
| 38 | ptr_t * GC_CONST GC_auobjfreelist_ptr = GC_auobjfreelist;
|
|---|
| 39 | # endif
|
|---|
| 40 |
|
|---|
| 41 |
|
|---|
| 42 | GC_PTR GC_generic_or_special_malloc(lb,knd)
|
|---|
| 43 | word lb;
|
|---|
| 44 | int knd;
|
|---|
| 45 | {
|
|---|
| 46 | switch(knd) {
|
|---|
| 47 | # ifdef STUBBORN_ALLOC
|
|---|
| 48 | case STUBBORN:
|
|---|
| 49 | return(GC_malloc_stubborn((size_t)lb));
|
|---|
| 50 | # endif
|
|---|
| 51 | case PTRFREE:
|
|---|
| 52 | return(GC_malloc_atomic((size_t)lb));
|
|---|
| 53 | case NORMAL:
|
|---|
| 54 | return(GC_malloc((size_t)lb));
|
|---|
| 55 | case UNCOLLECTABLE:
|
|---|
| 56 | return(GC_malloc_uncollectable((size_t)lb));
|
|---|
| 57 | # ifdef ATOMIC_UNCOLLECTABLE
|
|---|
| 58 | case AUNCOLLECTABLE:
|
|---|
| 59 | return(GC_malloc_atomic_uncollectable((size_t)lb));
|
|---|
| 60 | # endif /* ATOMIC_UNCOLLECTABLE */
|
|---|
| 61 | default:
|
|---|
| 62 | return(GC_generic_malloc(lb,knd));
|
|---|
| 63 | }
|
|---|
| 64 | }
|
|---|
| 65 |
|
|---|
| 66 |
|
|---|
| 67 | /* Change the size of the block pointed to by p to contain at least */
|
|---|
| 68 | /* lb bytes. The object may be (and quite likely will be) moved. */
|
|---|
| 69 | /* The kind (e.g. atomic) is the same as that of the old. */
|
|---|
| 70 | /* Shrinking of large blocks is not implemented well. */
|
|---|
| 71 | # ifdef __STDC__
|
|---|
| 72 | GC_PTR GC_realloc(GC_PTR p, size_t lb)
|
|---|
| 73 | # else
|
|---|
| 74 | GC_PTR GC_realloc(p,lb)
|
|---|
| 75 | GC_PTR p;
|
|---|
| 76 | size_t lb;
|
|---|
| 77 | # endif
|
|---|
| 78 | {
|
|---|
| 79 | register struct hblk * h;
|
|---|
| 80 | register hdr * hhdr;
|
|---|
| 81 | register word sz; /* Current size in bytes */
|
|---|
| 82 | register word orig_sz; /* Original sz in bytes */
|
|---|
| 83 | int obj_kind;
|
|---|
| 84 |
|
|---|
| 85 | if (p == 0) return(GC_malloc(lb)); /* Required by ANSI */
|
|---|
| 86 | h = HBLKPTR(p);
|
|---|
| 87 | hhdr = HDR(h);
|
|---|
| 88 | sz = hhdr -> hb_sz;
|
|---|
| 89 | obj_kind = hhdr -> hb_obj_kind;
|
|---|
| 90 | sz = WORDS_TO_BYTES(sz);
|
|---|
| 91 | orig_sz = sz;
|
|---|
| 92 |
|
|---|
| 93 | if (sz > MAXOBJBYTES) {
|
|---|
| 94 | /* Round it up to the next whole heap block */
|
|---|
| 95 | register word descr;
|
|---|
| 96 |
|
|---|
| 97 | sz = (sz+HBLKSIZE-1) & (~HBLKMASK);
|
|---|
| 98 | hhdr -> hb_sz = BYTES_TO_WORDS(sz);
|
|---|
| 99 | descr = GC_obj_kinds[obj_kind].ok_descriptor;
|
|---|
| 100 | if (GC_obj_kinds[obj_kind].ok_relocate_descr) descr += sz;
|
|---|
| 101 | hhdr -> hb_descr = descr;
|
|---|
| 102 | if (IS_UNCOLLECTABLE(obj_kind)) GC_non_gc_bytes += (sz - orig_sz);
|
|---|
| 103 | /* Extra area is already cleared by GC_alloc_large_and_clear. */
|
|---|
| 104 | }
|
|---|
| 105 | if (ADD_SLOP(lb) <= sz) {
|
|---|
| 106 | if (lb >= (sz >> 1)) {
|
|---|
| 107 | # ifdef STUBBORN_ALLOC
|
|---|
| 108 | if (obj_kind == STUBBORN) GC_change_stubborn(p);
|
|---|
| 109 | # endif
|
|---|
| 110 | if (orig_sz > lb) {
|
|---|
| 111 | /* Clear unneeded part of object to avoid bogus pointer */
|
|---|
| 112 | /* tracing. */
|
|---|
| 113 | /* Safe for stubborn objects. */
|
|---|
| 114 | BZERO(((ptr_t)p) + lb, orig_sz - lb);
|
|---|
| 115 | }
|
|---|
| 116 | return(p);
|
|---|
| 117 | } else {
|
|---|
| 118 | /* shrink */
|
|---|
| 119 | GC_PTR result =
|
|---|
| 120 | GC_generic_or_special_malloc((word)lb, obj_kind);
|
|---|
| 121 |
|
|---|
| 122 | if (result == 0) return(0);
|
|---|
| 123 | /* Could also return original object. But this */
|
|---|
| 124 | /* gives the client warning of imminent disaster. */
|
|---|
| 125 | BCOPY(p, result, lb);
|
|---|
| 126 | # ifndef IGNORE_FREE
|
|---|
| 127 | GC_free(p);
|
|---|
| 128 | # endif
|
|---|
| 129 | return(result);
|
|---|
| 130 | }
|
|---|
| 131 | } else {
|
|---|
| 132 | /* grow */
|
|---|
| 133 | GC_PTR result =
|
|---|
| 134 | GC_generic_or_special_malloc((word)lb, obj_kind);
|
|---|
| 135 |
|
|---|
| 136 | if (result == 0) return(0);
|
|---|
| 137 | BCOPY(p, result, sz);
|
|---|
| 138 | # ifndef IGNORE_FREE
|
|---|
| 139 | GC_free(p);
|
|---|
| 140 | # endif
|
|---|
| 141 | return(result);
|
|---|
| 142 | }
|
|---|
| 143 | }
|
|---|
| 144 |
|
|---|
| 145 | # if defined(REDIRECT_MALLOC) || defined(REDIRECT_REALLOC)
|
|---|
| 146 | # ifdef __STDC__
|
|---|
| 147 | GC_PTR realloc(GC_PTR p, size_t lb)
|
|---|
| 148 | # else
|
|---|
| 149 | GC_PTR realloc(p,lb)
|
|---|
| 150 | GC_PTR p;
|
|---|
| 151 | size_t lb;
|
|---|
| 152 | # endif
|
|---|
| 153 | {
|
|---|
| 154 | # ifdef REDIRECT_REALLOC
|
|---|
| 155 | return(REDIRECT_REALLOC(p, lb));
|
|---|
| 156 | # else
|
|---|
| 157 | return(GC_realloc(p, lb));
|
|---|
| 158 | # endif
|
|---|
| 159 | }
|
|---|
| 160 | # endif /* REDIRECT_MALLOC */
|
|---|
| 161 |
|
|---|
| 162 |
|
|---|
| 163 | /* The same thing, except caller does not hold allocation lock. */
|
|---|
| 164 | /* We avoid holding allocation lock while we clear memory. */
|
|---|
| 165 | ptr_t GC_generic_malloc_ignore_off_page(lb, k)
|
|---|
| 166 | register size_t lb;
|
|---|
| 167 | register int k;
|
|---|
| 168 | {
|
|---|
| 169 | register ptr_t result;
|
|---|
| 170 | word lw;
|
|---|
| 171 | word n_blocks;
|
|---|
| 172 | GC_bool init;
|
|---|
| 173 | DCL_LOCK_STATE;
|
|---|
| 174 |
|
|---|
| 175 | if (SMALL_OBJ(lb))
|
|---|
| 176 | return(GC_generic_malloc((word)lb, k));
|
|---|
| 177 | lw = ROUNDED_UP_WORDS(lb);
|
|---|
| 178 | n_blocks = OBJ_SZ_TO_BLOCKS(lw);
|
|---|
| 179 | init = GC_obj_kinds[k].ok_init;
|
|---|
| 180 | GC_INVOKE_FINALIZERS();
|
|---|
| 181 | DISABLE_SIGNALS();
|
|---|
| 182 | LOCK();
|
|---|
| 183 | result = (ptr_t)GC_alloc_large(lw, k, IGNORE_OFF_PAGE);
|
|---|
| 184 | if (0 != result) {
|
|---|
| 185 | if (GC_debugging_started) {
|
|---|
| 186 | BZERO(result, n_blocks * HBLKSIZE);
|
|---|
| 187 | } else {
|
|---|
| 188 | # ifdef THREADS
|
|---|
| 189 | /* Clear any memory that might be used for GC descriptors */
|
|---|
| 190 | /* before we release the lock. */
|
|---|
| 191 | ((word *)result)[0] = 0;
|
|---|
| 192 | ((word *)result)[1] = 0;
|
|---|
| 193 | ((word *)result)[lw-1] = 0;
|
|---|
| 194 | ((word *)result)[lw-2] = 0;
|
|---|
| 195 | # endif
|
|---|
| 196 | }
|
|---|
| 197 | }
|
|---|
| 198 | GC_words_allocd += lw;
|
|---|
| 199 | UNLOCK();
|
|---|
| 200 | ENABLE_SIGNALS();
|
|---|
| 201 | if (0 == result) {
|
|---|
| 202 | return((*GC_oom_fn)(lb));
|
|---|
| 203 | } else {
|
|---|
| 204 | if (init & !GC_debugging_started) {
|
|---|
| 205 | BZERO(result, n_blocks * HBLKSIZE);
|
|---|
| 206 | }
|
|---|
| 207 | return(result);
|
|---|
| 208 | }
|
|---|
| 209 | }
|
|---|
| 210 |
|
|---|
| 211 | # if defined(__STDC__) || defined(__cplusplus)
|
|---|
| 212 | void * GC_malloc_ignore_off_page(size_t lb)
|
|---|
| 213 | # else
|
|---|
| 214 | char * GC_malloc_ignore_off_page(lb)
|
|---|
| 215 | register size_t lb;
|
|---|
| 216 | # endif
|
|---|
| 217 | {
|
|---|
| 218 | return((GC_PTR)GC_generic_malloc_ignore_off_page(lb, NORMAL));
|
|---|
| 219 | }
|
|---|
| 220 |
|
|---|
| 221 | # if defined(__STDC__) || defined(__cplusplus)
|
|---|
| 222 | void * GC_malloc_atomic_ignore_off_page(size_t lb)
|
|---|
| 223 | # else
|
|---|
| 224 | char * GC_malloc_atomic_ignore_off_page(lb)
|
|---|
| 225 | register size_t lb;
|
|---|
| 226 | # endif
|
|---|
| 227 | {
|
|---|
| 228 | return((GC_PTR)GC_generic_malloc_ignore_off_page(lb, PTRFREE));
|
|---|
| 229 | }
|
|---|
| 230 |
|
|---|
| 231 | /* Increment GC_words_allocd from code that doesn't have direct access */
|
|---|
| 232 | /* to GC_arrays. */
|
|---|
| 233 | # ifdef __STDC__
|
|---|
| 234 | void GC_incr_words_allocd(size_t n)
|
|---|
| 235 | {
|
|---|
| 236 | GC_words_allocd += n;
|
|---|
| 237 | }
|
|---|
| 238 |
|
|---|
| 239 | /* The same for GC_mem_freed. */
|
|---|
| 240 | void GC_incr_mem_freed(size_t n)
|
|---|
| 241 | {
|
|---|
| 242 | GC_mem_freed += n;
|
|---|
| 243 | }
|
|---|
| 244 | # endif /* __STDC__ */
|
|---|
| 245 |
|
|---|
| 246 | /* Analogous to the above, but assumes a small object size, and */
|
|---|
| 247 | /* bypasses MERGE_SIZES mechanism. Used by gc_inline.h. */
|
|---|
| 248 | ptr_t GC_generic_malloc_words_small_inner(lw, k)
|
|---|
| 249 | register word lw;
|
|---|
| 250 | register int k;
|
|---|
| 251 | {
|
|---|
| 252 | register ptr_t op;
|
|---|
| 253 | register ptr_t *opp;
|
|---|
| 254 | register struct obj_kind * kind = GC_obj_kinds + k;
|
|---|
| 255 |
|
|---|
| 256 | opp = &(kind -> ok_freelist[lw]);
|
|---|
| 257 | if( (op = *opp) == 0 ) {
|
|---|
| 258 | if (!GC_is_initialized) {
|
|---|
| 259 | GC_init_inner();
|
|---|
| 260 | }
|
|---|
| 261 | if (kind -> ok_reclaim_list != 0 || GC_alloc_reclaim_list(kind)) {
|
|---|
| 262 | op = GC_clear_stack(GC_allocobj((word)lw, k));
|
|---|
| 263 | }
|
|---|
| 264 | if (op == 0) {
|
|---|
| 265 | UNLOCK();
|
|---|
| 266 | ENABLE_SIGNALS();
|
|---|
| 267 | return ((*GC_oom_fn)(WORDS_TO_BYTES(lw)));
|
|---|
| 268 | }
|
|---|
| 269 | }
|
|---|
| 270 | *opp = obj_link(op);
|
|---|
| 271 | obj_link(op) = 0;
|
|---|
| 272 | GC_words_allocd += lw;
|
|---|
| 273 | return((ptr_t)op);
|
|---|
| 274 | }
|
|---|
| 275 |
|
|---|
| 276 | /* Analogous to the above, but assumes a small object size, and */
|
|---|
| 277 | /* bypasses MERGE_SIZES mechanism. Used by gc_inline.h. */
|
|---|
| 278 | #ifdef __STDC__
|
|---|
| 279 | ptr_t GC_generic_malloc_words_small(size_t lw, int k)
|
|---|
| 280 | #else
|
|---|
| 281 | ptr_t GC_generic_malloc_words_small(lw, k)
|
|---|
| 282 | register word lw;
|
|---|
| 283 | register int k;
|
|---|
| 284 | #endif
|
|---|
| 285 | {
|
|---|
| 286 | register ptr_t op;
|
|---|
| 287 | DCL_LOCK_STATE;
|
|---|
| 288 |
|
|---|
| 289 | GC_INVOKE_FINALIZERS();
|
|---|
| 290 | DISABLE_SIGNALS();
|
|---|
| 291 | LOCK();
|
|---|
| 292 | op = GC_generic_malloc_words_small_inner(lw, k);
|
|---|
| 293 | UNLOCK();
|
|---|
| 294 | ENABLE_SIGNALS();
|
|---|
| 295 | return((ptr_t)op);
|
|---|
| 296 | }
|
|---|
| 297 |
|
|---|
| 298 | #if defined(THREADS) && !defined(SRC_M3)
|
|---|
| 299 |
|
|---|
| 300 | extern signed_word GC_mem_found; /* Protected by GC lock. */
|
|---|
| 301 |
|
|---|
| 302 | #ifdef PARALLEL_MARK
|
|---|
| 303 | volatile signed_word GC_words_allocd_tmp = 0;
|
|---|
| 304 | /* Number of words of memory allocated since */
|
|---|
| 305 | /* we released the GC lock. Instead of */
|
|---|
| 306 | /* reacquiring the GC lock just to add this in, */
|
|---|
| 307 | /* we add it in the next time we reacquire */
|
|---|
| 308 | /* the lock. (Atomically adding it doesn't */
|
|---|
| 309 | /* work, since we would have to atomically */
|
|---|
| 310 | /* update it in GC_malloc, which is too */
|
|---|
| 311 | /* expensive. */
|
|---|
| 312 | #endif /* PARALLEL_MARK */
|
|---|
| 313 |
|
|---|
| 314 | /* See reclaim.c: */
|
|---|
| 315 | extern ptr_t GC_reclaim_generic();
|
|---|
| 316 |
|
|---|
| 317 | /* Return a list of 1 or more objects of the indicated size, linked */
|
|---|
| 318 | /* through the first word in the object. This has the advantage that */
|
|---|
| 319 | /* it acquires the allocation lock only once, and may greatly reduce */
|
|---|
| 320 | /* time wasted contending for the allocation lock. Typical usage would */
|
|---|
| 321 | /* be in a thread that requires many items of the same size. It would */
|
|---|
| 322 | /* keep its own free list in thread-local storage, and call */
|
|---|
| 323 | /* GC_malloc_many or friends to replenish it. (We do not round up */
|
|---|
| 324 | /* object sizes, since a call indicates the intention to consume many */
|
|---|
| 325 | /* objects of exactly this size.) */
|
|---|
| 326 | /* We return the free-list by assigning it to *result, since it is */
|
|---|
| 327 | /* not safe to return, e.g. a linked list of pointer-free objects, */
|
|---|
| 328 | /* since the collector would not retain the entire list if it were */
|
|---|
| 329 | /* invoked just as we were returning. */
|
|---|
| 330 | /* Note that the client should usually clear the link field. */
|
|---|
| 331 | void GC_generic_malloc_many(lb, k, result)
|
|---|
| 332 | register word lb;
|
|---|
| 333 | register int k;
|
|---|
| 334 | ptr_t *result;
|
|---|
| 335 | {
|
|---|
| 336 | ptr_t op;
|
|---|
| 337 | ptr_t p;
|
|---|
| 338 | ptr_t *opp;
|
|---|
| 339 | word lw;
|
|---|
| 340 | word my_words_allocd = 0;
|
|---|
| 341 | struct obj_kind * ok = &(GC_obj_kinds[k]);
|
|---|
| 342 | DCL_LOCK_STATE;
|
|---|
| 343 |
|
|---|
| 344 | # if defined(GATHERSTATS) || defined(PARALLEL_MARK)
|
|---|
| 345 | # define COUNT_ARG , &my_words_allocd
|
|---|
| 346 | # else
|
|---|
| 347 | # define COUNT_ARG
|
|---|
| 348 | # define NEED_TO_COUNT
|
|---|
| 349 | # endif
|
|---|
| 350 | if (!SMALL_OBJ(lb)) {
|
|---|
| 351 | op = GC_generic_malloc(lb, k);
|
|---|
| 352 | if(0 != op) obj_link(op) = 0;
|
|---|
| 353 | *result = op;
|
|---|
| 354 | return;
|
|---|
| 355 | }
|
|---|
| 356 | lw = ALIGNED_WORDS(lb);
|
|---|
| 357 | GC_INVOKE_FINALIZERS();
|
|---|
| 358 | DISABLE_SIGNALS();
|
|---|
| 359 | LOCK();
|
|---|
| 360 | if (!GC_is_initialized) GC_init_inner();
|
|---|
| 361 | /* Do our share of marking work */
|
|---|
| 362 | if (GC_incremental && !GC_dont_gc) {
|
|---|
| 363 | ENTER_GC();
|
|---|
| 364 | GC_collect_a_little_inner(1);
|
|---|
| 365 | EXIT_GC();
|
|---|
| 366 | }
|
|---|
| 367 | /* First see if we can reclaim a page of objects waiting to be */
|
|---|
| 368 | /* reclaimed. */
|
|---|
| 369 | {
|
|---|
| 370 | struct hblk ** rlh = ok -> ok_reclaim_list;
|
|---|
| 371 | struct hblk * hbp;
|
|---|
| 372 | hdr * hhdr;
|
|---|
| 373 |
|
|---|
| 374 | rlh += lw;
|
|---|
| 375 | while ((hbp = *rlh) != 0) {
|
|---|
| 376 | hhdr = HDR(hbp);
|
|---|
| 377 | *rlh = hhdr -> hb_next;
|
|---|
| 378 | # ifdef PARALLEL_MARK
|
|---|
| 379 | {
|
|---|
| 380 | signed_word my_words_allocd_tmp = GC_words_allocd_tmp;
|
|---|
| 381 |
|
|---|
| 382 | GC_ASSERT(my_words_allocd_tmp >= 0);
|
|---|
| 383 | /* We only decrement it while holding the GC lock. */
|
|---|
| 384 | /* Thus we can't accidentally adjust it down in more */
|
|---|
| 385 | /* than one thread simultaneously. */
|
|---|
| 386 | if (my_words_allocd_tmp != 0) {
|
|---|
| 387 | (void)GC_atomic_add(
|
|---|
| 388 | (volatile GC_word *)(&GC_words_allocd_tmp),
|
|---|
| 389 | (GC_word)(-my_words_allocd_tmp));
|
|---|
| 390 | GC_words_allocd += my_words_allocd_tmp;
|
|---|
| 391 | }
|
|---|
| 392 | }
|
|---|
| 393 | GC_acquire_mark_lock();
|
|---|
| 394 | ++ GC_fl_builder_count;
|
|---|
| 395 | UNLOCK();
|
|---|
| 396 | ENABLE_SIGNALS();
|
|---|
| 397 | GC_release_mark_lock();
|
|---|
| 398 | # endif
|
|---|
| 399 | op = GC_reclaim_generic(hbp, hhdr, lw,
|
|---|
| 400 | ok -> ok_init, 0 COUNT_ARG);
|
|---|
| 401 | if (op != 0) {
|
|---|
| 402 | # ifdef NEED_TO_COUNT
|
|---|
| 403 | /* We are neither gathering statistics, nor marking in */
|
|---|
| 404 | /* parallel. Thus GC_reclaim_generic doesn't count */
|
|---|
| 405 | /* for us. */
|
|---|
| 406 | for (p = op; p != 0; p = obj_link(p)) {
|
|---|
| 407 | my_words_allocd += lw;
|
|---|
| 408 | }
|
|---|
| 409 | # endif
|
|---|
| 410 | # if defined(GATHERSTATS)
|
|---|
| 411 | /* We also reclaimed memory, so we need to adjust */
|
|---|
| 412 | /* that count. */
|
|---|
| 413 | /* This should be atomic, so the results may be */
|
|---|
| 414 | /* inaccurate. */
|
|---|
| 415 | GC_mem_found += my_words_allocd;
|
|---|
| 416 | # endif
|
|---|
| 417 | # ifdef PARALLEL_MARK
|
|---|
| 418 | *result = op;
|
|---|
| 419 | (void)GC_atomic_add(
|
|---|
| 420 | (volatile GC_word *)(&GC_words_allocd_tmp),
|
|---|
| 421 | (GC_word)(my_words_allocd));
|
|---|
| 422 | GC_acquire_mark_lock();
|
|---|
| 423 | -- GC_fl_builder_count;
|
|---|
| 424 | if (GC_fl_builder_count == 0) GC_notify_all_builder();
|
|---|
| 425 | GC_release_mark_lock();
|
|---|
| 426 | (void) GC_clear_stack(0);
|
|---|
| 427 | return;
|
|---|
| 428 | # else
|
|---|
| 429 | GC_words_allocd += my_words_allocd;
|
|---|
| 430 | goto out;
|
|---|
| 431 | # endif
|
|---|
| 432 | }
|
|---|
| 433 | # ifdef PARALLEL_MARK
|
|---|
| 434 | GC_acquire_mark_lock();
|
|---|
| 435 | -- GC_fl_builder_count;
|
|---|
| 436 | if (GC_fl_builder_count == 0) GC_notify_all_builder();
|
|---|
| 437 | GC_release_mark_lock();
|
|---|
| 438 | DISABLE_SIGNALS();
|
|---|
| 439 | LOCK();
|
|---|
| 440 | /* GC lock is needed for reclaim list access. We */
|
|---|
| 441 | /* must decrement fl_builder_count before reaquiring GC */
|
|---|
| 442 | /* lock. Hopefully this path is rare. */
|
|---|
| 443 | # endif
|
|---|
| 444 | }
|
|---|
| 445 | }
|
|---|
| 446 | /* Next try to use prefix of global free list if there is one. */
|
|---|
| 447 | /* We don't refill it, but we need to use it up before allocating */
|
|---|
| 448 | /* a new block ourselves. */
|
|---|
| 449 | opp = &(GC_obj_kinds[k].ok_freelist[lw]);
|
|---|
| 450 | if ( (op = *opp) != 0 ) {
|
|---|
| 451 | *opp = 0;
|
|---|
| 452 | my_words_allocd = 0;
|
|---|
| 453 | for (p = op; p != 0; p = obj_link(p)) {
|
|---|
| 454 | my_words_allocd += lw;
|
|---|
| 455 | if (my_words_allocd >= BODY_SZ) {
|
|---|
| 456 | *opp = obj_link(p);
|
|---|
| 457 | obj_link(p) = 0;
|
|---|
| 458 | break;
|
|---|
| 459 | }
|
|---|
| 460 | }
|
|---|
| 461 | GC_words_allocd += my_words_allocd;
|
|---|
| 462 | goto out;
|
|---|
| 463 | }
|
|---|
| 464 | /* Next try to allocate a new block worth of objects of this size. */
|
|---|
| 465 | {
|
|---|
| 466 | struct hblk *h = GC_allochblk(lw, k, 0);
|
|---|
| 467 | if (h != 0) {
|
|---|
| 468 | if (IS_UNCOLLECTABLE(k)) GC_set_hdr_marks(HDR(h));
|
|---|
| 469 | GC_words_allocd += BYTES_TO_WORDS(HBLKSIZE)
|
|---|
| 470 | - BYTES_TO_WORDS(HBLKSIZE) % lw;
|
|---|
| 471 | # ifdef PARALLEL_MARK
|
|---|
| 472 | GC_acquire_mark_lock();
|
|---|
| 473 | ++ GC_fl_builder_count;
|
|---|
| 474 | UNLOCK();
|
|---|
| 475 | ENABLE_SIGNALS();
|
|---|
| 476 | GC_release_mark_lock();
|
|---|
| 477 | # endif
|
|---|
| 478 |
|
|---|
| 479 | op = GC_build_fl(h, lw, ok -> ok_init, 0);
|
|---|
| 480 | # ifdef PARALLEL_MARK
|
|---|
| 481 | *result = op;
|
|---|
| 482 | GC_acquire_mark_lock();
|
|---|
| 483 | -- GC_fl_builder_count;
|
|---|
| 484 | if (GC_fl_builder_count == 0) GC_notify_all_builder();
|
|---|
| 485 | GC_release_mark_lock();
|
|---|
| 486 | (void) GC_clear_stack(0);
|
|---|
| 487 | return;
|
|---|
| 488 | # else
|
|---|
| 489 | goto out;
|
|---|
| 490 | # endif
|
|---|
| 491 | }
|
|---|
| 492 | }
|
|---|
| 493 |
|
|---|
| 494 | /* As a last attempt, try allocating a single object. Note that */
|
|---|
| 495 | /* this may trigger a collection or expand the heap. */
|
|---|
| 496 | op = GC_generic_malloc_inner(lb, k);
|
|---|
| 497 | if (0 != op) obj_link(op) = 0;
|
|---|
| 498 |
|
|---|
| 499 | out:
|
|---|
| 500 | *result = op;
|
|---|
| 501 | UNLOCK();
|
|---|
| 502 | ENABLE_SIGNALS();
|
|---|
| 503 | (void) GC_clear_stack(0);
|
|---|
| 504 | }
|
|---|
| 505 |
|
|---|
| 506 | GC_PTR GC_malloc_many(size_t lb)
|
|---|
| 507 | {
|
|---|
| 508 | ptr_t result;
|
|---|
| 509 | GC_generic_malloc_many(lb, NORMAL, &result);
|
|---|
| 510 | return result;
|
|---|
| 511 | }
|
|---|
| 512 |
|
|---|
| 513 | /* Note that the "atomic" version of this would be unsafe, since the */
|
|---|
| 514 | /* links would not be seen by the collector. */
|
|---|
| 515 | # endif
|
|---|
| 516 |
|
|---|
| 517 | /* Allocate lb bytes of pointerful, traced, but not collectable data */
|
|---|
| 518 | # ifdef __STDC__
|
|---|
| 519 | GC_PTR GC_malloc_uncollectable(size_t lb)
|
|---|
| 520 | # else
|
|---|
| 521 | GC_PTR GC_malloc_uncollectable(lb)
|
|---|
| 522 | size_t lb;
|
|---|
| 523 | # endif
|
|---|
| 524 | {
|
|---|
| 525 | register ptr_t op;
|
|---|
| 526 | register ptr_t *opp;
|
|---|
| 527 | register word lw;
|
|---|
| 528 | DCL_LOCK_STATE;
|
|---|
| 529 |
|
|---|
| 530 | if( SMALL_OBJ(lb) ) {
|
|---|
| 531 | # ifdef MERGE_SIZES
|
|---|
| 532 | if (EXTRA_BYTES != 0 && lb != 0) lb--;
|
|---|
| 533 | /* We don't need the extra byte, since this won't be */
|
|---|
| 534 | /* collected anyway. */
|
|---|
| 535 | lw = GC_size_map[lb];
|
|---|
| 536 | # else
|
|---|
| 537 | lw = ALIGNED_WORDS(lb);
|
|---|
| 538 | # endif
|
|---|
| 539 | opp = &(GC_uobjfreelist[lw]);
|
|---|
| 540 | FASTLOCK();
|
|---|
| 541 | if( FASTLOCK_SUCCEEDED() && (op = *opp) != 0 ) {
|
|---|
| 542 | /* See above comment on signals. */
|
|---|
| 543 | *opp = obj_link(op);
|
|---|
| 544 | obj_link(op) = 0;
|
|---|
| 545 | GC_words_allocd += lw;
|
|---|
| 546 | /* Mark bit ws already set on free list. It will be */
|
|---|
| 547 | /* cleared only temporarily during a collection, as a */
|
|---|
| 548 | /* result of the normal free list mark bit clearing. */
|
|---|
| 549 | GC_non_gc_bytes += WORDS_TO_BYTES(lw);
|
|---|
| 550 | FASTUNLOCK();
|
|---|
| 551 | return((GC_PTR) op);
|
|---|
| 552 | }
|
|---|
| 553 | FASTUNLOCK();
|
|---|
| 554 | op = (ptr_t)GC_generic_malloc((word)lb, UNCOLLECTABLE);
|
|---|
| 555 | } else {
|
|---|
| 556 | op = (ptr_t)GC_generic_malloc((word)lb, UNCOLLECTABLE);
|
|---|
| 557 | }
|
|---|
| 558 | if (0 == op) return(0);
|
|---|
| 559 | /* We don't need the lock here, since we have an undisguised */
|
|---|
| 560 | /* pointer. We do need to hold the lock while we adjust */
|
|---|
| 561 | /* mark bits. */
|
|---|
| 562 | {
|
|---|
| 563 | register struct hblk * h;
|
|---|
| 564 |
|
|---|
| 565 | h = HBLKPTR(op);
|
|---|
| 566 | lw = HDR(h) -> hb_sz;
|
|---|
| 567 |
|
|---|
| 568 | DISABLE_SIGNALS();
|
|---|
| 569 | LOCK();
|
|---|
| 570 | GC_set_mark_bit(op);
|
|---|
| 571 | GC_non_gc_bytes += WORDS_TO_BYTES(lw);
|
|---|
| 572 | UNLOCK();
|
|---|
| 573 | ENABLE_SIGNALS();
|
|---|
| 574 | return((GC_PTR) op);
|
|---|
| 575 | }
|
|---|
| 576 | }
|
|---|
| 577 |
|
|---|
| 578 | # ifdef ATOMIC_UNCOLLECTABLE
|
|---|
| 579 | /* Allocate lb bytes of pointerfree, untraced, uncollectable data */
|
|---|
| 580 | /* This is normally roughly equivalent to the system malloc. */
|
|---|
| 581 | /* But it may be useful if malloc is redefined. */
|
|---|
| 582 | # ifdef __STDC__
|
|---|
| 583 | GC_PTR GC_malloc_atomic_uncollectable(size_t lb)
|
|---|
| 584 | # else
|
|---|
| 585 | GC_PTR GC_malloc_atomic_uncollectable(lb)
|
|---|
| 586 | size_t lb;
|
|---|
| 587 | # endif
|
|---|
| 588 | {
|
|---|
| 589 | register ptr_t op;
|
|---|
| 590 | register ptr_t *opp;
|
|---|
| 591 | register word lw;
|
|---|
| 592 | DCL_LOCK_STATE;
|
|---|
| 593 |
|
|---|
| 594 | if( SMALL_OBJ(lb) ) {
|
|---|
| 595 | # ifdef MERGE_SIZES
|
|---|
| 596 | if (EXTRA_BYTES != 0 && lb != 0) lb--;
|
|---|
| 597 | /* We don't need the extra byte, since this won't be */
|
|---|
| 598 | /* collected anyway. */
|
|---|
| 599 | lw = GC_size_map[lb];
|
|---|
| 600 | # else
|
|---|
| 601 | lw = ALIGNED_WORDS(lb);
|
|---|
| 602 | # endif
|
|---|
| 603 | opp = &(GC_auobjfreelist[lw]);
|
|---|
| 604 | FASTLOCK();
|
|---|
| 605 | if( FASTLOCK_SUCCEEDED() && (op = *opp) != 0 ) {
|
|---|
| 606 | /* See above comment on signals. */
|
|---|
| 607 | *opp = obj_link(op);
|
|---|
| 608 | obj_link(op) = 0;
|
|---|
| 609 | GC_words_allocd += lw;
|
|---|
| 610 | /* Mark bit was already set while object was on free list. */
|
|---|
| 611 | GC_non_gc_bytes += WORDS_TO_BYTES(lw);
|
|---|
| 612 | FASTUNLOCK();
|
|---|
| 613 | return((GC_PTR) op);
|
|---|
| 614 | }
|
|---|
| 615 | FASTUNLOCK();
|
|---|
| 616 | op = (ptr_t)GC_generic_malloc((word)lb, AUNCOLLECTABLE);
|
|---|
| 617 | } else {
|
|---|
| 618 | op = (ptr_t)GC_generic_malloc((word)lb, AUNCOLLECTABLE);
|
|---|
| 619 | }
|
|---|
| 620 | if (0 == op) return(0);
|
|---|
| 621 | /* We don't need the lock here, since we have an undisguised */
|
|---|
| 622 | /* pointer. We do need to hold the lock while we adjust */
|
|---|
| 623 | /* mark bits. */
|
|---|
| 624 | {
|
|---|
| 625 | register struct hblk * h;
|
|---|
| 626 |
|
|---|
| 627 | h = HBLKPTR(op);
|
|---|
| 628 | lw = HDR(h) -> hb_sz;
|
|---|
| 629 |
|
|---|
| 630 | DISABLE_SIGNALS();
|
|---|
| 631 | LOCK();
|
|---|
| 632 | GC_set_mark_bit(op);
|
|---|
| 633 | GC_non_gc_bytes += WORDS_TO_BYTES(lw);
|
|---|
| 634 | UNLOCK();
|
|---|
| 635 | ENABLE_SIGNALS();
|
|---|
| 636 | return((GC_PTR) op);
|
|---|
| 637 | }
|
|---|
| 638 | }
|
|---|
| 639 |
|
|---|
| 640 | #endif /* ATOMIC_UNCOLLECTABLE */
|
|---|