mirror of
https://git.savannah.gnu.org/git/guile.git
synced 2025-05-10 15:50:50 +02:00
Mostly implementation-independent inline allocation
This is a step towards separate compilation of the GC without losing performance. Only remaining task is the write barrier.
This commit is contained in:
parent
4d8a7169d0
commit
a75842be90
4 changed files with 225 additions and 62 deletions
40
bdw.h
40
bdw.h
|
@ -45,6 +45,34 @@ static inline size_t gc_inline_freelist_object_size(size_t idx) {
|
|||
return (idx + 1U) * GC_INLINE_GRANULE_BYTES;
|
||||
}
|
||||
|
||||
static inline enum gc_allocator_kind gc_allocator_kind(void) {
|
||||
return GC_ALLOCATOR_INLINE_FREELIST;
|
||||
}
|
||||
static inline size_t gc_allocator_small_granule_size(void) {
|
||||
return GC_INLINE_GRANULE_BYTES;
|
||||
}
|
||||
static inline size_t gc_allocator_large_threshold(void) {
|
||||
return 256;
|
||||
}
|
||||
|
||||
static inline size_t gc_allocator_allocation_pointer_offset(void) {
|
||||
abort();
|
||||
}
|
||||
static inline size_t gc_allocator_allocation_limit_offset(void) {
|
||||
abort();
|
||||
}
|
||||
|
||||
static inline size_t gc_allocator_freelist_offset(size_t size) {
|
||||
GC_ASSERT(size);
|
||||
return sizeof(void*) * gc_inline_bytes_to_freelist_index(size);
|
||||
}
|
||||
|
||||
static inline void gc_allocator_inline_success(struct mutator *mut,
|
||||
struct gc_ref obj,
|
||||
uintptr_t aligned_size) {}
|
||||
static inline void gc_allocator_inline_failure(struct mutator *mut,
|
||||
uintptr_t aligned_size) {}
|
||||
|
||||
// The values of these must match the internal POINTERLESS and NORMAL
|
||||
// definitions in libgc, for which unfortunately there are no external
|
||||
// definitions. Alack.
|
||||
|
@ -80,12 +108,14 @@ allocate_small(void **freelist, size_t idx, enum gc_inline_kind kind) {
|
|||
return head;
|
||||
}
|
||||
|
||||
static inline void* gc_allocate(struct mutator *mut, size_t size) {
|
||||
static void* gc_allocate_large(struct mutator *mut, size_t size) {
|
||||
return GC_malloc(size);
|
||||
}
|
||||
|
||||
static void* gc_allocate_small(struct mutator *mut, size_t size) {
|
||||
GC_ASSERT(size != 0);
|
||||
GC_ASSERT(size <= gc_allocator_large_threshold());
|
||||
size_t idx = gc_inline_bytes_to_freelist_index(size);
|
||||
|
||||
if (UNLIKELY(idx >= GC_INLINE_FREELIST_COUNT))
|
||||
return GC_malloc(size);
|
||||
|
||||
return allocate_small(&mut->freelists[idx], idx, GC_INLINE_KIND_NORMAL);
|
||||
}
|
||||
|
||||
|
|
96
gc-api.h
96
gc-api.h
|
@ -23,6 +23,10 @@ struct gc_option {
|
|||
double value;
|
||||
};
|
||||
|
||||
struct gc_mutator {
|
||||
void *user_data;
|
||||
};
|
||||
|
||||
// FIXME: Conflict with bdw-gc GC_API. Switch prefix?
|
||||
#ifndef GC_API_
|
||||
#define GC_API_ static
|
||||
|
@ -38,8 +42,96 @@ GC_API_ void gc_finish_for_thread(struct mutator *mut);
|
|||
GC_API_ void* gc_call_without_gc(struct mutator *mut, void* (*f)(void*),
|
||||
void *data) GC_NEVER_INLINE;
|
||||
|
||||
GC_API_ inline void* gc_allocate(struct mutator *mut, size_t bytes);
|
||||
GC_API_ void* gc_allocate_small(struct mutator *mut, size_t bytes) GC_NEVER_INLINE;
|
||||
GC_API_ void* gc_allocate_large(struct mutator *mut, size_t bytes) GC_NEVER_INLINE;
|
||||
static inline void* gc_allocate(struct mutator *mut, size_t bytes) GC_ALWAYS_INLINE;
|
||||
// FIXME: remove :P
|
||||
GC_API_ inline void* gc_allocate_pointerless(struct mutator *mut, size_t bytes);
|
||||
static inline void* gc_allocate_pointerless(struct mutator *mut, size_t bytes);
|
||||
|
||||
enum gc_allocator_kind {
|
||||
GC_ALLOCATOR_INLINE_BUMP_POINTER,
|
||||
GC_ALLOCATOR_INLINE_FREELIST,
|
||||
GC_ALLOCATOR_INLINE_NONE
|
||||
};
|
||||
|
||||
static inline enum gc_allocator_kind gc_allocator_kind(void) GC_ALWAYS_INLINE;
|
||||
|
||||
static inline size_t gc_allocator_large_threshold(void) GC_ALWAYS_INLINE;
|
||||
|
||||
static inline size_t gc_allocator_small_granule_size(void) GC_ALWAYS_INLINE;
|
||||
|
||||
static inline size_t gc_allocator_allocation_pointer_offset(void) GC_ALWAYS_INLINE;
|
||||
static inline size_t gc_allocator_allocation_limit_offset(void) GC_ALWAYS_INLINE;
|
||||
|
||||
static inline size_t gc_allocator_freelist_offset(size_t size) GC_ALWAYS_INLINE;
|
||||
|
||||
static inline void gc_allocator_inline_success(struct mutator *mut,
|
||||
struct gc_ref obj,
|
||||
uintptr_t aligned_size);
|
||||
static inline void gc_allocator_inline_failure(struct mutator *mut,
|
||||
uintptr_t aligned_size);
|
||||
|
||||
static inline void*
|
||||
gc_allocate_bump_pointer(struct mutator *mut, size_t size) GC_ALWAYS_INLINE;
|
||||
static inline void* gc_allocate_bump_pointer(struct mutator *mut, size_t size) {
|
||||
GC_ASSERT(size <= gc_allocator_large_threshold());
|
||||
|
||||
size_t granule_size = gc_allocator_small_granule_size();
|
||||
size_t hp_offset = gc_allocator_allocation_pointer_offset();
|
||||
size_t limit_offset = gc_allocator_allocation_limit_offset();
|
||||
|
||||
uintptr_t base_addr = (uintptr_t)mut;
|
||||
uintptr_t *hp_loc = (uintptr_t*)(base_addr + hp_offset);
|
||||
uintptr_t *limit_loc = (uintptr_t*)(base_addr + limit_offset);
|
||||
|
||||
size = (size + granule_size - 1) & ~(granule_size - 1);
|
||||
uintptr_t hp = *hp_loc;
|
||||
uintptr_t limit = *limit_loc;
|
||||
uintptr_t new_hp = hp + size;
|
||||
|
||||
if (GC_UNLIKELY (new_hp > limit)) {
|
||||
gc_allocator_inline_failure(mut, size);
|
||||
return gc_allocate_small(mut, size);
|
||||
}
|
||||
|
||||
gc_allocator_inline_success(mut, gc_ref(hp), size);
|
||||
|
||||
*hp_loc = new_hp;
|
||||
return (void*)hp;
|
||||
}
|
||||
|
||||
static inline void* gc_allocate_freelist(struct mutator *mut,
|
||||
size_t size) GC_ALWAYS_INLINE;
|
||||
static inline void* gc_allocate_freelist(struct mutator *mut, size_t size) {
|
||||
GC_ASSERT(size <= gc_allocator_large_threshold());
|
||||
|
||||
size_t freelist_offset = gc_allocator_freelist_offset(size);
|
||||
uintptr_t base_addr = (uintptr_t)mut;
|
||||
void **freelist_loc = (void**)(base_addr + freelist_offset);
|
||||
|
||||
void *head = *freelist_loc;
|
||||
if (GC_UNLIKELY(!head))
|
||||
return gc_allocate_small(mut, size);
|
||||
|
||||
*freelist_loc = *(void**)head;
|
||||
return head;
|
||||
}
|
||||
|
||||
static inline void* gc_allocate(struct mutator *mut, size_t size) {
|
||||
GC_ASSERT(size != 0);
|
||||
if (size > gc_allocator_large_threshold())
|
||||
return gc_allocate_large(mut, size);
|
||||
|
||||
switch (gc_allocator_kind()) {
|
||||
case GC_ALLOCATOR_INLINE_BUMP_POINTER:
|
||||
return gc_allocate_bump_pointer(mut, size);
|
||||
case GC_ALLOCATOR_INLINE_FREELIST:
|
||||
return gc_allocate_freelist(mut, size);
|
||||
case GC_ALLOCATOR_INLINE_NONE:
|
||||
return gc_allocate_small(mut, size);
|
||||
default:
|
||||
abort();
|
||||
}
|
||||
}
|
||||
|
||||
#endif // GC_API_H_
|
||||
|
|
51
semi.h
51
semi.h
|
@ -29,6 +29,43 @@ struct mutator {
|
|||
struct handle *roots;
|
||||
};
|
||||
|
||||
static const uintptr_t ALIGNMENT = 8;
|
||||
static const size_t LARGE_OBJECT_THRESHOLD = 8192;
|
||||
|
||||
static inline void clear_memory(uintptr_t addr, size_t size) {
|
||||
memset((char*)addr, 0, size);
|
||||
}
|
||||
|
||||
static inline enum gc_allocator_kind gc_allocator_kind(void) {
|
||||
return GC_ALLOCATOR_INLINE_BUMP_POINTER;
|
||||
}
|
||||
static inline size_t gc_allocator_small_granule_size(void) {
|
||||
return ALIGNMENT;
|
||||
}
|
||||
static inline size_t gc_allocator_large_threshold(void) {
|
||||
return LARGE_OBJECT_THRESHOLD;
|
||||
}
|
||||
|
||||
static inline size_t gc_allocator_allocation_pointer_offset(void) {
|
||||
return offsetof(struct semi_space, hp);
|
||||
}
|
||||
static inline size_t gc_allocator_allocation_limit_offset(void) {
|
||||
return offsetof(struct semi_space, limit);
|
||||
}
|
||||
|
||||
static inline size_t gc_allocator_freelist_offset(size_t size) {
|
||||
abort();
|
||||
}
|
||||
|
||||
static inline void gc_allocator_inline_success(struct mutator *mut,
|
||||
struct gc_ref obj,
|
||||
uintptr_t aligned_size) {
|
||||
// FIXME: Allow allocator to avoid clearing memory?
|
||||
clear_memory(gc_ref_value(obj), aligned_size);
|
||||
}
|
||||
static inline void gc_allocator_inline_failure(struct mutator *mut,
|
||||
uintptr_t aligned_size) {}
|
||||
|
||||
static inline struct heap* mutator_heap(struct mutator *mut) {
|
||||
return &mut->heap;
|
||||
}
|
||||
|
@ -42,16 +79,10 @@ static inline struct semi_space* mutator_semi_space(struct mutator *mut) {
|
|||
return heap_semi_space(mutator_heap(mut));
|
||||
}
|
||||
|
||||
static const uintptr_t ALIGNMENT = 8;
|
||||
|
||||
static uintptr_t align_up(uintptr_t addr, size_t align) {
|
||||
return (addr + align - 1) & ~(align-1);
|
||||
}
|
||||
|
||||
static inline void clear_memory(uintptr_t addr, size_t size) {
|
||||
memset((char*)addr, 0, size);
|
||||
}
|
||||
|
||||
static void collect(struct mutator *mut) GC_NEVER_INLINE;
|
||||
static void collect_for_alloc(struct mutator *mut, size_t bytes) GC_NEVER_INLINE;
|
||||
|
||||
|
@ -171,8 +202,7 @@ static void collect_for_alloc(struct mutator *mut, size_t bytes) {
|
|||
}
|
||||
}
|
||||
|
||||
static const size_t LARGE_OBJECT_THRESHOLD = 8192;
|
||||
static void* allocate_large(struct mutator *mut, size_t size) {
|
||||
static void* gc_allocate_large(struct mutator *mut, size_t size) {
|
||||
struct heap *heap = mutator_heap(mut);
|
||||
struct large_object_space *space = heap_large_object_space(heap);
|
||||
struct semi_space *semi_space = heap_semi_space(heap);
|
||||
|
@ -198,10 +228,7 @@ static void* allocate_large(struct mutator *mut, size_t size) {
|
|||
return ret;
|
||||
}
|
||||
|
||||
static inline void* gc_allocate(struct mutator *mut, size_t size) {
|
||||
if (size >= LARGE_OBJECT_THRESHOLD)
|
||||
return allocate_large(mut, size);
|
||||
|
||||
static void* gc_allocate_small(struct mutator *mut, size_t size) {
|
||||
struct semi_space *space = mutator_semi_space(mut);
|
||||
while (1) {
|
||||
uintptr_t addr = space->hp;
|
||||
|
|
100
whippet.h
100
whippet.h
|
@ -369,6 +369,44 @@ static inline struct heap* mutator_heap(struct mutator *mutator) {
|
|||
return mutator->heap;
|
||||
}
|
||||
|
||||
static inline enum gc_allocator_kind gc_allocator_kind(void) {
|
||||
return GC_ALLOCATOR_INLINE_BUMP_POINTER;
|
||||
}
|
||||
static inline size_t gc_allocator_small_granule_size(void) {
|
||||
return GRANULE_SIZE;
|
||||
}
|
||||
static inline size_t gc_allocator_large_threshold(void) {
|
||||
return LARGE_OBJECT_THRESHOLD;
|
||||
}
|
||||
|
||||
static inline size_t gc_allocator_allocation_pointer_offset(void) {
|
||||
return offsetof(struct mutator, alloc);
|
||||
}
|
||||
static inline size_t gc_allocator_allocation_limit_offset(void) {
|
||||
return offsetof(struct mutator, sweep);
|
||||
}
|
||||
|
||||
static inline size_t gc_allocator_freelist_offset(size_t size) {
|
||||
abort();
|
||||
}
|
||||
|
||||
static inline void gc_allocator_inline_success(struct mutator *mut,
|
||||
struct gc_ref obj,
|
||||
uintptr_t aligned_size) {
|
||||
uint8_t *metadata = object_metadata_byte(gc_ref_heap_object(obj));
|
||||
size_t granules = aligned_size >> GRANULE_SIZE_LOG_2;
|
||||
if (granules == 1) {
|
||||
metadata[0] = METADATA_BYTE_YOUNG | METADATA_BYTE_END;
|
||||
} else {
|
||||
metadata[0] = METADATA_BYTE_YOUNG;
|
||||
if (granules > 2)
|
||||
memset(metadata + 1, 0, granules - 2);
|
||||
metadata[granules - 1] = METADATA_BYTE_END;
|
||||
}
|
||||
}
|
||||
static inline void gc_allocator_inline_failure(struct mutator *mut,
|
||||
uintptr_t aligned_size) {}
|
||||
|
||||
static inline void clear_memory(uintptr_t addr, size_t size) {
|
||||
memset((char*)addr, 0, size);
|
||||
}
|
||||
|
@ -1756,11 +1794,10 @@ static void trigger_collection(struct mutator *mut) {
|
|||
heap_unlock(heap);
|
||||
}
|
||||
|
||||
static void* allocate_large(struct mutator *mut, size_t granules) {
|
||||
static void* gc_allocate_large(struct mutator *mut, size_t size) {
|
||||
struct heap *heap = mutator_heap(mut);
|
||||
struct large_object_space *space = heap_large_object_space(heap);
|
||||
|
||||
size_t size = granules * GRANULE_SIZE;
|
||||
size_t npages = large_object_space_npages(space, size);
|
||||
|
||||
mark_space_request_release_memory(heap_mark_space(heap),
|
||||
|
@ -1782,58 +1819,35 @@ static void* allocate_large(struct mutator *mut, size_t granules) {
|
|||
return ret;
|
||||
}
|
||||
|
||||
static void* allocate_small_slow(struct mutator *mut, size_t granules) GC_NEVER_INLINE;
|
||||
static void* allocate_small_slow(struct mutator *mut, size_t granules) {
|
||||
while (1) {
|
||||
size_t hole = next_hole(mut);
|
||||
if (hole >= granules) {
|
||||
clear_memory(mut->alloc, hole * GRANULE_SIZE);
|
||||
break;
|
||||
}
|
||||
if (!hole)
|
||||
trigger_collection(mut);
|
||||
}
|
||||
struct gcobj* ret = (struct gcobj*)mut->alloc;
|
||||
mut->alloc += granules * GRANULE_SIZE;
|
||||
return ret;
|
||||
}
|
||||
|
||||
static inline void* allocate_small(struct mutator *mut, size_t granules) {
|
||||
GC_ASSERT(granules > 0); // allocating 0 granules would be silly
|
||||
static void* gc_allocate_small(struct mutator *mut, size_t size) {
|
||||
GC_ASSERT(size > 0); // allocating 0 bytes would be silly
|
||||
GC_ASSERT(size <= gc_allocator_large_threshold());
|
||||
size = align_up(size, GRANULE_SIZE);
|
||||
uintptr_t alloc = mut->alloc;
|
||||
uintptr_t sweep = mut->sweep;
|
||||
uintptr_t new_alloc = alloc + granules * GRANULE_SIZE;
|
||||
uintptr_t new_alloc = alloc + size;
|
||||
struct gcobj *obj;
|
||||
if (new_alloc <= sweep) {
|
||||
mut->alloc = new_alloc;
|
||||
obj = (struct gcobj *)alloc;
|
||||
} else {
|
||||
obj = allocate_small_slow(mut, granules);
|
||||
}
|
||||
uint8_t *metadata = object_metadata_byte(obj);
|
||||
if (granules == 1) {
|
||||
metadata[0] = METADATA_BYTE_YOUNG | METADATA_BYTE_END;
|
||||
} else {
|
||||
metadata[0] = METADATA_BYTE_YOUNG;
|
||||
if (granules > 2)
|
||||
memset(metadata + 1, 0, granules - 2);
|
||||
metadata[granules - 1] = METADATA_BYTE_END;
|
||||
size_t granules = size >> GRANULE_SIZE_LOG_2;
|
||||
while (1) {
|
||||
size_t hole = next_hole(mut);
|
||||
if (hole >= granules) {
|
||||
clear_memory(mut->alloc, hole * GRANULE_SIZE);
|
||||
break;
|
||||
}
|
||||
if (!hole)
|
||||
trigger_collection(mut);
|
||||
}
|
||||
obj = (struct gcobj*)mut->alloc;
|
||||
mut->alloc += size;
|
||||
}
|
||||
gc_allocator_inline_success(mut, gc_ref_from_heap_object(obj), size);
|
||||
return obj;
|
||||
}
|
||||
|
||||
static inline void* allocate_medium(struct mutator *mut, size_t granules) {
|
||||
return allocate_small(mut, granules);
|
||||
}
|
||||
|
||||
static inline void* gc_allocate(struct mutator *mut, size_t size) {
|
||||
size_t granules = size_to_granules(size);
|
||||
if (granules <= MEDIUM_OBJECT_GRANULE_THRESHOLD)
|
||||
return allocate_small(mut, granules);
|
||||
if (granules <= LARGE_OBJECT_GRANULE_THRESHOLD)
|
||||
return allocate_medium(mut, granules);
|
||||
return allocate_large(mut, granules);
|
||||
}
|
||||
static inline void* gc_allocate_pointerless(struct mutator *mut, size_t size) {
|
||||
return gc_allocate(mut, size);
|
||||
}
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue