1
Fork 0
mirror of https://git.savannah.gnu.org/git/guile.git synced 2025-05-09 15:10:29 +02:00
guile/serial-marker.h
2022-03-11 11:48:26 +01:00

154 lines
3.8 KiB
C

#ifndef SERIAL_TRACE_H
#define SERIAL_TRACE_H
#include <sys/mman.h>
#include <unistd.h>
#include "assert.h"
#include "debug.h"
struct mark_queue {
size_t size;
size_t read;
size_t write;
uintptr_t *buf;
};
static const size_t mark_queue_max_size =
(1ULL << (sizeof(uintptr_t) * 8 - 1)) / sizeof(uintptr_t);
static const size_t mark_queue_release_byte_threshold = 1 * 1024 * 1024;
static void*
mark_queue_alloc(size_t size) {
void *mem = mmap(NULL, size * sizeof(uintptr_t), PROT_READ|PROT_WRITE,
MAP_PRIVATE|MAP_ANONYMOUS, -1, 0);
if (mem == MAP_FAILED) {
perror("Failed to grow mark queue");
DEBUG("Failed to allocate %zu bytes", size);
return NULL;
}
return mem;
}
static int
mark_queue_init(struct mark_queue *q) {
q->size = getpagesize() / sizeof(uintptr_t);
q->read = 0;
q->write = 0;
q->buf = mark_queue_alloc(q->size);
return !!q->buf;
}
static inline uintptr_t
mark_queue_get(struct mark_queue *q, size_t idx) {
return q->buf[idx & (q->size - 1)];
}
static inline void
mark_queue_put(struct mark_queue *q, size_t idx, uintptr_t x) {
q->buf[idx & (q->size - 1)] = x;
}
static int mark_queue_grow(struct mark_queue *q) NEVER_INLINE;
static int
mark_queue_grow(struct mark_queue *q) {
size_t old_size = q->size;
uintptr_t *old_buf = q->buf;
if (old_size >= mark_queue_max_size) {
DEBUG("mark queue already at max size of %zu bytes", old_size);
return 0;
}
size_t new_size = old_size * 2;
uintptr_t *new_buf = mark_queue_alloc(new_size);
if (!new_buf)
return 0;
size_t old_mask = old_size - 1;
size_t new_mask = new_size - 1;
for (size_t i = q->read; i < q->write; i++)
new_buf[i & new_mask] = old_buf[i & old_mask];
munmap(old_buf, old_size * sizeof(uintptr_t));
q->size = new_size;
q->buf = new_buf;
return 1;
}
static inline void
mark_queue_push(struct mark_queue *q, void *p) {
if (UNLIKELY(q->write - q->read == q->size)) {
if (!mark_queue_grow(q))
abort();
}
mark_queue_put(q, q->write++, (uintptr_t)p);
}
static inline void*
mark_queue_pop(struct mark_queue *q) {
if (UNLIKELY(q->read == q->write))
return NULL;
return (void*)mark_queue_get(q, q->read++);
}
static void
mark_queue_release(struct mark_queue *q) {
size_t byte_size = q->size * sizeof(uintptr_t);
if (byte_size >= mark_queue_release_byte_threshold)
madvise(q->buf, byte_size, MADV_DONTNEED);
q->read = q->write = 0;
}
static void
mark_queue_destroy(struct mark_queue *q) {
size_t byte_size = q->size * sizeof(uintptr_t);
munmap(q->buf, byte_size);
}
struct marker {
struct mark_queue queue;
};
struct context;
static inline struct marker* context_marker(struct context *cx);
static int
marker_init(struct context *cx) {
return mark_queue_init(&context_marker(cx)->queue);
}
static void marker_prepare(struct context *cx) {}
static void marker_release(struct context *cx) {
mark_queue_release(&context_marker(cx)->queue);
}
struct gcobj;
static inline void marker_visit(void **loc, void *mark_data) ALWAYS_INLINE;
static inline void marker_trace(struct context *cx,
void (*)(struct context *, struct gcobj *))
ALWAYS_INLINE;
static inline int mark_object(struct context *cx,
struct gcobj *obj) ALWAYS_INLINE;
static inline void
marker_visit(void **loc, void *mark_data) {
struct context *cx = mark_data;
struct gcobj *obj = *loc;
if (obj && mark_object(cx, obj))
mark_queue_push(&context_marker(cx)->queue, obj);
}
static inline void
marker_visit_root(void **loc, struct context *cx) {
marker_visit(loc, cx);
}
static inline void
marker_trace(struct context *cx,
void (*process)(struct context *, struct gcobj *)) {
struct gcobj *obj;
while ((obj = mark_queue_pop(&context_marker(cx)->queue)))
process(cx, obj);
}
#endif // SERIAL_MARK_H