hardened_malloc/h_malloc.c

1745 lines
51 KiB
C
Raw Normal View History

2018-08-22 00:53:22 +05:30
#include <assert.h>
#include <errno.h>
#include <stdatomic.h>
#include <stdbool.h>
#include <stdlib.h>
#include <string.h>
2019-03-26 10:58:27 +05:30
#if N_ARENA > 1
2019-03-26 00:19:50 +05:30
#include <threads.h>
2019-03-26 10:58:27 +05:30
#endif
2018-08-22 00:53:22 +05:30
2018-12-05 12:52:10 +05:30
#include <malloc.h>
2018-08-22 00:53:22 +05:30
#include <pthread.h>
#include <sys/mman.h>
#include <sys/utsname.h>
2018-10-19 05:05:21 +05:30
#include <unistd.h>
2018-08-22 00:53:22 +05:30
#include "third_party/libdivide.h"
2018-12-05 12:51:28 +05:30
#include "h_malloc.h"
#include "mutex.h"
#include "memory.h"
2018-09-02 11:33:27 +05:30
#include "pages.h"
2018-08-22 00:53:22 +05:30
#include "random.h"
#include "util.h"
#define SLAB_QUARANTINE (SLAB_QUARANTINE_RANDOM_LENGTH > 0 || SLAB_QUARANTINE_QUEUE_LENGTH > 0)
2019-01-06 11:22:25 +05:30
#define MREMAP_MOVE_THRESHOLD (32 * 1024 * 1024)
2018-08-22 00:53:22 +05:30
static_assert(sizeof(void *) == 8, "64-bit only");
2018-09-07 10:03:51 +05:30
static_assert(!WRITE_AFTER_FREE_CHECK || ZERO_ON_FREE, "WRITE_AFTER_FREE_CHECK depends on ZERO_ON_FREE");
2019-01-06 11:22:25 +05:30
static_assert(SLAB_QUARANTINE_RANDOM_LENGTH >= 0 && SLAB_QUARANTINE_RANDOM_LENGTH <= 65536,
"invalid slab quarantine random length");
static_assert(SLAB_QUARANTINE_QUEUE_LENGTH >= 0 && SLAB_QUARANTINE_QUEUE_LENGTH <= 65536,
"invalid slab quarantine queue length");
static_assert(REGION_QUARANTINE_RANDOM_LENGTH >= 0 && REGION_QUARANTINE_RANDOM_LENGTH <= 65536,
"invalid region quarantine random length");
static_assert(REGION_QUARANTINE_QUEUE_LENGTH >= 0 && REGION_QUARANTINE_QUEUE_LENGTH <= 65536,
"invalid region quarantine queue length");
static_assert(FREE_SLABS_QUARANTINE_RANDOM_LENGTH >= 0 && FREE_SLABS_QUARANTINE_RANDOM_LENGTH <= 65536,
"invalid free slabs quarantine random length");
static_assert(REGION_QUARANTINE_SKIP_THRESHOLD >= 0,
"invalid region quarantine skip threshold");
static_assert(MREMAP_MOVE_THRESHOLD >= REGION_QUARANTINE_SKIP_THRESHOLD,
"mremap move threshold must be above region quarantine limit");
// either sizeof(u64) or 0
static const size_t canary_size = SLAB_CANARY ? sizeof(u64) : 0;
2019-03-26 00:19:50 +05:30
static_assert(N_ARENA >= 1, "must have at least 1 arena");
static_assert(N_ARENA <= 4, "currently only support up to 4 arenas (as an initial arbitrary limitation)");
#define CACHELINE_SIZE 64
2018-08-22 00:53:22 +05:30
2019-03-26 00:19:50 +05:30
#if N_ARENA > 1
__attribute__((tls_model("initial-exec")))
static thread_local unsigned thread_arena = N_ARENA;
#else
static const unsigned thread_arena = 0;
#endif
2018-08-22 00:53:22 +05:30
static union {
struct {
2018-11-05 12:49:04 +05:30
void *_Atomic slab_region_start;
2018-08-22 00:53:22 +05:30
void *slab_region_end;
2019-03-26 00:19:50 +05:30
struct size_class *size_class_metadata[N_ARENA];
struct region_allocator *region_allocator;
2018-10-15 07:24:23 +05:30
struct region_metadata *regions[2];
#ifdef USE_PKEY
int metadata_pkey;
bool pkey_state_preserved_on_fork;
#endif
2018-08-22 00:53:22 +05:30
};
char padding[PAGE_SIZE];
} ro __attribute__((aligned(PAGE_SIZE)));
2018-08-22 00:53:22 +05:30
2018-11-05 12:49:04 +05:30
static inline void *get_slab_region_start() {
return atomic_load_explicit(&ro.slab_region_start, memory_order_acquire);
}
#define SLAB_METADATA_COUNT
2018-08-22 00:53:22 +05:30
struct slab_metadata {
u64 bitmap[4];
2018-08-22 00:53:22 +05:30
struct slab_metadata *next;
struct slab_metadata *prev;
u64 canary_value;
#ifdef SLAB_METADATA_COUNT
u16 count;
#endif
2018-11-06 02:49:50 +05:30
#if SLAB_QUARANTINE
u64 quarantine_bitmap[4];
2018-11-06 02:49:50 +05:30
#endif
2018-08-22 00:53:22 +05:30
};
static const size_t min_align = 16;
2018-08-22 00:53:22 +05:30
static const size_t max_slab_size_class = 16384;
2019-01-03 00:52:28 +05:30
#define MIN_SLAB_SIZE_CLASS_SHIFT 4
#define MAX_SLAB_SIZE_CLASS_SHIFT 14
2018-08-22 00:53:22 +05:30
static const u16 size_classes[] = {
/* 0 */ 0,
2018-08-22 00:53:22 +05:30
/* 16 */ 16, 32, 48, 64, 80, 96, 112, 128,
/* 32 */ 160, 192, 224, 256,
/* 64 */ 320, 384, 448, 512,
/* 128 */ 640, 768, 896, 1024,
/* 256 */ 1280, 1536, 1792, 2048,
/* 512 */ 2560, 3072, 3584, 4096,
/* 1024 */ 5120, 6144, 7168, 8192,
/* 2048 */ 10240, 12288, 14336, 16384
};
static const u16 size_class_slots[] = {
/* 0 */ 256,
/* 16 */ 256, 128, 85, 64, 51, 42, 36, 64,
/* 32 */ 51, 64, 54, 64,
2018-08-22 00:53:22 +05:30
/* 64 */ 64, 64, 64, 64,
/* 128 */ 64, 64, 64, 64,
/* 256 */ 16, 16, 16, 16,
/* 512 */ 8, 8, 8, 8,
/* 1024 */ 8, 8, 8, 8,
/* 2048 */ 6, 5, 4, 4
2018-08-22 00:53:22 +05:30
};
static const char *const size_class_labels[] = {
/* 0 */ "malloc 0",
/* 16 */ "malloc 16", "malloc 32", "malloc 48", "malloc 64", "malloc 80", "malloc 96", "malloc 112", "malloc 128",
/* 32 */ "malloc 160", "malloc 192", "malloc 224", "malloc 256",
/* 64 */ "malloc 320", "malloc 384", "malloc 448", "malloc 512",
/* 128 */ "malloc 640", "malloc 768", "malloc 896", "malloc 1024",
/* 256 */ "malloc 1280", "malloc 1536", "malloc 1792", "malloc 2048",
/* 512 */ "malloc 2560", "malloc 3072", "malloc 3584", "malloc 4096",
/* 1024 */ "malloc 5120", "malloc 6144", "malloc 7168", "malloc 8192",
/* 2048 */ "malloc 10240", "malloc 12288", "malloc 14336", "malloc 16384"
};
2018-08-22 00:53:22 +05:30
#define N_SIZE_CLASSES (sizeof(size_classes) / sizeof(size_classes[0]))
struct size_info {
size_t size;
size_t class;
};
2018-08-30 00:43:53 +05:30
static inline struct size_info get_size_info(size_t size) {
if (size == 0) {
return (struct size_info){0, 0};
}
if (size <= 128) {
return (struct size_info){(size + 15) & ~15, ((size - 1) >> 4) + 1};
}
for (unsigned class = 9; class < N_SIZE_CLASSES; class++) {
size_t real_size = size_classes[class];
2018-08-22 00:53:22 +05:30
if (size <= real_size) {
return (struct size_info){real_size, class};
2018-08-22 00:53:22 +05:30
}
}
fatal_error("invalid size for slabs");
}
// alignment must be a power of 2 <= PAGE_SIZE since slabs are only page aligned
static inline struct size_info get_size_info_align(size_t size, size_t alignment) {
for (unsigned class = 1; class < N_SIZE_CLASSES; class++) {
size_t real_size = size_classes[class];
if (size <= real_size && !(real_size & (alignment - 1))) {
return (struct size_info){real_size, class};
}
}
fatal_error("invalid size for slabs");
}
2018-08-22 00:53:22 +05:30
static size_t get_slab_size(size_t slots, size_t size) {
return PAGE_CEILING(slots * size);
}
2018-09-01 08:12:34 +05:30
// limit on the number of cached empty slabs before attempting purging instead
static const size_t max_empty_slabs_total = 64 * 1024;
struct __attribute__((aligned(CACHELINE_SIZE))) size_class {
struct mutex lock;
2018-10-15 00:39:55 +05:30
2018-08-22 00:53:22 +05:30
void *class_region_start;
struct slab_metadata *slab_info;
2018-10-15 00:39:55 +05:30
struct libdivide_u32_t size_divisor;
struct libdivide_u64_t slab_size_divisor;
2018-09-01 08:12:34 +05:30
#if SLAB_QUARANTINE_RANDOM_LENGTH > 0
2019-01-03 00:52:28 +05:30
void *quarantine_random[SLAB_QUARANTINE_RANDOM_LENGTH << (MAX_SLAB_SIZE_CLASS_SHIFT - MIN_SLAB_SIZE_CLASS_SHIFT)];
2018-11-06 02:49:50 +05:30
#endif
#if SLAB_QUARANTINE_QUEUE_LENGTH > 0
2019-01-03 00:52:28 +05:30
void *quarantine_queue[SLAB_QUARANTINE_QUEUE_LENGTH << (MAX_SLAB_SIZE_CLASS_SHIFT - MIN_SLAB_SIZE_CLASS_SHIFT)];
2018-11-06 02:49:50 +05:30
size_t quarantine_queue_index;
#endif
2018-09-01 08:12:34 +05:30
// slabs with at least one allocated slot and at least one free slot
//
// LIFO doubly-linked list
2018-08-22 00:53:22 +05:30
struct slab_metadata *partial_slabs;
2018-09-01 08:12:34 +05:30
// slabs without allocated slots that are cached for near-term usage
//
// LIFO singly-linked list
2018-08-30 14:25:39 +05:30
struct slab_metadata *empty_slabs;
2018-09-01 08:12:34 +05:30
size_t empty_slabs_total; // length * slab_size
// slabs without allocated slots that are purged and memory protected
//
// FIFO singly-linked list
struct slab_metadata *free_slabs_head;
struct slab_metadata *free_slabs_tail;
struct slab_metadata *free_slabs_quarantine[FREE_SLABS_QUARANTINE_RANDOM_LENGTH];
2018-09-01 08:12:34 +05:30
#if STATS
2019-04-07 08:36:47 +05:30
u64 nmalloc; // may wrap (per jemalloc API)
u64 ndalloc; // may wrap (per jemalloc API)
size_t allocated;
size_t slab_allocated;
#endif
2018-08-25 09:51:02 +05:30
struct random_state rng;
size_t metadata_allocated;
size_t metadata_count;
size_t metadata_count_unguarded;
};
2018-08-22 00:53:22 +05:30
2018-12-05 19:53:05 +05:30
#define CLASS_REGION_SIZE (size_t)CONFIG_CLASS_REGION_SIZE
#define REAL_CLASS_REGION_SIZE (CLASS_REGION_SIZE * 2)
2019-03-26 00:19:50 +05:30
#define ARENA_SIZE (REAL_CLASS_REGION_SIZE * N_SIZE_CLASSES)
static const size_t slab_region_size = ARENA_SIZE * N_ARENA;
2018-08-22 00:53:22 +05:30
static_assert(PAGE_SIZE == 4096, "bitmap handling will need adjustment for other page sizes");
static void *get_slab(struct size_class *c, size_t slab_size, struct slab_metadata *metadata) {
size_t index = metadata - c->slab_info;
return (char *)c->class_region_start + (index * slab_size);
}
2018-10-19 01:31:04 +05:30
#define MAX_METADATA_MAX (CLASS_REGION_SIZE / PAGE_SIZE)
2018-08-22 00:53:22 +05:30
static size_t get_metadata_max(size_t slab_size) {
return CLASS_REGION_SIZE / slab_size;
2018-08-22 00:53:22 +05:30
}
static struct slab_metadata *alloc_metadata(struct size_class *c, size_t slab_size, bool non_zero_size) {
2018-09-07 04:23:06 +05:30
if (unlikely(c->metadata_count >= c->metadata_allocated)) {
2018-08-22 00:53:22 +05:30
size_t metadata_max = get_metadata_max(slab_size);
2018-09-07 04:23:06 +05:30
if (c->metadata_count >= metadata_max) {
errno = ENOMEM;
2018-08-22 00:53:22 +05:30
return NULL;
}
size_t allocate = max(c->metadata_allocated * 2, PAGE_SIZE / sizeof(struct slab_metadata));
2018-08-22 00:53:22 +05:30
if (allocate > metadata_max) {
allocate = metadata_max;
}
if (memory_protect_rw_metadata(c->slab_info, allocate * sizeof(struct slab_metadata))) {
2018-08-22 00:53:22 +05:30
return NULL;
}
c->metadata_allocated = allocate;
}
struct slab_metadata *metadata = c->slab_info + c->metadata_count;
void *slab = get_slab(c, slab_size, metadata);
if (non_zero_size && memory_protect_rw(slab, slab_size)) {
return NULL;
}
2018-08-22 00:53:22 +05:30
c->metadata_count++;
c->metadata_count_unguarded++;
if (c->metadata_count_unguarded >= GUARD_SLABS_INTERVAL) {
2018-09-07 04:23:06 +05:30
c->metadata_count++;
c->metadata_count_unguarded = 0;
2018-09-07 04:23:06 +05:30
}
2018-08-22 00:53:22 +05:30
return metadata;
}
static void set_slot(struct slab_metadata *metadata, size_t index) {
size_t bucket = index / 64;
metadata->bitmap[bucket] |= 1UL << (index - bucket * 64);
#ifdef SLAB_METADATA_COUNT
metadata->count++;
#endif
2018-08-22 00:53:22 +05:30
}
static void clear_slot(struct slab_metadata *metadata, size_t index) {
size_t bucket = index / 64;
metadata->bitmap[bucket] &= ~(1UL << (index - bucket * 64));
#ifdef SLAB_METADATA_COUNT
metadata->count--;
#endif
2018-08-22 00:53:22 +05:30
}
static bool get_slot(struct slab_metadata *metadata, size_t index) {
size_t bucket = index / 64;
return (metadata->bitmap[bucket] >> (index - bucket * 64)) & 1UL;
2018-08-22 00:53:22 +05:30
}
2018-11-06 02:49:50 +05:30
#if SLAB_QUARANTINE
static void set_quarantine(struct slab_metadata *metadata, size_t index) {
size_t bucket = index / 64;
metadata->quarantine_bitmap[bucket] |= 1UL << (index - bucket * 64);
2018-11-06 02:49:50 +05:30
}
static void clear_quarantine(struct slab_metadata *metadata, size_t index) {
size_t bucket = index / 64;
metadata->quarantine_bitmap[bucket] &= ~(1UL << (index - bucket * 64));
2018-11-06 02:49:50 +05:30
}
static bool get_quarantine(struct slab_metadata *metadata, size_t index) {
size_t bucket = index / 64;
return (metadata->quarantine_bitmap[bucket] >> (index - bucket * 64)) & 1UL;
2018-11-06 02:49:50 +05:30
}
#endif
static u64 get_mask(size_t slots) {
2018-08-22 00:53:22 +05:30
return slots < 64 ? ~0UL << slots : 0;
}
2018-08-25 12:32:39 +05:30
static size_t get_free_slot(struct random_state *rng, size_t slots, struct slab_metadata *metadata) {
if (SLOT_RANDOMIZE) {
// randomize start location for linear search (uniform random choice is too slow)
unsigned random_index = get_random_u16_uniform(rng, slots);
unsigned first_bitmap = random_index / 64;
u64 random_split = ~(~0UL << (random_index - first_bitmap * 64));
2018-10-31 03:39:07 +05:30
unsigned i = first_bitmap;
u64 masked = metadata->bitmap[i];
masked |= random_split;
for (;;) {
if (i == slots / 64) {
masked |= get_mask(slots - i * 64);
}
2018-10-31 02:07:23 +05:30
if (masked != ~0UL) {
return ffzl(masked) - 1 + i * 64;
}
2018-10-31 03:39:07 +05:30
i = i == (slots - 1) / 64 ? 0 : i + 1;
masked = metadata->bitmap[i];
}
2018-10-31 03:39:07 +05:30
} else {
for (unsigned i = 0; i <= (slots - 1) / 64; i++) {
u64 masked = metadata->bitmap[i];
if (i == (slots - 1) / 64) {
masked |= get_mask(slots - i * 64);
}
2018-10-31 03:39:07 +05:30
if (masked != ~0UL) {
return ffzl(masked) - 1 + i * 64;
}
}
2018-08-25 12:32:39 +05:30
}
fatal_error("no zero bits");
2018-08-22 00:53:22 +05:30
}
static bool has_free_slots(size_t slots, struct slab_metadata *metadata) {
#ifdef SLAB_METADATA_COUNT
return metadata->count < slots;
#else
if (slots <= 64) {
u64 masked = metadata->bitmap[0] | get_mask(slots);
return masked != ~0UL;
}
if (slots <= 128) {
u64 masked = metadata->bitmap[1] | get_mask(slots - 64);
return metadata->bitmap[0] != ~0UL || masked != ~0UL;
}
if (slots <= 192) {
u64 masked = metadata->bitmap[2] | get_mask(slots - 128);
return metadata->bitmap[0] != ~0UL || metadata->bitmap[1] != ~0UL || masked != ~0UL;
2018-08-25 12:32:39 +05:30
}
u64 masked = metadata->bitmap[3] | get_mask(slots - 192);
return metadata->bitmap[0] != ~0UL || metadata->bitmap[1] != ~0UL || metadata->bitmap[2] != ~0UL || masked != ~0UL;
#endif
2018-08-22 00:53:22 +05:30
}
static bool is_free_slab(struct slab_metadata *metadata) {
#ifdef SLAB_METADATA_COUNT
return !metadata->count;
#else
return !metadata->bitmap[0] && !metadata->bitmap[1] && !metadata->bitmap[2] &&
!metadata->bitmap[3];
#endif
2018-08-22 00:53:22 +05:30
}
static struct slab_metadata *get_metadata(struct size_class *c, void *p) {
2018-08-22 00:53:22 +05:30
size_t offset = (char *)p - (char *)c->class_region_start;
size_t index = libdivide_u64_do(offset, &c->slab_size_divisor);
// still caught without this check either as a read access violation or "double free"
if (index >= c->metadata_allocated) {
fatal_error("invalid free within a slab yet to be used");
}
2018-08-22 00:53:22 +05:30
return c->slab_info + index;
}
static void *slot_pointer(size_t size, void *slab, size_t slot) {
return (char *)slab + slot * size;
}
static void write_after_free_check(const char *p, size_t size) {
2018-09-07 09:47:22 +05:30
if (!WRITE_AFTER_FREE_CHECK) {
return;
}
for (size_t i = 0; i < size; i += sizeof(u64)) {
if (*(const u64 *)(const void *)(p + i)) {
2018-09-07 09:47:22 +05:30
fatal_error("detected write after free");
}
}
}
static const u64 canary_mask = __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ ?
0xffffffffffffff00UL :
0x00ffffffffffffffUL;
2018-10-04 02:39:57 +05:30
static void set_canary(struct slab_metadata *metadata, void *p, size_t size) {
memcpy((char *)p + size - canary_size, &metadata->canary_value, canary_size);
}
2018-10-31 04:52:56 +05:30
static u64 get_random_canary(struct random_state *rng) {
return get_random_u64(rng) & canary_mask;
}
static inline void *allocate_small(size_t requested_size) {
2018-08-22 00:53:22 +05:30
struct size_info info = get_size_info(requested_size);
size_t size = info.size ? info.size : 16;
2019-03-26 00:19:50 +05:30
#if N_ARENA > 1
if (unlikely(thread_arena == N_ARENA)) {
thread_arena = hash_page(&thread_arena) % N_ARENA;
}
#endif
struct size_class *c = &ro.size_class_metadata[thread_arena][info.class];
2018-08-22 00:53:22 +05:30
size_t slots = size_class_slots[info.class];
size_t slab_size = get_slab_size(slots, size);
mutex_lock(&c->lock);
2018-08-22 00:53:22 +05:30
if (c->partial_slabs == NULL) {
2018-08-30 14:25:39 +05:30
if (c->empty_slabs != NULL) {
struct slab_metadata *metadata = c->empty_slabs;
c->empty_slabs = c->empty_slabs->next;
c->empty_slabs_total -= slab_size;
2018-08-22 00:53:22 +05:30
metadata->next = NULL;
2018-08-22 00:53:22 +05:30
metadata->prev = NULL;
c->partial_slabs = metadata;
void *slab = get_slab(c, slab_size, metadata);
size_t slot = get_free_slot(&c->rng, slots, metadata);
set_slot(metadata, slot);
void *p = slot_pointer(size, slab, slot);
if (requested_size) {
write_after_free_check(p, size - canary_size);
set_canary(metadata, p, size);
}
#if STATS
c->allocated += size;
2019-04-07 08:36:47 +05:30
c->nmalloc++;
#endif
mutex_unlock(&c->lock);
return p;
}
if (c->free_slabs_head != NULL) {
2018-09-01 08:12:34 +05:30
struct slab_metadata *metadata = c->free_slabs_head;
2018-10-31 04:52:56 +05:30
metadata->canary_value = get_random_canary(&c->rng);
void *slab = get_slab(c, slab_size, metadata);
if (requested_size && memory_protect_rw(slab, slab_size)) {
mutex_unlock(&c->lock);
return NULL;
}
#if STATS
c->slab_allocated += slab_size;
#endif
2018-09-01 08:12:34 +05:30
c->free_slabs_head = c->free_slabs_head->next;
if (c->free_slabs_head == NULL) {
c->free_slabs_tail = NULL;
}
metadata->next = NULL;
metadata->prev = NULL;
c->partial_slabs = metadata;
size_t slot = get_free_slot(&c->rng, slots, metadata);
set_slot(metadata, slot);
void *p = slot_pointer(size, slab, slot);
if (requested_size) {
set_canary(metadata, p, size);
}
#if STATS
c->allocated += size;
2019-04-07 08:36:47 +05:30
c->nmalloc++;
#endif
mutex_unlock(&c->lock);
return p;
2018-08-22 00:53:22 +05:30
}
struct slab_metadata *metadata = alloc_metadata(c, slab_size, requested_size);
if (unlikely(metadata == NULL)) {
mutex_unlock(&c->lock);
2018-08-22 00:53:22 +05:30
return NULL;
}
#if STATS
c->slab_allocated += slab_size;
#endif
2018-10-31 04:52:56 +05:30
metadata->canary_value = get_random_canary(&c->rng);
2018-08-22 00:53:22 +05:30
c->partial_slabs = metadata;
void *slab = get_slab(c, slab_size, metadata);
size_t slot = get_free_slot(&c->rng, slots, metadata);
set_slot(metadata, slot);
void *p = slot_pointer(size, slab, slot);
if (requested_size) {
set_canary(metadata, p, size);
}
2018-08-22 00:53:22 +05:30
#if STATS
c->allocated += size;
2019-04-07 08:36:47 +05:30
c->nmalloc++;
#endif
mutex_unlock(&c->lock);
return p;
2018-08-22 00:53:22 +05:30
}
struct slab_metadata *metadata = c->partial_slabs;
2018-08-25 12:32:39 +05:30
size_t slot = get_free_slot(&c->rng, slots, metadata);
2018-08-22 00:53:22 +05:30
set_slot(metadata, slot);
if (!has_free_slots(slots, metadata)) {
c->partial_slabs = c->partial_slabs->next;
if (c->partial_slabs) {
c->partial_slabs->prev = NULL;
}
}
void *slab = get_slab(c, slab_size, metadata);
void *p = slot_pointer(size, slab, slot);
if (requested_size) {
write_after_free_check(p, size - canary_size);
set_canary(metadata, p, size);
}
2018-08-22 00:53:22 +05:30
#if STATS
c->allocated += size;
2019-04-07 08:36:47 +05:30
c->nmalloc++;
#endif
mutex_unlock(&c->lock);
2018-08-22 00:53:22 +05:30
return p;
}
2019-03-26 00:19:50 +05:30
struct slab_size_class_info {
unsigned arena;
size_t class;
};
static struct slab_size_class_info slab_size_class(const void *p) {
2018-12-05 13:57:09 +05:30
size_t offset = (const char *)p - (const char *)ro.slab_region_start;
2019-03-26 00:19:50 +05:30
unsigned arena = 0;
if (N_ARENA > 1) {
arena = offset / ARENA_SIZE;
offset -= arena * ARENA_SIZE;
}
return (struct slab_size_class_info){arena, offset / REAL_CLASS_REGION_SIZE};
2018-08-22 00:53:22 +05:30
}
2018-12-05 13:57:09 +05:30
static size_t slab_usable_size(const void *p) {
2019-03-26 00:19:50 +05:30
return size_classes[slab_size_class(p).class];
2018-08-22 00:53:22 +05:30
}
2018-09-01 08:12:34 +05:30
static void enqueue_free_slab(struct size_class *c, struct slab_metadata *metadata) {
metadata->next = NULL;
static_assert(FREE_SLABS_QUARANTINE_RANDOM_LENGTH < (u16)-1, "free slabs quarantine too large");
size_t index = get_random_u16_uniform(&c->rng, FREE_SLABS_QUARANTINE_RANDOM_LENGTH);
2018-10-13 23:25:44 +05:30
struct slab_metadata *substitute = c->free_slabs_quarantine[index];
c->free_slabs_quarantine[index] = metadata;
if (substitute == NULL) {
return;
}
2018-09-01 08:12:34 +05:30
if (c->free_slabs_tail != NULL) {
2018-10-13 23:25:44 +05:30
c->free_slabs_tail->next = substitute;
2018-09-01 08:12:34 +05:30
} else {
2018-10-13 23:25:44 +05:30
c->free_slabs_head = substitute;
2018-09-01 08:12:34 +05:30
}
2018-10-13 23:25:44 +05:30
c->free_slabs_tail = substitute;
2018-09-01 08:12:34 +05:30
}
static inline void deallocate_small(void *p, const size_t *expected_size) {
2019-03-26 00:19:50 +05:30
struct slab_size_class_info size_class_info = slab_size_class(p);
size_t class = size_class_info.class;
2018-08-22 00:53:22 +05:30
2019-03-26 00:19:50 +05:30
struct size_class *c = &ro.size_class_metadata[size_class_info.arena][class];
2018-08-22 00:53:22 +05:30
size_t size = size_classes[class];
2018-09-19 02:58:52 +05:30
if (expected_size && size != *expected_size) {
fatal_error("sized deallocation mismatch (small)");
2018-09-19 02:58:52 +05:30
}
bool is_zero_size = size == 0;
if (is_zero_size) {
size = 16;
}
2018-08-22 00:53:22 +05:30
size_t slots = size_class_slots[class];
size_t slab_size = get_slab_size(slots, size);
mutex_lock(&c->lock);
#if STATS
c->allocated -= size;
2019-04-07 08:36:47 +05:30
c->ndalloc++;
#endif
2018-08-22 00:53:22 +05:30
struct slab_metadata *metadata = get_metadata(c, p);
void *slab = get_slab(c, slab_size, metadata);
size_t slot = libdivide_u32_do((char *)p - (char *)slab, &c->size_divisor);
if (slot_pointer(size, slab, slot) != p) {
fatal_error("invalid unaligned free");
}
2018-08-24 14:21:52 +05:30
if (!get_slot(metadata, slot)) {
fatal_error("double free");
}
if (!is_zero_size) {
if (canary_size) {
u64 canary_value;
memcpy(&canary_value, (char *)p + size - canary_size, canary_size);
if (unlikely(canary_value != metadata->canary_value)) {
fatal_error("canary corrupted");
}
}
if (ZERO_ON_FREE) {
memset(p, 0, size - canary_size);
}
}
2018-11-06 02:49:50 +05:30
#if SLAB_QUARANTINE
if (get_quarantine(metadata, slot)) {
fatal_error("double free (quarantine)");
}
set_quarantine(metadata, slot);
2019-01-03 00:52:28 +05:30
size_t quarantine_shift = __builtin_clzl(size) - (63 - MAX_SLAB_SIZE_CLASS_SHIFT);
#if SLAB_QUARANTINE_RANDOM_LENGTH > 0
2019-01-03 00:52:28 +05:30
size_t slab_quarantine_random_length = SLAB_QUARANTINE_RANDOM_LENGTH << quarantine_shift;
size_t random_index = get_random_u16_uniform(&c->rng, slab_quarantine_random_length);
void *random_substitute = c->quarantine_random[random_index];
2018-11-06 02:49:50 +05:30
c->quarantine_random[random_index] = p;
if (random_substitute == NULL) {
2018-11-06 02:49:50 +05:30
mutex_unlock(&c->lock);
return;
}
p = random_substitute;
2018-11-06 02:49:50 +05:30
#endif
#if SLAB_QUARANTINE_QUEUE_LENGTH > 0
2019-01-03 00:52:28 +05:30
size_t slab_quarantine_queue_length = SLAB_QUARANTINE_QUEUE_LENGTH << quarantine_shift;
void *queue_substitute = c->quarantine_queue[c->quarantine_queue_index];
2018-11-06 02:49:50 +05:30
c->quarantine_queue[c->quarantine_queue_index] = p;
2019-01-03 00:52:28 +05:30
c->quarantine_queue_index = (c->quarantine_queue_index + 1) % slab_quarantine_queue_length;
2018-11-06 02:49:50 +05:30
if (queue_substitute == NULL) {
2018-11-06 02:49:50 +05:30
mutex_unlock(&c->lock);
return;
}
p = queue_substitute;
2018-11-06 02:49:50 +05:30
#endif
metadata = get_metadata(c, p);
slab = get_slab(c, slab_size, metadata);
slot = libdivide_u32_do((char *)p - (char *)slab, &c->size_divisor);
clear_quarantine(metadata, slot);
#endif
2018-08-22 00:53:22 +05:30
if (!has_free_slots(slots, metadata)) {
metadata->next = c->partial_slabs;
metadata->prev = NULL;
if (c->partial_slabs) {
c->partial_slabs->prev = metadata;
}
c->partial_slabs = metadata;
}
clear_slot(metadata, slot);
if (is_free_slab(metadata)) {
if (metadata->prev) {
metadata->prev->next = metadata->next;
} else {
c->partial_slabs = metadata->next;
}
if (metadata->next) {
metadata->next->prev = metadata->prev;
}
metadata->prev = NULL;
if (c->empty_slabs_total + slab_size > max_empty_slabs_total) {
if (!memory_map_fixed(slab, slab_size)) {
2019-03-24 08:29:59 +05:30
memory_set_name(slab, slab_size, size_class_labels[class]);
#if STATS
c->slab_allocated -= slab_size;
#endif
2018-09-01 08:12:34 +05:30
enqueue_free_slab(c, metadata);
mutex_unlock(&c->lock);
return;
}
// handle out-of-memory by just putting it into the empty slabs list
}
metadata->next = c->empty_slabs;
2018-08-30 14:25:39 +05:30
c->empty_slabs = metadata;
c->empty_slabs_total += slab_size;
2018-08-22 00:53:22 +05:30
}
mutex_unlock(&c->lock);
2018-08-22 00:53:22 +05:30
}
2018-10-15 07:24:23 +05:30
struct region_metadata {
2018-08-22 00:53:22 +05:30
void *p;
size_t size;
2018-08-26 14:12:01 +05:30
size_t guard_size;
2018-08-22 00:53:22 +05:30
};
struct quarantine_info {
void *p;
size_t size;
};
#define INITIAL_REGION_TABLE_SIZE 256
#define MAX_REGION_TABLE_SIZE (CLASS_REGION_SIZE / PAGE_SIZE / sizeof(struct region_metadata))
2018-08-22 00:53:22 +05:30
struct region_allocator {
struct mutex lock;
2018-10-15 07:24:23 +05:30
struct region_metadata *regions;
size_t total;
size_t free;
#if STATS
size_t allocated;
#endif
struct quarantine_info quarantine_random[REGION_QUARANTINE_RANDOM_LENGTH];
struct quarantine_info quarantine_queue[REGION_QUARANTINE_QUEUE_LENGTH];
size_t quarantine_queue_index;
struct random_state rng;
};
2018-10-09 01:20:31 +05:30
2018-10-19 01:31:04 +05:30
struct __attribute__((aligned(PAGE_SIZE))) slab_info_mapping {
struct slab_metadata slab_info[MAX_METADATA_MAX];
};
2018-10-16 04:38:20 +05:30
struct __attribute__((aligned(PAGE_SIZE))) allocator_state {
2019-03-26 00:19:50 +05:30
struct size_class size_class_metadata[N_ARENA][N_SIZE_CLASSES];
struct region_allocator region_allocator;
2018-10-16 04:38:20 +05:30
// padding until next page boundary for mprotect
struct region_metadata regions_a[MAX_REGION_TABLE_SIZE] __attribute__((aligned(PAGE_SIZE)));
2018-10-16 04:38:20 +05:30
// padding until next page boundary for mprotect
struct region_metadata regions_b[MAX_REGION_TABLE_SIZE] __attribute__((aligned(PAGE_SIZE)));
2018-10-16 04:38:20 +05:30
// padding until next page boundary for mprotect
2019-03-26 00:19:50 +05:30
struct slab_info_mapping slab_info_mapping[N_ARENA][N_SIZE_CLASSES];
2018-10-19 01:31:04 +05:30
// padding until next page boundary for mprotect
};
2018-10-09 01:20:31 +05:30
static void regions_quarantine_deallocate_pages(void *p, size_t size, size_t guard_size) {
if (size >= REGION_QUARANTINE_SKIP_THRESHOLD) {
deallocate_pages(p, size, guard_size);
return;
}
if (unlikely(memory_map_fixed(p, size))) {
deallocate_pages(p, size, guard_size);
return;
}
memory_set_name(p, size, "malloc large");
2018-10-09 01:20:31 +05:30
struct quarantine_info a =
(struct quarantine_info){(char *)p - guard_size, size + guard_size * 2};
struct region_allocator *ra = ro.region_allocator;
mutex_lock(&ra->lock);
size_t index = get_random_u64_uniform(&ra->rng, REGION_QUARANTINE_RANDOM_LENGTH);
struct quarantine_info b = ra->quarantine_random[index];
ra->quarantine_random[index] = a;
if (b.p == NULL) {
mutex_unlock(&ra->lock);
return;
}
a = ra->quarantine_queue[ra->quarantine_queue_index];
ra->quarantine_queue[ra->quarantine_queue_index] = b;
ra->quarantine_queue_index = (ra->quarantine_queue_index + 1) % REGION_QUARANTINE_QUEUE_LENGTH;
mutex_unlock(&ra->lock);
if (a.p != NULL) {
memory_unmap(a.p, a.size);
}
2018-10-09 01:20:31 +05:30
}
2018-08-22 00:53:22 +05:30
static int regions_grow(void) {
struct region_allocator *ra = ro.region_allocator;
2018-10-15 07:24:23 +05:30
if (ra->total > SIZE_MAX / sizeof(struct region_metadata) / 2) {
2018-08-22 00:53:22 +05:30
return 1;
}
size_t newtotal = ra->total * 2;
2018-10-15 07:24:23 +05:30
size_t newsize = newtotal * sizeof(struct region_metadata);
2018-08-22 00:53:22 +05:30
size_t mask = newtotal - 1;
if (newtotal > MAX_REGION_TABLE_SIZE) {
return 1;
}
2018-10-15 07:24:23 +05:30
struct region_metadata *p = ra->regions == ro.regions[0] ?
ro.regions[1] : ro.regions[0];
if (memory_protect_rw_metadata(p, newsize)) {
2018-08-22 00:53:22 +05:30
return 1;
}
for (size_t i = 0; i < ra->total; i++) {
void *q = ra->regions[i].p;
2018-08-22 00:53:22 +05:30
if (q != NULL) {
size_t index = hash_page(q) & mask;
while (p[index].p != NULL) {
index = (index - 1) & mask;
}
p[index] = ra->regions[i];
2018-08-22 00:53:22 +05:30
}
}
2018-10-15 07:24:23 +05:30
memory_map_fixed(ra->regions, ra->total * sizeof(struct region_metadata));
2019-03-21 00:48:34 +05:30
memory_set_name(ra->regions, ra->total * sizeof(struct region_metadata), "malloc allocator_state");
ra->free = ra->free + ra->total;
ra->total = newtotal;
ra->regions = p;
2018-08-22 00:53:22 +05:30
return 0;
}
2018-08-26 14:12:01 +05:30
static int regions_insert(void *p, size_t size, size_t guard_size) {
struct region_allocator *ra = ro.region_allocator;
if (ra->free * 4 < ra->total) {
2018-08-22 00:53:22 +05:30
if (regions_grow()) {
return 1;
}
}
size_t mask = ra->total - 1;
2018-08-22 00:53:22 +05:30
size_t index = hash_page(p) & mask;
void *q = ra->regions[index].p;
2018-08-22 00:53:22 +05:30
while (q != NULL) {
index = (index - 1) & mask;
q = ra->regions[index].p;
2018-08-22 00:53:22 +05:30
}
ra->regions[index].p = p;
ra->regions[index].size = size;
ra->regions[index].guard_size = guard_size;
ra->free--;
2018-08-22 00:53:22 +05:30
return 0;
}
2018-12-05 13:57:09 +05:30
static struct region_metadata *regions_find(const void *p) {
struct region_allocator *ra = ro.region_allocator;
size_t mask = ra->total - 1;
2018-08-22 00:53:22 +05:30
size_t index = hash_page(p) & mask;
void *r = ra->regions[index].p;
2018-08-22 00:53:22 +05:30
while (r != p && r != NULL) {
index = (index - 1) & mask;
r = ra->regions[index].p;
2018-08-22 00:53:22 +05:30
}
return (r == p && r != NULL) ? &ra->regions[index] : NULL;
2018-08-22 00:53:22 +05:30
}
2018-10-15 07:24:23 +05:30
static void regions_delete(struct region_metadata *region) {
struct region_allocator *ra = ro.region_allocator;
2018-08-22 00:53:22 +05:30
size_t mask = ra->total - 1;
2018-08-22 00:53:22 +05:30
ra->free++;
size_t i = region - ra->regions;
2018-08-22 00:53:22 +05:30
for (;;) {
ra->regions[i].p = NULL;
ra->regions[i].size = 0;
2018-08-22 00:53:22 +05:30
size_t j = i;
for (;;) {
i = (i - 1) & mask;
if (ra->regions[i].p == NULL) {
2018-08-22 00:53:22 +05:30
return;
}
size_t r = hash_page(ra->regions[i].p) & mask;
2018-08-22 00:53:22 +05:30
if ((i <= r && r < j) || (r < j && j < i) || (j < i && i <= r)) {
continue;
}
ra->regions[j] = ra->regions[i];
2018-08-22 00:53:22 +05:30
break;
}
}
}
2019-04-07 10:20:26 +05:30
int get_metadata_key(void) {
#ifdef USE_PKEY
return ro.metadata_pkey;
#else
return -1;
#endif
}
#ifdef USE_PKEY
static inline void thread_set_metadata_access(unsigned access) {
if (ro.metadata_pkey == -1) {
return;
}
pkey_set(ro.metadata_pkey, access);
}
#endif
static inline void thread_unseal_metadata(void) {
#ifdef USE_PKEY
thread_set_metadata_access(0);
#endif
}
static inline void thread_seal_metadata(void) {
#ifdef USE_PKEY
thread_set_metadata_access(PKEY_DISABLE_ACCESS);
#endif
}
static void full_lock(void) {
thread_unseal_metadata();
mutex_lock(&ro.region_allocator->lock);
2019-03-26 00:19:50 +05:30
for (unsigned arena = 0; arena < N_ARENA; arena++) {
for (unsigned class = 0; class < N_SIZE_CLASSES; class++) {
mutex_lock(&ro.size_class_metadata[arena][class].lock);
}
}
thread_seal_metadata();
}
static void full_unlock(void) {
thread_unseal_metadata();
mutex_unlock(&ro.region_allocator->lock);
2019-03-26 00:19:50 +05:30
for (unsigned arena = 0; arena < N_ARENA; arena++) {
for (unsigned class = 0; class < N_SIZE_CLASSES; class++) {
mutex_unlock(&ro.size_class_metadata[arena][class].lock);
}
}
thread_seal_metadata();
}
static void post_fork_child(void) {
thread_unseal_metadata();
#ifdef USE_PKEY
if (!ro.pkey_state_preserved_on_fork) {
// disable sealing to work around kernel bug causing fork to lose the pkey setup
memory_protect_rw(&ro, sizeof(ro));
ro.metadata_pkey = -1;
memory_protect_ro(&ro, sizeof(ro));
}
#endif
mutex_init(&ro.region_allocator->lock);
random_state_init(&ro.region_allocator->rng);
2019-03-26 00:19:50 +05:30
for (unsigned arena = 0; arena < N_ARENA; arena++) {
for (unsigned class = 0; class < N_SIZE_CLASSES; class++) {
struct size_class *c = &ro.size_class_metadata[arena][class];
mutex_init(&c->lock);
random_state_init(&c->rng);
}
}
thread_seal_metadata();
}
2018-09-03 05:28:57 +05:30
static inline bool is_init(void) {
2018-11-05 12:49:04 +05:30
return get_slab_region_start() != NULL;
2018-09-03 05:28:57 +05:30
}
static inline void enforce_init(void) {
if (!is_init()) {
fatal_error("invalid uninitialized allocator usage");
}
}
2018-08-22 00:53:22 +05:30
COLD static void init_slow_path(void) {
static struct mutex lock = MUTEX_INITIALIZER;
2018-08-22 00:53:22 +05:30
mutex_lock(&lock);
2018-08-22 00:53:22 +05:30
2018-09-03 05:28:57 +05:30
if (is_init()) {
mutex_unlock(&lock);
2018-08-22 00:53:22 +05:30
return;
}
#ifdef USE_PKEY
ro.metadata_pkey = pkey_alloc(0, 0);
// pkey state is not preserved on fork before Linux 5.0 unless the patch was backported
struct utsname uts;
if (uname(&uts) == 0) {
unsigned long version = strtoul(uts.release, NULL, 10);
if (version >= 5) {
ro.pkey_state_preserved_on_fork = true;
}
}
#endif
2018-08-29 20:29:37 +05:30
if (sysconf(_SC_PAGESIZE) != PAGE_SIZE) {
fatal_error("page size mismatch");
}
2019-03-23 08:47:26 +05:30
struct random_state *rng = allocate_pages(sizeof(struct random_state), PAGE_SIZE, true, "malloc init rng");
if (rng == NULL) {
fatal_error("failed to allocate init rng");
}
random_state_init(rng);
size_t metadata_guard_size =
(get_random_u64_uniform(rng, REAL_CLASS_REGION_SIZE / PAGE_SIZE) + 1) * PAGE_SIZE;
struct allocator_state *allocator_state =
2019-03-23 08:47:26 +05:30
allocate_pages(sizeof(struct allocator_state), metadata_guard_size, false, "malloc allocator_state");
if (allocator_state == NULL) {
fatal_error("failed to reserve allocator state");
}
if (memory_protect_rw_metadata(allocator_state, offsetof(struct allocator_state, regions_a))) {
fatal_error("failed to unprotect allocator state");
}
ro.region_allocator = &allocator_state->region_allocator;
struct region_allocator *ra = ro.region_allocator;
mutex_init(&ra->lock);
random_state_init(&ra->rng);
ro.regions[0] = allocator_state->regions_a;
ro.regions[1] = allocator_state->regions_b;
ra->regions = ro.regions[0];
ra->total = INITIAL_REGION_TABLE_SIZE;
ra->free = INITIAL_REGION_TABLE_SIZE;
if (memory_protect_rw_metadata(ra->regions, ra->total * sizeof(struct region_metadata))) {
fatal_error("failed to unprotect memory for regions table");
2018-08-22 00:53:22 +05:30
}
2018-11-05 12:49:04 +05:30
void *slab_region_start = memory_map(slab_region_size);
if (slab_region_start == NULL) {
2018-08-22 00:53:22 +05:30
fatal_error("failed to allocate slab region");
}
2018-11-05 12:49:04 +05:30
ro.slab_region_end = (char *)slab_region_start + slab_region_size;
2019-03-24 08:24:21 +05:30
memory_set_name(slab_region_start, slab_region_size, "malloc slab region gap");
2018-08-22 00:53:22 +05:30
2019-03-26 00:19:50 +05:30
for (unsigned arena = 0; arena < N_ARENA; arena++) {
ro.size_class_metadata[arena] = allocator_state->size_class_metadata[arena];
for (unsigned class = 0; class < N_SIZE_CLASSES; class++) {
struct size_class *c = &ro.size_class_metadata[arena][class];
2018-08-22 00:53:22 +05:30
2019-03-26 00:19:50 +05:30
mutex_init(&c->lock);
random_state_init(&c->rng);
2018-08-25 09:51:02 +05:30
2019-03-26 00:19:50 +05:30
size_t bound = (REAL_CLASS_REGION_SIZE - CLASS_REGION_SIZE) / PAGE_SIZE - 1;
size_t gap = (get_random_u64_uniform(rng, bound) + 1) * PAGE_SIZE;
c->class_region_start = (char *)slab_region_start + ARENA_SIZE * arena + REAL_CLASS_REGION_SIZE * class + gap;
memory_set_name(c->class_region_start, CLASS_REGION_SIZE, size_class_labels[class]);
2018-08-22 00:53:22 +05:30
2019-03-26 00:19:50 +05:30
size_t size = size_classes[class];
if (size == 0) {
size = 16;
}
c->size_divisor = libdivide_u32_gen(size);
size_t slab_size = get_slab_size(size_class_slots[class], size);
c->slab_size_divisor = libdivide_u64_gen(slab_size);
c->slab_info = allocator_state->slab_info_mapping[arena][class].slab_info;
}
2018-08-22 00:53:22 +05:30
}
deallocate_pages(rng, sizeof(struct random_state), PAGE_SIZE);
2018-11-05 12:49:04 +05:30
atomic_store_explicit(&ro.slab_region_start, slab_region_start, memory_order_release);
2018-08-22 00:53:22 +05:30
if (memory_protect_ro(&ro, sizeof(ro))) {
2018-08-22 00:53:22 +05:30
fatal_error("failed to protect allocator data");
}
memory_set_name(&ro, sizeof(ro), "malloc read-only after init");
2018-08-22 00:53:22 +05:30
mutex_unlock(&lock);
// may allocate, so wait until the allocator is initialized to avoid deadlocking
if (atfork(full_lock, full_unlock, post_fork_child)) {
fatal_error("pthread_atfork failed");
}
2018-08-22 00:53:22 +05:30
}
static inline void init(void) {
2018-09-11 23:42:27 +05:30
if (unlikely(!is_init())) {
init_slow_path();
2018-08-22 00:53:22 +05:30
}
}
// trigger early initialization to set up pthread_atfork and protect state as soon as possible
COLD __attribute__((constructor(101))) static void trigger_early_init(void) {
// avoid calling init directly to skip it if this isn't the malloc implementation
h_free(h_malloc(16));
}
2018-08-26 14:12:01 +05:30
static size_t get_guard_size(struct random_state *state, size_t size) {
2018-10-07 00:47:55 +05:30
return (get_random_u64_uniform(state, size / PAGE_SIZE / GUARD_SIZE_DIVISOR) + 1) * PAGE_SIZE;
2018-08-26 14:12:01 +05:30
}
2018-08-22 00:53:22 +05:30
static void *allocate(size_t size) {
if (size <= max_slab_size_class) {
return allocate_small(size);
2018-08-22 00:53:22 +05:30
}
struct region_allocator *ra = ro.region_allocator;
mutex_lock(&ra->lock);
size_t guard_size = get_guard_size(&ra->rng, size);
mutex_unlock(&ra->lock);
2018-08-26 14:12:01 +05:30
2019-03-23 08:47:26 +05:30
void *p = allocate_pages(size, guard_size, true, "malloc large");
2018-08-22 00:53:22 +05:30
if (p == NULL) {
return NULL;
}
mutex_lock(&ra->lock);
2018-08-26 14:12:01 +05:30
if (regions_insert(p, size, guard_size)) {
mutex_unlock(&ra->lock);
2018-08-26 14:12:01 +05:30
deallocate_pages(p, size, guard_size);
2018-08-22 00:53:22 +05:30
return NULL;
}
#if STATS
ra->allocated += size;
#endif
mutex_unlock(&ra->lock);
2018-08-22 00:53:22 +05:30
return p;
}
static void deallocate_large(void *p, const size_t *expected_size) {
enforce_init();
thread_unseal_metadata();
2018-08-22 00:53:22 +05:30
struct region_allocator *ra = ro.region_allocator;
mutex_lock(&ra->lock);
2018-10-15 07:24:23 +05:30
struct region_metadata *region = regions_find(p);
2018-08-22 00:53:22 +05:30
if (region == NULL) {
fatal_error("invalid free");
}
size_t size = region->size;
2018-09-19 02:58:52 +05:30
if (expected_size && size != *expected_size) {
fatal_error("sized deallocation mismatch (large)");
2018-09-19 02:58:52 +05:30
}
2018-08-26 14:12:01 +05:30
size_t guard_size = region->guard_size;
2018-08-22 00:53:22 +05:30
regions_delete(region);
#if STATS
ra->allocated -= size;
#endif
mutex_unlock(&ra->lock);
2018-08-22 00:53:22 +05:30
2018-10-09 01:20:31 +05:30
regions_quarantine_deallocate_pages(p, size, guard_size);
2018-08-22 00:53:22 +05:30
}
static size_t adjust_size_for_canaries(size_t size) {
if (size > 0 && size <= max_slab_size_class) {
return size + canary_size;
}
return size;
}
static int alloc_aligned(void **memptr, size_t alignment, size_t size, size_t min_alignment) {
if ((alignment - 1) & alignment || alignment < min_alignment) {
return EINVAL;
}
if (alignment <= PAGE_SIZE) {
if (size <= max_slab_size_class && alignment > min_align) {
size = get_size_info_align(size, alignment).size;
}
void *p = allocate(size);
if (p == NULL) {
return ENOMEM;
}
*memptr = p;
return 0;
}
struct region_allocator *ra = ro.region_allocator;
mutex_lock(&ra->lock);
size_t guard_size = get_guard_size(&ra->rng, size);
mutex_unlock(&ra->lock);
void *p = allocate_pages_aligned(size, alignment, guard_size, "malloc large");
if (p == NULL) {
return ENOMEM;
}
mutex_lock(&ra->lock);
if (regions_insert(p, size, guard_size)) {
mutex_unlock(&ra->lock);
deallocate_pages(p, size, guard_size);
return ENOMEM;
}
mutex_unlock(&ra->lock);
*memptr = p;
return 0;
}
static void *alloc_aligned_simple(size_t alignment, size_t size) {
void *ptr;
int ret = alloc_aligned(&ptr, alignment, size, 1);
if (ret) {
errno = ret;
return NULL;
}
return ptr;
}
2018-08-22 00:53:22 +05:30
EXPORT void *h_malloc(size_t size) {
init();
thread_unseal_metadata();
size = adjust_size_for_canaries(size);
void *p = allocate(size);
thread_seal_metadata();
return p;
2018-08-22 00:53:22 +05:30
}
EXPORT void *h_calloc(size_t nmemb, size_t size) {
size_t total_size;
if (unlikely(__builtin_mul_overflow(nmemb, size, &total_size))) {
2018-08-22 00:53:22 +05:30
errno = ENOMEM;
return NULL;
}
init();
thread_unseal_metadata();
total_size = adjust_size_for_canaries(total_size);
2018-09-07 10:03:51 +05:30
void *p = allocate(total_size);
thread_seal_metadata();
if (!ZERO_ON_FREE && likely(p != NULL) && total_size && total_size <= max_slab_size_class) {
2018-09-07 10:03:51 +05:30
memset(p, 0, total_size - canary_size);
}
return p;
2018-08-22 00:53:22 +05:30
}
EXPORT void *h_realloc(void *old, size_t size) {
if (old == NULL) {
init();
thread_unseal_metadata();
size = adjust_size_for_canaries(size);
void *p = allocate(size);
thread_seal_metadata();
return p;
2018-08-22 00:53:22 +05:30
}
size = adjust_size_for_canaries(size);
2018-08-22 00:53:22 +05:30
size_t old_size;
2018-11-05 12:49:04 +05:30
if (old >= get_slab_region_start() && old < ro.slab_region_end) {
2018-08-22 00:53:22 +05:30
old_size = slab_usable_size(old);
if (size <= max_slab_size_class && get_size_info(size).size == old_size) {
return old;
}
thread_unseal_metadata();
2018-08-22 00:53:22 +05:30
} else {
enforce_init();
thread_unseal_metadata();
struct region_allocator *ra = ro.region_allocator;
mutex_lock(&ra->lock);
2018-10-15 07:24:23 +05:30
struct region_metadata *region = regions_find(old);
2018-08-22 00:53:22 +05:30
if (region == NULL) {
fatal_error("invalid realloc");
}
old_size = region->size;
size_t old_guard_size = region->guard_size;
2018-08-22 00:53:22 +05:30
if (PAGE_CEILING(old_size) == PAGE_CEILING(size)) {
region->size = size;
mutex_unlock(&ra->lock);
thread_seal_metadata();
2018-08-22 00:53:22 +05:30
return old;
}
mutex_unlock(&ra->lock);
2018-08-28 19:19:18 +05:30
size_t old_rounded_size = PAGE_CEILING(old_size);
size_t rounded_size = PAGE_CEILING(size);
if (size > max_slab_size_class) {
// in-place shrink
if (size < old_size) {
void *new_end = (char *)old + rounded_size;
if (memory_map_fixed(new_end, old_guard_size)) {
thread_seal_metadata();
return NULL;
}
memory_set_name(new_end, old_guard_size, "malloc large");
void *new_guard_end = (char *)new_end + old_guard_size;
2018-10-09 01:20:31 +05:30
regions_quarantine_deallocate_pages(new_guard_end, old_rounded_size - rounded_size, 0);
mutex_lock(&ra->lock);
2018-10-15 07:24:23 +05:30
struct region_metadata *region = regions_find(old);
if (region == NULL) {
fatal_error("invalid realloc");
}
region->size = size;
mutex_unlock(&ra->lock);
thread_seal_metadata();
return old;
}
// in-place growth
void *guard_end = (char *)old + old_rounded_size + old_guard_size;
size_t extra = rounded_size - old_rounded_size;
if (!memory_remap((char *)old + old_rounded_size, old_guard_size, old_guard_size + extra)) {
if (memory_protect_rw((char *)old + old_rounded_size, extra)) {
memory_unmap(guard_end, extra);
} else {
mutex_lock(&ra->lock);
2018-10-15 07:24:23 +05:30
struct region_metadata *region = regions_find(old);
if (region == NULL) {
fatal_error("invalid realloc");
}
region->size = size;
mutex_unlock(&ra->lock);
thread_seal_metadata();
return old;
}
}
2018-10-19 00:55:14 +05:30
size_t copy_size = min(size, old_size);
if (copy_size >= MREMAP_MOVE_THRESHOLD) {
void *new = allocate(size);
if (new == NULL) {
thread_seal_metadata();
return NULL;
}
2018-08-28 19:19:18 +05:30
mutex_lock(&ra->lock);
2018-10-15 07:24:23 +05:30
struct region_metadata *region = regions_find(old);
if (region == NULL) {
fatal_error("invalid realloc");
}
regions_delete(region);
mutex_unlock(&ra->lock);
2018-08-28 19:19:18 +05:30
if (memory_remap_fixed(old, old_size, new, size)) {
memcpy(new, old, copy_size);
deallocate_pages(old, old_size, old_guard_size);
} else {
memory_unmap((char *)old - old_guard_size, old_guard_size);
memory_unmap((char *)old + PAGE_CEILING(old_size), old_guard_size);
}
thread_seal_metadata();
return new;
2018-08-28 19:19:18 +05:30
}
}
2018-08-22 00:53:22 +05:30
}
void *new = allocate(size);
if (new == NULL) {
thread_seal_metadata();
2018-08-22 00:53:22 +05:30
return NULL;
}
2018-10-19 00:55:14 +05:30
size_t copy_size = min(size, old_size);
if (size > 0 && size <= max_slab_size_class) {
copy_size -= canary_size;
}
2018-08-22 00:53:22 +05:30
memcpy(new, old, copy_size);
if (old_size <= max_slab_size_class) {
2018-09-19 02:58:52 +05:30
deallocate_small(old, NULL);
} else {
2018-09-19 02:58:52 +05:30
deallocate_large(old, NULL);
}
thread_seal_metadata();
2018-08-22 00:53:22 +05:30
return new;
}
EXPORT int h_posix_memalign(void **memptr, size_t alignment, size_t size) {
init();
thread_unseal_metadata();
size = adjust_size_for_canaries(size);
int ret = alloc_aligned(memptr, alignment, size, sizeof(void *));
thread_seal_metadata();
return ret;
2018-08-22 00:53:22 +05:30
}
EXPORT void *h_aligned_alloc(size_t alignment, size_t size) {
init();
thread_unseal_metadata();
size = adjust_size_for_canaries(size);
void *p = alloc_aligned_simple(alignment, size);
thread_seal_metadata();
return p;
2018-08-22 00:53:22 +05:30
}
EXPORT void *h_memalign(size_t alignment, size_t size) ALIAS(h_aligned_alloc);
2018-08-22 00:53:22 +05:30
#ifndef __ANDROID__
2018-08-22 00:53:22 +05:30
EXPORT void *h_valloc(size_t size) {
init();
thread_unseal_metadata();
size = adjust_size_for_canaries(size);
void *p = alloc_aligned_simple(PAGE_SIZE, size);
thread_seal_metadata();
return p;
2018-08-22 00:53:22 +05:30
}
EXPORT void *h_pvalloc(size_t size) {
size = PAGE_CEILING(size);
if (!size) {
2018-08-22 00:53:22 +05:30
errno = ENOMEM;
return NULL;
}
init();
thread_unseal_metadata();
size = adjust_size_for_canaries(size);
void *p = alloc_aligned_simple(PAGE_SIZE, size);
thread_seal_metadata();
return p;
2018-08-22 00:53:22 +05:30
}
#endif
2018-08-22 00:53:22 +05:30
EXPORT void h_free(void *p) {
if (p == NULL) {
return;
}
2018-11-05 12:49:04 +05:30
if (p >= get_slab_region_start() && p < ro.slab_region_end) {
thread_unseal_metadata();
2018-09-19 02:58:52 +05:30
deallocate_small(p, NULL);
thread_seal_metadata();
return;
}
2018-09-19 02:58:52 +05:30
deallocate_large(p, NULL);
thread_seal_metadata();
2018-08-22 00:53:22 +05:30
}
#ifdef __GLIBC__
2018-08-30 00:36:49 +05:30
EXPORT void h_cfree(void *ptr) ALIAS(h_free);
#endif
2018-08-22 00:53:22 +05:30
2018-09-19 02:58:52 +05:30
EXPORT void h_free_sized(void *p, size_t expected_size) {
if (p == NULL) {
return;
}
2018-11-05 12:49:04 +05:30
if (p >= get_slab_region_start() && p < ro.slab_region_end) {
thread_unseal_metadata();
2018-09-19 02:58:52 +05:30
expected_size = get_size_info(adjust_size_for_canaries(expected_size)).size;
deallocate_small(p, &expected_size);
thread_seal_metadata();
2018-09-19 02:58:52 +05:30
return;
}
deallocate_large(p, &expected_size);
thread_seal_metadata();
2018-09-19 02:58:52 +05:30
}
2018-12-05 13:57:09 +05:30
EXPORT size_t h_malloc_usable_size(H_MALLOC_USABLE_SIZE_CONST void *p) {
2018-08-22 00:53:22 +05:30
if (p == NULL) {
return 0;
}
2018-11-05 12:49:04 +05:30
if (p >= get_slab_region_start() && p < ro.slab_region_end) {
size_t size = slab_usable_size(p);
return size ? size - canary_size : 0;
2018-08-22 00:53:22 +05:30
}
enforce_init();
thread_unseal_metadata();
struct region_allocator *ra = ro.region_allocator;
mutex_lock(&ra->lock);
2018-10-15 07:24:23 +05:30
struct region_metadata *region = regions_find(p);
2018-08-22 00:53:22 +05:30
if (p == NULL) {
fatal_error("invalid malloc_usable_size");
}
size_t size = region->size;
mutex_unlock(&ra->lock);
2018-08-22 00:53:22 +05:30
thread_seal_metadata();
2018-08-22 00:53:22 +05:30
return size;
}
EXPORT size_t h_malloc_object_size(void *p) {
if (p == NULL) {
return 0;
}
2018-11-05 12:49:04 +05:30
void *slab_region_start = get_slab_region_start();
if (p >= slab_region_start && p < ro.slab_region_end) {
size_t size = slab_usable_size(p);
return size ? size - canary_size : 0;
}
2018-11-05 12:49:04 +05:30
if (unlikely(slab_region_start == NULL)) {
return 0;
}
thread_unseal_metadata();
struct region_allocator *ra = ro.region_allocator;
mutex_lock(&ra->lock);
2018-10-15 07:24:23 +05:30
struct region_metadata *region = regions_find(p);
size_t size = p == NULL ? SIZE_MAX : region->size;
mutex_unlock(&ra->lock);
thread_seal_metadata();
return size;
}
EXPORT size_t h_malloc_object_size_fast(void *p) {
if (p == NULL) {
return 0;
}
2018-11-05 12:49:04 +05:30
void *slab_region_start = get_slab_region_start();
if (p >= slab_region_start && p < ro.slab_region_end) {
size_t size = slab_usable_size(p);
return size ? size - canary_size : 0;
}
2018-11-05 12:49:04 +05:30
if (unlikely(slab_region_start == NULL)) {
return 0;
}
return SIZE_MAX;
}
2018-08-22 00:53:22 +05:30
EXPORT int h_mallopt(UNUSED int param, UNUSED int value) {
#ifdef __ANDROID__
if (param == M_PURGE) {
h_malloc_trim(0);
return 1;
}
#endif
2018-08-22 00:53:22 +05:30
return 0;
}
EXPORT int h_malloc_trim(UNUSED size_t pad) {
2018-09-07 12:23:15 +05:30
if (unlikely(!is_init())) {
2018-08-22 00:53:22 +05:30
return 0;
}
thread_unseal_metadata();
bool is_trimmed = false;
2019-03-26 00:19:50 +05:30
for (unsigned arena = 0; arena < N_ARENA; arena++) {
// skip zero byte size class since there's nothing to change
for (unsigned class = 1; class < N_SIZE_CLASSES; class++) {
struct size_class *c = &ro.size_class_metadata[arena][class];
size_t slab_size = get_slab_size(size_class_slots[class], size_classes[class]);
mutex_lock(&c->lock);
struct slab_metadata *iterator = c->empty_slabs;
while (iterator) {
void *slab = get_slab(c, slab_size, iterator);
if (memory_map_fixed(slab, slab_size)) {
break;
}
memory_set_name(slab, slab_size, size_class_labels[class]);
#if STATS
c->slab_allocated -= slab_size;
#endif
2019-03-26 00:19:50 +05:30
struct slab_metadata *trimmed = iterator;
iterator = iterator->next;
c->empty_slabs_total -= slab_size;
2019-03-26 00:19:50 +05:30
enqueue_free_slab(c, trimmed);
2019-03-26 00:19:50 +05:30
is_trimmed = true;
}
c->empty_slabs = iterator;
mutex_unlock(&c->lock);
}
2018-08-22 00:53:22 +05:30
}
thread_seal_metadata();
return is_trimmed;
2018-08-22 00:53:22 +05:30
}
EXPORT void h_malloc_stats(void) {}
2018-09-07 00:05:08 +05:30
#if defined(__GLIBC__) || defined(__ANDROID__)
2018-08-22 00:53:22 +05:30
EXPORT struct mallinfo h_mallinfo(void) {
struct mallinfo info = {0};
// glibc mallinfo type definition and implementation are both broken
#if STATS && !defined(__GLIBC__)
struct region_allocator *ra = ro.region_allocator;
mutex_lock(&ra->lock);
info.hblkhd += ra->allocated;
info.uordblks += ra->allocated;
mutex_unlock(&ra->lock);
for (unsigned arena = 0; arena < N_ARENA; arena++) {
for (unsigned class = 0; class < N_SIZE_CLASSES; class++) {
struct size_class *c = &ro.size_class_metadata[arena][class];
mutex_lock(&c->lock);
info.hblkhd += c->slab_allocated;
info.uordblks += c->allocated;
mutex_unlock(&c->lock);
}
}
info.fordblks = info.hblkhd - info.uordblks;
info.usmblks = info.hblkhd;
#endif
return info;
2018-08-22 00:53:22 +05:30
}
2018-09-07 00:05:08 +05:30
#endif
2018-08-22 00:53:22 +05:30
#ifdef __GLIBC__
2018-08-22 00:53:22 +05:30
EXPORT int h_malloc_info(UNUSED int options, UNUSED FILE *fp) {
errno = ENOSYS;
return -1;
}
COLD EXPORT void *h_malloc_get_state(void) {
return NULL;
}
COLD EXPORT int h_malloc_set_state(UNUSED void *state) {
return -2;
}
#endif
2018-09-02 14:38:45 +05:30
#ifdef __ANDROID__
EXPORT size_t __mallinfo_narenas(void) {
// Consider region allocator to be an arena with index N_ARENA.
return N_ARENA + 1;
2018-09-02 14:38:45 +05:30
}
EXPORT size_t __mallinfo_nbins(void) {
return N_SIZE_CLASSES;
2018-09-02 14:38:45 +05:30
}
// This internal Android API uses mallinfo in a non-standard way to implement malloc_info:
//
// hblkhd: total mapped memory as usual
// ordblks: large allocations
// uordblks: huge allocations
// fsmblks: small allocations
// (other fields are unused)
2018-09-02 14:38:45 +05:30
EXPORT struct mallinfo __mallinfo_arena_info(UNUSED size_t arena) {
struct mallinfo info = {0};
#if STATS
if (arena < N_ARENA) {
for (unsigned class = 0; class < N_SIZE_CLASSES; class++) {
struct size_class *c = &ro.size_class_metadata[arena][class];
mutex_lock(&c->lock);
info.hblkhd += c->slab_allocated;
info.fsmblks += c->allocated;
mutex_unlock(&c->lock);
}
} else if (arena == N_ARENA) {
struct region_allocator *ra = ro.region_allocator;
mutex_lock(&ra->lock);
info.hblkhd = ra->allocated;
// our large allocations are roughly comparable to jemalloc huge allocations
info.uordblks = ra->allocated;
mutex_unlock(&ra->lock);
}
#endif
return info;
2018-09-02 14:38:45 +05:30
}
// This internal Android API uses mallinfo in a non-standard way to implement malloc_info:
//
// ordblks: total allocated space
2019-04-07 08:36:47 +05:30
// uordblks: nmalloc
// fordblks: ndalloc
// (other fields are unused)
2018-09-02 14:38:45 +05:30
EXPORT struct mallinfo __mallinfo_bin_info(UNUSED size_t arena, UNUSED size_t bin) {
struct mallinfo info = {0};
#if STATS
if (arena < N_ARENA && bin < N_SIZE_CLASSES) {
struct size_class *c = &ro.size_class_metadata[arena][bin];
mutex_lock(&c->lock);
info.ordblks = c->allocated;
2019-04-07 08:36:47 +05:30
info.uordblks = c->nmalloc;
info.fordblks = c->ndalloc;
mutex_unlock(&c->lock);
}
#endif
return info;
2018-09-02 14:38:45 +05:30
}
COLD EXPORT int h_iterate(UNUSED uintptr_t base, UNUSED size_t size,
UNUSED void (*callback)(uintptr_t ptr, size_t size, void *arg),
UNUSED void *arg) {
fatal_error("not implemented");
}
COLD EXPORT void h_malloc_disable(void) {
init();
full_lock();
2018-09-02 14:38:45 +05:30
}
COLD EXPORT void h_malloc_enable(void) {
2018-11-12 16:07:20 +05:30
enforce_init();
full_unlock();
2018-09-02 14:38:45 +05:30
}
#endif