volatile int free_lock[2];
} mal;
+int __malloc_replaced;
+
/* Synchronization tools */
static inline void lock(volatile int *lk)
}
#endif
-void *__expand_heap(size_t *);
-
static struct chunk *expand_heap(size_t n)
{
static int heap_lock[2];
return 1;
}
-static void bin_chunk(struct chunk *);
-
static void trim(struct chunk *self, size_t n)
{
size_t n1 = CHUNK_SIZE(self);
next->psize = n1-n | C_INUSE;
self->csize = n | C_INUSE;
- bin_chunk(split);
+ __bin_chunk(split);
}
void *malloc(size_t n)
}
n *= m;
void *p = malloc(n);
- if (!p || IS_MMAPPED(MEM_TO_CHUNK(p)))
- return p;
- if (n >= PAGE_SIZE)
- n = mal0_clear(p, PAGE_SIZE, n);
+ if (!p) return p;
+ if (!__malloc_replaced) {
+ if (IS_MMAPPED(MEM_TO_CHUNK(p)))
+ return p;
+ if (n >= PAGE_SIZE)
+ n = mal0_clear(p, PAGE_SIZE, n);
+ }
return memset(p, 0, n);
}
return new;
}
-static void bin_chunk(struct chunk *self)
+void __bin_chunk(struct chunk *self)
{
struct chunk *next = NEXT_CHUNK(self);
size_t final_size, new_size, size;
if (IS_MMAPPED(self))
unmap_chunk(self);
else
- bin_chunk(self);
+ __bin_chunk(self);
}
void __malloc_donate(char *start, char *end)
struct chunk *c = MEM_TO_CHUNK(start), *n = MEM_TO_CHUNK(end);
c->psize = n->csize = C_INUSE;
c->csize = n->psize = C_INUSE | (end-start);
- bin_chunk(c);
+ __bin_chunk(c);
}