483 #define WIN32_LEAN_AND_MEAN 486 #define HAVE_MORECORE 0 487 #define LACKS_UNISTD_H 488 #define LACKS_SYS_PARAM_H 489 #define LACKS_SYS_MMAN_H 490 #define LACKS_STRING_H 491 #define LACKS_STRINGS_H 492 #define LACKS_SYS_TYPES_H 493 #define LACKS_ERRNO_H 494 #define MALLOC_FAILURE_ACTION 495 #define MMAP_CLEARS 0 498 #if defined(DARWIN) || defined(_DARWIN) 500 #ifndef HAVE_MORECORE 501 #define HAVE_MORECORE 0 506 #ifndef LACKS_SYS_TYPES_H 507 #include <sys/types.h> 511 #define MAX_SIZE_T (~(size_t)0) 514 #define ONLY_MSPACES 0 523 #ifndef MALLOC_ALIGNMENT 524 #define MALLOC_ALIGNMENT ((size_t)8U) 530 #define ABORT abort() 532 #ifndef ABORT_ON_ASSERT_FAILURE 533 #define ABORT_ON_ASSERT_FAILURE 1 535 #ifndef PROCEED_ON_ERROR 536 #define PROCEED_ON_ERROR 0 548 #define MMAP_CLEARS 1 552 #define HAVE_MREMAP 1 554 #define HAVE_MREMAP 0 557 #ifndef MALLOC_FAILURE_ACTION 558 #define MALLOC_FAILURE_ACTION errno = ENOMEM; 560 #ifndef HAVE_MORECORE 562 #define HAVE_MORECORE 0 564 #define HAVE_MORECORE 1 568 #define MORECORE_CONTIGUOUS 0 571 #define MORECORE sbrk 573 #ifndef MORECORE_CONTIGUOUS 574 #define MORECORE_CONTIGUOUS 1 577 #ifndef DEFAULT_GRANULARITY 578 #if MORECORE_CONTIGUOUS 579 #define DEFAULT_GRANULARITY (0) 581 #define DEFAULT_GRANULARITY ((size_t)64U * (size_t)1024U) 584 #ifndef DEFAULT_TRIM_THRESHOLD 585 #ifndef MORECORE_CANNOT_TRIM 586 #define DEFAULT_TRIM_THRESHOLD ((size_t)2U * (size_t)1024U * (size_t)1024U) 588 #define DEFAULT_TRIM_THRESHOLD MAX_SIZE_T 591 #ifndef DEFAULT_MMAP_THRESHOLD 593 #define DEFAULT_MMAP_THRESHOLD ((size_t)256U * (size_t)1024U) 595 #define DEFAULT_MMAP_THRESHOLD MAX_SIZE_T 598 #ifndef USE_BUILTIN_FFS 599 #define USE_BUILTIN_FFS 0 601 #ifndef USE_DEV_RANDOM 602 #define USE_DEV_RANDOM 0 605 #define NO_MALLINFO 0 607 #ifndef MALLINFO_FIELD_TYPE 608 #define MALLINFO_FIELD_TYPE size_t 618 #define M_TRIM_THRESHOLD (-1) 619 #define M_GRANULARITY (-2) 620 #define M_MMAP_THRESHOLD (-3) 649 #ifdef HAVE_USR_INCLUDE_MALLOC_H 650 #include "/usr/include/malloc.h" 679 #ifndef USE_DL_PREFIX 680 #define dlcalloc calloc 682 #define dlmalloc malloc 683 #define dlmemalign memalign 684 #define dlrealloc realloc 685 #define dlvalloc valloc 686 #define dlpvalloc pvalloc 687 #define dlmallinfo mallinfo 688 #define dlmallopt mallopt 689 #define dlmalloc_trim malloc_trim 690 #define dlmalloc_stats malloc_stats 691 #define dlmalloc_usable_size malloc_usable_size 692 #define dlmalloc_footprint malloc_footprint 693 #define dlmalloc_max_footprint malloc_max_footprint 694 #define dlindependent_calloc independent_calloc 695 #define dlindependent_comalloc independent_comalloc 1038 typedef void *mspace;
1051 mspace create_mspace (
size_t capacity,
int locked);
1059 size_t destroy_mspace (mspace msp);
1070 mspace create_mspace_with_base (
void *base,
size_t capacity,
int locked);
1076 void *mspace_malloc (mspace msp,
size_t bytes);
1086 void mspace_free (mspace msp,
void *mem);
1097 void *mspace_realloc (mspace msp,
void *mem,
size_t newsize);
1103 void *mspace_calloc (mspace msp,
size_t n_elements,
size_t elem_size);
1109 void *mspace_memalign (mspace msp,
size_t alignment,
size_t bytes);
1115 void **mspace_independent_calloc (mspace msp,
size_t n_elements,
size_t elem_size,
void *chunks[]);
1121 void **mspace_independent_comalloc (mspace msp,
size_t n_elements,
size_t sizes[],
void *chunks[]);
1127 size_t mspace_footprint (mspace msp);
1133 size_t mspace_max_footprint (mspace msp);
1141 struct mallinfo mspace_mallinfo (mspace msp);
1148 void mspace_malloc_stats (mspace msp);
1154 int mspace_trim (mspace msp,
size_t pad);
1159 int mspace_mallopt (
int,
int);
1180 #pragma warning( disable : 4146 ) 1185 #ifndef LACKS_ERRNO_H 1191 #ifndef LACKS_STDLIB_H 1195 #if ABORT_ON_ASSERT_FAILURE 1196 #define assert(x) if(!(x)) ABORT 1206 #ifndef LACKS_STRING_H 1210 #ifndef LACKS_STRINGS_H 1211 #include <strings.h> 1215 #ifndef LACKS_SYS_MMAN_H 1216 #include <sys/mman.h> 1218 #ifndef LACKS_FCNTL_H 1223 #ifndef LACKS_UNISTD_H 1226 #if !defined(__FreeBSD__) && !defined(__OpenBSD__) && !defined(__NetBSD__) 1227 extern void *sbrk (ptrdiff_t);
1233 #ifndef malloc_getpagesize 1234 # ifdef _SC_PAGESIZE 1235 # ifndef _SC_PAGE_SIZE 1236 # define _SC_PAGE_SIZE _SC_PAGESIZE 1239 # ifdef _SC_PAGE_SIZE 1240 # define malloc_getpagesize sysconf(_SC_PAGE_SIZE) 1242 # if defined(BSD) || defined(DGUX) || defined(HAVE_GETPAGESIZE) 1243 extern size_t getpagesize ();
1244 # define malloc_getpagesize getpagesize() 1247 # define malloc_getpagesize getpagesize() 1249 # ifndef LACKS_SYS_PARAM_H 1250 # include <sys/param.h> 1252 # ifdef EXEC_PAGESIZE 1253 # define malloc_getpagesize EXEC_PAGESIZE 1257 # define malloc_getpagesize NBPG 1259 # define malloc_getpagesize (NBPG * CLSIZE) 1263 # define malloc_getpagesize NBPC 1266 # define malloc_getpagesize PAGESIZE 1268 # define malloc_getpagesize ((size_t)4096U) 1282 #define SIZE_T_SIZE (sizeof(size_t)) 1283 #define SIZE_T_BITSIZE (sizeof(size_t) << 3) 1287 #define SIZE_T_ZERO ((size_t)0) 1288 #define SIZE_T_ONE ((size_t)1) 1289 #define SIZE_T_TWO ((size_t)2) 1290 #define TWO_SIZE_T_SIZES (SIZE_T_SIZE<<1) 1291 #define FOUR_SIZE_T_SIZES (SIZE_T_SIZE<<2) 1292 #define SIX_SIZE_T_SIZES (FOUR_SIZE_T_SIZES+TWO_SIZE_T_SIZES) 1293 #define HALF_MAX_SIZE_T (MAX_SIZE_T / 2U) 1296 #define CHUNK_ALIGN_MASK (MALLOC_ALIGNMENT - SIZE_T_ONE) 1299 #define is_aligned(A) (((size_t)((A)) & (CHUNK_ALIGN_MASK)) == 0) 1302 #define align_offset(A)\ 1303 ((((size_t)(A) & CHUNK_ALIGN_MASK) == 0)? 0 :\ 1304 ((MALLOC_ALIGNMENT - ((size_t)(A) & CHUNK_ALIGN_MASK)) & CHUNK_ALIGN_MASK)) 1316 #define MFAIL ((void*)(MAX_SIZE_T)) 1317 #define CMFAIL ((char*)(MFAIL)) 1320 #define IS_MMAPPED_BIT (SIZE_T_ZERO) 1321 #define USE_MMAP_BIT (SIZE_T_ZERO) 1322 #define CALL_MMAP(s) MFAIL 1323 #define CALL_MUNMAP(a, s) (-1) 1324 #define DIRECT_MMAP(s) MFAIL 1327 #define IS_MMAPPED_BIT (SIZE_T_ONE) 1328 #define USE_MMAP_BIT (SIZE_T_ONE) 1330 #ifndef USE_MALLOC_INSTEAD 1331 #define USE_MALLOC_INSTEAD 0 1334 #if USE_MALLOC_INSTEAD 1335 #define CALL_MMAP(s) system_malloc(s) 1336 #define CALL_MUNMAP(a,s) system_free(a) 1337 #define DIRECT_MMAP(s) system_malloc(s) 1340 #define CALL_MUNMAP(a, s) munmap((a), (s)) 1341 #define MMAP_PROT (PROT_READ|PROT_WRITE) 1342 #if !defined(MAP_ANONYMOUS) && defined(MAP_ANON) 1343 #define MAP_ANONYMOUS MAP_ANON 1345 #ifdef MAP_ANONYMOUS 1346 #define MMAP_FLAGS (MAP_PRIVATE|MAP_ANONYMOUS) 1347 #define CALL_MMAP(s) mmap(0, (s), MMAP_PROT, MMAP_FLAGS, -1, 0) 1353 #define MMAP_FLAGS (MAP_PRIVATE) 1355 #define CALL_MMAP(s) ((dev_zero_fd < 0) ? \ 1356 (dev_zero_fd = open("/dev/zero", O_RDWR), \ 1357 mmap(0, (s), MMAP_PROT, MMAP_FLAGS, dev_zero_fd, 0)) : \ 1358 mmap(0, (s), MMAP_PROT, MMAP_FLAGS, dev_zero_fd, 0)) 1361 #define DIRECT_MMAP(s) CALL_MMAP(s) 1366 win32mmap (
size_t size)
1368 void *ptr = VirtualAlloc (0, size, MEM_RESERVE | MEM_COMMIT, PAGE_READWRITE);
1369 return (ptr != 0) ? ptr :
MFAIL;
1374 win32direct_mmap (
size_t size)
1376 void *ptr = VirtualAlloc (0, size, MEM_RESERVE | MEM_COMMIT | MEM_TOP_DOWN,
1378 return (ptr != 0) ? ptr :
MFAIL;
1383 win32munmap (
void *ptr,
size_t size)
1385 MEMORY_BASIC_INFORMATION minfo;
1389 if (VirtualQuery (cptr, &minfo,
sizeof (minfo)) == 0)
1391 if (minfo.BaseAddress != cptr || minfo.AllocationBase != cptr ||
1392 minfo.State != MEM_COMMIT || minfo.RegionSize > size)
1394 if (VirtualFree (cptr, 0, MEM_RELEASE) == 0)
1396 cptr += minfo.RegionSize;
1397 size -= minfo.RegionSize;
1402 #define CALL_MMAP(s) win32mmap(s) 1403 #define CALL_MUNMAP(a, s) win32munmap((a), (s)) 1404 #define DIRECT_MMAP(s) win32direct_mmap(s) 1409 #if HAVE_MMAP && HAVE_MREMAP 1410 #define CALL_MREMAP(addr, osz, nsz, mv) mremap((addr), (osz), (nsz), (mv)) 1412 #define CALL_MREMAP(addr, osz, nsz, mv) MFAIL 1416 #define CALL_MORECORE(S) MORECORE(S) 1418 #define CALL_MORECORE(S) MFAIL 1422 #define USE_NONCONTIGUOUS_BIT (4U) 1425 #define EXTERN_BIT (8U) 1448 #include <pthread.h> 1449 #define MLOCK_T pthread_mutex_t 1450 #define INITIAL_LOCK(l) pthread_mutex_init(l, NULL) 1451 #define ACQUIRE_LOCK(l) pthread_mutex_lock(l) 1452 #define RELEASE_LOCK(l) pthread_mutex_unlock(l) 1455 static MLOCK_T morecore_mutex = PTHREAD_MUTEX_INITIALIZER;
1458 static MLOCK_T magic_init_mutex = PTHREAD_MUTEX_INITIALIZER;
1466 #define MLOCK_T long 1468 win32_acquire_lock (MLOCK_T * sl)
1472 #ifdef InterlockedCompareExchangePointer 1473 if (!InterlockedCompareExchange (sl, 1, 0))
1476 if (!InterlockedCompareExchange ((
void **) sl, (
void *) 1, (
void *) 0))
1484 win32_release_lock (MLOCK_T * sl)
1486 InterlockedExchange (sl, 0);
1489 #define INITIAL_LOCK(l) *(l)=0 1490 #define ACQUIRE_LOCK(l) win32_acquire_lock(l) 1491 #define RELEASE_LOCK(l) win32_release_lock(l) 1493 static MLOCK_T morecore_mutex;
1495 static MLOCK_T magic_init_mutex;
1498 #define USE_LOCK_BIT (2U) 1500 #define USE_LOCK_BIT (0U) 1501 #define INITIAL_LOCK(l) 1504 #if USE_LOCKS && HAVE_MORECORE 1505 #define ACQUIRE_MORECORE_LOCK() ACQUIRE_LOCK(&morecore_mutex); 1506 #define RELEASE_MORECORE_LOCK() RELEASE_LOCK(&morecore_mutex); 1508 #define ACQUIRE_MORECORE_LOCK() 1509 #define RELEASE_MORECORE_LOCK() 1513 #define ACQUIRE_MAGIC_INIT_LOCK() ACQUIRE_LOCK(&magic_init_mutex); 1514 #define RELEASE_MAGIC_INIT_LOCK() RELEASE_LOCK(&magic_init_mutex); 1516 #define ACQUIRE_MAGIC_INIT_LOCK() 1517 #define RELEASE_MAGIC_INIT_LOCK() 1675 #define MCHUNK_SIZE (sizeof(mchunk)) 1678 #define CHUNK_OVERHEAD (TWO_SIZE_T_SIZES) 1680 #define CHUNK_OVERHEAD (SIZE_T_SIZE) 1684 #define MMAP_CHUNK_OVERHEAD (TWO_SIZE_T_SIZES) 1686 #define MMAP_FOOT_PAD (FOUR_SIZE_T_SIZES) 1689 #define MIN_CHUNK_SIZE\ 1690 ((MCHUNK_SIZE + CHUNK_ALIGN_MASK) & ~CHUNK_ALIGN_MASK) 1693 #define chunk2mem(p) ((void*)((char*)(p) + TWO_SIZE_T_SIZES)) 1694 #define mem2chunk(mem) ((mchunkptr)((char*)(mem) - TWO_SIZE_T_SIZES)) 1696 #define align_as_chunk(A) (mchunkptr)((A) + align_offset(chunk2mem(A))) 1699 #define MAX_REQUEST ((-MIN_CHUNK_SIZE) << 2) 1700 #define MIN_REQUEST (MIN_CHUNK_SIZE - CHUNK_OVERHEAD - SIZE_T_ONE) 1703 #define pad_request(req) \ 1704 (((req) + CHUNK_OVERHEAD + CHUNK_ALIGN_MASK) & ~CHUNK_ALIGN_MASK) 1707 #define request2size(req) \ 1708 (((req) < MIN_REQUEST)? MIN_CHUNK_SIZE : pad_request(req)) 1721 #define PINUSE_BIT (SIZE_T_ONE) 1722 #define CINUSE_BIT (SIZE_T_TWO) 1723 #define INUSE_BITS (PINUSE_BIT|CINUSE_BIT) 1726 #define FENCEPOST_HEAD (INUSE_BITS|SIZE_T_SIZE) 1729 #define cinuse(p) ((p)->head & CINUSE_BIT) 1730 #define pinuse(p) ((p)->head & PINUSE_BIT) 1731 #define chunksize(p) ((p)->head & ~(INUSE_BITS)) 1733 #define clear_pinuse(p) ((p)->head &= ~PINUSE_BIT) 1734 #define clear_cinuse(p) ((p)->head &= ~CINUSE_BIT) 1737 #define chunk_plus_offset(p, s) ((mchunkptr)(((char*)(p)) + (s))) 1738 #define chunk_minus_offset(p, s) ((mchunkptr)(((char*)(p)) - (s))) 1741 #define next_chunk(p) ((mchunkptr)( ((char*)(p)) + ((p)->head & ~INUSE_BITS))) 1742 #define prev_chunk(p) ((mchunkptr)( ((char*)(p)) - ((p)->prev_foot) )) 1745 #define next_pinuse(p) ((next_chunk(p)->head) & PINUSE_BIT) 1748 #define get_foot(p, s) (((mchunkptr)((char*)(p) + (s)))->prev_foot) 1749 #define set_foot(p, s) (((mchunkptr)((char*)(p) + (s)))->prev_foot = (s)) 1752 #define set_size_and_pinuse_of_free_chunk(p, s)\ 1753 ((p)->head = (s|PINUSE_BIT), set_foot(p, s)) 1756 #define set_free_with_pinuse(p, s, n)\ 1757 (clear_pinuse(n), set_size_and_pinuse_of_free_chunk(p, s)) 1759 #define is_mmapped(p)\ 1760 (!((p)->head & PINUSE_BIT) && ((p)->prev_foot & IS_MMAPPED_BIT)) 1763 #define overhead_for(p)\ 1764 (is_mmapped(p)? MMAP_CHUNK_OVERHEAD : CHUNK_OVERHEAD) 1768 #define calloc_must_clear(p) (!is_mmapped(p)) 1770 #define calloc_must_clear(p) (1) 1882 #define leftmost_child(t) ((t)->child[0] != 0? (t)->child[0] : (t)->child[1]) 1949 #define is_mmapped_segment(S) ((S)->sflags & IS_MMAPPED_BIT) 1950 #define is_extern_segment(S) ((S)->sflags & EXTERN_BIT) 2031 #define NSMALLBINS (32U) 2032 #define NTREEBINS (32U) 2033 #define SMALLBIN_SHIFT (3U) 2034 #define SMALLBIN_WIDTH (SIZE_T_ONE << SMALLBIN_SHIFT) 2035 #define TREEBIN_SHIFT (8U) 2036 #define MIN_LARGE_SIZE (SIZE_T_ONE << TREEBIN_SHIFT) 2037 #define MAX_SMALL_SIZE (MIN_LARGE_SIZE - SIZE_T_ONE) 2038 #define MAX_SMALL_REQUEST (MAX_SMALL_SIZE - CHUNK_ALIGN_MASK - CHUNK_OVERHEAD) 2087 #define is_global(M) ((M) == &_gm_) 2088 #define is_initialized(M) ((M)->top != 0) 2094 #define use_lock(M) ((M)->mflags & USE_LOCK_BIT) 2095 #define enable_lock(M) ((M)->mflags |= USE_LOCK_BIT) 2096 #define disable_lock(M) ((M)->mflags &= ~USE_LOCK_BIT) 2098 #define use_mmap(M) ((M)->mflags & USE_MMAP_BIT) 2099 #define enable_mmap(M) ((M)->mflags |= USE_MMAP_BIT) 2100 #define disable_mmap(M) ((M)->mflags &= ~USE_MMAP_BIT) 2102 #define use_noncontiguous(M) ((M)->mflags & USE_NONCONTIGUOUS_BIT) 2103 #define disable_contiguous(M) ((M)->mflags |= USE_NONCONTIGUOUS_BIT) 2105 #define set_lock(M,L)\ 2106 ((M)->mflags = (L)?\ 2107 ((M)->mflags | USE_LOCK_BIT) :\ 2108 ((M)->mflags & ~USE_LOCK_BIT)) 2111 #define page_align(S)\ 2112 (((S) + (mparams.page_size)) & ~(mparams.page_size - SIZE_T_ONE)) 2115 #define granularity_align(S)\ 2116 (((S) + (mparams.granularity)) & ~(mparams.granularity - SIZE_T_ONE)) 2118 #define is_page_aligned(S)\ 2119 (((size_t)(S) & (mparams.page_size - SIZE_T_ONE)) == 0) 2120 #define is_granularity_aligned(S)\ 2121 (((size_t)(S) & (mparams.granularity - SIZE_T_ONE)) == 0) 2124 #define segment_holds(S, A)\ 2125 ((char*)(A) >= S->base && (char*)(A) < S->base + S->size) 2131 msegmentptr sp = &m->
seg;
2134 if (addr >= sp->
base && addr < sp->base + sp->
size)
2136 if ((sp = sp->
next) == 0)
2145 msegmentptr sp = &m->
seg;
2148 if ((
char *) sp >= ss->
base && (
char *) sp < ss->base + ss->
size)
2150 if ((sp = sp->
next) == 0)
2155 #ifndef MORECORE_CANNOT_TRIM 2156 #define should_trim(M,s) ((s) > (M)->trim_check) 2158 #define should_trim(M,s) (0) 2166 #define TOP_FOOT_SIZE\ 2167 (align_offset(chunk2mem(0))+pad_request(sizeof(struct malloc_segment))+MIN_CHUNK_SIZE) 2181 #define GLOBALLY_INITIALIZE() (mparams.page_size == 0 && init_mparams()) 2183 #define PREACTION(M) ((GLOBALLY_INITIALIZE() || use_lock(M))? ACQUIRE_LOCK(&(M)->mutex) : 0) 2184 #define POSTACTION(M) { if (use_lock(M)) RELEASE_LOCK(&(M)->mutex); } 2188 #define PREACTION(M) (0) 2192 #define POSTACTION(M) 2205 #if PROCEED_ON_ERROR 2208 int malloc_corruption_error_count;
2211 static void reset_on_error (mstate m);
2213 #define CORRUPTION_ERROR_ACTION(m) reset_on_error(m) 2214 #define USAGE_ERROR_ACTION(m, p) 2218 #ifndef CORRUPTION_ERROR_ACTION 2219 #define CORRUPTION_ERROR_ACTION(m) ABORT 2222 #ifndef USAGE_ERROR_ACTION 2223 #define USAGE_ERROR_ACTION(m,p) ABORT 2232 #define check_free_chunk(M,P) 2233 #define check_inuse_chunk(M,P) 2234 #define check_malloced_chunk(M,P,N) 2235 #define check_mmapped_chunk(M,P) 2236 #define check_malloc_state(M) 2237 #define check_top_chunk(M,P) 2240 #define check_free_chunk(M,P) do_check_free_chunk(M,P) 2241 #define check_inuse_chunk(M,P) do_check_inuse_chunk(M,P) 2242 #define check_top_chunk(M,P) do_check_top_chunk(M,P) 2243 #define check_malloced_chunk(M,P,N) do_check_malloced_chunk(M,P,N) 2244 #define check_mmapped_chunk(M,P) do_check_mmapped_chunk(M,P) 2245 #define check_malloc_state(M) do_check_malloc_state(M) 2247 static void do_check_any_chunk (mstate m, mchunkptr
p);
2248 static void do_check_top_chunk (mstate m, mchunkptr
p);
2249 static void do_check_mmapped_chunk (mstate m, mchunkptr
p);
2250 static void do_check_inuse_chunk (mstate m, mchunkptr
p);
2251 static void do_check_free_chunk (mstate m, mchunkptr
p);
2252 static void do_check_malloced_chunk (mstate m,
void *mem,
size_t s);
2253 static void do_check_tree (mstate m, tchunkptr t);
2254 static void do_check_treebin (mstate m,
bindex_t i);
2255 static void do_check_smallbin (mstate m,
bindex_t i);
2256 static void do_check_malloc_state (mstate m);
2257 static int bin_find (mstate m, mchunkptr x);
2258 static size_t traverse_and_check (mstate m);
2263 #define is_small(s) (((s) >> SMALLBIN_SHIFT) < NSMALLBINS) 2264 #define small_index(s) ((s) >> SMALLBIN_SHIFT) 2265 #define small_index2size(i) ((i) << SMALLBIN_SHIFT) 2266 #define MIN_SMALL_INDEX (small_index(MIN_CHUNK_SIZE)) 2269 #define smallbin_at(M, i) ((sbinptr)((char*)&((M)->smallbins[(i)<<1]))) 2270 #define treebin_at(M,i) (&((M)->treebins[i])) 2273 #if defined(__GNUC__) && defined(i386) 2274 #define compute_tree_index(S, I)\ 2276 size_t X = S >> TREEBIN_SHIFT;\ 2279 else if (X > 0xFFFF)\ 2283 __asm__("bsrl %1,%0\n\t" : "=r" (K) : "rm" (X));\ 2284 I = (bindex_t)((K << 1) + ((S >> (K + (TREEBIN_SHIFT-1)) & 1)));\ 2288 #define compute_tree_index(S, I)\ 2290 size_t X = S >> TREEBIN_SHIFT;\ 2293 else if (X > 0xFFFF)\ 2296 unsigned int Y = (unsigned int)X;\ 2297 unsigned int N = ((Y - 0x100) >> 16) & 8;\ 2298 unsigned int K = (((Y <<= N) - 0x1000) >> 16) & 4;\ 2300 N += K = (((Y <<= K) - 0x4000) >> 16) & 2;\ 2301 K = 14 - N + ((Y <<= K) >> 15);\ 2302 I = (K << 1) + ((S >> (K + (TREEBIN_SHIFT-1)) & 1));\ 2308 #define bit_for_tree_index(i) \ 2309 (i == NTREEBINS-1)? (SIZE_T_BITSIZE-1) : (((i) >> 1) + TREEBIN_SHIFT - 2) 2312 #define leftshift_for_tree_index(i) \ 2313 ((i == NTREEBINS-1)? 0 : \ 2314 ((SIZE_T_BITSIZE-SIZE_T_ONE) - (((i) >> 1) + TREEBIN_SHIFT - 2))) 2317 #define minsize_for_tree_index(i) \ 2318 ((SIZE_T_ONE << (((i) >> 1) + TREEBIN_SHIFT)) | \ 2319 (((size_t)((i) & SIZE_T_ONE)) << (((i) >> 1) + TREEBIN_SHIFT - 1))) 2325 #define idx2bit(i) ((binmap_t)(1) << (i)) 2328 #define mark_smallmap(M,i) ((M)->smallmap |= idx2bit(i)) 2329 #define clear_smallmap(M,i) ((M)->smallmap &= ~idx2bit(i)) 2330 #define smallmap_is_marked(M,i) ((M)->smallmap & idx2bit(i)) 2332 #define mark_treemap(M,i) ((M)->treemap |= idx2bit(i)) 2333 #define clear_treemap(M,i) ((M)->treemap &= ~idx2bit(i)) 2334 #define treemap_is_marked(M,i) ((M)->treemap & idx2bit(i)) 2338 #if defined(__GNUC__) && defined(i386) 2339 #define compute_bit2idx(X, I)\ 2342 __asm__("bsfl %1,%0\n\t" : "=r" (J) : "rm" (X));\ 2348 #define compute_bit2idx(X, I) I = ffs(X)-1 2351 #define compute_bit2idx(X, I)\ 2353 unsigned int Y = X - 1;\ 2354 unsigned int K = Y >> (16-4) & 16;\ 2355 unsigned int N = K; Y >>= K;\ 2356 N += K = Y >> (8-3) & 8; Y >>= K;\ 2357 N += K = Y >> (4-2) & 4; Y >>= K;\ 2358 N += K = Y >> (2-1) & 2; Y >>= K;\ 2359 N += K = Y >> (1-0) & 1; Y >>= K;\ 2360 I = (bindex_t)(N + Y);\ 2366 #define least_bit(x) ((x) & -(x)) 2369 #define left_bits(x) ((x<<1) | -(x<<1)) 2372 #define same_or_left_bits(x) ((x) | -(x)) 2405 #define ok_address(M, a) ((char*)(a) >= (M)->least_addr) 2407 #define ok_next(p, n) ((char*)(p) < (char*)(n)) 2409 #define ok_cinuse(p) cinuse(p) 2411 #define ok_pinuse(p) pinuse(p) 2414 #define ok_address(M, a) (1) 2415 #define ok_next(b, n) (1) 2416 #define ok_cinuse(p) (1) 2417 #define ok_pinuse(p) (1) 2420 #if (FOOTERS && !INSECURE) 2422 #define ok_magic(M) ((M)->magic == mparams.magic) 2424 #define ok_magic(M) (1) 2430 #if defined(__GNUC__) && __GNUC__ >= 3 2431 #define RTCHECK(e) __builtin_expect(e, 1) 2433 #define RTCHECK(e) (e) 2436 #define RTCHECK(e) (1) 2443 #define mark_inuse_foot(M,p,s) 2446 #define set_inuse(M,p,s)\ 2447 ((p)->head = (((p)->head & PINUSE_BIT)|s|CINUSE_BIT),\ 2448 ((mchunkptr)(((char*)(p)) + (s)))->head |= PINUSE_BIT) 2451 #define set_inuse_and_pinuse(M,p,s)\ 2452 ((p)->head = (s|PINUSE_BIT|CINUSE_BIT),\ 2453 ((mchunkptr)(((char*)(p)) + (s)))->head |= PINUSE_BIT) 2456 #define set_size_and_pinuse_of_inuse_chunk(M, p, s)\ 2457 ((p)->head = (s|PINUSE_BIT|CINUSE_BIT)) 2462 #define mark_inuse_foot(M,p,s)\ 2463 (((mchunkptr)((char*)(p) + (s)))->prev_foot = ((size_t)(M) ^ mparams.magic)) 2465 #define get_mstate_for(p)\ 2466 ((mstate)(((mchunkptr)((char*)(p) +\ 2467 (chunksize(p))))->prev_foot ^ mparams.magic)) 2469 #define set_inuse(M,p,s)\ 2470 ((p)->head = (((p)->head & PINUSE_BIT)|s|CINUSE_BIT),\ 2471 (((mchunkptr)(((char*)(p)) + (s)))->head |= PINUSE_BIT), \ 2472 mark_inuse_foot(M,p,s)) 2474 #define set_inuse_and_pinuse(M,p,s)\ 2475 ((p)->head = (s|PINUSE_BIT|CINUSE_BIT),\ 2476 (((mchunkptr)(((char*)(p)) + (s)))->head |= PINUSE_BIT),\ 2477 mark_inuse_foot(M,p,s)) 2479 #define set_size_and_pinuse_of_inuse_chunk(M, p, s)\ 2480 ((p)->head = (s|PINUSE_BIT|CINUSE_BIT),\ 2481 mark_inuse_foot(M, p, s)) 2497 #if MORECORE_CONTIGUOUS 2503 #if (FOOTERS && !INSECURE) 2507 unsigned char buf[
sizeof (size_t)];
2509 if ((fd = open (
"/dev/urandom", O_RDONLY)) >= 0 && read (fd, buf,
sizeof (buf)) ==
sizeof (buf))
2511 s = *((
size_t *) buf);
2516 s = (size_t) (time (0) ^ (size_t) 0x55555555U);
2523 s = (size_t) 0x58585858U;
2540 SYSTEM_INFO system_info;
2541 GetSystemInfo (&system_info);
2553 if ((
sizeof (
size_t) !=
sizeof (
char *)) ||
2555 (
sizeof (
int) < 4) ||
2570 size_t val = (size_t) value;
2572 switch (param_number)
2598 do_check_any_chunk (mstate m, mchunkptr
p)
2606 do_check_top_chunk (mstate m, mchunkptr p)
2622 do_check_mmapped_chunk (mstate m, mchunkptr p)
2638 do_check_inuse_chunk (mstate m, mchunkptr p)
2640 do_check_any_chunk (m, p);
2646 do_check_mmapped_chunk (m, p);
2651 do_check_free_chunk (mstate m, mchunkptr p)
2655 do_check_any_chunk (m, p);
2659 if (p != m->
dv && p != m->
top)
2678 do_check_malloced_chunk (mstate m,
void *mem,
size_t s)
2684 do_check_inuse_chunk (m, p);
2695 do_check_tree (mstate m, tchunkptr t)
2710 do_check_any_chunk (m, ((mchunkptr) u));
2711 assert (u->index == tindex);
2719 assert (u->child[0] == 0);
2720 assert (u->child[1] == 0);
2727 assert (u->parent->child[0] == u || u->parent->child[1] == u || *((tbinptr *) (u->parent)) == u);
2728 if (u->child[0] != 0)
2730 assert (u->child[0]->parent == u);
2731 assert (u->child[0] != u);
2732 do_check_tree (m, u->child[0]);
2734 if (u->child[1] != 0)
2736 assert (u->child[1]->parent == u);
2737 assert (u->child[1] != u);
2738 do_check_tree (m, u->child[1]);
2740 if (u->child[0] != 0 && u->child[1] != 0)
2757 int empty = (m->
treemap & (1U <<
i)) == 0;
2761 do_check_tree (m, t);
2766 do_check_smallbin (mstate m,
bindex_t i)
2769 mchunkptr p = b->
bk;
2770 unsigned int empty = (m->
smallmap & (1U <<
i)) == 0;
2775 for (; p != b; p = p->
bk)
2780 do_check_free_chunk (m, p);
2787 do_check_inuse_chunk (m, q);
2794 bin_find (mstate m, mchunkptr x)
2809 while ((p = p->
fd) != b);
2830 if (u == (tchunkptr) x)
2833 while ((u = u->
fd) != t);
2842 traverse_and_check (mstate m)
2847 msegmentptr s = &m->
seg;
2852 mchunkptr lastq = 0;
2859 assert (!bin_find (m, q));
2860 do_check_inuse_chunk (m, q);
2864 assert (q == m->
dv || bin_find (m, q));
2866 do_check_free_chunk (m, q);
2879 do_check_malloc_state (mstate m)
2885 do_check_smallbin (m, i);
2887 do_check_treebin (m, i);
2891 do_check_any_chunk (m, m->
dv);
2899 do_check_top_chunk (m, m->
top);
2905 total = traverse_and_check (m);
2917 struct mallinfo nm = { 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 };
2926 msegmentptr s = &m->
seg;
2970 msegmentptr s = &m->
seg;
2988 fprintf (stderr,
"max system bytes = %10lu\n", (
unsigned long) (maxfp));
2989 fprintf (stderr,
"system bytes = %10lu\n", (
unsigned long) (fp));
2990 fprintf (stderr,
"in use bytes = %10lu\n", (
unsigned long) (used));
3006 #define insert_small_chunk(M, P, S) {\ 3007 bindex_t I = (bindex_t)small_index(S);\ 3008 mchunkptr B = smallbin_at(M, I);\ 3010 assert(S >= MIN_CHUNK_SIZE);\ 3011 if (!smallmap_is_marked(M, I))\ 3012 mark_smallmap(M, I);\ 3013 else if (RTCHECK(ok_address(M, B->fd)))\ 3016 CORRUPTION_ERROR_ACTION(M);\ 3025 #define unlink_small_chunk(M, P, S) {\ 3026 mchunkptr F = P->fd;\ 3027 mchunkptr B = P->bk;\ 3028 bindex_t I = small_index(S);\ 3031 assert(chunksize(P) == small_index2size(I));\ 3033 clear_smallmap(M, I);\ 3034 else if (RTCHECK((F == smallbin_at(M,I) || ok_address(M, F)) &&\ 3035 (B == smallbin_at(M,I) || ok_address(M, B)))) {\ 3040 CORRUPTION_ERROR_ACTION(M);\ 3045 #define unlink_first_small_chunk(M, B, P, I) {\ 3046 mchunkptr F = P->fd;\ 3049 assert(chunksize(P) == small_index2size(I));\ 3051 clear_smallmap(M, I);\ 3052 else if (RTCHECK(ok_address(M, F))) {\ 3057 CORRUPTION_ERROR_ACTION(M);\ 3063 #define replace_dv(M, P, S) {\ 3064 size_t DVS = M->dvsize;\ 3066 mchunkptr DV = M->dv;\ 3067 assert(is_small(DVS));\ 3068 insert_small_chunk(M, DV, DVS);\ 3077 #define insert_large_chunk(M, X, S) {\ 3080 compute_tree_index(S, I);\ 3081 H = treebin_at(M, I);\ 3083 X->child[0] = X->child[1] = 0;\ 3084 if (!treemap_is_marked(M, I)) {\ 3085 mark_treemap(M, I);\ 3087 X->parent = (tchunkptr)H;\ 3092 size_t K = S << leftshift_for_tree_index(I);\ 3094 if (chunksize(T) != S) {\ 3095 tchunkptr* C = &(T->child[(K >> (SIZE_T_BITSIZE-SIZE_T_ONE)) & 1]);\ 3099 else if (RTCHECK(ok_address(M, C))) {\ 3106 CORRUPTION_ERROR_ACTION(M);\ 3111 tchunkptr F = T->fd;\ 3112 if (RTCHECK(ok_address(M, T) && ok_address(M, F))) {\ 3120 CORRUPTION_ERROR_ACTION(M);\ 3145 #define unlink_large_chunk(M, X) {\ 3146 tchunkptr XP = X->parent;\ 3149 tchunkptr F = X->fd;\ 3151 if (RTCHECK(ok_address(M, F))) {\ 3156 CORRUPTION_ERROR_ACTION(M);\ 3161 if (((R = *(RP = &(X->child[1]))) != 0) ||\ 3162 ((R = *(RP = &(X->child[0]))) != 0)) {\ 3164 while ((*(CP = &(R->child[1])) != 0) ||\ 3165 (*(CP = &(R->child[0])) != 0)) {\ 3168 if (RTCHECK(ok_address(M, RP)))\ 3171 CORRUPTION_ERROR_ACTION(M);\ 3176 tbinptr* H = treebin_at(M, X->index);\ 3178 if ((*H = R) == 0) \ 3179 clear_treemap(M, X->index);\ 3181 else if (RTCHECK(ok_address(M, XP))) {\ 3182 if (XP->child[0] == X) \ 3188 CORRUPTION_ERROR_ACTION(M);\ 3190 if (RTCHECK(ok_address(M, R))) {\ 3193 if ((C0 = X->child[0]) != 0) {\ 3194 if (RTCHECK(ok_address(M, C0))) {\ 3199 CORRUPTION_ERROR_ACTION(M);\ 3201 if ((C1 = X->child[1]) != 0) {\ 3202 if (RTCHECK(ok_address(M, C1))) {\ 3207 CORRUPTION_ERROR_ACTION(M);\ 3211 CORRUPTION_ERROR_ACTION(M);\ 3218 #define insert_chunk(M, P, S)\ 3219 if (is_small(S)) insert_small_chunk(M, P, S)\ 3220 else { tchunkptr TP = (tchunkptr)(P); insert_large_chunk(M, TP, S); } 3222 #define unlink_chunk(M, P, S)\ 3223 if (is_small(S)) unlink_small_chunk(M, P, S)\ 3224 else { tchunkptr TP = (tchunkptr)(P); unlink_large_chunk(M, TP); } 3230 #define internal_malloc(m, b) mspace_malloc(m, b) 3231 #define internal_free(m, mem) mspace_free(m,mem); 3234 #define internal_malloc(m, b)\ 3235 (m == gm)? dlmalloc(b) : mspace_malloc(m, b) 3236 #define internal_free(m, mem)\ 3237 if (m == gm) dlfree(mem); else mspace_free(m,mem); 3239 #define internal_malloc(m, b) dlmalloc(b) 3240 #define internal_free(m, mem) dlfree(mem) 3260 #if defined(ENABLE_SEPARATE_MMAP_EVENT_TRACE) 3272 mchunkptr p = (
mchunkptr) (mm + offset);
3279 if (mm < m->least_addr)
3285 #if defined(ENABLE_SEPARATE_MMAP_EVENT_TRACE) 3308 #if defined(ENABLE_SEPARATE_MMAP_EVENT_TRACE) 3320 char *cp = (
char *)
CALL_MREMAP ((
char *) oldp - offset,
3321 oldmmsize, newmmsize, 1);
3324 mchunkptr newp = (
mchunkptr) (cp + offset);
3331 if (cp < m->least_addr)
3370 bin->
fd = bin->
bk = bin;
3374 #if PROCEED_ON_ERROR 3378 reset_on_error (mstate m)
3381 ++malloc_corruption_error_count;
3401 size_t psize = (
char *) oldfirst - (
char *)
p;
3403 size_t qsize = psize - nb;
3406 assert ((
char *) oldfirst > (
char *) q);
3411 if (oldfirst == m->
top)
3413 size_t tsize = m->
topsize += qsize;
3418 else if (oldfirst == m->
dv)
3420 size_t dsize = m->
dvsize += qsize;
3448 char *old_top = (
char *) m->
top;
3450 char *old_end = oldsp->
base + oldsp->size;
3454 char *asp = rawsp + offset;
3459 mchunkptr p = tnext;
3480 if ((
char *) (&(nextp->
head)) < old_end)
3491 size_t psize = csp - old_top;
3552 asize += (
page_align ((
size_t) base) - (size_t) base);
3611 char *mp = (
char *) (
CALL_MMAP (rsize));
3634 size_t ssize = end - br;
3670 msegmentptr sp = &m->
seg;
3671 while (sp != 0 && tbase != sp->
base + sp->
size)
3681 if (tbase < m->least_addr)
3684 while (sp != 0 && sp->
base != tbase + tsize)
3688 char *oldbase = sp->
base;
3698 if (nb < m->topsize)
3700 size_t rsize = m->
topsize -= nb;
3701 mchunkptr p = m->
top;
3721 size_t released = 0;
3722 msegmentptr pred = &m->
seg;
3723 msegmentptr sp = pred->
next;
3726 char *base = sp->
base;
3727 size_t size = sp->
size;
3728 msegmentptr next = sp->
next;
3770 size_t released = 0;
3788 size_t newsize = sp->
size - extra;
3809 if (rel_br !=
CMFAIL && new_br < old_br)
3810 released = old_br - new_br;
3819 sp->
size -= released;
3835 return (released != 0) ? 1 : 0;
3862 if ((rsize = trem) == 0)
3867 if (rt != 0 && rt != t)
3878 if (t == 0 && v == 0)
3902 if (v != 0 && rsize < (
size_t) (m->
dvsize - nb))
3998 else if (oldsize >= nb)
4000 size_t rsize = oldsize - nb;
4010 else if (next == m->
top && oldsize + m->
topsize > nb)
4013 size_t newsize = oldsize + m->
topsize;
4014 size_t newtopsize = newsize - nb;
4047 memcpy (newmem, oldmem, (oc < bytes) ? oc : bytes);
4065 if ((alignment & (alignment -
SIZE_T_ONE)) != 0)
4068 while (a < alignment)
4093 if ((((
size_t) (mem)) % alignment) != 0)
4103 char *br = (
char *)
mem2chunk ((
size_t) (((size_t) (mem + alignment -
SIZE_T_ONE)) & -alignment));
4104 char *pos = ((size_t) (br - (
char *) (
p)) >=
MIN_CHUNK_SIZE) ? br : br + alignment;
4106 size_t leadsize = pos - (
char *) (p);
4107 size_t newsize =
chunksize (p) - leadsize;
4129 size_t remainder_size = size - nb;
4132 set_inuse (m, remainder, remainder_size);
4158 ialloc (mstate m,
size_t n_elements,
size_t * sizes,
int opts,
void *chunks[])
4169 size_t element_size;
4170 size_t contents_size;
4174 size_t remainder_size;
4176 mchunkptr array_chunk;
4184 if (n_elements == 0)
4192 if (n_elements == 0)
4195 array_size =
request2size (n_elements * (
sizeof (
void *)));
4202 contents_size = n_elements * element_size;
4208 for (i = 0; i != n_elements; ++
i)
4212 size = contents_size + array_size;
4236 memset ((
size_t *) mem, 0, remainder_size -
SIZE_T_SIZE - array_size);
4242 size_t array_chunk_size;
4244 array_chunk_size = remainder_size - contents_size;
4245 marray = (
void **) (
chunk2mem (array_chunk));
4247 remainder_size = contents_size;
4254 if (i != n_elements - 1)
4256 if (element_size != 0)
4257 size = element_size;
4260 remainder_size -= size;
4272 if (marray != chunks)
4275 if (element_size != 0)
4277 assert (remainder_size == element_size);
4285 for (i = 0; i != n_elements; ++
i)
4335 smallbits =
gm->smallmap >> idx;
4337 if ((smallbits & 0x3U) != 0)
4340 idx += ~smallbits & 1;
4351 else if (nb >
gm->dvsize)
4400 if (nb <= gm->dvsize)
4402 size_t rsize =
gm->dvsize - nb;
4403 mchunkptr p =
gm->dv;
4413 size_t dvs =
gm->dvsize;
4423 else if (nb < gm->topsize)
4425 size_t rsize =
gm->topsize -= nb;
4426 mchunkptr p =
gm->top;
4459 mstate
fm = get_mstate_for (p);
4480 prevsize &= ~IS_MMAPPED_BIT;
4482 if (
CALL_MUNMAP ((
char *) p - prevsize, psize) == 0)
4513 if (next == fm->
top)
4515 size_t tsize = fm->
topsize += psize;
4527 else if (next == fm->
dv)
4529 size_t dsize = fm->
dvsize += psize;
4570 if (n_elements != 0)
4572 req = n_elements * elem_size;
4573 if (((n_elements | elem_size) & ~(
size_t) 0xffff) && (req / n_elements != elem_size))
4578 memset (mem, 0, req);
4587 #ifdef REALLOC_ZERO_BYTES_FREES 4599 mstate m = get_mstate_for (
mem2chunk (oldmem));
4619 size_t sz = elem_size;
4620 return ialloc (
gm, n_elements, &sz, 3, chunks);
4626 return ialloc (
gm, n_elements, sizes, 0, chunks);
4662 return gm->footprint;
4668 return gm->max_footprint;
4710 init_user_mstate (
char *tbase,
size_t tsize)
4716 memset (m, 0, msize);
4732 create_mspace (
size_t capacity,
int locked)
4742 char *tbase = (
char *) (
CALL_MMAP (tsize));
4745 m = init_user_mstate (tbase, tsize);
4754 create_mspace_with_base (
void *base,
size_t capacity,
int locked)
4762 m = init_user_mstate ((
char *) base, capacity);
4770 destroy_mspace (mspace msp)
4773 mstate ms = (
mstate) msp;
4776 msegmentptr sp = &ms->
seg;
4779 char *base = sp->
base;
4780 size_t size = sp->
size;
4801 mspace_malloc (mspace msp,
size_t bytes)
4803 mstate ms = (
mstate) msp;
4821 if ((smallbits & 0x3U) != 0)
4824 idx += ~smallbits & 1;
4835 else if (nb > ms->
dvsize)
4884 if (nb <= ms->dvsize)
4886 size_t rsize = ms->
dvsize - nb;
4887 mchunkptr p = ms->
dv;
4907 else if (nb < ms->topsize)
4909 size_t rsize = ms->
topsize -= nb;
4910 mchunkptr p = ms->
top;
4931 mspace_free (mspace msp,
void *mem)
4937 mstate
fm = get_mstate_for (p);
4939 mstate fm = (
mstate) msp;
4958 prevsize &= ~IS_MMAPPED_BIT;
4960 #if defined(ENABLE_SEPARATE_MMAP_EVENT_TRACE) 4963 void *ptr = (
char *) p - prevsize;
4969 if (
CALL_MUNMAP ((
char *) p - prevsize, psize) == 0)
5000 if (next == fm->
top)
5002 size_t tsize = fm->
topsize += psize;
5014 else if (next == fm->
dv)
5016 size_t dsize = fm->
dvsize += psize;
5050 mspace_calloc (mspace msp,
size_t n_elements,
size_t elem_size)
5054 mstate ms = (
mstate) msp;
5060 if (n_elements != 0)
5062 req = n_elements * elem_size;
5063 if (((n_elements | elem_size) & ~(
size_t) 0xffff) && (req / n_elements != elem_size))
5068 memset (mem, 0, req);
5073 mspace_realloc (mspace msp,
void *oldmem,
size_t bytes)
5076 return mspace_malloc (msp, bytes);
5077 #ifdef REALLOC_ZERO_BYTES_FREES 5080 mspace_free (msp, oldmem);
5088 mstate ms = get_mstate_for (p);
5090 mstate ms = (
mstate) msp;
5102 mspace_memalign (mspace msp,
size_t alignment,
size_t bytes)
5104 mstate ms = (
mstate) msp;
5114 mspace_independent_calloc (mspace msp,
size_t n_elements,
size_t elem_size,
void *chunks[])
5116 size_t sz = elem_size;
5117 mstate ms = (
mstate) msp;
5123 return ialloc (ms, n_elements, &sz, 3, chunks);
5127 mspace_independent_comalloc (mspace msp,
size_t n_elements,
size_t sizes[],
void *chunks[])
5129 mstate ms = (
mstate) msp;
5135 return ialloc (ms, n_elements, sizes, 0, chunks);
5139 mspace_trim (mspace msp,
size_t pad)
5142 mstate ms = (
mstate) msp;
5159 mspace_malloc_stats (mspace msp)
5161 mstate ms = (
mstate) msp;
5173 mspace_footprint (mspace msp)
5176 mstate ms = (
mstate) msp;
5190 mspace_max_footprint (mspace msp)
5193 mstate ms = (
mstate) msp;
5208 mspace_mallinfo (mspace msp)
5210 mstate ms = (
mstate) msp;
5220 mspace_mallopt (
int param_number,
int value)
#define treemap_is_marked(M, i)
static struct malloc_params mparams
#define MAX_SMALL_REQUEST
#define check_inuse_chunk(M, P)
#define mark_inuse_foot(M, p, s)
#define check_malloced_chunk(M, P, N)
#define dlmalloc_usable_size
static void * sys_alloc(mstate m, size_t nb)
#define set_inuse(M, p, s)
#define CORRUPTION_ERROR_ACTION(m)
MALLINFO_FIELD_TYPE arena
#define granularity_align(S)
MALLINFO_FIELD_TYPE hblks
static msegmentptr segment_holding(mstate m, char *addr)
static void * internal_memalign(mstate m, size_t alignment, size_t bytes)
#define USAGE_ERROR_ACTION(m, p)
#define MORECORE_CONTIGUOUS
#define is_extern_segment(S)
struct malloc_tree_chunk * parent
#define small_index2size(i)
#define chunk_minus_offset(p, s)
#define unlink_first_small_chunk(M, B, P, I)
#define is_page_aligned(S)
static void * tmalloc_small(mstate m, size_t nb)
tbinptr treebins[NTREEBINS]
MALLINFO_FIELD_TYPE ordblks
static int change_mparam(int param_number, int value)
#define ACQUIRE_MAGIC_INIT_LOCK()
#define set_size_and_pinuse_of_free_chunk(p, s)
struct malloc_tree_chunk * tchunkptr
struct malloc_chunk * mchunkptr
#define internal_malloc(m, b)
struct malloc_tree_chunk * fd
static void * tmalloc_large(mstate m, size_t nb)
#define use_noncontiguous(M)
static void ** ialloc(mstate m, size_t n_elements, size_t *sizes, int opts, void *chunks[])
struct malloc_state * mstate
static void add_segment(mstate m, char *tbase, size_t tsize, flag_t mmapped)
#define CALL_MREMAP(addr, osz, nsz, mv)
static mchunkptr mmap_resize(mstate m, mchunkptr oldp, size_t nb)
struct malloc_segment * next
static void mmap_called(void *m, void *ptr, MMAP_TRACE_H *h)
#define dlmalloc_max_footprint
static void * mmap_alloc(mstate m, size_t nb)
#define RELEASE_MORECORE_LOCK()
mchunkptr smallbins[(NSMALLBINS+1)*2]
static void * prepend_alloc(mstate m, char *newbase, char *oldbase, size_t nb)
#define minsize_for_tree_index(i)
#define dlindependent_calloc
#define check_free_chunk(M, P)
#define FOUR_SIZE_T_SIZES
#define USE_NONCONTIGUOUS_BIT
#define dlindependent_comalloc
#define segment_holds(S, A)
#define USE_MALLOC_INSTEAD
#define align_as_chunk(A)
#define malloc_getpagesize
#define chunk_plus_offset(p, s)
static struct mallinfo internal_mallinfo(mstate m)
static int has_segment_link(mstate m, msegmentptr ss)
#define ACQUIRE_MORECORE_LOCK()
struct malloc_segment * msegmentptr
MALLINFO_FIELD_TYPE fordblks
#define compute_tree_index(S, I)
#define CALL_MUNMAP(a, s)
#define leftshift_for_tree_index(i)
static int init_mparams(void)
#define compute_bit2idx(X, I)
static void * internal_realloc(mstate m, void *oldmem, size_t bytes)
struct malloc_tree_chunk * tbinptr
MALLINFO_FIELD_TYPE fsmblks
#define unlink_chunk(M, P, S)
#define internal_free(m, mem)
static void internal_malloc_stats(mstate m)
MALLINFO_FIELD_TYPE keepcost
struct malloc_tree_chunk * bk
#define request2size(req)
#define DEFAULT_GRANULARITY
#define DEFAULT_TRIM_THRESHOLD
MALLINFO_FIELD_TYPE hblkhd
static void init_top(mstate m, mchunkptr p, size_t psize)
static void munmap_is_to_be_called(void *m, void *ptr, MMAP_TRACE_H *h)
#define check_top_chunk(M, P)
struct malloc_tree_chunk * child[2]
#define insert_large_chunk(M, X, S)
#define MALLOC_FAILURE_ACTION
#define MMAP_TRACE_H_SIZE
#define is_initialized(M)
#define smallbin_at(M, i)
#define unlink_large_chunk(M, X)
#define insert_chunk(M, P, S)
static size_t release_unused_segments(mstate m)
#define dlmalloc_footprint
#define set_size_and_pinuse_of_inuse_chunk(M, p, s)
#define set_inuse_and_pinuse(M, p, s)
static struct malloc_state _gm_
#define disable_contiguous(M)
#define smallmap_is_marked(M, i)
#define DEFAULT_MMAP_THRESHOLD
static void init_bins(mstate m)
#define should_trim(M, s)
#define replace_dv(M, P, S)
#define check_mmapped_chunk(M, P)
#define calloc_must_clear(p)
#define is_mmapped_segment(S)
#define leftmost_child(t)
struct malloc_chunk * sbinptr
static int sys_trim(mstate m, size_t pad)
#define set_free_with_pinuse(p, s, n)
#define RELEASE_MAGIC_INIT_LOCK()
#define check_malloc_state(M)
#define MALLINFO_FIELD_TYPE
MALLINFO_FIELD_TYPE uordblks
MALLINFO_FIELD_TYPE smblks
MALLINFO_FIELD_TYPE usmblks