| /redis-3.2.3/deps/jemalloc/test/unit/ |
| H A D | rtree.c | 21 for (i = 1; i <= (sizeof(uintptr_t) << 3); i++) { in TEST_BEGIN() 37 for (i = 1; i <= (sizeof(uintptr_t) << 3); i++) { in TEST_BEGIN() 61 for (i = 1; i < (sizeof(uintptr_t) << 3); i++) { in TEST_BEGIN() 62 uintptr_t keys[] = {0, 1, in TEST_BEGIN() 63 (((uintptr_t)1) << (sizeof(uintptr_t)*8-i)) - 1}; in TEST_BEGIN() 70 for (j = 0; j < sizeof(keys)/sizeof(uintptr_t); j++) { in TEST_BEGIN() 73 for (k = 0; k < sizeof(keys)/sizeof(uintptr_t); k++) { in TEST_BEGIN() 82 (((uintptr_t)1) << (sizeof(uintptr_t)*8-i)), false), in TEST_BEGIN() 102 for (i = 1; i <= (sizeof(uintptr_t) << 3); i++) { in TEST_BEGIN() 103 uintptr_t keys[NSET]; in TEST_BEGIN() [all …]
|
| H A D | tsd.c | 58 data_t d = (data_t)(uintptr_t)arg; in malloc_tsd_externs() 72 assert_x_eq(*data_tsd_get(), (data_t)(uintptr_t)arg, in malloc_tsd_externs()
|
| H A D | rb.c | 36 ret = (((uintptr_t)a) > ((uintptr_t)b)) in node_cmp() 37 - (((uintptr_t)a) < ((uintptr_t)b)); in node_cmp()
|
| /redis-3.2.3/deps/jemalloc/src/ |
| H A D | chunk_dss.c | 117 cpad = (void *)((uintptr_t)dss_max + gap_size); in chunk_alloc_dss() 118 ret = (void *)ALIGNMENT_CEILING((uintptr_t)dss_max, in chunk_alloc_dss() 120 cpad_size = (uintptr_t)ret - (uintptr_t)cpad; in chunk_alloc_dss() 121 dss_next = (void *)((uintptr_t)ret + size); in chunk_alloc_dss() 122 if ((uintptr_t)ret < (uintptr_t)dss_max || in chunk_alloc_dss() 123 (uintptr_t)dss_next < (uintptr_t)dss_max) { in chunk_alloc_dss() 165 if ((uintptr_t)chunk >= (uintptr_t)dss_base in chunk_in_dss() 166 && (uintptr_t)chunk < (uintptr_t)dss_max) in chunk_in_dss()
|
| H A D | extent.c | 30 uintptr_t a_addr = (uintptr_t)extent_node_addr_get(a); in extent_szad_comp() 31 uintptr_t b_addr = (uintptr_t)extent_node_addr_get(b); in extent_szad_comp() 46 uintptr_t a_addr = (uintptr_t)extent_node_addr_get(a); in extent_ad_comp() 47 uintptr_t b_addr = (uintptr_t)extent_node_addr_get(b); in extent_ad_comp()
|
| H A D | huge.c | 154 memset((void *)((uintptr_t)ptr + usize), 0x5a, sdiff); in huge_ralloc_no_move_similar() 177 memset((void *)((uintptr_t)ptr + oldsize), 0, in huge_ralloc_no_move_similar() 181 memset((void *)((uintptr_t)ptr + oldsize), 0xa5, usize - in huge_ralloc_no_move_similar() 212 huge_dalloc_junk((void *)((uintptr_t)ptr + usize), in huge_ralloc_no_move_shrink() 217 CHUNK_ADDR2BASE((uintptr_t)ptr + usize), in huge_ralloc_no_move_shrink() 219 CHUNK_ADDR2OFFSET((uintptr_t)ptr + usize), sdiff); in huge_ralloc_no_move_shrink() 266 memset((void *)((uintptr_t)ptr + oldsize), 0, in huge_ralloc_no_move_expand() 270 memset((void *)((uintptr_t)ptr + in huge_ralloc_no_move_expand() 275 memset((void *)((uintptr_t)ptr + oldsize), 0xa5, usize - in huge_ralloc_no_move_expand() 434 huge_prof_tctx_set(ptr, (prof_tctx_t *)(uintptr_t)1U); in huge_prof_tctx_reset()
|
| H A D | arena.c | 81 uintptr_t a_miscelm = (uintptr_t)a; in arena_run_comp() 82 uintptr_t b_miscelm = (uintptr_t)b; in arena_run_comp() 175 uintptr_t a_miscelm = (uintptr_t)a; in arena_avail_comp() 187 uintptr_t b_miscelm = (uintptr_t)b; in arena_avail_comp() 305 ret = (void *)((uintptr_t)rpages + (uintptr_t)bin_info->reg0_offset + in arena_run_reg_alloc() 315 size_t pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE; in arena_run_reg_dalloc() 2399 pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE; in arena_prof_promoted() 2462 if ((uintptr_t)run < (uintptr_t)bin->runcur) { in arena_bin_lower_run() 2483 pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE; in arena_dalloc_bin_locked_impl() 2567 size_t pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE; in arena_dalloc_large_locked_impl() [all …]
|
| H A D | chunk.c | 147 if (rtree_set(&chunks_rtree, (uintptr_t)chunk, node)) in chunk_register() 173 err = rtree_set(&chunks_rtree, (uintptr_t)chunk, NULL); in chunk_deregister() 238 leadsize = ALIGNMENT_CEILING((uintptr_t)extent_node_addr_get(node), in chunk_recycle() 239 alignment) - (uintptr_t)extent_node_addr_get(node); in chunk_recycle() 243 ret = (void *)((uintptr_t)extent_node_addr_get(node) + leadsize); in chunk_recycle() 291 extent_node_init(node, arena, (void *)((uintptr_t)(ret) + size), in chunk_recycle() 314 size_t *p = (size_t *)(uintptr_t)ret; in chunk_recycle() 481 extent_node_init(&key, arena, (void *)((uintptr_t)chunk + size), 0, in chunk_record() 624 return (pages_commit((void *)((uintptr_t)chunk + (uintptr_t)offset), in chunk_commit_default() 633 return (pages_decommit((void *)((uintptr_t)chunk + (uintptr_t)offset), in chunk_decommit_default() [all …]
|
| H A D | base.c | 62 addr = (void *)((uintptr_t)addr + nsize); in base_chunk_alloc() 110 extent_node_addr_set(node, (void *)((uintptr_t)ret + csize)); in base_alloc() 121 base_resident += PAGE_CEILING((uintptr_t)ret + csize) - in base_alloc() 122 PAGE_CEILING((uintptr_t)ret); in base_alloc()
|
| H A D | jemalloc.c | 784 *klen_p = (uintptr_t)opts - 1 - (uintptr_t)*k_p; in malloc_conf_next() 815 *vlen_p = (uintptr_t)opts - 1 - (uintptr_t)*v_p; in malloc_conf_next() 819 *vlen_p = (uintptr_t)opts - (uintptr_t)*v_p; in malloc_conf_next() 1382 if (unlikely((uintptr_t)tctx != (uintptr_t)1U)) in imalloc_prof() 1472 if (unlikely((uintptr_t)tctx != (uintptr_t)1U)) in imemalign_prof() 1603 if (unlikely((uintptr_t)tctx != (uintptr_t)1U)) in icalloc_prof() 1713 if (unlikely((uintptr_t)tctx != (uintptr_t)1U)) in irealloc_prof() 2010 if (likely((uintptr_t)tctx == (uintptr_t)1U)) in imallocx_prof() 2012 else if ((uintptr_t)tctx > (uintptr_t)1U) { in imallocx_prof() 2123 if (unlikely((uintptr_t)tctx != (uintptr_t)1U)) { in irallocx_prof() [all …]
|
| H A D | chunk_mmap.c | 22 leadsize = ALIGNMENT_CEILING((uintptr_t)pages, alignment) - in chunk_alloc_mmap_slow() 23 (uintptr_t)pages; in chunk_alloc_mmap_slow()
|
| H A D | pages.c | 72 void *ret = (void *)((uintptr_t)addr + leadsize); in pages_trim() 94 pages_unmap((void *)((uintptr_t)ret + size), trailsize); in pages_trim()
|
| H A D | quarantine.c | 8 #define QUARANTINE_STATE_REINCARNATED ((quarantine_t *)(uintptr_t)1) 9 #define QUARANTINE_STATE_PURGATORY ((quarantine_t *)(uintptr_t)2)
|
| /redis-3.2.3/deps/jemalloc/include/jemalloc/internal/ |
| H A D | rtree.h | 110 unsigned rtree_start_level(rtree_t *rtree, uintptr_t key); 111 uintptr_t rtree_subkey(rtree_t *rtree, uintptr_t key, unsigned level); 124 extent_node_t *rtree_get(rtree_t *rtree, uintptr_t key, bool dependent); 130 rtree_start_level(rtree_t *rtree, uintptr_t key) in rtree_start_level() 143 JEMALLOC_INLINE uintptr_t 144 rtree_subkey(rtree_t *rtree, uintptr_t key, unsigned level) in rtree_subkey() 156 return ((uintptr_t)node > (uintptr_t)RTREE_NODE_INITIALIZING); in rtree_node_valid() 235 rtree_get(rtree_t *rtree, uintptr_t key, bool dependent) in rtree_get() 237 uintptr_t subkey; in rtree_get() 263 rtree_set(rtree_t *rtree, uintptr_t key, const extent_node_t *val) in rtree_set() [all …]
|
| H A D | prof.h | 51 #define PROF_TDATA_STATE_REINCARNATED ((prof_tdata_t *)(uintptr_t)1) 52 #define PROF_TDATA_STATE_PURGATORY ((prof_tdata_t *)(uintptr_t)2) 439 if ((uintptr_t)tdata <= (uintptr_t)PROF_TDATA_STATE_MAX) in prof_sample_accum_update() 471 ret = (prof_tctx_t *)(uintptr_t)1U; in prof_alloc_prep() 489 if (unlikely((uintptr_t)tctx > (uintptr_t)1U)) in prof_malloc() 492 prof_tctx_set(ptr, usize, (prof_tctx_t *)(uintptr_t)1U); in prof_malloc() 503 assert(ptr != NULL || (uintptr_t)tctx <= (uintptr_t)1U); in prof_realloc() 515 tctx = (prof_tctx_t *)(uintptr_t)1U; in prof_realloc() 519 sampled = ((uintptr_t)tctx > (uintptr_t)1U); in prof_realloc() 520 old_sampled = ((uintptr_t)old_tctx > (uintptr_t)1U); in prof_realloc() [all …]
|
| H A D | arena.h | 596 size_t pageind = ((uintptr_t)miscelm - ((uintptr_t)chunk + in arena_miscelm_to_pageind() 962 pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE; in arena_ptr_small_binind_get() 977 assert(((uintptr_t)ptr - ((uintptr_t)rpages + in arena_ptr_small_binind_get() 1007 assert((uintptr_t)ptr >= (uintptr_t)rpages + in arena_run_regind() 1014 diff = (unsigned)((uintptr_t)ptr - (uintptr_t)rpages - in arena_run_regind() 1080 size_t pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE; in arena_prof_tctx_get() 1140 (uintptr_t)old_tctx > (uintptr_t)1U))) { in arena_prof_tctx_reset() 1146 pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> in arena_prof_tctx_reset() 1216 pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE; in arena_salloc() 1263 pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE; in arena_dalloc() [all …]
|
| H A D | chunk.h | 12 ((void *)((uintptr_t)(a) & ~chunksize_mask)) 16 ((size_t)((uintptr_t)(a) & chunksize_mask)) 91 return (rtree_get(&chunks_rtree, (uintptr_t)ptr, dependent)); in chunk_lookup()
|
| H A D | rb.h | 61 (a_node)->a_field.rbn_right_red = (a_type *) (((uintptr_t) a_right) \ 62 | (((uintptr_t) (a_node)->a_field.rbn_right_red) & ((size_t)1))); \ 67 ((bool) (((uintptr_t) (a_node)->a_field.rbn_right_red) \ 75 (a_node)->a_field.rbn_right_red = (a_type *) (((uintptr_t) \ 483 for (pathp--; (uintptr_t)pathp >= (uintptr_t)path; pathp--) { \ 627 for (pathp--; (uintptr_t)pathp >= (uintptr_t)path; pathp--) { \ 668 assert((uintptr_t)pathp > (uintptr_t)path); \ 800 assert((uintptr_t)pathp > (uintptr_t)path); \
|
| H A D | valgrind.h | 48 (void *)((uintptr_t)ptr + \ 68 (void *)((uintptr_t)ptr + \
|
| H A D | tcache.h | 14 #define TCACHE_STATE_DISABLED ((tcache_t *)(uintptr_t)1) 15 #define TCACHE_STATE_REINCARNATED ((tcache_t *)(uintptr_t)2) 16 #define TCACHE_STATE_PURGATORY ((tcache_t *)(uintptr_t)3) 337 size_t pageind = (((uintptr_t)ret - (uintptr_t)chunk) >> in tcache_alloc_large()
|
| H A D | jemalloc_internal_macros.h | 52 # define __DECONST(type, var) ((type)(uintptr_t)(const void *)(var))
|
| /redis-3.2.3/deps/jemalloc/test/integration/ |
| H A D | rallocx.c | 99 memset((void *)((uintptr_t)q+psz), FILL_BYTE, in TEST_BEGIN() 128 (void *)((uintptr_t)q & (align-1)), in TEST_BEGIN() 154 (void *)((uintptr_t)q & ((ZU(1) << lg_align)-1)), in TEST_BEGIN() 165 (void *)((uintptr_t)q+sz-MAX_VALIDATE), in TEST_BEGIN()
|
| H A D | MALLOCX_ARENA.c | 16 unsigned thread_ind = (unsigned)(uintptr_t)arg; in thd_start() 55 (void *)(uintptr_t)i); in TEST_BEGIN()
|
| /redis-3.2.3/deps/jemalloc/include/msvc_compat/C99/ |
| H A D | stdint.h | 116 typedef unsigned __int64 uintptr_t; typedef 119 typedef _W64 unsigned int uintptr_t; typedef
|
| /redis-3.2.3/deps/jemalloc/test/src/ |
| H A D | thd.c | 20 *ret = (void *)(uintptr_t)exit_code; in thd_join()
|