Home
last modified time | relevance | path

Searched refs:uintptr_t (Results 1 – 25 of 35) sorted by relevance

12

/redis-3.2.3/deps/jemalloc/test/unit/
H A Drtree.c21 for (i = 1; i <= (sizeof(uintptr_t) << 3); i++) { in TEST_BEGIN()
37 for (i = 1; i <= (sizeof(uintptr_t) << 3); i++) { in TEST_BEGIN()
61 for (i = 1; i < (sizeof(uintptr_t) << 3); i++) { in TEST_BEGIN()
62 uintptr_t keys[] = {0, 1, in TEST_BEGIN()
63 (((uintptr_t)1) << (sizeof(uintptr_t)*8-i)) - 1}; in TEST_BEGIN()
70 for (j = 0; j < sizeof(keys)/sizeof(uintptr_t); j++) { in TEST_BEGIN()
73 for (k = 0; k < sizeof(keys)/sizeof(uintptr_t); k++) { in TEST_BEGIN()
82 (((uintptr_t)1) << (sizeof(uintptr_t)*8-i)), false), in TEST_BEGIN()
102 for (i = 1; i <= (sizeof(uintptr_t) << 3); i++) { in TEST_BEGIN()
103 uintptr_t keys[NSET]; in TEST_BEGIN()
[all …]
H A Dtsd.c58 data_t d = (data_t)(uintptr_t)arg; in malloc_tsd_externs()
72 assert_x_eq(*data_tsd_get(), (data_t)(uintptr_t)arg, in malloc_tsd_externs()
H A Drb.c36 ret = (((uintptr_t)a) > ((uintptr_t)b)) in node_cmp()
37 - (((uintptr_t)a) < ((uintptr_t)b)); in node_cmp()
/redis-3.2.3/deps/jemalloc/src/
H A Dchunk_dss.c117 cpad = (void *)((uintptr_t)dss_max + gap_size); in chunk_alloc_dss()
118 ret = (void *)ALIGNMENT_CEILING((uintptr_t)dss_max, in chunk_alloc_dss()
120 cpad_size = (uintptr_t)ret - (uintptr_t)cpad; in chunk_alloc_dss()
121 dss_next = (void *)((uintptr_t)ret + size); in chunk_alloc_dss()
122 if ((uintptr_t)ret < (uintptr_t)dss_max || in chunk_alloc_dss()
123 (uintptr_t)dss_next < (uintptr_t)dss_max) { in chunk_alloc_dss()
165 if ((uintptr_t)chunk >= (uintptr_t)dss_base in chunk_in_dss()
166 && (uintptr_t)chunk < (uintptr_t)dss_max) in chunk_in_dss()
H A Dextent.c30 uintptr_t a_addr = (uintptr_t)extent_node_addr_get(a); in extent_szad_comp()
31 uintptr_t b_addr = (uintptr_t)extent_node_addr_get(b); in extent_szad_comp()
46 uintptr_t a_addr = (uintptr_t)extent_node_addr_get(a); in extent_ad_comp()
47 uintptr_t b_addr = (uintptr_t)extent_node_addr_get(b); in extent_ad_comp()
H A Dhuge.c154 memset((void *)((uintptr_t)ptr + usize), 0x5a, sdiff); in huge_ralloc_no_move_similar()
177 memset((void *)((uintptr_t)ptr + oldsize), 0, in huge_ralloc_no_move_similar()
181 memset((void *)((uintptr_t)ptr + oldsize), 0xa5, usize - in huge_ralloc_no_move_similar()
212 huge_dalloc_junk((void *)((uintptr_t)ptr + usize), in huge_ralloc_no_move_shrink()
217 CHUNK_ADDR2BASE((uintptr_t)ptr + usize), in huge_ralloc_no_move_shrink()
219 CHUNK_ADDR2OFFSET((uintptr_t)ptr + usize), sdiff); in huge_ralloc_no_move_shrink()
266 memset((void *)((uintptr_t)ptr + oldsize), 0, in huge_ralloc_no_move_expand()
270 memset((void *)((uintptr_t)ptr + in huge_ralloc_no_move_expand()
275 memset((void *)((uintptr_t)ptr + oldsize), 0xa5, usize - in huge_ralloc_no_move_expand()
434 huge_prof_tctx_set(ptr, (prof_tctx_t *)(uintptr_t)1U); in huge_prof_tctx_reset()
H A Darena.c81 uintptr_t a_miscelm = (uintptr_t)a; in arena_run_comp()
82 uintptr_t b_miscelm = (uintptr_t)b; in arena_run_comp()
175 uintptr_t a_miscelm = (uintptr_t)a; in arena_avail_comp()
187 uintptr_t b_miscelm = (uintptr_t)b; in arena_avail_comp()
305 ret = (void *)((uintptr_t)rpages + (uintptr_t)bin_info->reg0_offset + in arena_run_reg_alloc()
315 size_t pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE; in arena_run_reg_dalloc()
2399 pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE; in arena_prof_promoted()
2462 if ((uintptr_t)run < (uintptr_t)bin->runcur) { in arena_bin_lower_run()
2483 pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE; in arena_dalloc_bin_locked_impl()
2567 size_t pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE; in arena_dalloc_large_locked_impl()
[all …]
H A Dchunk.c147 if (rtree_set(&chunks_rtree, (uintptr_t)chunk, node)) in chunk_register()
173 err = rtree_set(&chunks_rtree, (uintptr_t)chunk, NULL); in chunk_deregister()
238 leadsize = ALIGNMENT_CEILING((uintptr_t)extent_node_addr_get(node), in chunk_recycle()
239 alignment) - (uintptr_t)extent_node_addr_get(node); in chunk_recycle()
243 ret = (void *)((uintptr_t)extent_node_addr_get(node) + leadsize); in chunk_recycle()
291 extent_node_init(node, arena, (void *)((uintptr_t)(ret) + size), in chunk_recycle()
314 size_t *p = (size_t *)(uintptr_t)ret; in chunk_recycle()
481 extent_node_init(&key, arena, (void *)((uintptr_t)chunk + size), 0, in chunk_record()
624 return (pages_commit((void *)((uintptr_t)chunk + (uintptr_t)offset), in chunk_commit_default()
633 return (pages_decommit((void *)((uintptr_t)chunk + (uintptr_t)offset), in chunk_decommit_default()
[all …]
H A Dbase.c62 addr = (void *)((uintptr_t)addr + nsize); in base_chunk_alloc()
110 extent_node_addr_set(node, (void *)((uintptr_t)ret + csize)); in base_alloc()
121 base_resident += PAGE_CEILING((uintptr_t)ret + csize) - in base_alloc()
122 PAGE_CEILING((uintptr_t)ret); in base_alloc()
H A Djemalloc.c784 *klen_p = (uintptr_t)opts - 1 - (uintptr_t)*k_p; in malloc_conf_next()
815 *vlen_p = (uintptr_t)opts - 1 - (uintptr_t)*v_p; in malloc_conf_next()
819 *vlen_p = (uintptr_t)opts - (uintptr_t)*v_p; in malloc_conf_next()
1382 if (unlikely((uintptr_t)tctx != (uintptr_t)1U)) in imalloc_prof()
1472 if (unlikely((uintptr_t)tctx != (uintptr_t)1U)) in imemalign_prof()
1603 if (unlikely((uintptr_t)tctx != (uintptr_t)1U)) in icalloc_prof()
1713 if (unlikely((uintptr_t)tctx != (uintptr_t)1U)) in irealloc_prof()
2010 if (likely((uintptr_t)tctx == (uintptr_t)1U)) in imallocx_prof()
2012 else if ((uintptr_t)tctx > (uintptr_t)1U) { in imallocx_prof()
2123 if (unlikely((uintptr_t)tctx != (uintptr_t)1U)) { in irallocx_prof()
[all …]
H A Dchunk_mmap.c22 leadsize = ALIGNMENT_CEILING((uintptr_t)pages, alignment) - in chunk_alloc_mmap_slow()
23 (uintptr_t)pages; in chunk_alloc_mmap_slow()
H A Dpages.c72 void *ret = (void *)((uintptr_t)addr + leadsize); in pages_trim()
94 pages_unmap((void *)((uintptr_t)ret + size), trailsize); in pages_trim()
H A Dquarantine.c8 #define QUARANTINE_STATE_REINCARNATED ((quarantine_t *)(uintptr_t)1)
9 #define QUARANTINE_STATE_PURGATORY ((quarantine_t *)(uintptr_t)2)
/redis-3.2.3/deps/jemalloc/include/jemalloc/internal/
H A Drtree.h110 unsigned rtree_start_level(rtree_t *rtree, uintptr_t key);
111 uintptr_t rtree_subkey(rtree_t *rtree, uintptr_t key, unsigned level);
124 extent_node_t *rtree_get(rtree_t *rtree, uintptr_t key, bool dependent);
130 rtree_start_level(rtree_t *rtree, uintptr_t key) in rtree_start_level()
143 JEMALLOC_INLINE uintptr_t
144 rtree_subkey(rtree_t *rtree, uintptr_t key, unsigned level) in rtree_subkey()
156 return ((uintptr_t)node > (uintptr_t)RTREE_NODE_INITIALIZING); in rtree_node_valid()
235 rtree_get(rtree_t *rtree, uintptr_t key, bool dependent) in rtree_get()
237 uintptr_t subkey; in rtree_get()
263 rtree_set(rtree_t *rtree, uintptr_t key, const extent_node_t *val) in rtree_set()
[all …]
H A Dprof.h51 #define PROF_TDATA_STATE_REINCARNATED ((prof_tdata_t *)(uintptr_t)1)
52 #define PROF_TDATA_STATE_PURGATORY ((prof_tdata_t *)(uintptr_t)2)
439 if ((uintptr_t)tdata <= (uintptr_t)PROF_TDATA_STATE_MAX) in prof_sample_accum_update()
471 ret = (prof_tctx_t *)(uintptr_t)1U; in prof_alloc_prep()
489 if (unlikely((uintptr_t)tctx > (uintptr_t)1U)) in prof_malloc()
492 prof_tctx_set(ptr, usize, (prof_tctx_t *)(uintptr_t)1U); in prof_malloc()
503 assert(ptr != NULL || (uintptr_t)tctx <= (uintptr_t)1U); in prof_realloc()
515 tctx = (prof_tctx_t *)(uintptr_t)1U; in prof_realloc()
519 sampled = ((uintptr_t)tctx > (uintptr_t)1U); in prof_realloc()
520 old_sampled = ((uintptr_t)old_tctx > (uintptr_t)1U); in prof_realloc()
[all …]
H A Darena.h596 size_t pageind = ((uintptr_t)miscelm - ((uintptr_t)chunk + in arena_miscelm_to_pageind()
962 pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE; in arena_ptr_small_binind_get()
977 assert(((uintptr_t)ptr - ((uintptr_t)rpages + in arena_ptr_small_binind_get()
1007 assert((uintptr_t)ptr >= (uintptr_t)rpages + in arena_run_regind()
1014 diff = (unsigned)((uintptr_t)ptr - (uintptr_t)rpages - in arena_run_regind()
1080 size_t pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE; in arena_prof_tctx_get()
1140 (uintptr_t)old_tctx > (uintptr_t)1U))) { in arena_prof_tctx_reset()
1146 pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> in arena_prof_tctx_reset()
1216 pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE; in arena_salloc()
1263 pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE; in arena_dalloc()
[all …]
H A Dchunk.h12 ((void *)((uintptr_t)(a) & ~chunksize_mask))
16 ((size_t)((uintptr_t)(a) & chunksize_mask))
91 return (rtree_get(&chunks_rtree, (uintptr_t)ptr, dependent)); in chunk_lookup()
H A Drb.h61 (a_node)->a_field.rbn_right_red = (a_type *) (((uintptr_t) a_right) \
62 | (((uintptr_t) (a_node)->a_field.rbn_right_red) & ((size_t)1))); \
67 ((bool) (((uintptr_t) (a_node)->a_field.rbn_right_red) \
75 (a_node)->a_field.rbn_right_red = (a_type *) (((uintptr_t) \
483 for (pathp--; (uintptr_t)pathp >= (uintptr_t)path; pathp--) { \
627 for (pathp--; (uintptr_t)pathp >= (uintptr_t)path; pathp--) { \
668 assert((uintptr_t)pathp > (uintptr_t)path); \
800 assert((uintptr_t)pathp > (uintptr_t)path); \
H A Dvalgrind.h48 (void *)((uintptr_t)ptr + \
68 (void *)((uintptr_t)ptr + \
H A Dtcache.h14 #define TCACHE_STATE_DISABLED ((tcache_t *)(uintptr_t)1)
15 #define TCACHE_STATE_REINCARNATED ((tcache_t *)(uintptr_t)2)
16 #define TCACHE_STATE_PURGATORY ((tcache_t *)(uintptr_t)3)
337 size_t pageind = (((uintptr_t)ret - (uintptr_t)chunk) >> in tcache_alloc_large()
H A Djemalloc_internal_macros.h52 # define __DECONST(type, var) ((type)(uintptr_t)(const void *)(var))
/redis-3.2.3/deps/jemalloc/test/integration/
H A Drallocx.c99 memset((void *)((uintptr_t)q+psz), FILL_BYTE, in TEST_BEGIN()
128 (void *)((uintptr_t)q & (align-1)), in TEST_BEGIN()
154 (void *)((uintptr_t)q & ((ZU(1) << lg_align)-1)), in TEST_BEGIN()
165 (void *)((uintptr_t)q+sz-MAX_VALIDATE), in TEST_BEGIN()
H A DMALLOCX_ARENA.c16 unsigned thread_ind = (unsigned)(uintptr_t)arg; in thd_start()
55 (void *)(uintptr_t)i); in TEST_BEGIN()
/redis-3.2.3/deps/jemalloc/include/msvc_compat/C99/
H A Dstdint.h116 typedef unsigned __int64 uintptr_t; typedef
119 typedef _W64 unsigned int uintptr_t; typedef
/redis-3.2.3/deps/jemalloc/test/src/
H A Dthd.c20 *ret = (void *)(uintptr_t)exit_code; in thd_join()

12