151c0b2f7Stbbdev /* 2b15aabb3Stbbdev Copyright (c) 2005-2021 Intel Corporation 351c0b2f7Stbbdev 451c0b2f7Stbbdev Licensed under the Apache License, Version 2.0 (the "License"); 551c0b2f7Stbbdev you may not use this file except in compliance with the License. 651c0b2f7Stbbdev You may obtain a copy of the License at 751c0b2f7Stbbdev 851c0b2f7Stbbdev http://www.apache.org/licenses/LICENSE-2.0 951c0b2f7Stbbdev 1051c0b2f7Stbbdev Unless required by applicable law or agreed to in writing, software 1151c0b2f7Stbbdev distributed under the License is distributed on an "AS IS" BASIS, 1251c0b2f7Stbbdev WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 1351c0b2f7Stbbdev See the License for the specific language governing permissions and 1451c0b2f7Stbbdev limitations under the License. 1551c0b2f7Stbbdev */ 1651c0b2f7Stbbdev 1749e08aacStbbdev #include "oneapi/tbb/version.h" 1851c0b2f7Stbbdev 1949e08aacStbbdev #include "oneapi/tbb/detail/_exception.h" 2049e08aacStbbdev #include "oneapi/tbb/detail/_assert.h" 2149e08aacStbbdev #include "oneapi/tbb/detail/_utils.h" 228827ea7dSLong Nguyen #include "oneapi/tbb/tbb_allocator.h" // Is this OK? 238827ea7dSLong Nguyen #include "oneapi/tbb/cache_aligned_allocator.h" 2451c0b2f7Stbbdev 2551c0b2f7Stbbdev #include "dynamic_link.h" 2651c0b2f7Stbbdev #include "misc.h" 2751c0b2f7Stbbdev 2851c0b2f7Stbbdev #include <cstdlib> 2951c0b2f7Stbbdev 30ce476173SJulien Schueller #ifdef _WIN32 31ce476173SJulien Schueller #include <windows.h> 3251c0b2f7Stbbdev #else 3351c0b2f7Stbbdev #include <dlfcn.h> 34ce476173SJulien Schueller #endif 3551c0b2f7Stbbdev 36*b7a062e2SAndrey Semashev #if (!defined(_WIN32) && !defined(_WIN64)) || defined(__CYGWIN__) 37*b7a062e2SAndrey Semashev #include <stdlib.h> // posix_memalign, free 38*b7a062e2SAndrey Semashev // With glibc, uClibc and musl on Linux and bionic on Android it is safe to use memalign(), as the allocated memory 39*b7a062e2SAndrey Semashev // can be freed with free(). It is also better to use memalign() since posix_memalign() is just a wrapper on top of 40*b7a062e2SAndrey Semashev // memalign() and it offers nothing but overhead due to inconvenient interface. This is likely the case with other 41*b7a062e2SAndrey Semashev // standard libraries as well, and more libraries can be added to the preprocessor check below. Unfortunately, we 42*b7a062e2SAndrey Semashev // can't detect musl, so we simply enable memalign() on Linux and Android in general. 43*b7a062e2SAndrey Semashev #if defined(linux) || defined(__linux) || defined(__linux__) || defined(__ANDROID__) 44*b7a062e2SAndrey Semashev #include <malloc.h> // memalign 45*b7a062e2SAndrey Semashev #define __TBB_USE_MEMALIGN 46*b7a062e2SAndrey Semashev #else 47*b7a062e2SAndrey Semashev #define __TBB_USE_POSIX_MEMALIGN 48*b7a062e2SAndrey Semashev #endif 49*b7a062e2SAndrey Semashev #elif defined(_MSC_VER) || defined(__MINGW32__) 50*b7a062e2SAndrey Semashev #include <malloc.h> // _aligned_malloc, _aligned_free 51*b7a062e2SAndrey Semashev #define __TBB_USE_MSVC_ALIGNED_MALLOC 52*b7a062e2SAndrey Semashev #endif 53*b7a062e2SAndrey Semashev 5451c0b2f7Stbbdev #if __TBB_WEAK_SYMBOLS_PRESENT 5551c0b2f7Stbbdev 5651c0b2f7Stbbdev #pragma weak scalable_malloc 5751c0b2f7Stbbdev #pragma weak scalable_free 5851c0b2f7Stbbdev #pragma weak scalable_aligned_malloc 5951c0b2f7Stbbdev #pragma weak scalable_aligned_free 6051c0b2f7Stbbdev 6151c0b2f7Stbbdev extern "C" { 6251c0b2f7Stbbdev void* scalable_malloc(std::size_t); 6351c0b2f7Stbbdev void scalable_free(void*); 6451c0b2f7Stbbdev void* scalable_aligned_malloc(std::size_t, std::size_t); 6551c0b2f7Stbbdev void scalable_aligned_free(void*); 6651c0b2f7Stbbdev } 6751c0b2f7Stbbdev 6851c0b2f7Stbbdev #endif /* __TBB_WEAK_SYMBOLS_PRESENT */ 6951c0b2f7Stbbdev 7051c0b2f7Stbbdev namespace tbb { 7151c0b2f7Stbbdev namespace detail { 7251c0b2f7Stbbdev namespace r1 { 7351c0b2f7Stbbdev 7451c0b2f7Stbbdev //! Initialization routine used for first indirect call via allocate_handler. 7551c0b2f7Stbbdev static void* initialize_allocate_handler(std::size_t size); 7651c0b2f7Stbbdev 7751c0b2f7Stbbdev //! Handler for memory allocation 788e7f9e14SAlex using allocate_handler_type = void* (*)(std::size_t size); 798e7f9e14SAlex static std::atomic<allocate_handler_type> allocate_handler{ &initialize_allocate_handler }; 808e7f9e14SAlex allocate_handler_type allocate_handler_unsafe = nullptr; 8151c0b2f7Stbbdev 8251c0b2f7Stbbdev //! Handler for memory deallocation 8351c0b2f7Stbbdev static void (*deallocate_handler)(void* pointer) = nullptr; 8451c0b2f7Stbbdev 8551c0b2f7Stbbdev //! Initialization routine used for first indirect call via cache_aligned_allocate_handler. 8651c0b2f7Stbbdev static void* initialize_cache_aligned_allocate_handler(std::size_t n, std::size_t alignment); 8751c0b2f7Stbbdev 88*b7a062e2SAndrey Semashev //! Allocates overaligned memory using standard memory allocator. It is used when scalable_allocator is not available. 8951c0b2f7Stbbdev static void* std_cache_aligned_allocate(std::size_t n, std::size_t alignment); 9051c0b2f7Stbbdev 91*b7a062e2SAndrey Semashev //! Deallocates overaligned memory using standard memory allocator. It is used when scalable_allocator is not available. 9251c0b2f7Stbbdev static void std_cache_aligned_deallocate(void* p); 9351c0b2f7Stbbdev 9451c0b2f7Stbbdev //! Handler for padded memory allocation 958e7f9e14SAlex using cache_aligned_allocate_handler_type = void* (*)(std::size_t n, std::size_t alignment); 968e7f9e14SAlex static std::atomic<cache_aligned_allocate_handler_type> cache_aligned_allocate_handler{ &initialize_cache_aligned_allocate_handler }; 978e7f9e14SAlex cache_aligned_allocate_handler_type cache_aligned_allocate_handler_unsafe = nullptr; 9851c0b2f7Stbbdev 9951c0b2f7Stbbdev //! Handler for padded memory deallocation 10051c0b2f7Stbbdev static void (*cache_aligned_deallocate_handler)(void* p) = nullptr; 10151c0b2f7Stbbdev 10251c0b2f7Stbbdev //! Table describing how to link the handlers. 10351c0b2f7Stbbdev static const dynamic_link_descriptor MallocLinkTable[] = { 1048e7f9e14SAlex DLD(scalable_malloc, allocate_handler_unsafe), 10551c0b2f7Stbbdev DLD(scalable_free, deallocate_handler), 1068e7f9e14SAlex DLD(scalable_aligned_malloc, cache_aligned_allocate_handler_unsafe), 10751c0b2f7Stbbdev DLD(scalable_aligned_free, cache_aligned_deallocate_handler), 10851c0b2f7Stbbdev }; 10951c0b2f7Stbbdev 11051c0b2f7Stbbdev 11151c0b2f7Stbbdev #if TBB_USE_DEBUG 11251c0b2f7Stbbdev #define DEBUG_SUFFIX "_debug" 11351c0b2f7Stbbdev #else 11451c0b2f7Stbbdev #define DEBUG_SUFFIX 11551c0b2f7Stbbdev #endif /* TBB_USE_DEBUG */ 11651c0b2f7Stbbdev 11751c0b2f7Stbbdev // MALLOCLIB_NAME is the name of the oneTBB memory allocator library. 11851c0b2f7Stbbdev #if _WIN32||_WIN64 11951c0b2f7Stbbdev #define MALLOCLIB_NAME "tbbmalloc" DEBUG_SUFFIX ".dll" 12051c0b2f7Stbbdev #elif __APPLE__ 12151c0b2f7Stbbdev #define MALLOCLIB_NAME "libtbbmalloc" DEBUG_SUFFIX ".dylib" 12251c0b2f7Stbbdev #elif __FreeBSD__ || __NetBSD__ || __OpenBSD__ || __sun || _AIX || __ANDROID__ 12351c0b2f7Stbbdev #define MALLOCLIB_NAME "libtbbmalloc" DEBUG_SUFFIX ".so" 124734f0bc0SPablo Romero #elif __unix__ // Note that order of these #elif's is important! 12551c0b2f7Stbbdev #define MALLOCLIB_NAME "libtbbmalloc" DEBUG_SUFFIX ".so.2" 12651c0b2f7Stbbdev #else 12751c0b2f7Stbbdev #error Unknown OS 12851c0b2f7Stbbdev #endif 12951c0b2f7Stbbdev 13051c0b2f7Stbbdev //! Initialize the allocation/free handler pointers. 13151c0b2f7Stbbdev /** Caller is responsible for ensuring this routine is called exactly once. 13251c0b2f7Stbbdev The routine attempts to dynamically link with the TBB memory allocator. 13351c0b2f7Stbbdev If that allocator is not found, it links to malloc and free. */ 13451c0b2f7Stbbdev void initialize_handler_pointers() { 13551c0b2f7Stbbdev __TBB_ASSERT(allocate_handler == &initialize_allocate_handler, NULL); 13651c0b2f7Stbbdev bool success = dynamic_link(MALLOCLIB_NAME, MallocLinkTable, 4); 13751c0b2f7Stbbdev if(!success) { 13851c0b2f7Stbbdev // If unsuccessful, set the handlers to the default routines. 13951c0b2f7Stbbdev // This must be done now, and not before FillDynamicLinks runs, because if other 14051c0b2f7Stbbdev // threads call the handlers, we want them to go through the DoOneTimeInitializations logic, 14151c0b2f7Stbbdev // which forces them to wait. 1428e7f9e14SAlex allocate_handler_unsafe = &std::malloc; 14351c0b2f7Stbbdev deallocate_handler = &std::free; 1448e7f9e14SAlex cache_aligned_allocate_handler_unsafe = &std_cache_aligned_allocate; 14551c0b2f7Stbbdev cache_aligned_deallocate_handler = &std_cache_aligned_deallocate; 14651c0b2f7Stbbdev } 14751c0b2f7Stbbdev 1488e7f9e14SAlex allocate_handler.store(allocate_handler_unsafe, std::memory_order_release); 1498e7f9e14SAlex cache_aligned_allocate_handler.store(cache_aligned_allocate_handler_unsafe, std::memory_order_release); 1508e7f9e14SAlex 15151c0b2f7Stbbdev PrintExtraVersionInfo( "ALLOCATOR", success?"scalable_malloc":"malloc" ); 15251c0b2f7Stbbdev } 15351c0b2f7Stbbdev 15451c0b2f7Stbbdev static std::once_flag initialization_state; 15551c0b2f7Stbbdev void initialize_cache_aligned_allocator() { 15651c0b2f7Stbbdev std::call_once(initialization_state, &initialize_handler_pointers); 15751c0b2f7Stbbdev } 15851c0b2f7Stbbdev 15951c0b2f7Stbbdev //! Executed on very first call through allocate_handler 16051c0b2f7Stbbdev static void* initialize_allocate_handler(std::size_t size) { 16151c0b2f7Stbbdev initialize_cache_aligned_allocator(); 16251c0b2f7Stbbdev __TBB_ASSERT(allocate_handler != &initialize_allocate_handler, NULL); 16351c0b2f7Stbbdev return (*allocate_handler)(size); 16451c0b2f7Stbbdev } 16551c0b2f7Stbbdev 16651c0b2f7Stbbdev //! Executed on very first call through cache_aligned_allocate_handler 16751c0b2f7Stbbdev static void* initialize_cache_aligned_allocate_handler(std::size_t bytes, std::size_t alignment) { 16851c0b2f7Stbbdev initialize_cache_aligned_allocator(); 16951c0b2f7Stbbdev __TBB_ASSERT(cache_aligned_allocate_handler != &initialize_cache_aligned_allocate_handler, NULL); 17051c0b2f7Stbbdev return (*cache_aligned_allocate_handler)(bytes, alignment); 17151c0b2f7Stbbdev } 17251c0b2f7Stbbdev 17351c0b2f7Stbbdev // TODO: use CPUID to find actual line size, though consider backward compatibility 17451c0b2f7Stbbdev // nfs - no false sharing 17551c0b2f7Stbbdev static constexpr std::size_t nfs_size = 128; 17651c0b2f7Stbbdev 17751c0b2f7Stbbdev std::size_t __TBB_EXPORTED_FUNC cache_line_size() { 17851c0b2f7Stbbdev return nfs_size; 17951c0b2f7Stbbdev } 18051c0b2f7Stbbdev 18151c0b2f7Stbbdev void* __TBB_EXPORTED_FUNC cache_aligned_allocate(std::size_t size) { 18251c0b2f7Stbbdev const std::size_t cache_line_size = nfs_size; 18351c0b2f7Stbbdev __TBB_ASSERT(is_power_of_two(cache_line_size), "must be power of two"); 18451c0b2f7Stbbdev 18551c0b2f7Stbbdev // Check for overflow 18651c0b2f7Stbbdev if (size + cache_line_size < size) { 18751c0b2f7Stbbdev throw_exception(exception_id::bad_alloc); 18851c0b2f7Stbbdev } 18951c0b2f7Stbbdev // scalable_aligned_malloc considers zero size request an error, and returns NULL 19051c0b2f7Stbbdev if (size == 0) size = 1; 19151c0b2f7Stbbdev 1928e7f9e14SAlex void* result = cache_aligned_allocate_handler.load(std::memory_order_acquire)(size, cache_line_size); 19351c0b2f7Stbbdev if (!result) { 19451c0b2f7Stbbdev throw_exception(exception_id::bad_alloc); 19551c0b2f7Stbbdev } 19651c0b2f7Stbbdev __TBB_ASSERT(is_aligned(result, cache_line_size), "The returned address isn't aligned"); 19751c0b2f7Stbbdev return result; 19851c0b2f7Stbbdev } 19951c0b2f7Stbbdev 20051c0b2f7Stbbdev void __TBB_EXPORTED_FUNC cache_aligned_deallocate(void* p) { 20151c0b2f7Stbbdev __TBB_ASSERT(cache_aligned_deallocate_handler, "Initialization has not been yet."); 20251c0b2f7Stbbdev (*cache_aligned_deallocate_handler)(p); 20351c0b2f7Stbbdev } 20451c0b2f7Stbbdev 20551c0b2f7Stbbdev static void* std_cache_aligned_allocate(std::size_t bytes, std::size_t alignment) { 206*b7a062e2SAndrey Semashev #if defined(__TBB_USE_MEMALIGN) 207*b7a062e2SAndrey Semashev return memalign(alignment, bytes); 208*b7a062e2SAndrey Semashev #elif defined(__TBB_USE_POSIX_MEMALIGN) 209*b7a062e2SAndrey Semashev void* p = nullptr; 210*b7a062e2SAndrey Semashev int res = posix_memalign(&p, alignment, bytes); 211*b7a062e2SAndrey Semashev if (res != 0) 212*b7a062e2SAndrey Semashev p = nullptr; 213*b7a062e2SAndrey Semashev return p; 214*b7a062e2SAndrey Semashev #elif defined(__TBB_USE_MSVC_ALIGNED_MALLOC) 215*b7a062e2SAndrey Semashev return _aligned_malloc(bytes, alignment); 216*b7a062e2SAndrey Semashev #else 21751c0b2f7Stbbdev // TODO: make it common with cache_aligned_resource 21851c0b2f7Stbbdev std::size_t space = alignment + bytes; 21951c0b2f7Stbbdev std::uintptr_t base = reinterpret_cast<std::uintptr_t>(std::malloc(space)); 22051c0b2f7Stbbdev if (!base) { 22151c0b2f7Stbbdev return nullptr; 22251c0b2f7Stbbdev } 22351c0b2f7Stbbdev std::uintptr_t result = (base + nfs_size) & ~(nfs_size - 1); 22451c0b2f7Stbbdev // Round up to the next cache line (align the base address) 22551c0b2f7Stbbdev __TBB_ASSERT((result - base) >= sizeof(std::uintptr_t), "Cannot store a base pointer to the header"); 22651c0b2f7Stbbdev __TBB_ASSERT(space - (result - base) >= bytes, "Not enough space for the storage"); 22751c0b2f7Stbbdev 22851c0b2f7Stbbdev // Record where block actually starts. 22951c0b2f7Stbbdev (reinterpret_cast<std::uintptr_t*>(result))[-1] = base; 23051c0b2f7Stbbdev return reinterpret_cast<void*>(result); 231*b7a062e2SAndrey Semashev #endif 23251c0b2f7Stbbdev } 23351c0b2f7Stbbdev 23451c0b2f7Stbbdev static void std_cache_aligned_deallocate(void* p) { 235*b7a062e2SAndrey Semashev #if defined(__TBB_USE_MEMALIGN) || defined(__TBB_USE_POSIX_MEMALIGN) 236*b7a062e2SAndrey Semashev free(p); 237*b7a062e2SAndrey Semashev #elif defined(__TBB_USE_MSVC_ALIGNED_MALLOC) 238*b7a062e2SAndrey Semashev _aligned_free(p); 239*b7a062e2SAndrey Semashev #else 24051c0b2f7Stbbdev if (p) { 24151c0b2f7Stbbdev __TBB_ASSERT(reinterpret_cast<std::uintptr_t>(p) >= 0x4096, "attempt to free block not obtained from cache_aligned_allocator"); 24251c0b2f7Stbbdev // Recover where block actually starts 24351c0b2f7Stbbdev std::uintptr_t base = (reinterpret_cast<std::uintptr_t*>(p))[-1]; 24451c0b2f7Stbbdev __TBB_ASSERT(((base + nfs_size) & ~(nfs_size - 1)) == reinterpret_cast<std::uintptr_t>(p), "Incorrect alignment or not allocated by std_cache_aligned_deallocate?"); 24551c0b2f7Stbbdev std::free(reinterpret_cast<void*>(base)); 24651c0b2f7Stbbdev } 247*b7a062e2SAndrey Semashev #endif 24851c0b2f7Stbbdev } 24951c0b2f7Stbbdev 25051c0b2f7Stbbdev void* __TBB_EXPORTED_FUNC allocate_memory(std::size_t size) { 2518e7f9e14SAlex void* result = allocate_handler.load(std::memory_order_acquire)(size); 25251c0b2f7Stbbdev if (!result) { 25351c0b2f7Stbbdev throw_exception(exception_id::bad_alloc); 25451c0b2f7Stbbdev } 25551c0b2f7Stbbdev return result; 25651c0b2f7Stbbdev } 25751c0b2f7Stbbdev 25851c0b2f7Stbbdev void __TBB_EXPORTED_FUNC deallocate_memory(void* p) { 25951c0b2f7Stbbdev if (p) { 26051c0b2f7Stbbdev __TBB_ASSERT(deallocate_handler, "Initialization has not been yet."); 26151c0b2f7Stbbdev (*deallocate_handler)(p); 26251c0b2f7Stbbdev } 26351c0b2f7Stbbdev } 26451c0b2f7Stbbdev 26551c0b2f7Stbbdev bool __TBB_EXPORTED_FUNC is_tbbmalloc_used() { 2668e7f9e14SAlex auto handler_snapshot = allocate_handler.load(std::memory_order_acquire); 2678e7f9e14SAlex if (handler_snapshot == &initialize_allocate_handler) { 2688e7f9e14SAlex initialize_cache_aligned_allocator(); 26951c0b2f7Stbbdev } 2708e7f9e14SAlex handler_snapshot = allocate_handler.load(std::memory_order_relaxed); 2718e7f9e14SAlex __TBB_ASSERT(handler_snapshot != &initialize_allocate_handler && deallocate_handler != nullptr, NULL); 27251c0b2f7Stbbdev // Cast to void avoids type mismatch errors on some compilers (e.g. __IBMCPP__) 2738e7f9e14SAlex __TBB_ASSERT((reinterpret_cast<void*>(handler_snapshot) == reinterpret_cast<void*>(&std::malloc)) == (reinterpret_cast<void*>(deallocate_handler) == reinterpret_cast<void*>(&std::free)), 27451c0b2f7Stbbdev "Both shim pointers must refer to routines from the same package (either TBB or CRT)"); 2758e7f9e14SAlex return reinterpret_cast<void*>(handler_snapshot) == reinterpret_cast<void*>(&std::malloc); 27651c0b2f7Stbbdev } 27751c0b2f7Stbbdev 27851c0b2f7Stbbdev } // namespace r1 27951c0b2f7Stbbdev } // namespace detail 28051c0b2f7Stbbdev } // namespace tbb 281