X-Git-Url: http://info.iut-bm.univ-fcomte.fr/pub/gitweb/simgrid.git/blobdiff_plain/9d1a9a4b871895531b7e70f313691ef75dc47a96..84402e8e2ee2a2d0bef25fdceb0a263ed8b471f6:/src/smpi/internals/smpi_shared.cpp diff --git a/src/smpi/internals/smpi_shared.cpp b/src/smpi/internals/smpi_shared.cpp index 29f3b27171..c510fc5d63 100644 --- a/src/smpi/internals/smpi_shared.cpp +++ b/src/smpi/internals/smpi_shared.cpp @@ -1,4 +1,4 @@ -/* Copyright (c) 2007-2019. The SimGrid Team. All rights reserved. */ +/* Copyright (c) 2007-2020. The SimGrid Team. All rights reserved. */ /* This program is free software; you can redistribute it and/or modify it * under the terms of the license (GNU LGPL) which comes with this package. */ @@ -59,8 +59,6 @@ XBT_LOG_NEW_DEFAULT_SUBCATEGORY(smpi_shared, smpi, "Logging specific to SMPI (shared memory macros)"); -#define PTR_STRLEN (2 + 2 * sizeof(void*) + 1) - namespace{ /** Some location in the source code * @@ -109,35 +107,11 @@ void smpi_shared_destroy() calls.clear(); } -static size_t shm_size(int fd) { - struct stat st; - - if(fstat(fd, &st) < 0) { - xbt_die("Could not stat fd %d: %s", fd, strerror(errno)); - } - return static_cast(st.st_size); -} - #ifndef WIN32 -static void* shm_map(int fd, size_t size, shared_data_key_type* data) { - char loc[PTR_STRLEN]; +static void* shm_map(int fd, size_t size, shared_data_key_type* data) +{ + void* mem = smpi_temp_shm_mmap(fd, size); shared_metadata_t meta; - - if(size > shm_size(fd) && (ftruncate(fd, static_cast(size)) < 0)) { - xbt_die("Could not truncate fd %d to %zu: %s", fd, size, strerror(errno)); - } - - void* mem = mmap(nullptr, size, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0); - if(mem == MAP_FAILED) { - xbt_die("Failed to map fd %d with size %zu: %s\n" - "If you are running a lot of ranks, you may be exceeding the amount of mappings allowed per process.\n" - "On Linux systems, change this value with sudo sysctl -w vm.max_map_count=newvalue (default value: 65536)\n" - "Please see " - "https://simgrid.org/doc/latest/Configuring_SimGrid.html#configuring-the-user-code-virtualization for more " - "information.", - fd, size, strerror(errno)); - } - snprintf(loc, PTR_STRLEN, "%p", mem); meta.size = size; meta.data = data; meta.allocated_ptr = mem; @@ -155,28 +129,15 @@ static void *smpi_shared_malloc_local(size_t size, const char *file, int line) auto data = res.first; if (res.second) { // The new element was inserted. - // Generate a shared memory name from the address of the shared_data: - char shmname[32]; // cannot be longer than PSHMNAMLEN = 31 on macOS (shm_open raises ENAMETOOLONG otherwise) - snprintf(shmname, 31, "/shmalloc%p", &*data); - int fd = shm_open(shmname, O_RDWR | O_CREAT | O_EXCL, S_IRUSR | S_IWUSR | S_IRGRP | S_IROTH); - if (fd < 0) { - if (errno == EEXIST) - xbt_die("Please cleanup /dev/shm/%s", shmname); - else - xbt_die("An unhandled error occurred while opening %s. shm_open: %s", shmname, strerror(errno)); - } - data->second.fd = fd; + int fd = smpi_temp_shm_get(); + data->second.fd = fd; data->second.count = 1; mem = shm_map(fd, size, &*data); - if (shm_unlink(shmname) < 0) { - XBT_WARN("Could not early unlink %s. shm_unlink: %s", shmname, strerror(errno)); - } - XBT_DEBUG("Mapping %s at %p through %d", shmname, mem, fd); } else { mem = shm_map(data->second.fd, size, &*data); data->second.count++; } - XBT_DEBUG("Shared malloc %zu in %p (metadata at %p)", size, mem, &*data); + XBT_DEBUG("Shared malloc %zu in %p through %d (metadata at %p)", size, mem, data->second.fd, &*data); return mem; } @@ -247,7 +208,7 @@ void* smpi_shared_malloc_partial(size_t size, size_t* shared_block_offsets, int smpi_shared_malloc_bogusfile = mkstemp(name); XBT_DEBUG("bogusfile : %s\n", name); unlink(name); - char* dumb = new char[smpi_shared_malloc_blocksize](); // zero initialized + const char* dumb = new char[smpi_shared_malloc_blocksize](); // zero initialized ssize_t err = write(smpi_shared_malloc_bogusfile, dumb, smpi_shared_malloc_blocksize); if(err<0) xbt_die("Could not write bogus file for shared malloc"); @@ -279,8 +240,7 @@ void* smpi_shared_malloc_partial(size_t size, size_t* shared_block_offsets, int for (size_t offset = start_block_offset; offset < stop_block_offset; offset += smpi_shared_malloc_blocksize) { XBT_DEBUG("\t\tglobal shared allocation, mmap block offset %zx", offset); void* pos = (void*)((unsigned long)mem + offset); - void* res = mmap(pos, smpi_shared_malloc_blocksize, PROT_READ | PROT_WRITE, mmap_flag, - huge_fd, 0); + const void* res = mmap(pos, smpi_shared_malloc_blocksize, PROT_READ | PROT_WRITE, mmap_flag, huge_fd, 0); xbt_assert(res == pos, "Could not map folded virtual memory (%s). Do you perhaps need to increase the " "size of the mapped file using --cfg=smpi/shared-malloc-blocksize:newvalue (default 1048576) ? " "You can also try using the sysctl vm.max_map_count. " @@ -293,8 +253,9 @@ void* smpi_shared_malloc_partial(size_t size, size_t* shared_block_offsets, int if(low_page_start_offset < low_page_stop_offset) { XBT_DEBUG("\t\tglobal shared allocation, mmap block start"); void* pos = (void*)((unsigned long)mem + low_page_start_offset); - void* res = mmap(pos, low_page_stop_offset-low_page_start_offset, PROT_READ | PROT_WRITE, mmap_base_flag, // not a full huge page - smpi_shared_malloc_bogusfile, 0); + const void* res = mmap(pos, low_page_stop_offset - low_page_start_offset, PROT_READ | PROT_WRITE, + mmap_base_flag, // not a full huge page + smpi_shared_malloc_bogusfile, 0); xbt_assert(res == pos, "Could not map folded virtual memory (%s). Do you perhaps need to increase the " "size of the mapped file using --cfg=smpi/shared-malloc-blocksize:newvalue (default 1048576) ?" "You can also try using the sysctl vm.max_map_count", @@ -305,8 +266,9 @@ void* smpi_shared_malloc_partial(size_t size, size_t* shared_block_offsets, int size_t high_page_stop_offset = stop_offset == size ? size : ALIGN_DOWN((int64_t)stop_offset, PAGE_SIZE); if(high_page_stop_offset > stop_block_offset) { void* pos = (void*)((unsigned long)mem + stop_block_offset); - void* res = mmap(pos, high_page_stop_offset-stop_block_offset, PROT_READ | PROT_WRITE, mmap_base_flag, // not a full huge page - smpi_shared_malloc_bogusfile, 0); + const void* res = mmap(pos, high_page_stop_offset - stop_block_offset, PROT_READ | PROT_WRITE, + mmap_base_flag, // not a full huge page + smpi_shared_malloc_bogusfile, 0); xbt_assert(res == pos, "Could not map folded virtual memory (%s). Do you perhaps need to increase the " "size of the mapped file using --cfg=smpi/shared-malloc-blocksize:newvalue (default 1048576) ?" "You can also try using the sysctl vm.max_map_count", @@ -342,28 +304,29 @@ void* smpi_shared_malloc_partial(size_t size, size_t* shared_block_offsets, int return mem; } - -void *smpi_shared_malloc_intercept(size_t size, const char *file, int line) { - if( simgrid::config::get_value("smpi/auto-shared-malloc-thresh") == 0 || size < simgrid::config::get_value("smpi/auto-shared-malloc-thresh")) +void* smpi_shared_malloc_intercept(size_t size, const char* file, int line) +{ + if( smpi_cfg_auto_shared_malloc_thresh() == 0 || size < smpi_cfg_auto_shared_malloc_thresh()) return ::operator new(size); else return smpi_shared_malloc(size, file, line); } -void* smpi_shared_calloc_intercept(size_t num_elm, size_t elem_size, const char* file, int line){ - if( simgrid::config::get_value("smpi/auto-shared-malloc-thresh") == 0 || elem_size*num_elm < simgrid::config::get_value("smpi/auto-shared-malloc-thresh")){ +void* smpi_shared_calloc_intercept(size_t num_elm, size_t elem_size, const char* file, int line) +{ + if( smpi_cfg_auto_shared_malloc_thresh() == 0 || elem_size*num_elm < smpi_cfg_auto_shared_malloc_thresh()){ void* ptr = ::operator new(elem_size*num_elm); memset(ptr, 0, elem_size*num_elm); return ptr; } else return smpi_shared_malloc(elem_size*num_elm, file, line); - } -void *smpi_shared_malloc(size_t size, const char *file, int line) { - if (size > 0 && smpi_cfg_shared_malloc == SharedMallocType::LOCAL) { +void* smpi_shared_malloc(size_t size, const char* file, int line) +{ + if (size > 0 && smpi_cfg_shared_malloc() == SharedMallocType::LOCAL) { return smpi_shared_malloc_local(size, file, line); - } else if (smpi_cfg_shared_malloc == SharedMallocType::GLOBAL) { + } else if (smpi_cfg_shared_malloc() == SharedMallocType::GLOBAL) { int nb_shared_blocks = 1; size_t shared_block_offsets[2] = {0, size}; return smpi_shared_malloc_partial(size, shared_block_offsets, nb_shared_blocks); @@ -376,7 +339,7 @@ int smpi_is_shared(const void* ptr, std::vector> &priv private_blocks.clear(); // being paranoid if (allocs_metadata.empty()) return 0; - if (smpi_cfg_shared_malloc == SharedMallocType::LOCAL || smpi_cfg_shared_malloc == SharedMallocType::GLOBAL) { + if (smpi_cfg_shared_malloc() == SharedMallocType::LOCAL || smpi_cfg_shared_malloc() == SharedMallocType::GLOBAL) { auto low = allocs_metadata.lower_bound(ptr); if (low != allocs_metadata.end() && low->first == ptr) { private_blocks = low->second.private_blocks; @@ -440,9 +403,7 @@ std::vector> merge_private_blocks(const std::vectorcount); } - } else if (smpi_cfg_shared_malloc == SharedMallocType::GLOBAL) { + } else if (smpi_cfg_shared_malloc() == SharedMallocType::GLOBAL) { auto meta = allocs_metadata.find(ptr); if (meta != allocs_metadata.end()){ meta->second.data->second.count--;