X-Git-Url: http://info.iut-bm.univ-fcomte.fr/pub/gitweb/simgrid.git/blobdiff_plain/f3ae712a1b95294052b6e8136d0f0f2d4b30e6eb..6f065c16d0eca6e4d75c009fc578e59c30b3f392:/src/xbt/mmalloc/mm_module.c diff --git a/src/xbt/mmalloc/mm_module.c b/src/xbt/mmalloc/mm_module.c index 8ee23eb6b3..a2c36cddb6 100644 --- a/src/xbt/mmalloc/mm_module.c +++ b/src/xbt/mmalloc/mm_module.c @@ -1,4 +1,4 @@ -/* Initialization for access to a mmap'd malloc managed region. */ +/* Initialization for acces s to a mmap'd malloc managed region. */ /* Copyright (c) 2012-2014. The SimGrid Team. * All rights reserved. */ @@ -72,6 +72,11 @@ On failure returns NULL. */ xbt_mheap_t xbt_mheap_new(int fd, void *baseaddr) +{ + return xbt_mheap_new_options(fd, baseaddr, 0); +} + +xbt_mheap_t xbt_mheap_new_options(int fd, void *baseaddr, int options) { struct mdesc mtemp; xbt_mheap_t mdp; @@ -169,6 +174,7 @@ xbt_mheap_t xbt_mheap_new(int fd, void *baseaddr) mdp->base = mdp->breakval = mdp->top = baseaddr; mdp->next_mdesc = NULL; mdp->refcount = 1; + mdp->options = options; /* If we have not been passed a valid open file descriptor for the file to map to, then we go for an anonymous map */ @@ -326,7 +332,7 @@ void *mmalloc_preinit(void) if (__mmalloc_default_mdp == NULL) { unsigned long mask = ~((unsigned long)xbt_pagesize - 1); void *addr = (void*)(((unsigned long)sbrk(0) + HEAP_OFFSET) & mask); - __mmalloc_default_mdp = xbt_mheap_new(-1, addr); + __mmalloc_default_mdp = xbt_mheap_new_options(-1, addr, XBT_MHEAP_OPTION_MEMSET); /* Fixme? only the default mdp in protected against forks */ // This is mandated to protect the mmalloced areas through forks. Think of tesh. // Nah, removing the mutex isn't a good idea either for tesh @@ -356,11 +362,11 @@ size_t mmalloc_get_bytes_used(xbt_mheap_t heap){ int bytes = 0; while(i<=((struct mdesc *)heap)->heaplimit){ - if(((struct mdesc *)heap)->heapinfo[i].type == 0){ + if(((struct mdesc *)heap)->heapinfo[i].type == MMALLOC_TYPE_UNFRAGMENTED){ if(((struct mdesc *)heap)->heapinfo[i].busy_block.busy_size > 0) bytes += ((struct mdesc *)heap)->heapinfo[i].busy_block.busy_size; - }else if(((struct mdesc *)heap)->heapinfo[i].type > 0){ + } else if(((struct mdesc *)heap)->heapinfo[i].type > 0){ for(j=0; j < (size_t) (BLOCKSIZE >> ((struct mdesc *)heap)->heapinfo[i].type); j++){ if(((struct mdesc *)heap)->heapinfo[i].busy_frag.frag_size[j] > 0) bytes += ((struct mdesc *)heap)->heapinfo[i].busy_frag.frag_size[j]; @@ -375,9 +381,9 @@ size_t mmalloc_get_bytes_used(xbt_mheap_t heap){ ssize_t mmalloc_get_busy_size(xbt_mheap_t heap, void *ptr){ ssize_t block = ((char*)ptr - (char*)(heap->heapbase)) / BLOCKSIZE + 1; - if(heap->heapinfo[block].type == -1) + if(heap->heapinfo[block].type < 0) return -1; - else if(heap->heapinfo[block].type == 0) + else if(heap->heapinfo[block].type == MMALLOC_TYPE_UNFRAGMENTED) return heap->heapinfo[block].busy_block.busy_size; else{ ssize_t frag = ((uintptr_t) (ADDR2UINT (ptr) % (BLOCKSIZE))) >> heap->heapinfo[block].type; @@ -385,3 +391,32 @@ ssize_t mmalloc_get_busy_size(xbt_mheap_t heap, void *ptr){ } } + +void mmcheck(xbt_mheap_t heap) {return; + if (!heap->heapinfo) + return; + malloc_info* heapinfo = NULL; + for (size_t i=1; i < heap->heaplimit; i += mmalloc_get_increment(heapinfo)) { + heapinfo = heap->heapinfo + i; + switch (heapinfo->type) { + case MMALLOC_TYPE_HEAPINFO: + case MMALLOC_TYPE_FREE: + if (heapinfo->free_block.size==0) { + xbt_die("Block size == 0"); + } + break; + case MMALLOC_TYPE_UNFRAGMENTED: + if (heapinfo->busy_block.size==0) { + xbt_die("Block size == 0"); + } + if (heapinfo->busy_block.busy_size==0 && heapinfo->busy_block.size!=0) { + xbt_die("Empty busy block"); + } + break; + default: + if (heapinfo->type<0) { + xbt_die("Unkown mmalloc block type."); + } + } + } +}