X-Git-Url: http://info.iut-bm.univ-fcomte.fr/pub/gitweb/simgrid.git/blobdiff_plain/6af348b861fc72e45173e89c5f9540cefb6ee252..f8abe8974ac5f49ab6282c4584899333dd9611a6:/src/xbt/mmalloc/mmprivate.h diff --git a/src/xbt/mmalloc/mmprivate.h b/src/xbt/mmalloc/mmprivate.h index 4854410fe6..db3ee33f61 100644 --- a/src/xbt/mmalloc/mmprivate.h +++ b/src/xbt/mmalloc/mmprivate.h @@ -1,6 +1,6 @@ /* Declarations for `mmalloc' and friends. */ -/* Copyright (c) 2010-2018. The SimGrid Team. All rights reserved. */ +/* Copyright (c) 2010-2019. The SimGrid Team. All rights reserved. */ /* This program is free software; you can redistribute it and/or modify it * under the terms of the license (GNU LGPL) which comes with this package. */ @@ -18,7 +18,6 @@ #include "swag.h" #include "src/internal_config.h" -#include "xbt/xbt_os_thread.h" #include "xbt/mmalloc.h" #include "xbt/ex.h" #include "xbt/dynar.h" @@ -52,14 +51,13 @@ #define BLOCKIFY(SIZE) (((SIZE) + BLOCKSIZE - 1) / BLOCKSIZE) /* We keep fragment-specific meta-data for introspection purposes, and these - * information are kept in fixed lenght arrays. Here is the computation of + * information are kept in fixed length arrays. Here is the computation of * that size. * * Never make SMALLEST_POSSIBLE_MALLOC smaller than sizeof(list) because we * need to enlist the free fragments. */ -//#define SMALLEST_POSSIBLE_MALLOC (sizeof(struct list)) #define SMALLEST_POSSIBLE_MALLOC (16*sizeof(struct list)) #define MAX_FRAGMENT_PER_BLOCK (BLOCKSIZE / SMALLEST_POSSIBLE_MALLOC) @@ -96,7 +94,7 @@ #define ADDRESS(B) ((void*) (((ADDR2UINT(B)) - 1) * BLOCKSIZE + (char*) mdp -> heapbase)) -SG_BEGIN_DECL() +SG_BEGIN_DECL /* Doubly linked lists of free fragments. */ struct list { @@ -154,14 +152,11 @@ typedef struct { struct { size_t nfree; /* Free fragments in a fragmented block. */ ssize_t frag_size[MAX_FRAGMENT_PER_BLOCK]; - //void *bt[MAX_FRAGMENT_PER_BLOCK][XBT_BACKTRACE_SIZE]; /* Where it was malloced (or realloced lastly) */ int ignore[MAX_FRAGMENT_PER_BLOCK]; } busy_frag; struct { size_t size; /* Size (in blocks) of a large cluster. */ size_t busy_size; /* Actually used space, in bytes */ - //void *bt[XBT_BACKTRACE_SIZE]; /* Where it was malloced (or realloced lastly) */ - //int bt_size; int ignore; } busy_block; /* Heap information for a free block (that may be the first of a free cluster). */ @@ -231,7 +226,7 @@ struct mdesc { /* @brief List of all blocks containing free fragments of a given size. * - * The array indice is the log2 of requested size. + * The array index is the log2 of requested size. * Actually only the sizes 8->11 seem to be used, but who cares? */ s_xbt_swag_t fraghead[BLOCKLOG]; @@ -291,13 +286,13 @@ XBT_PUBLIC void* mmorecore(struct mdesc* mdp, ssize_t size); * in a model-checking enabled tree. Without this protection, our malloc * implementation will not like multi-threading AT ALL. */ -#define LOCK(mdp) pthread_mutex_lock(&mdp->mutex) -#define UNLOCK(mdp) pthread_mutex_unlock(&mdp->mutex) +#define LOCK(mdp) pthread_mutex_lock(&(mdp)->mutex) +#define UNLOCK(mdp) pthread_mutex_unlock(&(mdp)->mutex) XBT_PRIVATE int malloc_use_mmalloc(void); XBT_PRIVATE size_t mmalloc_get_bytes_used_remote(size_t heaplimit, const malloc_info* heapinfo); -SG_END_DECL() +SG_END_DECL #endif