A
lgorithmique
N
umérique
D
istribuée
Public GIT Repository
projects
/
simgrid.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Fix some doxygen warnings.
[simgrid.git]
/
src
/
xbt
/
mallocator.c
diff --git
a/src/xbt/mallocator.c
b/src/xbt/mallocator.c
index
f898c39
..
9b7bd8e
100644
(file)
--- a/
src/xbt/mallocator.c
+++ b/
src/xbt/mallocator.c
@@
-1,17
+1,24
@@
/* mallocator - recycle objects to avoid malloc() / free() */
/* mallocator - recycle objects to avoid malloc() / free() */
-/* Copyright (c) 2006-201
4
. The SimGrid Team.
+/* Copyright (c) 2006-201
8
. The SimGrid Team.
* All rights reserved. */
/* This program is free software; you can redistribute it and/or modify it
* under the terms of the license (GNU LGPL) which comes with this package. */
* All rights reserved. */
/* This program is free software; you can redistribute it and/or modify it
* under the terms of the license (GNU LGPL) which comes with this package. */
-#include "src/internal_config.h"
#include "xbt/mallocator.h"
#include "xbt/mallocator.h"
+#include "mallocator_private.h"
+#include "simgrid/modelchecker.h" /* kill mallocators when model-checking is enabled */
+#include "src/internal_config.h"
#include "xbt/asserts.h"
#include "xbt/sysdep.h"
#include "xbt/asserts.h"
#include "xbt/sysdep.h"
-#include "mc/mc.h" /* kill mallocators when model-checking is enabled */
-#include "mallocator_private.h"
+
+#ifndef MIN
+#define MIN(a, b) ((a) < (b) ? (a) : (b))
+#endif
+#ifndef MAX
+#define MAX(a, b) ((a) > (b) ? (a) : (b))
+#endif
XBT_LOG_NEW_DEFAULT_SUBCATEGORY(xbt_mallocator, xbt, "Mallocators");
XBT_LOG_NEW_DEFAULT_SUBCATEGORY(xbt_mallocator, xbt, "Mallocators");
@@
-46,8
+53,7
@@
static inline void lock_reset(xbt_mallocator_t m)
static inline void lock_acquire(xbt_mallocator_t m)
{
if (initialization_done > 1) {
static inline void lock_acquire(xbt_mallocator_t m)
{
if (initialization_done > 1) {
- int *lock = &m->lock;
- while (__sync_lock_test_and_set(lock, 1))
+ while (__atomic_test_and_set(&m->lock, __ATOMIC_ACQUIRE))
/* nop */;
}
}
/* nop */;
}
}
@@
-55,7
+61,7
@@
static inline void lock_acquire(xbt_mallocator_t m)
static inline void lock_release(xbt_mallocator_t m)
{
if (initialization_done > 1)
static inline void lock_release(xbt_mallocator_t m)
{
if (initialization_done > 1)
- __
sync_lock_release(&m->lock
);
+ __
atomic_clear(&m->lock, __ATOMIC_RELEASE
);
}
/**
}
/**
@@
-71,7
+77,7
@@
void xbt_mallocator_initialization_is_done(int protect)
/** used by the module to know if it's time to activate the mallocators yet */
static inline int xbt_mallocator_is_active(void) {
/** used by the module to know if it's time to activate the mallocators yet */
static inline int xbt_mallocator_is_active(void) {
-#if HAVE_MALLOCATOR
+#if
SIMGRID_
HAVE_MALLOCATOR
return initialization_done && !MC_is_active();
#else
return 0;
return initialization_done && !MC_is_active();
#else
return 0;
@@
-154,8
+160,6
@@
void *xbt_mallocator_get(xbt_mallocator_t m)
if (m->current_size <= 0) {
/* No object is ready yet. Create a bunch of them to try to group the
* mallocs on the same memory pages (to help the cache lines) */
if (m->current_size <= 0) {
/* No object is ready yet. Create a bunch of them to try to group the
* mallocs on the same memory pages (to help the cache lines) */
-
- /* XBT_DEBUG("Create a new object for mallocator %p (size:%d/%d)", m, m->current_size, m->max_size); */
int i;
int amount = MIN(m->max_size / 2, 1000);
for (i = 0; i < amount; i++)
int i;
int amount = MIN(m->max_size / 2, 1000);
for (i = 0; i < amount; i++)
@@
-164,7
+168,6
@@
void *xbt_mallocator_get(xbt_mallocator_t m)
}
/* there is at least an available object, now */
}
/* there is at least an available object, now */
- /* XBT_DEBUG("Reuse an old object for mallocator %p (size:%d/%d)", m, m->current_size, m->max_size); */
object = m->objects[--m->current_size];
lock_release(m);
} else {
object = m->objects[--m->current_size];
lock_release(m);
} else {
@@
-202,14
+205,11
@@
void xbt_mallocator_release(xbt_mallocator_t m, void *object)
lock_acquire(m);
if (m->current_size < m->max_size) {
/* there is enough place to push the object */
lock_acquire(m);
if (m->current_size < m->max_size) {
/* there is enough place to push the object */
- /* XBT_DEBUG("Store deleted object in mallocator %p for further use (size:%d/%d)",
- m, m->current_size, m->max_size); */
m->objects[m->current_size++] = object;
lock_release(m);
} else {
lock_release(m);
/* otherwise we don't have a choice, we must free the object */
m->objects[m->current_size++] = object;
lock_release(m);
} else {
lock_release(m);
/* otherwise we don't have a choice, we must free the object */
- /* XBT_DEBUG("Free deleted object: mallocator %p is full (size:%d/%d)", m, m->current_size, m->max_size); */
m->free_f(object);
}
} else {
m->free_f(object);
}
} else {