+ mdp -> heapstats.bytes_used += 1 << log;
+ mdp -> heapstats.chunks_free--;
+ mdp -> heapstats.bytes_free -= 1 << log;
+
+ } else {
+ /* No free fragments of the desired size, so get a new block
+ and break it into fragments, returning the first. */
+
+ result = mmalloc(mdp, BLOCKSIZE); // does not return NULL
+ block = BLOCK(result);
+
+ mdp->heapinfo[block].type = log;
+ /* Link all fragments but the first as free, and add the block to the swag of blocks containing free frags */
+ for (i = 1; i < (size_t) (BLOCKSIZE >> log); ++i) {
+ mdp->heapinfo[block].busy_frag.frag_size[i] = -1;
+ mdp->heapinfo[block].busy_frag.ignore[i] = 0;
+ }
+ mdp->heapinfo[block].busy_frag.nfree = i - 1;
+ mdp->heapinfo[block].freehook.prev = NULL;
+ mdp->heapinfo[block].freehook.next = NULL;
+
+ xbt_swag_insert(&mdp->heapinfo[block], &(mdp->fraghead[log]));
+
+ /* mark the fragment returned as busy */
+ mdp->heapinfo[block].busy_frag.frag_size[0] = requested_size;
+ mdp->heapinfo[block].busy_frag.ignore[0] = 0;
+ //xbt_backtrace_no_malloc(mdp->heapinfo[block].busy_frag.bt[0],XBT_BACKTRACE_SIZE);
+ //xbt_libunwind_backtrace(mdp->heapinfo[block].busy_frag.bt[0],XBT_BACKTRACE_SIZE);
+
+ /* update stats */
+ mdp -> heapstats.chunks_free += (BLOCKSIZE >> log) - 1;
+ mdp -> heapstats.bytes_free += BLOCKSIZE - (1 << log);
+ mdp -> heapstats.bytes_used -= BLOCKSIZE - (1 << log);
+ }
+ } else {
+ /* Large allocation to receive one or more blocks.
+ Search the free list in a circle starting at the last place visited.
+ If we loop completely around without finding a large enough
+ space we will have to get more memory from the system. */
+ blocks = BLOCKIFY(size);
+ start = block = MALLOC_SEARCH_START;
+ while (mdp->heapinfo[block].free_block.size < blocks) {
+ if (mdp->heapinfo[block].type >=0) { // Don't trust xbt_die and friends in malloc-level library, you fool!
+ fprintf(stderr,"Internal error: found a free block not marked as such (block=%lu type=%lu). Please report this bug.\n",(unsigned long)block,(unsigned long)mdp->heapinfo[block].type);
+ abort();
+ }
+
+ block = mdp->heapinfo[block].free_block.next;
+ if (block == start) {
+ /* Need to get more from the system. Check to see if
+ the new core will be contiguous with the final free
+ block; if so we don't need to get as much. */
+ block = mdp->heapinfo[0].free_block.prev;
+ lastblocks = mdp->heapinfo[block].free_block.size;
+ if (mdp->heaplimit != 0 &&
+ block + lastblocks == mdp->heaplimit &&
+ mmorecore(mdp, 0) == ADDRESS(block + lastblocks) &&
+ (register_morecore(mdp, (blocks - lastblocks) * BLOCKSIZE)) != NULL) {
+ /* Which block we are extending (the `final free
+ block' referred to above) might have changed, if
+ it got combined with a freed info table. */
+ block = mdp->heapinfo[0].free_block.prev;
+
+ mdp->heapinfo[block].free_block.size += (blocks - lastblocks);
+ continue;
+ }
+ result = register_morecore(mdp, blocks * BLOCKSIZE);
+
+ block = BLOCK(result);
+ for (it=0;it<blocks;it++){
+ mdp->heapinfo[block+it].type = 0;
+ mdp->heapinfo[block+it].busy_block.busy_size = 0;
+ mdp->heapinfo[block+it].busy_block.ignore = 0;
+ }
+ mdp->heapinfo[block].busy_block.size = blocks;
+ mdp->heapinfo[block].busy_block.busy_size = requested_size;
+ //mdp->heapinfo[block].busy_block.bt_size=xbt_backtrace_no_malloc(mdp->heapinfo[block].busy_block.bt,XBT_BACKTRACE_SIZE);
+ //mdp->heapinfo[block].busy_block.bt_size = xbt_libunwind_backtrace(mdp->heapinfo[block].busy_block.bt,XBT_BACKTRACE_SIZE);
+ mdp -> heapstats.chunks_used++;
+ mdp -> heapstats.bytes_used += blocks * BLOCKSIZE;
+
+ return result;
+ }
+ /* Need large block(s), but found some in the existing heap */