X-Git-Url: http://bilbo.iut-bm.univ-fcomte.fr/pub/gitweb/simgrid.git/blobdiff_plain/6ced0985aaa3ec81be59edb34c1753a9c9b9751a..2e60fe3cfd5cf5305888fcca0ae19700d808bb23:/src/xbt/mmalloc/mmprivate.h diff --git a/src/xbt/mmalloc/mmprivate.h b/src/xbt/mmalloc/mmprivate.h index 97841313e8..2d7af4a2b6 100644 --- a/src/xbt/mmalloc/mmprivate.h +++ b/src/xbt/mmalloc/mmprivate.h @@ -1,6 +1,6 @@ /* Declarations for `mmalloc' and friends. */ -/* Copyright (c) 2010-2019. The SimGrid Team. All rights reserved. */ +/* Copyright (c) 2010-2020. The SimGrid Team. All rights reserved. */ /* This program is free software; you can redistribute it and/or modify it * under the terms of the license (GNU LGPL) which comes with this package. */ @@ -25,13 +25,6 @@ #include #include -#ifndef MIN -#define MIN(a, b) ((a) < (b) ? (a) : (b)) -#endif -#ifndef MAX -#define MAX(a, b) ((a) > (b) ? (a) : (b)) -#endif - #ifdef HAVE_LIMITS_H # include #else @@ -58,14 +51,13 @@ #define BLOCKIFY(SIZE) (((SIZE) + BLOCKSIZE - 1) / BLOCKSIZE) /* We keep fragment-specific meta-data for introspection purposes, and these - * information are kept in fixed lenght arrays. Here is the computation of + * information are kept in fixed length arrays. Here is the computation of * that size. * * Never make SMALLEST_POSSIBLE_MALLOC smaller than sizeof(list) because we * need to enlist the free fragments. */ -//#define SMALLEST_POSSIBLE_MALLOC (sizeof(struct list)) #define SMALLEST_POSSIBLE_MALLOC (16*sizeof(struct list)) #define MAX_FRAGMENT_PER_BLOCK (BLOCKSIZE / SMALLEST_POSSIBLE_MALLOC) @@ -102,7 +94,7 @@ #define ADDRESS(B) ((void*) (((ADDR2UINT(B)) - 1) * BLOCKSIZE + (char*) mdp -> heapbase)) -SG_BEGIN_DECL() +SG_BEGIN_DECL /* Doubly linked lists of free fragments. */ struct list { @@ -160,14 +152,11 @@ typedef struct { struct { size_t nfree; /* Free fragments in a fragmented block. */ ssize_t frag_size[MAX_FRAGMENT_PER_BLOCK]; - //void *bt[MAX_FRAGMENT_PER_BLOCK][XBT_BACKTRACE_SIZE]; /* Where it was malloced (or realloced lastly) */ int ignore[MAX_FRAGMENT_PER_BLOCK]; } busy_frag; struct { size_t size; /* Size (in blocks) of a large cluster. */ size_t busy_size; /* Actually used space, in bytes */ - //void *bt[XBT_BACKTRACE_SIZE]; /* Where it was malloced (or realloced lastly) */ - //int bt_size; int ignore; } busy_block; /* Heap information for a free block (that may be the first of a free cluster). */ @@ -187,7 +176,6 @@ typedef struct { * if such a file exists. * */ struct mdesc { - /** @brief Mutex locking the access to the heap */ pthread_mutex_t mutex; @@ -237,7 +225,7 @@ struct mdesc { /* @brief List of all blocks containing free fragments of a given size. * - * The array indice is the log2 of requested size. + * The array index is the log2 of requested size. * Actually only the sizes 8->11 seem to be used, but who cares? */ s_xbt_swag_t fraghead[BLOCKLOG]; @@ -272,7 +260,6 @@ struct mdesc { /* @brief Instrumentation */ struct mstats heapstats; - }; /* Bits to look at in the malloc descriptor flags word */ @@ -287,7 +274,7 @@ XBT_PUBLIC_DATA struct mdesc* __mmalloc_default_mdp; /* Remap a mmalloc region that was previously mapped. */ -XBT_PUBLIC void* __mmalloc_remap_core(xbt_mheap_t mdp); +XBT_PUBLIC void* __mmalloc_remap_core(const s_xbt_mheap_t* mdp); XBT_PUBLIC void* mmorecore(struct mdesc* mdp, ssize_t size); @@ -297,13 +284,13 @@ XBT_PUBLIC void* mmorecore(struct mdesc* mdp, ssize_t size); * in a model-checking enabled tree. Without this protection, our malloc * implementation will not like multi-threading AT ALL. */ -#define LOCK(mdp) pthread_mutex_lock(&mdp->mutex) -#define UNLOCK(mdp) pthread_mutex_unlock(&mdp->mutex) +#define LOCK(mdp) pthread_mutex_lock(&(mdp)->mutex) +#define UNLOCK(mdp) pthread_mutex_unlock(&(mdp)->mutex) XBT_PRIVATE int malloc_use_mmalloc(void); XBT_PRIVATE size_t mmalloc_get_bytes_used_remote(size_t heaplimit, const malloc_info* heapinfo); -SG_END_DECL() +SG_END_DECL #endif