| |
| /*--------------------------------------------------------------------*/ |
| /*--- An implementation of malloc/free for the client. ---*/ |
| /*--- vg_clientmalloc.c ---*/ |
| /*--------------------------------------------------------------------*/ |
| |
| /* |
| This file is part of Valgrind, an x86 protected-mode emulator |
| designed for debugging and profiling binaries on x86-Unixes. |
| |
| Copyright (C) 2000-2002 Julian Seward |
| jseward@acm.org |
| Julian_Seward@muraroa.demon.co.uk |
| |
| This program is free software; you can redistribute it and/or |
| modify it under the terms of the GNU General Public License as |
| published by the Free Software Foundation; either version 2 of the |
| License, or (at your option) any later version. |
| |
| This program is distributed in the hope that it will be useful, but |
| WITHOUT ANY WARRANTY; without even the implied warranty of |
| MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| General Public License for more details. |
| |
| You should have received a copy of the GNU General Public License |
| along with this program; if not, write to the Free Software |
| Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA |
| 02111-1307, USA. |
| |
| The GNU General Public License is contained in the file LICENSE. |
| */ |
| |
| #include "vg_include.h" |
| |
| |
| /*------------------------------------------------------------*/ |
| /*--- Defns ---*/ |
| /*------------------------------------------------------------*/ |
| |
| /* #define DEBUG_CLIENTMALLOC */ |
| |
| /* Holds malloc'd but not freed blocks. */ |
| #define VG_MALLOCLIST_NO(aa) (((UInt)(aa)) % VG_N_MALLOCLISTS) |
| static ShadowChunk* vg_malloclist[VG_N_MALLOCLISTS]; |
| static Bool vg_client_malloc_init_done = False; |
| |
| /* Holds blocks after freeing. */ |
| static ShadowChunk* vg_freed_list_start = NULL; |
| static ShadowChunk* vg_freed_list_end = NULL; |
| static Int vg_freed_list_volume = 0; |
| |
| /* Stats ... */ |
| static UInt vg_cmalloc_n_mallocs = 0; |
| static UInt vg_cmalloc_n_frees = 0; |
| static UInt vg_cmalloc_bs_mallocd = 0; |
| |
| static UInt vg_mlist_frees = 0; |
| static UInt vg_mlist_tries = 0; |
| |
| |
| /*------------------------------------------------------------*/ |
| /*--- Fns ---*/ |
| /*------------------------------------------------------------*/ |
| |
| /* Allocate a suitably-sized array, copy all the malloc-d block |
| shadows into it, and return both the array and the size of it. |
| This is used by the memory-leak detector. |
| */ |
| ShadowChunk** VG_(get_malloc_shadows) ( /*OUT*/ UInt* n_shadows ) |
| { |
| UInt i, scn; |
| ShadowChunk** arr; |
| ShadowChunk* sc; |
| *n_shadows = 0; |
| for (scn = 0; scn < VG_N_MALLOCLISTS; scn++) { |
| for (sc = vg_malloclist[scn]; sc != NULL; sc = sc->next) { |
| (*n_shadows)++; |
| } |
| } |
| if (*n_shadows == 0) return NULL; |
| |
| arr = VG_(malloc)( VG_AR_PRIVATE, |
| *n_shadows * sizeof(ShadowChunk*) ); |
| |
| i = 0; |
| for (scn = 0; scn < VG_N_MALLOCLISTS; scn++) { |
| for (sc = vg_malloclist[scn]; sc != NULL; sc = sc->next) { |
| arr[i++] = sc; |
| } |
| } |
| vg_assert(i == *n_shadows); |
| return arr; |
| } |
| |
| static void client_malloc_init ( void ) |
| { |
| UInt ml_no; |
| if (vg_client_malloc_init_done) return; |
| for (ml_no = 0; ml_no < VG_N_MALLOCLISTS; ml_no++) |
| vg_malloclist[ml_no] = NULL; |
| vg_client_malloc_init_done = True; |
| } |
| |
| |
| static __attribute__ ((unused)) |
| Int count_freelist ( void ) |
| { |
| ShadowChunk* sc; |
| Int n = 0; |
| for (sc = vg_freed_list_start; sc != NULL; sc = sc->next) |
| n++; |
| return n; |
| } |
| |
| static __attribute__ ((unused)) |
| Int count_malloclists ( void ) |
| { |
| ShadowChunk* sc; |
| UInt ml_no; |
| Int n = 0; |
| for (ml_no = 0; ml_no < VG_N_MALLOCLISTS; ml_no++) |
| for (sc = vg_malloclist[ml_no]; sc != NULL; sc = sc->next) |
| n++; |
| return n; |
| } |
| |
| static __attribute__ ((unused)) |
| void freelist_sanity ( void ) |
| { |
| ShadowChunk* sc; |
| Int n = 0; |
| /* VG_(printf)("freelist sanity\n"); */ |
| for (sc = vg_freed_list_start; sc != NULL; sc = sc->next) |
| n += sc->size; |
| vg_assert(n == vg_freed_list_volume); |
| } |
| |
| /* Remove sc from malloc list # sc. It is an unchecked error for |
| sc not to be present in the list. |
| */ |
| static void remove_from_malloclist ( UInt ml_no, ShadowChunk* sc ) |
| { |
| ShadowChunk *sc1, *sc2; |
| if (sc == vg_malloclist[ml_no]) { |
| vg_malloclist[ml_no] = vg_malloclist[ml_no]->next; |
| } else { |
| sc1 = vg_malloclist[ml_no]; |
| vg_assert(sc1 != NULL); |
| sc2 = sc1->next; |
| while (sc2 != sc) { |
| vg_assert(sc2 != NULL); |
| sc1 = sc2; |
| sc2 = sc2->next; |
| } |
| vg_assert(sc1->next == sc); |
| vg_assert(sc2 == sc); |
| sc1->next = sc2->next; |
| } |
| } |
| |
| |
| /* Put a shadow chunk on the freed blocks queue, possibly freeing up |
| some of the oldest blocks in the queue at the same time. */ |
| |
| static void add_to_freed_queue ( ShadowChunk* sc ) |
| { |
| ShadowChunk* sc1; |
| |
| /* Put it at the end of the freed list */ |
| if (vg_freed_list_end == NULL) { |
| vg_assert(vg_freed_list_start == NULL); |
| vg_freed_list_end = vg_freed_list_start = sc; |
| vg_freed_list_volume = sc->size; |
| } else { |
| vg_assert(vg_freed_list_end->next == NULL); |
| vg_freed_list_end->next = sc; |
| vg_freed_list_end = sc; |
| vg_freed_list_volume += sc->size; |
| } |
| sc->next = NULL; |
| |
| /* Release enough of the oldest blocks to bring the free queue |
| volume below vg_clo_freelist_vol. */ |
| |
| while (vg_freed_list_volume > VG_(clo_freelist_vol)) { |
| /* freelist_sanity(); */ |
| vg_assert(vg_freed_list_start != NULL); |
| vg_assert(vg_freed_list_end != NULL); |
| |
| sc1 = vg_freed_list_start; |
| vg_freed_list_volume -= sc1->size; |
| /* VG_(printf)("volume now %d\n", vg_freed_list_volume); */ |
| vg_assert(vg_freed_list_volume >= 0); |
| |
| if (vg_freed_list_start == vg_freed_list_end) { |
| vg_freed_list_start = vg_freed_list_end = NULL; |
| } else { |
| vg_freed_list_start = sc1->next; |
| } |
| sc1->next = NULL; /* just paranoia */ |
| VG_(free)(VG_AR_CLIENT, (void*)(sc1->data)); |
| VG_(free)(VG_AR_PRIVATE, sc1); |
| } |
| } |
| |
| |
| /* Allocate a user-chunk of size bytes. Also allocate its shadow |
| block, make the shadow block point at the user block. Put the |
| shadow chunk on the appropriate list, and set all memory |
| protections correctly. */ |
| |
| static ShadowChunk* client_malloc_shadow ( UInt align, UInt size, |
| VgAllocKind kind ) |
| { |
| ShadowChunk* sc; |
| Addr p; |
| UInt ml_no; |
| |
| # ifdef DEBUG_CLIENTMALLOC |
| VG_(printf)("[m %d, f %d (%d)] client_malloc_shadow ( al %d, sz %d )\n", |
| count_malloclists(), |
| count_freelist(), vg_freed_list_volume, |
| align, size ); |
| # endif |
| |
| if (align == 0) |
| p = (Addr)VG_(malloc)(VG_AR_CLIENT, size); |
| else |
| p = (Addr)VG_(malloc_aligned)(VG_AR_CLIENT, align, size); |
| |
| sc = VG_(malloc)(VG_AR_PRIVATE, sizeof(ShadowChunk)); |
| sc->where = VG_(get_ExeContext)(True); |
| sc->size = size; |
| sc->allockind = kind; |
| sc->data = p; |
| ml_no = VG_MALLOCLIST_NO(p); |
| sc->next = vg_malloclist[ml_no]; |
| vg_malloclist[ml_no] = sc; |
| |
| VGM_(make_writable)(p, size); |
| VGM_(make_noaccess)(p + size, |
| VG_AR_CLIENT_REDZONE_SZB); |
| VGM_(make_noaccess)(p - VG_AR_CLIENT_REDZONE_SZB, |
| VG_AR_CLIENT_REDZONE_SZB); |
| |
| return sc; |
| } |
| |
| |
| /* Allocate memory, noticing whether or not we are doing the full |
| instrumentation thing. */ |
| |
| void* VG_(client_malloc) ( UInt size, UInt raw_alloc_kind ) |
| { |
| ShadowChunk* sc; |
| VgAllocKind kind; |
| |
| VGP_PUSHCC(VgpCliMalloc); |
| client_malloc_init(); |
| # ifdef DEBUG_CLIENTMALLOC |
| VG_(printf)("[m %d, f %d (%d)] client_malloc ( %d, %x )\n", |
| count_malloclists(), |
| count_freelist(), vg_freed_list_volume, |
| size, raw_alloc_kind ); |
| # endif |
| if (!VG_(clo_instrument)) { |
| VGP_POPCC; |
| return VG_(malloc) ( VG_AR_CLIENT, size ); |
| } |
| switch (raw_alloc_kind) { |
| case 0x4002: kind = Vg_AllocNewVec; break; |
| case 0x4001: kind = Vg_AllocNew; break; |
| case 0x4000: /* malloc */ |
| case 6666: /* calloc */ |
| kind = Vg_AllocMalloc; break; |
| default: /* should not happen */ |
| /* therefore we make sure it doesn't -- JRS */ |
| VG_(panic)("VG_(client_malloc): raw_alloc_kind"); |
| break; /*NOTREACHED*/ |
| } |
| sc = client_malloc_shadow ( 0, size, kind ); |
| VGP_POPCC; |
| return (void*)(sc->data); |
| } |
| |
| |
| void* VG_(client_memalign) ( UInt align, UInt size ) |
| { |
| ShadowChunk* sc; |
| VGP_PUSHCC(VgpCliMalloc); |
| client_malloc_init(); |
| # ifdef DEBUG_CLIENTMALLOC |
| VG_(printf)("[m %d, f %d (%d)] client_memalign ( al %d, sz %d )\n", |
| count_malloclists(), |
| count_freelist(), vg_freed_list_volume, |
| align, size ); |
| # endif |
| if (!VG_(clo_instrument)) { |
| VGP_POPCC; |
| return VG_(malloc_aligned) ( VG_AR_CLIENT, align, size ); |
| } |
| sc = client_malloc_shadow ( align, size, Vg_AllocMalloc ); |
| VGP_POPCC; |
| return (void*)(sc->data); |
| } |
| |
| |
| void VG_(client_free) ( void* ptrV, UInt raw_alloc_kind ) |
| { |
| ShadowChunk* sc; |
| UInt ml_no; |
| VgAllocKind kind; |
| |
| VGP_PUSHCC(VgpCliMalloc); |
| client_malloc_init(); |
| # ifdef DEBUG_CLIENTMALLOC |
| VG_(printf)("[m %d, f %d (%d)] client_free ( %p, %x )\n", |
| count_malloclists(), |
| count_freelist(), vg_freed_list_volume, |
| ptrV, raw_alloc_kind ); |
| # endif |
| if (!VG_(clo_instrument)) { |
| VGP_POPCC; |
| VG_(free) ( VG_AR_CLIENT, ptrV ); |
| return; |
| } |
| |
| /* first, see if ptrV is one vg_client_malloc gave out. */ |
| ml_no = VG_MALLOCLIST_NO(ptrV); |
| vg_mlist_frees++; |
| for (sc = vg_malloclist[ml_no]; sc != NULL; sc = sc->next) { |
| vg_mlist_tries++; |
| if ((Addr)ptrV == sc->data) |
| break; |
| } |
| |
| if (sc == NULL) { |
| VG_(record_free_error) ( (Addr)ptrV ); |
| VGP_POPCC; |
| return; |
| } |
| |
| switch (raw_alloc_kind) { |
| case 0x5002: kind = Vg_AllocNewVec; break; |
| case 0x5001: kind = Vg_AllocNew; break; |
| case 0x5000: |
| default: |
| kind = Vg_AllocMalloc; |
| /* should only happen if bug in client code */ |
| break; |
| } |
| |
| /* check if its a matching free() / delete / delete [] */ |
| if (kind != sc->allockind) |
| VG_(record_freemismatch_error) ( (Addr) ptrV ); |
| |
| /* Remove the shadow chunk from the mallocd list. */ |
| remove_from_malloclist ( ml_no, sc ); |
| |
| /* Declare it inaccessible. */ |
| VGM_(make_noaccess) ( sc->data - VG_AR_CLIENT_REDZONE_SZB, |
| sc->size + 2*VG_AR_CLIENT_REDZONE_SZB ); |
| VGM_(make_noaccess) ( (Addr)sc, sizeof(ShadowChunk) ); |
| sc->where = VG_(get_ExeContext)(True); |
| |
| /* Put it out of harm's way for a while. */ |
| add_to_freed_queue ( sc ); |
| VGP_POPCC; |
| } |
| |
| |
| |
| void* VG_(client_calloc) ( UInt nmemb, UInt size1 ) |
| { |
| ShadowChunk* sc; |
| Addr p; |
| UInt size, i, ml_no; |
| |
| VGP_PUSHCC(VgpCliMalloc); |
| client_malloc_init(); |
| |
| # ifdef DEBUG_CLIENTMALLOC |
| VG_(printf)("[m %d, f %d (%d)] client_calloc ( %d, %d )\n", |
| count_malloclists(), |
| count_freelist(), vg_freed_list_volume, |
| nmemb, size1 ); |
| # endif |
| |
| if (!VG_(clo_instrument)) { |
| VGP_POPCC; |
| return VG_(calloc) ( VG_AR_CLIENT, nmemb, size1 ); |
| } |
| |
| size = nmemb * size1; |
| p = (Addr)VG_(malloc)(VG_AR_CLIENT, size); |
| sc = VG_(malloc)(VG_AR_PRIVATE, sizeof(ShadowChunk)); |
| sc->where = VG_(get_ExeContext)(True); |
| sc->size = size; |
| sc->allockind = Vg_AllocMalloc; /* its a lie - but true. eat this :) */ |
| sc->data = p; |
| ml_no = VG_MALLOCLIST_NO(p); |
| sc->next = vg_malloclist[ml_no]; |
| vg_malloclist[ml_no] = sc; |
| |
| VGM_(make_readable)(p, size); |
| VGM_(make_noaccess)(p + size, |
| VG_AR_CLIENT_REDZONE_SZB); |
| VGM_(make_noaccess)(p - VG_AR_CLIENT_REDZONE_SZB, |
| VG_AR_CLIENT_REDZONE_SZB); |
| |
| for (i = 0; i < size; i++) ((UChar*)p)[i] = 0; |
| |
| VGP_POPCC; |
| return (void*)p; |
| } |
| |
| |
| void* VG_(client_realloc) ( void* ptrV, UInt size_new ) |
| { |
| ShadowChunk *sc, *sc_new; |
| UInt i, ml_no; |
| |
| VGP_PUSHCC(VgpCliMalloc); |
| client_malloc_init(); |
| |
| # ifdef DEBUG_CLIENTMALLOC |
| VG_(printf)("[m %d, f %d (%d)] client_realloc ( %p, %d )\n", |
| count_malloclists(), |
| count_freelist(), vg_freed_list_volume, |
| ptrV, size_new ); |
| # endif |
| |
| if (!VG_(clo_instrument)) { |
| vg_assert(ptrV != NULL && size_new != 0); |
| VGP_POPCC; |
| return VG_(realloc) ( VG_AR_CLIENT, ptrV, size_new ); |
| } |
| |
| /* First try and find the block. */ |
| ml_no = VG_MALLOCLIST_NO(ptrV); |
| for (sc = vg_malloclist[ml_no]; sc != NULL; sc = sc->next) { |
| if ((Addr)ptrV == sc->data) |
| break; |
| } |
| |
| if (sc == NULL) { |
| VG_(record_free_error) ( (Addr)ptrV ); |
| /* Perhaps we should keep going regardless. */ |
| VGP_POPCC; |
| return NULL; |
| } |
| |
| if (sc->allockind != Vg_AllocMalloc) { |
| /* can not realloc a range that was allocated with new or new [] */ |
| VG_(record_freemismatch_error) ( (Addr)ptrV ); |
| /* but keep going anyway */ |
| } |
| |
| if (sc->size == size_new) { |
| /* size unchanged */ |
| VGP_POPCC; |
| return ptrV; |
| } |
| if (sc->size > size_new) { |
| /* new size is smaller */ |
| VGM_(make_noaccess)( sc->data + size_new, |
| sc->size - size_new ); |
| sc->size = size_new; |
| VGP_POPCC; |
| return ptrV; |
| } else { |
| /* new size is bigger */ |
| sc_new = client_malloc_shadow ( 0, size_new, Vg_AllocMalloc ); |
| for (i = 0; i < sc->size; i++) |
| ((UChar*)(sc_new->data))[i] = ((UChar*)(sc->data))[i]; |
| VGM_(copy_address_range_perms) ( |
| sc->data, sc_new->data, sc->size ); |
| remove_from_malloclist ( VG_MALLOCLIST_NO(sc->data), sc ); |
| VGM_(make_noaccess) ( sc->data - VG_AR_CLIENT_REDZONE_SZB, |
| sc->size + 2*VG_AR_CLIENT_REDZONE_SZB ); |
| VGM_(make_noaccess) ( (Addr)sc, sizeof(ShadowChunk) ); |
| add_to_freed_queue ( sc ); |
| VGP_POPCC; |
| return (void*)sc_new->data; |
| } |
| } |
| |
| |
| void VG_(clientmalloc_done) ( void ) |
| { |
| UInt nblocks, nbytes, ml_no; |
| ShadowChunk* sc; |
| |
| client_malloc_init(); |
| |
| nblocks = nbytes = 0; |
| |
| for (ml_no = 0; ml_no < VG_N_MALLOCLISTS; ml_no++) { |
| for (sc = vg_malloclist[ml_no]; sc != NULL; sc = sc->next) { |
| nblocks ++; |
| nbytes += sc->size; |
| } |
| } |
| |
| if (VG_(clo_verbosity) == 0) |
| return; |
| |
| VG_(message)(Vg_UserMsg, |
| "malloc/free: in use at exit: %d bytes in %d blocks.", |
| nbytes, nblocks); |
| VG_(message)(Vg_UserMsg, |
| "malloc/free: %d allocs, %d frees, %d bytes allocated.", |
| vg_cmalloc_n_mallocs, |
| vg_cmalloc_n_frees, vg_cmalloc_bs_mallocd); |
| if (!VG_(clo_leak_check)) |
| VG_(message)(Vg_UserMsg, |
| "For a detailed leak analysis, rerun with: --leak-check=yes"); |
| if (0) |
| VG_(message)(Vg_DebugMsg, |
| "free search: %d tries, %d frees", |
| vg_mlist_tries, |
| vg_mlist_frees ); |
| if (VG_(clo_verbosity) > 1) |
| VG_(message)(Vg_UserMsg, ""); |
| } |
| |
| |
| /* Describe an address as best you can, for error messages, |
| putting the result in ai. */ |
| |
| void VG_(describe_addr) ( Addr a, AddrInfo* ai ) |
| { |
| ShadowChunk* sc; |
| UInt ml_no; |
| Bool ok; |
| |
| /* Perhaps it's a user-def'd block ? */ |
| ok = VG_(client_perm_maybe_describe)( a, ai ); |
| if (ok) |
| return; |
| /* Perhaps it's on the stack? */ |
| if (VG_(is_plausible_stack_addr)(a) |
| && a >= (Addr)VG_(baseBlock)[VGOFF_(m_esp)]) { |
| ai->akind = Stack; |
| return; |
| } |
| /* Search for a freed block which might bracket it. */ |
| for (sc = vg_freed_list_start; sc != NULL; sc = sc->next) { |
| if (sc->data - VG_AR_CLIENT_REDZONE_SZB <= a |
| && a < sc->data + sc->size + VG_AR_CLIENT_REDZONE_SZB) { |
| ai->akind = Freed; |
| ai->blksize = sc->size; |
| ai->rwoffset = (Int)(a) - (Int)(sc->data); |
| ai->lastchange = sc->where; |
| return; |
| } |
| } |
| /* Search for a mallocd block which might bracket it. */ |
| for (ml_no = 0; ml_no < VG_N_MALLOCLISTS; ml_no++) { |
| for (sc = vg_malloclist[ml_no]; sc != NULL; sc = sc->next) { |
| if (sc->data - VG_AR_CLIENT_REDZONE_SZB <= a |
| && a < sc->data + sc->size + VG_AR_CLIENT_REDZONE_SZB) { |
| ai->akind = Mallocd; |
| ai->blksize = sc->size; |
| ai->rwoffset = (Int)(a) - (Int)(sc->data); |
| ai->lastchange = sc->where; |
| return; |
| } |
| } |
| } |
| /* Clueless ... */ |
| ai->akind = Unknown; |
| return; |
| } |
| |
| /*------------------------------------------------------------*/ |
| /*--- Replace the C library versions with our own. Hairy. ---*/ |
| /*------------------------------------------------------------*/ |
| |
| /* Below are new versions of malloc, __builtin_new, free, |
| __builtin_delete, calloc and realloc. |
| |
| malloc, __builtin_new, free, __builtin_delete, calloc and realloc |
| can be entered either on the real CPU or the simulated one. If on |
| the real one, this is because the dynamic linker is running the |
| static initialisers for C++, before starting up Valgrind itself. |
| In this case it is safe to route calls through to |
| VG_(malloc)/vg_free, since that is self-initialising. |
| |
| Once Valgrind is initialised, vg_running_on_simd_CPU becomes True. |
| The call needs to be transferred from the simulated CPU back to the |
| real one and routed to the vg_client_* functions. To do that, the |
| args are passed to vg_trap_here, which the simulator detects. The |
| bogus epilogue fn call is to guarantee that gcc doesn't tailcall |
| vg_trap_here, since that would cause the simulator's detection to |
| fail -- it only checks the targets of call transfers, not jumps. |
| And of course we have to be sure gcc won't inline either the |
| vg_trap_here or vg_bogus_epilogue. Ha ha ha. What a mess. |
| */ |
| |
| /* Place afterwards to guarantee it won't get inlined ... */ |
| static UInt vg_trap_here_WRAPPER ( UInt arg1, UInt arg2, UInt what_to_do ); |
| static void vg_bogus_epilogue ( void ); |
| |
| /* ALL calls to malloc wind up here. */ |
| void* malloc ( UInt n ) |
| { |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)("malloc[simd=%d](%d)", |
| (UInt)VG_(running_on_simd_CPU), n ); |
| |
| if (VG_(clo_sloppy_malloc)) { while ((n % 4) > 0) n++; } |
| |
| vg_cmalloc_n_mallocs ++; |
| vg_cmalloc_bs_mallocd += n; |
| |
| if (VG_(running_on_simd_CPU)) { |
| UInt v = vg_trap_here_WRAPPER ( 0, n, 0x4000 ); |
| vg_bogus_epilogue(); |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)(" = %p\n", v ); |
| return (void*)v; |
| } else { |
| void* v = VG_(malloc)(VG_AR_CLIENT, n); |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)(" = %p\n", v ); |
| return (void*)v; |
| } |
| } |
| |
| void* __builtin_new ( UInt n ) |
| { |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)("__builtin_new[simd=%d](%d)", |
| (UInt)VG_(running_on_simd_CPU), n ); |
| |
| if (VG_(clo_sloppy_malloc)) { while ((n % 4) > 0) n++; } |
| |
| vg_cmalloc_n_mallocs++; |
| vg_cmalloc_bs_mallocd += n; |
| |
| if (VG_(running_on_simd_CPU)) { |
| UInt v = vg_trap_here_WRAPPER ( 0, n, 0x4001 ); |
| vg_bogus_epilogue(); |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)(" = %p\n", v ); |
| return (void*)v; |
| } else { |
| void* v = VG_(malloc)(VG_AR_CLIENT, n); |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)(" = %p\n", v ); |
| return v; |
| } |
| } |
| |
| void* __builtin_vec_new ( Int n ) |
| { |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)("__builtin_vec_new[simd=%d](%d)", |
| (UInt)VG_(running_on_simd_CPU), n ); |
| |
| if (VG_(clo_sloppy_malloc)) { while ((n % 4) > 0) n++; } |
| |
| vg_cmalloc_n_mallocs++; |
| vg_cmalloc_bs_mallocd += n; |
| |
| if (VG_(running_on_simd_CPU)) { |
| UInt v = vg_trap_here_WRAPPER ( 0, n, 0x4002 ); |
| vg_bogus_epilogue(); |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)(" = %p\n", v ); |
| return (void*)v; |
| } else { |
| void* v = VG_(malloc)(VG_AR_CLIENT, n); |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)(" = %p\n", v ); |
| return v; |
| } |
| } |
| |
| void free ( void* p ) |
| { |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)("free[simd=%d](%p)\n", |
| (UInt)VG_(running_on_simd_CPU), p ); |
| vg_cmalloc_n_frees ++; |
| |
| if (p == NULL) |
| return; |
| if (VG_(running_on_simd_CPU)) { |
| (void)vg_trap_here_WRAPPER ( 0, (UInt)p, 0x5000 ); |
| vg_bogus_epilogue(); |
| } else { |
| VG_(free)(VG_AR_CLIENT, p); |
| } |
| } |
| |
| void __builtin_delete ( void* p ) |
| { |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)("__builtin_delete[simd=%d](%p)\n", |
| (UInt)VG_(running_on_simd_CPU), p ); |
| vg_cmalloc_n_frees ++; |
| |
| if (p == NULL) |
| return; |
| if (VG_(running_on_simd_CPU)) { |
| (void)vg_trap_here_WRAPPER ( 0, (UInt)p, 0x5001 ); |
| vg_bogus_epilogue(); |
| } else { |
| VG_(free)(VG_AR_CLIENT, p); |
| } |
| } |
| |
| void __builtin_vec_delete ( void* p ) |
| { |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)("__builtin_vec_delete[simd=%d](%p)\n", |
| (UInt)VG_(running_on_simd_CPU), p ); |
| vg_cmalloc_n_frees ++; |
| |
| if (p == NULL) |
| return; |
| if (VG_(running_on_simd_CPU)) { |
| (void)vg_trap_here_WRAPPER ( 0, (UInt)p, 0x5002 ); |
| vg_bogus_epilogue(); |
| } else { |
| VG_(free)(VG_AR_CLIENT, p); |
| } |
| } |
| |
| void* calloc ( UInt nmemb, UInt size ) |
| { |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)("calloc[simd=%d](%d,%d)", |
| (UInt)VG_(running_on_simd_CPU), nmemb, size ); |
| vg_cmalloc_n_mallocs ++; |
| vg_cmalloc_bs_mallocd += size * nmemb; |
| |
| if (VG_(running_on_simd_CPU)) { |
| UInt v = vg_trap_here_WRAPPER ( nmemb, size, 6666 ); |
| vg_bogus_epilogue(); |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)(" = %p\n", v ); |
| return (void*)v; |
| } else { |
| void* v = VG_(calloc)(VG_AR_CLIENT, nmemb, size); |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)(" = %p\n", v ); |
| return v; |
| } |
| } |
| |
| void* realloc ( void* ptrV, UInt new_size ) |
| { |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)("realloc[simd=%d](%p,%d)", |
| (UInt)VG_(running_on_simd_CPU), ptrV, new_size ); |
| |
| if (VG_(clo_sloppy_malloc)) |
| { while ((new_size % 4) > 0) new_size++; } |
| |
| vg_cmalloc_n_frees ++; |
| vg_cmalloc_n_mallocs ++; |
| vg_cmalloc_bs_mallocd += new_size; |
| |
| if (ptrV == NULL) |
| return malloc(new_size); |
| if (new_size == 0) { |
| free(ptrV); |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)(" = 0\n" ); |
| return NULL; |
| } |
| if (VG_(running_on_simd_CPU)) { |
| UInt v = vg_trap_here_WRAPPER ( (UInt)ptrV, new_size, 7777 ); |
| vg_bogus_epilogue(); |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)(" = %p\n", v ); |
| return (void*)v; |
| } else { |
| void* v = VG_(realloc)(VG_AR_CLIENT, ptrV, new_size); |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)(" = %p\n", v ); |
| return v; |
| } |
| } |
| |
| void* memalign ( Int alignment, Int n ) |
| { |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)("memalign[simd=%d](al %d, size %d)", |
| (UInt)VG_(running_on_simd_CPU), alignment, n ); |
| |
| if (VG_(clo_sloppy_malloc)) { while ((n % 4) > 0) n++; } |
| |
| vg_cmalloc_n_mallocs ++; |
| vg_cmalloc_bs_mallocd += n; |
| |
| if (VG_(running_on_simd_CPU)) { |
| UInt v = vg_trap_here_WRAPPER ( alignment, n, 8888 ); |
| vg_bogus_epilogue(); |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)(" = %p\n", v ); |
| return (void*)v; |
| } else { |
| void* v = VG_(malloc_aligned)(VG_AR_CLIENT, alignment, n); |
| if (VG_(clo_trace_malloc)) |
| VG_(printf)(" = %p\n", v ); |
| return (void*)v; |
| } |
| } |
| |
| void* valloc ( Int size ) |
| { |
| return memalign(VKI_BYTES_PER_PAGE, size); |
| } |
| |
| |
| /* Various compatibility wrapper functions, for glibc and libstdc++. */ |
| void cfree ( void* p ) |
| { |
| free ( p ); |
| } |
| |
| void* mallinfo ( void ) |
| { |
| VG_(message)(Vg_UserMsg, |
| "Warning: incorrectly-handled call to mallinfo()"); |
| return NULL; |
| } |
| |
| |
| |
| int mallopt ( int cmd, int value ) |
| { |
| /* In glibc-2.2.4, 1 denoted a successful return value for mallopt */ |
| return 1; |
| } |
| |
| |
| /* Bomb out if we get any of these. */ |
| void pvalloc ( void ) |
| { VG_(panic)("call to pvalloc\n"); } |
| |
| void malloc_stats ( void ) |
| { VG_(panic)("call to malloc_stats\n"); } |
| void malloc_usable_size ( void ) |
| { VG_(panic)("call to malloc_usable_size\n"); } |
| void malloc_trim ( void ) |
| { VG_(panic)("call to malloc_trim\n"); } |
| void malloc_get_state ( void ) |
| { VG_(panic)("call to malloc_get_state\n"); } |
| void malloc_set_state ( void ) |
| { VG_(panic)("call to malloc_set_state\n"); } |
| |
| |
| int __posix_memalign ( void **memptr, UInt alignment, UInt size ) |
| { |
| void *mem; |
| |
| /* Test whether the SIZE argument is valid. It must be a power of |
| two multiple of sizeof (void *). */ |
| if (size % sizeof (void *) != 0 || (size & (size - 1)) != 0) |
| return 22 /*EINVAL*/; |
| |
| mem = memalign (alignment, size); |
| |
| if (mem != NULL) { |
| *memptr = mem; |
| return 0; |
| } |
| |
| return 12 /*ENOMEM*/; |
| } |
| |
| |
| /*------------------------------------------------------------*/ |
| /*--- Magic supporting hacks. ---*/ |
| /*------------------------------------------------------------*/ |
| |
| extern UInt VG_(trap_here) ( UInt arg1, UInt arg2, UInt what_to_do ); |
| |
| static |
| UInt vg_trap_here_WRAPPER ( UInt arg1, UInt arg2, UInt what_to_do ) |
| { |
| /* The point of this idiocy is to make a plain, ordinary call to |
| vg_trap_here which vg_dispatch_when_CALL can spot. Left to |
| itself, with -fpic, gcc generates "call vg_trap_here@PLT" which |
| doesn't get spotted, for whatever reason. I guess I could check |
| _all_ control flow transfers, but that would be an undesirable |
| performance overhead. |
| |
| If you compile without -fpic, gcc generates the obvious call |
| insn, so the wrappers below will work if they just call |
| vg_trap_here. But I don't want to rule out building with -fpic, |
| hence this hack. Sigh. |
| */ |
| UInt v; |
| |
| # define WHERE_TO VG_(trap_here) |
| # define STRINGIFY(xx) __STRING(xx) |
| |
| asm("# call to vg_trap_here\n" |
| "\t pushl %3\n" |
| "\t pushl %2\n" |
| "\t pushl %1\n" |
| "\t call " STRINGIFY(WHERE_TO) "\n" |
| "\t addl $12, %%esp\n" |
| "\t movl %%eax, %0\n" |
| : "=r" (v) |
| : "r" (arg1), "r" (arg2), "r" (what_to_do) |
| : "eax", "esp", "cc", "memory"); |
| return v; |
| |
| # undef WHERE_TO |
| # undef STRINGIFY |
| } |
| |
| /* Last, but not least ... */ |
| void vg_bogus_epilogue ( void ) |
| { |
| /* Runs on simulated CPU only. */ |
| } |
| |
| UInt VG_(trap_here) ( UInt arg1, UInt arg2, UInt what_to_do ) |
| { |
| /* Calls to this fn are detected in vg_dispatch.S and are handled |
| specially. So this fn should never be entered. */ |
| VG_(panic)("vg_trap_here called!"); |
| return 0; /*NOTREACHED*/ |
| } |
| |
| |
| /*--------------------------------------------------------------------*/ |
| /*--- end vg_clientmalloc.c ---*/ |
| /*--------------------------------------------------------------------*/ |