blob: 399291d682149359ffa068aee887793b7442acad [file] [log] [blame]
/*--------------------------------------------------------------------*/
/*--- C startup stuff, reached from vg_startup.S. ---*/
/*--- vg_main.c ---*/
/*--------------------------------------------------------------------*/
/*
This file is part of Valgrind, an x86 protected-mode emulator
designed for debugging and profiling binaries on x86-Unixes.
Copyright (C) 2000-2002 Julian Seward
jseward@acm.org
This program is free software; you can redistribute it and/or
modify it under the terms of the GNU General Public License as
published by the Free Software Foundation; either version 2 of the
License, or (at your option) any later version.
This program is distributed in the hope that it will be useful, but
WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program; if not, write to the Free Software
Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA
02111-1307, USA.
The GNU General Public License is contained in the file LICENSE.
*/
#include "vg_include.h"
#include "vg_constants.h"
/* ---------------------------------------------------------------------
Compute offsets into baseBlock. See comments in vg_include.h.
------------------------------------------------------------------ */
/* The variables storing offsets. */
#define INVALID_OFFSET (-1)
Int VGOFF_(m_eax) = INVALID_OFFSET;
Int VGOFF_(m_ecx) = INVALID_OFFSET;
Int VGOFF_(m_edx) = INVALID_OFFSET;
Int VGOFF_(m_ebx) = INVALID_OFFSET;
Int VGOFF_(m_esp) = INVALID_OFFSET;
Int VGOFF_(m_ebp) = INVALID_OFFSET;
Int VGOFF_(m_esi) = INVALID_OFFSET;
Int VGOFF_(m_edi) = INVALID_OFFSET;
Int VGOFF_(m_eflags) = INVALID_OFFSET;
Int VGOFF_(m_fpustate) = INVALID_OFFSET;
Int VGOFF_(m_eip) = INVALID_OFFSET;
Int VGOFF_(spillslots) = INVALID_OFFSET;
Int VGOFF_(sh_eax) = INVALID_OFFSET;
Int VGOFF_(sh_ecx) = INVALID_OFFSET;
Int VGOFF_(sh_edx) = INVALID_OFFSET;
Int VGOFF_(sh_ebx) = INVALID_OFFSET;
Int VGOFF_(sh_esp) = INVALID_OFFSET;
Int VGOFF_(sh_ebp) = INVALID_OFFSET;
Int VGOFF_(sh_esi) = INVALID_OFFSET;
Int VGOFF_(sh_edi) = INVALID_OFFSET;
Int VGOFF_(sh_eflags) = INVALID_OFFSET;
Int VGOFF_(helper_idiv_64_32) = INVALID_OFFSET;
Int VGOFF_(helper_div_64_32) = INVALID_OFFSET;
Int VGOFF_(helper_idiv_32_16) = INVALID_OFFSET;
Int VGOFF_(helper_div_32_16) = INVALID_OFFSET;
Int VGOFF_(helper_idiv_16_8) = INVALID_OFFSET;
Int VGOFF_(helper_div_16_8) = INVALID_OFFSET;
Int VGOFF_(helper_imul_32_64) = INVALID_OFFSET;
Int VGOFF_(helper_mul_32_64) = INVALID_OFFSET;
Int VGOFF_(helper_imul_16_32) = INVALID_OFFSET;
Int VGOFF_(helper_mul_16_32) = INVALID_OFFSET;
Int VGOFF_(helper_imul_8_16) = INVALID_OFFSET;
Int VGOFF_(helper_mul_8_16) = INVALID_OFFSET;
Int VGOFF_(helper_CLD) = INVALID_OFFSET;
Int VGOFF_(helper_STD) = INVALID_OFFSET;
Int VGOFF_(helper_get_dirflag) = INVALID_OFFSET;
Int VGOFF_(helper_shldl) = INVALID_OFFSET;
Int VGOFF_(helper_shldw) = INVALID_OFFSET;
Int VGOFF_(helper_shrdl) = INVALID_OFFSET;
Int VGOFF_(helper_shrdw) = INVALID_OFFSET;
Int VGOFF_(helper_RDTSC) = INVALID_OFFSET;
Int VGOFF_(helper_CPUID) = INVALID_OFFSET;
Int VGOFF_(helper_BSWAP) = INVALID_OFFSET;
Int VGOFF_(helper_bsf) = INVALID_OFFSET;
Int VGOFF_(helper_bsr) = INVALID_OFFSET;
Int VGOFF_(helper_fstsw_AX) = INVALID_OFFSET;
Int VGOFF_(helper_SAHF) = INVALID_OFFSET;
Int VGOFF_(helper_DAS) = INVALID_OFFSET;
Int VGOFF_(helper_DAA) = INVALID_OFFSET;
Int VGOFF_(helper_value_check4_fail) = INVALID_OFFSET;
Int VGOFF_(helper_value_check2_fail) = INVALID_OFFSET;
Int VGOFF_(helper_value_check1_fail) = INVALID_OFFSET;
Int VGOFF_(helper_value_check0_fail) = INVALID_OFFSET;
Int VGOFF_(helperc_LOADV4) = INVALID_OFFSET;
Int VGOFF_(helperc_LOADV2) = INVALID_OFFSET;
Int VGOFF_(helperc_LOADV1) = INVALID_OFFSET;
Int VGOFF_(helperc_STOREV4) = INVALID_OFFSET;
Int VGOFF_(helperc_STOREV2) = INVALID_OFFSET;
Int VGOFF_(helperc_STOREV1) = INVALID_OFFSET;
Int VGOFF_(handle_esp_assignment) = INVALID_OFFSET;
Int VGOFF_(fpu_write_check) = INVALID_OFFSET;
Int VGOFF_(fpu_read_check) = INVALID_OFFSET;
Int VGOFF_(cachesim_log_non_mem_instr) = INVALID_OFFSET;
Int VGOFF_(cachesim_log_mem_instr) = INVALID_OFFSET;
/* This is the actual defn of baseblock. */
UInt VG_(baseBlock)[VG_BASEBLOCK_WORDS];
/* Words. */
static Int baB_off = 0;
/* Returns the offset, in words. */
static Int alloc_BaB ( Int words )
{
Int off = baB_off;
baB_off += words;
if (baB_off >= VG_BASEBLOCK_WORDS)
VG_(panic)( "alloc_BaB: baseBlock is too small");
return off;
}
/* Allocate 1 word in baseBlock and set it to the given value. */
static Int alloc_BaB_1_set ( Addr a )
{
Int off = alloc_BaB(1);
VG_(baseBlock)[off] = (UInt)a;
return off;
}
/* Here we assign actual offsets. It's important to get the most
popular referents within 128 bytes of the start, so we can take
advantage of short addressing modes relative to %ebp. Popularity
of offsets was measured on 22 Feb 02 running a KDE application, and
the slots rearranged accordingly, with a 1.5% reduction in total
size of translations. */
static void vg_init_baseBlock ( void )
{
baB_off = 0;
/* Those with offsets under 128 are carefully chosen. */
/* WORD offsets in this column */
/* 0 */ VGOFF_(m_eax) = alloc_BaB(1);
/* 1 */ VGOFF_(m_ecx) = alloc_BaB(1);
/* 2 */ VGOFF_(m_edx) = alloc_BaB(1);
/* 3 */ VGOFF_(m_ebx) = alloc_BaB(1);
/* 4 */ VGOFF_(m_esp) = alloc_BaB(1);
/* 5 */ VGOFF_(m_ebp) = alloc_BaB(1);
/* 6 */ VGOFF_(m_esi) = alloc_BaB(1);
/* 7 */ VGOFF_(m_edi) = alloc_BaB(1);
/* 8 */ VGOFF_(m_eflags) = alloc_BaB(1);
/* 9 */ VGOFF_(sh_eax) = alloc_BaB(1);
/* 10 */ VGOFF_(sh_ecx) = alloc_BaB(1);
/* 11 */ VGOFF_(sh_edx) = alloc_BaB(1);
/* 12 */ VGOFF_(sh_ebx) = alloc_BaB(1);
/* 13 */ VGOFF_(sh_esp) = alloc_BaB(1);
/* 14 */ VGOFF_(sh_ebp) = alloc_BaB(1);
/* 15 */ VGOFF_(sh_esi) = alloc_BaB(1);
/* 16 */ VGOFF_(sh_edi) = alloc_BaB(1);
/* 17 */ VGOFF_(sh_eflags) = alloc_BaB(1);
/* 17a */
VGOFF_(cachesim_log_non_mem_instr)
= alloc_BaB_1_set( (Addr) & VG_(cachesim_log_non_mem_instr) );
/* 17b */
VGOFF_(cachesim_log_mem_instr)
= alloc_BaB_1_set( (Addr) & VG_(cachesim_log_mem_instr) );
/* 18 */
VGOFF_(helper_value_check4_fail)
= alloc_BaB_1_set( (Addr) & VG_(helper_value_check4_fail) );
/* 19 */
VGOFF_(helper_value_check0_fail)
= alloc_BaB_1_set( (Addr) & VG_(helper_value_check0_fail) );
/* 20 */
VGOFF_(helperc_STOREV4)
= alloc_BaB_1_set( (Addr) & VG_(helperc_STOREV4) );
/* 21 */
VGOFF_(helperc_STOREV1)
= alloc_BaB_1_set( (Addr) & VG_(helperc_STOREV1) );
/* 22 */
VGOFF_(helperc_LOADV4)
= alloc_BaB_1_set( (Addr) & VG_(helperc_LOADV4) );
/* 23 */
VGOFF_(helperc_LOADV1)
= alloc_BaB_1_set( (Addr) & VG_(helperc_LOADV1) );
/* 24 */
VGOFF_(handle_esp_assignment)
= alloc_BaB_1_set( (Addr) & VGM_(handle_esp_assignment) );
/* 25 */
VGOFF_(m_eip) = alloc_BaB(1);
/* There are currently 24 spill slots */
/* 26 .. 49 This overlaps the magic boundary at >= 32 words, but
most spills are to low numbered spill slots, so the ones above
the boundary don't see much action. */
VGOFF_(spillslots) = alloc_BaB(VG_MAX_SPILLSLOTS);
/* These two pushed beyond the boundary because 2-byte transactions
are rare. */
/* 50 */
VGOFF_(helperc_STOREV2)
= alloc_BaB_1_set( (Addr) & VG_(helperc_STOREV2) );
/* 51 */
VGOFF_(helperc_LOADV2)
= alloc_BaB_1_set( (Addr) & VG_(helperc_LOADV2) );
/* 52 */
VGOFF_(fpu_write_check)
= alloc_BaB_1_set( (Addr) & VGM_(fpu_write_check) );
/* 53 */
VGOFF_(fpu_read_check)
= alloc_BaB_1_set( (Addr) & VGM_(fpu_read_check) );
/* Actually I don't think these two are ever used. */
/* 54 */
VGOFF_(helper_value_check2_fail)
= alloc_BaB_1_set( (Addr) & VG_(helper_value_check2_fail) );
/* 55 */
VGOFF_(helper_value_check1_fail)
= alloc_BaB_1_set( (Addr) & VG_(helper_value_check1_fail) );
/* I gave up counting at this point. Since they're way above the
short-amode-boundary, there's no point. */
VGOFF_(m_fpustate) = alloc_BaB(VG_SIZE_OF_FPUSTATE_W);
VGOFF_(helper_idiv_64_32)
= alloc_BaB_1_set( (Addr) & VG_(helper_idiv_64_32) );
VGOFF_(helper_div_64_32)
= alloc_BaB_1_set( (Addr) & VG_(helper_div_64_32) );
VGOFF_(helper_idiv_32_16)
= alloc_BaB_1_set( (Addr) & VG_(helper_idiv_32_16) );
VGOFF_(helper_div_32_16)
= alloc_BaB_1_set( (Addr) & VG_(helper_div_32_16) );
VGOFF_(helper_idiv_16_8)
= alloc_BaB_1_set( (Addr) & VG_(helper_idiv_16_8) );
VGOFF_(helper_div_16_8)
= alloc_BaB_1_set( (Addr) & VG_(helper_div_16_8) );
VGOFF_(helper_imul_32_64)
= alloc_BaB_1_set( (Addr) & VG_(helper_imul_32_64) );
VGOFF_(helper_mul_32_64)
= alloc_BaB_1_set( (Addr) & VG_(helper_mul_32_64) );
VGOFF_(helper_imul_16_32)
= alloc_BaB_1_set( (Addr) & VG_(helper_imul_16_32) );
VGOFF_(helper_mul_16_32)
= alloc_BaB_1_set( (Addr) & VG_(helper_mul_16_32) );
VGOFF_(helper_imul_8_16)
= alloc_BaB_1_set( (Addr) & VG_(helper_imul_8_16) );
VGOFF_(helper_mul_8_16)
= alloc_BaB_1_set( (Addr) & VG_(helper_mul_8_16) );
VGOFF_(helper_CLD)
= alloc_BaB_1_set( (Addr) & VG_(helper_CLD) );
VGOFF_(helper_STD)
= alloc_BaB_1_set( (Addr) & VG_(helper_STD) );
VGOFF_(helper_get_dirflag)
= alloc_BaB_1_set( (Addr) & VG_(helper_get_dirflag) );
VGOFF_(helper_shldl)
= alloc_BaB_1_set( (Addr) & VG_(helper_shldl) );
VGOFF_(helper_shldw)
= alloc_BaB_1_set( (Addr) & VG_(helper_shldw) );
VGOFF_(helper_shrdl)
= alloc_BaB_1_set( (Addr) & VG_(helper_shrdl) );
VGOFF_(helper_shrdw)
= alloc_BaB_1_set( (Addr) & VG_(helper_shrdw) );
VGOFF_(helper_RDTSC)
= alloc_BaB_1_set( (Addr) & VG_(helper_RDTSC) );
VGOFF_(helper_CPUID)
= alloc_BaB_1_set( (Addr) & VG_(helper_CPUID) );
VGOFF_(helper_bsf)
= alloc_BaB_1_set( (Addr) & VG_(helper_bsf) );
VGOFF_(helper_bsr)
= alloc_BaB_1_set( (Addr) & VG_(helper_bsr) );
VGOFF_(helper_fstsw_AX)
= alloc_BaB_1_set( (Addr) & VG_(helper_fstsw_AX) );
VGOFF_(helper_SAHF)
= alloc_BaB_1_set( (Addr) & VG_(helper_SAHF) );
VGOFF_(helper_DAS)
= alloc_BaB_1_set( (Addr) & VG_(helper_DAS) );
VGOFF_(helper_DAA)
= alloc_BaB_1_set( (Addr) & VG_(helper_DAA) );
}
/* ---------------------------------------------------------------------
Global entities which are not referenced from generated code.
------------------------------------------------------------------ */
/* The stack on which Valgrind runs. We can't use the same stack as
the simulatee -- that's an important design decision. */
UInt VG_(stack)[10000];
/* Ditto our signal delivery stack. */
UInt VG_(sigstack)[10000];
/* Saving stuff across system calls. */
UInt VG_(real_fpu_state_saved_over_syscall_d1)[VG_SIZE_OF_FPUSTATE_W];
UInt VG_(real_fpu_state_saved_over_syscall_d2)[VG_SIZE_OF_FPUSTATE_W];
Addr VG_(esp_saved_over_syscall_d1);
Addr VG_(esp_saved_over_syscall_d2);
/* Counts downwards in vg_run_innerloop. */
UInt VG_(dispatch_ctr);
/* 64-bit counter for the number of basic blocks done. */
ULong VG_(bbs_done);
/* 64-bit counter for the number of bbs to go before a debug exit. */
ULong VG_(bbs_to_go);
/* Produce debugging output? */
Bool VG_(disassemble) = False;
/* The current LRU epoch. */
UInt VG_(current_epoch) = 0;
/* This is the ThreadId of the last thread the scheduler ran. */
ThreadId VG_(last_run_tid) = 0;
/* ---------------------------------------------------------------------
Counters, for informational purposes only.
------------------------------------------------------------------ */
/* Number of lookups which miss the fast tt helper. */
UInt VG_(tt_fast_misses) = 0;
/* Counts for LRU informational messages. */
/* Number and total o/t size of new translations this epoch. */
UInt VG_(this_epoch_in_count) = 0;
UInt VG_(this_epoch_in_osize) = 0;
UInt VG_(this_epoch_in_tsize) = 0;
/* Number and total o/t size of discarded translations this epoch. */
UInt VG_(this_epoch_out_count) = 0;
UInt VG_(this_epoch_out_osize) = 0;
UInt VG_(this_epoch_out_tsize) = 0;
/* Number and total o/t size of translations overall. */
UInt VG_(overall_in_count) = 0;
UInt VG_(overall_in_osize) = 0;
UInt VG_(overall_in_tsize) = 0;
/* Number and total o/t size of discards overall. */
UInt VG_(overall_out_count) = 0;
UInt VG_(overall_out_osize) = 0;
UInt VG_(overall_out_tsize) = 0;
/* The number of LRU-clearings of TT/TC. */
UInt VG_(number_of_lrus) = 0;
/* Counts pertaining to the register allocator. */
/* total number of uinstrs input to reg-alloc */
UInt VG_(uinstrs_prealloc) = 0;
/* total number of uinstrs added due to spill code */
UInt VG_(uinstrs_spill) = 0;
/* number of bbs requiring spill code */
UInt VG_(translations_needing_spill) = 0;
/* total of register ranks over all translations */
UInt VG_(total_reg_rank) = 0;
/* Counts pertaining to internal sanity checking. */
UInt VG_(sanity_fast_count) = 0;
UInt VG_(sanity_slow_count) = 0;
/* Counts pertaining to the scheduler. */
UInt VG_(num_scheduling_events_MINOR) = 0;
UInt VG_(num_scheduling_events_MAJOR) = 0;
/* ---------------------------------------------------------------------
Values derived from command-line options.
------------------------------------------------------------------ */
Bool VG_(clo_check_addrVs);
Bool VG_(clo_GDB_attach);
Int VG_(sanity_level);
Int VG_(clo_verbosity);
Bool VG_(clo_demangle);
Bool VG_(clo_leak_check);
Bool VG_(clo_show_reachable);
Int VG_(clo_leak_resolution);
Bool VG_(clo_sloppy_malloc);
Bool VG_(clo_partial_loads_ok);
Bool VG_(clo_trace_children);
Int VG_(clo_logfile_fd);
Int VG_(clo_freelist_vol);
Bool VG_(clo_workaround_gcc296_bugs);
Int VG_(clo_n_suppressions);
Char* VG_(clo_suppressions)[VG_CLO_MAX_SFILES];
Bool VG_(clo_single_step);
Bool VG_(clo_optimise);
Bool VG_(clo_instrument);
Bool VG_(clo_cleanup);
Bool VG_(clo_cachesim);
Int VG_(clo_smc_check);
Bool VG_(clo_trace_syscalls);
Bool VG_(clo_trace_signals);
Bool VG_(clo_trace_symtab);
Bool VG_(clo_trace_malloc);
Bool VG_(clo_trace_sched);
Int VG_(clo_trace_pthread_level);
ULong VG_(clo_stop_after);
Int VG_(clo_dump_error);
Int VG_(clo_backtrace_size);
Char* VG_(clo_weird_hacks);
/* This Bool is needed by wrappers in vg_clientmalloc.c to decide how
to behave. Initially we say False. */
Bool VG_(running_on_simd_CPU) = False;
/* Holds client's %esp at the point we gained control. */
Addr VG_(esp_at_startup);
/* As deduced from VG_(esp_at_startup), the client's argc, argv[] and
envp[] as extracted from the client's stack at startup-time. */
Int VG_(client_argc);
Char** VG_(client_argv);
Char** VG_(client_envp);
/* A place into which to copy the value of env var VG_ARGS, so we
don't have to modify the original. */
static Char vg_cmdline_copy[M_VG_CMDLINE_STRLEN];
/* ---------------------------------------------------------------------
Processing of command-line options.
------------------------------------------------------------------ */
static void bad_option ( Char* opt )
{
VG_(shutdown_logging)();
VG_(clo_logfile_fd) = 2; /* stderr */
VG_(printf)("valgrind.so: Bad option `%s'; aborting.\n", opt);
VG_(exit)(1);
}
static void config_error ( Char* msg )
{
VG_(shutdown_logging)();
VG_(clo_logfile_fd) = 2; /* stderr */
VG_(printf)("valgrind.so: Startup or configuration error:\n\t%s\n", msg);
VG_(printf)("valgrind.so: Unable to start up properly. Giving up.\n");
VG_(exit)(1);
}
static void args_grok_error ( Char* msg )
{
VG_(shutdown_logging)();
VG_(clo_logfile_fd) = 2; /* stderr */
VG_(printf)("valgrind.so: When searching for "
"client's argc/argc/envp:\n\t%s\n", msg);
config_error("couldn't find client's argc/argc/envp");
}
static void process_cmd_line_options ( void )
{
UChar* argv[M_VG_CMDLINE_OPTS];
UInt argc;
UChar* p;
UChar* str;
Int i, eventually_logfile_fd, ctr;
# define ISSPACE(cc) ((cc) == ' ' || (cc) == '\t' || (cc) == '\n')
# define STREQ(s1,s2) (0==VG_(strcmp_ws)((s1),(s2)))
# define STREQN(nn,s1,s2) (0==VG_(strncmp_ws)((s1),(s2),(nn)))
/* Set defaults. */
VG_(clo_check_addrVs) = True;
VG_(clo_GDB_attach) = False;
VG_(sanity_level) = 1;
VG_(clo_verbosity) = 1;
VG_(clo_demangle) = True;
VG_(clo_leak_check) = False;
VG_(clo_show_reachable) = False;
VG_(clo_leak_resolution) = 2;
VG_(clo_sloppy_malloc) = False;
VG_(clo_partial_loads_ok) = True;
VG_(clo_trace_children) = False;
VG_(clo_logfile_fd) = 2; /* stderr */
VG_(clo_freelist_vol) = 1000000;
VG_(clo_workaround_gcc296_bugs) = False;
VG_(clo_n_suppressions) = 0;
VG_(clo_single_step) = False;
VG_(clo_optimise) = True;
VG_(clo_instrument) = True;
VG_(clo_cleanup) = True;
VG_(clo_smc_check) = /* VG_CLO_SMC_SOME */ VG_CLO_SMC_NONE;
VG_(clo_trace_syscalls) = False;
VG_(clo_trace_signals) = False;
VG_(clo_trace_symtab) = False;
VG_(clo_trace_malloc) = False;
VG_(clo_trace_sched) = False;
VG_(clo_trace_pthread_level) = 0;
VG_(clo_stop_after) = 1000000000000LL;
VG_(clo_dump_error) = 0;
VG_(clo_backtrace_size) = 4;
VG_(clo_weird_hacks) = NULL;
eventually_logfile_fd = VG_(clo_logfile_fd);
/* Once logging is started, we can safely send messages pertaining
to failures in initialisation. */
VG_(startup_logging)();
/* (Suggested by Fabrice Bellard ... )
We look for the Linux ELF table and go down until we find the
envc & envp. It is not full proof, but these structures should
change less often than the libc ones. */
{
UInt* sp = 0; /* bogus init to keep gcc -O happy */
/* locate the top of the stack */
if (VG_STACK_MATCHES_BASE( VG_(esp_at_startup),
VG_STARTUP_STACK_BASE_1 )) {
sp = (UInt*)VG_STARTUP_STACK_BASE_1;
} else
if (VG_STACK_MATCHES_BASE( VG_(esp_at_startup),
VG_STARTUP_STACK_BASE_2 )) {
sp = (UInt*)VG_STARTUP_STACK_BASE_2;
} else {
args_grok_error(
"startup %esp is not near any VG_STARTUP_STACK_BASE_*\n "
"constants defined in vg_include.h. You should investigate."
);
}
/* we locate: NEW_AUX_ENT(1, AT_PAGESZ, ELF_EXEC_PAGESIZE) in
the elf interpreter table */
sp -= 2;
while (sp[0] != VKI_AT_PAGESZ || sp[1] != 4096) {
/* VG_(printf)("trying %p\n", sp); */
sp--;
}
if (sp[2] == VKI_AT_BASE
&& sp[0] == VKI_AT_PAGESZ
&& sp[-2] == VKI_AT_PHNUM
&& sp[-4] == VKI_AT_PHENT
&& sp[-6] == VKI_AT_PHDR
&& sp[-6-1] == 0) {
if (0)
VG_(printf)("Looks like you've got a 2.2.X kernel here.\n");
sp -= 6;
} else
if (sp[2] == VKI_AT_CLKTCK
&& sp[0] == VKI_AT_PAGESZ
&& sp[-2] == VKI_AT_HWCAP
&& sp[-2-1] == 0) {
if (0)
VG_(printf)("Looks like you've got a 2.4.X kernel here.\n");
sp -= 2;
} else
if (sp[2] == VKI_AT_CLKTCK
&& sp[0] == VKI_AT_PAGESZ
&& sp[-2] == VKI_AT_HWCAP
&& sp[-2-20-1] == 0) {
if (0)
VG_(printf)("Looks like you've got a early 2.4.X kernel here.\n");
sp -= 22;
} else
args_grok_error(
"ELF frame does not look like 2.2.X or 2.4.X.\n "
"See kernel sources linux/fs/binfmt_elf.c to make sense of this."
);
sp--;
if (*sp != 0)
args_grok_error("can't find NULL at end of env[]");
/* sp now points to NULL at the end of env[] */
ctr = 0;
while (True) {
sp --;
if (*sp == 0) break;
if (++ctr >= 1000)
args_grok_error(
"suspiciously many (1000) env[] entries; giving up");
}
/* sp now points to NULL at the end of argv[] */
VG_(client_envp) = (Char**)(sp+1);
ctr = 0;
VG_(client_argc) = 0;
while (True) {
sp--;
if (*sp == VG_(client_argc))
break;
VG_(client_argc)++;
if (++ctr >= 1000)
args_grok_error(
"suspiciously many (1000) argv[] entries; giving up");
}
VG_(client_argv) = (Char**)(sp+1);
}
/* Now that VG_(client_envp) has been set, we can extract the args
for Valgrind itself. Copy into global var so that we don't have to
write zeroes to the getenv'd value itself. */
str = VG_(getenv)("VG_ARGS");
argc = 0;
if (!str) {
config_error("Can't read options from env var VG_ARGS.");
}
if (VG_(strlen)(str) >= M_VG_CMDLINE_STRLEN-1) {
config_error("Command line length exceeds M_CMDLINE_STRLEN.");
}
VG_(strcpy)(vg_cmdline_copy, str);
str = NULL;
p = &vg_cmdline_copy[0];
while (True) {
while (ISSPACE(*p)) { *p = 0; p++; }
if (*p == 0) break;
if (argc < M_VG_CMDLINE_OPTS-1) {
argv[argc] = p; argc++;
} else {
config_error(
"Found more than M_CMDLINE_OPTS command-line opts.");
}
while (*p != 0 && !ISSPACE(*p)) p++;
}
for (i = 0; i < argc; i++) {
if (STREQ(argv[i], "-v") || STREQ(argv[i], "--verbose"))
VG_(clo_verbosity)++;
else if (STREQ(argv[i], "-q") || STREQ(argv[i], "--quiet"))
VG_(clo_verbosity)--;
else if (STREQ(argv[i], "--check-addrVs=yes"))
VG_(clo_check_addrVs) = True;
else if (STREQ(argv[i], "--check-addrVs=no"))
VG_(clo_check_addrVs) = False;
else if (STREQ(argv[i], "--gdb-attach=yes"))
VG_(clo_GDB_attach) = True;
else if (STREQ(argv[i], "--gdb-attach=no"))
VG_(clo_GDB_attach) = False;
else if (STREQ(argv[i], "--demangle=yes"))
VG_(clo_demangle) = True;
else if (STREQ(argv[i], "--demangle=no"))
VG_(clo_demangle) = False;
else if (STREQ(argv[i], "--partial-loads-ok=yes"))
VG_(clo_partial_loads_ok) = True;
else if (STREQ(argv[i], "--partial-loads-ok=no"))
VG_(clo_partial_loads_ok) = False;
else if (STREQ(argv[i], "--leak-check=yes"))
VG_(clo_leak_check) = True;
else if (STREQ(argv[i], "--leak-check=no"))
VG_(clo_leak_check) = False;
else if (STREQ(argv[i], "--show-reachable=yes"))
VG_(clo_show_reachable) = True;
else if (STREQ(argv[i], "--show-reachable=no"))
VG_(clo_show_reachable) = False;
else if (STREQ(argv[i], "--leak-resolution=low"))
VG_(clo_leak_resolution) = 2;
else if (STREQ(argv[i], "--leak-resolution=med"))
VG_(clo_leak_resolution) = 4;
else if (STREQ(argv[i], "--leak-resolution=high"))
VG_(clo_leak_resolution) = VG_DEEPEST_BACKTRACE;
else if (STREQ(argv[i], "--sloppy-malloc=yes"))
VG_(clo_sloppy_malloc) = True;
else if (STREQ(argv[i], "--sloppy-malloc=no"))
VG_(clo_sloppy_malloc) = False;
else if (STREQ(argv[i], "--trace-children=yes"))
VG_(clo_trace_children) = True;
else if (STREQ(argv[i], "--trace-children=no"))
VG_(clo_trace_children) = False;
else if (STREQ(argv[i], "--workaround-gcc296-bugs=yes"))
VG_(clo_workaround_gcc296_bugs) = True;
else if (STREQ(argv[i], "--workaround-gcc296-bugs=no"))
VG_(clo_workaround_gcc296_bugs) = False;
else if (STREQN(15, argv[i], "--sanity-level="))
VG_(sanity_level) = (Int)VG_(atoll)(&argv[i][15]);
else if (STREQN(13, argv[i], "--logfile-fd="))
eventually_logfile_fd = (Int)VG_(atoll)(&argv[i][13]);
else if (STREQN(15, argv[i], "--freelist-vol=")) {
VG_(clo_freelist_vol) = (Int)VG_(atoll)(&argv[i][15]);
if (VG_(clo_freelist_vol) < 0) VG_(clo_freelist_vol) = 2;
}
else if (STREQN(15, argv[i], "--suppressions=")) {
if (VG_(clo_n_suppressions) >= VG_CLO_MAX_SFILES) {
VG_(message)(Vg_UserMsg, "Too many logfiles specified.");
VG_(message)(Vg_UserMsg,
"Increase VG_CLO_MAX_SFILES and recompile.");
bad_option(argv[i]);
}
VG_(clo_suppressions)[VG_(clo_n_suppressions)] = &argv[i][15];
VG_(clo_n_suppressions)++;
}
else if (STREQ(argv[i], "--single-step=yes"))
VG_(clo_single_step) = True;
else if (STREQ(argv[i], "--single-step=no"))
VG_(clo_single_step) = False;
else if (STREQ(argv[i], "--optimise=yes"))
VG_(clo_optimise) = True;
else if (STREQ(argv[i], "--optimise=no"))
VG_(clo_optimise) = False;
else if (STREQ(argv[i], "--instrument=yes"))
VG_(clo_instrument) = True;
else if (STREQ(argv[i], "--instrument=no"))
VG_(clo_instrument) = False;
else if (STREQ(argv[i], "--cleanup=yes"))
VG_(clo_cleanup) = True;
else if (STREQ(argv[i], "--cleanup=no"))
VG_(clo_cleanup) = False;
else if (STREQ(argv[i], "--cachesim=yes"))
VG_(clo_cachesim) = True;
else if (STREQ(argv[i], "--cachesim=no"))
VG_(clo_cachesim) = False;
else if (STREQ(argv[i], "--smc-check=none"))
VG_(clo_smc_check) = VG_CLO_SMC_NONE;
else if (STREQ(argv[i], "--smc-check=some"))
VG_(clo_smc_check) = VG_CLO_SMC_SOME;
else if (STREQ(argv[i], "--smc-check=all"))
VG_(clo_smc_check) = VG_CLO_SMC_ALL;
else if (STREQ(argv[i], "--trace-syscalls=yes"))
VG_(clo_trace_syscalls) = True;
else if (STREQ(argv[i], "--trace-syscalls=no"))
VG_(clo_trace_syscalls) = False;
else if (STREQ(argv[i], "--trace-signals=yes"))
VG_(clo_trace_signals) = True;
else if (STREQ(argv[i], "--trace-signals=no"))
VG_(clo_trace_signals) = False;
else if (STREQ(argv[i], "--trace-symtab=yes"))
VG_(clo_trace_symtab) = True;
else if (STREQ(argv[i], "--trace-symtab=no"))
VG_(clo_trace_symtab) = False;
else if (STREQ(argv[i], "--trace-malloc=yes"))
VG_(clo_trace_malloc) = True;
else if (STREQ(argv[i], "--trace-malloc=no"))
VG_(clo_trace_malloc) = False;
else if (STREQ(argv[i], "--trace-sched=yes"))
VG_(clo_trace_sched) = True;
else if (STREQ(argv[i], "--trace-sched=no"))
VG_(clo_trace_sched) = False;
else if (STREQ(argv[i], "--trace-pthread=none"))
VG_(clo_trace_pthread_level) = 0;
else if (STREQ(argv[i], "--trace-pthread=some"))
VG_(clo_trace_pthread_level) = 1;
else if (STREQ(argv[i], "--trace-pthread=all"))
VG_(clo_trace_pthread_level) = 2;
else if (STREQN(14, argv[i], "--weird-hacks="))
VG_(clo_weird_hacks) = &argv[i][14];
else if (STREQN(13, argv[i], "--stop-after="))
VG_(clo_stop_after) = VG_(atoll)(&argv[i][13]);
else if (STREQN(13, argv[i], "--dump-error="))
VG_(clo_dump_error) = (Int)VG_(atoll)(&argv[i][13]);
else if (STREQN(14, argv[i], "--num-callers=")) {
/* Make sure it's sane. */
VG_(clo_backtrace_size) = (Int)VG_(atoll)(&argv[i][14]);
if (VG_(clo_backtrace_size) < 2)
VG_(clo_backtrace_size) = 2;
if (VG_(clo_backtrace_size) >= VG_DEEPEST_BACKTRACE)
VG_(clo_backtrace_size) = VG_DEEPEST_BACKTRACE;
}
else
bad_option(argv[i]);
}
# undef ISSPACE
# undef STREQ
# undef STREQN
if (VG_(clo_verbosity < 0))
VG_(clo_verbosity) = 0;
if (VG_(clo_GDB_attach) && VG_(clo_trace_children)) {
VG_(message)(Vg_UserMsg, "");
VG_(message)(Vg_UserMsg,
"--gdb-attach=yes conflicts with --trace-children=yes");
VG_(message)(Vg_UserMsg,
"Please choose one or the other, but not both.");
bad_option("--gdb-attach=yes and --trace-children=yes");
}
VG_(clo_logfile_fd) = eventually_logfile_fd;
/* Don't do memory checking if simulating the cache. */
if (VG_(clo_cachesim)) {
VG_(clo_instrument) = False;
}
if (VG_(clo_verbosity > 0)) {
if (VG_(clo_cachesim)) {
VG_(message)(Vg_UserMsg,
"cachegrind-%s, an I1/D1/L2 cache profiler for x86 GNU/Linux.",
VERSION);
} else {
VG_(message)(Vg_UserMsg,
"valgrind-%s, a memory error detector for x86 GNU/Linux.",
VERSION);
}
}
if (VG_(clo_verbosity > 0))
VG_(message)(Vg_UserMsg,
"Copyright (C) 2000-2002, and GNU GPL'd, by Julian Seward.");
if (VG_(clo_verbosity) > 1) {
VG_(message)(Vg_UserMsg, "Startup, with flags:");
for (i = 0; i < argc; i++) {
VG_(message)(Vg_UserMsg, " %s", argv[i]);
}
}
if (VG_(clo_n_suppressions) == 0 && !VG_(clo_cachesim)) {
config_error("No error-suppression files were specified.");
}
}
/* ---------------------------------------------------------------------
Copying to/from m_state_static.
------------------------------------------------------------------ */
UInt VG_(m_state_static) [8 /* int regs, in Intel order */
+ 1 /* %eflags */
+ 1 /* %eip */
+ VG_SIZE_OF_FPUSTATE_W /* FPU state */
];
void VG_(copy_baseBlock_to_m_state_static) ( void )
{
Int i;
VG_(m_state_static)[ 0/4] = VG_(baseBlock)[VGOFF_(m_eax)];
VG_(m_state_static)[ 4/4] = VG_(baseBlock)[VGOFF_(m_ecx)];
VG_(m_state_static)[ 8/4] = VG_(baseBlock)[VGOFF_(m_edx)];
VG_(m_state_static)[12/4] = VG_(baseBlock)[VGOFF_(m_ebx)];
VG_(m_state_static)[16/4] = VG_(baseBlock)[VGOFF_(m_esp)];
VG_(m_state_static)[20/4] = VG_(baseBlock)[VGOFF_(m_ebp)];
VG_(m_state_static)[24/4] = VG_(baseBlock)[VGOFF_(m_esi)];
VG_(m_state_static)[28/4] = VG_(baseBlock)[VGOFF_(m_edi)];
VG_(m_state_static)[32/4] = VG_(baseBlock)[VGOFF_(m_eflags)];
VG_(m_state_static)[36/4] = VG_(baseBlock)[VGOFF_(m_eip)];
for (i = 0; i < VG_SIZE_OF_FPUSTATE_W; i++)
VG_(m_state_static)[40/4 + i]
= VG_(baseBlock)[VGOFF_(m_fpustate) + i];
}
void VG_(copy_m_state_static_to_baseBlock) ( void )
{
Int i;
VG_(baseBlock)[VGOFF_(m_eax)] = VG_(m_state_static)[ 0/4];
VG_(baseBlock)[VGOFF_(m_ecx)] = VG_(m_state_static)[ 4/4];
VG_(baseBlock)[VGOFF_(m_edx)] = VG_(m_state_static)[ 8/4];
VG_(baseBlock)[VGOFF_(m_ebx)] = VG_(m_state_static)[12/4];
VG_(baseBlock)[VGOFF_(m_esp)] = VG_(m_state_static)[16/4];
VG_(baseBlock)[VGOFF_(m_ebp)] = VG_(m_state_static)[20/4];
VG_(baseBlock)[VGOFF_(m_esi)] = VG_(m_state_static)[24/4];
VG_(baseBlock)[VGOFF_(m_edi)] = VG_(m_state_static)[28/4];
VG_(baseBlock)[VGOFF_(m_eflags)] = VG_(m_state_static)[32/4];
VG_(baseBlock)[VGOFF_(m_eip)] = VG_(m_state_static)[36/4];
for (i = 0; i < VG_SIZE_OF_FPUSTATE_W; i++)
VG_(baseBlock)[VGOFF_(m_fpustate) + i]
= VG_(m_state_static)[40/4 + i];
}
/* ---------------------------------------------------------------------
Show accumulated counts.
------------------------------------------------------------------ */
static void vg_show_counts ( void )
{
VG_(message)(Vg_DebugMsg,
" lru: %d epochs, %d clearings.",
VG_(current_epoch),
VG_(number_of_lrus) );
VG_(message)(Vg_DebugMsg,
"translate: new %d (%d -> %d), discard %d (%d -> %d).",
VG_(overall_in_count),
VG_(overall_in_osize),
VG_(overall_in_tsize),
VG_(overall_out_count),
VG_(overall_out_osize),
VG_(overall_out_tsize) );
VG_(message)(Vg_DebugMsg,
" dispatch: %lu basic blocks, %d/%d sched events, %d tt_fast misses.",
VG_(bbs_done), VG_(num_scheduling_events_MAJOR),
VG_(num_scheduling_events_MINOR),
VG_(tt_fast_misses));
VG_(message)(Vg_DebugMsg,
"reg-alloc: %d t-req-spill, "
"%d+%d orig+spill uis, %d total-reg-r.",
VG_(translations_needing_spill),
VG_(uinstrs_prealloc),
VG_(uinstrs_spill),
VG_(total_reg_rank) );
VG_(message)(Vg_DebugMsg,
" sanity: %d cheap, %d expensive checks.",
VG_(sanity_fast_count),
VG_(sanity_slow_count) );
}
/* ---------------------------------------------------------------------
Main!
------------------------------------------------------------------ */
/* Where we jump to once Valgrind has got control, and the real
machine's state has been copied to the m_state_static. */
void VG_(main) ( void )
{
Int i;
VgSchedReturnCode src;
ThreadState* tst;
/* Set up our stack sanity-check words. */
for (i = 0; i < 10; i++) {
VG_(stack)[i] = (UInt)(&VG_(stack)[i]) ^ 0xA4B3C2D1;
VG_(stack)[10000-1-i] = (UInt)(&VG_(stack)[10000-i-1]) ^ 0xABCD4321;
}
/* Set up baseBlock offsets and copy the saved machine's state into
it. */
vg_init_baseBlock();
VG_(copy_m_state_static_to_baseBlock)();
/* Process Valgrind's command-line opts (from env var VG_OPTS). */
process_cmd_line_options();
/* Initialise the scheduler, and copy the client's state from
baseBlock into VG_(threads)[1]. This has to come before signal
initialisations. */
VG_(scheduler_init)();
/* Initialise the signal handling subsystem, temporarily parking
the saved blocking-mask in saved_sigmask. */
VG_(sigstartup_actions)();
/* Perhaps we're profiling Valgrind? */
# ifdef VG_PROFILE
VGP_(init_profiling)();
# endif
/* Start calibration of our RDTSC-based clock. */
VG_(start_rdtsc_calibration)();
/* Hook to delay things long enough so we can get the pid and
attach GDB in another shell. */
/* {extern unsigned int sleep(unsigned int seconds); sleep(10);} */
if (VG_(clo_instrument) || VG_(clo_cachesim)) {
VGP_PUSHCC(VgpInitAudit);
VGM_(init_memory_audit)();
VGP_POPCC;
}
VGP_PUSHCC(VgpReadSyms);
VG_(read_symbols)();
VGP_POPCC;
/* End calibration of our RDTSC-based clock, leaving it as long as
we can. */
VG_(end_rdtsc_calibration)();
/* This should come after init_memory_audit; otherwise the latter
carefully sets up the permissions maps to cover the anonymous
mmaps for the translation table and translation cache, which
wastes > 20M of virtual address space. */
VG_(init_tt_tc)();
if (VG_(clo_verbosity) == 1) {
VG_(message)(Vg_UserMsg,
"For more details, rerun with: -v");
}
/* Now it is safe for malloc et al in vg_clientmalloc.c to act
instrumented-ly. */
VG_(running_on_simd_CPU) = True;
if (VG_(clo_instrument)) {
VGM_(make_readable) ( (Addr)&VG_(running_on_simd_CPU), 1 );
VGM_(make_readable) ( (Addr)&VG_(clo_instrument), 1 );
VGM_(make_readable) ( (Addr)&VG_(clo_trace_malloc), 1 );
VGM_(make_readable) ( (Addr)&VG_(clo_sloppy_malloc), 1 );
}
if (VG_(clo_cachesim))
VG_(init_cachesim)();
if (VG_(clo_verbosity) > 0)
VG_(message)(Vg_UserMsg, "");
VG_(bbs_to_go) = VG_(clo_stop_after);
/* Run! */
VGP_PUSHCC(VgpSched);
src = VG_(scheduler)();
VGP_POPCC;
if (VG_(clo_verbosity) > 0)
VG_(message)(Vg_UserMsg, "");
if (src == VgSrc_Deadlock) {
VG_(message)(Vg_UserMsg,
"Warning: pthread scheduler exited due to deadlock");
}
if (VG_(clo_instrument)) {
VG_(show_all_errors)();
VG_(clientmalloc_done)();
if (VG_(clo_verbosity) == 1) {
VG_(message)(Vg_UserMsg,
"For counts of detected errors, rerun with: -v");
}
if (VG_(clo_leak_check)) VG_(detect_memory_leaks)();
}
VG_(running_on_simd_CPU) = False;
if (VG_(clo_cachesim))
VG_(show_cachesim_results)(VG_(client_argc), VG_(client_argv));
VG_(do_sanity_checks)( True /*include expensive checks*/ );
if (VG_(clo_verbosity) > 1)
vg_show_counts();
if (0) {
VG_(message)(Vg_DebugMsg, "");
VG_(message)(Vg_DebugMsg,
"------ Valgrind's internal memory use stats follow ------" );
VG_(mallocSanityCheckAll)();
VG_(show_all_arena_stats)();
VG_(message)(Vg_DebugMsg,
"------ Valgrind's ExeContext management stats follow ------" );
VG_(show_ExeContext_stats)();
VG_(message)(Vg_DebugMsg,
"------ Valgrind's client block stats follow ---------------" );
VG_(show_client_block_stats)();
}
# ifdef VG_PROFILE
VGP_(done_profiling)();
# endif
VG_(done_prof_mem)();
VG_(shutdown_logging)();
/* Remove valgrind.so from a LD_PRELOAD=... string so child
processes don't get traced into. Also mess up $libdir/valgrind
so that our libpthread.so disappears from view. */
if (!VG_(clo_trace_children)) {
VG_(mash_LD_PRELOAD_and_LD_LIBRARY_PATH)(
VG_(getenv)("LD_PRELOAD"),
VG_(getenv)("LD_LIBRARY_PATH")
);
}
/* Decide how to exit. This depends on what the scheduler
returned. */
switch (src) {
case VgSrc_ExitSyscall: /* the normal way out */
vg_assert(VG_(last_run_tid) > 0
&& VG_(last_run_tid) < VG_N_THREADS);
tst = & VG_(threads)[VG_(last_run_tid)];
vg_assert(tst->status == VgTs_Runnable);
/* The thread's %EBX will hold the arg to exit(), so we just
do exit with that arg. */
VG_(exit)( tst->m_ebx );
/* NOT ALIVE HERE! */
VG_(panic)("entered the afterlife in vg_main() -- ExitSyscall");
break; /* what the hell :) */
case VgSrc_Deadlock:
/* Just exit now. No point in continuing. */
VG_(exit)(0);
VG_(panic)("entered the afterlife in vg_main() -- Deadlock");
break;
case VgSrc_BbsDone:
/* Tricky; we have to try and switch back to the real CPU.
This is all very dodgy and won't work at all in the
presence of threads, or if the client happened to be
running a signal handler. */
/* Prepare to restore state to the real CPU. */
VG_(load_thread_state)(1 /* root thread */ );
VG_(copy_baseBlock_to_m_state_static)();
/* This pushes a return address on the simulator's stack,
which is abandoned. We call vg_sigshutdown_actions() at
the end of vg_switch_to_real_CPU(), so as to ensure that
the original stack and machine state is restored before
the real signal mechanism is restored. */
VG_(switch_to_real_CPU)();
default:
VG_(panic)("vg_main(): unexpected scheduler return code");
}
}
/* Debugging thing .. can be called from assembly with OYNK macro. */
void VG_(oynk) ( Int n )
{
OINK(n);
}
/* Find "valgrind.so" in a LD_PRELOAD=... string, and convert it to
"valgrinq.so", which doesn't do anything. This is used to avoid
tracing into child processes. To make this work the build system
also supplies a dummy file, "valgrinq.so".
Also look for $(libdir)/lib/valgrind in LD_LIBRARY_PATH and change
it to $(libdir)/lib/valgrinq, so as to make our libpthread.so
disappear.
*/
void VG_(mash_LD_PRELOAD_and_LD_LIBRARY_PATH) ( Char* ld_preload_str,
Char* ld_library_path_str )
{
Char* p_prel = NULL;
Char* p_path = NULL;
Int what = 0;
if (ld_preload_str == NULL || ld_library_path_str == NULL)
goto mutancy;
/* VG_(printf)("%s %s\n", ld_preload_str, ld_library_path_str); */
p_prel = VG_(strstr)(ld_preload_str, "valgrind.so");
p_path = VG_(strstr)(ld_library_path_str, VG_LIBDIR);
if (p_prel == NULL) {
/* perhaps already happened? */
what = 1;
if (VG_(strstr)(ld_preload_str, "valgrinq.so") == NULL)
goto mutancy;
if (VG_(strstr)(ld_library_path_str, "lib/valgrinq") == NULL)
goto mutancy;
return;
}
what = 2;
if (p_path == NULL) goto mutancy;
/* in LD_PRELOAD, turn valgrind.so into valgrinq.so. */
what = 3;
if (p_prel[7] != 'd') goto mutancy;
p_prel[7] = 'q';
/* in LD_LIBRARY_PATH, turn $libdir/valgrind (as configure'd) from
.../lib/valgrind .../lib/valgrinq, which doesn't exist,
so that our own libpthread.so goes out of scope. */
p_path += VG_(strlen)(VG_LIBDIR);
what = 4;
if (p_path[0] != '/') goto mutancy;
p_path++; /* step over / */
what = 5;
if (p_path[7] != 'd') goto mutancy;
p_path[7] = 'q';
return;
mutancy:
VG_(printf)(
"\nVG_(mash_LD_PRELOAD_and_LD_LIBRARY_PATH): internal error:\n"
" what = %d\n"
" ld_preload_str = `%s'\n"
" ld_library_path_str = `%s'\n"
" p_prel = `%s'\n"
" p_path = `%s'\n"
" VG_LIBDIR = `%s'\n",
what, ld_preload_str, ld_library_path_str, p_prel, p_path, VG_LIBDIR );
VG_(panic)("VG_(mash_LD_PRELOAD_and_LD_LIBRARY_PATH) failed\n");
}
/* RUNS ON THE CLIENT'S STACK, but on the real CPU. Start GDB and get
it to attach to this process. Called if the user requests this
service after an error has been shown, so she can poke around and
look at parameters, memory, etc. You can't meaningfully get GDB to
continue the program, though; to continue, quit GDB. */
extern void VG_(start_GDB_whilst_on_client_stack) ( void )
{
Int res;
UChar buf[100];
VG_(sprintf)(buf,
"/usr/bin/gdb -nw /proc/%d/exe %d",
VG_(getpid)(), VG_(getpid)());
VG_(message)(Vg_UserMsg, "starting GDB with cmd: %s", buf);
res = VG_(system)(buf);
if (res == 0) {
VG_(message)(Vg_UserMsg, "");
VG_(message)(Vg_UserMsg,
"GDB has detached. Valgrind regains control. We continue.");
} else {
VG_(message)(Vg_UserMsg, "Apparently failed!");
VG_(message)(Vg_UserMsg, "");
}
}
/* Print some helpful-ish text about unimplemented things, and give
up. */
void VG_(unimplemented) ( Char* msg )
{
VG_(message)(Vg_UserMsg, "");
VG_(message)(Vg_UserMsg,
"Valgrind detected that your program requires");
VG_(message)(Vg_UserMsg,
"the following unimplemented functionality:");
VG_(message)(Vg_UserMsg, " %s", msg);
VG_(message)(Vg_UserMsg,
"This may be because the functionality is hard to implement,");
VG_(message)(Vg_UserMsg,
"or because no reasonable program would behave this way,");
VG_(message)(Vg_UserMsg,
"or because nobody has yet needed it. In any case, let me know");
VG_(message)(Vg_UserMsg,
"(jseward@acm.org) and/or try to work around the problem, if you can.");
VG_(message)(Vg_UserMsg,
"");
VG_(message)(Vg_UserMsg,
"Valgrind has to exit now. Sorry. Bye!");
VG_(message)(Vg_UserMsg,
"");
VG_(pp_sched_status)();
VG_(exit)(1);
}
void VG_(nvidia_moan) ( void)
{
VG_(message)(Vg_UserMsg,
"The following failure _might_ be caused by linking to NVidia's\n "
"libGL.so, so avoiding it, if you can, _might_ help you. For example,\n "
"re-build any Qt libraries you are using without OpenGL support.");
}
/*--------------------------------------------------------------------*/
/*--- end vg_main.c ---*/
/*--------------------------------------------------------------------*/