blob: 18bea183179037d5c0a0be31ef8cdf684bb96cea [file] [log] [blame]
/*--------------------------------------------------------------------*/
/*--- A replacement for the standard libpthread.so. ---*/
/*--- vg_libpthread.c ---*/
/*--------------------------------------------------------------------*/
/*
This file is part of Valgrind, an extensible x86 protected-mode
emulator for monitoring program execution on x86-Unixes.
Copyright (C) 2000-2002 Julian Seward
jseward@acm.org
This program is free software; you can redistribute it and/or
modify it under the terms of the GNU General Public License as
published by the Free Software Foundation; either version 2 of the
License, or (at your option) any later version.
This program is distributed in the hope that it will be useful, but
WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program; if not, write to the Free Software
Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA
02111-1307, USA.
The GNU General Public License is contained in the file COPYING.
*/
/* ALL THIS CODE RUNS ON THE SIMULATED CPU.
This is a replacement for the standard libpthread.so. It is loaded
as part of the client's image (if required) and directs pthread
calls through to Valgrind's request mechanism.
A couple of caveats.
1. Since it's a binary-compatible replacement for an existing library,
we must take care to used exactly the same data layouts, etc, as
the standard pthread.so does.
2. Since this runs as part of the client, there are no specific
restrictions on what headers etc we can include, so long as
this libpthread.so does not end up having dependencies on .so's
which the real one doesn't.
Later ... it appears we cannot call file-related stuff in libc here,
perhaps fair enough. Be careful what you call from here. Even exit()
doesn't work (gives infinite recursion and then stack overflow); hence
myexit(). Also fprintf doesn't seem safe.
*/
#include "valgrind.h" /* For the request-passing mechanism */
#include "vg_include.h" /* For the VG_USERREQ__* constants */
#define __USE_UNIX98
#include <sys/types.h>
#include <pthread.h>
#undef __USE_UNIX98
#include <unistd.h>
#include <string.h>
#ifdef GLIBC_2_1
#include <sys/time.h>
#endif
#include <sys/stat.h>
#include <sys/poll.h>
#include <stdio.h>
/* ---------------------------------------------------------------------
Forwardses.
------------------------------------------------------------------ */
static void wait_for_fd_to_be_readable_or_erring ( int fd );
static void wait_for_fd_to_be_writable_or_erring ( int fd );
static
int my_do_syscall1 ( int syscallno, int arg1 );
static
int my_do_syscall2 ( int syscallno,
int arg1, int arg2 );
static
int my_do_syscall3 ( int syscallno,
int arg1, int arg2, int arg3 );
static
__inline__
int is_kerror ( int res )
{
if (res >= -4095 && res <= -1)
return 1;
else
return 0;
}
#ifdef GLIBC_2_3
/* kludge by JRS (not from glibc) ... */
typedef void* __locale_t;
/* Copied from locale/locale.h in glibc-2.2.93 sources */
/* This value can be passed to `uselocale' and may be returned by
it. Passing this value to any other function has undefined
behavior. */
# define LC_GLOBAL_LOCALE ((__locale_t) -1L)
extern __locale_t __uselocale ( __locale_t );
#endif
static
void init_libc_tsd_keys ( void );
/* ---------------------------------------------------------------------
Helpers. We have to be pretty self-sufficient.
------------------------------------------------------------------ */
/* Number of times any given error message is printed. */
#define N_MOANS 3
/* Extract from Valgrind the value of VG_(clo_trace_pthread_level).
Returns 0 (none) if not running on Valgrind. */
static
int get_pt_trace_level ( void )
{
int res;
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__GET_PTHREAD_TRACE_LEVEL,
0, 0, 0, 0);
return res;
}
static
void my_exit ( int arg )
{
my_do_syscall1(__NR_exit, arg);
/*NOTREACHED*/
}
/* Apparently unused.
static
void my_write ( int fd, const void *buf, int count )
{
my_do_syscall3(__NR_write, fd, (int)buf, count );
}
*/
/* We need this guy -- it's in valgrind.so. */
extern void VG_(startup) ( void );
/* Just start up Valgrind if it's not already going. VG_(startup)()
detects and ignores second and subsequent calls. */
static __inline__
void ensure_valgrind ( char* caller )
{
VG_(startup)();
}
/* While we're at it ... hook our own startup function into this
game. */
__asm__ (
".section .init\n"
"\tcall vgPlain_startup"
);
static
__attribute__((noreturn))
void barf ( char* str )
{
int res;
char buf[1000];
buf[0] = 0;
strcat(buf, "\nvalgrind's libpthread.so: ");
strcat(buf, str);
strcat(buf, "\n\n");
VALGRIND_MAGIC_SEQUENCE(res, 0, /* irrelevant default */
VG_USERREQ__LOGMESSAGE, buf, 0, 0, 0);
my_exit(1);
/* We have to persuade gcc into believing this doesn't return. */
while (1) { };
}
static void cat_n_send ( char* pre, char* msg )
{
char buf[1000];
int res;
if (get_pt_trace_level() >= 0) {
snprintf(buf, sizeof(buf), "%s%s", pre, msg );
buf[sizeof(buf)-1] = '\0';
VALGRIND_MAGIC_SEQUENCE(res, 0, /* irrelevant default */
VG_USERREQ__LOGMESSAGE, buf, 0, 0, 0);
}
}
static void ignored ( char* msg )
{
cat_n_send ( "valgrind's libpthread.so: IGNORED call to: ", msg );
}
static void kludged ( char* msg )
{
cat_n_send ( "valgrind's libpthread.so: KLUDGED call to: ", msg );
}
__attribute__((noreturn))
void vgPlain_unimp ( char* what )
{
cat_n_send (
"valgrind's libpthread.so: UNIMPLEMENTED FUNCTION: ", what );
barf("Please report this bug to me at: jseward@acm.org");
}
static
void my_assert_fail ( Char* expr, Char* file, Int line, Char* fn )
{
char buf[1000];
static Bool entered = False;
if (entered)
my_exit(2);
entered = True;
sprintf(buf, "\n%s: %s:%d (%s): Assertion `%s' failed.\n",
"valgrind", file, line, fn, expr );
cat_n_send ( "", buf );
sprintf(buf, "Please report this bug to me at: %s\n\n",
VG_EMAIL_ADDR);
cat_n_send ( "", buf );
my_exit(1);
}
#define MY__STRING(__str) #__str
#define my_assert(expr) \
((void) ((expr) ? 0 : \
(my_assert_fail (MY__STRING(expr), \
__FILE__, __LINE__, \
__PRETTY_FUNCTION__), 0)))
static
void my_free ( void* ptr )
{
int res;
VALGRIND_MAGIC_SEQUENCE(res, (-1) /* default */,
VG_USERREQ__FREE, ptr, 0, 0, 0);
my_assert(res == 0);
}
static
void* my_malloc ( int nbytes )
{
void* res;
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__MALLOC, nbytes, 0, 0, 0);
my_assert(res != (void*)0);
return res;
}
/* ---------------------------------------------------------------------
Pass pthread_ calls to Valgrind's request mechanism.
------------------------------------------------------------------ */
#include <errno.h>
#include <sys/time.h> /* gettimeofday */
/* ---------------------------------------------------
Ummm ..
------------------------------------------------ */
static
void pthread_error ( const char* msg )
{
int res;
VALGRIND_MAGIC_SEQUENCE(res, 0,
VG_USERREQ__PTHREAD_ERROR,
msg, 0, 0, 0);
}
/* ---------------------------------------------------
Here so it can be inlined without complaint.
------------------------------------------------ */
__inline__
pthread_t pthread_self(void)
{
int tid;
ensure_valgrind("pthread_self");
VALGRIND_MAGIC_SEQUENCE(tid, 0 /* default */,
VG_USERREQ__PTHREAD_GET_THREADID,
0, 0, 0, 0);
if (tid < 1 || tid >= VG_N_THREADS)
barf("pthread_self: invalid ThreadId");
return tid;
}
/* ---------------------------------------------------
THREAD ATTRIBUTES
------------------------------------------------ */
int pthread_attr_init(pthread_attr_t *attr)
{
/* Just initialise the fields which we might look at. */
attr->__detachstate = PTHREAD_CREATE_JOINABLE;
/* Linuxthreads sets this field to the value __getpagesize(), so I
guess the following is OK. */
attr->__guardsize = VKI_BYTES_PER_PAGE; return 0;
}
int pthread_attr_setdetachstate(pthread_attr_t *attr, int detachstate)
{
if (detachstate != PTHREAD_CREATE_JOINABLE
&& detachstate != PTHREAD_CREATE_DETACHED) {
pthread_error("pthread_attr_setdetachstate: "
"detachstate is invalid");
return EINVAL;
}
attr->__detachstate = detachstate;
return 0;
}
int pthread_attr_getdetachstate(const pthread_attr_t *attr, int *detachstate)
{
*detachstate = attr->__detachstate;
return 0;
}
int pthread_attr_setinheritsched(pthread_attr_t *attr, int inherit)
{
static int moans = N_MOANS;
if (moans-- > 0)
ignored("pthread_attr_setinheritsched");
return 0;
}
__attribute__((weak))
int pthread_attr_setstacksize (pthread_attr_t *__attr,
size_t __stacksize)
{
size_t limit;
char buf[1024];
ensure_valgrind("pthread_attr_setstacksize");
limit = VG_PTHREAD_STACK_SIZE - VG_AR_CLIENT_STACKBASE_REDZONE_SZB
- 1000; /* paranoia */
if (__stacksize < limit)
return 0;
snprintf(buf, sizeof(buf), "pthread_attr_setstacksize: "
"requested size %d >= VG_PTHREAD_STACK_SIZE\n "
"edit vg_include.h and rebuild.", __stacksize);
buf[sizeof(buf)-1] = '\0'; /* Make sure it is zero terminated */
barf(buf);
}
/* This is completely bogus. */
int pthread_attr_getschedparam(const pthread_attr_t *attr,
struct sched_param *param)
{
static int moans = N_MOANS;
if (moans-- > 0)
kludged("pthread_attr_getschedparam");
# ifdef HAVE_SCHED_PRIORITY
if (param) param->sched_priority = 0; /* who knows */
# else
if (param) param->__sched_priority = 0; /* who knows */
# endif
return 0;
}
int pthread_attr_setschedparam(pthread_attr_t *attr,
const struct sched_param *param)
{
static int moans = N_MOANS;
if (moans-- > 0)
ignored("pthread_attr_setschedparam");
return 0;
}
int pthread_attr_destroy(pthread_attr_t *attr)
{
static int moans = N_MOANS;
if (moans-- > 0)
ignored("pthread_attr_destroy");
return 0;
}
/* These are no-ops, as with LinuxThreads. */
int pthread_attr_setscope ( pthread_attr_t *attr, int scope )
{
ensure_valgrind("pthread_attr_setscope");
if (scope == PTHREAD_SCOPE_SYSTEM)
return 0;
pthread_error("pthread_attr_setscope: "
"invalid or unsupported scope");
if (scope == PTHREAD_SCOPE_PROCESS)
return ENOTSUP;
return EINVAL;
}
int pthread_attr_getscope ( const pthread_attr_t *attr, int *scope )
{
ensure_valgrind("pthread_attr_setscope");
if (scope)
*scope = PTHREAD_SCOPE_SYSTEM;
return 0;
}
/* Pretty bogus. Avoid if possible. */
int pthread_getattr_np (pthread_t thread, pthread_attr_t *attr)
{
int detached;
size_t limit;
ensure_valgrind("pthread_getattr_np");
kludged("pthread_getattr_np");
limit = VG_PTHREAD_STACK_SIZE - VG_AR_CLIENT_STACKBASE_REDZONE_SZB
- 1000; /* paranoia */
attr->__detachstate = PTHREAD_CREATE_JOINABLE;
attr->__schedpolicy = SCHED_OTHER;
attr->__schedparam.sched_priority = 0;
attr->__inheritsched = PTHREAD_EXPLICIT_SCHED;
attr->__scope = PTHREAD_SCOPE_SYSTEM;
attr->__guardsize = VKI_BYTES_PER_PAGE;
attr->__stackaddr = NULL;
attr->__stackaddr_set = 0;
attr->__stacksize = limit;
VALGRIND_MAGIC_SEQUENCE(detached, (-1) /* default */,
VG_USERREQ__SET_OR_GET_DETACH,
2 /* get */, thread, 0, 0);
my_assert(detached == 0 || detached == 1);
if (detached)
attr->__detachstate = PTHREAD_CREATE_DETACHED;
return 0;
}
/* Bogus ... */
__attribute__((weak))
int pthread_attr_getstackaddr ( const pthread_attr_t * attr,
void ** stackaddr )
{
ensure_valgrind("pthread_attr_getstackaddr");
kludged("pthread_attr_getstackaddr");
if (stackaddr)
*stackaddr = NULL;
return 0;
}
/* Not bogus (!) */
__attribute__((weak))
int pthread_attr_getstacksize ( const pthread_attr_t * _attr,
size_t * __stacksize )
{
size_t limit;
ensure_valgrind("pthread_attr_getstacksize");
limit = VG_PTHREAD_STACK_SIZE - VG_AR_CLIENT_STACKBASE_REDZONE_SZB
- 1000; /* paranoia */
if (__stacksize)
*__stacksize = limit;
return 0;
}
int pthread_attr_setschedpolicy(pthread_attr_t *attr, int policy)
{
if (policy != SCHED_OTHER && policy != SCHED_FIFO && policy != SCHED_RR)
return EINVAL;
attr->__schedpolicy = policy;
return 0;
}
int pthread_attr_getschedpolicy(const pthread_attr_t *attr, int *policy)
{
*policy = attr->__schedpolicy;
return 0;
}
/* This is completely bogus. We reject all attempts to change it from
VKI_BYTES_PER_PAGE. I don't have a clue what it's for so it seems
safest to be paranoid. */
__attribute__((weak))
int pthread_attr_setguardsize(pthread_attr_t *attr, size_t guardsize)
{
static int moans = N_MOANS;
if (guardsize == VKI_BYTES_PER_PAGE)
return 0;
if (moans-- > 0)
ignored("pthread_attr_setguardsize: ignoring guardsize != 4096");
return 0;
}
/* A straight copy of the LinuxThreads code. */
__attribute__((weak))
int pthread_attr_getguardsize(const pthread_attr_t *attr, size_t *guardsize)
{
*guardsize = attr->__guardsize;
return 0;
}
/* Again, like LinuxThreads. */
static int concurrency_current_level = 0;
__attribute__((weak))
int pthread_setconcurrency(int new_level)
{
if (new_level < 0)
return EINVAL;
else {
concurrency_current_level = new_level;
return 0;
}
}
__attribute__((weak))
int pthread_getconcurrency(void)
{
return concurrency_current_level;
}
/* ---------------------------------------------------
Helper functions for running a thread
and for clearing up afterwards.
------------------------------------------------ */
/* All exiting threads eventually pass through here, bearing the
return value, or PTHREAD_CANCELED, in ret_val. */
static
__attribute__((noreturn))
void thread_exit_wrapper ( void* ret_val )
{
int detached, res;
CleanupEntry cu;
pthread_key_t key;
void** specifics_ptr;
/* Run this thread's cleanup handlers. */
while (1) {
VALGRIND_MAGIC_SEQUENCE(res, (-1) /* default */,
VG_USERREQ__CLEANUP_POP,
&cu, 0, 0, 0);
if (res == -1) break; /* stack empty */
my_assert(res == 0);
if (0) printf("running exit cleanup handler");
cu.fn ( cu.arg );
}
/* Run this thread's key finalizers. Really this should be run
PTHREAD_DESTRUCTOR_ITERATIONS times. */
for (key = 0; key < VG_N_THREAD_KEYS; key++) {
VALGRIND_MAGIC_SEQUENCE(res, (-2) /* default */,
VG_USERREQ__GET_KEY_D_AND_S,
key, &cu, 0, 0 );
if (res == 0) {
/* valid key */
if (cu.fn && cu.arg)
cu.fn /* destructor for key */
( cu.arg /* specific for key for this thread */ );
continue;
}
my_assert(res == -1);
}
/* Free up my specifics space, if any. */
VALGRIND_MAGIC_SEQUENCE(specifics_ptr, 3 /* default */,
VG_USERREQ__PTHREAD_GETSPECIFIC_PTR,
pthread_self(), 0, 0, 0);
my_assert(specifics_ptr != (void**)3);
my_assert(specifics_ptr != (void**)1); /* 1 means invalid thread */
if (specifics_ptr != NULL)
my_free(specifics_ptr);
/* Decide on my final disposition. */
VALGRIND_MAGIC_SEQUENCE(detached, (-1) /* default */,
VG_USERREQ__SET_OR_GET_DETACH,
2 /* get */, pthread_self(), 0, 0);
my_assert(detached == 0 || detached == 1);
if (detached) {
/* Detached; I just quit right now. */
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__QUIT, 0, 0, 0, 0);
} else {
/* Not detached; so I wait for a joiner. */
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__WAIT_JOINER, ret_val, 0, 0, 0);
}
/* NOTREACHED */
barf("thread_exit_wrapper: still alive?!");
}
/* This function is a wrapper function for running a thread. It runs
the root function specified in pthread_create, and then, should the
root function return a value, it arranges to run the thread's
cleanup handlers and exit correctly. */
/* Struct used to convey info from pthread_create to thread_wrapper.
Must be careful not to pass to the child thread any pointers to
objects which might be on the parent's stack. */
typedef
struct {
int attr__detachstate;
void* (*root_fn) ( void* );
void* arg;
}
NewThreadInfo;
/* This is passed to the VG_USERREQ__APPLY_IN_NEW_THREAD and so must
not return. Note that this runs in the new thread, not the
parent. */
static
__attribute__((noreturn))
void thread_wrapper ( NewThreadInfo* info )
{
int attr__detachstate;
void* (*root_fn) ( void* );
void* arg;
void* ret_val;
attr__detachstate = info->attr__detachstate;
root_fn = info->root_fn;
arg = info->arg;
/* Free up the arg block that pthread_create malloced. */
my_free(info);
/* Minimally observe the attributes supplied. */
if (attr__detachstate != PTHREAD_CREATE_DETACHED
&& attr__detachstate != PTHREAD_CREATE_JOINABLE)
pthread_error("thread_wrapper: invalid attr->__detachstate");
if (attr__detachstate == PTHREAD_CREATE_DETACHED)
pthread_detach(pthread_self());
# ifdef GLIBC_2_3
/* Set this thread's locale to the global (default) locale. A hack
in support of glibc-2.3. This does the biz for the all new
threads; the root thread is done with a horrible hack in
init_libc_tsd_keys() below.
*/
__uselocale(LC_GLOBAL_LOCALE);
# endif
/* The root function might not return. But if it does we simply
move along to thread_exit_wrapper. All other ways out for the
thread (cancellation, or calling pthread_exit) lead there
too. */
ret_val = root_fn(arg);
thread_exit_wrapper(ret_val);
/* NOTREACHED */
}
/* ---------------------------------------------------
THREADs
------------------------------------------------ */
__attribute__((weak))
int pthread_yield ( void )
{
int res;
ensure_valgrind("pthread_yield");
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__PTHREAD_YIELD, 0, 0, 0, 0);
return 0;
}
int pthread_equal(pthread_t thread1, pthread_t thread2)
{
return thread1 == thread2 ? 1 : 0;
}
/* Bundle up the args into a malloc'd block and create a new thread
consisting of thread_wrapper() applied to said malloc'd block. */
int
pthread_create (pthread_t *__restrict __thredd,
__const pthread_attr_t *__restrict __attr,
void *(*__start_routine) (void *),
void *__restrict __arg)
{
int tid_child;
NewThreadInfo* info;
ensure_valgrind("pthread_create");
/* make sure the tsd keys, and hence locale info, are initialised
before we get into complications making new threads. */
init_libc_tsd_keys();
/* Allocate space for the arg block. thread_wrapper will free
it. */
info = my_malloc(sizeof(NewThreadInfo));
my_assert(info != NULL);
if (__attr)
info->attr__detachstate = __attr->__detachstate;
else
info->attr__detachstate = PTHREAD_CREATE_JOINABLE;
info->root_fn = __start_routine;
info->arg = __arg;
VALGRIND_MAGIC_SEQUENCE(tid_child, VG_INVALID_THREADID /* default */,
VG_USERREQ__APPLY_IN_NEW_THREAD,
&thread_wrapper, info, 0, 0);
my_assert(tid_child != VG_INVALID_THREADID);
if (__thredd)
*__thredd = tid_child;
return 0; /* success */
}
int
pthread_join (pthread_t __th, void **__thread_return)
{
int res;
ensure_valgrind("pthread_join");
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__PTHREAD_JOIN,
__th, __thread_return, 0, 0);
return res;
}
void pthread_exit(void *retval)
{
ensure_valgrind("pthread_exit");
/* Simple! */
thread_exit_wrapper(retval);
}
int pthread_detach(pthread_t th)
{
int res;
ensure_valgrind("pthread_detach");
/* First we enquire as to the current detach state. */
VALGRIND_MAGIC_SEQUENCE(res, (-2) /* default */,
VG_USERREQ__SET_OR_GET_DETACH,
2 /* get */, th, 0, 0);
if (res == -1) {
/* not found */
pthread_error("pthread_detach: "
"invalid target thread");
return ESRCH;
}
if (res == 1) {
/* already detached */
pthread_error("pthread_detach: "
"target thread is already detached");
return EINVAL;
}
if (res == 0) {
VALGRIND_MAGIC_SEQUENCE(res, (-2) /* default */,
VG_USERREQ__SET_OR_GET_DETACH,
1 /* set */, th, 0, 0);
my_assert(res == 0);
return 0;
}
barf("pthread_detach");
}
/* ---------------------------------------------------
CLEANUP STACKS
------------------------------------------------ */
void _pthread_cleanup_push (struct _pthread_cleanup_buffer *__buffer,
void (*__routine) (void *),
void *__arg)
{
int res;
CleanupEntry cu;
ensure_valgrind("_pthread_cleanup_push");
cu.fn = __routine;
cu.arg = __arg;
VALGRIND_MAGIC_SEQUENCE(res, (-1) /* default */,
VG_USERREQ__CLEANUP_PUSH,
&cu, 0, 0, 0);
my_assert(res == 0);
}
void _pthread_cleanup_push_defer (struct _pthread_cleanup_buffer *__buffer,
void (*__routine) (void *),
void *__arg)
{
/* As _pthread_cleanup_push, but first save the thread's original
cancellation type in __buffer and set it to Deferred. */
int orig_ctype;
ensure_valgrind("_pthread_cleanup_push_defer");
/* Set to Deferred, and put the old cancellation type in res. */
my_assert(-1 != PTHREAD_CANCEL_DEFERRED);
my_assert(-1 != PTHREAD_CANCEL_ASYNCHRONOUS);
my_assert(sizeof(struct _pthread_cleanup_buffer) >= sizeof(int));
VALGRIND_MAGIC_SEQUENCE(orig_ctype, (-1) /* default */,
VG_USERREQ__SET_CANCELTYPE,
PTHREAD_CANCEL_DEFERRED, 0, 0, 0);
my_assert(orig_ctype != -1);
*((int*)(__buffer)) = orig_ctype;
/* Now push the cleanup. */
_pthread_cleanup_push(NULL, __routine, __arg);
}
void _pthread_cleanup_pop (struct _pthread_cleanup_buffer *__buffer,
int __execute)
{
int res;
CleanupEntry cu;
ensure_valgrind("_pthread_cleanup_push");
cu.fn = cu.arg = NULL; /* paranoia */
VALGRIND_MAGIC_SEQUENCE(res, (-1) /* default */,
VG_USERREQ__CLEANUP_POP,
&cu, 0, 0, 0);
if (res == 0) {
/* pop succeeded */
if (__execute) {
cu.fn ( cu.arg );
}
return;
}
if (res == -1) {
/* stack underflow */
return;
}
barf("_pthread_cleanup_pop");
}
void _pthread_cleanup_pop_restore (struct _pthread_cleanup_buffer *__buffer,
int __execute)
{
int orig_ctype, fake_ctype;
/* As _pthread_cleanup_pop, but after popping/running the handler,
restore the thread's original cancellation type from the first
word of __buffer. */
_pthread_cleanup_pop(NULL, __execute);
orig_ctype = *((int*)(__buffer));
my_assert(orig_ctype == PTHREAD_CANCEL_DEFERRED
|| orig_ctype == PTHREAD_CANCEL_ASYNCHRONOUS);
my_assert(-1 != PTHREAD_CANCEL_DEFERRED);
my_assert(-1 != PTHREAD_CANCEL_ASYNCHRONOUS);
my_assert(sizeof(struct _pthread_cleanup_buffer) >= sizeof(int));
VALGRIND_MAGIC_SEQUENCE(fake_ctype, (-1) /* default */,
VG_USERREQ__SET_CANCELTYPE,
orig_ctype, 0, 0, 0);
my_assert(fake_ctype == PTHREAD_CANCEL_DEFERRED);
}
/* ---------------------------------------------------
MUTEX ATTRIBUTES
------------------------------------------------ */
int __pthread_mutexattr_init(pthread_mutexattr_t *attr)
{
attr->__mutexkind = PTHREAD_MUTEX_ERRORCHECK_NP;
return 0;
}
int __pthread_mutexattr_settype(pthread_mutexattr_t *attr, int type)
{
switch (type) {
# ifndef GLIBC_2_1
case PTHREAD_MUTEX_TIMED_NP:
case PTHREAD_MUTEX_ADAPTIVE_NP:
# endif
# ifdef GLIBC_2_1
case PTHREAD_MUTEX_FAST_NP:
# endif
case PTHREAD_MUTEX_RECURSIVE_NP:
case PTHREAD_MUTEX_ERRORCHECK_NP:
attr->__mutexkind = type;
return 0;
default:
pthread_error("pthread_mutexattr_settype: "
"invalid type");
return EINVAL;
}
}
int __pthread_mutexattr_destroy(pthread_mutexattr_t *attr)
{
return 0;
}
/* ---------------------------------------------------
MUTEXes
------------------------------------------------ */
int __pthread_mutex_init(pthread_mutex_t *mutex,
const pthread_mutexattr_t *mutexattr)
{
mutex->__m_count = 0;
mutex->__m_owner = (_pthread_descr)VG_INVALID_THREADID;
mutex->__m_kind = PTHREAD_MUTEX_ERRORCHECK_NP;
if (mutexattr)
mutex->__m_kind = mutexattr->__mutexkind;
return 0;
}
int __pthread_mutex_lock(pthread_mutex_t *mutex)
{
int res;
if (RUNNING_ON_VALGRIND) {
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__PTHREAD_MUTEX_LOCK,
mutex, 0, 0, 0);
return res;
} else {
/* Play at locking */
if (0)
kludged("prehistoric lock");
mutex->__m_owner = (_pthread_descr)1;
mutex->__m_count = 1;
mutex->__m_kind |= VG_PTHREAD_PREHISTORY;
return 0; /* success */
}
}
int __pthread_mutex_trylock(pthread_mutex_t *mutex)
{
int res;
if (RUNNING_ON_VALGRIND) {
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__PTHREAD_MUTEX_TRYLOCK,
mutex, 0, 0, 0);
return res;
} else {
/* Play at locking */
if (0)
kludged("prehistoric trylock");
mutex->__m_owner = (_pthread_descr)1;
mutex->__m_count = 1;
mutex->__m_kind |= VG_PTHREAD_PREHISTORY;
return 0; /* success */
}
}
int __pthread_mutex_unlock(pthread_mutex_t *mutex)
{
int res;
if (RUNNING_ON_VALGRIND) {
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__PTHREAD_MUTEX_UNLOCK,
mutex, 0, 0, 0);
return res;
} else {
/* Play at locking */
if (0)
kludged("prehistoric unlock");
mutex->__m_owner = 0;
mutex->__m_count = 0;
mutex->__m_kind &= ~VG_PTHREAD_PREHISTORY;
return 0; /* success */
}
}
int __pthread_mutex_destroy(pthread_mutex_t *mutex)
{
/* Valgrind doesn't hold any resources on behalf of the mutex, so no
need to involve it. */
if (mutex->__m_count > 0) {
/* Oh, the horror. glibc's internal use of pthreads "knows"
that destroying a lock does an implicit unlock. Make it
explicit. */
__pthread_mutex_unlock(mutex);
pthread_error("pthread_mutex_destroy: "
"mutex is still in use");
return EBUSY;
}
mutex->__m_count = 0;
mutex->__m_owner = (_pthread_descr)VG_INVALID_THREADID;
mutex->__m_kind = PTHREAD_MUTEX_ERRORCHECK_NP;
return 0;
}
/* ---------------------------------------------------
CONDITION VARIABLES
------------------------------------------------ */
/* LinuxThreads supports no attributes for conditions. Hence ... */
int pthread_condattr_init(pthread_condattr_t *attr)
{
return 0;
}
int pthread_condattr_destroy(pthread_condattr_t *attr)
{
return 0;
}
int pthread_cond_init( pthread_cond_t *cond,
const pthread_condattr_t *cond_attr)
{
cond->__c_waiting = (_pthread_descr)VG_INVALID_THREADID;
return 0;
}
int pthread_cond_destroy(pthread_cond_t *cond)
{
/* should check that no threads are waiting on this CV */
static int moans = N_MOANS;
if (moans-- > 0)
kludged("pthread_cond_destroy");
return 0;
}
/* ---------------------------------------------------
SCHEDULING
------------------------------------------------ */
/* This is completely bogus. */
int pthread_getschedparam(pthread_t target_thread,
int *policy,
struct sched_param *param)
{
static int moans = N_MOANS;
if (moans-- > 0)
kludged("pthread_getschedparam");
if (policy) *policy = SCHED_OTHER;
# ifdef HAVE_SCHED_PRIORITY
if (param) param->sched_priority = 0; /* who knows */
# else
if (param) param->__sched_priority = 0; /* who knows */
# endif
return 0;
}
int pthread_setschedparam(pthread_t target_thread,
int policy,
const struct sched_param *param)
{
static int moans = N_MOANS;
if (moans-- > 0)
ignored("pthread_setschedparam");
return 0;
}
int pthread_cond_wait(pthread_cond_t *cond, pthread_mutex_t *mutex)
{
int res;
ensure_valgrind("pthread_cond_wait");
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__PTHREAD_COND_WAIT,
cond, mutex, 0, 0);
return res;
}
int pthread_cond_timedwait ( pthread_cond_t *cond,
pthread_mutex_t *mutex,
const struct timespec *abstime )
{
int res;
unsigned int ms_now, ms_end;
struct timeval timeval_now;
unsigned long long int ull_ms_now_after_1970;
unsigned long long int ull_ms_end_after_1970;
ensure_valgrind("pthread_cond_timedwait");
VALGRIND_MAGIC_SEQUENCE(ms_now, 0xFFFFFFFF /* default */,
VG_USERREQ__READ_MILLISECOND_TIMER,
0, 0, 0, 0);
my_assert(ms_now != 0xFFFFFFFF);
res = gettimeofday(&timeval_now, NULL);
my_assert(res == 0);
ull_ms_now_after_1970
= 1000ULL * ((unsigned long long int)(timeval_now.tv_sec))
+ ((unsigned long long int)(timeval_now.tv_usec / 1000000));
ull_ms_end_after_1970
= 1000ULL * ((unsigned long long int)(abstime->tv_sec))
+ ((unsigned long long int)(abstime->tv_nsec / 1000000));
if (ull_ms_end_after_1970 < ull_ms_now_after_1970)
ull_ms_end_after_1970 = ull_ms_now_after_1970;
ms_end
= ms_now + (unsigned int)(ull_ms_end_after_1970 - ull_ms_now_after_1970);
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__PTHREAD_COND_TIMEDWAIT,
cond, mutex, ms_end, 0);
return res;
}
int pthread_cond_signal(pthread_cond_t *cond)
{
int res;
ensure_valgrind("pthread_cond_signal");
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__PTHREAD_COND_SIGNAL,
cond, 0, 0, 0);
return res;
}
int pthread_cond_broadcast(pthread_cond_t *cond)
{
int res;
ensure_valgrind("pthread_cond_broadcast");
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__PTHREAD_COND_BROADCAST,
cond, 0, 0, 0);
return res;
}
/* ---------------------------------------------------
CANCELLATION
------------------------------------------------ */
int pthread_setcancelstate(int state, int *oldstate)
{
int res;
ensure_valgrind("pthread_setcancelstate");
if (state != PTHREAD_CANCEL_ENABLE
&& state != PTHREAD_CANCEL_DISABLE) {
pthread_error("pthread_setcancelstate: "
"invalid state");
return EINVAL;
}
my_assert(-1 != PTHREAD_CANCEL_ENABLE);
my_assert(-1 != PTHREAD_CANCEL_DISABLE);
VALGRIND_MAGIC_SEQUENCE(res, (-1) /* default */,
VG_USERREQ__SET_CANCELSTATE,
state, 0, 0, 0);
my_assert(res != -1);
if (oldstate)
*oldstate = res;
return 0;
}
int pthread_setcanceltype(int type, int *oldtype)
{
int res;
ensure_valgrind("pthread_setcanceltype");
if (type != PTHREAD_CANCEL_DEFERRED
&& type != PTHREAD_CANCEL_ASYNCHRONOUS) {
pthread_error("pthread_setcanceltype: "
"invalid type");
return EINVAL;
}
my_assert(-1 != PTHREAD_CANCEL_DEFERRED);
my_assert(-1 != PTHREAD_CANCEL_ASYNCHRONOUS);
VALGRIND_MAGIC_SEQUENCE(res, (-1) /* default */,
VG_USERREQ__SET_CANCELTYPE,
type, 0, 0, 0);
my_assert(res != -1);
if (oldtype)
*oldtype = res;
return 0;
}
int pthread_cancel(pthread_t thread)
{
int res;
ensure_valgrind("pthread_cancel");
VALGRIND_MAGIC_SEQUENCE(res, (-1) /* default */,
VG_USERREQ__SET_CANCELPEND,
thread, &thread_exit_wrapper, 0, 0);
my_assert(res != -1);
return res;
}
static __inline__
void __my_pthread_testcancel(void)
{
int res;
ensure_valgrind("__my_pthread_testcancel");
VALGRIND_MAGIC_SEQUENCE(res, (-1) /* default */,
VG_USERREQ__TESTCANCEL,
0, 0, 0, 0);
my_assert(res == 0);
}
void pthread_testcancel ( void )
{
__my_pthread_testcancel();
}
/* Not really sure what this is for. I suspect for doing the POSIX
requirements for fork() and exec(). We do this internally anyway
whenever those syscalls are observed, so this could be superfluous,
but hey ...
*/
void __pthread_kill_other_threads_np ( void )
{
int res;
ensure_valgrind("__pthread_kill_other_threads_np");
VALGRIND_MAGIC_SEQUENCE(res, (-1) /* default */,
VG_USERREQ__NUKE_OTHER_THREADS,
0, 0, 0, 0);
my_assert(res == 0);
}
/* ---------------------------------------------------
SIGNALS
------------------------------------------------ */
#include <signal.h>
int pthread_sigmask(int how, const sigset_t *newmask,
sigset_t *oldmask)
{
int res;
/* A bit subtle, because the scheduler expects newmask and oldmask
to be vki_sigset_t* rather than sigset_t*, and the two are
different. Fortunately the first 64 bits of a sigset_t are
exactly a vki_sigset_t, so we just pass the pointers through
unmodified. Haaaack!
Also mash the how value so that the SIG_ constants from glibc
constants to VKI_ constants, so that the former do not have to
be included into vg_scheduler.c. */
ensure_valgrind("pthread_sigmask");
switch (how) {
case SIG_SETMASK: how = VKI_SIG_SETMASK; break;
case SIG_BLOCK: how = VKI_SIG_BLOCK; break;
case SIG_UNBLOCK: how = VKI_SIG_UNBLOCK; break;
default: pthread_error("pthread_sigmask: invalid how");
return EINVAL;
}
/* Crude check */
if (newmask == NULL)
return EFAULT;
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__PTHREAD_SIGMASK,
how, newmask, oldmask, 0);
/* The scheduler tells us of any memory violations. */
return res == 0 ? 0 : EFAULT;
}
int sigwait ( const sigset_t* set, int* sig )
{
int res;
ensure_valgrind("sigwait");
/* As with pthread_sigmask we deliberately confuse sigset_t with
vki_ksigset_t. */
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__SIGWAIT,
set, sig, 0, 0);
return res;
}
int pthread_kill(pthread_t thread, int signo)
{
int res;
ensure_valgrind("pthread_kill");
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__PTHREAD_KILL,
thread, signo, 0, 0);
return res;
}
/* Copied verbatim from Linuxthreads */
/* Redefine raise() to send signal to calling thread only,
as per POSIX 1003.1c */
int raise (int sig)
{
int retcode = pthread_kill(pthread_self(), sig);
if (retcode == 0) {
return 0;
} else {
errno = retcode;
return -1;
}
}
int pause ( void )
{
unsigned int n_orig, n_now;
struct vki_timespec nanosleep_interval;
ensure_valgrind("pause");
/* This is surely a cancellation point. */
__my_pthread_testcancel();
VALGRIND_MAGIC_SEQUENCE(n_orig, 0xFFFFFFFF /* default */,
VG_USERREQ__GET_N_SIGS_RETURNED,
0, 0, 0, 0);
my_assert(n_orig != 0xFFFFFFFF);
while (1) {
VALGRIND_MAGIC_SEQUENCE(n_now, 0xFFFFFFFF /* default */,
VG_USERREQ__GET_N_SIGS_RETURNED,
0, 0, 0, 0);
my_assert(n_now != 0xFFFFFFFF);
my_assert(n_now >= n_orig);
if (n_now != n_orig) break;
nanosleep_interval.tv_sec = 0;
nanosleep_interval.tv_nsec = 12 * 1000 * 1000; /* 12 milliseconds */
/* It's critical here that valgrind's nanosleep implementation
is nonblocking. */
(void)my_do_syscall2(__NR_nanosleep,
(int)(&nanosleep_interval), (int)NULL);
}
* (__errno_location()) = EINTR;
return -1;
}
/* ---------------------------------------------------
THREAD-SPECIFICs
------------------------------------------------ */
static
int key_is_valid (pthread_key_t key)
{
int res;
VALGRIND_MAGIC_SEQUENCE(res, 2 /* default */,
VG_USERREQ__PTHREAD_KEY_VALIDATE,
key, 0, 0, 0);
my_assert(res != 2);
return res;
}
/* Returns NULL if thread is invalid. Otherwise, if the thread
already has a specifics area, return that. Otherwise allocate it
one. */
static
void** get_or_allocate_specifics_ptr ( pthread_t thread )
{
int res, i;
void** specifics_ptr;
ensure_valgrind("get_or_allocate_specifics_ptr");
/* Returns zero if the thread has no specific_ptr. One if thread
is invalid. Otherwise, the specific_ptr value. This is
allocated with my_malloc and so is aligned and cannot be
confused with 1 or 3. */
VALGRIND_MAGIC_SEQUENCE(specifics_ptr, 3 /* default */,
VG_USERREQ__PTHREAD_GETSPECIFIC_PTR,
thread, 0, 0, 0);
my_assert(specifics_ptr != (void**)3);
if (specifics_ptr == (void**)1)
return NULL; /* invalid thread */
if (specifics_ptr != NULL)
return specifics_ptr; /* already has a specifics ptr. */
/* None yet ... allocate a new one. Should never fail. */
specifics_ptr = my_malloc( VG_N_THREAD_KEYS * sizeof(void*) );
my_assert(specifics_ptr != NULL);
VALGRIND_MAGIC_SEQUENCE(res, -1 /* default */,
VG_USERREQ__PTHREAD_SETSPECIFIC_PTR,
specifics_ptr, 0, 0, 0);
my_assert(res == 0);
/* POSIX sez: "Upon thread creation, the value NULL shall be
associated with all defined keys in the new thread." This
allocation is in effect a delayed allocation of the specific
data for a thread, at its first-use. Hence we initialise it
here. */
for (i = 0; i < VG_N_THREAD_KEYS; i++) {
specifics_ptr[i] = NULL;
}
return specifics_ptr;
}
int __pthread_key_create(pthread_key_t *key,
void (*destr_function) (void *))
{
void** specifics_ptr;
int res, i;
ensure_valgrind("pthread_key_create");
/* This writes *key if successful. It should never fail. */
VALGRIND_MAGIC_SEQUENCE(res, 1 /* default */,
VG_USERREQ__PTHREAD_KEY_CREATE,
key, destr_function, 0, 0);
my_assert(res == 0);
/* POSIX sez: "Upon key creation, the value NULL shall be
associated with the new key in all active threads." */
for (i = 0; i < VG_N_THREADS; i++) {
specifics_ptr = get_or_allocate_specifics_ptr(i);
/* we get NULL if i is an invalid thread. */
if (specifics_ptr != NULL)
specifics_ptr[*key] = NULL;
}
return res;
}
int pthread_key_delete(pthread_key_t key)
{
int res;
ensure_valgrind("pthread_key_create");
if (!key_is_valid(key))
return EINVAL;
VALGRIND_MAGIC_SEQUENCE(res, 0 /* default */,
VG_USERREQ__PTHREAD_KEY_DELETE,
key, 0, 0, 0);
my_assert(res == 0);
return 0;
}
int __pthread_setspecific(pthread_key_t key, const void *pointer)
{
void** specifics_ptr;
ensure_valgrind("pthread_setspecific");
if (!key_is_valid(key))
return EINVAL;
specifics_ptr = get_or_allocate_specifics_ptr(pthread_self());
specifics_ptr[key] = (void*)pointer;
return 0;
}
void * __pthread_getspecific(pthread_key_t key)
{
void** specifics_ptr;
ensure_valgrind("pthread_getspecific");
if (!key_is_valid(key))
return NULL;
specifics_ptr = get_or_allocate_specifics_ptr(pthread_self());
return specifics_ptr[key];
}
#ifdef GLIBC_2_3
static
void ** __pthread_getspecific_addr(pthread_key_t key)
{
void** specifics_ptr;
ensure_valgrind("pthread_getspecific_addr");
if (!key_is_valid(key))
return NULL;
specifics_ptr = get_or_allocate_specifics_ptr(pthread_self());
return &(specifics_ptr[key]);
}
#endif
/* ---------------------------------------------------
ONCEry
------------------------------------------------ */
static pthread_mutex_t once_masterlock = PTHREAD_MUTEX_INITIALIZER;
int __pthread_once ( pthread_once_t *once_control,
void (*init_routine) (void) )
{
int res;
ensure_valgrind("pthread_once");
res = __pthread_mutex_lock(&once_masterlock);
if (res != 0) {
barf("pthread_once: Looks like your program's "
"init routine calls back to pthread_once() ?!");
}
if (*once_control == 0) {
*once_control = 1;
init_routine();
}
__pthread_mutex_unlock(&once_masterlock);
return 0;
}
/* ---------------------------------------------------
MISC
------------------------------------------------ */
static pthread_mutex_t pthread_atfork_lock
= PTHREAD_MUTEX_INITIALIZER;
int __pthread_atfork ( void (*prepare)(void),
void (*parent)(void),
void (*child)(void) )
{
int n, res;
ForkHandlerEntry entry;
ensure_valgrind("pthread_atfork");
__pthread_mutex_lock(&pthread_atfork_lock);
/* Fetch old counter */
VALGRIND_MAGIC_SEQUENCE(n, -2 /* default */,
VG_USERREQ__GET_FHSTACK_USED,
0, 0, 0, 0);
my_assert(n >= 0 && n < VG_N_FORKHANDLERSTACK);
if (n == VG_N_FORKHANDLERSTACK-1)
barf("pthread_atfork: VG_N_FORKHANDLERSTACK is too low; "
"increase and recompile");
/* Add entry */
entry.prepare = *prepare;
entry.parent = *parent;
entry.child = *child;
VALGRIND_MAGIC_SEQUENCE(res, -2 /* default */,
VG_USERREQ__SET_FHSTACK_ENTRY,
n, &entry, 0, 0);
my_assert(res == 0);
/* Bump counter */
VALGRIND_MAGIC_SEQUENCE(res, -2 /* default */,
VG_USERREQ__SET_FHSTACK_USED,
n+1, 0, 0, 0);
my_assert(res == 0);
__pthread_mutex_unlock(&pthread_atfork_lock);
return 0;
}
__attribute__((weak))
void __pthread_initialize ( void )
{
ensure_valgrind("__pthread_initialize");
}
/* ---------------------------------------------------
LIBRARY-PRIVATE THREAD SPECIFIC STATE
------------------------------------------------ */
#include <resolv.h>
static int thread_specific_errno[VG_N_THREADS];
static int thread_specific_h_errno[VG_N_THREADS];
static struct __res_state
thread_specific_res_state[VG_N_THREADS];
int* __errno_location ( void )
{
int tid;
/* ensure_valgrind("__errno_location"); */
VALGRIND_MAGIC_SEQUENCE(tid, 1 /* default */,
VG_USERREQ__PTHREAD_GET_THREADID,
0, 0, 0, 0);
/* 'cos I'm paranoid ... */
if (tid < 1 || tid >= VG_N_THREADS)
barf("__errno_location: invalid ThreadId");
return & thread_specific_errno[tid];
}
int* __h_errno_location ( void )
{
int tid;
/* ensure_valgrind("__h_errno_location"); */
VALGRIND_MAGIC_SEQUENCE(tid, 1 /* default */,
VG_USERREQ__PTHREAD_GET_THREADID,
0, 0, 0, 0);
/* 'cos I'm paranoid ... */
if (tid < 1 || tid >= VG_N_THREADS)
barf("__h_errno_location: invalid ThreadId");
return & thread_specific_h_errno[tid];
}
#undef _res
extern struct __res_state _res;
struct __res_state* __res_state ( void )
{
int tid;
/* ensure_valgrind("__res_state"); */
VALGRIND_MAGIC_SEQUENCE(tid, 1 /* default */,
VG_USERREQ__PTHREAD_GET_THREADID,
0, 0, 0, 0);
/* 'cos I'm paranoid ... */
if (tid < 1 || tid >= VG_N_THREADS)
barf("__res_state: invalid ThreadId");
if (tid == 1)
return & _res;
return & thread_specific_res_state[tid];
}
/* ---------------------------------------------------
LIBC-PRIVATE SPECIFIC DATA
------------------------------------------------ */
/* Relies on assumption that initial private data is NULL. This
should be fixed somehow. */
/* The allowable keys (indices) (all 3 of them).
From sysdeps/pthread/bits/libc-tsd.h
*/
/* as per glibc anoncvs HEAD of 20021001. */
enum __libc_tsd_key_t { _LIBC_TSD_KEY_MALLOC = 0,
_LIBC_TSD_KEY_DL_ERROR,
_LIBC_TSD_KEY_RPC_VARS,
_LIBC_TSD_KEY_LOCALE,
_LIBC_TSD_KEY_CTYPE_B,
_LIBC_TSD_KEY_CTYPE_TOLOWER,
_LIBC_TSD_KEY_CTYPE_TOUPPER,
_LIBC_TSD_KEY_N };
/* Auto-initialising subsystem. libc_specifics_inited is set
after initialisation. libc_specifics_inited_mx guards it. */
static int libc_specifics_inited = 0;
static pthread_mutex_t libc_specifics_inited_mx = PTHREAD_MUTEX_INITIALIZER;
/* These are the keys we must initialise the first time. */
static pthread_key_t libc_specifics_keys[_LIBC_TSD_KEY_N];
/* Initialise the keys, if they are not already initialised. */
static
void init_libc_tsd_keys ( void )
{
int res, i;
pthread_key_t k;
/* Don't fall into deadlock if we get called again whilst we still
hold the lock, via the __uselocale() call herein. */
if (libc_specifics_inited != 0)
return;
/* Take the lock. */
res = __pthread_mutex_lock(&libc_specifics_inited_mx);
if (res != 0) barf("init_libc_tsd_keys: lock");
/* Now test again, to be sure there is no mistake. */
if (libc_specifics_inited != 0) {
res = __pthread_mutex_unlock(&libc_specifics_inited_mx);
if (res != 0) barf("init_libc_tsd_keys: unlock(1)");
return;
}
/* Actually do the initialisation. */
/* printf("INIT libc specifics\n"); */
for (i = 0; i < _LIBC_TSD_KEY_N; i++) {
res = __pthread_key_create(&k, NULL);
if (res != 0) barf("init_libc_tsd_keys: create");
libc_specifics_keys[i] = k;
}
/* Signify init done. */
libc_specifics_inited = 1;
# ifdef GLIBC_2_3
/* Set the initialising thread's locale to the global (default)
locale. A hack in support of glibc-2.3. This does the biz for
the root thread. For all other threads we run this in
thread_wrapper(), which does the real work of
pthread_create(). */
/* assert that we are the root thread. I don't know if this is
really a valid assertion to make; if it breaks I'll reconsider
it. */
my_assert(pthread_self() == 1);
__uselocale(LC_GLOBAL_LOCALE);
# endif
/* Unlock and return. */
res = __pthread_mutex_unlock(&libc_specifics_inited_mx);
if (res != 0) barf("init_libc_tsd_keys: unlock");
}
static int
libc_internal_tsd_set ( enum __libc_tsd_key_t key,
const void * pointer )
{
int res;
/* printf("SET SET SET key %d ptr %p\n", key, pointer); */
if (key < _LIBC_TSD_KEY_MALLOC || key >= _LIBC_TSD_KEY_N)
barf("libc_internal_tsd_set: invalid key");
init_libc_tsd_keys();
res = __pthread_setspecific(libc_specifics_keys[key], pointer);
if (res != 0) barf("libc_internal_tsd_set: setspecific failed");
return 0;
}
static void *
libc_internal_tsd_get ( enum __libc_tsd_key_t key )
{
void* v;
/* printf("GET GET GET key %d\n", key); */
if (key < _LIBC_TSD_KEY_MALLOC || key >= _LIBC_TSD_KEY_N)
barf("libc_internal_tsd_get: invalid key");
init_libc_tsd_keys();
v = __pthread_getspecific(libc_specifics_keys[key]);
/* if (v == NULL) barf("libc_internal_tsd_set: getspecific failed"); */
return v;
}
int (*__libc_internal_tsd_set)
(enum __libc_tsd_key_t key, const void * pointer)
= libc_internal_tsd_set;
void* (*__libc_internal_tsd_get)
(enum __libc_tsd_key_t key)
= libc_internal_tsd_get;
#ifdef GLIBC_2_3
/* This one was first spotted be me in the glibc-2.2.93 sources. */
static void**
libc_internal_tsd_address ( enum __libc_tsd_key_t key )
{
void** v;
/* printf("ADDR ADDR ADDR key %d\n", key); */
if (key < _LIBC_TSD_KEY_MALLOC || key >= _LIBC_TSD_KEY_N)
barf("libc_internal_tsd_address: invalid key");
init_libc_tsd_keys();
v = __pthread_getspecific_addr(libc_specifics_keys[key]);
return v;
}
void ** (*__libc_internal_tsd_address)
(enum __libc_tsd_key_t key)
= libc_internal_tsd_address;
#endif
/* ---------------------------------------------------------------------
These are here (I think) because they are deemed cancellation
points by POSIX. For the moment we'll simply pass the call along
to the corresponding thread-unaware (?) libc routine.
------------------------------------------------------------------ */
#include <stdlib.h>
#include <sys/types.h>
#include <sys/socket.h>
#ifdef GLIBC_2_1
extern
int __sigaction
(int signum,
const struct sigaction *act,
struct sigaction *oldact);
#else
extern
int __libc_sigaction
(int signum,
const struct sigaction *act,
struct sigaction *oldact);
#endif
int sigaction(int signum,
const struct sigaction *act,
struct sigaction *oldact)
{
__my_pthread_testcancel();
# ifdef GLIBC_2_1
return __sigaction(signum, act, oldact);
# else
return __libc_sigaction(signum, act, oldact);
# endif
}
extern
int __libc_connect(int sockfd,
const struct sockaddr *serv_addr,
socklen_t addrlen);
__attribute__((weak))
int connect(int sockfd,
const struct sockaddr *serv_addr,
socklen_t addrlen)
{
__my_pthread_testcancel();
return __libc_connect(sockfd, serv_addr, addrlen);
}
extern
int __libc_fcntl(int fd, int cmd, long arg);
__attribute__((weak))
int fcntl(int fd, int cmd, long arg)
{
__my_pthread_testcancel();
return __libc_fcntl(fd, cmd, arg);
}
extern
ssize_t __libc_write(int fd, const void *buf, size_t count);
__attribute__((weak))
ssize_t write(int fd, const void *buf, size_t count)
{
__my_pthread_testcancel();
return __libc_write(fd, buf, count);
}
extern
ssize_t __libc_read(int fd, void *buf, size_t count);
__attribute__((weak))
ssize_t read(int fd, void *buf, size_t count)
{
__my_pthread_testcancel();
return __libc_read(fd, buf, count);
}
/*
* Ugh, this is horrible but here goes:
*
* Open of a named pipe (fifo file) can block. In a threaded program,
* this means that the whole thing can block. We therefore need to
* make the open appear to block to the caller, but still keep polling
* for everyone else.
*
* There are four cases:
*
* - the caller asked for O_NONBLOCK. The easy one: we just do it.
*
* - the caller asked for a blocking O_RDONLY open. We open it with
* O_NONBLOCK and then use poll to wait for it to become ready.
*
* - the caller asked for a blocking O_WRONLY open. Unfortunately, this
* will fail with ENXIO when we make it non-blocking. Doubly
* unfortunate is that we can only rely on these semantics if it is
* actually a fifo file; the hack is that if we see that it is a
* O_WRONLY open and we get ENXIO, then stat the path and see if it
* actually is a fifo. This is racy, but it is the best we can do.
* If it is a fifo, then keep trying the open until it works; if not
* just return the error.
*
* - the caller asked for a blocking O_RDWR open. Well, under Linux,
* this never blocks, so we just clear the non-blocking flag and
* return.
*
* This code assumes that for whatever we open, O_NONBLOCK followed by
* a fcntl clearing O_NONBLOCK is the same as opening without
* O_NONBLOCK. Also assumes that stat and fstat have no side-effects.
*
* XXX Should probably put in special cases for some devices as well,
* like serial ports. Unfortunately they don't work like fifos, so
* this logic will become even more tortured. Wait until we really
* need it.
*/
static inline int _open(const char *pathname, int flags, mode_t mode,
int (*openp)(const char *, int, mode_t))
{
int fd;
struct stat st;
struct vki_timespec nanosleep_interval;
int saved_errno;
__my_pthread_testcancel();
/* Assume we can only get O_RDONLY, O_WRONLY or O_RDWR */
my_assert((flags & VKI_O_ACCMODE) != VKI_O_ACCMODE);
for(;;) {
fd = (*openp)(pathname, flags | VKI_O_NONBLOCK, mode);
/* return immediately if caller wanted nonblocking anyway */
if (flags & VKI_O_NONBLOCK)
return fd;
saved_errno = errno;
if (fd != -1)
break; /* open worked */
/* If we got ENXIO and we're opening WRONLY, and it turns out
to really be a FIFO, then poll waiting for open to succeed */
if (errno == ENXIO &&
(flags & VKI_O_ACCMODE) == VKI_O_WRONLY &&
(stat(pathname, &st) == 0 && S_ISFIFO(st.st_mode))) {
/* OK, we're opening a FIFO for writing; sleep and spin */
nanosleep_interval.tv_sec = 0;
nanosleep_interval.tv_nsec = 13 * 1000 * 1000; /* 13 milliseconds */
/* It's critical here that valgrind's nanosleep implementation
is nonblocking. */
(void)my_do_syscall2(__NR_nanosleep,
(int)(&nanosleep_interval), (int)NULL);
} else {
/* it was just an error */
errno = saved_errno;
return -1;
}
}
/* OK, we've got a nonblocking FD for a caller who wants blocking;
reset the flags to what they asked for */
fcntl(fd, VKI_F_SETFL, flags);
/* Return now if one of:
- we were opening O_RDWR (never blocks)
- we opened with O_WRONLY (polling already done)
- the thing we opened wasn't a FIFO after all (or fstat failed)
*/
if ((flags & VKI_O_ACCMODE) != VKI_O_RDONLY ||
(fstat(fd, &st) == -1 || !S_ISFIFO(st.st_mode))) {
errno = saved_errno;
return fd;
}
/* OK, drop into the poll loop looking for something to read on the fd */
my_assert((flags & VKI_O_ACCMODE) == VKI_O_RDONLY);
for(;;) {
struct pollfd pollfd;
int res;
pollfd.fd = fd;
pollfd.events = POLLIN;
pollfd.revents = 0;
res = my_do_syscall3(__NR_poll, (int)&pollfd, 1, 0);
my_assert(res == 0 || res == 1);
if (res == 1) {
/* OK, got it.
XXX This is wrong: we're waiting for either something to
read or a HUP on the file descriptor, but the semantics of
fifo open are that we should unblock as soon as someone
simply opens the other end, not that they write something.
With luck this won't matter in practice.
*/
my_assert(pollfd.revents & (POLLIN|POLLHUP));
break;
}
/* Still nobody home; sleep and spin */
nanosleep_interval.tv_sec = 0;
nanosleep_interval.tv_nsec = 13 * 1000 * 1000; /* 13 milliseconds */
/* It's critical here that valgrind's nanosleep implementation
is nonblocking. */
(void)my_do_syscall2(__NR_nanosleep,
(int)(&nanosleep_interval), (int)NULL);
}
errno = saved_errno;
return fd;
}
extern
int __libc_open64(const char *pathname, int flags, mode_t mode);
/* __attribute__((weak)) */
int open64(const char *pathname, int flags, mode_t mode)
{
return _open(pathname, flags, mode, __libc_open64);
}
extern
int __libc_open(const char *pathname, int flags, mode_t mode);
/* __attribute__((weak)) */
int open(const char *pathname, int flags, mode_t mode)
{
return _open(pathname, flags, mode, __libc_open);
}
extern
int __libc_close(int fd);
__attribute__((weak))
int close(int fd)
{
__my_pthread_testcancel();
return __libc_close(fd);
}
extern
int __libc_accept(int s, struct sockaddr *addr, socklen_t *addrlen);
int VGL_(accept)(int s, struct sockaddr *addr, socklen_t *addrlen)
{
__my_pthread_testcancel();
wait_for_fd_to_be_readable_or_erring(s);
__my_pthread_testcancel();
return __libc_accept(s, addr, addrlen);
}
extern
int __libc_recv(int s, void *buf, size_t len, int flags);
int VGL_(recv)(int s, void *buf, size_t len, int flags)
{
__my_pthread_testcancel();
wait_for_fd_to_be_readable_or_erring(s);
__my_pthread_testcancel();
return __libc_recv(s, buf, len, flags);
}
int VGL_(readv)(int fd, const struct iovec *iov, int count)
{
int res;
__my_pthread_testcancel();
wait_for_fd_to_be_readable_or_erring(fd);
__my_pthread_testcancel();
res = my_do_syscall3(__NR_readv, fd, (unsigned)iov, count);
if (is_kerror(res)) {
*(__errno_location()) = -res;
return -1;
}
return res;
}
int VGL_(writev)(int fd, struct iovec *iov, int count)
{
int res;
__my_pthread_testcancel();
wait_for_fd_to_be_writable_or_erring(fd);
__my_pthread_testcancel();
res = my_do_syscall3(__NR_writev, fd, (unsigned)iov, count);
if (is_kerror(res)) {
*(__errno_location()) = -res;
return -1;
}
return res;
}
extern
pid_t __libc_waitpid(pid_t pid, int *status, int options);
__attribute__((weak))
pid_t waitpid(pid_t pid, int *status, int options)
{
__my_pthread_testcancel();
return __libc_waitpid(pid, status, options);
}
extern
int __libc_nanosleep(const struct timespec *req, struct timespec *rem);
__attribute__((weak))
int nanosleep(const struct timespec *req, struct timespec *rem)
{
__my_pthread_testcancel();
return __libc_nanosleep(req, rem);
}
extern
int __libc_fsync(int fd);
__attribute__((weak))
int fsync(int fd)
{
__my_pthread_testcancel();
return __libc_fsync(fd);
}
extern
off_t __libc_lseek(int fildes, off_t offset, int whence);
__attribute__((weak))
off_t lseek(int fildes, off_t offset, int whence)
{
__my_pthread_testcancel();
return __libc_lseek(fildes, offset, whence);
}
extern
__off64_t __libc_lseek64(int fildes, __off64_t offset, int whence);
__attribute__((weak))
__off64_t lseek64(int fildes, __off64_t offset, int whence)
{
__my_pthread_testcancel();
return __libc_lseek64(fildes, offset, whence);
}
extern
ssize_t __libc_pread64 (int __fd, void *__buf, size_t __nbytes,
__off64_t __offset);
ssize_t __pread64 (int __fd, void *__buf, size_t __nbytes,
__off64_t __offset)
{
__my_pthread_testcancel();
return __libc_pread64(__fd, __buf, __nbytes, __offset);
}
extern
ssize_t __libc_pwrite64 (int __fd, const void *__buf, size_t __nbytes,
__off64_t __offset);
ssize_t __pwrite64 (int __fd, const void *__buf, size_t __nbytes,
__off64_t __offset)
{
__my_pthread_testcancel();
return __libc_pwrite64(__fd, __buf, __nbytes, __offset);
}
extern
ssize_t __libc_pwrite(int fd, const void *buf, size_t count, off_t offset);
__attribute__((weak))
ssize_t pwrite(int fd, const void *buf, size_t count, off_t offset)
{
__my_pthread_testcancel();
return __libc_pwrite(fd, buf, count, offset);
}
extern
ssize_t __libc_pread(int fd, void *buf, size_t count, off_t offset);
__attribute__((weak))
ssize_t pread(int fd, void *buf, size_t count, off_t offset)
{
__my_pthread_testcancel();
return __libc_pread(fd, buf, count, offset);
}
extern
void __libc_longjmp(jmp_buf env, int val) __attribute((noreturn));
/* not weak: __attribute__((weak)) */
void longjmp(jmp_buf env, int val)
{
__libc_longjmp(env, val);
}
extern void __libc_siglongjmp (sigjmp_buf env, int val)
__attribute__ ((noreturn));
void siglongjmp(sigjmp_buf env, int val)
{
kludged("siglongjmp (cleanup handlers are ignored)");
__libc_siglongjmp(env, val);
}
extern
int __libc_send(int s, const void *msg, size_t len, int flags);
__attribute__((weak))
int send(int s, const void *msg, size_t len, int flags)
{
__my_pthread_testcancel();
return __libc_send(s, msg, len, flags);
}
extern
int __libc_recv(int s, void *buf, size_t len, int flags);
__attribute__((weak))
int recv(int s, void *buf, size_t len, int flags)
{
__my_pthread_testcancel();
wait_for_fd_to_be_readable_or_erring(s);
__my_pthread_testcancel();
return __libc_recv(s, buf, len, flags);
}
extern
int __libc_sendmsg(int s, const struct msghdr *msg, int flags);
__attribute__((weak))
int sendmsg(int s, const struct msghdr *msg, int flags)
{
__my_pthread_testcancel();
return __libc_sendmsg(s, msg, flags);
}
extern
int __libc_recvmsg(int s, struct msghdr *msg, int flags);
__attribute__((weak))
int recvmsg(int s, struct msghdr *msg, int flags)
{
__my_pthread_testcancel();
return __libc_recvmsg(s, msg, flags);
}
extern
int __libc_recvfrom(int s, void *buf, size_t len, int flags,
struct sockaddr *from, socklen_t *fromlen);
__attribute__((weak))
int recvfrom(int s, void *buf, size_t len, int flags,
struct sockaddr *from, socklen_t *fromlen)
{
__my_pthread_testcancel();
wait_for_fd_to_be_readable_or_erring(s);
__my_pthread_testcancel();
return __libc_recvfrom(s, buf, len, flags, from, fromlen);
}
extern
int __libc_sendto(int s, const void *msg, size_t len, int flags,
const struct sockaddr *to, socklen_t tolen);
__attribute__((weak))
int sendto(int s, const void *msg, size_t len, int flags,
const struct sockaddr *to, socklen_t tolen)
{
__my_pthread_testcancel();
return __libc_sendto(s, msg, len, flags, to, tolen);
}
extern
int __libc_system(const char* str);
__attribute__((weak))
int system(const char* str)
{
__my_pthread_testcancel();
return __libc_system(str);
}
extern
pid_t __libc_wait(int *status);
__attribute__((weak))
pid_t wait(int *status)
{
__my_pthread_testcancel();
return __libc_wait(status);
}
extern
int __libc_msync(const void *start, size_t length, int flags);
__attribute__((weak))
int msync(const void *start, size_t length, int flags)
{
__my_pthread_testcancel();
return __libc_msync(start, length, flags);
}
/*--- fork and its helper ---*/
static
void run_fork_handlers ( int what )
{
ForkHandlerEntry entry;
int n_h, n_handlers, i, res;
my_assert(what == 0 || what == 1 || what == 2);
/* Fetch old counter */
VALGRIND_MAGIC_SEQUENCE(n_handlers, -2 /* default */,
VG_USERREQ__GET_FHSTACK_USED,
0, 0, 0, 0);
my_assert(n_handlers >= 0 && n_handlers < VG_N_FORKHANDLERSTACK);
/* Prepare handlers (what == 0) are called in opposite order of
calls to pthread_atfork. Parent and child handlers are called
in the same order as calls to pthread_atfork. */
if (what == 0)
n_h = n_handlers - 1;
else
n_h = 0;
for (i = 0; i < n_handlers; i++) {
VALGRIND_MAGIC_SEQUENCE(res, -2 /* default */,
VG_USERREQ__GET_FHSTACK_ENTRY,
n_h, &entry, 0, 0);
my_assert(res == 0);
switch (what) {
case 0: if (entry.prepare) entry.prepare();
n_h--; break;
case 1: if (entry.parent) entry.parent();
n_h++; break;
case 2: if (entry.child) entry.child();
n_h++; break;
default: barf("run_fork_handlers: invalid what");
}
}
if (what != 0 /* prepare */) {
/* Empty out the stack. */
VALGRIND_MAGIC_SEQUENCE(res, -2 /* default */,
VG_USERREQ__SET_FHSTACK_USED,
0, 0, 0, 0);
my_assert(res == 0);
}
}
extern
pid_t __libc_fork(void);
pid_t __fork(void)
{
pid_t pid;
__my_pthread_testcancel();
__pthread_mutex_lock(&pthread_atfork_lock);
run_fork_handlers(0 /* prepare */);
pid = __libc_fork();
if (pid == 0) {
/* I am the child */
run_fork_handlers(2 /* child */);
__pthread_mutex_unlock(&pthread_atfork_lock);
__pthread_mutex_init(&pthread_atfork_lock, NULL);
} else {
/* I am the parent */
run_fork_handlers(1 /* parent */);
__pthread_mutex_unlock(&pthread_atfork_lock);
}
return pid;
}
pid_t __vfork(void)
{
return __fork();
}
/* ---------------------------------------------------------------------
Nonblocking implementations of select() and poll(). This stuff will
surely rot your mind.
------------------------------------------------------------------ */
/*--------------------------------------------------*/
#include "vg_kerneliface.h"
static
int my_do_syscall1 ( int syscallno, int arg1 )
{
int __res;
__asm__ volatile ("pushl %%ebx; movl %%edx,%%ebx ; int $0x80 ; popl %%ebx"
: "=a" (__res)
: "0" (syscallno),
"d" (arg1) );
return __res;
}
static
int my_do_syscall2 ( int syscallno,
int arg1, int arg2 )
{
int __res;
__asm__ volatile ("pushl %%ebx; movl %%edx,%%ebx ; int $0x80 ; popl %%ebx"
: "=a" (__res)
: "0" (syscallno),
"d" (arg1),
"c" (arg2) );
return __res;
}
static
int my_do_syscall3 ( int syscallno,
int arg1, int arg2, int arg3 )
{
int __res;
__asm__ volatile ("pushl %%ebx; movl %%esi,%%ebx ; int $0x80 ; popl %%ebx"
: "=a" (__res)
: "0" (syscallno),
"S" (arg1),
"c" (arg2),
"d" (arg3) );
return __res;
}
static inline
int my_do_syscall5 ( int syscallno,
int arg1, int arg2, int arg3, int arg4, int arg5 )
{
int __res;
__asm__ volatile ("int $0x80"
: "=a" (__res)
: "0" (syscallno),
"b" (arg1),
"c" (arg2),
"d" (arg3),
"S" (arg4),
"D" (arg5));
return __res;
}
static
int do_syscall_select( int n,
vki_fd_set* readfds,
vki_fd_set* writefds,
vki_fd_set* exceptfds,
struct vki_timeval * timeout )
{
int res;
int args[5];
args[0] = n;
args[1] = (int)readfds;
args[2] = (int)writefds;
args[3] = (int)exceptfds;
args[4] = (int)timeout;
res = my_do_syscall1(__NR_select, (int)(&(args[0])) );
return res;
}
/* This is a wrapper round select(), which makes it thread-safe,
meaning that only this thread will block, rather than the entire
process. This wrapper in turn depends on nanosleep() not to block
the entire process, but I think (hope? suspect?) that POSIX
pthreads guarantees that to be the case.
Basic idea is: modify the timeout parameter to select so that it
returns immediately. Poll like this until select returns non-zero,
indicating something interesting happened, or until our time is up.
Space out the polls with nanosleeps of say 11 milliseconds, which
is required to be nonblocking; this allows other threads to run.
Assumes:
* (checked via my_assert) types fd_set and vki_fd_set are identical.
* (checked via my_assert) types timeval and vki_timeval are identical.
* (unchecked) libc error numbers (EINTR etc) are the negation of the
kernel's error numbers (VKI_EINTR etc).
*/
int VGL_(select) ( int n,
fd_set *rfds,
fd_set *wfds,
fd_set *xfds,
struct timeval *timeout )
{
unsigned int ms_now, ms_end;
int res;
fd_set rfds_copy;
fd_set wfds_copy;
fd_set xfds_copy;
struct vki_timeval t_now;
struct vki_timeval zero_timeout;
struct vki_timespec nanosleep_interval;
__my_pthread_testcancel();
/* gcc's complains about ms_end being used uninitialised -- classic
case it can't understand, where ms_end is both defined and used
only if timeout != NULL. Hence ... */
ms_end = 0;
/* We assume that the kernel and libc data layouts are identical
for the following types. These asserts provide a crude
check. */
if (sizeof(fd_set) != sizeof(vki_fd_set)
|| sizeof(struct timeval) != sizeof(struct vki_timeval))
barf("valgrind's hacky non-blocking select(): data sizes error");
/* Detect the current time and simultaneously find out if we are
running on Valgrind. */
VALGRIND_MAGIC_SEQUENCE(ms_now, 0xFFFFFFFF /* default */,
VG_USERREQ__READ_MILLISECOND_TIMER,
0, 0, 0, 0);
/* If a zero timeout specified, this call is harmless. Also go
this route if we're not running on Valgrind, for whatever
reason. */
if ( (timeout && timeout->tv_sec == 0 && timeout->tv_usec == 0)
|| (ms_now == 0xFFFFFFFF) ) {
res = do_syscall_select( n, (vki_fd_set*)rfds,
(vki_fd_set*)wfds,
(vki_fd_set*)xfds,
(struct vki_timeval*)timeout);
if (is_kerror(res)) {
* (__errno_location()) = -res;
return -1;
} else {
return res;
}
}
/* If a timeout was specified, set ms_end to be the end millisecond
counter [wallclock] time. */
if (timeout) {
res = my_do_syscall2(__NR_gettimeofday, (int)&t_now, (int)NULL);
my_assert(res == 0);
ms_end = ms_now;
ms_end += (timeout->tv_usec / 1000);
ms_end += (timeout->tv_sec * 1000);
/* Stay sane ... */
my_assert (ms_end >= ms_now);
}
/* fprintf(stderr, "MY_SELECT: before loop\n"); */
/* Either timeout == NULL, meaning wait indefinitely, or timeout !=
NULL, in which case ms_end holds the end time. */
while (1) {
/* First, do a return-immediately select(). */
/* These could be trashed each time round the loop, so restore
them each time. */
if (rfds) rfds_copy = *rfds;
if (wfds) wfds_copy = *wfds;
if (xfds) xfds_copy = *xfds;
zero_timeout.tv_sec = zero_timeout.tv_usec = 0;
res = do_syscall_select( n,
rfds ? (vki_fd_set*)(&rfds_copy) : NULL,
wfds ? (vki_fd_set*)(&wfds_copy) : NULL,
xfds ? (vki_fd_set*)(&xfds_copy) : NULL,
& zero_timeout );
if (is_kerror(res)) {
/* Some kind of error (including EINTR). Set errno and
return. The sets are unspecified in this case. */
* (__errno_location()) = -res;
return -1;
}
if (res > 0) {
/* one or more fds is ready. Copy out resulting sets and
return. */
if (rfds) *rfds = rfds_copy;
if (wfds) *wfds = wfds_copy;
if (xfds) *xfds = xfds_copy;
return res;
}
/* Nothing interesting happened, so we go to sleep for a
while. */
/* fprintf(stderr, "MY_SELECT: nanosleep\n"); */
/* nanosleep and go round again */
nanosleep_interval.tv_sec = 0;
nanosleep_interval.tv_nsec = 11 * 1000 * 1000; /* 11 milliseconds */
/* It's critical here that valgrind's nanosleep implementation
is nonblocking. */
res = my_do_syscall2(__NR_nanosleep,
(int)(&nanosleep_interval), (int)NULL);
if (res == -VKI_EINTR) {
/* The nanosleep was interrupted by a signal. So we do the
same. */
* (__errno_location()) = EINTR;
return -1;
}
/* Sleeping finished. If a finite timeout, check to see if it
has expired yet. */
if (timeout) {
VALGRIND_MAGIC_SEQUENCE(ms_now, 0xFFFFFFFF /* default */,
VG_USERREQ__READ_MILLISECOND_TIMER,
0, 0, 0, 0);
my_assert(ms_now != 0xFFFFFFFF);
if (ms_now >= ms_end) {
/* timeout; nothing interesting happened. */
if (rfds) FD_ZERO(rfds);
if (wfds) FD_ZERO(wfds);
if (xfds) FD_ZERO(xfds);
return 0;
}
}
}
}
#include <sys/poll.h>
#ifndef HAVE_NFDS_T
typedef unsigned long int nfds_t;
#endif
int VGL_(poll) (struct pollfd *__fds, nfds_t __nfds, int __timeout)
{
unsigned int ms_now, ms_end;
int res, i;
struct vki_timespec nanosleep_interval;
__my_pthread_testcancel();
ensure_valgrind("poll");
/* Detect the current time and simultaneously find out if we are
running on Valgrind. */
VALGRIND_MAGIC_SEQUENCE(ms_now, 0xFFFFFFFF /* default */,
VG_USERREQ__READ_MILLISECOND_TIMER,
0, 0, 0, 0);
if (/* CHECK SIZES FOR struct pollfd */
sizeof(struct timeval) != sizeof(struct vki_timeval))
barf("valgrind's hacky non-blocking poll(): data sizes error");
/* dummy initialisation to keep gcc -Wall happy */
ms_end = 0;
/* If a zero timeout specified, this call is harmless. Also do
this if not running on Valgrind. */
if (__timeout == 0 || ms_now == 0xFFFFFFFF) {
res = my_do_syscall3(__NR_poll, (int)__fds, __nfds, __timeout);
if (is_kerror(res)) {
* (__errno_location()) = -res;
return -1;
} else {
return res;
}
}
/* If a timeout was specified, set ms_end to be the end wallclock
time. Easy considering that __timeout is in milliseconds. */
if (__timeout > 0) {
ms_end = ms_now + (unsigned int)__timeout;
}
/* fprintf(stderr, "MY_POLL: before loop\n"); */
/* Either timeout < 0, meaning wait indefinitely, or timeout > 0,
in which case t_end holds the end time. */
my_assert(__timeout != 0);
while (1) {
/* Do a return-immediately poll. */
res = my_do_syscall3(__NR_poll, (int)__fds, __nfds, 0 );
if (is_kerror(res)) {
/* Some kind of error. Set errno and return. */
* (__errno_location()) = -res;
return -1;
}
if (res > 0) {
/* One or more fds is ready. Return now. */
return res;
}
/* Nothing interesting happened, so we go to sleep for a
while. */
/* fprintf(stderr, "MY_POLL: nanosleep\n"); */
/* nanosleep and go round again */
nanosleep_interval.tv_sec = 0;
nanosleep_interval.tv_nsec = 13 * 1000 * 1000; /* 13 milliseconds */
/* It's critical here that valgrind's nanosleep implementation
is nonblocking. */
res = my_do_syscall2(__NR_nanosleep,
(int)(&nanosleep_interval), (int)NULL);
if (res == -VKI_EINTR) {
/* The nanosleep was interrupted by a signal. So we do the
same. */
* (__errno_location()) = EINTR;
return -1;
}
/* Sleeping finished. If a finite timeout, check to see if it
has expired yet. */
if (__timeout > 0) {
VALGRIND_MAGIC_SEQUENCE(ms_now, 0xFFFFFFFF /* default */,
VG_USERREQ__READ_MILLISECOND_TIMER,
0, 0, 0, 0);
my_assert(ms_now != 0xFFFFFFFF);
if (ms_now >= ms_end) {
/* timeout; nothing interesting happened. */
for (i = 0; i < __nfds; i++)
__fds[i].revents = 0;
return 0;
}
}
}
}
/* Helper function used to make accept() non-blocking. Idea is to use
the above nonblocking poll() to make this thread ONLY wait for the
specified fd to become ready, and then return. */
/* Sigh -- a hack. We're not supposed to include this file directly;
should do it via /usr/include/fcntl.h, but that introduces a
varargs prototype for fcntl itself, which we can't mimic. */
#define _FCNTL_H
#include <bits/fcntl.h>
static void wait_for_fd_to_be_readable_or_erring ( int fd )
{
struct pollfd pfd;
int res;
/* fprintf(stderr, "wait_for_fd_to_be_readable_or_erring %d\n", fd); */
/* First check to see if the fd is nonblocking, and/or invalid. In
either case return immediately. */
res = __libc_fcntl(fd, F_GETFL, 0);
if (res == -1) return; /* fd is invalid somehow */
if (res & O_NONBLOCK) return; /* fd is nonblocking */
/* Ok, we'd better wait with poll. */
pfd.fd = fd;
pfd.events = POLLIN | POLLPRI | POLLERR | POLLHUP | POLLNVAL;
/* ... but not POLLOUT, you may notice. */
pfd.revents = 0;
(void)poll(&pfd, 1, -1 /* forever */);
}
static void wait_for_fd_to_be_writable_or_erring ( int fd )
{
struct pollfd pfd;
int res;
/* fprintf(stderr, "wait_for_fd_to_be_readable_or_erring %d\n", fd); */
/* First check to see if the fd is nonblocking, and/or invalid. In
either case return immediately. */
res = __libc_fcntl(fd, F_GETFL, 0);
if (res == -1) return; /* fd is invalid somehow */
if (res & O_NONBLOCK) return; /* fd is nonblocking */
/* Ok, we'd better wait with poll. */
pfd.fd = fd;
pfd.events = POLLOUT | POLLERR | POLLHUP | POLLNVAL;
pfd.revents = 0;
(void)poll(&pfd, 1, -1 /* forever */);
}
/* ---------------------------------------------------------------------
Hacky implementation of semaphores.
------------------------------------------------------------------ */
#include <semaphore.h>
/* This is a terrible way to do the remapping. Plan is to import an
AVL tree at some point. */
typedef
struct {
pthread_mutex_t se_mx;
pthread_cond_t se_cv;
int count;
}
vg_sem_t;
static pthread_mutex_t se_remap_mx = PTHREAD_MUTEX_INITIALIZER;
static int se_remap_used = 0;
static sem_t* se_remap_orig[VG_N_SEMAPHORES];
static vg_sem_t se_remap_new[VG_N_SEMAPHORES];
static vg_sem_t* se_remap ( sem_t* orig )
{
int res, i;
res = __pthread_mutex_lock(&se_remap_mx);
my_assert(res == 0);
for (i = 0; i < se_remap_used; i++) {
if (se_remap_orig[i] == orig)
break;
}
if (i == se_remap_used) {
if (se_remap_used == VG_N_SEMAPHORES) {
res = pthread_mutex_unlock(&se_remap_mx);
my_assert(res == 0);
barf("VG_N_SEMAPHORES is too low. Increase and recompile.");
}
se_remap_used++;
se_remap_orig[i] = orig;
/* printf("allocated semaphore %d\n", i); */
}
res = __pthread_mutex_unlock(&se_remap_mx);
my_assert(res == 0);
return &se_remap_new[i];
}
int sem_init(sem_t *sem, int pshared, unsigned int value)
{
int res;
vg_sem_t* vg_sem;
ensure_valgrind("sem_init");
if (pshared != 0) {
pthread_error("sem_init: unsupported pshared value");
errno = ENOSYS;
return -1;
}
vg_sem = se_remap(sem);
res = pthread_mutex_init(&vg_sem->se_mx, NULL);
my_assert(res == 0);
res = pthread_cond_init(&vg_sem->se_cv, NULL);
my_assert(res == 0);
vg_sem->count = value;
return 0;
}
int sem_wait ( sem_t* sem )
{
int res;
vg_sem_t* vg_sem;
ensure_valgrind("sem_wait");
vg_sem = se_remap(sem);
res = __pthread_mutex_lock(&vg_sem->se_mx);
my_assert(res == 0);
while (vg_sem->count == 0) {
res = pthread_cond_wait(&vg_sem->se_cv, &vg_sem->se_mx);
my_assert(res == 0);
}
vg_sem->count--;
res = __pthread_mutex_unlock(&vg_sem->se_mx);
my_assert(res == 0);
return 0;
}
int sem_post ( sem_t* sem )
{
int res;
vg_sem_t* vg_sem;
ensure_valgrind("sem_post");
vg_sem = se_remap(sem);
res = __pthread_mutex_lock(&vg_sem->se_mx);
my_assert(res == 0);
if (vg_sem->count == 0) {
vg_sem->count++;
res = pthread_cond_broadcast(&vg_sem->se_cv);
my_assert(res == 0);
} else {
vg_sem->count++;
}
res = __pthread_mutex_unlock(&vg_sem->se_mx);
my_assert(res == 0);
return 0;
}
int sem_trywait ( sem_t* sem )
{
int ret, res;
vg_sem_t* vg_sem;
ensure_valgrind("sem_trywait");
vg_sem = se_remap(sem);
res = __pthread_mutex_lock(&vg_sem->se_mx);
my_assert(res == 0);
if (vg_sem->count > 0) {
vg_sem->count--;
ret = 0;
} else {
ret = -1;
errno = EAGAIN;
}
res = __pthread_mutex_unlock(&vg_sem->se_mx);
my_assert(res == 0);
return ret;
}
int sem_getvalue(sem_t* sem, int * sval)
{
vg_sem_t* vg_sem;
ensure_valgrind("sem_trywait");
vg_sem = se_remap(sem);
*sval = vg_sem->count;
return 0;
}
int sem_destroy(sem_t * sem)
{
kludged("sem_destroy");
/* if someone waiting on this semaphore, errno = EBUSY, return -1 */
return 0;
}
int sem_timedwait(sem_t* sem, const struct timespec *abstime)
{
int res;
vg_sem_t* vg_sem;
ensure_valgrind("sem_timedwait");
vg_sem = se_remap(sem);
res = __pthread_mutex_lock(&vg_sem->se_mx);
my_assert(res == 0);
while ( vg_sem->count == 0 && res != ETIMEDOUT ) {
res = pthread_cond_timedwait(&vg_sem->se_cv, &vg_sem->se_mx, abstime);
}
if ( vg_sem->count > 0 ) {
vg_sem->count--;
res = __pthread_mutex_unlock(&vg_sem->se_mx);
my_assert(res == 0 );
return 0;
} else {
res = __pthread_mutex_unlock(&vg_sem->se_mx);
my_assert(res == 0 );
*(__errno_location()) = ETIMEDOUT;
return -1;
}
}
/* ---------------------------------------------------------------------
Reader-writer locks.
------------------------------------------------------------------ */
typedef
struct {
int initted; /* != 0 --> in use; sanity check only */
int prefer_w; /* != 0 --> prefer writer */
int nwait_r; /* # of waiting readers */
int nwait_w; /* # of waiting writers */
pthread_cond_t cv_r; /* for signalling readers */
pthread_cond_t cv_w; /* for signalling writers */
pthread_mutex_t mx;
int status;
/* allowed range for status: >= -1. -1 means 1 writer currently
active, >= 0 means N readers currently active. */
}
vg_rwlock_t;
static pthread_mutex_t rw_remap_mx = PTHREAD_MUTEX_INITIALIZER;
static int rw_remap_used = 0;
static pthread_rwlock_t* rw_remap_orig[VG_N_RWLOCKS];
static vg_rwlock_t rw_remap_new[VG_N_RWLOCKS];
static
void init_vg_rwlock ( vg_rwlock_t* vg_rwl )
{
int res = 0;
vg_rwl->initted = 1;
vg_rwl->prefer_w = 1;
vg_rwl->nwait_r = 0;
vg_rwl->nwait_w = 0;
vg_rwl->status = 0;
res = pthread_mutex_init(&vg_rwl->mx, NULL);
res |= pthread_cond_init(&vg_rwl->cv_r, NULL);
res |= pthread_cond_init(&vg_rwl->cv_w, NULL);
my_assert(res == 0);
}
/* Take the address of a LinuxThreads rwlock_t and return the shadow
address of our version. Further, if the LinuxThreads version
appears to have been statically initialised, do the same to the one
we allocate here. The pthread_rwlock_t.__rw_readers field is set
to zero by PTHREAD_RWLOCK_INITIALIZER, so we take zero as meaning
uninitialised and non-zero meaning initialised.
*/
static vg_rwlock_t* rw_remap ( pthread_rwlock_t* orig )
{
int res, i;
vg_rwlock_t* vg_rwl;
res = __pthread_mutex_lock(&rw_remap_mx);
my_assert(res == 0);
for (i = 0; i < rw_remap_used; i++) {
if (rw_remap_orig[i] == orig)
break;
}
if (i == rw_remap_used) {
if (rw_remap_used == VG_N_RWLOCKS) {
res = __pthread_mutex_unlock(&rw_remap_mx);
my_assert(res == 0);
barf("VG_N_RWLOCKS is too low. Increase and recompile.");
}
rw_remap_used++;
rw_remap_orig[i] = orig;
rw_remap_new[i].initted = 0;
if (0) printf("allocated rwlock %d\n", i);
}
res = __pthread_mutex_unlock(&rw_remap_mx);
my_assert(res == 0);
vg_rwl = &rw_remap_new[i];
/* Initialise the shadow, if required. */
if (orig->__rw_readers == 0) {
orig->__rw_readers = 1;
init_vg_rwlock(vg_rwl);
if (orig->__rw_kind == PTHREAD_RWLOCK_PREFER_READER_NP)
vg_rwl->prefer_w = 0;
}
return vg_rwl;
}
int pthread_rwlock_init ( pthread_rwlock_t* orig,
const pthread_rwlockattr_t* attr )
{
vg_rwlock_t* rwl;
if (0) printf ("pthread_rwlock_init\n");
/* Force the remapper to initialise the shadow. */
orig->__rw_readers = 0;
/* Install the lock preference; the remapper needs to know it. */
orig->__rw_kind = PTHREAD_RWLOCK_DEFAULT_NP;
if (attr)
orig->__rw_kind = attr->__lockkind;
rwl = rw_remap ( orig );
return 0;
}
static
void pthread_rwlock_rdlock_CANCEL_HDLR ( void* rwl_v )
{
vg_rwlock_t* rwl = (vg_rwlock_t*)rwl_v;
rwl->nwait_r--;
pthread_mutex_unlock (&rwl->mx);
}
int pthread_rwlock_rdlock ( pthread_rwlock_t* orig )
{
int res;
vg_rwlock_t* rwl;
if (0) printf ("pthread_rwlock_rdlock\n");
rwl = rw_remap ( orig );
res = __pthread_mutex_lock(&rwl->mx);
my_assert(res == 0);
if (!rwl->initted) {
res = __pthread_mutex_unlock(&rwl->mx);
my_assert(res == 0);
return EINVAL;
}
if (rwl->status < 0) {
my_assert(rwl->status == -1);
rwl->nwait_r++;
pthread_cleanup_push( pthread_rwlock_rdlock_CANCEL_HDLR, rwl );
while (1) {
if (rwl->status == 0) break;
res = pthread_cond_wait(&rwl->cv_r, &rwl->mx);
my_assert(res == 0);
}
pthread_cleanup_pop(0);
rwl->nwait_r--;
}
my_assert(rwl->status >= 0);
rwl->status++;
res = __pthread_mutex_unlock(&rwl->mx);
my_assert(res == 0);
return 0;
}
int pthread_rwlock_tryrdlock ( pthread_rwlock_t* orig )
{
int res;
vg_rwlock_t* rwl;
if (0) printf ("pthread_rwlock_tryrdlock\n");
rwl = rw_remap ( orig );
res = __pthread_mutex_lock(&rwl->mx);
my_assert(res == 0);
if (!rwl->initted) {
res = __pthread_mutex_unlock(&rwl->mx);
my_assert(res == 0);
return EINVAL;
}
if (rwl->status == -1) {
/* Writer active; we have to give up. */
res = __pthread_mutex_unlock(&rwl->mx);
my_assert(res == 0);
return EBUSY;
}
/* Success */
my_assert(rwl->status >= 0);
rwl->status++;
res = __pthread_mutex_unlock(&rwl->mx);
my_assert(res == 0);
return 0;
}
static
void pthread_rwlock_wrlock_CANCEL_HDLR ( void* rwl_v )
{
vg_rwlock_t* rwl = (vg_rwlock_t*)rwl_v;
rwl->nwait_w--;
pthread_mutex_unlock (&rwl->mx);
}
int pthread_rwlock_wrlock ( pthread_rwlock_t* orig )
{
int res;
vg_rwlock_t* rwl;
if (0) printf ("pthread_rwlock_wrlock\n");
rwl = rw_remap ( orig );
res = __pthread_mutex_lock(&rwl->mx);
my_assert(res == 0);
if (!rwl->initted) {
res = __pthread_mutex_unlock(&rwl->mx);
my_assert(res == 0);
return EINVAL;
}
if (rwl->status != 0) {
rwl->nwait_w++;
pthread_cleanup_push( pthread_rwlock_wrlock_CANCEL_HDLR, rwl );
while (1) {
if (rwl->status == 0) break;
res = pthread_cond_wait(&rwl->cv_w, &rwl->mx);
my_assert(res == 0);
}
pthread_cleanup_pop(0);
rwl->nwait_w--;
}
my_assert(rwl->status == 0);
rwl->status = -1;
res = __pthread_mutex_unlock(&rwl->mx);
my_assert(res == 0);
return 0;
}
int pthread_rwlock_trywrlock ( pthread_rwlock_t* orig )
{
int res;
vg_rwlock_t* rwl;
if (0) printf ("pthread_wrlock_trywrlock\n");
rwl = rw_remap ( orig );
res = __pthread_mutex_lock(&rwl->mx);
my_assert(res == 0);
if (!rwl->initted) {
res = __pthread_mutex_unlock(&rwl->mx);
my_assert(res == 0);
return EINVAL;
}
if (rwl->status != 0) {
/* Reader(s) or a writer active; we have to give up. */
res = __pthread_mutex_unlock(&rwl->mx);
my_assert(res == 0);
return EBUSY;
}
/* Success */
my_assert(rwl->status == 0);
rwl->status = -1;
res = __pthread_mutex_unlock(&rwl->mx);
my_assert(res == 0);
return 0;
}
int pthread_rwlock_unlock ( pthread_rwlock_t* orig )
{
int res;
vg_rwlock_t* rwl;
if (0) printf ("pthread_rwlock_unlock\n");
rwl = rw_remap ( orig );
rwl = rw_remap ( orig );
res = __pthread_mutex_lock(&rwl->mx);
my_assert(res == 0);
if (!rwl->initted) {
res = __pthread_mutex_unlock(&rwl->mx);
my_assert(res == 0);
return EINVAL;
}
if (rwl->status == 0) {
res = __pthread_mutex_unlock(&rwl->mx);
my_assert(res == 0);
return EPERM;
}
my_assert(rwl->status != 0);
if (rwl->status == -1) {
rwl->status = 0;
} else {
my_assert(rwl->status > 0);
rwl->status--;
}
my_assert(rwl->status >= 0);
if (rwl->prefer_w) {
/* Favour waiting writers, if any. */
if (rwl->nwait_w > 0) {
/* Writer(s) are waiting. */
if (rwl->status == 0) {
/* We can let a writer in. */
res = pthread_cond_signal(&rwl->cv_w);
my_assert(res == 0);
} else {
/* There are still readers active. Do nothing; eventually
they will disappear, at which point a writer will be
admitted. */
}
}
else
/* No waiting writers. */
if (rwl->nwait_r > 0) {
/* Let in a waiting reader. */
res = pthread_cond_signal(&rwl->cv_r);
my_assert(res == 0);
}
} else {
/* Favour waiting readers, if any. */
if (rwl->nwait_r > 0) {
/* Reader(s) are waiting; let one in. */
res = pthread_cond_signal(&rwl->cv_r);
my_assert(res == 0);
}
else
/* No waiting readers. */
if (rwl->nwait_w > 0 && rwl->status == 0) {
/* We have waiting writers and no active readers; let a
writer in. */
res = pthread_cond_signal(&rwl->cv_w);
my_assert(res == 0);
}
}
res = __pthread_mutex_unlock(&rwl->mx);
my_assert(res == 0);
return 0;
}
int pthread_rwlock_destroy ( pthread_rwlock_t *orig )
{
int res;
vg_rwlock_t* rwl;
if (0) printf ("pthread_rwlock_destroy\n");
rwl = rw_remap ( orig );
res = __pthread_mutex_lock(&rwl->mx);
my_assert(res == 0);
if (!rwl->initted) {
res = __pthread_mutex_unlock(&rwl->mx);
my_assert(res == 0);
return EINVAL;
}
if (rwl->status != 0 || rwl->nwait_r > 0 || rwl->nwait_w > 0) {
res = __pthread_mutex_unlock(&rwl->mx);
my_assert(res == 0);
return EBUSY;
}
rwl->initted = 0;
res = __pthread_mutex_unlock(&rwl->mx);
my_assert(res == 0);
return 0;
}
/* Copied directly from LinuxThreads. */
int
pthread_rwlockattr_init (pthread_rwlockattr_t *attr)
{
attr->__lockkind = 0;
attr->__pshared = PTHREAD_PROCESS_PRIVATE;
return 0;
}
/* Copied directly from LinuxThreads. */
int
pthread_rwlockattr_destroy (pthread_rwlockattr_t *attr)
{
return 0;
}
/* Copied directly from LinuxThreads. */
int
pthread_rwlockattr_setpshared (pthread_rwlockattr_t *attr, int pshared)
{
if (pshared != PTHREAD_PROCESS_PRIVATE && pshared != PTHREAD_PROCESS_SHARED)
return EINVAL;
/* For now it is not possible to shared a conditional variable. */
if (pshared != PTHREAD_PROCESS_PRIVATE)
return ENOSYS;
attr->__pshared = pshared;
return 0;
}
/* ---------------------------------------------------------------------
Make SYSV IPC not block everything
------------------------------------------------------------------ */
#include <sys/ipc.h>
#include <sys/msg.h>
#include <asm/ipc.h> /* for ipc_kludge */
static inline int sys_ipc(unsigned call, int first, int second, int third, void *ptr)
{
return my_do_syscall5(__NR_ipc, call, first, second, third, (int)ptr);
}
/* Turn a blocking msgsnd() into a polling non-blocking one, so that
other threads make progress */
int VGL_(msgsnd)(int msgid, const void *msgp, size_t msgsz, int msgflg)
{
struct vki_timespec nanosleep_interval;
int err;
ensure_valgrind("msgsnd");
nanosleep_interval.tv_sec = 0;
nanosleep_interval.tv_nsec = 13 * 1000 * 1000; /* 13 milliseconds */
if (msgflg & IPC_NOWAIT) {
/* If we aren't blocking anyway, just do it */
err = sys_ipc(11, msgid, msgsz, msgflg, (void *)msgp);
} else {
/* Otherwise poll on the queue to let other things run */
for(;;) {
err = sys_ipc(11, msgid, msgsz, msgflg | IPC_NOWAIT, (void *)msgp);
if (err != -EAGAIN)
break;
(void)my_do_syscall2(__NR_nanosleep,
(int)(&nanosleep_interval), (int)NULL);
}
}
if (is_kerror(err)) {
*(__errno_location()) = -err;
return -1;
}
return 0;
}
/* Turn a blocking msgrcv() into a polling non-blocking one, so that
other threads make progress */
int VGL_(msgrcv)( int msqid, void *msgp, size_t msgsz, long msgtyp, int msgflg )
{
struct vki_timespec nanosleep_interval;
int err;
struct ipc_kludge tmp;
ensure_valgrind("msgrcv");
nanosleep_interval.tv_sec = 0;
nanosleep_interval.tv_nsec = 13 * 1000 * 1000; /* 13 milliseconds */
tmp.msgp = msgp;
tmp.msgtyp = msgtyp;
if (msgflg & IPC_NOWAIT) {
/* If we aren't blocking anyway, just do it */
err = sys_ipc(12, msqid, msgsz, msgflg, &tmp );
} else {
/* Otherwise poll on the queue to let other things run */
for(;;) {
err = sys_ipc(12, msqid, msgsz, msgflg | IPC_NOWAIT, &tmp );
if (err != -ENOMSG)
break;
(void)my_do_syscall2(__NR_nanosleep,
(int)(&nanosleep_interval), (int)NULL);
}
}
if (is_kerror(err)) {
*(__errno_location()) = -err;
return -1;
}
return err;
}
/* ---------------------------------------------------------------------
B'stard.
------------------------------------------------------------------ */
# define strong_alias(name, aliasname) \
extern __typeof (name) aliasname __attribute__ ((alias (#name)));
# define weak_alias(name, aliasname) \
extern __typeof (name) aliasname __attribute__ ((weak, alias (#name)));
strong_alias(__pthread_mutex_lock, pthread_mutex_lock)
strong_alias(__pthread_mutex_trylock, pthread_mutex_trylock)
strong_alias(__pthread_mutex_unlock, pthread_mutex_unlock)
strong_alias(__pthread_mutexattr_init, pthread_mutexattr_init)
weak_alias(__pthread_mutexattr_settype, pthread_mutexattr_settype)
strong_alias(__pthread_mutex_init, pthread_mutex_init)
strong_alias(__pthread_mutexattr_destroy, pthread_mutexattr_destroy)
strong_alias(__pthread_mutex_destroy, pthread_mutex_destroy)
strong_alias(__pthread_once, pthread_once)
strong_alias(__pthread_atfork, pthread_atfork)
strong_alias(__pthread_key_create, pthread_key_create)
strong_alias(__pthread_getspecific, pthread_getspecific)
strong_alias(__pthread_setspecific, pthread_setspecific)
#ifndef GLIBC_2_1
strong_alias(sigaction, __sigaction)
#endif
strong_alias(close, __close)
strong_alias(fcntl, __fcntl)
strong_alias(lseek, __lseek)
strong_alias(open, __open)
strong_alias(open64, __open64)
strong_alias(read, __read)
strong_alias(wait, __wait)
strong_alias(write, __write)
strong_alias(connect, __connect)
strong_alias(send, __send)
weak_alias (__pread64, pread64)
weak_alias (__pwrite64, pwrite64)
weak_alias(__fork, fork)
weak_alias(__vfork, vfork)
weak_alias (__pthread_kill_other_threads_np, pthread_kill_other_threads_np)
/*--------------------------------------------------*/
weak_alias(pthread_rwlock_rdlock, __pthread_rwlock_rdlock)
weak_alias(pthread_rwlock_unlock, __pthread_rwlock_unlock)
weak_alias(pthread_rwlock_wrlock, __pthread_rwlock_wrlock)
weak_alias(pthread_rwlock_destroy, __pthread_rwlock_destroy)
weak_alias(pthread_rwlock_init, __pthread_rwlock_init)
weak_alias(pthread_rwlock_tryrdlock, __pthread_rwlock_tryrdlock)
weak_alias(pthread_rwlock_trywrlock, __pthread_rwlock_trywrlock)
/* I've no idea what these are, but they get called quite a lot.
Anybody know? */
#undef _IO_flockfile
void _IO_flockfile ( _IO_FILE * file )
{
pthread_mutex_lock(file->_lock);
}
weak_alias(_IO_flockfile, flockfile);
#undef _IO_funlockfile
void _IO_funlockfile ( _IO_FILE * file )
{
pthread_mutex_unlock(file->_lock);
}
weak_alias(_IO_funlockfile, funlockfile);
/* This doesn't seem to be needed to simulate libpthread.so's external
interface, but many people complain about its absence. */
strong_alias(__pthread_mutexattr_settype, __pthread_mutexattr_setkind_np)
weak_alias(__pthread_mutexattr_setkind_np, pthread_mutexattr_setkind_np)
/*--------------------------------------------------------------------*/
/*--- end vg_libpthread.c ---*/
/*--------------------------------------------------------------------*/