mirror of
https://git.FreeBSD.org/src.git
synced 2024-12-17 10:26:15 +00:00
0e9a260520
C runtime services, like printf(). Unfortunately, the multithread-safeness measures in the libc do not work in rtld environment. Rip the kernel printf() implementation and use it in the rtld instead of libc version. This printf does not require any shared global data and thus is mt-safe. Systematically use rtld_printf() and related functions, remove the calls to err(3). Note that stdio is still pulled from libc due to libmap implementaion using fopen(). This is safe but unoptimal, and can be changed later. Reported and tested by: pgj Diagnosed and reviewed by: kan (previous version) Approved by: re (bz)
387 lines
9.3 KiB
C
387 lines
9.3 KiB
C
/*-
|
|
* Copyright 1999, 2000 John D. Polstra.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
|
|
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
|
|
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
|
|
* IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
|
|
* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
|
|
* NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
|
|
* THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*
|
|
* from: FreeBSD: src/libexec/rtld-elf/sparc64/lockdflt.c,v 1.3 2002/10/09
|
|
* $FreeBSD$
|
|
*/
|
|
|
|
/*
|
|
* Thread locking implementation for the dynamic linker.
|
|
*
|
|
* We use the "simple, non-scalable reader-preference lock" from:
|
|
*
|
|
* J. M. Mellor-Crummey and M. L. Scott. "Scalable Reader-Writer
|
|
* Synchronization for Shared-Memory Multiprocessors." 3rd ACM Symp. on
|
|
* Principles and Practice of Parallel Programming, April 1991.
|
|
*
|
|
* In this algorithm the lock is a single word. Its low-order bit is
|
|
* set when a writer holds the lock. The remaining high-order bits
|
|
* contain a count of readers desiring the lock. The algorithm requires
|
|
* atomic "compare_and_store" and "add" operations, which we implement
|
|
* using assembly language sequences in "rtld_start.S".
|
|
*/
|
|
|
|
#include <sys/param.h>
|
|
#include <signal.h>
|
|
#include <stdlib.h>
|
|
#include <time.h>
|
|
|
|
#include "debug.h"
|
|
#include "rtld.h"
|
|
#include "rtld_machdep.h"
|
|
|
|
#define WAFLAG 0x1 /* A writer holds the lock */
|
|
#define RC_INCR 0x2 /* Adjusts count of readers desiring lock */
|
|
|
|
typedef struct Struct_Lock {
|
|
volatile u_int lock;
|
|
void *base;
|
|
} Lock;
|
|
|
|
static sigset_t fullsigmask, oldsigmask;
|
|
static int thread_flag;
|
|
|
|
static void *
|
|
def_lock_create()
|
|
{
|
|
void *base;
|
|
char *p;
|
|
uintptr_t r;
|
|
Lock *l;
|
|
|
|
/*
|
|
* Arrange for the lock to occupy its own cache line. First, we
|
|
* optimistically allocate just a cache line, hoping that malloc
|
|
* will give us a well-aligned block of memory. If that doesn't
|
|
* work, we allocate a larger block and take a well-aligned cache
|
|
* line from it.
|
|
*/
|
|
base = xmalloc(CACHE_LINE_SIZE);
|
|
p = (char *)base;
|
|
if ((uintptr_t)p % CACHE_LINE_SIZE != 0) {
|
|
free(base);
|
|
base = xmalloc(2 * CACHE_LINE_SIZE);
|
|
p = (char *)base;
|
|
if ((r = (uintptr_t)p % CACHE_LINE_SIZE) != 0)
|
|
p += CACHE_LINE_SIZE - r;
|
|
}
|
|
l = (Lock *)p;
|
|
l->base = base;
|
|
l->lock = 0;
|
|
return l;
|
|
}
|
|
|
|
static void
|
|
def_lock_destroy(void *lock)
|
|
{
|
|
Lock *l = (Lock *)lock;
|
|
|
|
free(l->base);
|
|
}
|
|
|
|
static void
|
|
def_rlock_acquire(void *lock)
|
|
{
|
|
Lock *l = (Lock *)lock;
|
|
|
|
atomic_add_acq_int(&l->lock, RC_INCR);
|
|
while (l->lock & WAFLAG)
|
|
; /* Spin */
|
|
}
|
|
|
|
static void
|
|
def_wlock_acquire(void *lock)
|
|
{
|
|
Lock *l = (Lock *)lock;
|
|
sigset_t tmp_oldsigmask;
|
|
|
|
for ( ; ; ) {
|
|
sigprocmask(SIG_BLOCK, &fullsigmask, &tmp_oldsigmask);
|
|
if (atomic_cmpset_acq_int(&l->lock, 0, WAFLAG))
|
|
break;
|
|
sigprocmask(SIG_SETMASK, &tmp_oldsigmask, NULL);
|
|
}
|
|
oldsigmask = tmp_oldsigmask;
|
|
}
|
|
|
|
static void
|
|
def_lock_release(void *lock)
|
|
{
|
|
Lock *l = (Lock *)lock;
|
|
|
|
if ((l->lock & WAFLAG) == 0)
|
|
atomic_add_rel_int(&l->lock, -RC_INCR);
|
|
else {
|
|
atomic_add_rel_int(&l->lock, -WAFLAG);
|
|
sigprocmask(SIG_SETMASK, &oldsigmask, NULL);
|
|
}
|
|
}
|
|
|
|
static int
|
|
def_thread_set_flag(int mask)
|
|
{
|
|
int old_val = thread_flag;
|
|
thread_flag |= mask;
|
|
return (old_val);
|
|
}
|
|
|
|
static int
|
|
def_thread_clr_flag(int mask)
|
|
{
|
|
int old_val = thread_flag;
|
|
thread_flag &= ~mask;
|
|
return (old_val);
|
|
}
|
|
|
|
/*
|
|
* Public interface exposed to the rest of the dynamic linker.
|
|
*/
|
|
static struct RtldLockInfo lockinfo;
|
|
static struct RtldLockInfo deflockinfo;
|
|
|
|
static __inline int
|
|
thread_mask_set(int mask)
|
|
{
|
|
return lockinfo.thread_set_flag(mask);
|
|
}
|
|
|
|
static __inline void
|
|
thread_mask_clear(int mask)
|
|
{
|
|
lockinfo.thread_clr_flag(mask);
|
|
}
|
|
|
|
#define RTLD_LOCK_CNT 3
|
|
struct rtld_lock {
|
|
void *handle;
|
|
int mask;
|
|
} rtld_locks[RTLD_LOCK_CNT];
|
|
|
|
rtld_lock_t rtld_bind_lock = &rtld_locks[0];
|
|
rtld_lock_t rtld_libc_lock = &rtld_locks[1];
|
|
rtld_lock_t rtld_phdr_lock = &rtld_locks[2];
|
|
|
|
void
|
|
rlock_acquire(rtld_lock_t lock, RtldLockState *lockstate)
|
|
{
|
|
|
|
if (lockstate == NULL)
|
|
return;
|
|
|
|
if (thread_mask_set(lock->mask) & lock->mask) {
|
|
dbg("rlock_acquire: recursed");
|
|
lockstate->lockstate = RTLD_LOCK_UNLOCKED;
|
|
return;
|
|
}
|
|
lockinfo.rlock_acquire(lock->handle);
|
|
lockstate->lockstate = RTLD_LOCK_RLOCKED;
|
|
}
|
|
|
|
void
|
|
wlock_acquire(rtld_lock_t lock, RtldLockState *lockstate)
|
|
{
|
|
|
|
if (lockstate == NULL)
|
|
return;
|
|
|
|
if (thread_mask_set(lock->mask) & lock->mask) {
|
|
dbg("wlock_acquire: recursed");
|
|
lockstate->lockstate = RTLD_LOCK_UNLOCKED;
|
|
return;
|
|
}
|
|
lockinfo.wlock_acquire(lock->handle);
|
|
lockstate->lockstate = RTLD_LOCK_WLOCKED;
|
|
}
|
|
|
|
void
|
|
lock_release(rtld_lock_t lock, RtldLockState *lockstate)
|
|
{
|
|
|
|
if (lockstate == NULL)
|
|
return;
|
|
|
|
switch (lockstate->lockstate) {
|
|
case RTLD_LOCK_UNLOCKED:
|
|
break;
|
|
case RTLD_LOCK_RLOCKED:
|
|
case RTLD_LOCK_WLOCKED:
|
|
thread_mask_clear(lock->mask);
|
|
lockinfo.lock_release(lock->handle);
|
|
break;
|
|
default:
|
|
assert(0);
|
|
}
|
|
}
|
|
|
|
void
|
|
lock_upgrade(rtld_lock_t lock, RtldLockState *lockstate)
|
|
{
|
|
|
|
if (lockstate == NULL)
|
|
return;
|
|
|
|
lock_release(lock, lockstate);
|
|
wlock_acquire(lock, lockstate);
|
|
}
|
|
|
|
void
|
|
lock_restart_for_upgrade(RtldLockState *lockstate)
|
|
{
|
|
|
|
if (lockstate == NULL)
|
|
return;
|
|
|
|
switch (lockstate->lockstate) {
|
|
case RTLD_LOCK_UNLOCKED:
|
|
case RTLD_LOCK_WLOCKED:
|
|
break;
|
|
case RTLD_LOCK_RLOCKED:
|
|
siglongjmp(lockstate->env, 1);
|
|
break;
|
|
default:
|
|
assert(0);
|
|
}
|
|
}
|
|
|
|
void
|
|
lockdflt_init()
|
|
{
|
|
int i;
|
|
|
|
deflockinfo.rtli_version = RTLI_VERSION;
|
|
deflockinfo.lock_create = def_lock_create;
|
|
deflockinfo.lock_destroy = def_lock_destroy;
|
|
deflockinfo.rlock_acquire = def_rlock_acquire;
|
|
deflockinfo.wlock_acquire = def_wlock_acquire;
|
|
deflockinfo.lock_release = def_lock_release;
|
|
deflockinfo.thread_set_flag = def_thread_set_flag;
|
|
deflockinfo.thread_clr_flag = def_thread_clr_flag;
|
|
deflockinfo.at_fork = NULL;
|
|
|
|
for (i = 0; i < RTLD_LOCK_CNT; i++) {
|
|
rtld_locks[i].mask = (1 << i);
|
|
rtld_locks[i].handle = NULL;
|
|
}
|
|
|
|
memcpy(&lockinfo, &deflockinfo, sizeof(lockinfo));
|
|
_rtld_thread_init(NULL);
|
|
/*
|
|
* Construct a mask to block all signals except traps which might
|
|
* conceivably be generated within the dynamic linker itself.
|
|
*/
|
|
sigfillset(&fullsigmask);
|
|
sigdelset(&fullsigmask, SIGILL);
|
|
sigdelset(&fullsigmask, SIGTRAP);
|
|
sigdelset(&fullsigmask, SIGABRT);
|
|
sigdelset(&fullsigmask, SIGEMT);
|
|
sigdelset(&fullsigmask, SIGFPE);
|
|
sigdelset(&fullsigmask, SIGBUS);
|
|
sigdelset(&fullsigmask, SIGSEGV);
|
|
sigdelset(&fullsigmask, SIGSYS);
|
|
}
|
|
|
|
/*
|
|
* Callback function to allow threads implementation to
|
|
* register their own locking primitives if the default
|
|
* one is not suitable.
|
|
* The current context should be the only context
|
|
* executing at the invocation time.
|
|
*/
|
|
void
|
|
_rtld_thread_init(struct RtldLockInfo *pli)
|
|
{
|
|
int flags, i;
|
|
void *locks[RTLD_LOCK_CNT];
|
|
|
|
/* disable all locking while this function is running */
|
|
flags = thread_mask_set(~0);
|
|
|
|
if (pli == NULL)
|
|
pli = &deflockinfo;
|
|
|
|
|
|
for (i = 0; i < RTLD_LOCK_CNT; i++)
|
|
if ((locks[i] = pli->lock_create()) == NULL)
|
|
break;
|
|
|
|
if (i < RTLD_LOCK_CNT) {
|
|
while (--i >= 0)
|
|
pli->lock_destroy(locks[i]);
|
|
abort();
|
|
}
|
|
|
|
for (i = 0; i < RTLD_LOCK_CNT; i++) {
|
|
if (rtld_locks[i].handle == NULL)
|
|
continue;
|
|
if (flags & rtld_locks[i].mask)
|
|
lockinfo.lock_release(rtld_locks[i].handle);
|
|
lockinfo.lock_destroy(rtld_locks[i].handle);
|
|
}
|
|
|
|
for (i = 0; i < RTLD_LOCK_CNT; i++) {
|
|
rtld_locks[i].handle = locks[i];
|
|
if (flags & rtld_locks[i].mask)
|
|
pli->wlock_acquire(rtld_locks[i].handle);
|
|
}
|
|
|
|
lockinfo.lock_create = pli->lock_create;
|
|
lockinfo.lock_destroy = pli->lock_destroy;
|
|
lockinfo.rlock_acquire = pli->rlock_acquire;
|
|
lockinfo.wlock_acquire = pli->wlock_acquire;
|
|
lockinfo.lock_release = pli->lock_release;
|
|
lockinfo.thread_set_flag = pli->thread_set_flag;
|
|
lockinfo.thread_clr_flag = pli->thread_clr_flag;
|
|
lockinfo.at_fork = pli->at_fork;
|
|
|
|
/* restore thread locking state, this time with new locks */
|
|
thread_mask_clear(~0);
|
|
thread_mask_set(flags);
|
|
dbg("_rtld_thread_init: done");
|
|
}
|
|
|
|
void
|
|
_rtld_atfork_pre(int *locks)
|
|
{
|
|
RtldLockState ls[2];
|
|
|
|
wlock_acquire(rtld_phdr_lock, &ls[0]);
|
|
rlock_acquire(rtld_bind_lock, &ls[1]);
|
|
|
|
/* XXXKIB: I am really sorry for this. */
|
|
locks[0] = ls[1].lockstate;
|
|
locks[2] = ls[0].lockstate;
|
|
}
|
|
|
|
void
|
|
_rtld_atfork_post(int *locks)
|
|
{
|
|
RtldLockState ls[2];
|
|
|
|
bzero(ls, sizeof(ls));
|
|
ls[0].lockstate = locks[2];
|
|
ls[1].lockstate = locks[0];
|
|
lock_release(rtld_bind_lock, &ls[1]);
|
|
lock_release(rtld_phdr_lock, &ls[0]);
|
|
}
|