2006-01-19 09:42:46 +08:00
|
|
|
/*
|
|
|
|
* Copyright (C) 2002- 2004 Jeff Dike (jdike@addtoit.com)
|
|
|
|
* Licensed under the GPL
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <stdlib.h>
|
|
|
|
#include <string.h>
|
|
|
|
#include <unistd.h>
|
|
|
|
#include <errno.h>
|
|
|
|
#include <signal.h>
|
|
|
|
#include <sched.h>
|
|
|
|
#include "ptrace_user.h"
|
|
|
|
#include <sys/wait.h>
|
|
|
|
#include <sys/mman.h>
|
|
|
|
#include <sys/user.h>
|
|
|
|
#include <sys/time.h>
|
2006-11-03 14:07:23 +08:00
|
|
|
#include <sys/syscall.h>
|
2006-01-19 09:42:46 +08:00
|
|
|
#include <asm/types.h>
|
|
|
|
#include "user.h"
|
|
|
|
#include "sysdep/ptrace.h"
|
|
|
|
#include "kern_util.h"
|
|
|
|
#include "skas.h"
|
|
|
|
#include "stub-data.h"
|
|
|
|
#include "mm_id.h"
|
|
|
|
#include "sysdep/sigcontext.h"
|
|
|
|
#include "sysdep/stub.h"
|
|
|
|
#include "os.h"
|
|
|
|
#include "proc_mm.h"
|
|
|
|
#include "skas_ptrace.h"
|
|
|
|
#include "chan_user.h"
|
|
|
|
#include "registers.h"
|
|
|
|
#include "mem.h"
|
|
|
|
#include "uml-config.h"
|
|
|
|
#include "process.h"
|
[PATCH] uml: implement soft interrupts
This patch implements soft interrupts. Interrupt enabling and disabling no
longer map to sigprocmask. Rather, a flag is set indicating whether
interrupts may be handled. If a signal comes in and interrupts are marked as
OK, then it is handled normally. If interrupts are marked as off, then the
signal handler simply returns after noting that a signal needs handling. When
interrupts are enabled later on, this pending signals flag is checked, and the
IRQ handlers are called at that point.
The point of this is to reduce the cost of local_irq_save et al, since they
are very much more common than the signals that they are enabling and
disabling. Soft interrupts produce a speed-up of ~25% on a kernel build.
Subtleties -
UML uses sigsetjmp/siglongjmp to switch contexts. sigsetjmp has been
wrapped in a save_flags-like macro which remembers the interrupt state at
setjmp time, and restores it when it is longjmp-ed back to.
The enable_signals function has to loop because the IRQ handler
disables interrupts before returning. enable_signals has to return with
signals enabled, and signals may come in between the disabling and the
return to enable_signals. So, it loops for as long as there are pending
signals, ensuring that signals are enabled when it finally returns, and
that there are no pending signals that need to be dealt with.
Signed-off-by: Jeff Dike <jdike@addtoit.com>
Cc: Paolo 'Blaisorblade' Giarrusso <blaisorblade@yahoo.it>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-01-19 09:42:49 +08:00
|
|
|
#include "longjmp.h"
|
2007-05-07 05:51:22 +08:00
|
|
|
#include "kern_constants.h"
|
2007-05-07 05:51:48 +08:00
|
|
|
#include "as-layout.h"
|
2006-01-19 09:42:46 +08:00
|
|
|
|
|
|
|
int is_skas_winch(int pid, int fd, void *data)
|
|
|
|
{
|
|
|
|
if(pid != os_getpgrp())
|
|
|
|
return(0);
|
|
|
|
|
|
|
|
register_winch_irq(-1, fd, -1, data);
|
|
|
|
return(1);
|
|
|
|
}
|
|
|
|
|
2007-05-07 05:51:29 +08:00
|
|
|
static int ptrace_dump_regs(int pid)
|
|
|
|
{
|
|
|
|
unsigned long regs[MAX_REG_NR];
|
|
|
|
int i;
|
|
|
|
|
|
|
|
if(ptrace(PTRACE_GETREGS, pid, 0, regs) < 0)
|
|
|
|
return -errno;
|
|
|
|
else {
|
|
|
|
printk("Stub registers -\n");
|
|
|
|
for(i = 0; i < ARRAY_SIZE(regs); i++)
|
|
|
|
printk("\t%d - %lx\n", i, regs[i]);
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2007-05-07 05:51:48 +08:00
|
|
|
/*
|
|
|
|
* Signals that are OK to receive in the stub - we'll just continue it.
|
|
|
|
* SIGWINCH will happen when UML is inside a detached screen.
|
|
|
|
*/
|
|
|
|
#define STUB_SIG_MASK ((1 << SIGVTALRM) | (1 << SIGWINCH))
|
|
|
|
|
|
|
|
/* Signals that the stub will finish with - anything else is an error */
|
|
|
|
#define STUB_DONE_MASK ((1 << SIGUSR1) | (1 << SIGTRAP))
|
|
|
|
|
|
|
|
void wait_stub_done(int pid)
|
2006-01-19 09:42:46 +08:00
|
|
|
{
|
|
|
|
int n, status, err;
|
|
|
|
|
2007-05-07 05:51:48 +08:00
|
|
|
while(1){
|
2006-01-19 09:42:46 +08:00
|
|
|
CATCH_EINTR(n = waitpid(pid, &status, WUNTRACED));
|
2007-05-07 05:51:48 +08:00
|
|
|
if((n < 0) || !WIFSTOPPED(status))
|
|
|
|
goto bad_wait;
|
|
|
|
|
|
|
|
if(((1 << WSTOPSIG(status)) & STUB_SIG_MASK) == 0)
|
|
|
|
break;
|
|
|
|
|
|
|
|
err = ptrace(PTRACE_CONT, pid, 0, 0);
|
2007-05-07 05:51:29 +08:00
|
|
|
if(err)
|
2007-05-07 05:51:48 +08:00
|
|
|
panic("wait_stub_done : continue failed, errno = %d\n",
|
|
|
|
errno);
|
2006-01-19 09:42:46 +08:00
|
|
|
}
|
2007-05-07 05:51:48 +08:00
|
|
|
|
|
|
|
if(((1 << WSTOPSIG(status)) & STUB_DONE_MASK) != 0)
|
|
|
|
return;
|
|
|
|
|
|
|
|
bad_wait:
|
|
|
|
err = ptrace_dump_regs(pid);
|
|
|
|
if(err)
|
|
|
|
printk("Failed to get registers from stub, errno = %d\n", -err);
|
|
|
|
panic("wait_stub_done : failed to wait for SIGUSR1/SIGTRAP, pid = %d, "
|
|
|
|
"n = %d, errno = %d, status = 0x%x\n", pid, n, errno, status);
|
2006-01-19 09:42:46 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
extern unsigned long current_stub_stack(void);
|
|
|
|
|
|
|
|
void get_skas_faultinfo(int pid, struct faultinfo * fi)
|
|
|
|
{
|
|
|
|
int err;
|
|
|
|
|
|
|
|
if(ptrace_faultinfo){
|
|
|
|
err = ptrace(PTRACE_FAULTINFO, pid, 0, fi);
|
|
|
|
if(err)
|
|
|
|
panic("get_skas_faultinfo - PTRACE_FAULTINFO failed, "
|
|
|
|
"errno = %d\n", errno);
|
|
|
|
|
|
|
|
/* Special handling for i386, which has different structs */
|
|
|
|
if (sizeof(struct ptrace_faultinfo) < sizeof(struct faultinfo))
|
|
|
|
memset((char *)fi + sizeof(struct ptrace_faultinfo), 0,
|
|
|
|
sizeof(struct faultinfo) -
|
|
|
|
sizeof(struct ptrace_faultinfo));
|
|
|
|
}
|
|
|
|
else {
|
2007-05-07 05:51:48 +08:00
|
|
|
err = ptrace(PTRACE_CONT, pid, 0, SIGSEGV);
|
|
|
|
if(err)
|
|
|
|
panic("Failed to continue stub, pid = %d, errno = %d\n",
|
|
|
|
pid, errno);
|
|
|
|
wait_stub_done(pid);
|
2006-01-19 09:42:46 +08:00
|
|
|
|
|
|
|
/* faultinfo is prepared by the stub-segv-handler at start of
|
|
|
|
* the stub stack page. We just have to copy it.
|
|
|
|
*/
|
|
|
|
memcpy(fi, (void *)current_stub_stack(), sizeof(*fi));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void handle_segv(int pid, union uml_pt_regs * regs)
|
|
|
|
{
|
|
|
|
get_skas_faultinfo(pid, ®s->skas.faultinfo);
|
|
|
|
segv(regs->skas.faultinfo, 0, 1, NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*To use the same value of using_sysemu as the caller, ask it that value (in local_using_sysemu)*/
|
|
|
|
static void handle_trap(int pid, union uml_pt_regs *regs, int local_using_sysemu)
|
|
|
|
{
|
|
|
|
int err, status;
|
|
|
|
|
|
|
|
/* Mark this as a syscall */
|
|
|
|
UPT_SYSCALL_NR(regs) = PT_SYSCALL_NR(regs->skas.regs);
|
|
|
|
|
|
|
|
if (!local_using_sysemu)
|
|
|
|
{
|
|
|
|
err = ptrace(PTRACE_POKEUSR, pid, PT_SYSCALL_NR_OFFSET,
|
|
|
|
__NR_getpid);
|
|
|
|
if(err < 0)
|
|
|
|
panic("handle_trap - nullifying syscall failed errno = %d\n",
|
|
|
|
errno);
|
|
|
|
|
|
|
|
err = ptrace(PTRACE_SYSCALL, pid, 0, 0);
|
|
|
|
if(err < 0)
|
|
|
|
panic("handle_trap - continuing to end of syscall failed, "
|
|
|
|
"errno = %d\n", errno);
|
|
|
|
|
|
|
|
CATCH_EINTR(err = waitpid(pid, &status, WUNTRACED));
|
|
|
|
if((err < 0) || !WIFSTOPPED(status) ||
|
2007-05-07 05:51:29 +08:00
|
|
|
(WSTOPSIG(status) != SIGTRAP + 0x80)){
|
|
|
|
err = ptrace_dump_regs(pid);
|
|
|
|
if(err)
|
|
|
|
printk("Failed to get registers from process, "
|
|
|
|
"errno = %d\n", -err);
|
2006-01-19 09:42:46 +08:00
|
|
|
panic("handle_trap - failed to wait at end of syscall, "
|
|
|
|
"errno = %d, status = %d\n", errno, status);
|
2007-05-07 05:51:29 +08:00
|
|
|
}
|
2006-01-19 09:42:46 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
handle_syscall(regs);
|
|
|
|
}
|
|
|
|
|
|
|
|
extern int __syscall_stub_start;
|
|
|
|
|
|
|
|
static int userspace_tramp(void *stack)
|
|
|
|
{
|
|
|
|
void *addr;
|
2006-09-26 14:33:05 +08:00
|
|
|
int err;
|
2006-01-19 09:42:46 +08:00
|
|
|
|
|
|
|
ptrace(PTRACE_TRACEME, 0, 0, 0);
|
|
|
|
|
2006-07-10 19:45:07 +08:00
|
|
|
init_new_thread_signals();
|
2006-09-26 14:33:05 +08:00
|
|
|
err = set_interval(1);
|
|
|
|
if(err)
|
|
|
|
panic("userspace_tramp - setting timer failed, errno = %d\n",
|
|
|
|
err);
|
2006-01-19 09:42:46 +08:00
|
|
|
|
|
|
|
if(!proc_mm){
|
|
|
|
/* This has a pte, but it can't be mapped in with the usual
|
|
|
|
* tlb_flush mechanism because this is part of that mechanism
|
|
|
|
*/
|
2006-01-19 09:42:48 +08:00
|
|
|
int fd;
|
|
|
|
__u64 offset;
|
|
|
|
fd = phys_mapping(to_phys(&__syscall_stub_start), &offset);
|
2007-05-07 05:51:22 +08:00
|
|
|
addr = mmap64((void *) UML_CONFIG_STUB_CODE, UM_KERN_PAGE_SIZE,
|
2006-01-19 09:42:48 +08:00
|
|
|
PROT_EXEC, MAP_FIXED | MAP_PRIVATE, fd, offset);
|
2006-01-19 09:42:46 +08:00
|
|
|
if(addr == MAP_FAILED){
|
2006-01-19 09:42:48 +08:00
|
|
|
printk("mapping mmap stub failed, errno = %d\n",
|
2006-01-19 09:42:46 +08:00
|
|
|
errno);
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
if(stack != NULL){
|
|
|
|
fd = phys_mapping(to_phys(stack), &offset);
|
2007-05-07 05:51:22 +08:00
|
|
|
addr = mmap((void *) UML_CONFIG_STUB_DATA,
|
|
|
|
UM_KERN_PAGE_SIZE, PROT_READ | PROT_WRITE,
|
2006-01-19 09:42:46 +08:00
|
|
|
MAP_FIXED | MAP_SHARED, fd, offset);
|
|
|
|
if(addr == MAP_FAILED){
|
2006-01-19 09:42:48 +08:00
|
|
|
printk("mapping segfault stack failed, "
|
2006-01-19 09:42:46 +08:00
|
|
|
"errno = %d\n", errno);
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2006-01-19 09:42:48 +08:00
|
|
|
if(!ptrace_faultinfo && (stack != NULL)){
|
2006-09-26 14:33:04 +08:00
|
|
|
struct sigaction sa;
|
|
|
|
|
2006-01-19 09:42:46 +08:00
|
|
|
unsigned long v = UML_CONFIG_STUB_CODE +
|
|
|
|
(unsigned long) stub_segv_handler -
|
|
|
|
(unsigned long) &__syscall_stub_start;
|
|
|
|
|
2007-05-07 05:51:22 +08:00
|
|
|
set_sigstack((void *) UML_CONFIG_STUB_DATA, UM_KERN_PAGE_SIZE);
|
2006-09-26 14:33:04 +08:00
|
|
|
sigemptyset(&sa.sa_mask);
|
|
|
|
sigaddset(&sa.sa_mask, SIGIO);
|
|
|
|
sigaddset(&sa.sa_mask, SIGWINCH);
|
|
|
|
sigaddset(&sa.sa_mask, SIGALRM);
|
|
|
|
sigaddset(&sa.sa_mask, SIGVTALRM);
|
|
|
|
sigaddset(&sa.sa_mask, SIGUSR1);
|
|
|
|
sa.sa_flags = SA_ONSTACK;
|
|
|
|
sa.sa_handler = (void *) v;
|
|
|
|
sa.sa_restorer = NULL;
|
|
|
|
if(sigaction(SIGSEGV, &sa, NULL) < 0)
|
|
|
|
panic("userspace_tramp - setting SIGSEGV handler "
|
|
|
|
"failed - errno = %d\n", errno);
|
2006-01-19 09:42:46 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
os_stop_process(os_getpid());
|
|
|
|
return(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Each element set once, and only accessed by a single processor anyway */
|
|
|
|
#undef NR_CPUS
|
|
|
|
#define NR_CPUS 1
|
|
|
|
int userspace_pid[NR_CPUS];
|
|
|
|
|
|
|
|
int start_userspace(unsigned long stub_stack)
|
|
|
|
{
|
|
|
|
void *stack;
|
|
|
|
unsigned long sp;
|
|
|
|
int pid, status, n, flags;
|
|
|
|
|
|
|
|
stack = mmap(NULL, PAGE_SIZE, PROT_READ | PROT_WRITE | PROT_EXEC,
|
|
|
|
MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
|
|
|
|
if(stack == MAP_FAILED)
|
|
|
|
panic("start_userspace : mmap failed, errno = %d", errno);
|
|
|
|
sp = (unsigned long) stack + PAGE_SIZE - sizeof(void *);
|
|
|
|
|
|
|
|
flags = CLONE_FILES | SIGCHLD;
|
|
|
|
if(proc_mm) flags |= CLONE_VM;
|
|
|
|
pid = clone(userspace_tramp, (void *) sp, flags, (void *) stub_stack);
|
|
|
|
if(pid < 0)
|
|
|
|
panic("start_userspace : clone failed, errno = %d", errno);
|
|
|
|
|
|
|
|
do {
|
|
|
|
CATCH_EINTR(n = waitpid(pid, &status, WUNTRACED));
|
|
|
|
if(n < 0)
|
|
|
|
panic("start_userspace : wait failed, errno = %d",
|
|
|
|
errno);
|
|
|
|
} while(WIFSTOPPED(status) && (WSTOPSIG(status) == SIGVTALRM));
|
|
|
|
|
|
|
|
if(!WIFSTOPPED(status) || (WSTOPSIG(status) != SIGSTOP))
|
|
|
|
panic("start_userspace : expected SIGSTOP, got status = %d",
|
|
|
|
status);
|
|
|
|
|
|
|
|
if (ptrace(PTRACE_OLDSETOPTIONS, pid, NULL, (void *)PTRACE_O_TRACESYSGOOD) < 0)
|
|
|
|
panic("start_userspace : PTRACE_OLDSETOPTIONS failed, errno=%d\n",
|
|
|
|
errno);
|
|
|
|
|
|
|
|
if(munmap(stack, PAGE_SIZE) < 0)
|
|
|
|
panic("start_userspace : munmap failed, errno = %d\n", errno);
|
|
|
|
|
|
|
|
return(pid);
|
|
|
|
}
|
|
|
|
|
|
|
|
void userspace(union uml_pt_regs *regs)
|
|
|
|
{
|
|
|
|
int err, status, op, pid = userspace_pid[0];
|
|
|
|
int local_using_sysemu; /*To prevent races if using_sysemu changes under us.*/
|
|
|
|
|
|
|
|
while(1){
|
|
|
|
restore_registers(pid, regs);
|
|
|
|
|
|
|
|
/* Now we set local_using_sysemu to be used for one loop */
|
|
|
|
local_using_sysemu = get_using_sysemu();
|
|
|
|
|
|
|
|
op = SELECT_PTRACE_OPERATION(local_using_sysemu, singlestepping(NULL));
|
|
|
|
|
|
|
|
err = ptrace(op, pid, 0, 0);
|
|
|
|
if(err)
|
|
|
|
panic("userspace - could not resume userspace process, "
|
|
|
|
"pid=%d, ptrace operation = %d, errno = %d\n",
|
2006-04-11 13:53:32 +08:00
|
|
|
pid, op, errno);
|
2006-01-19 09:42:46 +08:00
|
|
|
|
|
|
|
CATCH_EINTR(err = waitpid(pid, &status, WUNTRACED));
|
|
|
|
if(err < 0)
|
|
|
|
panic("userspace - waitpid failed, errno = %d\n",
|
|
|
|
errno);
|
|
|
|
|
|
|
|
regs->skas.is_user = 1;
|
|
|
|
save_registers(pid, regs);
|
|
|
|
UPT_SYSCALL_NR(regs) = -1; /* Assume: It's not a syscall */
|
|
|
|
|
|
|
|
if(WIFSTOPPED(status)){
|
2007-05-07 05:51:48 +08:00
|
|
|
int sig = WSTOPSIG(status);
|
|
|
|
switch(sig){
|
2006-01-19 09:42:46 +08:00
|
|
|
case SIGSEGV:
|
2007-05-07 05:51:48 +08:00
|
|
|
if(PTRACE_FULL_FAULTINFO || !ptrace_faultinfo){
|
|
|
|
get_skas_faultinfo(pid, ®s->skas.faultinfo);
|
|
|
|
(*sig_info[SIGSEGV])(SIGSEGV, regs);
|
|
|
|
}
|
2006-01-19 09:42:46 +08:00
|
|
|
else handle_segv(pid, regs);
|
|
|
|
break;
|
|
|
|
case SIGTRAP + 0x80:
|
|
|
|
handle_trap(pid, regs, local_using_sysemu);
|
|
|
|
break;
|
|
|
|
case SIGTRAP:
|
|
|
|
relay_signal(SIGTRAP, regs);
|
|
|
|
break;
|
|
|
|
case SIGIO:
|
|
|
|
case SIGVTALRM:
|
|
|
|
case SIGILL:
|
|
|
|
case SIGBUS:
|
|
|
|
case SIGFPE:
|
|
|
|
case SIGWINCH:
|
2007-05-07 05:51:48 +08:00
|
|
|
block_signals();
|
|
|
|
(*sig_info[sig])(sig, regs);
|
|
|
|
unblock_signals();
|
2006-01-19 09:42:46 +08:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
printk("userspace - child stopped with signal "
|
2007-05-07 05:51:48 +08:00
|
|
|
"%d\n", sig);
|
2006-01-19 09:42:46 +08:00
|
|
|
}
|
|
|
|
pid = userspace_pid[0];
|
|
|
|
interrupt_end();
|
|
|
|
|
|
|
|
/* Avoid -ERESTARTSYS handling in host */
|
|
|
|
if(PT_SYSCALL_NR_OFFSET != PT_SYSCALL_RET_OFFSET)
|
|
|
|
PT_SYSCALL_NR(regs->skas.regs) = -1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2007-05-07 05:51:48 +08:00
|
|
|
static unsigned long thread_regs[MAX_REG_NR];
|
|
|
|
static unsigned long thread_fp_regs[HOST_FP_SIZE];
|
|
|
|
|
|
|
|
static int __init init_thread_regs(void)
|
|
|
|
{
|
|
|
|
get_safe_registers(thread_regs, thread_fp_regs);
|
|
|
|
/* Set parent's instruction pointer to start of clone-stub */
|
|
|
|
thread_regs[REGS_IP_INDEX] = UML_CONFIG_STUB_CODE +
|
|
|
|
(unsigned long) stub_clone_handler -
|
|
|
|
(unsigned long) &__syscall_stub_start;
|
|
|
|
thread_regs[REGS_SP_INDEX] = UML_CONFIG_STUB_DATA + PAGE_SIZE -
|
|
|
|
sizeof(void *);
|
|
|
|
#ifdef __SIGNAL_FRAMESIZE
|
|
|
|
thread_regs[REGS_SP_INDEX] -= __SIGNAL_FRAMESIZE;
|
|
|
|
#endif
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
__initcall(init_thread_regs);
|
|
|
|
|
2006-01-19 09:42:46 +08:00
|
|
|
int copy_context_skas0(unsigned long new_stack, int pid)
|
|
|
|
{
|
|
|
|
int err;
|
|
|
|
unsigned long current_stack = current_stub_stack();
|
|
|
|
struct stub_data *data = (struct stub_data *) current_stack;
|
|
|
|
struct stub_data *child_data = (struct stub_data *) new_stack;
|
|
|
|
__u64 new_offset;
|
|
|
|
int new_fd = phys_mapping(to_phys((void *)new_stack), &new_offset);
|
|
|
|
|
|
|
|
/* prepare offset and fd of child's stack as argument for parent's
|
|
|
|
* and child's mmap2 calls
|
|
|
|
*/
|
|
|
|
*data = ((struct stub_data) { .offset = MMAP_OFFSET(new_offset),
|
|
|
|
.fd = new_fd,
|
|
|
|
.timer = ((struct itimerval)
|
|
|
|
{ { 0, 1000000 / hz() },
|
|
|
|
{ 0, 1000000 / hz() }})});
|
2007-05-07 05:51:48 +08:00
|
|
|
err = ptrace_setregs(pid, thread_regs);
|
2006-01-19 09:42:46 +08:00
|
|
|
if(err < 0)
|
|
|
|
panic("copy_context_skas0 : PTRACE_SETREGS failed, "
|
2006-05-02 03:16:00 +08:00
|
|
|
"pid = %d, errno = %d\n", pid, -err);
|
2006-01-19 09:42:46 +08:00
|
|
|
|
2007-05-07 05:51:48 +08:00
|
|
|
err = ptrace_setfpregs(pid, thread_fp_regs);
|
2006-02-08 04:58:43 +08:00
|
|
|
if(err < 0)
|
|
|
|
panic("copy_context_skas0 : PTRACE_SETFPREGS failed, "
|
2006-05-02 03:16:00 +08:00
|
|
|
"pid = %d, errno = %d\n", pid, -err);
|
2006-02-08 04:58:43 +08:00
|
|
|
|
2006-01-19 09:42:46 +08:00
|
|
|
/* set a well known return code for detection of child write failure */
|
|
|
|
child_data->err = 12345678;
|
|
|
|
|
|
|
|
/* Wait, until parent has finished its work: read child's pid from
|
|
|
|
* parent's stack, and check, if bad result.
|
|
|
|
*/
|
2007-05-07 05:51:48 +08:00
|
|
|
err = ptrace(PTRACE_CONT, pid, 0, 0);
|
|
|
|
if(err)
|
|
|
|
panic("Failed to continue new process, pid = %d, "
|
|
|
|
"errno = %d\n", pid, errno);
|
|
|
|
wait_stub_done(pid);
|
2006-01-19 09:42:46 +08:00
|
|
|
|
|
|
|
pid = data->err;
|
|
|
|
if(pid < 0)
|
|
|
|
panic("copy_context_skas0 - stub-parent reports error %d\n",
|
2006-05-02 03:16:00 +08:00
|
|
|
-pid);
|
2006-01-19 09:42:46 +08:00
|
|
|
|
|
|
|
/* Wait, until child has finished too: read child's result from
|
|
|
|
* child's stack and check it.
|
|
|
|
*/
|
2007-05-07 05:51:48 +08:00
|
|
|
wait_stub_done(pid);
|
2006-01-19 09:42:46 +08:00
|
|
|
if (child_data->err != UML_CONFIG_STUB_DATA)
|
2006-04-11 13:53:32 +08:00
|
|
|
panic("copy_context_skas0 - stub-child reports error %ld\n",
|
2006-01-19 09:42:46 +08:00
|
|
|
child_data->err);
|
|
|
|
|
|
|
|
if (ptrace(PTRACE_OLDSETOPTIONS, pid, NULL,
|
|
|
|
(void *)PTRACE_O_TRACESYSGOOD) < 0)
|
|
|
|
panic("copy_context_skas0 : PTRACE_OLDSETOPTIONS failed, "
|
|
|
|
"errno = %d\n", errno);
|
|
|
|
|
|
|
|
return pid;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This is used only, if stub pages are needed, while proc_mm is
|
2007-05-07 05:51:33 +08:00
|
|
|
* available. Opening /proc/mm creates a new mm_context, which lacks
|
2006-01-19 09:42:46 +08:00
|
|
|
* the stub-pages. Thus, we map them using /proc/mm-fd
|
|
|
|
*/
|
|
|
|
void map_stub_pages(int fd, unsigned long code,
|
|
|
|
unsigned long data, unsigned long stack)
|
|
|
|
{
|
|
|
|
struct proc_mm_op mmop;
|
|
|
|
int n;
|
2006-01-19 09:42:48 +08:00
|
|
|
__u64 code_offset;
|
|
|
|
int code_fd = phys_mapping(to_phys((void *) &__syscall_stub_start),
|
|
|
|
&code_offset);
|
2006-01-19 09:42:46 +08:00
|
|
|
|
|
|
|
mmop = ((struct proc_mm_op) { .op = MM_MMAP,
|
|
|
|
.u =
|
|
|
|
{ .mmap =
|
|
|
|
{ .addr = code,
|
|
|
|
.len = PAGE_SIZE,
|
|
|
|
.prot = PROT_EXEC,
|
|
|
|
.flags = MAP_FIXED | MAP_PRIVATE,
|
2006-01-19 09:42:48 +08:00
|
|
|
.fd = code_fd,
|
|
|
|
.offset = code_offset
|
2006-01-19 09:42:46 +08:00
|
|
|
} } });
|
2007-05-07 05:51:35 +08:00
|
|
|
CATCH_EINTR(n = write(fd, &mmop, sizeof(mmop)));
|
2007-03-06 17:42:19 +08:00
|
|
|
if(n != sizeof(mmop)){
|
2007-05-07 05:51:35 +08:00
|
|
|
n = errno;
|
2007-03-06 17:42:19 +08:00
|
|
|
printk("mmap args - addr = 0x%lx, fd = %d, offset = %llx\n",
|
|
|
|
code, code_fd, (unsigned long long) code_offset);
|
2006-01-19 09:42:46 +08:00
|
|
|
panic("map_stub_pages : /proc/mm map for code failed, "
|
2007-05-07 05:51:35 +08:00
|
|
|
"err = %d\n", n);
|
2007-03-06 17:42:19 +08:00
|
|
|
}
|
2006-01-19 09:42:46 +08:00
|
|
|
|
|
|
|
if ( stack ) {
|
|
|
|
__u64 map_offset;
|
|
|
|
int map_fd = phys_mapping(to_phys((void *)stack), &map_offset);
|
|
|
|
mmop = ((struct proc_mm_op)
|
|
|
|
{ .op = MM_MMAP,
|
|
|
|
.u =
|
|
|
|
{ .mmap =
|
|
|
|
{ .addr = data,
|
|
|
|
.len = PAGE_SIZE,
|
|
|
|
.prot = PROT_READ | PROT_WRITE,
|
|
|
|
.flags = MAP_FIXED | MAP_SHARED,
|
|
|
|
.fd = map_fd,
|
|
|
|
.offset = map_offset
|
|
|
|
} } });
|
2007-05-07 05:51:35 +08:00
|
|
|
CATCH_EINTR(n = write(fd, &mmop, sizeof(mmop)));
|
2006-01-19 09:42:46 +08:00
|
|
|
if(n != sizeof(mmop))
|
|
|
|
panic("map_stub_pages : /proc/mm map for data failed, "
|
2007-05-07 05:51:35 +08:00
|
|
|
"err = %d\n", errno);
|
2006-01-19 09:42:46 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
[PATCH] uml: thread creation tidying
fork on UML has always somewhat subtle. The underlying cause has been the
need to initialize a stack for the new process. The only portable way to
initialize a new stack is to set it as the alternate signal stack and take a
signal. The signal handler does whatever initialization is needed and jumps
back to the original stack, where the fork processing is finished. The basic
context switching mechanism is a jmp_buf for each process. You switch to a
new process by longjmping to its jmp_buf.
Now that UML has its own implementation of setjmp and longjmp, and I can poke
around inside a jmp_buf without fear that libc will change the structure, a
much simpler mechanism is possible. The jmpbuf can simply be initialized by
hand.
This eliminates -
the need to set up and remove the alternate signal stack
sending and handling a signal
the signal blocking needed around the stack switching, since
there is no stack switching
setting up the jmp_buf needed to jump back to the original
stack after the new one is set up
In addition, since jmp_buf is now defined by UML, and not by libc, it can be
embedded in the thread struct. This makes it unnecessary to have it exist on
the stack, where it used to be. It also simplifies interfaces, since the
switch jmp_buf used to be a void * inside the thread struct, and functions
which took it as an argument needed to define a jmp_buf variable and assign it
from the void *.
Signed-off-by: Jeff Dike <jdike@addtoit.com>
Cc: Paolo 'Blaisorblade' Giarrusso <blaisorblade@yahoo.it>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-09-27 16:50:40 +08:00
|
|
|
void new_thread(void *stack, jmp_buf *buf, void (*handler)(void))
|
2006-01-19 09:42:46 +08:00
|
|
|
{
|
[PATCH] uml: thread creation tidying
fork on UML has always somewhat subtle. The underlying cause has been the
need to initialize a stack for the new process. The only portable way to
initialize a new stack is to set it as the alternate signal stack and take a
signal. The signal handler does whatever initialization is needed and jumps
back to the original stack, where the fork processing is finished. The basic
context switching mechanism is a jmp_buf for each process. You switch to a
new process by longjmping to its jmp_buf.
Now that UML has its own implementation of setjmp and longjmp, and I can poke
around inside a jmp_buf without fear that libc will change the structure, a
much simpler mechanism is possible. The jmpbuf can simply be initialized by
hand.
This eliminates -
the need to set up and remove the alternate signal stack
sending and handling a signal
the signal blocking needed around the stack switching, since
there is no stack switching
setting up the jmp_buf needed to jump back to the original
stack after the new one is set up
In addition, since jmp_buf is now defined by UML, and not by libc, it can be
embedded in the thread struct. This makes it unnecessary to have it exist on
the stack, where it used to be. It also simplifies interfaces, since the
switch jmp_buf used to be a void * inside the thread struct, and functions
which took it as an argument needed to define a jmp_buf variable and assign it
from the void *.
Signed-off-by: Jeff Dike <jdike@addtoit.com>
Cc: Paolo 'Blaisorblade' Giarrusso <blaisorblade@yahoo.it>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-09-27 16:50:40 +08:00
|
|
|
(*buf)[0].JB_IP = (unsigned long) handler;
|
|
|
|
(*buf)[0].JB_SP = (unsigned long) stack +
|
|
|
|
(PAGE_SIZE << UML_CONFIG_KERNEL_STACK_ORDER) - sizeof(void *);
|
2006-01-19 09:42:46 +08:00
|
|
|
}
|
|
|
|
|
2006-02-08 04:58:43 +08:00
|
|
|
#define INIT_JMP_NEW_THREAD 0
|
[PATCH] uml: thread creation tidying
fork on UML has always somewhat subtle. The underlying cause has been the
need to initialize a stack for the new process. The only portable way to
initialize a new stack is to set it as the alternate signal stack and take a
signal. The signal handler does whatever initialization is needed and jumps
back to the original stack, where the fork processing is finished. The basic
context switching mechanism is a jmp_buf for each process. You switch to a
new process by longjmping to its jmp_buf.
Now that UML has its own implementation of setjmp and longjmp, and I can poke
around inside a jmp_buf without fear that libc will change the structure, a
much simpler mechanism is possible. The jmpbuf can simply be initialized by
hand.
This eliminates -
the need to set up and remove the alternate signal stack
sending and handling a signal
the signal blocking needed around the stack switching, since
there is no stack switching
setting up the jmp_buf needed to jump back to the original
stack after the new one is set up
In addition, since jmp_buf is now defined by UML, and not by libc, it can be
embedded in the thread struct. This makes it unnecessary to have it exist on
the stack, where it used to be. It also simplifies interfaces, since the
switch jmp_buf used to be a void * inside the thread struct, and functions
which took it as an argument needed to define a jmp_buf variable and assign it
from the void *.
Signed-off-by: Jeff Dike <jdike@addtoit.com>
Cc: Paolo 'Blaisorblade' Giarrusso <blaisorblade@yahoo.it>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-09-27 16:50:40 +08:00
|
|
|
#define INIT_JMP_CALLBACK 1
|
|
|
|
#define INIT_JMP_HALT 2
|
|
|
|
#define INIT_JMP_REBOOT 3
|
2006-01-19 09:42:46 +08:00
|
|
|
|
[PATCH] uml: thread creation tidying
fork on UML has always somewhat subtle. The underlying cause has been the
need to initialize a stack for the new process. The only portable way to
initialize a new stack is to set it as the alternate signal stack and take a
signal. The signal handler does whatever initialization is needed and jumps
back to the original stack, where the fork processing is finished. The basic
context switching mechanism is a jmp_buf for each process. You switch to a
new process by longjmping to its jmp_buf.
Now that UML has its own implementation of setjmp and longjmp, and I can poke
around inside a jmp_buf without fear that libc will change the structure, a
much simpler mechanism is possible. The jmpbuf can simply be initialized by
hand.
This eliminates -
the need to set up and remove the alternate signal stack
sending and handling a signal
the signal blocking needed around the stack switching, since
there is no stack switching
setting up the jmp_buf needed to jump back to the original
stack after the new one is set up
In addition, since jmp_buf is now defined by UML, and not by libc, it can be
embedded in the thread struct. This makes it unnecessary to have it exist on
the stack, where it used to be. It also simplifies interfaces, since the
switch jmp_buf used to be a void * inside the thread struct, and functions
which took it as an argument needed to define a jmp_buf variable and assign it
from the void *.
Signed-off-by: Jeff Dike <jdike@addtoit.com>
Cc: Paolo 'Blaisorblade' Giarrusso <blaisorblade@yahoo.it>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-09-27 16:50:40 +08:00
|
|
|
void switch_threads(jmp_buf *me, jmp_buf *you)
|
2006-01-19 09:42:46 +08:00
|
|
|
{
|
[PATCH] uml: thread creation tidying
fork on UML has always somewhat subtle. The underlying cause has been the
need to initialize a stack for the new process. The only portable way to
initialize a new stack is to set it as the alternate signal stack and take a
signal. The signal handler does whatever initialization is needed and jumps
back to the original stack, where the fork processing is finished. The basic
context switching mechanism is a jmp_buf for each process. You switch to a
new process by longjmping to its jmp_buf.
Now that UML has its own implementation of setjmp and longjmp, and I can poke
around inside a jmp_buf without fear that libc will change the structure, a
much simpler mechanism is possible. The jmpbuf can simply be initialized by
hand.
This eliminates -
the need to set up and remove the alternate signal stack
sending and handling a signal
the signal blocking needed around the stack switching, since
there is no stack switching
setting up the jmp_buf needed to jump back to the original
stack after the new one is set up
In addition, since jmp_buf is now defined by UML, and not by libc, it can be
embedded in the thread struct. This makes it unnecessary to have it exist on
the stack, where it used to be. It also simplifies interfaces, since the
switch jmp_buf used to be a void * inside the thread struct, and functions
which took it as an argument needed to define a jmp_buf variable and assign it
from the void *.
Signed-off-by: Jeff Dike <jdike@addtoit.com>
Cc: Paolo 'Blaisorblade' Giarrusso <blaisorblade@yahoo.it>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-09-27 16:50:40 +08:00
|
|
|
if(UML_SETJMP(me) == 0)
|
|
|
|
UML_LONGJMP(you, 1);
|
2006-01-19 09:42:46 +08:00
|
|
|
}
|
|
|
|
|
2006-04-19 13:21:41 +08:00
|
|
|
static jmp_buf initial_jmpbuf;
|
2006-01-19 09:42:46 +08:00
|
|
|
|
|
|
|
/* XXX Make these percpu */
|
|
|
|
static void (*cb_proc)(void *arg);
|
|
|
|
static void *cb_arg;
|
2006-04-19 13:21:41 +08:00
|
|
|
static jmp_buf *cb_back;
|
2006-01-19 09:42:46 +08:00
|
|
|
|
[PATCH] uml: thread creation tidying
fork on UML has always somewhat subtle. The underlying cause has been the
need to initialize a stack for the new process. The only portable way to
initialize a new stack is to set it as the alternate signal stack and take a
signal. The signal handler does whatever initialization is needed and jumps
back to the original stack, where the fork processing is finished. The basic
context switching mechanism is a jmp_buf for each process. You switch to a
new process by longjmping to its jmp_buf.
Now that UML has its own implementation of setjmp and longjmp, and I can poke
around inside a jmp_buf without fear that libc will change the structure, a
much simpler mechanism is possible. The jmpbuf can simply be initialized by
hand.
This eliminates -
the need to set up and remove the alternate signal stack
sending and handling a signal
the signal blocking needed around the stack switching, since
there is no stack switching
setting up the jmp_buf needed to jump back to the original
stack after the new one is set up
In addition, since jmp_buf is now defined by UML, and not by libc, it can be
embedded in the thread struct. This makes it unnecessary to have it exist on
the stack, where it used to be. It also simplifies interfaces, since the
switch jmp_buf used to be a void * inside the thread struct, and functions
which took it as an argument needed to define a jmp_buf variable and assign it
from the void *.
Signed-off-by: Jeff Dike <jdike@addtoit.com>
Cc: Paolo 'Blaisorblade' Giarrusso <blaisorblade@yahoo.it>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-09-27 16:50:40 +08:00
|
|
|
int start_idle_thread(void *stack, jmp_buf *switch_buf)
|
2006-01-19 09:42:46 +08:00
|
|
|
{
|
2006-07-14 15:24:02 +08:00
|
|
|
int n;
|
2006-01-19 09:42:46 +08:00
|
|
|
|
|
|
|
set_handler(SIGWINCH, (__sighandler_t) sig_handler,
|
|
|
|
SA_ONSTACK | SA_RESTART, SIGUSR1, SIGIO, SIGALRM,
|
|
|
|
SIGVTALRM, -1);
|
|
|
|
|
2007-05-07 05:51:40 +08:00
|
|
|
/*
|
|
|
|
* Can't use UML_SETJMP or UML_LONGJMP here because they save
|
|
|
|
* and restore signals, with the possible side-effect of
|
|
|
|
* trying to handle any signals which came when they were
|
|
|
|
* blocked, which can't be done on this stack.
|
|
|
|
* Signals must be blocked when jumping back here and restored
|
|
|
|
* after returning to the jumper.
|
|
|
|
*/
|
|
|
|
n = setjmp(initial_jmpbuf);
|
2006-01-19 09:42:46 +08:00
|
|
|
switch(n){
|
|
|
|
case INIT_JMP_NEW_THREAD:
|
[PATCH] uml: thread creation tidying
fork on UML has always somewhat subtle. The underlying cause has been the
need to initialize a stack for the new process. The only portable way to
initialize a new stack is to set it as the alternate signal stack and take a
signal. The signal handler does whatever initialization is needed and jumps
back to the original stack, where the fork processing is finished. The basic
context switching mechanism is a jmp_buf for each process. You switch to a
new process by longjmping to its jmp_buf.
Now that UML has its own implementation of setjmp and longjmp, and I can poke
around inside a jmp_buf without fear that libc will change the structure, a
much simpler mechanism is possible. The jmpbuf can simply be initialized by
hand.
This eliminates -
the need to set up and remove the alternate signal stack
sending and handling a signal
the signal blocking needed around the stack switching, since
there is no stack switching
setting up the jmp_buf needed to jump back to the original
stack after the new one is set up
In addition, since jmp_buf is now defined by UML, and not by libc, it can be
embedded in the thread struct. This makes it unnecessary to have it exist on
the stack, where it used to be. It also simplifies interfaces, since the
switch jmp_buf used to be a void * inside the thread struct, and functions
which took it as an argument needed to define a jmp_buf variable and assign it
from the void *.
Signed-off-by: Jeff Dike <jdike@addtoit.com>
Cc: Paolo 'Blaisorblade' Giarrusso <blaisorblade@yahoo.it>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-09-27 16:50:40 +08:00
|
|
|
(*switch_buf)[0].JB_IP = (unsigned long) new_thread_handler;
|
|
|
|
(*switch_buf)[0].JB_SP = (unsigned long) stack +
|
|
|
|
(PAGE_SIZE << UML_CONFIG_KERNEL_STACK_ORDER) -
|
|
|
|
sizeof(void *);
|
2006-01-19 09:42:46 +08:00
|
|
|
break;
|
|
|
|
case INIT_JMP_CALLBACK:
|
|
|
|
(*cb_proc)(cb_arg);
|
2007-05-07 05:51:40 +08:00
|
|
|
longjmp(*cb_back, 1);
|
2006-01-19 09:42:46 +08:00
|
|
|
break;
|
|
|
|
case INIT_JMP_HALT:
|
|
|
|
kmalloc_ok = 0;
|
|
|
|
return(0);
|
|
|
|
case INIT_JMP_REBOOT:
|
|
|
|
kmalloc_ok = 0;
|
|
|
|
return(1);
|
|
|
|
default:
|
|
|
|
panic("Bad sigsetjmp return in start_idle_thread - %d\n", n);
|
|
|
|
}
|
2007-05-07 05:51:40 +08:00
|
|
|
longjmp(*switch_buf, 1);
|
2006-01-19 09:42:46 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
void initial_thread_cb_skas(void (*proc)(void *), void *arg)
|
|
|
|
{
|
2006-04-19 13:21:41 +08:00
|
|
|
jmp_buf here;
|
2006-01-19 09:42:46 +08:00
|
|
|
|
|
|
|
cb_proc = proc;
|
|
|
|
cb_arg = arg;
|
|
|
|
cb_back = &here;
|
|
|
|
|
|
|
|
block_signals();
|
2006-07-14 15:24:02 +08:00
|
|
|
if(UML_SETJMP(&here) == 0)
|
2006-04-19 13:21:41 +08:00
|
|
|
UML_LONGJMP(&initial_jmpbuf, INIT_JMP_CALLBACK);
|
2006-01-19 09:42:46 +08:00
|
|
|
unblock_signals();
|
|
|
|
|
|
|
|
cb_proc = NULL;
|
|
|
|
cb_arg = NULL;
|
|
|
|
cb_back = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
void halt_skas(void)
|
|
|
|
{
|
|
|
|
block_signals();
|
2006-04-19 13:21:41 +08:00
|
|
|
UML_LONGJMP(&initial_jmpbuf, INIT_JMP_HALT);
|
2006-01-19 09:42:46 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
void reboot_skas(void)
|
|
|
|
{
|
|
|
|
block_signals();
|
2006-04-19 13:21:41 +08:00
|
|
|
UML_LONGJMP(&initial_jmpbuf, INIT_JMP_REBOOT);
|
2006-01-19 09:42:46 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
void switch_mm_skas(struct mm_id *mm_idp)
|
|
|
|
{
|
|
|
|
int err;
|
|
|
|
|
2007-05-09 17:33:22 +08:00
|
|
|
/* FIXME: need cpu pid in switch_mm_skas */
|
2006-01-19 09:42:46 +08:00
|
|
|
if(proc_mm){
|
|
|
|
err = ptrace(PTRACE_SWITCH_MM, userspace_pid[0], 0,
|
|
|
|
mm_idp->u.mm_fd);
|
|
|
|
if(err)
|
|
|
|
panic("switch_mm_skas - PTRACE_SWITCH_MM failed, "
|
|
|
|
"errno = %d\n", errno);
|
|
|
|
}
|
|
|
|
else userspace_pid[0] = mm_idp->u.pid;
|
|
|
|
}
|