2005-04-17 05:20:36 +07:00
|
|
|
/*
|
|
|
|
* linux/arch/arm/kernel/signal.c
|
|
|
|
*
|
|
|
|
* Copyright (C) 1995-2002 Russell King
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or modify
|
|
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
|
|
* published by the Free Software Foundation.
|
|
|
|
*/
|
|
|
|
#include <linux/errno.h>
|
|
|
|
#include <linux/signal.h>
|
|
|
|
#include <linux/ptrace.h>
|
|
|
|
#include <linux/personality.h>
|
2006-12-07 11:34:23 +07:00
|
|
|
#include <linux/freezer.h>
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2006-11-09 21:20:47 +07:00
|
|
|
#include <asm/elf.h>
|
2005-04-17 05:20:36 +07:00
|
|
|
#include <asm/cacheflush.h>
|
|
|
|
#include <asm/ucontext.h>
|
|
|
|
#include <asm/uaccess.h>
|
|
|
|
#include <asm/unistd.h>
|
|
|
|
|
|
|
|
#include "ptrace.h"
|
2005-06-23 02:26:05 +07:00
|
|
|
#include "signal.h"
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
#define _BLOCKABLE (~(sigmask(SIGKILL) | sigmask(SIGSTOP)))
|
|
|
|
|
|
|
|
/*
|
|
|
|
* For ARM syscalls, we encode the syscall number into the instruction.
|
|
|
|
*/
|
|
|
|
#define SWI_SYS_SIGRETURN (0xef000000|(__NR_sigreturn))
|
|
|
|
#define SWI_SYS_RT_SIGRETURN (0xef000000|(__NR_rt_sigreturn))
|
|
|
|
|
2006-01-19 05:38:47 +07:00
|
|
|
/*
|
|
|
|
* With EABI, the syscall number has to be loaded into r7.
|
|
|
|
*/
|
|
|
|
#define MOV_R7_NR_SIGRETURN (0xe3a07000 | (__NR_sigreturn - __NR_SYSCALL_BASE))
|
|
|
|
#define MOV_R7_NR_RT_SIGRETURN (0xe3a07000 | (__NR_rt_sigreturn - __NR_SYSCALL_BASE))
|
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
/*
|
|
|
|
* For Thumb syscalls, we pass the syscall number via r7. We therefore
|
|
|
|
* need two 16-bit instructions.
|
|
|
|
*/
|
|
|
|
#define SWI_THUMB_SIGRETURN (0xdf00 << 16 | 0x2700 | (__NR_sigreturn - __NR_SYSCALL_BASE))
|
|
|
|
#define SWI_THUMB_RT_SIGRETURN (0xdf00 << 16 | 0x2700 | (__NR_rt_sigreturn - __NR_SYSCALL_BASE))
|
|
|
|
|
2006-01-19 05:38:47 +07:00
|
|
|
const unsigned long sigreturn_codes[7] = {
|
|
|
|
MOV_R7_NR_SIGRETURN, SWI_SYS_SIGRETURN, SWI_THUMB_SIGRETURN,
|
|
|
|
MOV_R7_NR_RT_SIGRETURN, SWI_SYS_RT_SIGRETURN, SWI_THUMB_RT_SIGRETURN,
|
2005-04-17 05:20:36 +07:00
|
|
|
};
|
|
|
|
|
|
|
|
static int do_signal(sigset_t *oldset, struct pt_regs * regs, int syscall);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* atomically swap in the new signal mask, and wait for a signal.
|
|
|
|
*/
|
|
|
|
asmlinkage int sys_sigsuspend(int restart, unsigned long oldmask, old_sigset_t mask, struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
sigset_t saveset;
|
|
|
|
|
|
|
|
mask &= _BLOCKABLE;
|
|
|
|
spin_lock_irq(¤t->sighand->siglock);
|
|
|
|
saveset = current->blocked;
|
|
|
|
siginitset(¤t->blocked, mask);
|
|
|
|
recalc_sigpending();
|
|
|
|
spin_unlock_irq(¤t->sighand->siglock);
|
|
|
|
regs->ARM_r0 = -EINTR;
|
|
|
|
|
|
|
|
while (1) {
|
|
|
|
current->state = TASK_INTERRUPTIBLE;
|
|
|
|
schedule();
|
|
|
|
if (do_signal(&saveset, regs, 0))
|
|
|
|
return regs->ARM_r0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
asmlinkage int
|
|
|
|
sys_rt_sigsuspend(sigset_t __user *unewset, size_t sigsetsize, struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
sigset_t saveset, newset;
|
|
|
|
|
|
|
|
/* XXX: Don't preclude handling different sized sigset_t's. */
|
|
|
|
if (sigsetsize != sizeof(sigset_t))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (copy_from_user(&newset, unewset, sizeof(newset)))
|
|
|
|
return -EFAULT;
|
|
|
|
sigdelsetmask(&newset, ~_BLOCKABLE);
|
|
|
|
|
|
|
|
spin_lock_irq(¤t->sighand->siglock);
|
|
|
|
saveset = current->blocked;
|
|
|
|
current->blocked = newset;
|
|
|
|
recalc_sigpending();
|
|
|
|
spin_unlock_irq(¤t->sighand->siglock);
|
|
|
|
regs->ARM_r0 = -EINTR;
|
|
|
|
|
|
|
|
while (1) {
|
|
|
|
current->state = TASK_INTERRUPTIBLE;
|
|
|
|
schedule();
|
|
|
|
if (do_signal(&saveset, regs, 0))
|
|
|
|
return regs->ARM_r0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
asmlinkage int
|
|
|
|
sys_sigaction(int sig, const struct old_sigaction __user *act,
|
|
|
|
struct old_sigaction __user *oact)
|
|
|
|
{
|
|
|
|
struct k_sigaction new_ka, old_ka;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (act) {
|
|
|
|
old_sigset_t mask;
|
|
|
|
if (!access_ok(VERIFY_READ, act, sizeof(*act)) ||
|
|
|
|
__get_user(new_ka.sa.sa_handler, &act->sa_handler) ||
|
|
|
|
__get_user(new_ka.sa.sa_restorer, &act->sa_restorer))
|
|
|
|
return -EFAULT;
|
|
|
|
__get_user(new_ka.sa.sa_flags, &act->sa_flags);
|
|
|
|
__get_user(mask, &act->sa_mask);
|
|
|
|
siginitset(&new_ka.sa.sa_mask, mask);
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = do_sigaction(sig, act ? &new_ka : NULL, oact ? &old_ka : NULL);
|
|
|
|
|
|
|
|
if (!ret && oact) {
|
|
|
|
if (!access_ok(VERIFY_WRITE, oact, sizeof(*oact)) ||
|
|
|
|
__put_user(old_ka.sa.sa_handler, &oact->sa_handler) ||
|
|
|
|
__put_user(old_ka.sa.sa_restorer, &oact->sa_restorer))
|
|
|
|
return -EFAULT;
|
|
|
|
__put_user(old_ka.sa.sa_flags, &oact->sa_flags);
|
|
|
|
__put_user(old_ka.sa.sa_mask.sig[0], &oact->sa_mask);
|
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2006-06-28 04:56:18 +07:00
|
|
|
#ifdef CONFIG_CRUNCH
|
|
|
|
static int preserve_crunch_context(struct crunch_sigframe *frame)
|
|
|
|
{
|
|
|
|
char kbuf[sizeof(*frame) + 8];
|
|
|
|
struct crunch_sigframe *kframe;
|
|
|
|
|
|
|
|
/* the crunch context must be 64 bit aligned */
|
|
|
|
kframe = (struct crunch_sigframe *)((unsigned long)(kbuf + 8) & ~7);
|
|
|
|
kframe->magic = CRUNCH_MAGIC;
|
|
|
|
kframe->size = CRUNCH_STORAGE_SIZE;
|
|
|
|
crunch_task_copy(current_thread_info(), &kframe->storage);
|
|
|
|
return __copy_to_user(frame, kframe, sizeof(*frame));
|
|
|
|
}
|
|
|
|
|
|
|
|
static int restore_crunch_context(struct crunch_sigframe *frame)
|
|
|
|
{
|
|
|
|
char kbuf[sizeof(*frame) + 8];
|
|
|
|
struct crunch_sigframe *kframe;
|
|
|
|
|
|
|
|
/* the crunch context must be 64 bit aligned */
|
|
|
|
kframe = (struct crunch_sigframe *)((unsigned long)(kbuf + 8) & ~7);
|
|
|
|
if (__copy_from_user(kframe, frame, sizeof(*frame)))
|
|
|
|
return -1;
|
|
|
|
if (kframe->magic != CRUNCH_MAGIC ||
|
|
|
|
kframe->size != CRUNCH_STORAGE_SIZE)
|
|
|
|
return -1;
|
|
|
|
crunch_task_restore(current_thread_info(), &kframe->storage);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
#ifdef CONFIG_IWMMXT
|
|
|
|
|
|
|
|
static int preserve_iwmmxt_context(struct iwmmxt_sigframe *frame)
|
|
|
|
{
|
[PATCH] mm: arm ready for split ptlock
Prepare arm for the split page_table_lock: three issues.
Signal handling's preserve and restore of iwmmxt context currently involves
reading and writing that context to and from user space, while holding
page_table_lock to secure the user page(s) against kswapd. If we split the
lock, then the structure might span two pages, secured by to read into and
write from a kernel stack buffer, copying that out and in without locking (the
structure is 160 bytes in size, and here we're near the top of the kernel
stack). Or would the overhead be noticeable?
arm_syscall's cmpxchg emulation use pte_offset_map_lock, instead of
pte_offset_map and mm-wide page_table_lock; and strictly, it should now also
take mmap_sem before descending to pmd, to guard against another thread
munmapping, and the page table pulled out beneath this thread.
Updated two comments in fault-armv.c. adjust_pte is interesting, since its
modification of a pte in one part of the mm depends on the lock held when
calling update_mmu_cache for a pte in some other part of that mm. This can't
be done with a split page_table_lock (and we've already taken the lowest lock
in the hierarchy here): so we'll have to disable split on arm, unless
CONFIG_CPU_CACHE_VIPT to ensures adjust_pte never used.
Signed-off-by: Hugh Dickins <hugh@veritas.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2005-10-30 08:16:36 +07:00
|
|
|
char kbuf[sizeof(*frame) + 8];
|
|
|
|
struct iwmmxt_sigframe *kframe;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
/* the iWMMXt context must be 64 bit aligned */
|
[PATCH] mm: arm ready for split ptlock
Prepare arm for the split page_table_lock: three issues.
Signal handling's preserve and restore of iwmmxt context currently involves
reading and writing that context to and from user space, while holding
page_table_lock to secure the user page(s) against kswapd. If we split the
lock, then the structure might span two pages, secured by to read into and
write from a kernel stack buffer, copying that out and in without locking (the
structure is 160 bytes in size, and here we're near the top of the kernel
stack). Or would the overhead be noticeable?
arm_syscall's cmpxchg emulation use pte_offset_map_lock, instead of
pte_offset_map and mm-wide page_table_lock; and strictly, it should now also
take mmap_sem before descending to pmd, to guard against another thread
munmapping, and the page table pulled out beneath this thread.
Updated two comments in fault-armv.c. adjust_pte is interesting, since its
modification of a pte in one part of the mm depends on the lock held when
calling update_mmu_cache for a pte in some other part of that mm. This can't
be done with a split page_table_lock (and we've already taken the lowest lock
in the hierarchy here): so we'll have to disable split on arm, unless
CONFIG_CPU_CACHE_VIPT to ensures adjust_pte never used.
Signed-off-by: Hugh Dickins <hugh@veritas.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2005-10-30 08:16:36 +07:00
|
|
|
kframe = (struct iwmmxt_sigframe *)((unsigned long)(kbuf + 8) & ~7);
|
2006-06-25 05:46:21 +07:00
|
|
|
kframe->magic = IWMMXT_MAGIC;
|
|
|
|
kframe->size = IWMMXT_STORAGE_SIZE;
|
[PATCH] mm: arm ready for split ptlock
Prepare arm for the split page_table_lock: three issues.
Signal handling's preserve and restore of iwmmxt context currently involves
reading and writing that context to and from user space, while holding
page_table_lock to secure the user page(s) against kswapd. If we split the
lock, then the structure might span two pages, secured by to read into and
write from a kernel stack buffer, copying that out and in without locking (the
structure is 160 bytes in size, and here we're near the top of the kernel
stack). Or would the overhead be noticeable?
arm_syscall's cmpxchg emulation use pte_offset_map_lock, instead of
pte_offset_map and mm-wide page_table_lock; and strictly, it should now also
take mmap_sem before descending to pmd, to guard against another thread
munmapping, and the page table pulled out beneath this thread.
Updated two comments in fault-armv.c. adjust_pte is interesting, since its
modification of a pte in one part of the mm depends on the lock held when
calling update_mmu_cache for a pte in some other part of that mm. This can't
be done with a split page_table_lock (and we've already taken the lowest lock
in the hierarchy here): so we'll have to disable split on arm, unless
CONFIG_CPU_CACHE_VIPT to ensures adjust_pte never used.
Signed-off-by: Hugh Dickins <hugh@veritas.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2005-10-30 08:16:36 +07:00
|
|
|
iwmmxt_task_copy(current_thread_info(), &kframe->storage);
|
|
|
|
return __copy_to_user(frame, kframe, sizeof(*frame));
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static int restore_iwmmxt_context(struct iwmmxt_sigframe *frame)
|
|
|
|
{
|
[PATCH] mm: arm ready for split ptlock
Prepare arm for the split page_table_lock: three issues.
Signal handling's preserve and restore of iwmmxt context currently involves
reading and writing that context to and from user space, while holding
page_table_lock to secure the user page(s) against kswapd. If we split the
lock, then the structure might span two pages, secured by to read into and
write from a kernel stack buffer, copying that out and in without locking (the
structure is 160 bytes in size, and here we're near the top of the kernel
stack). Or would the overhead be noticeable?
arm_syscall's cmpxchg emulation use pte_offset_map_lock, instead of
pte_offset_map and mm-wide page_table_lock; and strictly, it should now also
take mmap_sem before descending to pmd, to guard against another thread
munmapping, and the page table pulled out beneath this thread.
Updated two comments in fault-armv.c. adjust_pte is interesting, since its
modification of a pte in one part of the mm depends on the lock held when
calling update_mmu_cache for a pte in some other part of that mm. This can't
be done with a split page_table_lock (and we've already taken the lowest lock
in the hierarchy here): so we'll have to disable split on arm, unless
CONFIG_CPU_CACHE_VIPT to ensures adjust_pte never used.
Signed-off-by: Hugh Dickins <hugh@veritas.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2005-10-30 08:16:36 +07:00
|
|
|
char kbuf[sizeof(*frame) + 8];
|
|
|
|
struct iwmmxt_sigframe *kframe;
|
|
|
|
|
|
|
|
/* the iWMMXt context must be 64 bit aligned */
|
|
|
|
kframe = (struct iwmmxt_sigframe *)((unsigned long)(kbuf + 8) & ~7);
|
|
|
|
if (__copy_from_user(kframe, frame, sizeof(*frame)))
|
|
|
|
return -1;
|
2006-06-25 05:46:21 +07:00
|
|
|
if (kframe->magic != IWMMXT_MAGIC ||
|
|
|
|
kframe->size != IWMMXT_STORAGE_SIZE)
|
[PATCH] mm: arm ready for split ptlock
Prepare arm for the split page_table_lock: three issues.
Signal handling's preserve and restore of iwmmxt context currently involves
reading and writing that context to and from user space, while holding
page_table_lock to secure the user page(s) against kswapd. If we split the
lock, then the structure might span two pages, secured by to read into and
write from a kernel stack buffer, copying that out and in without locking (the
structure is 160 bytes in size, and here we're near the top of the kernel
stack). Or would the overhead be noticeable?
arm_syscall's cmpxchg emulation use pte_offset_map_lock, instead of
pte_offset_map and mm-wide page_table_lock; and strictly, it should now also
take mmap_sem before descending to pmd, to guard against another thread
munmapping, and the page table pulled out beneath this thread.
Updated two comments in fault-armv.c. adjust_pte is interesting, since its
modification of a pte in one part of the mm depends on the lock held when
calling update_mmu_cache for a pte in some other part of that mm. This can't
be done with a split page_table_lock (and we've already taken the lowest lock
in the hierarchy here): so we'll have to disable split on arm, unless
CONFIG_CPU_CACHE_VIPT to ensures adjust_pte never used.
Signed-off-by: Hugh Dickins <hugh@veritas.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2005-10-30 08:16:36 +07:00
|
|
|
return -1;
|
|
|
|
iwmmxt_task_restore(current_thread_info(), &kframe->storage);
|
|
|
|
return 0;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Do a signal return; undo the signal stack. These are aligned to 64-bit.
|
|
|
|
*/
|
|
|
|
struct sigframe {
|
2006-06-16 02:13:25 +07:00
|
|
|
struct ucontext uc;
|
2006-01-19 05:38:47 +07:00
|
|
|
unsigned long retcode[2];
|
2005-04-17 05:20:36 +07:00
|
|
|
};
|
|
|
|
|
|
|
|
struct rt_sigframe {
|
|
|
|
struct siginfo info;
|
2006-06-16 02:18:25 +07:00
|
|
|
struct sigframe sig;
|
2005-04-17 05:20:36 +07:00
|
|
|
};
|
|
|
|
|
2006-06-16 02:23:02 +07:00
|
|
|
static int restore_sigframe(struct pt_regs *regs, struct sigframe __user *sf)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
2006-06-25 05:46:21 +07:00
|
|
|
struct aux_sigframe __user *aux;
|
2006-06-16 02:23:02 +07:00
|
|
|
sigset_t set;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
err = __copy_from_user(&set, &sf->uc.uc_sigmask, sizeof(set));
|
|
|
|
if (err == 0) {
|
|
|
|
sigdelsetmask(&set, ~_BLOCKABLE);
|
|
|
|
spin_lock_irq(¤t->sighand->siglock);
|
|
|
|
current->blocked = set;
|
|
|
|
recalc_sigpending();
|
|
|
|
spin_unlock_irq(¤t->sighand->siglock);
|
|
|
|
}
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2006-06-16 02:23:02 +07:00
|
|
|
__get_user_error(regs->ARM_r0, &sf->uc.uc_mcontext.arm_r0, err);
|
|
|
|
__get_user_error(regs->ARM_r1, &sf->uc.uc_mcontext.arm_r1, err);
|
|
|
|
__get_user_error(regs->ARM_r2, &sf->uc.uc_mcontext.arm_r2, err);
|
|
|
|
__get_user_error(regs->ARM_r3, &sf->uc.uc_mcontext.arm_r3, err);
|
|
|
|
__get_user_error(regs->ARM_r4, &sf->uc.uc_mcontext.arm_r4, err);
|
|
|
|
__get_user_error(regs->ARM_r5, &sf->uc.uc_mcontext.arm_r5, err);
|
|
|
|
__get_user_error(regs->ARM_r6, &sf->uc.uc_mcontext.arm_r6, err);
|
|
|
|
__get_user_error(regs->ARM_r7, &sf->uc.uc_mcontext.arm_r7, err);
|
|
|
|
__get_user_error(regs->ARM_r8, &sf->uc.uc_mcontext.arm_r8, err);
|
|
|
|
__get_user_error(regs->ARM_r9, &sf->uc.uc_mcontext.arm_r9, err);
|
|
|
|
__get_user_error(regs->ARM_r10, &sf->uc.uc_mcontext.arm_r10, err);
|
|
|
|
__get_user_error(regs->ARM_fp, &sf->uc.uc_mcontext.arm_fp, err);
|
|
|
|
__get_user_error(regs->ARM_ip, &sf->uc.uc_mcontext.arm_ip, err);
|
|
|
|
__get_user_error(regs->ARM_sp, &sf->uc.uc_mcontext.arm_sp, err);
|
|
|
|
__get_user_error(regs->ARM_lr, &sf->uc.uc_mcontext.arm_lr, err);
|
|
|
|
__get_user_error(regs->ARM_pc, &sf->uc.uc_mcontext.arm_pc, err);
|
|
|
|
__get_user_error(regs->ARM_cpsr, &sf->uc.uc_mcontext.arm_cpsr, err);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
err |= !valid_user_regs(regs);
|
|
|
|
|
2006-06-25 05:46:21 +07:00
|
|
|
aux = (struct aux_sigframe __user *) sf->uc.uc_regspace;
|
2006-06-28 04:56:18 +07:00
|
|
|
#ifdef CONFIG_CRUNCH
|
|
|
|
if (err == 0)
|
|
|
|
err |= restore_crunch_context(&aux->crunch);
|
|
|
|
#endif
|
2005-04-17 05:20:36 +07:00
|
|
|
#ifdef CONFIG_IWMMXT
|
|
|
|
if (err == 0 && test_thread_flag(TIF_USING_IWMMXT))
|
|
|
|
err |= restore_iwmmxt_context(&aux->iwmmxt);
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_VFP
|
|
|
|
// if (err == 0)
|
2006-06-16 02:23:02 +07:00
|
|
|
// err |= vfp_restore_state(&sf->aux.vfp);
|
2005-04-17 05:20:36 +07:00
|
|
|
#endif
|
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
asmlinkage int sys_sigreturn(struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
struct sigframe __user *frame;
|
|
|
|
|
|
|
|
/* Always make any pending restarted system calls return -EINTR */
|
|
|
|
current_thread_info()->restart_block.fn = do_no_restart_syscall;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Since we stacked the signal on a 64-bit boundary,
|
|
|
|
* then 'sp' should be word aligned here. If it's
|
|
|
|
* not, then the user is trying to mess with us.
|
|
|
|
*/
|
|
|
|
if (regs->ARM_sp & 7)
|
|
|
|
goto badframe;
|
|
|
|
|
|
|
|
frame = (struct sigframe __user *)regs->ARM_sp;
|
|
|
|
|
|
|
|
if (!access_ok(VERIFY_READ, frame, sizeof (*frame)))
|
|
|
|
goto badframe;
|
|
|
|
|
2006-06-16 02:23:02 +07:00
|
|
|
if (restore_sigframe(regs, frame))
|
2005-04-17 05:20:36 +07:00
|
|
|
goto badframe;
|
|
|
|
|
|
|
|
/* Send SIGTRAP if we're single-stepping */
|
|
|
|
if (current->ptrace & PT_SINGLESTEP) {
|
|
|
|
ptrace_cancel_bpt(current);
|
|
|
|
send_sig(SIGTRAP, current, 1);
|
|
|
|
}
|
|
|
|
|
|
|
|
return regs->ARM_r0;
|
|
|
|
|
|
|
|
badframe:
|
|
|
|
force_sig(SIGSEGV, current);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
asmlinkage int sys_rt_sigreturn(struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
struct rt_sigframe __user *frame;
|
|
|
|
|
|
|
|
/* Always make any pending restarted system calls return -EINTR */
|
|
|
|
current_thread_info()->restart_block.fn = do_no_restart_syscall;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Since we stacked the signal on a 64-bit boundary,
|
|
|
|
* then 'sp' should be word aligned here. If it's
|
|
|
|
* not, then the user is trying to mess with us.
|
|
|
|
*/
|
|
|
|
if (regs->ARM_sp & 7)
|
|
|
|
goto badframe;
|
|
|
|
|
|
|
|
frame = (struct rt_sigframe __user *)regs->ARM_sp;
|
|
|
|
|
|
|
|
if (!access_ok(VERIFY_READ, frame, sizeof (*frame)))
|
|
|
|
goto badframe;
|
|
|
|
|
2006-06-16 02:23:02 +07:00
|
|
|
if (restore_sigframe(regs, &frame->sig))
|
2005-04-17 05:20:36 +07:00
|
|
|
goto badframe;
|
|
|
|
|
2006-06-16 02:18:25 +07:00
|
|
|
if (do_sigaltstack(&frame->sig.uc.uc_stack, NULL, regs->ARM_sp) == -EFAULT)
|
2005-04-17 05:20:36 +07:00
|
|
|
goto badframe;
|
|
|
|
|
|
|
|
/* Send SIGTRAP if we're single-stepping */
|
|
|
|
if (current->ptrace & PT_SINGLESTEP) {
|
|
|
|
ptrace_cancel_bpt(current);
|
|
|
|
send_sig(SIGTRAP, current, 1);
|
|
|
|
}
|
|
|
|
|
|
|
|
return regs->ARM_r0;
|
|
|
|
|
|
|
|
badframe:
|
|
|
|
force_sig(SIGSEGV, current);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
2006-06-16 02:28:03 +07:00
|
|
|
setup_sigframe(struct sigframe __user *sf, struct pt_regs *regs, sigset_t *set)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
2006-06-25 05:46:21 +07:00
|
|
|
struct aux_sigframe __user *aux;
|
2005-04-17 05:20:36 +07:00
|
|
|
int err = 0;
|
|
|
|
|
2006-06-16 02:28:03 +07:00
|
|
|
__put_user_error(regs->ARM_r0, &sf->uc.uc_mcontext.arm_r0, err);
|
|
|
|
__put_user_error(regs->ARM_r1, &sf->uc.uc_mcontext.arm_r1, err);
|
|
|
|
__put_user_error(regs->ARM_r2, &sf->uc.uc_mcontext.arm_r2, err);
|
|
|
|
__put_user_error(regs->ARM_r3, &sf->uc.uc_mcontext.arm_r3, err);
|
|
|
|
__put_user_error(regs->ARM_r4, &sf->uc.uc_mcontext.arm_r4, err);
|
|
|
|
__put_user_error(regs->ARM_r5, &sf->uc.uc_mcontext.arm_r5, err);
|
|
|
|
__put_user_error(regs->ARM_r6, &sf->uc.uc_mcontext.arm_r6, err);
|
|
|
|
__put_user_error(regs->ARM_r7, &sf->uc.uc_mcontext.arm_r7, err);
|
|
|
|
__put_user_error(regs->ARM_r8, &sf->uc.uc_mcontext.arm_r8, err);
|
|
|
|
__put_user_error(regs->ARM_r9, &sf->uc.uc_mcontext.arm_r9, err);
|
|
|
|
__put_user_error(regs->ARM_r10, &sf->uc.uc_mcontext.arm_r10, err);
|
|
|
|
__put_user_error(regs->ARM_fp, &sf->uc.uc_mcontext.arm_fp, err);
|
|
|
|
__put_user_error(regs->ARM_ip, &sf->uc.uc_mcontext.arm_ip, err);
|
|
|
|
__put_user_error(regs->ARM_sp, &sf->uc.uc_mcontext.arm_sp, err);
|
|
|
|
__put_user_error(regs->ARM_lr, &sf->uc.uc_mcontext.arm_lr, err);
|
|
|
|
__put_user_error(regs->ARM_pc, &sf->uc.uc_mcontext.arm_pc, err);
|
|
|
|
__put_user_error(regs->ARM_cpsr, &sf->uc.uc_mcontext.arm_cpsr, err);
|
|
|
|
|
|
|
|
__put_user_error(current->thread.trap_no, &sf->uc.uc_mcontext.trap_no, err);
|
|
|
|
__put_user_error(current->thread.error_code, &sf->uc.uc_mcontext.error_code, err);
|
|
|
|
__put_user_error(current->thread.address, &sf->uc.uc_mcontext.fault_address, err);
|
|
|
|
__put_user_error(set->sig[0], &sf->uc.uc_mcontext.oldmask, err);
|
|
|
|
|
|
|
|
err |= __copy_to_user(&sf->uc.uc_sigmask, set, sizeof(*set));
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2006-06-25 05:46:21 +07:00
|
|
|
aux = (struct aux_sigframe __user *) sf->uc.uc_regspace;
|
2006-06-28 04:56:18 +07:00
|
|
|
#ifdef CONFIG_CRUNCH
|
|
|
|
if (err == 0)
|
|
|
|
err |= preserve_crunch_context(&aux->crunch);
|
|
|
|
#endif
|
2005-04-17 05:20:36 +07:00
|
|
|
#ifdef CONFIG_IWMMXT
|
|
|
|
if (err == 0 && test_thread_flag(TIF_USING_IWMMXT))
|
|
|
|
err |= preserve_iwmmxt_context(&aux->iwmmxt);
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_VFP
|
|
|
|
// if (err == 0)
|
2006-06-16 02:28:03 +07:00
|
|
|
// err |= vfp_save_state(&sf->aux.vfp);
|
2005-04-17 05:20:36 +07:00
|
|
|
#endif
|
2006-06-25 05:46:21 +07:00
|
|
|
__put_user_error(0, &aux->end_magic, err);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void __user *
|
|
|
|
get_sigframe(struct k_sigaction *ka, struct pt_regs *regs, int framesize)
|
|
|
|
{
|
|
|
|
unsigned long sp = regs->ARM_sp;
|
|
|
|
void __user *frame;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This is the X/Open sanctioned signal stack switching.
|
|
|
|
*/
|
|
|
|
if ((ka->sa.sa_flags & SA_ONSTACK) && !sas_ss_flags(sp))
|
|
|
|
sp = current->sas_ss_sp + current->sas_ss_size;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* ATPCS B01 mandates 8-byte alignment
|
|
|
|
*/
|
|
|
|
frame = (void __user *)((sp - framesize) & ~7);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Check that we can actually write to the signal frame.
|
|
|
|
*/
|
|
|
|
if (!access_ok(VERIFY_WRITE, frame, framesize))
|
|
|
|
frame = NULL;
|
|
|
|
|
|
|
|
return frame;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
setup_return(struct pt_regs *regs, struct k_sigaction *ka,
|
|
|
|
unsigned long __user *rc, void __user *frame, int usig)
|
|
|
|
{
|
|
|
|
unsigned long handler = (unsigned long)ka->sa.sa_handler;
|
|
|
|
unsigned long retcode;
|
|
|
|
int thumb = 0;
|
|
|
|
unsigned long cpsr = regs->ARM_cpsr & ~PSR_f;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Maybe we need to deliver a 32-bit signal to a 26-bit task.
|
|
|
|
*/
|
|
|
|
if (ka->sa.sa_flags & SA_THIRTYTWO)
|
|
|
|
cpsr = (cpsr & ~MODE_MASK) | USR_MODE;
|
|
|
|
|
|
|
|
#ifdef CONFIG_ARM_THUMB
|
|
|
|
if (elf_hwcap & HWCAP_THUMB) {
|
|
|
|
/*
|
|
|
|
* The LSB of the handler determines if we're going to
|
|
|
|
* be using THUMB or ARM mode for this signal handler.
|
|
|
|
*/
|
|
|
|
thumb = handler & 1;
|
|
|
|
|
|
|
|
if (thumb)
|
|
|
|
cpsr |= PSR_T_BIT;
|
|
|
|
else
|
|
|
|
cpsr &= ~PSR_T_BIT;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
if (ka->sa.sa_flags & SA_RESTORER) {
|
|
|
|
retcode = (unsigned long)ka->sa.sa_restorer;
|
|
|
|
} else {
|
2006-01-19 05:38:47 +07:00
|
|
|
unsigned int idx = thumb << 1;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
if (ka->sa.sa_flags & SA_SIGINFO)
|
2006-01-19 05:38:47 +07:00
|
|
|
idx += 3;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2006-01-19 05:38:47 +07:00
|
|
|
if (__put_user(sigreturn_codes[idx], rc) ||
|
|
|
|
__put_user(sigreturn_codes[idx+1], rc+1))
|
2005-04-17 05:20:36 +07:00
|
|
|
return 1;
|
|
|
|
|
2005-06-23 02:26:05 +07:00
|
|
|
if (cpsr & MODE32_BIT) {
|
|
|
|
/*
|
|
|
|
* 32-bit code can use the new high-page
|
|
|
|
* signal return code support.
|
|
|
|
*/
|
|
|
|
retcode = KERN_SIGRETURN_CODE + (idx << 2) + thumb;
|
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* Ensure that the instruction cache sees
|
|
|
|
* the return code written onto the stack.
|
|
|
|
*/
|
|
|
|
flush_icache_range((unsigned long)rc,
|
2006-01-19 05:38:47 +07:00
|
|
|
(unsigned long)(rc + 2));
|
2005-06-23 02:26:05 +07:00
|
|
|
|
|
|
|
retcode = ((unsigned long)rc) + thumb;
|
|
|
|
}
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
regs->ARM_r0 = usig;
|
|
|
|
regs->ARM_sp = (unsigned long)frame;
|
|
|
|
regs->ARM_lr = retcode;
|
|
|
|
regs->ARM_pc = handler;
|
|
|
|
regs->ARM_cpsr = cpsr;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
setup_frame(int usig, struct k_sigaction *ka, sigset_t *set, struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
struct sigframe __user *frame = get_sigframe(ka, regs, sizeof(*frame));
|
|
|
|
int err = 0;
|
|
|
|
|
|
|
|
if (!frame)
|
|
|
|
return 1;
|
|
|
|
|
2006-06-25 04:41:09 +07:00
|
|
|
/*
|
|
|
|
* Set uc.uc_flags to a value which sc.trap_no would never have.
|
|
|
|
*/
|
|
|
|
__put_user_error(0x5ac3c35a, &frame->uc.uc_flags, err);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2006-06-16 02:28:03 +07:00
|
|
|
err |= setup_sigframe(frame, regs, set);
|
2005-04-17 05:20:36 +07:00
|
|
|
if (err == 0)
|
2006-01-19 05:38:47 +07:00
|
|
|
err = setup_return(regs, ka, frame->retcode, frame, usig);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
setup_rt_frame(int usig, struct k_sigaction *ka, siginfo_t *info,
|
|
|
|
sigset_t *set, struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
struct rt_sigframe __user *frame = get_sigframe(ka, regs, sizeof(*frame));
|
|
|
|
stack_t stack;
|
|
|
|
int err = 0;
|
|
|
|
|
|
|
|
if (!frame)
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
err |= copy_siginfo_to_user(&frame->info, info);
|
|
|
|
|
2006-06-16 02:18:25 +07:00
|
|
|
__put_user_error(0, &frame->sig.uc.uc_flags, err);
|
|
|
|
__put_user_error(NULL, &frame->sig.uc.uc_link, err);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
memset(&stack, 0, sizeof(stack));
|
|
|
|
stack.ss_sp = (void __user *)current->sas_ss_sp;
|
|
|
|
stack.ss_flags = sas_ss_flags(regs->ARM_sp);
|
|
|
|
stack.ss_size = current->sas_ss_size;
|
2006-06-16 02:18:25 +07:00
|
|
|
err |= __copy_to_user(&frame->sig.uc.uc_stack, &stack, sizeof(stack));
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2006-06-16 02:28:03 +07:00
|
|
|
err |= setup_sigframe(&frame->sig, regs, set);
|
2005-04-17 05:20:36 +07:00
|
|
|
if (err == 0)
|
2006-06-16 02:18:25 +07:00
|
|
|
err = setup_return(regs, ka, frame->sig.retcode, frame, usig);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
if (err == 0) {
|
|
|
|
/*
|
|
|
|
* For realtime signals we must also set the second and third
|
|
|
|
* arguments for the signal handler.
|
|
|
|
* -- Peter Maydell <pmaydell@chiark.greenend.org.uk> 2000-12-06
|
|
|
|
*/
|
|
|
|
regs->ARM_r1 = (unsigned long)&frame->info;
|
2006-06-16 02:18:25 +07:00
|
|
|
regs->ARM_r2 = (unsigned long)&frame->sig.uc;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void restart_syscall(struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
regs->ARM_r0 = regs->ARM_ORIG_r0;
|
|
|
|
regs->ARM_pc -= thumb_mode(regs) ? 2 : 4;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* OK, we're invoking a handler
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
handle_signal(unsigned long sig, struct k_sigaction *ka,
|
|
|
|
siginfo_t *info, sigset_t *oldset,
|
|
|
|
struct pt_regs * regs, int syscall)
|
|
|
|
{
|
|
|
|
struct thread_info *thread = current_thread_info();
|
|
|
|
struct task_struct *tsk = current;
|
|
|
|
int usig = sig;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If we were from a system call, check for system call restarting...
|
|
|
|
*/
|
|
|
|
if (syscall) {
|
|
|
|
switch (regs->ARM_r0) {
|
|
|
|
case -ERESTART_RESTARTBLOCK:
|
|
|
|
case -ERESTARTNOHAND:
|
|
|
|
regs->ARM_r0 = -EINTR;
|
|
|
|
break;
|
|
|
|
case -ERESTARTSYS:
|
|
|
|
if (!(ka->sa.sa_flags & SA_RESTART)) {
|
|
|
|
regs->ARM_r0 = -EINTR;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
/* fallthrough */
|
|
|
|
case -ERESTARTNOINTR:
|
|
|
|
restart_syscall(regs);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* translate the signal
|
|
|
|
*/
|
|
|
|
if (usig < 32 && thread->exec_domain && thread->exec_domain->signal_invmap)
|
|
|
|
usig = thread->exec_domain->signal_invmap[usig];
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Set up the stack frame
|
|
|
|
*/
|
|
|
|
if (ka->sa.sa_flags & SA_SIGINFO)
|
|
|
|
ret = setup_rt_frame(usig, ka, info, oldset, regs);
|
|
|
|
else
|
|
|
|
ret = setup_frame(usig, ka, oldset, regs);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Check that the resulting registers are actually sane.
|
|
|
|
*/
|
|
|
|
ret |= !valid_user_regs(regs);
|
|
|
|
|
[PATCH] convert signal handling of NODEFER to act like other Unix boxes.
It has been reported that the way Linux handles NODEFER for signals is
not consistent with the way other Unix boxes handle it. I've written a
program to test the behavior of how this flag affects signals and had
several reports from people who ran this on various Unix boxes,
confirming that Linux seems to be unique on the way this is handled.
The way NODEFER affects signals on other Unix boxes is as follows:
1) If NODEFER is set, other signals in sa_mask are still blocked.
2) If NODEFER is set and the signal is in sa_mask, then the signal is
still blocked. (Note: this is the behavior of all tested but Linux _and_
NetBSD 2.0 *).
The way NODEFER affects signals on Linux:
1) If NODEFER is set, other signals are _not_ blocked regardless of
sa_mask (Even NetBSD doesn't do this).
2) If NODEFER is set and the signal is in sa_mask, then the signal being
handled is not blocked.
The patch converts signal handling in all current Linux architectures to
the way most Unix boxes work.
Unix boxes that were tested: DU4, AIX 5.2, Irix 6.5, NetBSD 2.0, SFU
3.5 on WinXP, AIX 5.3, Mac OSX, and of course Linux 2.6.13-rcX.
* NetBSD was the only other Unix to behave like Linux on point #2. The
main concern was brought up by point #1 which even NetBSD isn't like
Linux. So with this patch, we leave NetBSD as the lonely one that
behaves differently here with #2.
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2005-08-29 22:44:09 +07:00
|
|
|
if (ret != 0) {
|
2005-11-19 17:01:07 +07:00
|
|
|
force_sigsegv(sig, tsk);
|
|
|
|
return;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
2005-11-19 17:01:07 +07:00
|
|
|
/*
|
|
|
|
* Block the signal if we were successful.
|
|
|
|
*/
|
|
|
|
spin_lock_irq(&tsk->sighand->siglock);
|
|
|
|
sigorsets(&tsk->blocked, &tsk->blocked,
|
|
|
|
&ka->sa.sa_mask);
|
|
|
|
if (!(ka->sa.sa_flags & SA_NODEFER))
|
|
|
|
sigaddset(&tsk->blocked, sig);
|
|
|
|
recalc_sigpending();
|
|
|
|
spin_unlock_irq(&tsk->sighand->siglock);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Note that 'init' is a special process: it doesn't get signals it doesn't
|
|
|
|
* want to handle. Thus you cannot kill init even with a SIGKILL even by
|
|
|
|
* mistake.
|
|
|
|
*
|
|
|
|
* Note that we go through the signals twice: once to check the signals that
|
|
|
|
* the kernel can handle, and then we build all the user-level signal handling
|
|
|
|
* stack-frames in one go after that.
|
|
|
|
*/
|
|
|
|
static int do_signal(sigset_t *oldset, struct pt_regs *regs, int syscall)
|
|
|
|
{
|
|
|
|
struct k_sigaction ka;
|
|
|
|
siginfo_t info;
|
|
|
|
int signr;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We want the common case to go fast, which
|
|
|
|
* is why we may in certain cases get here from
|
|
|
|
* kernel mode. Just return without doing anything
|
|
|
|
* if so.
|
|
|
|
*/
|
|
|
|
if (!user_mode(regs))
|
|
|
|
return 0;
|
|
|
|
|
2005-06-26 17:27:21 +07:00
|
|
|
if (try_to_freeze())
|
2005-04-17 05:20:36 +07:00
|
|
|
goto no_signal;
|
|
|
|
|
|
|
|
if (current->ptrace & PT_SINGLESTEP)
|
|
|
|
ptrace_cancel_bpt(current);
|
|
|
|
|
|
|
|
signr = get_signal_to_deliver(&info, &ka, regs, NULL);
|
|
|
|
if (signr > 0) {
|
|
|
|
handle_signal(signr, &ka, &info, oldset, regs, syscall);
|
|
|
|
if (current->ptrace & PT_SINGLESTEP)
|
|
|
|
ptrace_set_bpt(current);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
no_signal:
|
|
|
|
/*
|
|
|
|
* No signal to deliver to the process - restart the syscall.
|
|
|
|
*/
|
|
|
|
if (syscall) {
|
|
|
|
if (regs->ARM_r0 == -ERESTART_RESTARTBLOCK) {
|
|
|
|
if (thumb_mode(regs)) {
|
2006-06-23 04:18:45 +07:00
|
|
|
regs->ARM_r7 = __NR_restart_syscall - __NR_SYSCALL_BASE;
|
2005-04-17 05:20:36 +07:00
|
|
|
regs->ARM_pc -= 2;
|
|
|
|
} else {
|
2006-06-23 04:18:45 +07:00
|
|
|
#if defined(CONFIG_AEABI) && !defined(CONFIG_OABI_COMPAT)
|
|
|
|
regs->ARM_r7 = __NR_restart_syscall;
|
|
|
|
regs->ARM_pc -= 4;
|
|
|
|
#else
|
2005-04-17 05:20:36 +07:00
|
|
|
u32 __user *usp;
|
2006-06-23 04:18:45 +07:00
|
|
|
u32 swival = __NR_restart_syscall;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
regs->ARM_sp -= 12;
|
|
|
|
usp = (u32 __user *)regs->ARM_sp;
|
|
|
|
|
2006-06-23 04:18:45 +07:00
|
|
|
/*
|
|
|
|
* Either we supports OABI only, or we have
|
|
|
|
* EABI with the OABI compat layer enabled.
|
|
|
|
* In the later case we don't know if user
|
|
|
|
* space is EABI or not, and if not we must
|
|
|
|
* not clobber r7. Always using the OABI
|
|
|
|
* syscall solves that issue and works for
|
|
|
|
* all those cases.
|
|
|
|
*/
|
2006-06-23 22:40:53 +07:00
|
|
|
swival = swival - __NR_SYSCALL_BASE + __NR_OABI_SYSCALL_BASE;
|
2006-06-23 04:18:45 +07:00
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
put_user(regs->ARM_pc, &usp[0]);
|
|
|
|
/* swi __NR_restart_syscall */
|
2006-06-23 04:18:45 +07:00
|
|
|
put_user(0xef000000 | swival, &usp[1]);
|
2005-04-17 05:20:36 +07:00
|
|
|
/* ldr pc, [sp], #12 */
|
|
|
|
put_user(0xe49df00c, &usp[2]);
|
|
|
|
|
|
|
|
flush_icache_range((unsigned long)usp,
|
|
|
|
(unsigned long)(usp + 3));
|
|
|
|
|
|
|
|
regs->ARM_pc = regs->ARM_sp + 4;
|
2006-06-23 04:18:45 +07:00
|
|
|
#endif
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
if (regs->ARM_r0 == -ERESTARTNOHAND ||
|
|
|
|
regs->ARM_r0 == -ERESTARTSYS ||
|
|
|
|
regs->ARM_r0 == -ERESTARTNOINTR) {
|
|
|
|
restart_syscall(regs);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (current->ptrace & PT_SINGLESTEP)
|
|
|
|
ptrace_set_bpt(current);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
asmlinkage void
|
|
|
|
do_notify_resume(struct pt_regs *regs, unsigned int thread_flags, int syscall)
|
|
|
|
{
|
|
|
|
if (thread_flags & _TIF_SIGPENDING)
|
|
|
|
do_signal(¤t->blocked, regs, syscall);
|
|
|
|
}
|