2 * linux/arch/arm/kernel/signal.c
4 * Copyright (C) 1995-2002 Russell King
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License version 2 as
8 * published by the Free Software Foundation.
10 #include <linux/config.h>
11 #include <linux/errno.h>
12 #include <linux/signal.h>
13 #include <linux/ptrace.h>
14 #include <linux/personality.h>
16 #include <asm/cacheflush.h>
17 #include <asm/ucontext.h>
18 #include <asm/uaccess.h>
19 #include <asm/unistd.h>
24 #define _BLOCKABLE (~(sigmask(SIGKILL) | sigmask(SIGSTOP)))
27 * For ARM syscalls, we encode the syscall number into the instruction.
29 #define SWI_SYS_SIGRETURN (0xef000000|(__NR_sigreturn))
30 #define SWI_SYS_RT_SIGRETURN (0xef000000|(__NR_rt_sigreturn))
33 * With EABI, the syscall number has to be loaded into r7.
35 #define MOV_R7_NR_SIGRETURN (0xe3a07000 | (__NR_sigreturn - __NR_SYSCALL_BASE))
36 #define MOV_R7_NR_RT_SIGRETURN (0xe3a07000 | (__NR_rt_sigreturn - __NR_SYSCALL_BASE))
39 * For Thumb syscalls, we pass the syscall number via r7. We therefore
40 * need two 16-bit instructions.
42 #define SWI_THUMB_SIGRETURN (0xdf00 << 16 | 0x2700 | (__NR_sigreturn - __NR_SYSCALL_BASE))
43 #define SWI_THUMB_RT_SIGRETURN (0xdf00 << 16 | 0x2700 | (__NR_rt_sigreturn - __NR_SYSCALL_BASE))
45 const unsigned long sigreturn_codes
[7] = {
46 MOV_R7_NR_SIGRETURN
, SWI_SYS_SIGRETURN
, SWI_THUMB_SIGRETURN
,
47 MOV_R7_NR_RT_SIGRETURN
, SWI_SYS_RT_SIGRETURN
, SWI_THUMB_RT_SIGRETURN
,
50 static int do_signal(sigset_t
*oldset
, struct pt_regs
* regs
, int syscall
);
53 * atomically swap in the new signal mask, and wait for a signal.
55 asmlinkage
int sys_sigsuspend(int restart
, unsigned long oldmask
, old_sigset_t mask
, struct pt_regs
*regs
)
60 spin_lock_irq(¤t
->sighand
->siglock
);
61 saveset
= current
->blocked
;
62 siginitset(¤t
->blocked
, mask
);
64 spin_unlock_irq(¤t
->sighand
->siglock
);
65 regs
->ARM_r0
= -EINTR
;
68 current
->state
= TASK_INTERRUPTIBLE
;
70 if (do_signal(&saveset
, regs
, 0))
76 sys_rt_sigsuspend(sigset_t __user
*unewset
, size_t sigsetsize
, struct pt_regs
*regs
)
78 sigset_t saveset
, newset
;
80 /* XXX: Don't preclude handling different sized sigset_t's. */
81 if (sigsetsize
!= sizeof(sigset_t
))
84 if (copy_from_user(&newset
, unewset
, sizeof(newset
)))
86 sigdelsetmask(&newset
, ~_BLOCKABLE
);
88 spin_lock_irq(¤t
->sighand
->siglock
);
89 saveset
= current
->blocked
;
90 current
->blocked
= newset
;
92 spin_unlock_irq(¤t
->sighand
->siglock
);
93 regs
->ARM_r0
= -EINTR
;
96 current
->state
= TASK_INTERRUPTIBLE
;
98 if (do_signal(&saveset
, regs
, 0))
104 sys_sigaction(int sig
, const struct old_sigaction __user
*act
,
105 struct old_sigaction __user
*oact
)
107 struct k_sigaction new_ka
, old_ka
;
112 if (!access_ok(VERIFY_READ
, act
, sizeof(*act
)) ||
113 __get_user(new_ka
.sa
.sa_handler
, &act
->sa_handler
) ||
114 __get_user(new_ka
.sa
.sa_restorer
, &act
->sa_restorer
))
116 __get_user(new_ka
.sa
.sa_flags
, &act
->sa_flags
);
117 __get_user(mask
, &act
->sa_mask
);
118 siginitset(&new_ka
.sa
.sa_mask
, mask
);
121 ret
= do_sigaction(sig
, act
? &new_ka
: NULL
, oact
? &old_ka
: NULL
);
124 if (!access_ok(VERIFY_WRITE
, oact
, sizeof(*oact
)) ||
125 __put_user(old_ka
.sa
.sa_handler
, &oact
->sa_handler
) ||
126 __put_user(old_ka
.sa
.sa_restorer
, &oact
->sa_restorer
))
128 __put_user(old_ka
.sa
.sa_flags
, &oact
->sa_flags
);
129 __put_user(old_ka
.sa
.sa_mask
.sig
[0], &oact
->sa_mask
);
136 static int preserve_crunch_context(struct crunch_sigframe
*frame
)
138 char kbuf
[sizeof(*frame
) + 8];
139 struct crunch_sigframe
*kframe
;
141 /* the crunch context must be 64 bit aligned */
142 kframe
= (struct crunch_sigframe
*)((unsigned long)(kbuf
+ 8) & ~7);
143 kframe
->magic
= CRUNCH_MAGIC
;
144 kframe
->size
= CRUNCH_STORAGE_SIZE
;
145 crunch_task_copy(current_thread_info(), &kframe
->storage
);
146 return __copy_to_user(frame
, kframe
, sizeof(*frame
));
149 static int restore_crunch_context(struct crunch_sigframe
*frame
)
151 char kbuf
[sizeof(*frame
) + 8];
152 struct crunch_sigframe
*kframe
;
154 /* the crunch context must be 64 bit aligned */
155 kframe
= (struct crunch_sigframe
*)((unsigned long)(kbuf
+ 8) & ~7);
156 if (__copy_from_user(kframe
, frame
, sizeof(*frame
)))
158 if (kframe
->magic
!= CRUNCH_MAGIC
||
159 kframe
->size
!= CRUNCH_STORAGE_SIZE
)
161 crunch_task_restore(current_thread_info(), &kframe
->storage
);
168 static int preserve_iwmmxt_context(struct iwmmxt_sigframe
*frame
)
170 char kbuf
[sizeof(*frame
) + 8];
171 struct iwmmxt_sigframe
*kframe
;
173 /* the iWMMXt context must be 64 bit aligned */
174 kframe
= (struct iwmmxt_sigframe
*)((unsigned long)(kbuf
+ 8) & ~7);
175 kframe
->magic
= IWMMXT_MAGIC
;
176 kframe
->size
= IWMMXT_STORAGE_SIZE
;
177 iwmmxt_task_copy(current_thread_info(), &kframe
->storage
);
178 return __copy_to_user(frame
, kframe
, sizeof(*frame
));
181 static int restore_iwmmxt_context(struct iwmmxt_sigframe
*frame
)
183 char kbuf
[sizeof(*frame
) + 8];
184 struct iwmmxt_sigframe
*kframe
;
186 /* the iWMMXt context must be 64 bit aligned */
187 kframe
= (struct iwmmxt_sigframe
*)((unsigned long)(kbuf
+ 8) & ~7);
188 if (__copy_from_user(kframe
, frame
, sizeof(*frame
)))
190 if (kframe
->magic
!= IWMMXT_MAGIC
||
191 kframe
->size
!= IWMMXT_STORAGE_SIZE
)
193 iwmmxt_task_restore(current_thread_info(), &kframe
->storage
);
200 * Do a signal return; undo the signal stack. These are aligned to 64-bit.
204 unsigned long retcode
[2];
212 static int restore_sigframe(struct pt_regs
*regs
, struct sigframe __user
*sf
)
214 struct aux_sigframe __user
*aux
;
218 err
= __copy_from_user(&set
, &sf
->uc
.uc_sigmask
, sizeof(set
));
220 sigdelsetmask(&set
, ~_BLOCKABLE
);
221 spin_lock_irq(¤t
->sighand
->siglock
);
222 current
->blocked
= set
;
224 spin_unlock_irq(¤t
->sighand
->siglock
);
227 __get_user_error(regs
->ARM_r0
, &sf
->uc
.uc_mcontext
.arm_r0
, err
);
228 __get_user_error(regs
->ARM_r1
, &sf
->uc
.uc_mcontext
.arm_r1
, err
);
229 __get_user_error(regs
->ARM_r2
, &sf
->uc
.uc_mcontext
.arm_r2
, err
);
230 __get_user_error(regs
->ARM_r3
, &sf
->uc
.uc_mcontext
.arm_r3
, err
);
231 __get_user_error(regs
->ARM_r4
, &sf
->uc
.uc_mcontext
.arm_r4
, err
);
232 __get_user_error(regs
->ARM_r5
, &sf
->uc
.uc_mcontext
.arm_r5
, err
);
233 __get_user_error(regs
->ARM_r6
, &sf
->uc
.uc_mcontext
.arm_r6
, err
);
234 __get_user_error(regs
->ARM_r7
, &sf
->uc
.uc_mcontext
.arm_r7
, err
);
235 __get_user_error(regs
->ARM_r8
, &sf
->uc
.uc_mcontext
.arm_r8
, err
);
236 __get_user_error(regs
->ARM_r9
, &sf
->uc
.uc_mcontext
.arm_r9
, err
);
237 __get_user_error(regs
->ARM_r10
, &sf
->uc
.uc_mcontext
.arm_r10
, err
);
238 __get_user_error(regs
->ARM_fp
, &sf
->uc
.uc_mcontext
.arm_fp
, err
);
239 __get_user_error(regs
->ARM_ip
, &sf
->uc
.uc_mcontext
.arm_ip
, err
);
240 __get_user_error(regs
->ARM_sp
, &sf
->uc
.uc_mcontext
.arm_sp
, err
);
241 __get_user_error(regs
->ARM_lr
, &sf
->uc
.uc_mcontext
.arm_lr
, err
);
242 __get_user_error(regs
->ARM_pc
, &sf
->uc
.uc_mcontext
.arm_pc
, err
);
243 __get_user_error(regs
->ARM_cpsr
, &sf
->uc
.uc_mcontext
.arm_cpsr
, err
);
245 err
|= !valid_user_regs(regs
);
247 aux
= (struct aux_sigframe __user
*) sf
->uc
.uc_regspace
;
250 err
|= restore_crunch_context(&aux
->crunch
);
253 if (err
== 0 && test_thread_flag(TIF_USING_IWMMXT
))
254 err
|= restore_iwmmxt_context(&aux
->iwmmxt
);
258 // err |= vfp_restore_state(&sf->aux.vfp);
264 asmlinkage
int sys_sigreturn(struct pt_regs
*regs
)
266 struct sigframe __user
*frame
;
268 /* Always make any pending restarted system calls return -EINTR */
269 current_thread_info()->restart_block
.fn
= do_no_restart_syscall
;
272 * Since we stacked the signal on a 64-bit boundary,
273 * then 'sp' should be word aligned here. If it's
274 * not, then the user is trying to mess with us.
276 if (regs
->ARM_sp
& 7)
279 frame
= (struct sigframe __user
*)regs
->ARM_sp
;
281 if (!access_ok(VERIFY_READ
, frame
, sizeof (*frame
)))
284 if (restore_sigframe(regs
, frame
))
287 /* Send SIGTRAP if we're single-stepping */
288 if (current
->ptrace
& PT_SINGLESTEP
) {
289 ptrace_cancel_bpt(current
);
290 send_sig(SIGTRAP
, current
, 1);
296 force_sig(SIGSEGV
, current
);
300 asmlinkage
int sys_rt_sigreturn(struct pt_regs
*regs
)
302 struct rt_sigframe __user
*frame
;
304 /* Always make any pending restarted system calls return -EINTR */
305 current_thread_info()->restart_block
.fn
= do_no_restart_syscall
;
308 * Since we stacked the signal on a 64-bit boundary,
309 * then 'sp' should be word aligned here. If it's
310 * not, then the user is trying to mess with us.
312 if (regs
->ARM_sp
& 7)
315 frame
= (struct rt_sigframe __user
*)regs
->ARM_sp
;
317 if (!access_ok(VERIFY_READ
, frame
, sizeof (*frame
)))
320 if (restore_sigframe(regs
, &frame
->sig
))
323 if (do_sigaltstack(&frame
->sig
.uc
.uc_stack
, NULL
, regs
->ARM_sp
) == -EFAULT
)
326 /* Send SIGTRAP if we're single-stepping */
327 if (current
->ptrace
& PT_SINGLESTEP
) {
328 ptrace_cancel_bpt(current
);
329 send_sig(SIGTRAP
, current
, 1);
335 force_sig(SIGSEGV
, current
);
340 setup_sigframe(struct sigframe __user
*sf
, struct pt_regs
*regs
, sigset_t
*set
)
342 struct aux_sigframe __user
*aux
;
345 __put_user_error(regs
->ARM_r0
, &sf
->uc
.uc_mcontext
.arm_r0
, err
);
346 __put_user_error(regs
->ARM_r1
, &sf
->uc
.uc_mcontext
.arm_r1
, err
);
347 __put_user_error(regs
->ARM_r2
, &sf
->uc
.uc_mcontext
.arm_r2
, err
);
348 __put_user_error(regs
->ARM_r3
, &sf
->uc
.uc_mcontext
.arm_r3
, err
);
349 __put_user_error(regs
->ARM_r4
, &sf
->uc
.uc_mcontext
.arm_r4
, err
);
350 __put_user_error(regs
->ARM_r5
, &sf
->uc
.uc_mcontext
.arm_r5
, err
);
351 __put_user_error(regs
->ARM_r6
, &sf
->uc
.uc_mcontext
.arm_r6
, err
);
352 __put_user_error(regs
->ARM_r7
, &sf
->uc
.uc_mcontext
.arm_r7
, err
);
353 __put_user_error(regs
->ARM_r8
, &sf
->uc
.uc_mcontext
.arm_r8
, err
);
354 __put_user_error(regs
->ARM_r9
, &sf
->uc
.uc_mcontext
.arm_r9
, err
);
355 __put_user_error(regs
->ARM_r10
, &sf
->uc
.uc_mcontext
.arm_r10
, err
);
356 __put_user_error(regs
->ARM_fp
, &sf
->uc
.uc_mcontext
.arm_fp
, err
);
357 __put_user_error(regs
->ARM_ip
, &sf
->uc
.uc_mcontext
.arm_ip
, err
);
358 __put_user_error(regs
->ARM_sp
, &sf
->uc
.uc_mcontext
.arm_sp
, err
);
359 __put_user_error(regs
->ARM_lr
, &sf
->uc
.uc_mcontext
.arm_lr
, err
);
360 __put_user_error(regs
->ARM_pc
, &sf
->uc
.uc_mcontext
.arm_pc
, err
);
361 __put_user_error(regs
->ARM_cpsr
, &sf
->uc
.uc_mcontext
.arm_cpsr
, err
);
363 __put_user_error(current
->thread
.trap_no
, &sf
->uc
.uc_mcontext
.trap_no
, err
);
364 __put_user_error(current
->thread
.error_code
, &sf
->uc
.uc_mcontext
.error_code
, err
);
365 __put_user_error(current
->thread
.address
, &sf
->uc
.uc_mcontext
.fault_address
, err
);
366 __put_user_error(set
->sig
[0], &sf
->uc
.uc_mcontext
.oldmask
, err
);
368 err
|= __copy_to_user(&sf
->uc
.uc_sigmask
, set
, sizeof(*set
));
370 aux
= (struct aux_sigframe __user
*) sf
->uc
.uc_regspace
;
373 err
|= preserve_crunch_context(&aux
->crunch
);
376 if (err
== 0 && test_thread_flag(TIF_USING_IWMMXT
))
377 err
|= preserve_iwmmxt_context(&aux
->iwmmxt
);
381 // err |= vfp_save_state(&sf->aux.vfp);
383 __put_user_error(0, &aux
->end_magic
, err
);
388 static inline void __user
*
389 get_sigframe(struct k_sigaction
*ka
, struct pt_regs
*regs
, int framesize
)
391 unsigned long sp
= regs
->ARM_sp
;
395 * This is the X/Open sanctioned signal stack switching.
397 if ((ka
->sa
.sa_flags
& SA_ONSTACK
) && !sas_ss_flags(sp
))
398 sp
= current
->sas_ss_sp
+ current
->sas_ss_size
;
401 * ATPCS B01 mandates 8-byte alignment
403 frame
= (void __user
*)((sp
- framesize
) & ~7);
406 * Check that we can actually write to the signal frame.
408 if (!access_ok(VERIFY_WRITE
, frame
, framesize
))
415 setup_return(struct pt_regs
*regs
, struct k_sigaction
*ka
,
416 unsigned long __user
*rc
, void __user
*frame
, int usig
)
418 unsigned long handler
= (unsigned long)ka
->sa
.sa_handler
;
419 unsigned long retcode
;
421 unsigned long cpsr
= regs
->ARM_cpsr
& ~PSR_f
;
424 * Maybe we need to deliver a 32-bit signal to a 26-bit task.
426 if (ka
->sa
.sa_flags
& SA_THIRTYTWO
)
427 cpsr
= (cpsr
& ~MODE_MASK
) | USR_MODE
;
429 #ifdef CONFIG_ARM_THUMB
430 if (elf_hwcap
& HWCAP_THUMB
) {
432 * The LSB of the handler determines if we're going to
433 * be using THUMB or ARM mode for this signal handler.
444 if (ka
->sa
.sa_flags
& SA_RESTORER
) {
445 retcode
= (unsigned long)ka
->sa
.sa_restorer
;
447 unsigned int idx
= thumb
<< 1;
449 if (ka
->sa
.sa_flags
& SA_SIGINFO
)
452 if (__put_user(sigreturn_codes
[idx
], rc
) ||
453 __put_user(sigreturn_codes
[idx
+1], rc
+1))
456 if (cpsr
& MODE32_BIT
) {
458 * 32-bit code can use the new high-page
459 * signal return code support.
461 retcode
= KERN_SIGRETURN_CODE
+ (idx
<< 2) + thumb
;
464 * Ensure that the instruction cache sees
465 * the return code written onto the stack.
467 flush_icache_range((unsigned long)rc
,
468 (unsigned long)(rc
+ 2));
470 retcode
= ((unsigned long)rc
) + thumb
;
475 regs
->ARM_sp
= (unsigned long)frame
;
476 regs
->ARM_lr
= retcode
;
477 regs
->ARM_pc
= handler
;
478 regs
->ARM_cpsr
= cpsr
;
484 setup_frame(int usig
, struct k_sigaction
*ka
, sigset_t
*set
, struct pt_regs
*regs
)
486 struct sigframe __user
*frame
= get_sigframe(ka
, regs
, sizeof(*frame
));
493 * Set uc.uc_flags to a value which sc.trap_no would never have.
495 __put_user_error(0x5ac3c35a, &frame
->uc
.uc_flags
, err
);
497 err
|= setup_sigframe(frame
, regs
, set
);
499 err
= setup_return(regs
, ka
, frame
->retcode
, frame
, usig
);
505 setup_rt_frame(int usig
, struct k_sigaction
*ka
, siginfo_t
*info
,
506 sigset_t
*set
, struct pt_regs
*regs
)
508 struct rt_sigframe __user
*frame
= get_sigframe(ka
, regs
, sizeof(*frame
));
515 err
|= copy_siginfo_to_user(&frame
->info
, info
);
517 __put_user_error(0, &frame
->sig
.uc
.uc_flags
, err
);
518 __put_user_error(NULL
, &frame
->sig
.uc
.uc_link
, err
);
520 memset(&stack
, 0, sizeof(stack
));
521 stack
.ss_sp
= (void __user
*)current
->sas_ss_sp
;
522 stack
.ss_flags
= sas_ss_flags(regs
->ARM_sp
);
523 stack
.ss_size
= current
->sas_ss_size
;
524 err
|= __copy_to_user(&frame
->sig
.uc
.uc_stack
, &stack
, sizeof(stack
));
526 err
|= setup_sigframe(&frame
->sig
, regs
, set
);
528 err
= setup_return(regs
, ka
, frame
->sig
.retcode
, frame
, usig
);
532 * For realtime signals we must also set the second and third
533 * arguments for the signal handler.
534 * -- Peter Maydell <pmaydell@chiark.greenend.org.uk> 2000-12-06
536 regs
->ARM_r1
= (unsigned long)&frame
->info
;
537 regs
->ARM_r2
= (unsigned long)&frame
->sig
.uc
;
543 static inline void restart_syscall(struct pt_regs
*regs
)
545 regs
->ARM_r0
= regs
->ARM_ORIG_r0
;
546 regs
->ARM_pc
-= thumb_mode(regs
) ? 2 : 4;
550 * OK, we're invoking a handler
553 handle_signal(unsigned long sig
, struct k_sigaction
*ka
,
554 siginfo_t
*info
, sigset_t
*oldset
,
555 struct pt_regs
* regs
, int syscall
)
557 struct thread_info
*thread
= current_thread_info();
558 struct task_struct
*tsk
= current
;
563 * If we were from a system call, check for system call restarting...
566 switch (regs
->ARM_r0
) {
567 case -ERESTART_RESTARTBLOCK
:
568 case -ERESTARTNOHAND
:
569 regs
->ARM_r0
= -EINTR
;
572 if (!(ka
->sa
.sa_flags
& SA_RESTART
)) {
573 regs
->ARM_r0
= -EINTR
;
577 case -ERESTARTNOINTR
:
578 restart_syscall(regs
);
583 * translate the signal
585 if (usig
< 32 && thread
->exec_domain
&& thread
->exec_domain
->signal_invmap
)
586 usig
= thread
->exec_domain
->signal_invmap
[usig
];
589 * Set up the stack frame
591 if (ka
->sa
.sa_flags
& SA_SIGINFO
)
592 ret
= setup_rt_frame(usig
, ka
, info
, oldset
, regs
);
594 ret
= setup_frame(usig
, ka
, oldset
, regs
);
597 * Check that the resulting registers are actually sane.
599 ret
|= !valid_user_regs(regs
);
602 force_sigsegv(sig
, tsk
);
607 * Block the signal if we were successful.
609 spin_lock_irq(&tsk
->sighand
->siglock
);
610 sigorsets(&tsk
->blocked
, &tsk
->blocked
,
612 if (!(ka
->sa
.sa_flags
& SA_NODEFER
))
613 sigaddset(&tsk
->blocked
, sig
);
615 spin_unlock_irq(&tsk
->sighand
->siglock
);
620 * Note that 'init' is a special process: it doesn't get signals it doesn't
621 * want to handle. Thus you cannot kill init even with a SIGKILL even by
624 * Note that we go through the signals twice: once to check the signals that
625 * the kernel can handle, and then we build all the user-level signal handling
626 * stack-frames in one go after that.
628 static int do_signal(sigset_t
*oldset
, struct pt_regs
*regs
, int syscall
)
630 struct k_sigaction ka
;
635 * We want the common case to go fast, which
636 * is why we may in certain cases get here from
637 * kernel mode. Just return without doing anything
640 if (!user_mode(regs
))
646 if (current
->ptrace
& PT_SINGLESTEP
)
647 ptrace_cancel_bpt(current
);
649 signr
= get_signal_to_deliver(&info
, &ka
, regs
, NULL
);
651 handle_signal(signr
, &ka
, &info
, oldset
, regs
, syscall
);
652 if (current
->ptrace
& PT_SINGLESTEP
)
653 ptrace_set_bpt(current
);
659 * No signal to deliver to the process - restart the syscall.
662 if (regs
->ARM_r0
== -ERESTART_RESTARTBLOCK
) {
663 if (thumb_mode(regs
)) {
664 regs
->ARM_r7
= __NR_restart_syscall
- __NR_SYSCALL_BASE
;
667 #if defined(CONFIG_AEABI) && !defined(CONFIG_OABI_COMPAT)
668 regs
->ARM_r7
= __NR_restart_syscall
;
672 u32 swival
= __NR_restart_syscall
;
675 usp
= (u32 __user
*)regs
->ARM_sp
;
678 * Either we supports OABI only, or we have
679 * EABI with the OABI compat layer enabled.
680 * In the later case we don't know if user
681 * space is EABI or not, and if not we must
682 * not clobber r7. Always using the OABI
683 * syscall solves that issue and works for
686 swival
= swival
- __NR_SYSCALL_BASE
+ __NR_OABI_SYSCALL_BASE
;
688 put_user(regs
->ARM_pc
, &usp
[0]);
689 /* swi __NR_restart_syscall */
690 put_user(0xef000000 | swival
, &usp
[1]);
691 /* ldr pc, [sp], #12 */
692 put_user(0xe49df00c, &usp
[2]);
694 flush_icache_range((unsigned long)usp
,
695 (unsigned long)(usp
+ 3));
697 regs
->ARM_pc
= regs
->ARM_sp
+ 4;
701 if (regs
->ARM_r0
== -ERESTARTNOHAND
||
702 regs
->ARM_r0
== -ERESTARTSYS
||
703 regs
->ARM_r0
== -ERESTARTNOINTR
) {
704 restart_syscall(regs
);
707 if (current
->ptrace
& PT_SINGLESTEP
)
708 ptrace_set_bpt(current
);
713 do_notify_resume(struct pt_regs
*regs
, unsigned int thread_flags
, int syscall
)
715 if (thread_flags
& _TIF_SIGPENDING
)
716 do_signal(¤t
->blocked
, regs
, syscall
);
This page took 0.047359 seconds and 6 git commands to generate.