forked from luck/tmp_suning_uos_patched
1ca97bb541
Cross-architecture equivalent of rdusp(); default is user_stack_pointer(current_pt_regs()) - that works for almost all platforms that have usp saved in pt_regs. The only exception from that is ia64 - we want memory stack, not the backing store for register one. Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
152 lines
5.4 KiB
C
152 lines
5.4 KiB
C
/*
|
|
* Copyright (C) 1998-2004 Hewlett-Packard Co
|
|
* David Mosberger-Tang <davidm@hpl.hp.com>
|
|
* Stephane Eranian <eranian@hpl.hp.com>
|
|
* Copyright (C) 2003 Intel Co
|
|
* Suresh Siddha <suresh.b.siddha@intel.com>
|
|
* Fenghua Yu <fenghua.yu@intel.com>
|
|
* Arun Sharma <arun.sharma@intel.com>
|
|
*
|
|
* 12/07/98 S. Eranian added pt_regs & switch_stack
|
|
* 12/21/98 D. Mosberger updated to match latest code
|
|
* 6/17/99 D. Mosberger added second unat member to "struct switch_stack"
|
|
*
|
|
*/
|
|
#ifndef _ASM_IA64_PTRACE_H
|
|
#define _ASM_IA64_PTRACE_H
|
|
|
|
#ifndef ASM_OFFSETS_C
|
|
#include <asm/asm-offsets.h>
|
|
#endif
|
|
#include <uapi/asm/ptrace.h>
|
|
|
|
/*
|
|
* Base-2 logarithm of number of pages to allocate per task structure
|
|
* (including register backing store and memory stack):
|
|
*/
|
|
#if defined(CONFIG_IA64_PAGE_SIZE_4KB)
|
|
# define KERNEL_STACK_SIZE_ORDER 3
|
|
#elif defined(CONFIG_IA64_PAGE_SIZE_8KB)
|
|
# define KERNEL_STACK_SIZE_ORDER 2
|
|
#elif defined(CONFIG_IA64_PAGE_SIZE_16KB)
|
|
# define KERNEL_STACK_SIZE_ORDER 1
|
|
#else
|
|
# define KERNEL_STACK_SIZE_ORDER 0
|
|
#endif
|
|
|
|
#define IA64_RBS_OFFSET ((IA64_TASK_SIZE + IA64_THREAD_INFO_SIZE + 31) & ~31)
|
|
#define IA64_STK_OFFSET ((1 << KERNEL_STACK_SIZE_ORDER)*PAGE_SIZE)
|
|
|
|
#define KERNEL_STACK_SIZE IA64_STK_OFFSET
|
|
|
|
#ifndef __ASSEMBLY__
|
|
|
|
#include <asm/current.h>
|
|
#include <asm/page.h>
|
|
|
|
/*
|
|
* We use the ia64_psr(regs)->ri to determine which of the three
|
|
* instructions in bundle (16 bytes) took the sample. Generate
|
|
* the canonical representation by adding to instruction pointer.
|
|
*/
|
|
# define instruction_pointer(regs) ((regs)->cr_iip + ia64_psr(regs)->ri)
|
|
|
|
static inline unsigned long user_stack_pointer(struct pt_regs *regs)
|
|
{
|
|
/* FIXME: should this be bspstore + nr_dirty regs? */
|
|
return regs->ar_bspstore;
|
|
}
|
|
|
|
static inline int is_syscall_success(struct pt_regs *regs)
|
|
{
|
|
return regs->r10 != -1;
|
|
}
|
|
|
|
static inline long regs_return_value(struct pt_regs *regs)
|
|
{
|
|
if (is_syscall_success(regs))
|
|
return regs->r8;
|
|
else
|
|
return -regs->r8;
|
|
}
|
|
|
|
/* Conserve space in histogram by encoding slot bits in address
|
|
* bits 2 and 3 rather than bits 0 and 1.
|
|
*/
|
|
#define profile_pc(regs) \
|
|
({ \
|
|
unsigned long __ip = instruction_pointer(regs); \
|
|
(__ip & ~3UL) + ((__ip & 3UL) << 2); \
|
|
})
|
|
/*
|
|
* Why not default? Because user_stack_pointer() on ia64 gives register
|
|
* stack backing store instead...
|
|
*/
|
|
#define current_user_stack_pointer() (current_pt_regs()->r12)
|
|
|
|
/* given a pointer to a task_struct, return the user's pt_regs */
|
|
# define task_pt_regs(t) (((struct pt_regs *) ((char *) (t) + IA64_STK_OFFSET)) - 1)
|
|
# define ia64_psr(regs) ((struct ia64_psr *) &(regs)->cr_ipsr)
|
|
# define user_mode(regs) (((struct ia64_psr *) &(regs)->cr_ipsr)->cpl != 0)
|
|
# define user_stack(task,regs) ((long) regs - (long) task == IA64_STK_OFFSET - sizeof(*regs))
|
|
# define fsys_mode(task,regs) \
|
|
({ \
|
|
struct task_struct *_task = (task); \
|
|
struct pt_regs *_regs = (regs); \
|
|
!user_mode(_regs) && user_stack(_task, _regs); \
|
|
})
|
|
|
|
/*
|
|
* System call handlers that, upon successful completion, need to return a negative value
|
|
* should call force_successful_syscall_return() right before returning. On architectures
|
|
* where the syscall convention provides for a separate error flag (e.g., alpha, ia64,
|
|
* ppc{,64}, sparc{,64}, possibly others), this macro can be used to ensure that the error
|
|
* flag will not get set. On architectures which do not support a separate error flag,
|
|
* the macro is a no-op and the spurious error condition needs to be filtered out by some
|
|
* other means (e.g., in user-level, by passing an extra argument to the syscall handler,
|
|
* or something along those lines).
|
|
*
|
|
* On ia64, we can clear the user's pt_regs->r8 to force a successful syscall.
|
|
*/
|
|
# define force_successful_syscall_return() (task_pt_regs(current)->r8 = 0)
|
|
|
|
struct task_struct; /* forward decl */
|
|
struct unw_frame_info; /* forward decl */
|
|
|
|
extern void ia64_do_show_stack (struct unw_frame_info *, void *);
|
|
extern unsigned long ia64_get_user_rbs_end (struct task_struct *, struct pt_regs *,
|
|
unsigned long *);
|
|
extern long ia64_peek (struct task_struct *, struct switch_stack *, unsigned long,
|
|
unsigned long, long *);
|
|
extern long ia64_poke (struct task_struct *, struct switch_stack *, unsigned long,
|
|
unsigned long, long);
|
|
extern void ia64_flush_fph (struct task_struct *);
|
|
extern void ia64_sync_fph (struct task_struct *);
|
|
extern void ia64_sync_krbs(void);
|
|
extern long ia64_sync_user_rbs (struct task_struct *, struct switch_stack *,
|
|
unsigned long, unsigned long);
|
|
|
|
/* get nat bits for scratch registers such that bit N==1 iff scratch register rN is a NaT */
|
|
extern unsigned long ia64_get_scratch_nat_bits (struct pt_regs *pt, unsigned long scratch_unat);
|
|
/* put nat bits for scratch registers such that scratch register rN is a NaT iff bit N==1 */
|
|
extern unsigned long ia64_put_scratch_nat_bits (struct pt_regs *pt, unsigned long nat);
|
|
|
|
extern void ia64_increment_ip (struct pt_regs *pt);
|
|
extern void ia64_decrement_ip (struct pt_regs *pt);
|
|
|
|
extern void ia64_ptrace_stop(void);
|
|
#define arch_ptrace_stop(code, info) \
|
|
ia64_ptrace_stop()
|
|
#define arch_ptrace_stop_needed(code, info) \
|
|
(!test_thread_flag(TIF_RESTORE_RSE))
|
|
|
|
extern void ptrace_attach_sync_user_rbs (struct task_struct *);
|
|
#define arch_ptrace_attach(child) \
|
|
ptrace_attach_sync_user_rbs(child)
|
|
|
|
#define arch_has_single_step() (1)
|
|
#define arch_has_block_step() (1)
|
|
|
|
#endif /* !__ASSEMBLY__ */
|
|
#endif /* _ASM_IA64_PTRACE_H */
|