blob: 059780a7d3d7898b750a9f6f7265a0cb2687d2fd [file] [log] [blame]
Linus Torvalds1da177e2005-04-16 15:20:36 -07001/*
2 * include/asm-alpha/processor.h
3 *
4 * Copyright (C) 1994 Linus Torvalds
5 */
6
7#ifndef __ASM_ALPHA_PROCESSOR_H
8#define __ASM_ALPHA_PROCESSOR_H
9
10#include <linux/personality.h> /* for ADDR_LIMIT_32BIT */
11
12/*
13 * Returns current instruction pointer ("program counter").
14 */
15#define current_text_addr() \
16 ({ void *__pc; __asm__ ("br %0,.+4" : "=r"(__pc)); __pc; })
17
18/*
19 * We have a 42-bit user address space: 4TB user VM...
20 */
21#define TASK_SIZE (0x40000000000UL)
22
23/* This decides where the kernel will search for a free chunk of vm
24 * space during mmap's.
25 */
26#define TASK_UNMAPPED_BASE \
27 ((current->personality & ADDR_LIMIT_32BIT) ? 0x40000000 : TASK_SIZE / 2)
28
29typedef struct {
30 unsigned long seg;
31} mm_segment_t;
32
33/* This is dead. Everything has been moved to thread_info. */
34struct thread_struct { };
35#define INIT_THREAD { }
36
37/* Return saved PC of a blocked thread. */
38struct task_struct;
39extern unsigned long thread_saved_pc(struct task_struct *);
40
41/* Do necessary setup to start up a newly executed thread. */
42extern void start_thread(struct pt_regs *, unsigned long, unsigned long);
43
44/* Free all resources held by a thread. */
45extern void release_thread(struct task_struct *);
46
47/* Prepare to copy thread state - unlazy all lazy status */
48#define prepare_to_copy(tsk) do { } while (0)
49
50/* Create a kernel thread without removing it from tasklists. */
51extern long kernel_thread(int (*fn)(void *), void *arg, unsigned long flags);
52
53unsigned long get_wchan(struct task_struct *p);
54
55/* See arch/alpha/kernel/ptrace.c for details. */
56#define PT_REG(reg) \
57 (PAGE_SIZE*2 - sizeof(struct pt_regs) + offsetof(struct pt_regs, reg))
58
59#define SW_REG(reg) \
60 (PAGE_SIZE*2 - sizeof(struct pt_regs) - sizeof(struct switch_stack) \
61 + offsetof(struct switch_stack, reg))
62
63#define KSTK_EIP(tsk) \
64 (*(unsigned long *)(PT_REG(pc) + (unsigned long) ((tsk)->thread_info)))
65
66#define KSTK_ESP(tsk) \
67 ((tsk) == current ? rdusp() : (tsk)->thread_info->pcb.usp)
68
69#define cpu_relax() barrier()
70
71#define ARCH_HAS_PREFETCH
72#define ARCH_HAS_PREFETCHW
73#define ARCH_HAS_SPINLOCK_PREFETCH
74
75#ifndef CONFIG_SMP
76/* Nothing to prefetch. */
77#define spin_lock_prefetch(lock) do { } while (0)
78#endif
79
80#if __GNUC__ > 3 || (__GNUC__ == 3 && __GNUC_MINOR__ >= 1)
81extern inline void prefetch(const void *ptr)
82{
83 __builtin_prefetch(ptr, 0, 3);
84}
85
86extern inline void prefetchw(const void *ptr)
87{
88 __builtin_prefetch(ptr, 1, 3);
89}
90
91#ifdef CONFIG_SMP
92extern inline void spin_lock_prefetch(const void *ptr)
93{
94 __builtin_prefetch(ptr, 1, 3);
95}
96#endif
97
98#else
99extern inline void prefetch(const void *ptr)
100{
101 __asm__ ("ldl $31,%0" : : "m"(*(char *)ptr));
102}
103
104extern inline void prefetchw(const void *ptr)
105{
106 __asm__ ("ldq $31,%0" : : "m"(*(char *)ptr));
107}
108
109#ifdef CONFIG_SMP
110extern inline void spin_lock_prefetch(const void *ptr)
111{
112 __asm__ ("ldq $31,%0" : : "m"(*(char *)ptr));
113}
114#endif
115
116#endif /* GCC 3.1 */
117
118#endif /* __ASM_ALPHA_PROCESSOR_H */