blob: 7af1e1109c49e1f5092f44914601f98624b96a5a [file] [log] [blame]
Linus Torvalds1da177e2005-04-16 15:20:36 -07001#ifndef _SPARC64_TLB_H
2#define _SPARC64_TLB_H
3
Linus Torvalds1da177e2005-04-16 15:20:36 -07004#include <linux/swap.h>
5#include <asm/pgalloc.h>
6#include <asm/tlbflush.h>
7#include <asm/mmu_context.h>
8
9#define TLB_BATCH_NR 192
10
11/*
12 * For UP we don't need to worry about TLB flush
13 * and page free order so much..
14 */
15#ifdef CONFIG_SMP
16 #define FREE_PTE_NR 506
17 #define tlb_fast_mode(bp) ((bp)->pages_nr == ~0U)
18#else
19 #define FREE_PTE_NR 1
20 #define tlb_fast_mode(bp) 1
21#endif
22
23struct mmu_gather {
24 struct mm_struct *mm;
25 unsigned int pages_nr;
26 unsigned int need_flush;
Hugh Dickins4d6ddfa2005-10-29 18:16:02 -070027 unsigned int fullmm;
Linus Torvalds1da177e2005-04-16 15:20:36 -070028 unsigned int tlb_nr;
Linus Torvalds1da177e2005-04-16 15:20:36 -070029 unsigned long vaddrs[TLB_BATCH_NR];
30 struct page *pages[FREE_PTE_NR];
31};
32
33DECLARE_PER_CPU(struct mmu_gather, mmu_gathers);
34
35#ifdef CONFIG_SMP
36extern void smp_flush_tlb_pending(struct mm_struct *,
37 unsigned long, unsigned long *);
38#endif
39
40extern void __flush_tlb_pending(unsigned long, unsigned long, unsigned long *);
41extern void flush_tlb_pending(void);
42
43static inline struct mmu_gather *tlb_gather_mmu(struct mm_struct *mm, unsigned int full_mm_flush)
44{
Hugh Dickins15a23ff2005-10-29 18:16:01 -070045 struct mmu_gather *mp = &get_cpu_var(mmu_gathers);
Linus Torvalds1da177e2005-04-16 15:20:36 -070046
47 BUG_ON(mp->tlb_nr);
48
49 mp->mm = mm;
50 mp->pages_nr = num_online_cpus() > 1 ? 0U : ~0U;
Hugh Dickins4d6ddfa2005-10-29 18:16:02 -070051 mp->fullmm = full_mm_flush;
Linus Torvalds1da177e2005-04-16 15:20:36 -070052
53 return mp;
54}
55
56
57static inline void tlb_flush_mmu(struct mmu_gather *mp)
58{
59 if (mp->need_flush) {
Hugh Dickins59871bc2005-11-07 14:12:08 -080060 free_pages_and_swap_cache(mp->pages, mp->pages_nr);
61 mp->pages_nr = 0;
Linus Torvalds1da177e2005-04-16 15:20:36 -070062 mp->need_flush = 0;
Linus Torvalds1da177e2005-04-16 15:20:36 -070063 }
64
65}
66
67#ifdef CONFIG_SMP
68extern void smp_flush_tlb_mm(struct mm_struct *mm);
69#define do_flush_tlb_mm(mm) smp_flush_tlb_mm(mm)
70#else
71#define do_flush_tlb_mm(mm) __flush_tlb_mm(CTX_HWBITS(mm->context), SECONDARY_CONTEXT)
72#endif
73
74static inline void tlb_finish_mmu(struct mmu_gather *mp, unsigned long start, unsigned long end)
75{
Linus Torvalds1da177e2005-04-16 15:20:36 -070076 tlb_flush_mmu(mp);
77
David S. Miller62dbec72005-11-07 14:09:58 -080078 if (mp->fullmm)
Hugh Dickins4d6ddfa2005-10-29 18:16:02 -070079 mp->fullmm = 0;
David S. Miller62dbec72005-11-07 14:09:58 -080080 else
Linus Torvalds1da177e2005-04-16 15:20:36 -070081 flush_tlb_pending();
82
83 /* keep the page table cache within bounds */
84 check_pgt_cache();
Hugh Dickins15a23ff2005-10-29 18:16:01 -070085
86 put_cpu_var(mmu_gathers);
Linus Torvalds1da177e2005-04-16 15:20:36 -070087}
88
Linus Torvalds1da177e2005-04-16 15:20:36 -070089static inline void tlb_remove_page(struct mmu_gather *mp, struct page *page)
90{
Linus Torvalds1da177e2005-04-16 15:20:36 -070091 if (tlb_fast_mode(mp)) {
92 free_page_and_swap_cache(page);
93 return;
94 }
Hugh Dickins59871bc2005-11-07 14:12:08 -080095 mp->need_flush = 1;
Linus Torvalds1da177e2005-04-16 15:20:36 -070096 mp->pages[mp->pages_nr++] = page;
97 if (mp->pages_nr >= FREE_PTE_NR)
98 tlb_flush_mmu(mp);
99}
100
101#define tlb_remove_tlb_entry(mp,ptep,addr) do { } while (0)
102#define pte_free_tlb(mp,ptepage) pte_free(ptepage)
103#define pmd_free_tlb(mp,pmdp) pmd_free(pmdp)
104#define pud_free_tlb(tlb,pudp) __pud_free_tlb(tlb,pudp)
105
106#define tlb_migrate_finish(mm) do { } while (0)
107#define tlb_start_vma(tlb, vma) do { } while (0)
108#define tlb_end_vma(tlb, vma) do { } while (0)
109
110#endif /* _SPARC64_TLB_H */