| #ifndef _BLACKFIN_PGTABLE_H |
| #define _BLACKFIN_PGTABLE_H |
| |
| #include <asm-generic/4level-fixup.h> |
| |
| #include <asm/page.h> |
| #include <asm/cplb.h> |
| |
| typedef pte_t *pte_addr_t; |
| /* |
| * Trivial page table functions. |
| */ |
| #define pgd_present(pgd) (1) |
| #define pgd_none(pgd) (0) |
| #define pgd_bad(pgd) (0) |
| #define pgd_clear(pgdp) |
| #define kern_addr_valid(addr) (1) |
| |
| #define pmd_offset(a, b) ((void *)0) |
| #define pmd_none(x) (!pmd_val(x)) |
| #define pmd_present(x) (pmd_val(x)) |
| #define pmd_clear(xp) do { set_pmd(xp, __pmd(0)); } while (0) |
| #define pmd_bad(x) (pmd_val(x) & ~PAGE_MASK) |
| |
| #define kern_addr_valid(addr) (1) |
| |
| #define PAGE_NONE __pgprot(0) /* these mean nothing to NO_MM */ |
| #define PAGE_SHARED __pgprot(0) /* these mean nothing to NO_MM */ |
| #define PAGE_COPY __pgprot(0) /* these mean nothing to NO_MM */ |
| #define PAGE_READONLY __pgprot(0) /* these mean nothing to NO_MM */ |
| #define PAGE_KERNEL __pgprot(0) /* these mean nothing to NO_MM */ |
| |
| extern void paging_init(void); |
| |
| #define __swp_type(x) (0) |
| #define __swp_offset(x) (0) |
| #define __swp_entry(typ,off) ((swp_entry_t) { ((typ) | ((off) << 7)) }) |
| #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) |
| #define __swp_entry_to_pte(x) ((pte_t) { (x).val }) |
| |
| static inline int pte_file(pte_t pte) |
| { |
| return 0; |
| } |
| |
| #define set_pte(pteptr, pteval) (*(pteptr) = pteval) |
| #define set_pte_at(mm, addr, ptep, pteval) set_pte(ptep, pteval) |
| |
| /* |
| * Page assess control based on Blackfin CPLB management |
| */ |
| #define _PAGE_RD (CPLB_USER_RD) |
| #define _PAGE_WR (CPLB_USER_WR) |
| #define _PAGE_USER (CPLB_USER_RD | CPLB_USER_WR) |
| #define _PAGE_ACCESSED CPLB_ALL_ACCESS |
| #define _PAGE_DIRTY (CPLB_DIRTY) |
| |
| #define PTE_BIT_FUNC(fn, op) \ |
| static inline pte_t pte_##fn(pte_t _pte) { _pte.pte op; return _pte; } |
| |
| PTE_BIT_FUNC(rdprotect, &= ~_PAGE_RD); |
| PTE_BIT_FUNC(mkread, |= _PAGE_RD); |
| PTE_BIT_FUNC(wrprotect, &= ~_PAGE_WR); |
| PTE_BIT_FUNC(mkwrite, |= _PAGE_WR); |
| PTE_BIT_FUNC(exprotect, &= ~_PAGE_USER); |
| PTE_BIT_FUNC(mkexec, |= _PAGE_USER); |
| PTE_BIT_FUNC(mkclean, &= ~_PAGE_DIRTY); |
| PTE_BIT_FUNC(mkdirty, |= _PAGE_DIRTY); |
| PTE_BIT_FUNC(mkold, &= ~_PAGE_ACCESSED); |
| PTE_BIT_FUNC(mkyoung, |= _PAGE_ACCESSED); |
| |
| /* |
| * ZERO_PAGE is a global shared page that is always zero: used |
| * for zero-mapped memory areas etc.. |
| */ |
| #define ZERO_PAGE(vaddr) (virt_to_page(0)) |
| |
| extern unsigned int kobjsize(const void *objp); |
| |
| #define swapper_pg_dir ((pgd_t *) 0) |
| /* |
| * No page table caches to initialise. |
| */ |
| #define pgtable_cache_init() do { } while (0) |
| #define io_remap_pfn_range remap_pfn_range |
| |
| /* |
| * All 32bit addresses are effectively valid for vmalloc... |
| * Sort of meaningless for non-VM targets. |
| */ |
| #define VMALLOC_START 0 |
| #define VMALLOC_END 0xffffffff |
| |
| #include <asm-generic/pgtable.h> |
| |
| #endif /* _BLACKFIN_PGTABLE_H */ |