Thomas Gleixner | b2bba72 | 2007-10-15 23:28:20 +0200 | [diff] [blame] | 1 | #ifndef _ASM_X86_CACHEFLUSH_H |
| 2 | #define _ASM_X86_CACHEFLUSH_H |
| 3 | |
| 4 | /* Keep includes the same across arches. */ |
| 5 | #include <linux/mm.h> |
| 6 | |
| 7 | /* Caches aren't brain-dead on the intel. */ |
| 8 | #define flush_cache_all() do { } while (0) |
| 9 | #define flush_cache_mm(mm) do { } while (0) |
| 10 | #define flush_cache_dup_mm(mm) do { } while (0) |
| 11 | #define flush_cache_range(vma, start, end) do { } while (0) |
| 12 | #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) |
| 13 | #define flush_dcache_page(page) do { } while (0) |
| 14 | #define flush_dcache_mmap_lock(mapping) do { } while (0) |
| 15 | #define flush_dcache_mmap_unlock(mapping) do { } while (0) |
| 16 | #define flush_icache_range(start, end) do { } while (0) |
| 17 | #define flush_icache_page(vma,pg) do { } while (0) |
| 18 | #define flush_icache_user_range(vma,pg,adr,len) do { } while (0) |
| 19 | #define flush_cache_vmap(start, end) do { } while (0) |
| 20 | #define flush_cache_vunmap(start, end) do { } while (0) |
| 21 | |
| 22 | #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ |
| 23 | memcpy(dst, src, len) |
| 24 | #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ |
| 25 | memcpy(dst, src, len) |
| 26 | |
| 27 | void global_flush_tlb(void); |
Arjan van de Ven | e1271f6 | 2008-01-30 13:34:06 +0100 | [diff] [blame] | 28 | int __deprecated_for_modules change_page_attr(struct page *page, int numpages, |
| 29 | pgprot_t prot); |
Thomas Gleixner | b2bba72 | 2007-10-15 23:28:20 +0200 | [diff] [blame] | 30 | int change_page_attr_addr(unsigned long addr, int numpages, pgprot_t prot); |
Arjan van de Ven | 75cbade | 2008-01-30 13:34:06 +0100 | [diff] [blame] | 31 | |
| 32 | int set_pages_uc(struct page *page, int numpages); |
| 33 | int set_pages_wb(struct page *page, int numpages); |
| 34 | int set_pages_x(struct page *page, int numpages); |
| 35 | int set_pages_nx(struct page *page, int numpages); |
| 36 | int set_pages_ro(struct page *page, int numpages); |
| 37 | int set_pages_rw(struct page *page, int numpages); |
| 38 | |
| 39 | int set_memory_uc(unsigned long addr, int numpages); |
| 40 | int set_memory_wb(unsigned long addr, int numpages); |
| 41 | int set_memory_x(unsigned long addr, int numpages); |
| 42 | int set_memory_nx(unsigned long addr, int numpages); |
| 43 | int set_memory_ro(unsigned long addr, int numpages); |
| 44 | int set_memory_rw(unsigned long addr, int numpages); |
Ingo Molnar | f62d0f0 | 2008-01-30 13:34:07 +0100 | [diff] [blame^] | 45 | int set_memory_np(unsigned long addr, int numpages); |
Arjan van de Ven | 75cbade | 2008-01-30 13:34:06 +0100 | [diff] [blame] | 46 | |
Keshavamurthy, Anil S | a9c55b3 | 2007-10-21 16:41:47 -0700 | [diff] [blame] | 47 | void clflush_cache_range(void *addr, int size); |
Thomas Gleixner | b2bba72 | 2007-10-15 23:28:20 +0200 | [diff] [blame] | 48 | |
Thomas Gleixner | b2bba72 | 2007-10-15 23:28:20 +0200 | [diff] [blame] | 49 | #ifdef CONFIG_DEBUG_RODATA |
| 50 | void mark_rodata_ro(void); |
| 51 | #endif |
| 52 | |
Thomas Gleixner | 96a388d | 2007-10-11 11:20:03 +0200 | [diff] [blame] | 53 | #endif |