Commit | Line | Data |
---|---|---|
1965aae3 PA |
1 | #ifndef _ASM_X86_PGTABLE_64_H |
2 | #define _ASM_X86_PGTABLE_64_H | |
1da177e4 | 3 | |
6df95fd7 | 4 | #include <linux/const.h> |
fb355149 JF |
5 | #include <asm/pgtable_64_types.h> |
6 | ||
9d291e78 VG |
7 | #ifndef __ASSEMBLY__ |
8 | ||
1da177e4 LT |
9 | /* |
10 | * This file contains the functions and defines necessary to modify and use | |
11 | * the x86-64 page table tree. | |
12 | */ | |
13 | #include <asm/processor.h> | |
1977f032 | 14 | #include <linux/bitops.h> |
1da177e4 | 15 | #include <linux/threads.h> |
1da177e4 LT |
16 | |
17 | extern pud_t level3_kernel_pgt[512]; | |
1da177e4 LT |
18 | extern pud_t level3_ident_pgt[512]; |
19 | extern pmd_t level2_kernel_pgt[512]; | |
084a2a4e JF |
20 | extern pmd_t level2_fixmap_pgt[512]; |
21 | extern pmd_t level2_ident_pgt[512]; | |
1da177e4 | 22 | extern pgd_t init_level4_pgt[]; |
1da177e4 | 23 | |
e3ebadd9 | 24 | #define swapper_pg_dir init_level4_pgt |
1da177e4 | 25 | |
1da177e4 | 26 | extern void paging_init(void); |
1da177e4 | 27 | |
7f94401e JP |
28 | #define pte_ERROR(e) \ |
29 | printk("%s:%d: bad pte %p(%016lx).\n", \ | |
30 | __FILE__, __LINE__, &(e), pte_val(e)) | |
31 | #define pmd_ERROR(e) \ | |
32 | printk("%s:%d: bad pmd %p(%016lx).\n", \ | |
33 | __FILE__, __LINE__, &(e), pmd_val(e)) | |
34 | #define pud_ERROR(e) \ | |
35 | printk("%s:%d: bad pud %p(%016lx).\n", \ | |
36 | __FILE__, __LINE__, &(e), pud_val(e)) | |
37 | #define pgd_ERROR(e) \ | |
38 | printk("%s:%d: bad pgd %p(%016lx).\n", \ | |
39 | __FILE__, __LINE__, &(e), pgd_val(e)) | |
1da177e4 | 40 | |
4891645e JF |
41 | struct mm_struct; |
42 | ||
0814e0ba EH |
43 | void set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte); |
44 | ||
45 | ||
4891645e JF |
46 | static inline void native_pte_clear(struct mm_struct *mm, unsigned long addr, |
47 | pte_t *ptep) | |
1da177e4 | 48 | { |
4891645e JF |
49 | *ptep = native_make_pte(0); |
50 | } | |
1da177e4 | 51 | |
4891645e | 52 | static inline void native_set_pte(pte_t *ptep, pte_t pte) |
1da177e4 | 53 | { |
4891645e JF |
54 | *ptep = pte; |
55 | } | |
1da177e4 | 56 | |
b65e6390 IM |
57 | static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte) |
58 | { | |
59 | native_set_pte(ptep, pte); | |
60 | } | |
61 | ||
db3eb96f AA |
62 | static inline void native_set_pmd(pmd_t *pmdp, pmd_t pmd) |
63 | { | |
64 | *pmdp = pmd; | |
65 | } | |
66 | ||
67 | static inline void native_pmd_clear(pmd_t *pmd) | |
68 | { | |
69 | native_set_pmd(pmd, native_make_pmd(0)); | |
70 | } | |
71 | ||
4891645e | 72 | static inline pte_t native_ptep_get_and_clear(pte_t *xp) |
1da177e4 | 73 | { |
4891645e JF |
74 | #ifdef CONFIG_SMP |
75 | return native_make_pte(xchg(&xp->pte, 0)); | |
76 | #else | |
7f94401e JP |
77 | /* native_local_ptep_get_and_clear, |
78 | but duplicated because of cyclic dependency */ | |
4891645e JF |
79 | pte_t ret = *xp; |
80 | native_pte_clear(NULL, 0, xp); | |
81 | return ret; | |
82 | #endif | |
1da177e4 LT |
83 | } |
84 | ||
db3eb96f | 85 | static inline pmd_t native_pmdp_get_and_clear(pmd_t *xp) |
1da177e4 | 86 | { |
db3eb96f AA |
87 | #ifdef CONFIG_SMP |
88 | return native_make_pmd(xchg(&xp->pmd, 0)); | |
89 | #else | |
90 | /* native_local_pmdp_get_and_clear, | |
91 | but duplicated because of cyclic dependency */ | |
92 | pmd_t ret = *xp; | |
93 | native_pmd_clear(xp); | |
94 | return ret; | |
95 | #endif | |
4891645e | 96 | } |
1da177e4 | 97 | |
4891645e | 98 | static inline void native_set_pud(pud_t *pudp, pud_t pud) |
1da177e4 | 99 | { |
4891645e | 100 | *pudp = pud; |
1da177e4 LT |
101 | } |
102 | ||
4891645e JF |
103 | static inline void native_pud_clear(pud_t *pud) |
104 | { | |
105 | native_set_pud(pud, native_make_pud(0)); | |
106 | } | |
61e06037 | 107 | |
4891645e JF |
108 | static inline void native_set_pgd(pgd_t *pgdp, pgd_t pgd) |
109 | { | |
110 | *pgdp = pgd; | |
111 | } | |
8c65b4a6 | 112 | |
7f94401e | 113 | static inline void native_pgd_clear(pgd_t *pgd) |
61e06037 | 114 | { |
4891645e | 115 | native_set_pgd(pgd, native_make_pgd(0)); |
61e06037 ZA |
116 | } |
117 | ||
6afb5157 HL |
118 | extern void sync_global_pgds(unsigned long start, unsigned long end); |
119 | ||
1da177e4 LT |
120 | /* |
121 | * Conversion functions: convert a page and protection to a page entry, | |
122 | * and a page entry and page directory to the page they refer to. | |
123 | */ | |
124 | ||
1da177e4 LT |
125 | /* |
126 | * Level 4 access. | |
127 | */ | |
e00fc542 | 128 | static inline int pgd_large(pgd_t pgd) { return 0; } |
e7a9b0b3 | 129 | #define mk_kernel_pgd(address) __pgd((address) | _KERNPG_TABLE) |
1da177e4 LT |
130 | |
131 | /* PUD - Level3 access */ | |
1da177e4 | 132 | |
1da177e4 | 133 | /* PMD - Level 2 access */ |
7f94401e JP |
134 | #define pte_to_pgoff(pte) ((pte_val((pte)) & PHYSICAL_PAGE_MASK) >> PAGE_SHIFT) |
135 | #define pgoff_to_pte(off) ((pte_t) { .pte = ((off) << PAGE_SHIFT) | \ | |
136 | _PAGE_FILE }) | |
1da177e4 LT |
137 | #define PTE_FILE_MAX_BITS __PHYSICAL_MASK_SHIFT |
138 | ||
139 | /* PTE - Level 1 access. */ | |
140 | ||
1da177e4 | 141 | /* x86-64 always has all page tables mapped. */ |
7f94401e | 142 | #define pte_offset_map(dir, address) pte_offset_kernel((dir), (address)) |
4e60c86b | 143 | #define pte_unmap(pte) ((void)(pte))/* NOP */ |
1da177e4 | 144 | |
4b3073e1 | 145 | #define update_mmu_cache(vma, address, ptep) do { } while (0) |
1da177e4 | 146 | |
1da177e4 | 147 | /* Encode and de-code a swap entry */ |
1796316a JB |
148 | #if _PAGE_BIT_FILE < _PAGE_BIT_PROTNONE |
149 | #define SWP_TYPE_BITS (_PAGE_BIT_FILE - _PAGE_BIT_PRESENT - 1) | |
150 | #define SWP_OFFSET_SHIFT (_PAGE_BIT_PROTNONE + 1) | |
151 | #else | |
152 | #define SWP_TYPE_BITS (_PAGE_BIT_PROTNONE - _PAGE_BIT_PRESENT - 1) | |
153 | #define SWP_OFFSET_SHIFT (_PAGE_BIT_FILE + 1) | |
154 | #endif | |
155 | ||
156 | #define MAX_SWAPFILES_CHECK() BUILD_BUG_ON(MAX_SWAPFILES_SHIFT > SWP_TYPE_BITS) | |
157 | ||
158 | #define __swp_type(x) (((x).val >> (_PAGE_BIT_PRESENT + 1)) \ | |
159 | & ((1U << SWP_TYPE_BITS) - 1)) | |
160 | #define __swp_offset(x) ((x).val >> SWP_OFFSET_SHIFT) | |
161 | #define __swp_entry(type, offset) ((swp_entry_t) { \ | |
162 | ((type) << (_PAGE_BIT_PRESENT + 1)) \ | |
163 | | ((offset) << SWP_OFFSET_SHIFT) }) | |
7f94401e | 164 | #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val((pte)) }) |
c8e5393a | 165 | #define __swp_entry_to_pte(x) ((pte_t) { .pte = (x).val }) |
1da177e4 | 166 | |
7f94401e | 167 | extern int kern_addr_valid(unsigned long addr); |
31eedd82 | 168 | extern void cleanup_highmap(void); |
1da177e4 | 169 | |
1da177e4 | 170 | #define HAVE_ARCH_UNMAPPED_AREA |
cc503c1b | 171 | #define HAVE_ARCH_UNMAPPED_AREA_TOPDOWN |
1da177e4 LT |
172 | |
173 | #define pgtable_cache_init() do { } while (0) | |
da8f153e | 174 | #define check_pgt_cache() do { } while (0) |
1da177e4 LT |
175 | |
176 | #define PAGE_AGP PAGE_KERNEL_NOCACHE | |
177 | #define HAVE_PAGE_AGP 1 | |
178 | ||
179 | /* fs/proc/kcore.c */ | |
180 | #define kc_vaddr_to_offset(v) ((v) & __VIRTUAL_MASK) | |
9063c61f | 181 | #define kc_offset_to_vaddr(o) ((o) | ~__VIRTUAL_MASK) |
1da177e4 | 182 | |
1da177e4 | 183 | #define __HAVE_ARCH_PTE_SAME |
5f6e8da7 AA |
184 | |
185 | #ifdef CONFIG_TRANSPARENT_HUGEPAGE | |
186 | static inline int pmd_trans_splitting(pmd_t pmd) | |
187 | { | |
188 | return pmd_val(pmd) & _PAGE_SPLITTING; | |
189 | } | |
190 | ||
191 | static inline int pmd_trans_huge(pmd_t pmd) | |
192 | { | |
193 | return pmd_val(pmd) & _PAGE_PSE; | |
194 | } | |
195 | #endif /* CONFIG_TRANSPARENT_HUGEPAGE */ | |
196 | ||
db3eb96f AA |
197 | #define mk_pmd(page, pgprot) pfn_pmd(page_to_pfn(page), (pgprot)) |
198 | ||
199 | #define __HAVE_ARCH_PMDP_SET_ACCESS_FLAGS | |
200 | extern int pmdp_set_access_flags(struct vm_area_struct *vma, | |
201 | unsigned long address, pmd_t *pmdp, | |
202 | pmd_t entry, int dirty); | |
203 | ||
204 | #define __HAVE_ARCH_PMDP_TEST_AND_CLEAR_YOUNG | |
205 | extern int pmdp_test_and_clear_young(struct vm_area_struct *vma, | |
206 | unsigned long addr, pmd_t *pmdp); | |
207 | ||
208 | #define __HAVE_ARCH_PMDP_CLEAR_YOUNG_FLUSH | |
209 | extern int pmdp_clear_flush_young(struct vm_area_struct *vma, | |
210 | unsigned long address, pmd_t *pmdp); | |
211 | ||
212 | ||
213 | #define __HAVE_ARCH_PMDP_SPLITTING_FLUSH | |
214 | extern void pmdp_splitting_flush(struct vm_area_struct *vma, | |
215 | unsigned long addr, pmd_t *pmdp); | |
216 | ||
217 | #define __HAVE_ARCH_PMD_WRITE | |
218 | static inline int pmd_write(pmd_t pmd) | |
219 | { | |
220 | return pmd_flags(pmd) & _PAGE_RW; | |
221 | } | |
222 | ||
223 | #define __HAVE_ARCH_PMDP_GET_AND_CLEAR | |
224 | static inline pmd_t pmdp_get_and_clear(struct mm_struct *mm, unsigned long addr, | |
225 | pmd_t *pmdp) | |
226 | { | |
227 | pmd_t pmd = native_pmdp_get_and_clear(pmdp); | |
228 | pmd_update(mm, addr, pmdp); | |
229 | return pmd; | |
230 | } | |
231 | ||
232 | #define __HAVE_ARCH_PMDP_SET_WRPROTECT | |
233 | static inline void pmdp_set_wrprotect(struct mm_struct *mm, | |
234 | unsigned long addr, pmd_t *pmdp) | |
235 | { | |
236 | clear_bit(_PAGE_BIT_RW, (unsigned long *)&pmdp->pmd); | |
237 | pmd_update(mm, addr, pmdp); | |
238 | } | |
239 | ||
240 | static inline int pmd_young(pmd_t pmd) | |
241 | { | |
242 | return pmd_flags(pmd) & _PAGE_ACCESSED; | |
243 | } | |
244 | ||
245 | static inline pmd_t pmd_set_flags(pmd_t pmd, pmdval_t set) | |
246 | { | |
247 | pmdval_t v = native_pmd_val(pmd); | |
248 | ||
249 | return native_make_pmd(v | set); | |
250 | } | |
251 | ||
252 | static inline pmd_t pmd_clear_flags(pmd_t pmd, pmdval_t clear) | |
253 | { | |
254 | pmdval_t v = native_pmd_val(pmd); | |
255 | ||
256 | return native_make_pmd(v & ~clear); | |
257 | } | |
258 | ||
259 | static inline pmd_t pmd_mkold(pmd_t pmd) | |
260 | { | |
261 | return pmd_clear_flags(pmd, _PAGE_ACCESSED); | |
262 | } | |
263 | ||
264 | static inline pmd_t pmd_wrprotect(pmd_t pmd) | |
265 | { | |
266 | return pmd_clear_flags(pmd, _PAGE_RW); | |
267 | } | |
268 | ||
269 | static inline pmd_t pmd_mkdirty(pmd_t pmd) | |
270 | { | |
271 | return pmd_set_flags(pmd, _PAGE_DIRTY); | |
272 | } | |
273 | ||
274 | static inline pmd_t pmd_mkhuge(pmd_t pmd) | |
275 | { | |
276 | return pmd_set_flags(pmd, _PAGE_PSE); | |
277 | } | |
278 | ||
279 | static inline pmd_t pmd_mkyoung(pmd_t pmd) | |
280 | { | |
281 | return pmd_set_flags(pmd, _PAGE_ACCESSED); | |
282 | } | |
283 | ||
284 | static inline pmd_t pmd_mkwrite(pmd_t pmd) | |
285 | { | |
286 | return pmd_set_flags(pmd, _PAGE_RW); | |
287 | } | |
288 | ||
9d291e78 | 289 | #endif /* !__ASSEMBLY__ */ |
1da177e4 | 290 | |
1965aae3 | 291 | #endif /* _ASM_X86_PGTABLE_64_H */ |