]>
Commit | Line | Data |
---|---|---|
26b6a3d9 AK |
1 | #ifndef _ASM_POWERPC_BOOK3S_64_HASH_H |
2 | #define _ASM_POWERPC_BOOK3S_64_HASH_H | |
c605782b BH |
3 | #ifdef __KERNEL__ |
4 | ||
e34aa03c AK |
5 | /* |
6 | * Common bits between 4K and 64K pages in a linux-style PTE. | |
1ec3f937 | 7 | * Additional bits may be defined in pgtable-hash64-*.h |
e34aa03c AK |
8 | * |
9 | * Note: We only support user read/write permissions. Supervisor always | |
10 | * have full read/write to pages above PAGE_OFFSET (pages below that | |
11 | * always use the user access permissions). | |
12 | * | |
13 | * We could create separate kernel read-only if we used the 3 PP bits | |
14 | * combinations that newer processors provide but we currently don't. | |
15 | */ | |
945537df | 16 | #define H_PAGE_BUSY 0x00800 /* software: PTE & hash are busy */ |
d2cf0050 | 17 | #define H_PTE_NONE_MASK _PAGE_HPTEFLAGS |
945537df AK |
18 | #define H_PAGE_F_GIX_SHIFT 57 |
19 | #define H_PAGE_F_GIX (7ul << 57) /* HPTE index within HPTEG */ | |
20 | #define H_PAGE_F_SECOND (1ul << 60) /* HPTE is in 2ndary HPTEG */ | |
21 | #define H_PAGE_HASHPTE (1ul << 61) /* PTE has associated HPTE */ | |
e34aa03c | 22 | |
371352ca AK |
23 | #ifdef CONFIG_PPC_64K_PAGES |
24 | #include <asm/book3s/64/hash-64k.h> | |
25 | #else | |
26 | #include <asm/book3s/64/hash-4k.h> | |
27 | #endif | |
28 | ||
29 | /* | |
30 | * Size of EA range mapped by our pagetables. | |
31 | */ | |
dd1842a2 AK |
32 | #define H_PGTABLE_EADDR_SIZE (H_PTE_INDEX_SIZE + H_PMD_INDEX_SIZE + \ |
33 | H_PUD_INDEX_SIZE + H_PGD_INDEX_SIZE + PAGE_SHIFT) | |
34 | #define H_PGTABLE_RANGE (ASM_CONST(1) << H_PGTABLE_EADDR_SIZE) | |
371352ca AK |
35 | |
36 | #ifdef CONFIG_TRANSPARENT_HUGEPAGE | |
dd1842a2 AK |
37 | /* |
38 | * only with hash we need to use the second half of pmd page table | |
39 | * to store pointer to deposited pgtable_t | |
40 | */ | |
41 | #define H_PMD_CACHE_INDEX (H_PMD_INDEX_SIZE + 1) | |
371352ca | 42 | #else |
dd1842a2 | 43 | #define H_PMD_CACHE_INDEX H_PMD_INDEX_SIZE |
371352ca AK |
44 | #endif |
45 | /* | |
46 | * Define the address range of the kernel non-linear virtual area | |
47 | */ | |
d6a9996e AK |
48 | #define H_KERN_VIRT_START ASM_CONST(0xD000000000000000) |
49 | #define H_KERN_VIRT_SIZE ASM_CONST(0x0000100000000000) | |
371352ca AK |
50 | |
51 | /* | |
52 | * The vmalloc space starts at the beginning of that region, and | |
53 | * occupies half of it on hash CPUs and a quarter of it on Book3E | |
54 | * (we keep a quarter for the virtual memmap) | |
55 | */ | |
d6a9996e AK |
56 | #define H_VMALLOC_START H_KERN_VIRT_START |
57 | #define H_VMALLOC_SIZE (H_KERN_VIRT_SIZE >> 1) | |
58 | #define H_VMALLOC_END (H_VMALLOC_START + H_VMALLOC_SIZE) | |
371352ca AK |
59 | |
60 | /* | |
61 | * Region IDs | |
62 | */ | |
63 | #define REGION_SHIFT 60UL | |
64 | #define REGION_MASK (0xfUL << REGION_SHIFT) | |
65 | #define REGION_ID(ea) (((unsigned long)(ea)) >> REGION_SHIFT) | |
66 | ||
d6a9996e | 67 | #define VMALLOC_REGION_ID (REGION_ID(H_VMALLOC_START)) |
371352ca AK |
68 | #define KERNEL_REGION_ID (REGION_ID(PAGE_OFFSET)) |
69 | #define VMEMMAP_REGION_ID (0xfUL) /* Server only */ | |
70 | #define USER_REGION_ID (0UL) | |
71 | ||
72 | /* | |
73 | * Defines the address of the vmemap area, in its own region on | |
74 | * hash table CPUs. | |
75 | */ | |
d6a9996e | 76 | #define H_VMEMMAP_BASE (VMEMMAP_REGION_ID << REGION_SHIFT) |
371352ca AK |
77 | |
78 | #ifdef CONFIG_PPC_MM_SLICES | |
79 | #define HAVE_ARCH_UNMAPPED_AREA | |
80 | #define HAVE_ARCH_UNMAPPED_AREA_TOPDOWN | |
81 | #endif /* CONFIG_PPC_MM_SLICES */ | |
8d1cf34e | 82 | |
c605782b BH |
83 | |
84 | /* PTEIDX nibble */ | |
85 | #define _PTEIDX_SECONDARY 0x8 | |
86 | #define _PTEIDX_GROUP_IX 0x7 | |
87 | ||
ac94ac79 AK |
88 | #define H_PMD_BAD_BITS (PTE_TABLE_SIZE-1) |
89 | #define H_PUD_BAD_BITS (PMD_TABLE_SIZE-1) | |
371352ca AK |
90 | |
91 | #ifndef __ASSEMBLY__ | |
ac94ac79 AK |
92 | #define hash__pmd_bad(pmd) (pmd_val(pmd) & H_PMD_BAD_BITS) |
93 | #define hash__pud_bad(pud) (pud_val(pud) & H_PUD_BAD_BITS) | |
94 | static inline int hash__pgd_bad(pgd_t pgd) | |
95 | { | |
96 | return (pgd_val(pgd) == 0); | |
97 | } | |
371352ca AK |
98 | |
99 | extern void hpte_need_flush(struct mm_struct *mm, unsigned long addr, | |
100 | pte_t *ptep, unsigned long pte, int huge); | |
c6a3c495 | 101 | extern unsigned long htab_convert_pte_flags(unsigned long pteflags); |
371352ca | 102 | /* Atomic PTE updates */ |
ac94ac79 AK |
103 | static inline unsigned long hash__pte_update(struct mm_struct *mm, |
104 | unsigned long addr, | |
105 | pte_t *ptep, unsigned long clr, | |
106 | unsigned long set, | |
107 | int huge) | |
371352ca | 108 | { |
5dc1ef85 AK |
109 | __be64 old_be, tmp_be; |
110 | unsigned long old; | |
371352ca AK |
111 | |
112 | __asm__ __volatile__( | |
113 | "1: ldarx %0,0,%3 # pte_update\n\ | |
5dc1ef85 | 114 | and. %1,%0,%6\n\ |
371352ca AK |
115 | bne- 1b \n\ |
116 | andc %1,%0,%4 \n\ | |
117 | or %1,%1,%7\n\ | |
118 | stdcx. %1,0,%3 \n\ | |
119 | bne- 1b" | |
5dc1ef85 AK |
120 | : "=&r" (old_be), "=&r" (tmp_be), "=m" (*ptep) |
121 | : "r" (ptep), "r" (cpu_to_be64(clr)), "m" (*ptep), | |
945537df | 122 | "r" (cpu_to_be64(H_PAGE_BUSY)), "r" (cpu_to_be64(set)) |
371352ca AK |
123 | : "cc" ); |
124 | /* huge pages use the old page table lock */ | |
125 | if (!huge) | |
126 | assert_pte_locked(mm, addr); | |
127 | ||
5dc1ef85 | 128 | old = be64_to_cpu(old_be); |
945537df | 129 | if (old & H_PAGE_HASHPTE) |
371352ca AK |
130 | hpte_need_flush(mm, addr, ptep, old, huge); |
131 | ||
132 | return old; | |
133 | } | |
134 | ||
371352ca AK |
135 | /* Set the dirty and/or accessed bits atomically in a linux PTE, this |
136 | * function doesn't need to flush the hash entry | |
137 | */ | |
ac94ac79 | 138 | static inline void hash__ptep_set_access_flags(pte_t *ptep, pte_t entry) |
371352ca | 139 | { |
5dc1ef85 AK |
140 | __be64 old, tmp, val, mask; |
141 | ||
c7d54842 | 142 | mask = cpu_to_be64(_PAGE_DIRTY | _PAGE_ACCESSED | _PAGE_READ | _PAGE_WRITE | |
5dc1ef85 | 143 | _PAGE_EXEC | _PAGE_SOFT_DIRTY); |
371352ca | 144 | |
5dc1ef85 | 145 | val = pte_raw(entry) & mask; |
371352ca AK |
146 | |
147 | __asm__ __volatile__( | |
148 | "1: ldarx %0,0,%4\n\ | |
5dc1ef85 | 149 | and. %1,%0,%6\n\ |
371352ca AK |
150 | bne- 1b \n\ |
151 | or %0,%3,%0\n\ | |
152 | stdcx. %0,0,%4\n\ | |
153 | bne- 1b" | |
154 | :"=&r" (old), "=&r" (tmp), "=m" (*ptep) | |
945537df | 155 | :"r" (val), "r" (ptep), "m" (*ptep), "r" (cpu_to_be64(H_PAGE_BUSY)) |
371352ca AK |
156 | :"cc"); |
157 | } | |
158 | ||
ac94ac79 | 159 | static inline int hash__pte_same(pte_t pte_a, pte_t pte_b) |
368ced78 | 160 | { |
ac94ac79 | 161 | return (((pte_raw(pte_a) ^ pte_raw(pte_b)) & ~cpu_to_be64(_PAGE_HPTEFLAGS)) == 0); |
368ced78 AK |
162 | } |
163 | ||
ac94ac79 | 164 | static inline int hash__pte_none(pte_t pte) |
ee3caed3 | 165 | { |
ac94ac79 | 166 | return (pte_val(pte) & ~H_PTE_NONE_MASK) == 0; |
ee3caed3 ME |
167 | } |
168 | ||
1ca72129 AK |
169 | /* This low level function performs the actual PTE insertion |
170 | * Setting the PTE depends on the MMU type and other factors. It's | |
171 | * an horrible mess that I'm not going to try to clean up now but | |
172 | * I'm keeping it in one place rather than spread around | |
173 | */ | |
ac94ac79 AK |
174 | static inline void hash__set_pte_at(struct mm_struct *mm, unsigned long addr, |
175 | pte_t *ptep, pte_t pte, int percpu) | |
1ca72129 AK |
176 | { |
177 | /* | |
178 | * Anything else just stores the PTE normally. That covers all 64-bit | |
179 | * cases, and 32-bit non-hash with 32-bit PTEs. | |
180 | */ | |
181 | *ptep = pte; | |
182 | } | |
183 | ||
371352ca AK |
184 | #ifdef CONFIG_TRANSPARENT_HUGEPAGE |
185 | extern void hpte_do_hugepage_flush(struct mm_struct *mm, unsigned long addr, | |
186 | pmd_t *pmdp, unsigned long old_pmd); | |
187 | #else | |
188 | static inline void hpte_do_hugepage_flush(struct mm_struct *mm, | |
189 | unsigned long addr, pmd_t *pmdp, | |
190 | unsigned long old_pmd) | |
191 | { | |
192 | WARN(1, "%s called with THP disabled\n", __func__); | |
193 | } | |
194 | #endif /* CONFIG_TRANSPARENT_HUGEPAGE */ | |
195 | ||
31a14fae AK |
196 | |
197 | extern int hash__map_kernel_page(unsigned long ea, unsigned long pa, | |
198 | unsigned long flags); | |
199 | extern int __meminit hash__vmemmap_create_mapping(unsigned long start, | |
200 | unsigned long page_size, | |
201 | unsigned long phys); | |
202 | extern void hash__vmemmap_remove_mapping(unsigned long start, | |
203 | unsigned long page_size); | |
32b53c01 RA |
204 | |
205 | int hash__create_section_mapping(unsigned long start, unsigned long end); | |
206 | int hash__remove_section_mapping(unsigned long start, unsigned long end); | |
207 | ||
371352ca | 208 | #endif /* !__ASSEMBLY__ */ |
c605782b | 209 | #endif /* __KERNEL__ */ |
26b6a3d9 | 210 | #endif /* _ASM_POWERPC_BOOK3S_64_HASH_H */ |