]> git.proxmox.com Git - mirror_ubuntu-bionic-kernel.git/blame - arch/metag/mm/hugetlbpage.c
Merge tag 'platform-drivers-x86-v4.15-2' of git://git.infradead.org/linux-platform...
[mirror_ubuntu-bionic-kernel.git] / arch / metag / mm / hugetlbpage.c
CommitLineData
b2441318 1// SPDX-License-Identifier: GPL-2.0
e624e95b
JH
2/*
3 * arch/metag/mm/hugetlbpage.c
4 *
5 * METAG HugeTLB page support.
6 *
7 * Cloned from SuperH
8 *
9 * Cloned from sparc64 by Paul Mundt.
10 *
11 * Copyright (C) 2002, 2003 David S. Miller (davem@redhat.com)
12 */
13
14#include <linux/init.h>
15#include <linux/fs.h>
16#include <linux/mm.h>
17#include <linux/hugetlb.h>
18#include <linux/pagemap.h>
19#include <linux/sysctl.h>
20
21#include <asm/mman.h>
22#include <asm/pgalloc.h>
23#include <asm/tlb.h>
24#include <asm/tlbflush.h>
25#include <asm/cacheflush.h>
26
27/*
28 * If the arch doesn't supply something else, assume that hugepage
29 * size aligned regions are ok without further preparation.
30 */
31int prepare_hugepage_range(struct file *file, unsigned long addr,
32 unsigned long len)
33{
34 struct mm_struct *mm = current->mm;
35 struct hstate *h = hstate_file(file);
36 struct vm_area_struct *vma;
37
38 if (len & ~huge_page_mask(h))
39 return -EINVAL;
40 if (addr & ~huge_page_mask(h))
41 return -EINVAL;
42 if (TASK_SIZE - len < addr)
43 return -EINVAL;
44
45 vma = find_vma(mm, ALIGN_HUGEPT(addr));
46 if (vma && !(vma->vm_flags & MAP_HUGETLB))
47 return -EINVAL;
48
49 vma = find_vma(mm, addr);
50 if (vma) {
51 if (addr + len > vma->vm_start)
52 return -EINVAL;
53 if (!(vma->vm_flags & MAP_HUGETLB) &&
54 (ALIGN_HUGEPT(addr + len) > vma->vm_start))
55 return -EINVAL;
56 }
57 return 0;
58}
59
60pte_t *huge_pte_alloc(struct mm_struct *mm,
61 unsigned long addr, unsigned long sz)
62{
63 pgd_t *pgd;
64 pud_t *pud;
65 pmd_t *pmd;
66 pte_t *pte;
67
68 pgd = pgd_offset(mm, addr);
69 pud = pud_offset(pgd, addr);
70 pmd = pmd_offset(pud, addr);
3ed3a4f0 71 pte = pte_alloc_map(mm, pmd, addr);
e624e95b
JH
72 pgd->pgd &= ~_PAGE_SZ_MASK;
73 pgd->pgd |= _PAGE_SZHUGE;
74
75 return pte;
76}
77
7868a208
PA
78pte_t *huge_pte_offset(struct mm_struct *mm,
79 unsigned long addr, unsigned long sz)
e624e95b
JH
80{
81 pgd_t *pgd;
82 pud_t *pud;
83 pmd_t *pmd;
84 pte_t *pte = NULL;
85
86 pgd = pgd_offset(mm, addr);
87 pud = pud_offset(pgd, addr);
88 pmd = pmd_offset(pud, addr);
89 pte = pte_offset_kernel(pmd, addr);
90
91 return pte;
92}
93
e624e95b
JH
94int pmd_huge(pmd_t pmd)
95{
96 return pmd_page_shift(pmd) > PAGE_SHIFT;
97}
98
99int pud_huge(pud_t pud)
100{
101 return 0;
102}
103
104struct page *follow_huge_pmd(struct mm_struct *mm, unsigned long address,
105 pmd_t *pmd, int write)
106{
107 return NULL;
108}
109
110#ifdef HAVE_ARCH_HUGETLB_UNMAPPED_AREA
111
112/*
113 * Look for an unmapped area starting after another hugetlb vma.
114 * There are guaranteed to be no huge pte's spare if all the huge pages are
115 * full size (4MB), so in that case compile out this search.
116 */
117#if HPAGE_SHIFT == HUGEPT_SHIFT
118static inline unsigned long
119hugetlb_get_unmapped_area_existing(unsigned long len)
120{
121 return 0;
122}
123#else
124static unsigned long
125hugetlb_get_unmapped_area_existing(unsigned long len)
126{
127 struct mm_struct *mm = current->mm;
128 struct vm_area_struct *vma;
129 unsigned long start_addr, addr;
130 int after_huge;
131
132 if (mm->context.part_huge) {
133 start_addr = mm->context.part_huge;
134 after_huge = 1;
135 } else {
136 start_addr = TASK_UNMAPPED_BASE;
137 after_huge = 0;
138 }
139new_search:
140 addr = start_addr;
141
142 for (vma = find_vma(mm, addr); ; vma = vma->vm_next) {
143 if ((!vma && !after_huge) || TASK_SIZE - len < addr) {
144 /*
145 * Start a new search - just in case we missed
146 * some holes.
147 */
148 if (start_addr != TASK_UNMAPPED_BASE) {
149 start_addr = TASK_UNMAPPED_BASE;
150 goto new_search;
151 }
152 return 0;
153 }
154 /* skip ahead if we've aligned right over some vmas */
155 if (vma && vma->vm_end <= addr)
156 continue;
157 /* space before the next vma? */
158 if (after_huge && (!vma || ALIGN_HUGEPT(addr + len)
159 <= vma->vm_start)) {
160 unsigned long end = addr + len;
161 if (end & HUGEPT_MASK)
162 mm->context.part_huge = end;
163 else if (addr == mm->context.part_huge)
164 mm->context.part_huge = 0;
165 return addr;
166 }
8bc613be 167 if (vma->vm_flags & MAP_HUGETLB) {
e624e95b
JH
168 /* space after a huge vma in 2nd level page table? */
169 if (vma->vm_end & HUGEPT_MASK) {
170 after_huge = 1;
171 /* no need to align to the next PT block */
172 addr = vma->vm_end;
173 continue;
174 }
175 }
176 after_huge = 0;
177 addr = ALIGN_HUGEPT(vma->vm_end);
178 }
179}
180#endif
181
182/* Do a full search to find an area without any nearby normal pages. */
183static unsigned long
184hugetlb_get_unmapped_area_new_pmd(unsigned long len)
185{
f75c28d8
JH
186 struct vm_unmapped_area_info info;
187
188 info.flags = 0;
189 info.length = len;
190 info.low_limit = TASK_UNMAPPED_BASE;
191 info.high_limit = TASK_SIZE;
192 info.align_mask = PAGE_MASK & HUGEPT_MASK;
193 info.align_offset = 0;
194 return vm_unmapped_area(&info);
e624e95b
JH
195}
196
197unsigned long
198hugetlb_get_unmapped_area(struct file *file, unsigned long addr,
199 unsigned long len, unsigned long pgoff, unsigned long flags)
200{
201 struct hstate *h = hstate_file(file);
202
203 if (len & ~huge_page_mask(h))
204 return -EINVAL;
205 if (len > TASK_SIZE)
206 return -ENOMEM;
207
208 if (flags & MAP_FIXED) {
209 if (prepare_hugepage_range(file, addr, len))
210 return -EINVAL;
211 return addr;
212 }
213
214 if (addr) {
215 addr = ALIGN(addr, huge_page_size(h));
216 if (!prepare_hugepage_range(file, addr, len))
217 return addr;
218 }
219
220 /*
221 * Look for an existing hugetlb vma with space after it (this is to to
222 * minimise fragmentation caused by huge pages.
223 */
224 addr = hugetlb_get_unmapped_area_existing(len);
225 if (addr)
226 return addr;
227
228 /*
229 * Find an unmapped naturally aligned set of 4MB blocks that we can use
230 * for huge pages.
231 */
f75c28d8 232 return hugetlb_get_unmapped_area_new_pmd(len);
e624e95b
JH
233}
234
235#endif /*HAVE_ARCH_HUGETLB_UNMAPPED_AREA*/
236
237/* necessary for boot time 4MB huge page allocation */
238static __init int setup_hugepagesz(char *opt)
239{
240 unsigned long ps = memparse(opt, &opt);
241 if (ps == (1 << HPAGE_SHIFT)) {
242 hugetlb_add_hstate(HPAGE_SHIFT - PAGE_SHIFT);
243 } else {
9cc3387f 244 hugetlb_bad_size();
e624e95b
JH
245 pr_err("hugepagesz: Unsupported page size %lu M\n",
246 ps >> 20);
247 return 0;
248 }
249 return 1;
250}
251__setup("hugepagesz=", setup_hugepagesz);