]> git.proxmox.com Git - mirror_ubuntu-artful-kernel.git/commitdiff
arm64: mm: Invalidate both kernel and user ASIDs when performing TLBI
authorWill Deacon <will.deacon@arm.com>
Thu, 10 Aug 2017 13:13:33 +0000 (14:13 +0100)
committerKhalid Elmously <khalid.elmously@canonical.com>
Tue, 27 Feb 2018 16:32:40 +0000 (11:32 -0500)
Commit 9b0de864b5bc upstream.

Since an mm has both a kernel and a user ASID, we need to ensure that
broadcast TLB maintenance targets both address spaces so that things
like CoW continue to work with the uaccess primitives in the kernel.

Reviewed-by: Mark Rutland <mark.rutland@arm.com>
Tested-by: Laura Abbott <labbott@redhat.com>
Tested-by: Shanker Donthineni <shankerd@codeaurora.org>
Signed-off-by: Will Deacon <will.deacon@arm.com>
Signed-off-by: Ard Biesheuvel <ard.biesheuvel@linaro.org>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
(cherry picked from commit d6c414013bb4c41c93fcef1ffa90df4f56c4cc53)

CVE-2017-5753
CVE-2017-5715
CVE-2017-5754

Signed-off-by: Paolo Pisati <paolo.pisati@canonical.com>
Acked-by: Brad Figg <brad.figg@canonical.com>
Acked-by: Kleber Sacilotto de Souza <kleber.souza@canonical.com>
Signed-off-by: Khalid Elmously <khalid.elmously@canonical.com>
arch/arm64/include/asm/tlbflush.h

index af1c76981911dbcb0d399f3552a8888ac2e0bc3d..9e82dd79c7db59d1da730af1b6ed5354599ad2f4 100644 (file)
@@ -23,6 +23,7 @@
 
 #include <linux/sched.h>
 #include <asm/cputype.h>
+#include <asm/mmu.h>
 
 /*
  * Raw TLBI operations.
 
 #define __tlbi(op, ...)                __TLBI_N(op, ##__VA_ARGS__, 1, 0)
 
+#define __tlbi_user(op, arg) do {                                              \
+       if (arm64_kernel_unmapped_at_el0())                                     \
+               __tlbi(op, (arg) | USER_ASID_FLAG);                             \
+} while (0)
+
 /*
  *     TLB Management
  *     ==============
@@ -115,6 +121,7 @@ static inline void flush_tlb_mm(struct mm_struct *mm)
 
        dsb(ishst);
        __tlbi(aside1is, asid);
+       __tlbi_user(aside1is, asid);
        dsb(ish);
 }
 
@@ -125,6 +132,7 @@ static inline void flush_tlb_page(struct vm_area_struct *vma,
 
        dsb(ishst);
        __tlbi(vale1is, addr);
+       __tlbi_user(vale1is, addr);
        dsb(ish);
 }
 
@@ -151,10 +159,13 @@ static inline void __flush_tlb_range(struct vm_area_struct *vma,
 
        dsb(ishst);
        for (addr = start; addr < end; addr += 1 << (PAGE_SHIFT - 12)) {
-               if (last_level)
+               if (last_level) {
                        __tlbi(vale1is, addr);
-               else
+                       __tlbi_user(vale1is, addr);
+               } else {
                        __tlbi(vae1is, addr);
+                       __tlbi_user(vae1is, addr);
+               }
        }
        dsb(ish);
 }
@@ -194,6 +205,7 @@ static inline void __flush_tlb_pgtable(struct mm_struct *mm,
        unsigned long addr = uaddr >> 12 | (ASID(mm) << 48);
 
        __tlbi(vae1is, addr);
+       __tlbi_user(vae1is, addr);
        dsb(ish);
 }