]>
Commit | Line | Data |
---|---|---|
1d737c8a ZX |
1 | #ifndef __KVM_X86_MMU_H |
2 | #define __KVM_X86_MMU_H | |
3 | ||
edf88417 | 4 | #include <linux/kvm_host.h> |
fc78f519 | 5 | #include "kvm_cache_regs.h" |
1d737c8a | 6 | |
8c6d6adc SY |
7 | #define PT64_PT_BITS 9 |
8 | #define PT64_ENT_PER_PAGE (1 << PT64_PT_BITS) | |
9 | #define PT32_PT_BITS 10 | |
10 | #define PT32_ENT_PER_PAGE (1 << PT32_PT_BITS) | |
11 | ||
12 | #define PT_WRITABLE_SHIFT 1 | |
13 | ||
14 | #define PT_PRESENT_MASK (1ULL << 0) | |
15 | #define PT_WRITABLE_MASK (1ULL << PT_WRITABLE_SHIFT) | |
16 | #define PT_USER_MASK (1ULL << 2) | |
17 | #define PT_PWT_MASK (1ULL << 3) | |
18 | #define PT_PCD_MASK (1ULL << 4) | |
1b7fcd32 AK |
19 | #define PT_ACCESSED_SHIFT 5 |
20 | #define PT_ACCESSED_MASK (1ULL << PT_ACCESSED_SHIFT) | |
8ea667f2 AK |
21 | #define PT_DIRTY_SHIFT 6 |
22 | #define PT_DIRTY_MASK (1ULL << PT_DIRTY_SHIFT) | |
8c6d6adc SY |
23 | #define PT_PAGE_SIZE_MASK (1ULL << 7) |
24 | #define PT_PAT_MASK (1ULL << 7) | |
25 | #define PT_GLOBAL_MASK (1ULL << 8) | |
26 | #define PT64_NX_SHIFT 63 | |
27 | #define PT64_NX_MASK (1ULL << PT64_NX_SHIFT) | |
28 | ||
29 | #define PT_PAT_SHIFT 7 | |
30 | #define PT_DIR_PAT_SHIFT 12 | |
31 | #define PT_DIR_PAT_MASK (1ULL << PT_DIR_PAT_SHIFT) | |
32 | ||
33 | #define PT32_DIR_PSE36_SIZE 4 | |
34 | #define PT32_DIR_PSE36_SHIFT 13 | |
35 | #define PT32_DIR_PSE36_MASK \ | |
36 | (((1ULL << PT32_DIR_PSE36_SIZE) - 1) << PT32_DIR_PSE36_SHIFT) | |
37 | ||
38 | #define PT64_ROOT_LEVEL 4 | |
39 | #define PT32_ROOT_LEVEL 2 | |
40 | #define PT32E_ROOT_LEVEL 3 | |
41 | ||
c9c54174 SY |
42 | #define PT_PDPE_LEVEL 3 |
43 | #define PT_DIRECTORY_LEVEL 2 | |
44 | #define PT_PAGE_TABLE_LEVEL 1 | |
45 | ||
1871c602 GN |
46 | #define PFERR_PRESENT_MASK (1U << 0) |
47 | #define PFERR_WRITE_MASK (1U << 1) | |
48 | #define PFERR_USER_MASK (1U << 2) | |
49 | #define PFERR_RSVD_MASK (1U << 3) | |
50 | #define PFERR_FETCH_MASK (1U << 4) | |
51 | ||
94d8b056 | 52 | int kvm_mmu_get_spte_hierarchy(struct kvm_vcpu *vcpu, u64 addr, u64 sptes[4]); |
ce88decf XG |
53 | void kvm_mmu_set_mmio_spte_mask(u64 mmio_mask); |
54 | int handle_mmio_page_fault_common(struct kvm_vcpu *vcpu, u64 addr, bool direct); | |
52fde8df | 55 | int kvm_init_shadow_mmu(struct kvm_vcpu *vcpu, struct kvm_mmu *context); |
94d8b056 | 56 | |
e0df7b9f DH |
57 | static inline unsigned int kvm_mmu_available_pages(struct kvm *kvm) |
58 | { | |
49d5ca26 DH |
59 | return kvm->arch.n_max_mmu_pages - |
60 | kvm->arch.n_used_mmu_pages; | |
e0df7b9f DH |
61 | } |
62 | ||
1d737c8a ZX |
63 | static inline void kvm_mmu_free_some_pages(struct kvm_vcpu *vcpu) |
64 | { | |
e0df7b9f | 65 | if (unlikely(kvm_mmu_available_pages(vcpu->kvm)< KVM_MIN_FREE_MMU_PAGES)) |
1d737c8a ZX |
66 | __kvm_mmu_free_some_pages(vcpu); |
67 | } | |
68 | ||
69 | static inline int kvm_mmu_reload(struct kvm_vcpu *vcpu) | |
70 | { | |
71 | if (likely(vcpu->arch.mmu.root_hpa != INVALID_PAGE)) | |
72 | return 0; | |
73 | ||
74 | return kvm_mmu_load(vcpu); | |
75 | } | |
76 | ||
43a3795a | 77 | static inline int is_present_gpte(unsigned long pte) |
20c466b5 DE |
78 | { |
79 | return pte & PT_PRESENT_MASK; | |
80 | } | |
81 | ||
bebb106a XG |
82 | static inline int is_writable_pte(unsigned long pte) |
83 | { | |
84 | return pte & PT_WRITABLE_MASK; | |
85 | } | |
86 | ||
87 | static inline bool is_write_protection(struct kvm_vcpu *vcpu) | |
88 | { | |
89 | return kvm_read_cr0_bits(vcpu, X86_CR0_WP); | |
90 | } | |
91 | ||
97d64b78 AK |
92 | /* |
93 | * Will a fault with a given page-fault error code (pfec) cause a permission | |
94 | * fault with the given access (in ACC_* format)? | |
95 | */ | |
96 | static inline bool permission_fault(struct kvm_mmu *mmu, unsigned pte_access, | |
97 | unsigned pfec) | |
bebb106a | 98 | { |
97d64b78 | 99 | return (mmu->permissions[pfec >> 1] >> pte_access) & 1; |
bebb106a | 100 | } |
97d64b78 | 101 | |
1d737c8a | 102 | #endif |