]> git.proxmox.com Git - mirror_ubuntu-eoan-kernel.git/commitdiff
KVM: PPC: Remove mmio_vsx_tx_sx_enabled in KVM MMIO emulation
authorSimon Guo <wei.guo.simon@gmail.com>
Mon, 28 May 2018 01:48:26 +0000 (09:48 +0800)
committerPaul Mackerras <paulus@ozlabs.org>
Wed, 18 Jul 2018 05:14:45 +0000 (15:14 +1000)
Originally PPC KVM MMIO emulation uses only 0~31#(5 bits) for VSR
reg number, and use mmio_vsx_tx_sx_enabled field together for
0~63# VSR regs.

Currently PPC KVM MMIO emulation is reimplemented with analyse_instr()
assistance.  analyse_instr() returns 0~63 for VSR register number, so
it is not necessary to use additional mmio_vsx_tx_sx_enabled field
any more.

This patch extends related reg bits (expand io_gpr to u16 from u8
and use 6 bits for VSR reg#), so that mmio_vsx_tx_sx_enabled can
be removed.

Signed-off-by: Simon Guo <wei.guo.simon@gmail.com>
Signed-off-by: Paul Mackerras <paulus@ozlabs.org>
arch/powerpc/include/asm/kvm_host.h
arch/powerpc/kvm/emulate_loadstore.c
arch/powerpc/kvm/powerpc.c

index fa4efa7e88f703b10b7a35a22947f1d98d2c5880..5b9e6608c5bf1e10668ad2043af10131b2159964 100644 (file)
@@ -672,7 +672,7 @@ struct kvm_vcpu_arch {
        gva_t vaddr_accessed;
        pgd_t *pgdir;
 
-       u8 io_gpr; /* GPR used as IO source/target */
+       u16 io_gpr; /* GPR used as IO source/target */
        u8 mmio_host_swabbed;
        u8 mmio_sign_extend;
        /* conversion between single and double precision */
@@ -688,7 +688,6 @@ struct kvm_vcpu_arch {
         */
        u8 mmio_vsx_copy_nums;
        u8 mmio_vsx_offset;
-       u8 mmio_vsx_tx_sx_enabled;
        u8 mmio_vmx_copy_nums;
        u8 mmio_vmx_offset;
        u8 mmio_copy_type;
@@ -801,14 +800,14 @@ struct kvm_vcpu_arch {
 #define KVMPPC_VCPU_BUSY_IN_HOST       2
 
 /* Values for vcpu->arch.io_gpr */
-#define KVM_MMIO_REG_MASK      0x001f
-#define KVM_MMIO_REG_EXT_MASK  0xffe0
+#define KVM_MMIO_REG_MASK      0x003f
+#define KVM_MMIO_REG_EXT_MASK  0xffc0
 #define KVM_MMIO_REG_GPR       0x0000
-#define KVM_MMIO_REG_FPR       0x0020
-#define KVM_MMIO_REG_QPR       0x0040
-#define KVM_MMIO_REG_FQPR      0x0060
-#define KVM_MMIO_REG_VSX       0x0080
-#define KVM_MMIO_REG_VMX       0x00c0
+#define KVM_MMIO_REG_FPR       0x0040
+#define KVM_MMIO_REG_QPR       0x0080
+#define KVM_MMIO_REG_FQPR      0x00c0
+#define KVM_MMIO_REG_VSX       0x0100
+#define KVM_MMIO_REG_VMX       0x0180
 
 #define __KVM_HAVE_ARCH_WQP
 #define __KVM_HAVE_CREATE_DEVICE
index afde788be14148bf1386b4442e103789dcb181ad..75dce1ef3bc83473e23689ded4c7bfdf72637a7e 100644 (file)
@@ -106,7 +106,6 @@ int kvmppc_emulate_loadstore(struct kvm_vcpu *vcpu)
         * if mmio_vsx_tx_sx_enabled == 1, copy data between
         * VSR[32..63] and memory
         */
-       vcpu->arch.mmio_vsx_tx_sx_enabled = get_tx_or_sx(inst);
        vcpu->arch.mmio_vsx_copy_nums = 0;
        vcpu->arch.mmio_vsx_offset = 0;
        vcpu->arch.mmio_copy_type = KVMPPC_VSX_COPY_NONE;
@@ -242,8 +241,8 @@ int kvmppc_emulate_loadstore(struct kvm_vcpu *vcpu)
                        }
 
                        emulated = kvmppc_handle_vsx_load(run, vcpu,
-                                       KVM_MMIO_REG_VSX | (op.reg & 0x1f),
-                                       io_size_each, 1, op.type & SIGNEXT);
+                                       KVM_MMIO_REG_VSX|op.reg, io_size_each,
+                                       1, op.type & SIGNEXT);
                        break;
                }
 #endif
@@ -363,7 +362,7 @@ int kvmppc_emulate_loadstore(struct kvm_vcpu *vcpu)
                        }
 
                        emulated = kvmppc_handle_vsx_store(run, vcpu,
-                                       op.reg & 0x1f, io_size_each, 1);
+                                       op.reg, io_size_each, 1);
                        break;
                }
 #endif
index 0e8c20c5eaace243850e255acf5be966c85f0a5c..b6fd7c5b8d1ab90fd63229671d6c7cc710628b1e 100644 (file)
@@ -880,10 +880,10 @@ static inline void kvmppc_set_vsr_dword(struct kvm_vcpu *vcpu,
        if (offset == -1)
                return;
 
-       if (vcpu->arch.mmio_vsx_tx_sx_enabled) {
-               val.vval = VCPU_VSX_VR(vcpu, index);
+       if (index >= 32) {
+               val.vval = VCPU_VSX_VR(vcpu, index - 32);
                val.vsxval[offset] = gpr;
-               VCPU_VSX_VR(vcpu, index) = val.vval;
+               VCPU_VSX_VR(vcpu, index - 32) = val.vval;
        } else {
                VCPU_VSX_FPR(vcpu, index, offset) = gpr;
        }
@@ -895,11 +895,11 @@ static inline void kvmppc_set_vsr_dword_dump(struct kvm_vcpu *vcpu,
        union kvmppc_one_reg val;
        int index = vcpu->arch.io_gpr & KVM_MMIO_REG_MASK;
 
-       if (vcpu->arch.mmio_vsx_tx_sx_enabled) {
-               val.vval = VCPU_VSX_VR(vcpu, index);
+       if (index >= 32) {
+               val.vval = VCPU_VSX_VR(vcpu, index - 32);
                val.vsxval[0] = gpr;
                val.vsxval[1] = gpr;
-               VCPU_VSX_VR(vcpu, index) = val.vval;
+               VCPU_VSX_VR(vcpu, index - 32) = val.vval;
        } else {
                VCPU_VSX_FPR(vcpu, index, 0) = gpr;
                VCPU_VSX_FPR(vcpu, index, 1) = gpr;
@@ -912,12 +912,12 @@ static inline void kvmppc_set_vsr_word_dump(struct kvm_vcpu *vcpu,
        union kvmppc_one_reg val;
        int index = vcpu->arch.io_gpr & KVM_MMIO_REG_MASK;
 
-       if (vcpu->arch.mmio_vsx_tx_sx_enabled) {
+       if (index >= 32) {
                val.vsx32val[0] = gpr;
                val.vsx32val[1] = gpr;
                val.vsx32val[2] = gpr;
                val.vsx32val[3] = gpr;
-               VCPU_VSX_VR(vcpu, index) = val.vval;
+               VCPU_VSX_VR(vcpu, index - 32) = val.vval;
        } else {
                val.vsx32val[0] = gpr;
                val.vsx32val[1] = gpr;
@@ -937,10 +937,10 @@ static inline void kvmppc_set_vsr_word(struct kvm_vcpu *vcpu,
        if (offset == -1)
                return;
 
-       if (vcpu->arch.mmio_vsx_tx_sx_enabled) {
-               val.vval = VCPU_VSX_VR(vcpu, index);
+       if (index >= 32) {
+               val.vval = VCPU_VSX_VR(vcpu, index - 32);
                val.vsx32val[offset] = gpr32;
-               VCPU_VSX_VR(vcpu, index) = val.vval;
+               VCPU_VSX_VR(vcpu, index - 32) = val.vval;
        } else {
                dword_offset = offset / 2;
                word_offset = offset % 2;
@@ -1361,10 +1361,10 @@ static inline int kvmppc_get_vsr_data(struct kvm_vcpu *vcpu, int rs, u64 *val)
                        break;
                }
 
-               if (!vcpu->arch.mmio_vsx_tx_sx_enabled) {
+               if (rs < 32) {
                        *val = VCPU_VSX_FPR(vcpu, rs, vsx_offset);
                } else {
-                       reg.vval = VCPU_VSX_VR(vcpu, rs);
+                       reg.vval = VCPU_VSX_VR(vcpu, rs - 32);
                        *val = reg.vsxval[vsx_offset];
                }
                break;
@@ -1378,13 +1378,13 @@ static inline int kvmppc_get_vsr_data(struct kvm_vcpu *vcpu, int rs, u64 *val)
                        break;
                }
 
-               if (!vcpu->arch.mmio_vsx_tx_sx_enabled) {
+               if (rs < 32) {
                        dword_offset = vsx_offset / 2;
                        word_offset = vsx_offset % 2;
                        reg.vsxval[0] = VCPU_VSX_FPR(vcpu, rs, dword_offset);
                        *val = reg.vsx32val[word_offset];
                } else {
-                       reg.vval = VCPU_VSX_VR(vcpu, rs);
+                       reg.vval = VCPU_VSX_VR(vcpu, rs - 32);
                        *val = reg.vsx32val[vsx_offset];
                }
                break;