From 53c76a6478d106c89c4ba48599a8f19b74480d13 Mon Sep 17 00:00:00 2001 From: vanjeff Date: Wed, 28 Nov 2012 04:49:48 +0000 Subject: [PATCH] Save/Restore missing volatile registers (XMM0-5) save/restore in Page Fault handler. Signed-off-by: Jeff Fan Reviewed-by: Jiewen Yao git-svn-id: https://edk2.svn.sourceforge.net/svnroot/edk2/trunk/edk2@13974 6f19259b-4bc3-4df7-8a09-765794883524 --- .../SmmBaseHelper/X64/PageFaultHandler.S | 24 ++++++++++++++++--- .../SmmBaseHelper/X64/PageFaultHandler.asm | 24 ++++++++++++++++--- 2 files changed, 42 insertions(+), 6 deletions(-) diff --git a/EdkCompatibilityPkg/Compatibility/SmmBaseHelper/X64/PageFaultHandler.S b/EdkCompatibilityPkg/Compatibility/SmmBaseHelper/X64/PageFaultHandler.S index d14b9114d8..f1b5ad7adb 100644 --- a/EdkCompatibilityPkg/Compatibility/SmmBaseHelper/X64/PageFaultHandler.S +++ b/EdkCompatibilityPkg/Compatibility/SmmBaseHelper/X64/PageFaultHandler.S @@ -1,6 +1,6 @@ #------------------------------------------------------------------------------ # -# Copyright (c) 2006 - 2011, Intel Corporation. All rights reserved.
+# Copyright (c) 2006 - 2012, Intel Corporation. All rights reserved.
# This program and the accompanying materials # are licensed and made available under the terms and conditions of the BSD License # which accompanies this distribution. The full text of the license may be found at @@ -28,10 +28,28 @@ ASM_PFX(PageFaultHandlerHook): pushq %r9 pushq %r10 pushq %r11 + + addq $-0x68, %rsp # reserve memory to store XMM registers and make address 16-byte alignment + movdqa %xmm0, 0(%rsp) + movdqa %xmm1, 0x10(%rsp) + movdqa %xmm2, 0x20(%rsp) + movdqa %xmm3, 0x30(%rsp) + movdqa %xmm4, 0x40(%rsp) + movdqa %xmm5, 0x50(%rsp) + addq $-0x20, %rsp call ASM_PFX(PageFaultHandler) addq $0x20, %rsp - testb %al, %al + + movdqa 0(%rsp), %xmm0 + movdqa 0x10(%rsp), %xmm1 + movdqa 0x20(%rsp), %xmm2 + movdqa 0x30(%rsp), %xmm3 + movdqa 0x40(%rsp), %xmm4 + movdqa 0x50(%rsp), %xmm5 + addq $0x68, %rsp + + testb %al, %al # set ZF flag popq %r11 popq %r10 popq %r9 @@ -39,7 +57,7 @@ ASM_PFX(PageFaultHandlerHook): popq %rdx popq %rcx popq %rax # restore all volatile registers - jnz L1 + jnz L1 # check ZF flag #ifdef __APPLE__ int $3 #else diff --git a/EdkCompatibilityPkg/Compatibility/SmmBaseHelper/X64/PageFaultHandler.asm b/EdkCompatibilityPkg/Compatibility/SmmBaseHelper/X64/PageFaultHandler.asm index d1c7bf972d..d58a04bcfb 100644 --- a/EdkCompatibilityPkg/Compatibility/SmmBaseHelper/X64/PageFaultHandler.asm +++ b/EdkCompatibilityPkg/Compatibility/SmmBaseHelper/X64/PageFaultHandler.asm @@ -1,6 +1,6 @@ ;------------------------------------------------------------------------------ ; -; Copyright (c) 2010 - 2011, Intel Corporation. All rights reserved.
+; Copyright (c) 2010 - 2012, Intel Corporation. All rights reserved.
; This program and the accompanying materials ; are licensed and made available under the terms and conditions of the BSD License ; which accompanies this distribution. The full text of the license may be found at @@ -32,10 +32,28 @@ PageFaultHandlerHook PROC push r9 push r10 push r11 + + add rsp, -10h * 6 - 8 ; reserve memory to store XMM registers and make address 16-byte alignment + movdqa [rsp], xmm0 + movdqa [rsp + 10h], xmm1 + movdqa [rsp + 20h], xmm2 + movdqa [rsp + 30h], xmm3 + movdqa [rsp + 40h], xmm4 + movdqa [rsp + 50h], xmm5 + add rsp, -20h call PageFaultHandler add rsp, 20h - test al, al + + movdqa xmm0, [rsp] + movdqa xmm1, [rsp + 10h] + movdqa xmm2, [rsp + 20h] + movdqa xmm3, [rsp + 30h] + movdqa xmm4, [rsp + 40h] + movdqa xmm5, [rsp + 50h] + add rsp, 10h * 6 + 8 + + test al, al ; set ZF flag pop r11 pop r10 pop r9 @@ -43,7 +61,7 @@ PageFaultHandlerHook PROC pop rdx pop rcx pop rax ; restore all volatile registers - jnz @F + jnz @F ; check ZF flag jmp mOriginalHandler @@: add rsp, 08h ; skip error code for PF -- 2.39.2