;------------------------------------------------------------------------------ ; ; Copyright (c) 2016 - 2022, Intel Corporation. All rights reserved.
; Copyright (c) 2020, AMD Incorporated. All rights reserved.
; SPDX-License-Identifier: BSD-2-Clause-Patent ; ; Module Name: ; ; SmiEntry.nasm ; ; Abstract: ; ; Code template of the SMI handler for a particular processor ; ;------------------------------------------------------------------------------- %include "StuffRsbNasm.inc" %include "Nasm.inc" %define MSR_IA32_S_CET 0x6A2 %define MSR_IA32_CET_SH_STK_EN 0x1 %define MSR_IA32_CET_WR_SHSTK_EN 0x2 %define MSR_IA32_CET_ENDBR_EN 0x4 %define MSR_IA32_CET_LEG_IW_EN 0x8 %define MSR_IA32_CET_NO_TRACK_EN 0x10 %define MSR_IA32_CET_SUPPRESS_DIS 0x20 %define MSR_IA32_CET_SUPPRESS 0x400 %define MSR_IA32_CET_TRACKER 0x800 %define MSR_IA32_PL0_SSP 0x6A4 %define CR4_CET 0x800000 %define MSR_IA32_MISC_ENABLE 0x1A0 %define MSR_EFER 0xc0000080 %define MSR_EFER_XD 0x800 ; ; Constants relating to PROCESSOR_SMM_DESCRIPTOR ; %define DSC_OFFSET 0xfb00 %define DSC_GDTPTR 0x30 %define DSC_GDTSIZ 0x38 %define DSC_CS 14 %define DSC_DS 16 %define DSC_SS 18 %define DSC_OTHERSEG 20 %define PROTECT_MODE_CS 0x8 %define PROTECT_MODE_DS 0x20 %define TSS_SEGMENT 0x40 extern ASM_PFX(SmiRendezvous) extern ASM_PFX(FeaturePcdGet (PcdCpuSmmStackGuard)) extern ASM_PFX(CpuSmmDebugEntry) extern ASM_PFX(CpuSmmDebugExit) global ASM_PFX(gcSmiHandlerTemplate) global ASM_PFX(gcSmiHandlerSize) global ASM_PFX(gPatchSmiCr3) global ASM_PFX(gPatchSmiStack) global ASM_PFX(gPatchSmbase) extern ASM_PFX(mXdSupported) global ASM_PFX(gPatchXdSupported) global ASM_PFX(gPatchMsrIa32MiscEnableSupported) extern ASM_PFX(gSmiHandlerIdtr) extern ASM_PFX(mCetSupported) global ASM_PFX(mPatchCetSupported) global ASM_PFX(mPatchCetPl0Ssp) global ASM_PFX(mPatchCetInterruptSsp) SECTION .text BITS 16 ASM_PFX(gcSmiHandlerTemplate): _SmiEntryPoint: mov bx, _GdtDesc - _SmiEntryPoint + 0x8000 mov ax,[cs:DSC_OFFSET + DSC_GDTSIZ] dec ax mov [cs:bx], ax mov eax, [cs:DSC_OFFSET + DSC_GDTPTR] mov [cs:bx + 2], eax mov ebp, eax ; ebp = GDT base o32 lgdt [cs:bx] ; lgdt fword ptr cs:[bx] mov ax, PROTECT_MODE_CS mov [cs:bx-0x2],ax mov edi, strict dword 0 ; source operand will be patched ASM_PFX(gPatchSmbase): lea eax, [edi + (@32bit - _SmiEntryPoint) + 0x8000] mov [cs:bx-0x6],eax mov ebx, cr0 and ebx, 0x9ffafff3 or ebx, 0x23 mov cr0, ebx jmp dword 0x0:0x0 _GdtDesc: DW 0 DD 0 BITS 32 @32bit: mov ax, PROTECT_MODE_DS o16 mov ds, ax o16 mov es, ax o16 mov fs, ax o16 mov gs, ax o16 mov ss, ax mov esp, strict dword 0 ; source operand will be patched ASM_PFX(gPatchSmiStack): mov eax, ASM_PFX(gSmiHandlerIdtr) lidt [eax] jmp ProtFlatMode ProtFlatMode: mov eax, strict dword 0 ; source operand will be patched ASM_PFX(gPatchSmiCr3): mov cr3, eax ; ; Need to test for CR4 specific bit support ; mov eax, 1 cpuid ; use CPUID to determine if specific CR4 bits are supported xor eax, eax ; Clear EAX test edx, BIT2 ; Check for DE capabilities jz .0 or eax, BIT3 .0: test edx, BIT6 ; Check for PAE capabilities jz .1 or eax, BIT5 .1: test edx, BIT7 ; Check for MCE capabilities jz .2 or eax, BIT6 .2: test edx, BIT24 ; Check for FXSR capabilities jz .3 or eax, BIT9 .3: test edx, BIT25 ; Check for SSE capabilities jz .4 or eax, BIT10 .4: ; as cr4.PGE is not set here, refresh cr3 mov cr4, eax ; in PreModifyMtrrs() to flush TLB. cmp byte [dword ASM_PFX(FeaturePcdGet (PcdCpuSmmStackGuard))], 0 jz .6 ; Load TSS mov byte [ebp + TSS_SEGMENT + 5], 0x89 ; clear busy flag mov eax, TSS_SEGMENT ltr ax .6: ; enable NXE if supported mov al, strict byte 1 ; source operand may be patched ASM_PFX(gPatchXdSupported): cmp al, 0 jz @SkipXd ; If MSR_IA32_MISC_ENABLE is supported, clear XD Disable bit mov al, strict byte 1 ; source operand may be patched ASM_PFX(gPatchMsrIa32MiscEnableSupported): cmp al, 1 jz MsrIa32MiscEnableSupported ; MSR_IA32_MISC_ENABLE not supported xor edx, edx push edx ; don't try to restore the XD Disable bit just before RSM jmp EnableNxe ; ; Check XD disable bit ; MsrIa32MiscEnableSupported: mov ecx, MSR_IA32_MISC_ENABLE rdmsr push edx ; save MSR_IA32_MISC_ENABLE[63-32] test edx, BIT2 ; MSR_IA32_MISC_ENABLE[34] jz EnableNxe and dx, 0xFFFB ; clear XD Disable bit if it is set wrmsr EnableNxe: mov ecx, MSR_EFER rdmsr or ax, MSR_EFER_XD ; enable NXE wrmsr jmp @XdDone @SkipXd: sub esp, 4 @XdDone: mov ebx, cr0 or ebx, 0x80010023 ; enable paging + WP + NE + MP + PE mov cr0, ebx lea ebx, [edi + DSC_OFFSET] mov ax, [ebx + DSC_DS] mov ds, eax mov ax, [ebx + DSC_OTHERSEG] mov es, eax mov fs, eax mov gs, eax mov ax, [ebx + DSC_SS] mov ss, eax mov ebx, [esp + 4] ; ebx <- CpuIndex ; enable CET if supported mov al, strict byte 1 ; source operand may be patched ASM_PFX(mPatchCetSupported): cmp al, 0 jz CetDone mov ecx, MSR_IA32_S_CET rdmsr push edx push eax mov ecx, MSR_IA32_PL0_SSP rdmsr push edx push eax mov ecx, MSR_IA32_S_CET mov eax, MSR_IA32_CET_SH_STK_EN xor edx, edx wrmsr mov ecx, MSR_IA32_PL0_SSP mov eax, strict dword 0 ; source operand will be patched ASM_PFX(mPatchCetPl0Ssp): xor edx, edx wrmsr mov ecx, cr0 btr ecx, 16 ; clear WP mov cr0, ecx mov [eax], eax ; reload SSP, and clear busyflag. xor ecx, ecx mov [eax + 4], ecx mov eax, strict dword 0 ; source operand will be patched ASM_PFX(mPatchCetInterruptSsp): cmp eax, 0 jz CetInterruptDone mov [eax], eax ; reload SSP, and clear busyflag. xor ecx, ecx mov [eax + 4], ecx CetInterruptDone: mov ecx, cr0 bts ecx, 16 ; set WP mov cr0, ecx mov eax, 0x668 | CR4_CET mov cr4, eax setssbsy CetDone: push ebx mov eax, ASM_PFX(CpuSmmDebugEntry) call eax add esp, 4 push ebx mov eax, ASM_PFX(SmiRendezvous) call eax add esp, 4 push ebx mov eax, ASM_PFX(CpuSmmDebugExit) call eax add esp, 4 mov eax, ASM_PFX(mCetSupported) mov al, [eax] cmp al, 0 jz CetDone2 mov eax, 0x668 mov cr4, eax ; disable CET mov ecx, MSR_IA32_PL0_SSP pop eax pop edx wrmsr mov ecx, MSR_IA32_S_CET pop eax pop edx wrmsr CetDone2: mov eax, ASM_PFX(mXdSupported) mov al, [eax] cmp al, 0 jz .7 pop edx ; get saved MSR_IA32_MISC_ENABLE[63-32] test edx, BIT2 jz .7 mov ecx, MSR_IA32_MISC_ENABLE rdmsr or dx, BIT2 ; set XD Disable bit if it was set before entering into SMM wrmsr .7: StuffRsb32 rsm ASM_PFX(gcSmiHandlerSize): DW $ - _SmiEntryPoint global ASM_PFX(PiSmmCpuSmiEntryFixupAddress) ASM_PFX(PiSmmCpuSmiEntryFixupAddress): ret