root/arch/x86/mm/mem_encrypt_boot.S

/* [<][>][^][v][top][bottom][index][help] */
   1 /* SPDX-License-Identifier: GPL-2.0-only */
   2 /*
   3  * AMD Memory Encryption Support
   4  *
   5  * Copyright (C) 2016 Advanced Micro Devices, Inc.
   6  *
   7  * Author: Tom Lendacky <thomas.lendacky@amd.com>
   8  */
   9 
  10 #include <linux/linkage.h>
  11 #include <asm/pgtable.h>
  12 #include <asm/page.h>
  13 #include <asm/processor-flags.h>
  14 #include <asm/msr-index.h>
  15 #include <asm/nospec-branch.h>
  16 
  17         .text
  18         .code64
  19 ENTRY(sme_encrypt_execute)
  20 
  21         /*
  22          * Entry parameters:
  23          *   RDI - virtual address for the encrypted mapping
  24          *   RSI - virtual address for the decrypted mapping
  25          *   RDX - length to encrypt
  26          *   RCX - virtual address of the encryption workarea, including:
  27          *     - stack page (PAGE_SIZE)
  28          *     - encryption routine page (PAGE_SIZE)
  29          *     - intermediate copy buffer (PMD_PAGE_SIZE)
  30          *    R8 - physcial address of the pagetables to use for encryption
  31          */
  32 
  33         push    %rbp
  34         movq    %rsp, %rbp              /* RBP now has original stack pointer */
  35 
  36         /* Set up a one page stack in the non-encrypted memory area */
  37         movq    %rcx, %rax              /* Workarea stack page */
  38         leaq    PAGE_SIZE(%rax), %rsp   /* Set new stack pointer */
  39         addq    $PAGE_SIZE, %rax        /* Workarea encryption routine */
  40 
  41         push    %r12
  42         movq    %rdi, %r10              /* Encrypted area */
  43         movq    %rsi, %r11              /* Decrypted area */
  44         movq    %rdx, %r12              /* Area length */
  45 
  46         /* Copy encryption routine into the workarea */
  47         movq    %rax, %rdi                              /* Workarea encryption routine */
  48         leaq    __enc_copy(%rip), %rsi                  /* Encryption routine */
  49         movq    $(.L__enc_copy_end - __enc_copy), %rcx  /* Encryption routine length */
  50         rep     movsb
  51 
  52         /* Setup registers for call */
  53         movq    %r10, %rdi              /* Encrypted area */
  54         movq    %r11, %rsi              /* Decrypted area */
  55         movq    %r8, %rdx               /* Pagetables used for encryption */
  56         movq    %r12, %rcx              /* Area length */
  57         movq    %rax, %r8               /* Workarea encryption routine */
  58         addq    $PAGE_SIZE, %r8         /* Workarea intermediate copy buffer */
  59 
  60         ANNOTATE_RETPOLINE_SAFE
  61         call    *%rax                   /* Call the encryption routine */
  62 
  63         pop     %r12
  64 
  65         movq    %rbp, %rsp              /* Restore original stack pointer */
  66         pop     %rbp
  67 
  68         ret
  69 ENDPROC(sme_encrypt_execute)
  70 
  71 ENTRY(__enc_copy)
  72 /*
  73  * Routine used to encrypt memory in place.
  74  *   This routine must be run outside of the kernel proper since
  75  *   the kernel will be encrypted during the process. So this
  76  *   routine is defined here and then copied to an area outside
  77  *   of the kernel where it will remain and run decrypted
  78  *   during execution.
  79  *
  80  *   On entry the registers must be:
  81  *     RDI - virtual address for the encrypted mapping
  82  *     RSI - virtual address for the decrypted mapping
  83  *     RDX - address of the pagetables to use for encryption
  84  *     RCX - length of area
  85  *      R8 - intermediate copy buffer
  86  *
  87  *     RAX - points to this routine
  88  *
  89  * The area will be encrypted by copying from the non-encrypted
  90  * memory space to an intermediate buffer and then copying from the
  91  * intermediate buffer back to the encrypted memory space. The physical
  92  * addresses of the two mappings are the same which results in the area
  93  * being encrypted "in place".
  94  */
  95         /* Enable the new page tables */
  96         mov     %rdx, %cr3
  97 
  98         /* Flush any global TLBs */
  99         mov     %cr4, %rdx
 100         andq    $~X86_CR4_PGE, %rdx
 101         mov     %rdx, %cr4
 102         orq     $X86_CR4_PGE, %rdx
 103         mov     %rdx, %cr4
 104 
 105         push    %r15
 106         push    %r12
 107 
 108         movq    %rcx, %r9               /* Save area length */
 109         movq    %rdi, %r10              /* Save encrypted area address */
 110         movq    %rsi, %r11              /* Save decrypted area address */
 111 
 112         /* Set the PAT register PA5 entry to write-protect */
 113         movl    $MSR_IA32_CR_PAT, %ecx
 114         rdmsr
 115         mov     %rdx, %r15              /* Save original PAT value */
 116         andl    $0xffff00ff, %edx       /* Clear PA5 */
 117         orl     $0x00000500, %edx       /* Set PA5 to WP */
 118         wrmsr
 119 
 120         wbinvd                          /* Invalidate any cache entries */
 121 
 122         /* Copy/encrypt up to 2MB at a time */
 123         movq    $PMD_PAGE_SIZE, %r12
 124 1:
 125         cmpq    %r12, %r9
 126         jnb     2f
 127         movq    %r9, %r12
 128 
 129 2:
 130         movq    %r11, %rsi              /* Source - decrypted area */
 131         movq    %r8, %rdi               /* Dest   - intermediate copy buffer */
 132         movq    %r12, %rcx
 133         rep     movsb
 134 
 135         movq    %r8, %rsi               /* Source - intermediate copy buffer */
 136         movq    %r10, %rdi              /* Dest   - encrypted area */
 137         movq    %r12, %rcx
 138         rep     movsb
 139 
 140         addq    %r12, %r11
 141         addq    %r12, %r10
 142         subq    %r12, %r9               /* Kernel length decrement */
 143         jnz     1b                      /* Kernel length not zero? */
 144 
 145         /* Restore PAT register */
 146         movl    $MSR_IA32_CR_PAT, %ecx
 147         rdmsr
 148         mov     %r15, %rdx              /* Restore original PAT value */
 149         wrmsr
 150 
 151         pop     %r12
 152         pop     %r15
 153 
 154         ret
 155 .L__enc_copy_end:
 156 ENDPROC(__enc_copy)

/* [<][>][^][v][top][bottom][index][help] */