pti.txt 8.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186
  1. Overview
  2. ========
  3. Page Table Isolation (pti, previously known as KAISER[1]) is a
  4. countermeasure against attacks on the shared user/kernel address
  5. space such as the "Meltdown" approach[2].
  6. To mitigate this class of attacks, we create an independent set of
  7. page tables for use only when running userspace applications. When
  8. the kernel is entered via syscalls, interrupts or exceptions, the
  9. page tables are switched to the full "kernel" copy. When the system
  10. switches back to user mode, the user copy is used again.
  11. The userspace page tables contain only a minimal amount of kernel
  12. data: only what is needed to enter/exit the kernel such as the
  13. entry/exit functions themselves and the interrupt descriptor table
  14. (IDT). There are a few strictly unnecessary things that get mapped
  15. such as the first C function when entering an interrupt (see
  16. comments in pti.c).
  17. This approach helps to ensure that side-channel attacks leveraging
  18. the paging structures do not function when PTI is enabled. It can be
  19. enabled by setting CONFIG_PAGE_TABLE_ISOLATION=y at compile time.
  20. Once enabled at compile-time, it can be disabled at boot with the
  21. 'nopti' or 'pti=' kernel parameters (see kernel-parameters.txt).
  22. Page Table Management
  23. =====================
  24. When PTI is enabled, the kernel manages two sets of page tables.
  25. The first set is very similar to the single set which is present in
  26. kernels without PTI. This includes a complete mapping of userspace
  27. that the kernel can use for things like copy_to_user().
  28. Although _complete_, the user portion of the kernel page tables is
  29. crippled by setting the NX bit in the top level. This ensures
  30. that any missed kernel->user CR3 switch will immediately crash
  31. userspace upon executing its first instruction.
  32. The userspace page tables map only the kernel data needed to enter
  33. and exit the kernel. This data is entirely contained in the 'struct
  34. cpu_entry_area' structure which is placed in the fixmap which gives
  35. each CPU's copy of the area a compile-time-fixed virtual address.
  36. For new userspace mappings, the kernel makes the entries in its
  37. page tables like normal. The only difference is when the kernel
  38. makes entries in the top (PGD) level. In addition to setting the
  39. entry in the main kernel PGD, a copy of the entry is made in the
  40. userspace page tables' PGD.
  41. This sharing at the PGD level also inherently shares all the lower
  42. layers of the page tables. This leaves a single, shared set of
  43. userspace page tables to manage. One PTE to lock, one set of
  44. accessed bits, dirty bits, etc...
  45. Overhead
  46. ========
  47. Protection against side-channel attacks is important. But,
  48. this protection comes at a cost:
  49. 1. Increased Memory Use
  50. a. Each process now needs an order-1 PGD instead of order-0.
  51. (Consumes an additional 4k per process).
  52. b. The 'cpu_entry_area' structure must be 2MB in size and 2MB
  53. aligned so that it can be mapped by setting a single PMD
  54. entry. This consumes nearly 2MB of RAM once the kernel
  55. is decompressed, but no space in the kernel image itself.
  56. 2. Runtime Cost
  57. a. CR3 manipulation to switch between the page table copies
  58. must be done at interrupt, syscall, and exception entry
  59. and exit (it can be skipped when the kernel is interrupted,
  60. though.) Moves to CR3 are on the order of a hundred
  61. cycles, and are required at every entry and exit.
  62. b. A "trampoline" must be used for SYSCALL entry. This
  63. trampoline depends on a smaller set of resources than the
  64. non-PTI SYSCALL entry code, so requires mapping fewer
  65. things into the userspace page tables. The downside is
  66. that stacks must be switched at entry time.
  67. c. Global pages are disabled for all kernel structures not
  68. mapped into both kernel and userspace page tables. This
  69. feature of the MMU allows different processes to share TLB
  70. entries mapping the kernel. Losing the feature means more
  71. TLB misses after a context switch. The actual loss of
  72. performance is very small, however, never exceeding 1%.
  73. d. Process Context IDentifiers (PCID) is a CPU feature that
  74. allows us to skip flushing the entire TLB when switching page
  75. tables by setting a special bit in CR3 when the page tables
  76. are changed. This makes switching the page tables (at context
  77. switch, or kernel entry/exit) cheaper. But, on systems with
  78. PCID support, the context switch code must flush both the user
  79. and kernel entries out of the TLB. The user PCID TLB flush is
  80. deferred until the exit to userspace, minimizing the cost.
  81. See intel.com/sdm for the gory PCID/INVPCID details.
  82. e. The userspace page tables must be populated for each new
  83. process. Even without PTI, the shared kernel mappings
  84. are created by copying top-level (PGD) entries into each
  85. new process. But, with PTI, there are now *two* kernel
  86. mappings: one in the kernel page tables that maps everything
  87. and one for the entry/exit structures. At fork(), we need to
  88. copy both.
  89. f. In addition to the fork()-time copying, there must also
  90. be an update to the userspace PGD any time a set_pgd() is done
  91. on a PGD used to map userspace. This ensures that the kernel
  92. and userspace copies always map the same userspace
  93. memory.
  94. g. On systems without PCID support, each CR3 write flushes
  95. the entire TLB. That means that each syscall, interrupt
  96. or exception flushes the TLB.
  97. h. INVPCID is a TLB-flushing instruction which allows flushing
  98. of TLB entries for non-current PCIDs. Some systems support
  99. PCIDs, but do not support INVPCID. On these systems, addresses
  100. can only be flushed from the TLB for the current PCID. When
  101. flushing a kernel address, we need to flush all PCIDs, so a
  102. single kernel address flush will require a TLB-flushing CR3
  103. write upon the next use of every PCID.
  104. Possible Future Work
  105. ====================
  106. 1. We can be more careful about not actually writing to CR3
  107. unless its value is actually changed.
  108. 2. Allow PTI to be enabled/disabled at runtime in addition to the
  109. boot-time switching.
  110. Testing
  111. ========
  112. To test stability of PTI, the following test procedure is recommended,
  113. ideally doing all of these in parallel:
  114. 1. Set CONFIG_DEBUG_ENTRY=y
  115. 2. Run several copies of all of the tools/testing/selftests/x86/ tests
  116. (excluding MPX and protection_keys) in a loop on multiple CPUs for
  117. several minutes. These tests frequently uncover corner cases in the
  118. kernel entry code. In general, old kernels might cause these tests
  119. themselves to crash, but they should never crash the kernel.
  120. 3. Run the 'perf' tool in a mode (top or record) that generates many
  121. frequent performance monitoring non-maskable interrupts (see "NMI"
  122. in /proc/interrupts). This exercises the NMI entry/exit code which
  123. is known to trigger bugs in code paths that did not expect to be
  124. interrupted, including nested NMIs. Using "-c" boosts the rate of
  125. NMIs, and using two -c with separate counters encourages nested NMIs
  126. and less deterministic behavior.
  127. while true; do perf record -c 10000 -e instructions,cycles -a sleep 10; done
  128. 4. Launch a KVM virtual machine.
  129. 5. Run 32-bit binaries on systems supporting the SYSCALL instruction.
  130. This has been a lightly-tested code path and needs extra scrutiny.
  131. Debugging
  132. =========
  133. Bugs in PTI cause a few different signatures of crashes
  134. that are worth noting here.
  135. * Failures of the selftests/x86 code. Usually a bug in one of the
  136. more obscure corners of entry_64.S
  137. * Crashes in early boot, especially around CPU bringup. Bugs
  138. in the trampoline code or mappings cause these.
  139. * Crashes at the first interrupt. Caused by bugs in entry_64.S,
  140. like screwing up a page table switch. Also caused by
  141. incorrectly mapping the IRQ handler entry code.
  142. * Crashes at the first NMI. The NMI code is separate from main
  143. interrupt handlers and can have bugs that do not affect
  144. normal interrupts. Also caused by incorrectly mapping NMI
  145. code. NMIs that interrupt the entry code must be very
  146. careful and can be the cause of crashes that show up when
  147. running perf.
  148. * Kernel crashes at the first exit to userspace. entry_64.S
  149. bugs, or failing to map some of the exit code.
  150. * Crashes at first interrupt that interrupts userspace. The paths
  151. in entry_64.S that return to userspace are sometimes separate
  152. from the ones that return to the kernel.
  153. * Double faults: overflowing the kernel stack because of page
  154. faults upon page faults. Caused by touching non-pti-mapped
  155. data in the entry code, or forgetting to switch to kernel
  156. CR3 before calling into C functions which are not pti-mapped.
  157. * Userspace segfaults early in boot, sometimes manifesting
  158. as mount(8) failing to mount the rootfs. These have
  159. tended to be TLB invalidation issues. Usually invalidating
  160. the wrong PCID, or otherwise missing an invalidation.
  161. 1. https://gruss.cc/files/kaiser.pdf
  162. 2. https://meltdownattack.com/meltdown.pdf