LLVM 20.0.0git
AArch64SpeculationHardening.cpp
Go to the documentation of this file.
1//===- AArch64SpeculationHardening.cpp - Harden Against Missspeculation --===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This file contains a pass to insert code to mitigate against side channel
10// vulnerabilities that may happen under control flow miss-speculation.
11//
12// The pass implements tracking of control flow miss-speculation into a "taint"
13// register. That taint register can then be used to mask off registers with
14// sensitive data when executing under miss-speculation, a.k.a. "transient
15// execution".
16// This pass is aimed at mitigating against SpectreV1-style vulnarabilities.
17//
18// It also implements speculative load hardening, i.e. using the taint register
19// to automatically mask off loaded data.
20//
21// As a possible follow-on improvement, also an intrinsics-based approach as
22// explained at https://lwn.net/Articles/759423/ could be implemented on top of
23// the current design.
24//
25// For AArch64, the following implementation choices are made to implement the
26// tracking of control flow miss-speculation into a taint register:
27// Some of these are different than the implementation choices made in
28// the similar pass implemented in X86SpeculativeLoadHardening.cpp, as
29// the instruction set characteristics result in different trade-offs.
30// - The speculation hardening is done after register allocation. With a
31// relative abundance of registers, one register is reserved (X16) to be
32// the taint register. X16 is expected to not clash with other register
33// reservation mechanisms with very high probability because:
34// . The AArch64 ABI doesn't guarantee X16 to be retained across any call.
35// . The only way to request X16 to be used as a programmer is through
36// inline assembly. In the rare case a function explicitly demands to
37// use X16/W16, this pass falls back to hardening against speculation
38// by inserting a DSB SYS/ISB barrier pair which will prevent control
39// flow speculation.
40// - It is easy to insert mask operations at this late stage as we have
41// mask operations available that don't set flags.
42// - The taint variable contains all-ones when no miss-speculation is detected,
43// and contains all-zeros when miss-speculation is detected. Therefore, when
44// masking, an AND instruction (which only changes the register to be masked,
45// no other side effects) can easily be inserted anywhere that's needed.
46// - The tracking of miss-speculation is done by using a data-flow conditional
47// select instruction (CSEL) to evaluate the flags that were also used to
48// make conditional branch direction decisions. Speculation of the CSEL
49// instruction can be limited with a CSDB instruction - so the combination of
50// CSEL + a later CSDB gives the guarantee that the flags as used in the CSEL
51// aren't speculated. When conditional branch direction gets miss-speculated,
52// the semantics of the inserted CSEL instruction is such that the taint
53// register will contain all zero bits.
54// One key requirement for this to work is that the conditional branch is
55// followed by an execution of the CSEL instruction, where the CSEL
56// instruction needs to use the same flags status as the conditional branch.
57// This means that the conditional branches must not be implemented as one
58// of the AArch64 conditional branches that do not use the flags as input
59// (CB(N)Z and TB(N)Z). This is implemented by ensuring in the instruction
60// selectors to not produce these instructions when speculation hardening
61// is enabled. This pass will assert if it does encounter such an instruction.
62// - On function call boundaries, the miss-speculation state is transferred from
63// the taint register X16 to be encoded in the SP register as value 0.
64//
65// For the aspect of automatically hardening loads, using the taint register,
66// (a.k.a. speculative load hardening, see
67// https://llvm.org/docs/SpeculativeLoadHardening.html), the following
68// implementation choices are made for AArch64:
69// - Many of the optimizations described at
70// https://llvm.org/docs/SpeculativeLoadHardening.html to harden fewer
71// loads haven't been implemented yet - but for some of them there are
72// FIXMEs in the code.
73// - loads that load into general purpose (X or W) registers get hardened by
74// masking the loaded data. For loads that load into other registers, the
75// address loaded from gets hardened. It is expected that hardening the
76// loaded data may be more efficient; but masking data in registers other
77// than X or W is not easy and may result in being slower than just
78// hardening the X address register loaded from.
79// - On AArch64, CSDB instructions are inserted between the masking of the
80// register and its first use, to ensure there's no non-control-flow
81// speculation that might undermine the hardening mechanism.
82//
83// Future extensions/improvements could be:
84// - Implement this functionality using full speculation barriers, akin to the
85// x86-slh-lfence option. This may be more useful for the intrinsics-based
86// approach than for the SLH approach to masking.
87// Note that this pass already inserts the full speculation barriers if the
88// function for some niche reason makes use of X16/W16.
89// - no indirect branch misprediction gets protected/instrumented; but this
90// could be done for some indirect branches, such as switch jump tables.
91//===----------------------------------------------------------------------===//
92
93#include "AArch64InstrInfo.h"
94#include "AArch64Subtarget.h"
96#include "llvm/ADT/BitVector.h"
106#include "llvm/IR/DebugLoc.h"
107#include "llvm/Pass.h"
108#include "llvm/Support/CodeGen.h"
109#include "llvm/Support/Debug.h"
111#include <cassert>
112
113using namespace llvm;
114
115#define DEBUG_TYPE "aarch64-speculation-hardening"
116
117#define AARCH64_SPECULATION_HARDENING_NAME "AArch64 speculation hardening pass"
118
119static cl::opt<bool> HardenLoads("aarch64-slh-loads", cl::Hidden,
120 cl::desc("Sanitize loads from memory."),
121 cl::init(true));
122
123namespace {
124
125class AArch64SpeculationHardening : public MachineFunctionPass {
126public:
127 const TargetInstrInfo *TII;
128 const TargetRegisterInfo *TRI;
129
130 static char ID;
131
132 AArch64SpeculationHardening() : MachineFunctionPass(ID) {
134 }
135
136 bool runOnMachineFunction(MachineFunction &Fn) override;
137
138 StringRef getPassName() const override {
140 }
141
142private:
143 unsigned MisspeculatingTaintReg;
144 unsigned MisspeculatingTaintReg32Bit;
145 bool UseControlFlowSpeculationBarrier;
146 BitVector RegsNeedingCSDBBeforeUse;
147 BitVector RegsAlreadyMasked;
148
149 bool functionUsesHardeningRegister(MachineFunction &MF) const;
150 bool instrumentControlFlow(MachineBasicBlock &MBB,
151 bool &UsesFullSpeculationBarrier);
152 bool endsWithCondControlFlow(MachineBasicBlock &MBB, MachineBasicBlock *&TBB,
153 MachineBasicBlock *&FBB,
154 AArch64CC::CondCode &CondCode) const;
155 void insertTrackingCode(MachineBasicBlock &SplitEdgeBB,
156 AArch64CC::CondCode &CondCode, DebugLoc DL) const;
157 void insertSPToRegTaintPropagation(MachineBasicBlock &MBB,
159 void insertRegToSPTaintPropagation(MachineBasicBlock &MBB,
161 unsigned TmpReg) const;
162 void insertFullSpeculationBarrier(MachineBasicBlock &MBB,
164 DebugLoc DL) const;
165
166 bool slhLoads(MachineBasicBlock &MBB);
167 bool makeGPRSpeculationSafe(MachineBasicBlock &MBB,
169 MachineInstr &MI, unsigned Reg);
170 bool lowerSpeculationSafeValuePseudos(MachineBasicBlock &MBB,
171 bool UsesFullSpeculationBarrier);
172 bool expandSpeculationSafeValue(MachineBasicBlock &MBB,
174 bool UsesFullSpeculationBarrier);
176 DebugLoc DL);
177};
178
179} // end anonymous namespace
180
181char AArch64SpeculationHardening::ID = 0;
182
183INITIALIZE_PASS(AArch64SpeculationHardening, "aarch64-speculation-hardening",
185
186bool AArch64SpeculationHardening::endsWithCondControlFlow(
188 AArch64CC::CondCode &CondCode) const {
189 SmallVector<MachineOperand, 1> analyzeBranchCondCode;
190 if (TII->analyzeBranch(MBB, TBB, FBB, analyzeBranchCondCode, false))
191 return false;
192
193 // Ignore if the BB ends in an unconditional branch/fall-through.
194 if (analyzeBranchCondCode.empty())
195 return false;
196
197 // If the BB ends with a single conditional branch, FBB will be set to
198 // nullptr (see API docs for TII->analyzeBranch). For the rest of the
199 // analysis we want the FBB block to be set always.
200 assert(TBB != nullptr);
201 if (FBB == nullptr)
202 FBB = MBB.getFallThrough();
203
204 // If both the true and the false condition jump to the same basic block,
205 // there isn't need for any protection - whether the branch is speculated
206 // correctly or not, we end up executing the architecturally correct code.
207 if (TBB == FBB)
208 return false;
209
210 assert(MBB.succ_size() == 2);
211 // translate analyzeBranchCondCode to CondCode.
212 assert(analyzeBranchCondCode.size() == 1 && "unknown Cond array format");
213 CondCode = AArch64CC::CondCode(analyzeBranchCondCode[0].getImm());
214 return true;
215}
216
217void AArch64SpeculationHardening::insertFullSpeculationBarrier(
219 DebugLoc DL) const {
220 // A full control flow speculation barrier consists of (DSB SYS + ISB)
221 BuildMI(MBB, MBBI, DL, TII->get(AArch64::DSB)).addImm(0xf);
222 BuildMI(MBB, MBBI, DL, TII->get(AArch64::ISB)).addImm(0xf);
223}
224
225void AArch64SpeculationHardening::insertTrackingCode(
226 MachineBasicBlock &SplitEdgeBB, AArch64CC::CondCode &CondCode,
227 DebugLoc DL) const {
228 if (UseControlFlowSpeculationBarrier) {
229 insertFullSpeculationBarrier(SplitEdgeBB, SplitEdgeBB.begin(), DL);
230 } else {
231 BuildMI(SplitEdgeBB, SplitEdgeBB.begin(), DL, TII->get(AArch64::CSELXr))
232 .addDef(MisspeculatingTaintReg)
233 .addUse(MisspeculatingTaintReg)
234 .addUse(AArch64::XZR)
235 .addImm(CondCode);
236 SplitEdgeBB.addLiveIn(AArch64::NZCV);
237 }
238}
239
240bool AArch64SpeculationHardening::instrumentControlFlow(
241 MachineBasicBlock &MBB, bool &UsesFullSpeculationBarrier) {
242 LLVM_DEBUG(dbgs() << "Instrument control flow tracking on MBB: " << MBB);
243
244 bool Modified = false;
245 MachineBasicBlock *TBB = nullptr;
246 MachineBasicBlock *FBB = nullptr;
248
249 if (!endsWithCondControlFlow(MBB, TBB, FBB, CondCode)) {
250 LLVM_DEBUG(dbgs() << "... doesn't end with CondControlFlow\n");
251 } else {
252 // Now insert:
253 // "CSEL MisSpeculatingR, MisSpeculatingR, XZR, cond" on the True edge and
254 // "CSEL MisSpeculatingR, MisSpeculatingR, XZR, Invertcond" on the False
255 // edge.
257
258 MachineBasicBlock *SplitEdgeTBB = MBB.SplitCriticalEdge(TBB, *this);
259 MachineBasicBlock *SplitEdgeFBB = MBB.SplitCriticalEdge(FBB, *this);
260
261 assert(SplitEdgeTBB != nullptr);
262 assert(SplitEdgeFBB != nullptr);
263
264 DebugLoc DL;
265 if (MBB.instr_end() != MBB.instr_begin())
266 DL = (--MBB.instr_end())->getDebugLoc();
267
268 insertTrackingCode(*SplitEdgeTBB, CondCode, DL);
269 insertTrackingCode(*SplitEdgeFBB, InvCondCode, DL);
270
271 LLVM_DEBUG(dbgs() << "SplitEdgeTBB: " << *SplitEdgeTBB << "\n");
272 LLVM_DEBUG(dbgs() << "SplitEdgeFBB: " << *SplitEdgeFBB << "\n");
273 Modified = true;
274 }
275
276 // Perform correct code generation around function calls and before returns.
277 // The below variables record the return/terminator instructions and the call
278 // instructions respectively; including which register is available as a
279 // temporary register just before the recorded instructions.
282 // if a temporary register is not available for at least one of the
283 // instructions for which we need to transfer taint to the stack pointer, we
284 // need to insert a full speculation barrier.
285 // TmpRegisterNotAvailableEverywhere tracks that condition.
286 bool TmpRegisterNotAvailableEverywhere = false;
287
288 RegScavenger RS;
290
291 for (MachineBasicBlock::iterator I = MBB.end(); I != MBB.begin(); ) {
292 MachineInstr &MI = *--I;
293 if (!MI.isReturn() && !MI.isCall())
294 continue;
295
296 // The RegScavenger represents registers available *after* the MI
297 // instruction pointed to by RS.getCurrentPosition().
298 // We need to have a register that is available *before* the MI is executed.
299 if (I == MBB.begin())
301 else
302 RS.backward(I);
303 // FIXME: The below just finds *a* unused register. Maybe code could be
304 // optimized more if this looks for the register that isn't used for the
305 // longest time around this place, to enable more scheduling freedom. Not
306 // sure if that would actually result in a big performance difference
307 // though. Maybe RegisterScavenger::findSurvivorBackwards has some logic
308 // already to do this - but it's unclear if that could easily be used here.
309 Register TmpReg = RS.FindUnusedReg(&AArch64::GPR64commonRegClass);
310 LLVM_DEBUG(dbgs() << "RS finds "
311 << ((TmpReg == 0) ? "no register " : "register ");
312 if (TmpReg != 0) dbgs() << printReg(TmpReg, TRI) << " ";
313 dbgs() << "to be available at MI " << MI);
314 if (TmpReg == 0)
315 TmpRegisterNotAvailableEverywhere = true;
316 if (MI.isReturn())
317 ReturnInstructions.push_back({&MI, TmpReg});
318 else if (MI.isCall())
319 CallInstructions.push_back({&MI, TmpReg});
320 }
321
322 if (TmpRegisterNotAvailableEverywhere) {
323 // When a temporary register is not available everywhere in this basic
324 // basic block where a propagate-taint-to-sp operation is needed, just
325 // emit a full speculation barrier at the start of this basic block, which
326 // renders the taint/speculation tracking in this basic block unnecessary.
327 insertFullSpeculationBarrier(MBB, MBB.begin(),
328 (MBB.begin())->getDebugLoc());
329 UsesFullSpeculationBarrier = true;
330 Modified = true;
331 } else {
332 for (auto MI_Reg : ReturnInstructions) {
333 assert(MI_Reg.second != 0);
335 dbgs()
336 << " About to insert Reg to SP taint propagation with temp register "
337 << printReg(MI_Reg.second, TRI)
338 << " on instruction: " << *MI_Reg.first);
339 insertRegToSPTaintPropagation(MBB, MI_Reg.first, MI_Reg.second);
340 Modified = true;
341 }
342
343 for (auto MI_Reg : CallInstructions) {
344 assert(MI_Reg.second != 0);
345 LLVM_DEBUG(dbgs() << " About to insert Reg to SP and back taint "
346 "propagation with temp register "
347 << printReg(MI_Reg.second, TRI)
348 << " around instruction: " << *MI_Reg.first);
349 // Just after the call:
350 insertSPToRegTaintPropagation(
351 MBB, std::next((MachineBasicBlock::iterator)MI_Reg.first));
352 // Just before the call:
353 insertRegToSPTaintPropagation(MBB, MI_Reg.first, MI_Reg.second);
354 Modified = true;
355 }
356 }
357 return Modified;
358}
359
360void AArch64SpeculationHardening::insertSPToRegTaintPropagation(
362 // If full control flow speculation barriers are used, emit a control flow
363 // barrier to block potential miss-speculation in flight coming in to this
364 // function.
365 if (UseControlFlowSpeculationBarrier) {
366 insertFullSpeculationBarrier(MBB, MBBI, DebugLoc());
367 return;
368 }
369
370 // CMP SP, #0 === SUBS xzr, SP, #0
371 BuildMI(MBB, MBBI, DebugLoc(), TII->get(AArch64::SUBSXri))
372 .addDef(AArch64::XZR)
373 .addUse(AArch64::SP)
374 .addImm(0)
375 .addImm(0); // no shift
376 // CSETM x16, NE === CSINV x16, xzr, xzr, EQ
377 BuildMI(MBB, MBBI, DebugLoc(), TII->get(AArch64::CSINVXr))
378 .addDef(MisspeculatingTaintReg)
379 .addUse(AArch64::XZR)
380 .addUse(AArch64::XZR)
382}
383
384void AArch64SpeculationHardening::insertRegToSPTaintPropagation(
386 unsigned TmpReg) const {
387 // If full control flow speculation barriers are used, there will not be
388 // miss-speculation when returning from this function, and therefore, also
389 // no need to encode potential miss-speculation into the stack pointer.
390 if (UseControlFlowSpeculationBarrier)
391 return;
392
393 // mov Xtmp, SP === ADD Xtmp, SP, #0
394 BuildMI(MBB, MBBI, DebugLoc(), TII->get(AArch64::ADDXri))
395 .addDef(TmpReg)
396 .addUse(AArch64::SP)
397 .addImm(0)
398 .addImm(0); // no shift
399 // and Xtmp, Xtmp, TaintReg === AND Xtmp, Xtmp, TaintReg, #0
400 BuildMI(MBB, MBBI, DebugLoc(), TII->get(AArch64::ANDXrs))
403 .addUse(MisspeculatingTaintReg, RegState::Kill)
404 .addImm(0);
405 // mov SP, Xtmp === ADD SP, Xtmp, #0
406 BuildMI(MBB, MBBI, DebugLoc(), TII->get(AArch64::ADDXri))
407 .addDef(AArch64::SP)
408 .addUse(TmpReg, RegState::Kill)
409 .addImm(0)
410 .addImm(0); // no shift
411}
412
413bool AArch64SpeculationHardening::functionUsesHardeningRegister(
414 MachineFunction &MF) const {
415 for (MachineBasicBlock &MBB : MF) {
416 for (MachineInstr &MI : MBB) {
417 // treat function calls specially, as the hardening register does not
418 // need to remain live across function calls.
419 if (MI.isCall())
420 continue;
421 if (MI.readsRegister(MisspeculatingTaintReg, TRI) ||
422 MI.modifiesRegister(MisspeculatingTaintReg, TRI))
423 return true;
424 }
425 }
426 return false;
427}
428
429// Make GPR register Reg speculation-safe by putting it through the
430// SpeculationSafeValue pseudo instruction, if we can't prove that
431// the value in the register has already been hardened.
432bool AArch64SpeculationHardening::makeGPRSpeculationSafe(
434 unsigned Reg) {
435 assert(AArch64::GPR32allRegClass.contains(Reg) ||
436 AArch64::GPR64allRegClass.contains(Reg));
437
438 // Loads cannot directly load a value into the SP (nor WSP).
439 // Therefore, if Reg is SP or WSP, it is because the instruction loads from
440 // the stack through the stack pointer.
441 //
442 // Since the stack pointer is never dynamically controllable, don't harden it.
443 if (Reg == AArch64::SP || Reg == AArch64::WSP)
444 return false;
445
446 // Do not harden the register again if already hardened before.
447 if (RegsAlreadyMasked[Reg])
448 return false;
449
450 const bool Is64Bit = AArch64::GPR64allRegClass.contains(Reg);
451 LLVM_DEBUG(dbgs() << "About to harden register : " << Reg << "\n");
452 BuildMI(MBB, MBBI, MI.getDebugLoc(),
453 TII->get(Is64Bit ? AArch64::SpeculationSafeValueX
454 : AArch64::SpeculationSafeValueW))
455 .addDef(Reg)
456 .addUse(Reg);
457 RegsAlreadyMasked.set(Reg);
458 return true;
459}
460
461bool AArch64SpeculationHardening::slhLoads(MachineBasicBlock &MBB) {
462 bool Modified = false;
463
464 LLVM_DEBUG(dbgs() << "slhLoads running on MBB: " << MBB);
465
466 RegsAlreadyMasked.reset();
467
470 for (; MBBI != E; MBBI = NextMBBI) {
471 MachineInstr &MI = *MBBI;
472 NextMBBI = std::next(MBBI);
473 // Only harden loaded values or addresses used in loads.
474 if (!MI.mayLoad())
475 continue;
476
477 LLVM_DEBUG(dbgs() << "About to harden: " << MI);
478
479 // For general purpose register loads, harden the registers loaded into.
480 // For other loads, harden the address loaded from.
481 // Masking the loaded value is expected to result in less performance
482 // overhead, as the load can still execute speculatively in comparison to
483 // when the address loaded from gets masked. However, masking is only
484 // easy to do efficiently on GPR registers, so for loads into non-GPR
485 // registers (e.g. floating point loads), mask the address loaded from.
486 bool AllDefsAreGPR = llvm::all_of(MI.defs(), [&](MachineOperand &Op) {
487 return Op.isReg() && (AArch64::GPR32allRegClass.contains(Op.getReg()) ||
488 AArch64::GPR64allRegClass.contains(Op.getReg()));
489 });
490 // FIXME: it might be a worthwhile optimization to not mask loaded
491 // values if all the registers involved in address calculation are already
492 // hardened, leading to this load not able to execute on a miss-speculated
493 // path.
494 bool HardenLoadedData = AllDefsAreGPR;
495 bool HardenAddressLoadedFrom = !HardenLoadedData;
496
497 // First remove registers from AlreadyMaskedRegisters if their value is
498 // updated by this instruction - it makes them contain a new value that is
499 // not guaranteed to already have been masked.
500 for (MachineOperand Op : MI.defs())
501 for (MCRegAliasIterator AI(Op.getReg(), TRI, true); AI.isValid(); ++AI)
502 RegsAlreadyMasked.reset(*AI);
503
504 // FIXME: loads from the stack with an immediate offset from the stack
505 // pointer probably shouldn't be hardened, which could result in a
506 // significant optimization. See section "Don’t check loads from
507 // compile-time constant stack offsets", in
508 // https://llvm.org/docs/SpeculativeLoadHardening.html
509
510 if (HardenLoadedData)
511 for (auto Def : MI.defs()) {
512 if (Def.isDead())
513 // Do not mask a register that is not used further.
514 continue;
515 // FIXME: For pre/post-increment addressing modes, the base register
516 // used in address calculation is also defined by this instruction.
517 // It might be a worthwhile optimization to not harden that
518 // base register increment/decrement when the increment/decrement is
519 // an immediate.
520 Modified |= makeGPRSpeculationSafe(MBB, NextMBBI, MI, Def.getReg());
521 }
522
523 if (HardenAddressLoadedFrom)
524 for (auto Use : MI.uses()) {
525 if (!Use.isReg())
526 continue;
527 Register Reg = Use.getReg();
528 // Some loads of floating point data have implicit defs/uses on a
529 // super register of that floating point data. Some examples:
530 // $s0 = LDRSui $sp, 22, implicit-def $q0
531 // $q0 = LD1i64 $q0, 1, renamable $x0
532 // We need to filter out these uses for non-GPR register which occur
533 // because the load partially fills a non-GPR register with the loaded
534 // data. Just skipping all non-GPR registers is safe (for now) as all
535 // AArch64 load instructions only use GPR registers to perform the
536 // address calculation. FIXME: However that might change once we can
537 // produce SVE gather instructions.
538 if (!(AArch64::GPR32allRegClass.contains(Reg) ||
539 AArch64::GPR64allRegClass.contains(Reg)))
540 continue;
541 Modified |= makeGPRSpeculationSafe(MBB, MBBI, MI, Reg);
542 }
543 }
544 return Modified;
545}
546
547/// \brief If MBBI references a pseudo instruction that should be expanded
548/// here, do the expansion and return true. Otherwise return false.
549bool AArch64SpeculationHardening::expandSpeculationSafeValue(
551 bool UsesFullSpeculationBarrier) {
552 MachineInstr &MI = *MBBI;
553 unsigned Opcode = MI.getOpcode();
554 bool Is64Bit = true;
555
556 switch (Opcode) {
557 default:
558 break;
559 case AArch64::SpeculationSafeValueW:
560 Is64Bit = false;
561 [[fallthrough]];
562 case AArch64::SpeculationSafeValueX:
563 // Just remove the SpeculationSafe pseudo's if control flow
564 // miss-speculation isn't happening because we're already inserting barriers
565 // to guarantee that.
566 if (!UseControlFlowSpeculationBarrier && !UsesFullSpeculationBarrier) {
567 Register DstReg = MI.getOperand(0).getReg();
568 Register SrcReg = MI.getOperand(1).getReg();
569 // Mark this register and all its aliasing registers as needing to be
570 // value speculation hardened before its next use, by using a CSDB
571 // barrier instruction.
572 for (MachineOperand Op : MI.defs())
573 for (MCRegAliasIterator AI(Op.getReg(), TRI, true); AI.isValid(); ++AI)
574 RegsNeedingCSDBBeforeUse.set(*AI);
575
576 // Mask off with taint state.
577 BuildMI(MBB, MBBI, MI.getDebugLoc(),
578 Is64Bit ? TII->get(AArch64::ANDXrs) : TII->get(AArch64::ANDWrs))
579 .addDef(DstReg)
580 .addUse(SrcReg, RegState::Kill)
581 .addUse(Is64Bit ? MisspeculatingTaintReg
582 : MisspeculatingTaintReg32Bit)
583 .addImm(0);
584 }
585 MI.eraseFromParent();
586 return true;
587 }
588 return false;
589}
590
591bool AArch64SpeculationHardening::insertCSDB(MachineBasicBlock &MBB,
593 DebugLoc DL) {
594 assert(!UseControlFlowSpeculationBarrier && "No need to insert CSDBs when "
595 "control flow miss-speculation "
596 "is already blocked");
597 // insert data value speculation barrier (CSDB)
598 BuildMI(MBB, MBBI, DL, TII->get(AArch64::HINT)).addImm(0x14);
599 RegsNeedingCSDBBeforeUse.reset();
600 return true;
601}
602
603bool AArch64SpeculationHardening::lowerSpeculationSafeValuePseudos(
604 MachineBasicBlock &MBB, bool UsesFullSpeculationBarrier) {
605 bool Modified = false;
606
607 RegsNeedingCSDBBeforeUse.reset();
608
609 // The following loop iterates over all instructions in the basic block,
610 // and performs 2 operations:
611 // 1. Insert a CSDB at this location if needed.
612 // 2. Expand the SpeculationSafeValuePseudo if the current instruction is
613 // one.
614 //
615 // The insertion of the CSDB is done as late as possible (i.e. just before
616 // the use of a masked register), in the hope that that will reduce the
617 // total number of CSDBs in a block when there are multiple masked registers
618 // in the block.
620 DebugLoc DL;
621 while (MBBI != E) {
622 MachineInstr &MI = *MBBI;
623 DL = MI.getDebugLoc();
624 MachineBasicBlock::iterator NMBBI = std::next(MBBI);
625
626 // First check if a CSDB needs to be inserted due to earlier registers
627 // that were masked and that are used by the next instruction.
628 // Also emit the barrier on any potential control flow changes.
629 bool NeedToEmitBarrier = false;
630 if (RegsNeedingCSDBBeforeUse.any() && (MI.isCall() || MI.isTerminator()))
631 NeedToEmitBarrier = true;
632 if (!NeedToEmitBarrier)
633 for (MachineOperand Op : MI.uses())
634 if (Op.isReg() && RegsNeedingCSDBBeforeUse[Op.getReg()]) {
635 NeedToEmitBarrier = true;
636 break;
637 }
638
639 if (NeedToEmitBarrier && !UsesFullSpeculationBarrier)
640 Modified |= insertCSDB(MBB, MBBI, DL);
641
642 Modified |=
643 expandSpeculationSafeValue(MBB, MBBI, UsesFullSpeculationBarrier);
644
645 MBBI = NMBBI;
646 }
647
648 if (RegsNeedingCSDBBeforeUse.any() && !UsesFullSpeculationBarrier)
649 Modified |= insertCSDB(MBB, MBBI, DL);
650
651 return Modified;
652}
653
654bool AArch64SpeculationHardening::runOnMachineFunction(MachineFunction &MF) {
655 if (!MF.getFunction().hasFnAttribute(Attribute::SpeculativeLoadHardening))
656 return false;
657
658 MisspeculatingTaintReg = AArch64::X16;
659 MisspeculatingTaintReg32Bit = AArch64::W16;
662 RegsNeedingCSDBBeforeUse.resize(TRI->getNumRegs());
663 RegsAlreadyMasked.resize(TRI->getNumRegs());
664 UseControlFlowSpeculationBarrier = functionUsesHardeningRegister(MF);
665
666 bool Modified = false;
667
668 // Step 1: Enable automatic insertion of SpeculationSafeValue.
669 if (HardenLoads) {
671 dbgs() << "***** AArch64SpeculationHardening - automatic insertion of "
672 "SpeculationSafeValue intrinsics *****\n");
673 for (auto &MBB : MF)
674 Modified |= slhLoads(MBB);
675 }
676
677 // 2. Add instrumentation code to function entry and exits.
679 dbgs()
680 << "***** AArch64SpeculationHardening - track control flow *****\n");
681
683 EntryBlocks.push_back(&MF.front());
684 for (const LandingPadInfo &LPI : MF.getLandingPads())
685 EntryBlocks.push_back(LPI.LandingPadBlock);
686 for (auto *Entry : EntryBlocks)
687 insertSPToRegTaintPropagation(
688 *Entry, Entry->SkipPHIsLabelsAndDebug(Entry->begin()));
689
690 // 3. Add instrumentation code to every basic block.
691 for (auto &MBB : MF) {
692 bool UsesFullSpeculationBarrier = false;
693 Modified |= instrumentControlFlow(MBB, UsesFullSpeculationBarrier);
694 Modified |=
695 lowerSpeculationSafeValuePseudos(MBB, UsesFullSpeculationBarrier);
696 }
697
698 return Modified;
699}
700
701/// \brief Returns an instance of the pseudo instruction expansion pass.
703 return new AArch64SpeculationHardening();
704}
#define AARCH64_SPECULATION_HARDENING_NAME
static cl::opt< bool > HardenLoads("aarch64-slh-loads", cl::Hidden, cl::desc("Sanitize loads from memory."), cl::init(true))
MachineBasicBlock & MBB
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
MachineBasicBlock MachineBasicBlock::iterator MBBI
This file implements the BitVector class.
#define LLVM_DEBUG(X)
Definition: Debug.h:101
const HexagonInstrInfo * TII
IRTranslator LLVM IR MI
#define I(x, y, z)
Definition: MD5.cpp:58
static DebugLoc getDebugLoc(MachineBasicBlock::instr_iterator FirstMI, MachineBasicBlock::instr_iterator LastMI)
Return the first found DebugLoc that has a DILocation, given a range of instructions.
unsigned const TargetRegisterInfo * TRI
#define INITIALIZE_PASS(passName, arg, name, cfg, analysis)
Definition: PassSupport.h:38
const SmallVectorImpl< MachineOperand > MachineBasicBlock * TBB
This file declares the machine register scavenger class.
assert(ImpDefSCC.getReg()==AMDGPU::SCC &&ImpDefSCC.isDef())
This file defines the SmallVector class.
static bool contains(SmallPtrSetImpl< ConstantExpr * > &Cache, ConstantExpr *Expr, Constant *C)
Definition: Value.cpp:469
This class represents an Operation in the Expression.
A debug info location.
Definition: DebugLoc.h:33
FunctionPass class - This class is used to implement most global optimizations.
Definition: Pass.h:310
bool hasFnAttribute(Attribute::AttrKind Kind) const
Return true if the function has the attribute.
Definition: Function.cpp:743
bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB, MachineBasicBlock *&FBB, SmallVectorImpl< MachineOperand > &Cond, bool AllowModify) const override
Analyze the branching code at the end of MBB, returning true if it cannot be understood (e....
MCRegAliasIterator enumerates all registers aliasing Reg.
instr_iterator instr_begin()
MachineBasicBlock * getFallThrough(bool JumpToFallThrough=true)
Return the fallthrough block if the block can implicitly transfer control to the block after it by fa...
unsigned succ_size() const
instr_iterator instr_end()
void addLiveIn(MCRegister PhysReg, LaneBitmask LaneMask=LaneBitmask::getAll())
Adds the specified register as a live in.
MachineBasicBlock * SplitCriticalEdge(MachineBasicBlock *Succ, Pass &P, std::vector< SparseBitVector<> > *LiveInSets=nullptr)
Split the critical edge from this block to the given successor block, and return the newly created bl...
MachineFunctionPass - This class adapts the FunctionPass interface to allow convenient creation of pa...
virtual bool runOnMachineFunction(MachineFunction &MF)=0
runOnMachineFunction - This method must be overloaded to perform the desired machine code transformat...
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
Function & getFunction()
Return the LLVM function that this machine code represents.
const std::vector< LandingPadInfo > & getLandingPads() const
Return a reference to the landing pad info for the current function.
const MachineBasicBlock & front() const
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & addUse(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a virtual register use operand.
const MachineInstrBuilder & addDef(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a virtual register definition operand.
Representation of each machine instruction.
Definition: MachineInstr.h:69
MachineOperand class - Representation of each machine instruction operand.
static PassRegistry * getPassRegistry()
getPassRegistry - Access the global registry object, which is automatically initialized at applicatio...
virtual StringRef getPassName() const
getPassName - Return a nice clean name for a pass.
Definition: Pass.cpp:81
void enterBasicBlockEnd(MachineBasicBlock &MBB)
Start tracking liveness from the end of basic block MBB.
Register FindUnusedReg(const TargetRegisterClass *RC) const
Find an unused register of the specified register class.
void backward()
Update internal register state and move MBB iterator backwards.
void enterBasicBlock(MachineBasicBlock &MBB)
Start tracking liveness from the begin of basic block MBB.
Wrapper class representing virtual and physical registers.
Definition: Register.h:19
bool empty() const
Definition: SmallVector.h:94
size_t size() const
Definition: SmallVector.h:91
void push_back(const T &Elt)
Definition: SmallVector.h:426
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
Definition: SmallVector.h:1209
StringRef - Represent a constant reference to a string, i.e.
Definition: StringRef.h:50
TargetInstrInfo - Interface to description of machine instruction set.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
virtual const TargetRegisterInfo * getRegisterInfo() const
getRegisterInfo - If register information is available, return it.
virtual const TargetInstrInfo * getInstrInfo() const
A Use represents the edge between a Value definition and its users.
Definition: Use.h:43
static CondCode getInvertedCondCode(CondCode Code)
@ Entry
Definition: COFF.h:826
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
Definition: CallingConv.h:24
CondCode
ISD::CondCode enum - These are ordered carefully to make the bitfields below work out,...
Definition: ISDOpcodes.h:1603
@ Renamable
Register that may be renamed.
@ Kill
The last use of a register.
Reg
All possible values of the reg field in the ModR/M byte.
initializer< Ty > init(const Ty &Val)
Definition: CommandLine.h:443
NodeAddr< DefNode * > Def
Definition: RDFGraph.h:384
This is an optimization pass for GlobalISel generic memory operations.
Definition: AddressRanges.h:18
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
Definition: STLExtras.h:1722
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
void initializeAArch64SpeculationHardeningPass(PassRegistry &)
raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
Definition: Debug.cpp:163
FunctionPass * createAArch64SpeculationHardeningPass()
Returns an instance of the pseudo instruction expansion pass.
Printable printReg(Register Reg, const TargetRegisterInfo *TRI=nullptr, unsigned SubIdx=0, const MachineRegisterInfo *MRI=nullptr)
Prints virtual and physical registers with or without a TRI instance.
This structure is used to retain landing pad info for the current function.