| 1 | //===------------ BPFCheckAndAdjustIR.cpp - Check and Adjust IR -----------===// |
| 2 | // |
| 3 | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
| 4 | // See https://llvm.org/LICENSE.txt for license information. |
| 5 | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
| 6 | // |
| 7 | //===----------------------------------------------------------------------===// |
| 8 | // |
| 9 | // Check IR and adjust IR for verifier friendly codes. |
| 10 | // The following are done for IR checking: |
| 11 | // - no relocation globals in PHI node. |
| 12 | // The following are done for IR adjustment: |
| 13 | // - remove __builtin_bpf_passthrough builtins. Target independent IR |
| 14 | // optimizations are done and those builtins can be removed. |
| 15 | // - remove llvm.bpf.getelementptr.and.load builtins. |
| 16 | // - remove llvm.bpf.getelementptr.and.store builtins. |
| 17 | // - for loads and stores with base addresses from non-zero address space |
| 18 | // cast base address to zero address space (support for BPF address spaces). |
| 19 | // |
| 20 | //===----------------------------------------------------------------------===// |
| 21 | |
| 22 | #include "BPF.h" |
| 23 | #include "BPFCORE.h" |
| 24 | #include "llvm/Analysis/LoopInfo.h" |
| 25 | #include "llvm/IR/GlobalVariable.h" |
| 26 | #include "llvm/IR/IRBuilder.h" |
| 27 | #include "llvm/IR/Instruction.h" |
| 28 | #include "llvm/IR/Instructions.h" |
| 29 | #include "llvm/IR/IntrinsicsBPF.h" |
| 30 | #include "llvm/IR/Module.h" |
| 31 | #include "llvm/IR/Type.h" |
| 32 | #include "llvm/IR/Value.h" |
| 33 | #include "llvm/Pass.h" |
| 34 | #include "llvm/Transforms/Utils/BasicBlockUtils.h" |
| 35 | |
| 36 | #define DEBUG_TYPE "bpf-check-and-opt-ir" |
| 37 | |
| 38 | using namespace llvm; |
| 39 | |
| 40 | namespace { |
| 41 | |
| 42 | class BPFCheckAndAdjustIR final : public ModulePass { |
| 43 | bool runOnModule(Module &F) override; |
| 44 | |
| 45 | public: |
| 46 | static char ID; |
| 47 | BPFCheckAndAdjustIR() : ModulePass(ID) {} |
| 48 | virtual void getAnalysisUsage(AnalysisUsage &AU) const override; |
| 49 | |
| 50 | private: |
| 51 | void checkIR(Module &M); |
| 52 | bool adjustIR(Module &M); |
| 53 | bool removePassThroughBuiltin(Module &M); |
| 54 | bool removeCompareBuiltin(Module &M); |
| 55 | bool sinkMinMax(Module &M); |
| 56 | bool removeGEPBuiltins(Module &M); |
| 57 | bool insertASpaceCasts(Module &M); |
| 58 | }; |
| 59 | } // End anonymous namespace |
| 60 | |
| 61 | char BPFCheckAndAdjustIR::ID = 0; |
| 62 | INITIALIZE_PASS(BPFCheckAndAdjustIR, DEBUG_TYPE, "BPF Check And Adjust IR" , |
| 63 | false, false) |
| 64 | |
| 65 | ModulePass *llvm::createBPFCheckAndAdjustIR() { |
| 66 | return new BPFCheckAndAdjustIR(); |
| 67 | } |
| 68 | |
| 69 | void BPFCheckAndAdjustIR::checkIR(Module &M) { |
| 70 | // Ensure relocation global won't appear in PHI node |
| 71 | // This may happen if the compiler generated the following code: |
| 72 | // B1: |
| 73 | // g1 = @llvm.skb_buff:0:1... |
| 74 | // ... |
| 75 | // goto B_COMMON |
| 76 | // B2: |
| 77 | // g2 = @llvm.skb_buff:0:2... |
| 78 | // ... |
| 79 | // goto B_COMMON |
| 80 | // B_COMMON: |
| 81 | // g = PHI(g1, g2) |
| 82 | // x = load g |
| 83 | // ... |
| 84 | // If anything likes the above "g = PHI(g1, g2)", issue a fatal error. |
| 85 | for (Function &F : M) |
| 86 | for (auto &BB : F) |
| 87 | for (auto &I : BB) { |
| 88 | PHINode *PN = dyn_cast<PHINode>(Val: &I); |
| 89 | if (!PN || PN->use_empty()) |
| 90 | continue; |
| 91 | for (int i = 0, e = PN->getNumIncomingValues(); i < e; ++i) { |
| 92 | auto *GV = dyn_cast<GlobalVariable>(Val: PN->getIncomingValue(i)); |
| 93 | if (!GV) |
| 94 | continue; |
| 95 | if (GV->hasAttribute(Kind: BPFCoreSharedInfo::AmaAttr) || |
| 96 | GV->hasAttribute(Kind: BPFCoreSharedInfo::TypeIdAttr)) |
| 97 | report_fatal_error(reason: "relocation global in PHI node" ); |
| 98 | } |
| 99 | } |
| 100 | } |
| 101 | |
| 102 | bool BPFCheckAndAdjustIR::removePassThroughBuiltin(Module &M) { |
| 103 | // Remove __builtin_bpf_passthrough()'s which are used to prevent |
| 104 | // certain IR optimizations. Now major IR optimizations are done, |
| 105 | // remove them. |
| 106 | bool Changed = false; |
| 107 | CallInst *ToBeDeleted = nullptr; |
| 108 | for (Function &F : M) |
| 109 | for (auto &BB : F) |
| 110 | for (auto &I : BB) { |
| 111 | if (ToBeDeleted) { |
| 112 | ToBeDeleted->eraseFromParent(); |
| 113 | ToBeDeleted = nullptr; |
| 114 | } |
| 115 | |
| 116 | auto *Call = dyn_cast<CallInst>(Val: &I); |
| 117 | if (!Call) |
| 118 | continue; |
| 119 | auto *GV = dyn_cast<GlobalValue>(Val: Call->getCalledOperand()); |
| 120 | if (!GV) |
| 121 | continue; |
| 122 | if (!GV->getName().starts_with(Prefix: "llvm.bpf.passthrough" )) |
| 123 | continue; |
| 124 | Changed = true; |
| 125 | Value *Arg = Call->getArgOperand(i: 1); |
| 126 | Call->replaceAllUsesWith(V: Arg); |
| 127 | ToBeDeleted = Call; |
| 128 | } |
| 129 | return Changed; |
| 130 | } |
| 131 | |
| 132 | bool BPFCheckAndAdjustIR::removeCompareBuiltin(Module &M) { |
| 133 | // Remove __builtin_bpf_compare()'s which are used to prevent |
| 134 | // certain IR optimizations. Now major IR optimizations are done, |
| 135 | // remove them. |
| 136 | bool Changed = false; |
| 137 | CallInst *ToBeDeleted = nullptr; |
| 138 | for (Function &F : M) |
| 139 | for (auto &BB : F) |
| 140 | for (auto &I : BB) { |
| 141 | if (ToBeDeleted) { |
| 142 | ToBeDeleted->eraseFromParent(); |
| 143 | ToBeDeleted = nullptr; |
| 144 | } |
| 145 | |
| 146 | auto *Call = dyn_cast<CallInst>(Val: &I); |
| 147 | if (!Call) |
| 148 | continue; |
| 149 | auto *GV = dyn_cast<GlobalValue>(Val: Call->getCalledOperand()); |
| 150 | if (!GV) |
| 151 | continue; |
| 152 | if (!GV->getName().starts_with(Prefix: "llvm.bpf.compare" )) |
| 153 | continue; |
| 154 | |
| 155 | Changed = true; |
| 156 | Value *Arg0 = Call->getArgOperand(i: 0); |
| 157 | Value *Arg1 = Call->getArgOperand(i: 1); |
| 158 | Value *Arg2 = Call->getArgOperand(i: 2); |
| 159 | |
| 160 | auto OpVal = cast<ConstantInt>(Val: Arg0)->getValue().getZExtValue(); |
| 161 | CmpInst::Predicate Opcode = (CmpInst::Predicate)OpVal; |
| 162 | |
| 163 | auto *ICmp = new ICmpInst(Opcode, Arg1, Arg2); |
| 164 | ICmp->insertBefore(InsertPos: Call->getIterator()); |
| 165 | |
| 166 | Call->replaceAllUsesWith(V: ICmp); |
| 167 | ToBeDeleted = Call; |
| 168 | } |
| 169 | return Changed; |
| 170 | } |
| 171 | |
| 172 | struct MinMaxSinkInfo { |
| 173 | ICmpInst *ICmp; |
| 174 | Value *Other; |
| 175 | ICmpInst::Predicate Predicate; |
| 176 | CallInst *MinMax; |
| 177 | ZExtInst *ZExt; |
| 178 | SExtInst *SExt; |
| 179 | |
| 180 | MinMaxSinkInfo(ICmpInst *ICmp, Value *Other, ICmpInst::Predicate Predicate) |
| 181 | : ICmp(ICmp), Other(Other), Predicate(Predicate), MinMax(nullptr), |
| 182 | ZExt(nullptr), SExt(nullptr) {} |
| 183 | }; |
| 184 | |
| 185 | static bool sinkMinMaxInBB(BasicBlock &BB, |
| 186 | const std::function<bool(Instruction *)> &Filter) { |
| 187 | // Check if V is: |
| 188 | // (fn %a %b) or (ext (fn %a %b)) |
| 189 | // Where: |
| 190 | // ext := sext | zext |
| 191 | // fn := smin | umin | smax | umax |
| 192 | auto IsMinMaxCall = [=](Value *V, MinMaxSinkInfo &Info) { |
| 193 | if (auto *ZExt = dyn_cast<ZExtInst>(Val: V)) { |
| 194 | V = ZExt->getOperand(i_nocapture: 0); |
| 195 | Info.ZExt = ZExt; |
| 196 | } else if (auto *SExt = dyn_cast<SExtInst>(Val: V)) { |
| 197 | V = SExt->getOperand(i_nocapture: 0); |
| 198 | Info.SExt = SExt; |
| 199 | } |
| 200 | |
| 201 | auto *Call = dyn_cast<CallInst>(Val: V); |
| 202 | if (!Call) |
| 203 | return false; |
| 204 | |
| 205 | auto *Called = dyn_cast<Function>(Val: Call->getCalledOperand()); |
| 206 | if (!Called) |
| 207 | return false; |
| 208 | |
| 209 | switch (Called->getIntrinsicID()) { |
| 210 | case Intrinsic::smin: |
| 211 | case Intrinsic::umin: |
| 212 | case Intrinsic::smax: |
| 213 | case Intrinsic::umax: |
| 214 | break; |
| 215 | default: |
| 216 | return false; |
| 217 | } |
| 218 | |
| 219 | if (!Filter(Call)) |
| 220 | return false; |
| 221 | |
| 222 | Info.MinMax = Call; |
| 223 | |
| 224 | return true; |
| 225 | }; |
| 226 | |
| 227 | auto ZeroOrSignExtend = [](IRBuilder<> &Builder, Value *V, |
| 228 | MinMaxSinkInfo &Info) { |
| 229 | if (Info.SExt) { |
| 230 | if (Info.SExt->getType() == V->getType()) |
| 231 | return V; |
| 232 | return Builder.CreateSExt(V, DestTy: Info.SExt->getType()); |
| 233 | } |
| 234 | if (Info.ZExt) { |
| 235 | if (Info.ZExt->getType() == V->getType()) |
| 236 | return V; |
| 237 | return Builder.CreateZExt(V, DestTy: Info.ZExt->getType()); |
| 238 | } |
| 239 | return V; |
| 240 | }; |
| 241 | |
| 242 | bool Changed = false; |
| 243 | SmallVector<MinMaxSinkInfo, 2> SinkList; |
| 244 | |
| 245 | // Check BB for instructions like: |
| 246 | // insn := (icmp %a (fn ...)) | (icmp (fn ...) %a) |
| 247 | // |
| 248 | // Where: |
| 249 | // fn := min | max | (sext (min ...)) | (sext (max ...)) |
| 250 | // |
| 251 | // Put such instructions to SinkList. |
| 252 | for (Instruction &I : BB) { |
| 253 | ICmpInst *ICmp = dyn_cast<ICmpInst>(Val: &I); |
| 254 | if (!ICmp) |
| 255 | continue; |
| 256 | if (!ICmp->isRelational()) |
| 257 | continue; |
| 258 | MinMaxSinkInfo First(ICmp, ICmp->getOperand(i_nocapture: 1), |
| 259 | ICmpInst::getSwappedPredicate(pred: ICmp->getPredicate())); |
| 260 | MinMaxSinkInfo Second(ICmp, ICmp->getOperand(i_nocapture: 0), ICmp->getPredicate()); |
| 261 | bool FirstMinMax = IsMinMaxCall(ICmp->getOperand(i_nocapture: 0), First); |
| 262 | bool SecondMinMax = IsMinMaxCall(ICmp->getOperand(i_nocapture: 1), Second); |
| 263 | if (!(FirstMinMax ^ SecondMinMax)) |
| 264 | continue; |
| 265 | SinkList.push_back(Elt: FirstMinMax ? First : Second); |
| 266 | } |
| 267 | |
| 268 | // Iterate SinkList and replace each (icmp ...) with corresponding |
| 269 | // `x < a && x < b` or similar expression. |
| 270 | for (auto &Info : SinkList) { |
| 271 | ICmpInst *ICmp = Info.ICmp; |
| 272 | CallInst *MinMax = Info.MinMax; |
| 273 | Intrinsic::ID IID = MinMax->getCalledFunction()->getIntrinsicID(); |
| 274 | ICmpInst::Predicate P = Info.Predicate; |
| 275 | if (ICmpInst::isSigned(predicate: P) && IID != Intrinsic::smin && |
| 276 | IID != Intrinsic::smax) |
| 277 | continue; |
| 278 | |
| 279 | IRBuilder<> Builder(ICmp); |
| 280 | Value *X = Info.Other; |
| 281 | Value *A = ZeroOrSignExtend(Builder, MinMax->getArgOperand(i: 0), Info); |
| 282 | Value *B = ZeroOrSignExtend(Builder, MinMax->getArgOperand(i: 1), Info); |
| 283 | bool IsMin = IID == Intrinsic::smin || IID == Intrinsic::umin; |
| 284 | bool IsMax = IID == Intrinsic::smax || IID == Intrinsic::umax; |
| 285 | bool IsLess = ICmpInst::isLE(P) || ICmpInst::isLT(P); |
| 286 | bool IsGreater = ICmpInst::isGE(P) || ICmpInst::isGT(P); |
| 287 | assert(IsMin ^ IsMax); |
| 288 | assert(IsLess ^ IsGreater); |
| 289 | |
| 290 | Value *Replacement; |
| 291 | Value *LHS = Builder.CreateICmp(P, LHS: X, RHS: A); |
| 292 | Value *RHS = Builder.CreateICmp(P, LHS: X, RHS: B); |
| 293 | if ((IsLess && IsMin) || (IsGreater && IsMax)) |
| 294 | // x < min(a, b) -> x < a && x < b |
| 295 | // x > max(a, b) -> x > a && x > b |
| 296 | Replacement = Builder.CreateLogicalAnd(Cond1: LHS, Cond2: RHS); |
| 297 | else |
| 298 | // x > min(a, b) -> x > a || x > b |
| 299 | // x < max(a, b) -> x < a || x < b |
| 300 | Replacement = Builder.CreateLogicalOr(Cond1: LHS, Cond2: RHS); |
| 301 | |
| 302 | ICmp->replaceAllUsesWith(V: Replacement); |
| 303 | |
| 304 | Instruction *ToRemove[] = {ICmp, Info.ZExt, Info.SExt, MinMax}; |
| 305 | for (Instruction *I : ToRemove) |
| 306 | if (I && I->use_empty()) |
| 307 | I->eraseFromParent(); |
| 308 | |
| 309 | Changed = true; |
| 310 | } |
| 311 | |
| 312 | return Changed; |
| 313 | } |
| 314 | |
| 315 | // Do the following transformation: |
| 316 | // |
| 317 | // x < min(a, b) -> x < a && x < b |
| 318 | // x > min(a, b) -> x > a || x > b |
| 319 | // x < max(a, b) -> x < a || x < b |
| 320 | // x > max(a, b) -> x > a && x > b |
| 321 | // |
| 322 | // Such patterns are introduced by LICM.cpp:hoistMinMax() |
| 323 | // transformation and might lead to BPF verification failures for |
| 324 | // older kernels. |
| 325 | // |
| 326 | // To minimize "collateral" changes only do it for icmp + min/max |
| 327 | // calls when icmp is inside a loop and min/max is outside of that |
| 328 | // loop. |
| 329 | // |
| 330 | // Verification failure happens when: |
| 331 | // - RHS operand of some `icmp LHS, RHS` is replaced by some RHS1; |
| 332 | // - verifier can recognize RHS as a constant scalar in some context; |
| 333 | // - verifier can't recognize RHS1 as a constant scalar in the same |
| 334 | // context; |
| 335 | // |
| 336 | // The "constant scalar" is not a compile time constant, but a register |
| 337 | // that holds a scalar value known to verifier at some point in time |
| 338 | // during abstract interpretation. |
| 339 | // |
| 340 | // See also: |
| 341 | // https://lore.kernel.org/bpf/20230406164505.1046801-1-yhs@fb.com/ |
| 342 | bool BPFCheckAndAdjustIR::sinkMinMax(Module &M) { |
| 343 | bool Changed = false; |
| 344 | |
| 345 | for (Function &F : M) { |
| 346 | if (F.isDeclaration()) |
| 347 | continue; |
| 348 | |
| 349 | LoopInfo &LI = getAnalysis<LoopInfoWrapperPass>(F).getLoopInfo(); |
| 350 | for (Loop *L : LI) |
| 351 | for (BasicBlock *BB : L->blocks()) { |
| 352 | // Filter out instructions coming from the same loop |
| 353 | Loop *BBLoop = LI.getLoopFor(BB); |
| 354 | auto OtherLoopFilter = [&](Instruction *I) { |
| 355 | return LI.getLoopFor(BB: I->getParent()) != BBLoop; |
| 356 | }; |
| 357 | Changed |= sinkMinMaxInBB(BB&: *BB, Filter: OtherLoopFilter); |
| 358 | } |
| 359 | } |
| 360 | |
| 361 | return Changed; |
| 362 | } |
| 363 | |
| 364 | void BPFCheckAndAdjustIR::getAnalysisUsage(AnalysisUsage &AU) const { |
| 365 | AU.addRequired<LoopInfoWrapperPass>(); |
| 366 | } |
| 367 | |
| 368 | static void unrollGEPLoad(CallInst *Call) { |
| 369 | auto [GEP, Load] = BPFPreserveStaticOffsetPass::reconstructLoad(Call); |
| 370 | GEP->insertBefore(InsertPos: Call->getIterator()); |
| 371 | Load->insertBefore(InsertPos: Call->getIterator()); |
| 372 | Call->replaceAllUsesWith(V: Load); |
| 373 | Call->eraseFromParent(); |
| 374 | } |
| 375 | |
| 376 | static void unrollGEPStore(CallInst *Call) { |
| 377 | auto [GEP, Store] = BPFPreserveStaticOffsetPass::reconstructStore(Call); |
| 378 | GEP->insertBefore(InsertPos: Call->getIterator()); |
| 379 | Store->insertBefore(InsertPos: Call->getIterator()); |
| 380 | Call->eraseFromParent(); |
| 381 | } |
| 382 | |
| 383 | static bool removeGEPBuiltinsInFunc(Function &F) { |
| 384 | SmallVector<CallInst *> GEPLoads; |
| 385 | SmallVector<CallInst *> GEPStores; |
| 386 | for (auto &BB : F) |
| 387 | for (auto &Insn : BB) |
| 388 | if (auto *Call = dyn_cast<CallInst>(Val: &Insn)) |
| 389 | if (auto *Called = Call->getCalledFunction()) |
| 390 | switch (Called->getIntrinsicID()) { |
| 391 | case Intrinsic::bpf_getelementptr_and_load: |
| 392 | GEPLoads.push_back(Elt: Call); |
| 393 | break; |
| 394 | case Intrinsic::bpf_getelementptr_and_store: |
| 395 | GEPStores.push_back(Elt: Call); |
| 396 | break; |
| 397 | } |
| 398 | |
| 399 | if (GEPLoads.empty() && GEPStores.empty()) |
| 400 | return false; |
| 401 | |
| 402 | for_each(Range&: GEPLoads, F: unrollGEPLoad); |
| 403 | for_each(Range&: GEPStores, F: unrollGEPStore); |
| 404 | |
| 405 | return true; |
| 406 | } |
| 407 | |
| 408 | // Rewrites the following builtins: |
| 409 | // - llvm.bpf.getelementptr.and.load |
| 410 | // - llvm.bpf.getelementptr.and.store |
| 411 | // As (load (getelementptr ...)) or (store (getelementptr ...)). |
| 412 | bool BPFCheckAndAdjustIR::removeGEPBuiltins(Module &M) { |
| 413 | bool Changed = false; |
| 414 | for (auto &F : M) |
| 415 | Changed = removeGEPBuiltinsInFunc(F) || Changed; |
| 416 | return Changed; |
| 417 | } |
| 418 | |
| 419 | // Wrap ToWrap with cast to address space zero: |
| 420 | // - if ToWrap is a getelementptr, |
| 421 | // wrap it's base pointer instead and return a copy; |
| 422 | // - if ToWrap is Instruction, insert address space cast |
| 423 | // immediately after ToWrap; |
| 424 | // - if ToWrap is not an Instruction (function parameter |
| 425 | // or a global value), insert address space cast at the |
| 426 | // beginning of the Function F; |
| 427 | // - use Cache to avoid inserting too many casts; |
| 428 | static Value *aspaceWrapValue(DenseMap<Value *, Value *> &Cache, Function *F, |
| 429 | Value *ToWrap) { |
| 430 | auto It = Cache.find(Val: ToWrap); |
| 431 | if (It != Cache.end()) |
| 432 | return It->getSecond(); |
| 433 | |
| 434 | if (auto *GEP = dyn_cast<GetElementPtrInst>(Val: ToWrap)) { |
| 435 | Value *Ptr = GEP->getPointerOperand(); |
| 436 | Value *WrappedPtr = aspaceWrapValue(Cache, F, ToWrap: Ptr); |
| 437 | auto *GEPTy = cast<PointerType>(Val: GEP->getType()); |
| 438 | auto *NewGEP = GEP->clone(); |
| 439 | NewGEP->insertAfter(InsertPos: GEP->getIterator()); |
| 440 | NewGEP->mutateType(Ty: PointerType::getUnqual(C&: GEPTy->getContext())); |
| 441 | NewGEP->setOperand(i: GEP->getPointerOperandIndex(), Val: WrappedPtr); |
| 442 | NewGEP->setName(GEP->getName()); |
| 443 | Cache[ToWrap] = NewGEP; |
| 444 | return NewGEP; |
| 445 | } |
| 446 | |
| 447 | IRBuilder IB(F->getContext()); |
| 448 | if (Instruction *InsnPtr = dyn_cast<Instruction>(Val: ToWrap)) |
| 449 | IB.SetInsertPoint(*InsnPtr->getInsertionPointAfterDef()); |
| 450 | else |
| 451 | IB.SetInsertPoint(F->getEntryBlock().getFirstInsertionPt()); |
| 452 | auto *ASZeroPtrTy = IB.getPtrTy(AddrSpace: 0); |
| 453 | auto *ACast = IB.CreateAddrSpaceCast(V: ToWrap, DestTy: ASZeroPtrTy, Name: ToWrap->getName()); |
| 454 | Cache[ToWrap] = ACast; |
| 455 | return ACast; |
| 456 | } |
| 457 | |
| 458 | // Wrap a pointer operand OpNum of instruction I |
| 459 | // with cast to address space zero |
| 460 | static void aspaceWrapOperand(DenseMap<Value *, Value *> &Cache, Instruction *I, |
| 461 | unsigned OpNum) { |
| 462 | Value *OldOp = I->getOperand(i: OpNum); |
| 463 | if (OldOp->getType()->getPointerAddressSpace() == 0) |
| 464 | return; |
| 465 | |
| 466 | Value *NewOp = aspaceWrapValue(Cache, F: I->getFunction(), ToWrap: OldOp); |
| 467 | I->setOperand(i: OpNum, Val: NewOp); |
| 468 | // Check if there are any remaining users of old GEP, |
| 469 | // delete those w/o users |
| 470 | for (;;) { |
| 471 | auto *OldGEP = dyn_cast<GetElementPtrInst>(Val: OldOp); |
| 472 | if (!OldGEP) |
| 473 | break; |
| 474 | if (!OldGEP->use_empty()) |
| 475 | break; |
| 476 | OldOp = OldGEP->getPointerOperand(); |
| 477 | OldGEP->eraseFromParent(); |
| 478 | } |
| 479 | } |
| 480 | |
| 481 | // Support for BPF address spaces: |
| 482 | // - for each function in the module M, update pointer operand of |
| 483 | // each memory access instruction (load/store/cmpxchg/atomicrmw) |
| 484 | // by casting it from non-zero address space to zero address space, e.g: |
| 485 | // |
| 486 | // (load (ptr addrspace (N) %p) ...) |
| 487 | // -> (load (addrspacecast ptr addrspace (N) %p to ptr)) |
| 488 | // |
| 489 | // - assign section with name .addr_space.N for globals defined in |
| 490 | // non-zero address space N |
| 491 | bool BPFCheckAndAdjustIR::insertASpaceCasts(Module &M) { |
| 492 | bool Changed = false; |
| 493 | for (Function &F : M) { |
| 494 | DenseMap<Value *, Value *> CastsCache; |
| 495 | for (BasicBlock &BB : F) { |
| 496 | for (Instruction &I : BB) { |
| 497 | unsigned PtrOpNum; |
| 498 | |
| 499 | if (auto *LD = dyn_cast<LoadInst>(Val: &I)) |
| 500 | PtrOpNum = LD->getPointerOperandIndex(); |
| 501 | else if (auto *ST = dyn_cast<StoreInst>(Val: &I)) |
| 502 | PtrOpNum = ST->getPointerOperandIndex(); |
| 503 | else if (auto *CmpXchg = dyn_cast<AtomicCmpXchgInst>(Val: &I)) |
| 504 | PtrOpNum = CmpXchg->getPointerOperandIndex(); |
| 505 | else if (auto *RMW = dyn_cast<AtomicRMWInst>(Val: &I)) |
| 506 | PtrOpNum = RMW->getPointerOperandIndex(); |
| 507 | else |
| 508 | continue; |
| 509 | |
| 510 | aspaceWrapOperand(Cache&: CastsCache, I: &I, OpNum: PtrOpNum); |
| 511 | } |
| 512 | } |
| 513 | Changed |= !CastsCache.empty(); |
| 514 | } |
| 515 | // Merge all globals within same address space into single |
| 516 | // .addr_space.<addr space no> section |
| 517 | for (GlobalVariable &G : M.globals()) { |
| 518 | if (G.getAddressSpace() == 0 || G.hasSection()) |
| 519 | continue; |
| 520 | SmallString<16> SecName; |
| 521 | raw_svector_ostream OS(SecName); |
| 522 | OS << ".addr_space." << G.getAddressSpace(); |
| 523 | G.setSection(SecName); |
| 524 | // Prevent having separate section for constants |
| 525 | G.setConstant(false); |
| 526 | } |
| 527 | return Changed; |
| 528 | } |
| 529 | |
| 530 | bool BPFCheckAndAdjustIR::adjustIR(Module &M) { |
| 531 | bool Changed = removePassThroughBuiltin(M); |
| 532 | Changed = removeCompareBuiltin(M) || Changed; |
| 533 | Changed = sinkMinMax(M) || Changed; |
| 534 | Changed = removeGEPBuiltins(M) || Changed; |
| 535 | Changed = insertASpaceCasts(M) || Changed; |
| 536 | return Changed; |
| 537 | } |
| 538 | |
| 539 | bool BPFCheckAndAdjustIR::runOnModule(Module &M) { |
| 540 | checkIR(M); |
| 541 | return adjustIR(M); |
| 542 | } |
| 543 | |