1 | //===-- AMDGPURegBankLegalizeHelper.cpp -----------------------------------===// |
2 | // |
3 | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
4 | // See https://llvm.org/LICENSE.txt for license information. |
5 | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
6 | // |
7 | //===----------------------------------------------------------------------===// |
8 | // |
9 | /// Implements actual lowering algorithms for each ID that can be used in |
10 | /// Rule.OperandMapping. Similar to legalizer helper but with register banks. |
11 | // |
12 | //===----------------------------------------------------------------------===// |
13 | |
14 | #include "AMDGPURegBankLegalizeHelper.h" |
15 | #include "AMDGPUGlobalISelUtils.h" |
16 | #include "AMDGPUInstrInfo.h" |
17 | #include "AMDGPURegBankLegalizeRules.h" |
18 | #include "AMDGPURegisterBankInfo.h" |
19 | #include "GCNSubtarget.h" |
20 | #include "MCTargetDesc/AMDGPUMCTargetDesc.h" |
21 | #include "llvm/CodeGen/GlobalISel/GenericMachineInstrs.h" |
22 | #include "llvm/CodeGen/GlobalISel/MachineIRBuilder.h" |
23 | #include "llvm/CodeGen/MachineInstr.h" |
24 | #include "llvm/CodeGen/MachineUniformityAnalysis.h" |
25 | #include "llvm/IR/IntrinsicsAMDGPU.h" |
26 | #include "llvm/Support/AMDGPUAddrSpace.h" |
27 | |
28 | #define DEBUG_TYPE "amdgpu-regbanklegalize" |
29 | |
30 | using namespace llvm; |
31 | using namespace AMDGPU; |
32 | |
33 | RegBankLegalizeHelper::RegBankLegalizeHelper( |
34 | MachineIRBuilder &B, const MachineUniformityInfo &MUI, |
35 | const RegisterBankInfo &RBI, const RegBankLegalizeRules &RBLRules) |
36 | : ST(B.getMF().getSubtarget<GCNSubtarget>()), B(B), MRI(*B.getMRI()), |
37 | MUI(MUI), RBI(RBI), RBLRules(RBLRules), |
38 | SgprRB(&RBI.getRegBank(ID: AMDGPU::SGPRRegBankID)), |
39 | VgprRB(&RBI.getRegBank(ID: AMDGPU::VGPRRegBankID)), |
40 | VccRB(&RBI.getRegBank(ID: AMDGPU::VCCRegBankID)) {} |
41 | |
42 | void RegBankLegalizeHelper::findRuleAndApplyMapping(MachineInstr &MI) { |
43 | const SetOfRulesForOpcode &RuleSet = RBLRules.getRulesForOpc(MI); |
44 | const RegBankLLTMapping &Mapping = RuleSet.findMappingForMI(MI, MRI, MUI); |
45 | |
46 | SmallSet<Register, 4> WaterfallSgprs; |
47 | unsigned OpIdx = 0; |
48 | if (Mapping.DstOpMapping.size() > 0) { |
49 | B.setInsertPt(MBB&: *MI.getParent(), II: std::next(x: MI.getIterator())); |
50 | applyMappingDst(MI, OpIdx, MethodIDs: Mapping.DstOpMapping); |
51 | } |
52 | if (Mapping.SrcOpMapping.size() > 0) { |
53 | B.setInstr(MI); |
54 | applyMappingSrc(MI, OpIdx, MethodIDs: Mapping.SrcOpMapping, SgprWaterfallOperandRegs&: WaterfallSgprs); |
55 | } |
56 | |
57 | lower(MI, Mapping, SgprWaterfallOperandRegs&: WaterfallSgprs); |
58 | } |
59 | |
60 | void RegBankLegalizeHelper::splitLoad(MachineInstr &MI, |
61 | ArrayRef<LLT> LLTBreakdown, LLT MergeTy) { |
62 | MachineFunction &MF = B.getMF(); |
63 | assert(MI.getNumMemOperands() == 1); |
64 | MachineMemOperand &BaseMMO = **MI.memoperands_begin(); |
65 | Register Dst = MI.getOperand(i: 0).getReg(); |
66 | const RegisterBank *DstRB = MRI.getRegBankOrNull(Reg: Dst); |
67 | Register Base = MI.getOperand(i: 1).getReg(); |
68 | LLT PtrTy = MRI.getType(Reg: Base); |
69 | const RegisterBank *PtrRB = MRI.getRegBankOrNull(Reg: Base); |
70 | LLT OffsetTy = LLT::scalar(SizeInBits: PtrTy.getSizeInBits()); |
71 | SmallVector<Register, 4> LoadPartRegs; |
72 | |
73 | unsigned ByteOffset = 0; |
74 | for (LLT PartTy : LLTBreakdown) { |
75 | Register BasePlusOffset; |
76 | if (ByteOffset == 0) { |
77 | BasePlusOffset = Base; |
78 | } else { |
79 | auto Offset = B.buildConstant(Res: {PtrRB, OffsetTy}, Val: ByteOffset); |
80 | BasePlusOffset = B.buildPtrAdd(Res: {PtrRB, PtrTy}, Op0: Base, Op1: Offset).getReg(Idx: 0); |
81 | } |
82 | auto *OffsetMMO = MF.getMachineMemOperand(MMO: &BaseMMO, Offset: ByteOffset, Ty: PartTy); |
83 | auto LoadPart = B.buildLoad(Res: {DstRB, PartTy}, Addr: BasePlusOffset, MMO&: *OffsetMMO); |
84 | LoadPartRegs.push_back(Elt: LoadPart.getReg(Idx: 0)); |
85 | ByteOffset += PartTy.getSizeInBytes(); |
86 | } |
87 | |
88 | if (!MergeTy.isValid()) { |
89 | // Loads are of same size, concat or merge them together. |
90 | B.buildMergeLikeInstr(Res: Dst, Ops: LoadPartRegs); |
91 | } else { |
92 | // Loads are not all of same size, need to unmerge them to smaller pieces |
93 | // of MergeTy type, then merge pieces to Dst. |
94 | SmallVector<Register, 4> MergeTyParts; |
95 | for (Register Reg : LoadPartRegs) { |
96 | if (MRI.getType(Reg) == MergeTy) { |
97 | MergeTyParts.push_back(Elt: Reg); |
98 | } else { |
99 | auto Unmerge = B.buildUnmerge(Attrs: {.RCOrRB: DstRB, .Ty: MergeTy}, Op: Reg); |
100 | for (unsigned i = 0; i < Unmerge->getNumOperands() - 1; ++i) |
101 | MergeTyParts.push_back(Elt: Unmerge.getReg(Idx: i)); |
102 | } |
103 | } |
104 | B.buildMergeLikeInstr(Res: Dst, Ops: MergeTyParts); |
105 | } |
106 | MI.eraseFromParent(); |
107 | } |
108 | |
109 | void RegBankLegalizeHelper::widenLoad(MachineInstr &MI, LLT WideTy, |
110 | LLT MergeTy) { |
111 | MachineFunction &MF = B.getMF(); |
112 | assert(MI.getNumMemOperands() == 1); |
113 | MachineMemOperand &BaseMMO = **MI.memoperands_begin(); |
114 | Register Dst = MI.getOperand(i: 0).getReg(); |
115 | const RegisterBank *DstRB = MRI.getRegBankOrNull(Reg: Dst); |
116 | Register Base = MI.getOperand(i: 1).getReg(); |
117 | |
118 | MachineMemOperand *WideMMO = MF.getMachineMemOperand(MMO: &BaseMMO, Offset: 0, Ty: WideTy); |
119 | auto WideLoad = B.buildLoad(Res: {DstRB, WideTy}, Addr: Base, MMO&: *WideMMO); |
120 | |
121 | if (WideTy.isScalar()) { |
122 | B.buildTrunc(Res: Dst, Op: WideLoad); |
123 | } else { |
124 | SmallVector<Register, 4> MergeTyParts; |
125 | auto Unmerge = B.buildUnmerge(Attrs: {.RCOrRB: DstRB, .Ty: MergeTy}, Op: WideLoad); |
126 | |
127 | LLT DstTy = MRI.getType(Reg: Dst); |
128 | unsigned NumElts = DstTy.getSizeInBits() / MergeTy.getSizeInBits(); |
129 | for (unsigned i = 0; i < NumElts; ++i) { |
130 | MergeTyParts.push_back(Elt: Unmerge.getReg(Idx: i)); |
131 | } |
132 | B.buildMergeLikeInstr(Res: Dst, Ops: MergeTyParts); |
133 | } |
134 | MI.eraseFromParent(); |
135 | } |
136 | |
137 | void RegBankLegalizeHelper::lowerVccExtToSel(MachineInstr &MI) { |
138 | Register Dst = MI.getOperand(i: 0).getReg(); |
139 | LLT Ty = MRI.getType(Reg: Dst); |
140 | Register Src = MI.getOperand(i: 1).getReg(); |
141 | unsigned Opc = MI.getOpcode(); |
142 | int TrueExtCst = Opc == G_SEXT ? -1 : 1; |
143 | if (Ty == S32 || Ty == S16) { |
144 | auto True = B.buildConstant(Res: {VgprRB, Ty}, Val: TrueExtCst); |
145 | auto False = B.buildConstant(Res: {VgprRB, Ty}, Val: 0); |
146 | B.buildSelect(Res: Dst, Tst: Src, Op0: True, Op1: False); |
147 | } else if (Ty == S64) { |
148 | auto True = B.buildConstant(Res: {VgprRB_S32}, Val: TrueExtCst); |
149 | auto False = B.buildConstant(Res: {VgprRB_S32}, Val: 0); |
150 | auto Lo = B.buildSelect(Res: {VgprRB_S32}, Tst: Src, Op0: True, Op1: False); |
151 | MachineInstrBuilder Hi; |
152 | switch (Opc) { |
153 | case G_SEXT: |
154 | Hi = Lo; |
155 | break; |
156 | case G_ZEXT: |
157 | Hi = False; |
158 | break; |
159 | case G_ANYEXT: |
160 | Hi = B.buildUndef(Res: {VgprRB_S32}); |
161 | break; |
162 | default: |
163 | llvm_unreachable("Opcode not supported" ); |
164 | } |
165 | |
166 | B.buildMergeValues(Res: Dst, Ops: {Lo.getReg(Idx: 0), Hi.getReg(Idx: 0)}); |
167 | } else { |
168 | llvm_unreachable("Type not supported" ); |
169 | } |
170 | |
171 | MI.eraseFromParent(); |
172 | } |
173 | |
174 | std::pair<Register, Register> RegBankLegalizeHelper::unpackZExt(Register Reg) { |
175 | auto PackedS32 = B.buildBitcast(Dst: SgprRB_S32, Src: Reg); |
176 | auto Mask = B.buildConstant(Res: SgprRB_S32, Val: 0x0000ffff); |
177 | auto Lo = B.buildAnd(Dst: SgprRB_S32, Src0: PackedS32, Src1: Mask); |
178 | auto Hi = B.buildLShr(Dst: SgprRB_S32, Src0: PackedS32, Src1: B.buildConstant(Res: SgprRB_S32, Val: 16)); |
179 | return {Lo.getReg(Idx: 0), Hi.getReg(Idx: 0)}; |
180 | } |
181 | |
182 | std::pair<Register, Register> RegBankLegalizeHelper::unpackSExt(Register Reg) { |
183 | auto PackedS32 = B.buildBitcast(Dst: SgprRB_S32, Src: Reg); |
184 | auto Lo = B.buildSExtInReg(Res: SgprRB_S32, Op: PackedS32, ImmOp: 16); |
185 | auto Hi = B.buildAShr(Dst: SgprRB_S32, Src0: PackedS32, Src1: B.buildConstant(Res: SgprRB_S32, Val: 16)); |
186 | return {Lo.getReg(Idx: 0), Hi.getReg(Idx: 0)}; |
187 | } |
188 | |
189 | std::pair<Register, Register> RegBankLegalizeHelper::unpackAExt(Register Reg) { |
190 | auto PackedS32 = B.buildBitcast(Dst: SgprRB_S32, Src: Reg); |
191 | auto Lo = PackedS32; |
192 | auto Hi = B.buildLShr(Dst: SgprRB_S32, Src0: PackedS32, Src1: B.buildConstant(Res: SgprRB_S32, Val: 16)); |
193 | return {Lo.getReg(Idx: 0), Hi.getReg(Idx: 0)}; |
194 | } |
195 | |
196 | void RegBankLegalizeHelper::lowerUnpackBitShift(MachineInstr &MI) { |
197 | Register Lo, Hi; |
198 | switch (MI.getOpcode()) { |
199 | case AMDGPU::G_SHL: { |
200 | auto [Val0, Val1] = unpackAExt(Reg: MI.getOperand(i: 1).getReg()); |
201 | auto [Amt0, Amt1] = unpackAExt(Reg: MI.getOperand(i: 2).getReg()); |
202 | Lo = B.buildInstr(Opc: MI.getOpcode(), DstOps: {SgprRB_S32}, SrcOps: {Val0, Amt0}).getReg(Idx: 0); |
203 | Hi = B.buildInstr(Opc: MI.getOpcode(), DstOps: {SgprRB_S32}, SrcOps: {Val1, Amt1}).getReg(Idx: 0); |
204 | break; |
205 | } |
206 | case AMDGPU::G_LSHR: { |
207 | auto [Val0, Val1] = unpackZExt(Reg: MI.getOperand(i: 1).getReg()); |
208 | auto [Amt0, Amt1] = unpackZExt(Reg: MI.getOperand(i: 2).getReg()); |
209 | Lo = B.buildInstr(Opc: MI.getOpcode(), DstOps: {SgprRB_S32}, SrcOps: {Val0, Amt0}).getReg(Idx: 0); |
210 | Hi = B.buildInstr(Opc: MI.getOpcode(), DstOps: {SgprRB_S32}, SrcOps: {Val1, Amt1}).getReg(Idx: 0); |
211 | break; |
212 | } |
213 | case AMDGPU::G_ASHR: { |
214 | auto [Val0, Val1] = unpackSExt(Reg: MI.getOperand(i: 1).getReg()); |
215 | auto [Amt0, Amt1] = unpackSExt(Reg: MI.getOperand(i: 2).getReg()); |
216 | Lo = B.buildAShr(Dst: SgprRB_S32, Src0: Val0, Src1: Amt0).getReg(Idx: 0); |
217 | Hi = B.buildAShr(Dst: SgprRB_S32, Src0: Val1, Src1: Amt1).getReg(Idx: 0); |
218 | break; |
219 | } |
220 | default: |
221 | llvm_unreachable("Unpack lowering not implemented" ); |
222 | } |
223 | B.buildBuildVectorTrunc(Res: MI.getOperand(i: 0).getReg(), Ops: {Lo, Hi}); |
224 | MI.eraseFromParent(); |
225 | } |
226 | |
227 | static bool isSignedBFE(MachineInstr &MI) { |
228 | if (GIntrinsic *GI = dyn_cast<GIntrinsic>(Val: &MI)) |
229 | return (GI->is(ID: Intrinsic::amdgcn_sbfe)); |
230 | |
231 | return MI.getOpcode() == AMDGPU::G_SBFX; |
232 | } |
233 | |
234 | void RegBankLegalizeHelper::lowerV_BFE(MachineInstr &MI) { |
235 | Register Dst = MI.getOperand(i: 0).getReg(); |
236 | assert(MRI.getType(Dst) == LLT::scalar(64)); |
237 | bool Signed = isSignedBFE(MI); |
238 | unsigned FirstOpnd = isa<GIntrinsic>(Val: MI) ? 2 : 1; |
239 | // Extract bitfield from Src, LSBit is the least-significant bit for the |
240 | // extraction (field offset) and Width is size of bitfield. |
241 | Register Src = MI.getOperand(i: FirstOpnd).getReg(); |
242 | Register LSBit = MI.getOperand(i: FirstOpnd + 1).getReg(); |
243 | Register Width = MI.getOperand(i: FirstOpnd + 2).getReg(); |
244 | // Comments are for signed bitfield extract, similar for unsigned. x is sign |
245 | // bit. s is sign, l is LSB and y are remaining bits of bitfield to extract. |
246 | |
247 | // Src >> LSBit Hi|Lo: x?????syyyyyyl??? -> xxxx?????syyyyyyl |
248 | unsigned SHROpc = Signed ? AMDGPU::G_ASHR : AMDGPU::G_LSHR; |
249 | auto SHRSrc = B.buildInstr(Opc: SHROpc, DstOps: {{VgprRB, S64}}, SrcOps: {Src, LSBit}); |
250 | |
251 | auto ConstWidth = getIConstantVRegValWithLookThrough(VReg: Width, MRI); |
252 | |
253 | // Expand to Src >> LSBit << (64 - Width) >> (64 - Width) |
254 | // << (64 - Width): Hi|Lo: xxxx?????syyyyyyl -> syyyyyyl000000000 |
255 | // >> (64 - Width): Hi|Lo: syyyyyyl000000000 -> ssssssssssyyyyyyl |
256 | if (!ConstWidth) { |
257 | auto Amt = B.buildSub(Dst: VgprRB_S32, Src0: B.buildConstant(Res: SgprRB_S32, Val: 64), Src1: Width); |
258 | auto SignBit = B.buildShl(Dst: {VgprRB, S64}, Src0: SHRSrc, Src1: Amt); |
259 | B.buildInstr(Opc: SHROpc, DstOps: {Dst}, SrcOps: {SignBit, Amt}); |
260 | MI.eraseFromParent(); |
261 | return; |
262 | } |
263 | |
264 | uint64_t WidthImm = ConstWidth->Value.getZExtValue(); |
265 | auto UnmergeSHRSrc = B.buildUnmerge(Attrs: VgprRB_S32, Op: SHRSrc); |
266 | Register SHRSrcLo = UnmergeSHRSrc.getReg(Idx: 0); |
267 | Register SHRSrcHi = UnmergeSHRSrc.getReg(Idx: 1); |
268 | auto Zero = B.buildConstant(Res: {VgprRB, S32}, Val: 0); |
269 | unsigned BFXOpc = Signed ? AMDGPU::G_SBFX : AMDGPU::G_UBFX; |
270 | |
271 | if (WidthImm <= 32) { |
272 | // SHRSrc Hi|Lo: ????????|???syyyl -> ????????|ssssyyyl |
273 | auto Lo = B.buildInstr(Opc: BFXOpc, DstOps: {VgprRB_S32}, SrcOps: {SHRSrcLo, Zero, Width}); |
274 | MachineInstrBuilder Hi; |
275 | if (Signed) { |
276 | // SHRSrc Hi|Lo: ????????|ssssyyyl -> ssssssss|ssssyyyl |
277 | Hi = B.buildAShr(Dst: VgprRB_S32, Src0: Lo, Src1: B.buildConstant(Res: VgprRB_S32, Val: 31)); |
278 | } else { |
279 | // SHRSrc Hi|Lo: ????????|000syyyl -> 00000000|000syyyl |
280 | Hi = Zero; |
281 | } |
282 | B.buildMergeLikeInstr(Res: Dst, Ops: {Lo, Hi}); |
283 | } else { |
284 | auto Amt = B.buildConstant(Res: VgprRB_S32, Val: WidthImm - 32); |
285 | // SHRSrc Hi|Lo: ??????sy|yyyyyyyl -> sssssssy|yyyyyyyl |
286 | auto Hi = B.buildInstr(Opc: BFXOpc, DstOps: {VgprRB_S32}, SrcOps: {SHRSrcHi, Zero, Amt}); |
287 | B.buildMergeLikeInstr(Res: Dst, Ops: {SHRSrcLo, Hi}); |
288 | } |
289 | |
290 | MI.eraseFromParent(); |
291 | } |
292 | |
293 | void RegBankLegalizeHelper::lowerS_BFE(MachineInstr &MI) { |
294 | Register DstReg = MI.getOperand(i: 0).getReg(); |
295 | LLT Ty = MRI.getType(Reg: DstReg); |
296 | bool Signed = isSignedBFE(MI); |
297 | unsigned FirstOpnd = isa<GIntrinsic>(Val: MI) ? 2 : 1; |
298 | Register Src = MI.getOperand(i: FirstOpnd).getReg(); |
299 | Register LSBit = MI.getOperand(i: FirstOpnd + 1).getReg(); |
300 | Register Width = MI.getOperand(i: FirstOpnd + 2).getReg(); |
301 | // For uniform bit field extract there are 4 available instructions, but |
302 | // LSBit(field offset) and Width(size of bitfield) need to be packed in S32, |
303 | // field offset in low and size in high 16 bits. |
304 | |
305 | // Src1 Hi16|Lo16 = Size|FieldOffset |
306 | auto Mask = B.buildConstant(Res: SgprRB_S32, Val: maskTrailingOnes<unsigned>(N: 6)); |
307 | auto FieldOffset = B.buildAnd(Dst: SgprRB_S32, Src0: LSBit, Src1: Mask); |
308 | auto Size = B.buildShl(Dst: SgprRB_S32, Src0: Width, Src1: B.buildConstant(Res: SgprRB_S32, Val: 16)); |
309 | auto Src1 = B.buildOr(Dst: SgprRB_S32, Src0: FieldOffset, Src1: Size); |
310 | unsigned Opc32 = Signed ? AMDGPU::S_BFE_I32 : AMDGPU::S_BFE_U32; |
311 | unsigned Opc64 = Signed ? AMDGPU::S_BFE_I64 : AMDGPU::S_BFE_U64; |
312 | unsigned Opc = Ty == S32 ? Opc32 : Opc64; |
313 | |
314 | // Select machine instruction, because of reg class constraining, insert |
315 | // copies from reg class to reg bank. |
316 | auto S_BFE = B.buildInstr(Opc, DstOps: {{SgprRB, Ty}}, |
317 | SrcOps: {B.buildCopy(Res: Ty, Op: Src), B.buildCopy(Res: S32, Op: Src1)}); |
318 | if (!constrainSelectedInstRegOperands(I&: *S_BFE, TII: *ST.getInstrInfo(), |
319 | TRI: *ST.getRegisterInfo(), RBI)) |
320 | llvm_unreachable("failed to constrain BFE" ); |
321 | |
322 | B.buildCopy(Res: DstReg, Op: S_BFE->getOperand(i: 0).getReg()); |
323 | MI.eraseFromParent(); |
324 | } |
325 | |
326 | void RegBankLegalizeHelper::lowerSplitTo32(MachineInstr &MI) { |
327 | Register Dst = MI.getOperand(i: 0).getReg(); |
328 | LLT DstTy = MRI.getType(Reg: Dst); |
329 | assert(DstTy == V4S16 || DstTy == V2S32 || DstTy == S64); |
330 | LLT Ty = DstTy == V4S16 ? V2S16 : S32; |
331 | auto Op1 = B.buildUnmerge(Attrs: {.RCOrRB: VgprRB, .Ty: Ty}, Op: MI.getOperand(i: 1).getReg()); |
332 | auto Op2 = B.buildUnmerge(Attrs: {.RCOrRB: VgprRB, .Ty: Ty}, Op: MI.getOperand(i: 2).getReg()); |
333 | unsigned Opc = MI.getOpcode(); |
334 | auto Flags = MI.getFlags(); |
335 | auto Lo = |
336 | B.buildInstr(Opc, DstOps: {{VgprRB, Ty}}, SrcOps: {Op1.getReg(Idx: 0), Op2.getReg(Idx: 0)}, Flags); |
337 | auto Hi = |
338 | B.buildInstr(Opc, DstOps: {{VgprRB, Ty}}, SrcOps: {Op1.getReg(Idx: 1), Op2.getReg(Idx: 1)}, Flags); |
339 | B.buildMergeLikeInstr(Res: Dst, Ops: {Lo, Hi}); |
340 | MI.eraseFromParent(); |
341 | } |
342 | |
343 | void RegBankLegalizeHelper::lowerSplitTo32Select(MachineInstr &MI) { |
344 | Register Dst = MI.getOperand(i: 0).getReg(); |
345 | LLT DstTy = MRI.getType(Reg: Dst); |
346 | assert(DstTy == V4S16 || DstTy == V2S32 || DstTy == S64 || |
347 | (DstTy.isPointer() && DstTy.getSizeInBits() == 64)); |
348 | LLT Ty = DstTy == V4S16 ? V2S16 : S32; |
349 | auto Op2 = B.buildUnmerge(Attrs: {.RCOrRB: VgprRB, .Ty: Ty}, Op: MI.getOperand(i: 2).getReg()); |
350 | auto Op3 = B.buildUnmerge(Attrs: {.RCOrRB: VgprRB, .Ty: Ty}, Op: MI.getOperand(i: 3).getReg()); |
351 | Register Cond = MI.getOperand(i: 1).getReg(); |
352 | auto Flags = MI.getFlags(); |
353 | auto Lo = |
354 | B.buildSelect(Res: {VgprRB, Ty}, Tst: Cond, Op0: Op2.getReg(Idx: 0), Op1: Op3.getReg(Idx: 0), Flags); |
355 | auto Hi = |
356 | B.buildSelect(Res: {VgprRB, Ty}, Tst: Cond, Op0: Op2.getReg(Idx: 1), Op1: Op3.getReg(Idx: 1), Flags); |
357 | |
358 | B.buildMergeLikeInstr(Res: Dst, Ops: {Lo, Hi}); |
359 | MI.eraseFromParent(); |
360 | } |
361 | |
362 | void RegBankLegalizeHelper::lowerSplitTo32SExtInReg(MachineInstr &MI) { |
363 | auto Op1 = B.buildUnmerge(Attrs: VgprRB_S32, Op: MI.getOperand(i: 1).getReg()); |
364 | int Amt = MI.getOperand(i: 2).getImm(); |
365 | Register Lo, Hi; |
366 | // Hi|Lo: s sign bit, ?/x bits changed/not changed by sign-extend |
367 | if (Amt <= 32) { |
368 | auto Freeze = B.buildFreeze(Dst: VgprRB_S32, Src: Op1.getReg(Idx: 0)); |
369 | if (Amt == 32) { |
370 | // Hi|Lo: ????????|sxxxxxxx -> ssssssss|sxxxxxxx |
371 | Lo = Freeze.getReg(Idx: 0); |
372 | } else { |
373 | // Hi|Lo: ????????|???sxxxx -> ssssssss|ssssxxxx |
374 | Lo = B.buildSExtInReg(Res: VgprRB_S32, Op: Freeze, ImmOp: Amt).getReg(Idx: 0); |
375 | } |
376 | |
377 | auto SignExtCst = B.buildConstant(Res: SgprRB_S32, Val: 31); |
378 | Hi = B.buildAShr(Dst: VgprRB_S32, Src0: Lo, Src1: SignExtCst).getReg(Idx: 0); |
379 | } else { |
380 | // Hi|Lo: ?????sxx|xxxxxxxx -> ssssssxx|xxxxxxxx |
381 | Lo = Op1.getReg(Idx: 0); |
382 | Hi = B.buildSExtInReg(Res: VgprRB_S32, Op: Op1.getReg(Idx: 1), ImmOp: Amt - 32).getReg(Idx: 0); |
383 | } |
384 | |
385 | B.buildMergeLikeInstr(Res: MI.getOperand(i: 0).getReg(), Ops: {Lo, Hi}); |
386 | MI.eraseFromParent(); |
387 | } |
388 | |
389 | void RegBankLegalizeHelper::lower(MachineInstr &MI, |
390 | const RegBankLLTMapping &Mapping, |
391 | SmallSet<Register, 4> &WaterfallSgprs) { |
392 | |
393 | switch (Mapping.LoweringMethod) { |
394 | case DoNotLower: |
395 | return; |
396 | case VccExtToSel: |
397 | return lowerVccExtToSel(MI); |
398 | case UniExtToSel: { |
399 | LLT Ty = MRI.getType(Reg: MI.getOperand(i: 0).getReg()); |
400 | auto True = B.buildConstant(Res: {SgprRB, Ty}, |
401 | Val: MI.getOpcode() == AMDGPU::G_SEXT ? -1 : 1); |
402 | auto False = B.buildConstant(Res: {SgprRB, Ty}, Val: 0); |
403 | // Input to G_{Z|S}EXT is 'Legalizer legal' S1. Most common case is compare. |
404 | // We are making select here. S1 cond was already 'any-extended to S32' + |
405 | // 'AND with 1 to clean high bits' by Sgpr32AExtBoolInReg. |
406 | B.buildSelect(Res: MI.getOperand(i: 0).getReg(), Tst: MI.getOperand(i: 1).getReg(), Op0: True, |
407 | Op1: False); |
408 | MI.eraseFromParent(); |
409 | return; |
410 | } |
411 | case UnpackBitShift: |
412 | return lowerUnpackBitShift(MI); |
413 | case Ext32To64: { |
414 | const RegisterBank *RB = MRI.getRegBank(Reg: MI.getOperand(i: 0).getReg()); |
415 | MachineInstrBuilder Hi; |
416 | switch (MI.getOpcode()) { |
417 | case AMDGPU::G_ZEXT: { |
418 | Hi = B.buildConstant(Res: {RB, S32}, Val: 0); |
419 | break; |
420 | } |
421 | case AMDGPU::G_SEXT: { |
422 | // Replicate sign bit from 32-bit extended part. |
423 | auto ShiftAmt = B.buildConstant(Res: {RB, S32}, Val: 31); |
424 | Hi = B.buildAShr(Dst: {RB, S32}, Src0: MI.getOperand(i: 1).getReg(), Src1: ShiftAmt); |
425 | break; |
426 | } |
427 | case AMDGPU::G_ANYEXT: { |
428 | Hi = B.buildUndef(Res: {RB, S32}); |
429 | break; |
430 | } |
431 | default: |
432 | llvm_unreachable("Unsuported Opcode in Ext32To64" ); |
433 | } |
434 | |
435 | B.buildMergeLikeInstr(Res: MI.getOperand(i: 0).getReg(), |
436 | Ops: {MI.getOperand(i: 1).getReg(), Hi}); |
437 | MI.eraseFromParent(); |
438 | return; |
439 | } |
440 | case UniCstExt: { |
441 | uint64_t ConstVal = MI.getOperand(i: 1).getCImm()->getZExtValue(); |
442 | B.buildConstant(Res: MI.getOperand(i: 0).getReg(), Val: ConstVal); |
443 | |
444 | MI.eraseFromParent(); |
445 | return; |
446 | } |
447 | case VgprToVccCopy: { |
448 | Register Src = MI.getOperand(i: 1).getReg(); |
449 | LLT Ty = MRI.getType(Reg: Src); |
450 | // Take lowest bit from each lane and put it in lane mask. |
451 | // Lowering via compare, but we need to clean high bits first as compare |
452 | // compares all bits in register. |
453 | Register BoolSrc = MRI.createVirtualRegister(RegAttr: {.RCOrRB: VgprRB, .Ty: Ty}); |
454 | if (Ty == S64) { |
455 | auto Src64 = B.buildUnmerge(Attrs: VgprRB_S32, Op: Src); |
456 | auto One = B.buildConstant(Res: VgprRB_S32, Val: 1); |
457 | auto AndLo = B.buildAnd(Dst: VgprRB_S32, Src0: Src64.getReg(Idx: 0), Src1: One); |
458 | auto Zero = B.buildConstant(Res: VgprRB_S32, Val: 0); |
459 | auto AndHi = B.buildAnd(Dst: VgprRB_S32, Src0: Src64.getReg(Idx: 1), Src1: Zero); |
460 | B.buildMergeLikeInstr(Res: BoolSrc, Ops: {AndLo, AndHi}); |
461 | } else { |
462 | assert(Ty == S32 || Ty == S16); |
463 | auto One = B.buildConstant(Res: {VgprRB, Ty}, Val: 1); |
464 | B.buildAnd(Dst: BoolSrc, Src0: Src, Src1: One); |
465 | } |
466 | auto Zero = B.buildConstant(Res: {VgprRB, Ty}, Val: 0); |
467 | B.buildICmp(Pred: CmpInst::ICMP_NE, Res: MI.getOperand(i: 0).getReg(), Op0: BoolSrc, Op1: Zero); |
468 | MI.eraseFromParent(); |
469 | return; |
470 | } |
471 | case V_BFE: |
472 | return lowerV_BFE(MI); |
473 | case S_BFE: |
474 | return lowerS_BFE(MI); |
475 | case SplitTo32: |
476 | return lowerSplitTo32(MI); |
477 | case SplitTo32Select: |
478 | return lowerSplitTo32Select(MI); |
479 | case SplitTo32SExtInReg: |
480 | return lowerSplitTo32SExtInReg(MI); |
481 | case SplitLoad: { |
482 | LLT DstTy = MRI.getType(Reg: MI.getOperand(i: 0).getReg()); |
483 | unsigned Size = DstTy.getSizeInBits(); |
484 | // Even split to 128-bit loads |
485 | if (Size > 128) { |
486 | LLT B128; |
487 | if (DstTy.isVector()) { |
488 | LLT EltTy = DstTy.getElementType(); |
489 | B128 = LLT::fixed_vector(NumElements: 128 / EltTy.getSizeInBits(), ScalarTy: EltTy); |
490 | } else { |
491 | B128 = LLT::scalar(SizeInBits: 128); |
492 | } |
493 | if (Size / 128 == 2) |
494 | splitLoad(MI, LLTBreakdown: {B128, B128}); |
495 | else if (Size / 128 == 4) |
496 | splitLoad(MI, LLTBreakdown: {B128, B128, B128, B128}); |
497 | else { |
498 | LLVM_DEBUG(dbgs() << "MI: " ; MI.dump();); |
499 | llvm_unreachable("SplitLoad type not supported for MI" ); |
500 | } |
501 | } |
502 | // 64 and 32 bit load |
503 | else if (DstTy == S96) |
504 | splitLoad(MI, LLTBreakdown: {S64, S32}, MergeTy: S32); |
505 | else if (DstTy == V3S32) |
506 | splitLoad(MI, LLTBreakdown: {V2S32, S32}, MergeTy: S32); |
507 | else if (DstTy == V6S16) |
508 | splitLoad(MI, LLTBreakdown: {V4S16, V2S16}, MergeTy: V2S16); |
509 | else { |
510 | LLVM_DEBUG(dbgs() << "MI: " ; MI.dump();); |
511 | llvm_unreachable("SplitLoad type not supported for MI" ); |
512 | } |
513 | break; |
514 | } |
515 | case WidenLoad: { |
516 | LLT DstTy = MRI.getType(Reg: MI.getOperand(i: 0).getReg()); |
517 | if (DstTy == S96) |
518 | widenLoad(MI, WideTy: S128); |
519 | else if (DstTy == V3S32) |
520 | widenLoad(MI, WideTy: V4S32, MergeTy: S32); |
521 | else if (DstTy == V6S16) |
522 | widenLoad(MI, WideTy: V8S16, MergeTy: V2S16); |
523 | else { |
524 | LLVM_DEBUG(dbgs() << "MI: " ; MI.dump();); |
525 | llvm_unreachable("WidenLoad type not supported for MI" ); |
526 | } |
527 | break; |
528 | } |
529 | } |
530 | |
531 | // TODO: executeInWaterfallLoop(... WaterfallSgprs) |
532 | } |
533 | |
534 | LLT RegBankLegalizeHelper::getTyFromID(RegBankLLTMappingApplyID ID) { |
535 | switch (ID) { |
536 | case Vcc: |
537 | case UniInVcc: |
538 | return LLT::scalar(SizeInBits: 1); |
539 | case Sgpr16: |
540 | case Vgpr16: |
541 | return LLT::scalar(SizeInBits: 16); |
542 | case Sgpr32: |
543 | case Sgpr32Trunc: |
544 | case Sgpr32AExt: |
545 | case Sgpr32AExtBoolInReg: |
546 | case Sgpr32SExt: |
547 | case Sgpr32ZExt: |
548 | case UniInVgprS32: |
549 | case Vgpr32: |
550 | case Vgpr32SExt: |
551 | case Vgpr32ZExt: |
552 | return LLT::scalar(SizeInBits: 32); |
553 | case Sgpr64: |
554 | case Vgpr64: |
555 | return LLT::scalar(SizeInBits: 64); |
556 | case Sgpr128: |
557 | case Vgpr128: |
558 | return LLT::scalar(SizeInBits: 128); |
559 | case VgprP0: |
560 | return LLT::pointer(AddressSpace: 0, SizeInBits: 64); |
561 | case SgprP1: |
562 | case VgprP1: |
563 | return LLT::pointer(AddressSpace: 1, SizeInBits: 64); |
564 | case SgprP3: |
565 | case VgprP3: |
566 | return LLT::pointer(AddressSpace: 3, SizeInBits: 32); |
567 | case SgprP4: |
568 | case VgprP4: |
569 | return LLT::pointer(AddressSpace: 4, SizeInBits: 64); |
570 | case SgprP5: |
571 | case VgprP5: |
572 | return LLT::pointer(AddressSpace: 5, SizeInBits: 32); |
573 | case SgprV2S16: |
574 | case VgprV2S16: |
575 | case UniInVgprV2S16: |
576 | return LLT::fixed_vector(NumElements: 2, ScalarSizeInBits: 16); |
577 | case SgprV2S32: |
578 | case VgprV2S32: |
579 | return LLT::fixed_vector(NumElements: 2, ScalarSizeInBits: 32); |
580 | case SgprV4S32: |
581 | case VgprV4S32: |
582 | case UniInVgprV4S32: |
583 | return LLT::fixed_vector(NumElements: 4, ScalarSizeInBits: 32); |
584 | default: |
585 | return LLT(); |
586 | } |
587 | } |
588 | |
589 | LLT RegBankLegalizeHelper::getBTyFromID(RegBankLLTMappingApplyID ID, LLT Ty) { |
590 | switch (ID) { |
591 | case SgprB32: |
592 | case VgprB32: |
593 | case UniInVgprB32: |
594 | if (Ty == LLT::scalar(SizeInBits: 32) || Ty == LLT::fixed_vector(NumElements: 2, ScalarSizeInBits: 16) || |
595 | isAnyPtr(Ty, Width: 32)) |
596 | return Ty; |
597 | return LLT(); |
598 | case SgprPtr32: |
599 | case VgprPtr32: |
600 | return isAnyPtr(Ty, Width: 32) ? Ty : LLT(); |
601 | case SgprPtr64: |
602 | case VgprPtr64: |
603 | return isAnyPtr(Ty, Width: 64) ? Ty : LLT(); |
604 | case SgprPtr128: |
605 | case VgprPtr128: |
606 | return isAnyPtr(Ty, Width: 128) ? Ty : LLT(); |
607 | case SgprB64: |
608 | case VgprB64: |
609 | case UniInVgprB64: |
610 | if (Ty == LLT::scalar(SizeInBits: 64) || Ty == LLT::fixed_vector(NumElements: 2, ScalarSizeInBits: 32) || |
611 | Ty == LLT::fixed_vector(NumElements: 4, ScalarSizeInBits: 16) || isAnyPtr(Ty, Width: 64)) |
612 | return Ty; |
613 | return LLT(); |
614 | case SgprB96: |
615 | case VgprB96: |
616 | case UniInVgprB96: |
617 | if (Ty == LLT::scalar(SizeInBits: 96) || Ty == LLT::fixed_vector(NumElements: 3, ScalarSizeInBits: 32) || |
618 | Ty == LLT::fixed_vector(NumElements: 6, ScalarSizeInBits: 16)) |
619 | return Ty; |
620 | return LLT(); |
621 | case SgprB128: |
622 | case VgprB128: |
623 | case UniInVgprB128: |
624 | if (Ty == LLT::scalar(SizeInBits: 128) || Ty == LLT::fixed_vector(NumElements: 4, ScalarSizeInBits: 32) || |
625 | Ty == LLT::fixed_vector(NumElements: 2, ScalarSizeInBits: 64) || isAnyPtr(Ty, Width: 128)) |
626 | return Ty; |
627 | return LLT(); |
628 | case SgprB256: |
629 | case VgprB256: |
630 | case UniInVgprB256: |
631 | if (Ty == LLT::scalar(SizeInBits: 256) || Ty == LLT::fixed_vector(NumElements: 8, ScalarSizeInBits: 32) || |
632 | Ty == LLT::fixed_vector(NumElements: 4, ScalarSizeInBits: 64) || Ty == LLT::fixed_vector(NumElements: 16, ScalarSizeInBits: 16)) |
633 | return Ty; |
634 | return LLT(); |
635 | case SgprB512: |
636 | case VgprB512: |
637 | case UniInVgprB512: |
638 | if (Ty == LLT::scalar(SizeInBits: 512) || Ty == LLT::fixed_vector(NumElements: 16, ScalarSizeInBits: 32) || |
639 | Ty == LLT::fixed_vector(NumElements: 8, ScalarSizeInBits: 64)) |
640 | return Ty; |
641 | return LLT(); |
642 | default: |
643 | return LLT(); |
644 | } |
645 | } |
646 | |
647 | const RegisterBank * |
648 | RegBankLegalizeHelper::getRegBankFromID(RegBankLLTMappingApplyID ID) { |
649 | switch (ID) { |
650 | case Vcc: |
651 | return VccRB; |
652 | case Sgpr16: |
653 | case Sgpr32: |
654 | case Sgpr64: |
655 | case Sgpr128: |
656 | case SgprP1: |
657 | case SgprP3: |
658 | case SgprP4: |
659 | case SgprP5: |
660 | case SgprPtr32: |
661 | case SgprPtr64: |
662 | case SgprPtr128: |
663 | case SgprV2S16: |
664 | case SgprV2S32: |
665 | case SgprV4S32: |
666 | case SgprB32: |
667 | case SgprB64: |
668 | case SgprB96: |
669 | case SgprB128: |
670 | case SgprB256: |
671 | case SgprB512: |
672 | case UniInVcc: |
673 | case UniInVgprS32: |
674 | case UniInVgprV2S16: |
675 | case UniInVgprV4S32: |
676 | case UniInVgprB32: |
677 | case UniInVgprB64: |
678 | case UniInVgprB96: |
679 | case UniInVgprB128: |
680 | case UniInVgprB256: |
681 | case UniInVgprB512: |
682 | case Sgpr32Trunc: |
683 | case Sgpr32AExt: |
684 | case Sgpr32AExtBoolInReg: |
685 | case Sgpr32SExt: |
686 | case Sgpr32ZExt: |
687 | return SgprRB; |
688 | case Vgpr16: |
689 | case Vgpr32: |
690 | case Vgpr64: |
691 | case Vgpr128: |
692 | case VgprP0: |
693 | case VgprP1: |
694 | case VgprP3: |
695 | case VgprP4: |
696 | case VgprP5: |
697 | case VgprPtr32: |
698 | case VgprPtr64: |
699 | case VgprPtr128: |
700 | case VgprV2S16: |
701 | case VgprV2S32: |
702 | case VgprV4S32: |
703 | case VgprB32: |
704 | case VgprB64: |
705 | case VgprB96: |
706 | case VgprB128: |
707 | case VgprB256: |
708 | case VgprB512: |
709 | case Vgpr32SExt: |
710 | case Vgpr32ZExt: |
711 | return VgprRB; |
712 | default: |
713 | return nullptr; |
714 | } |
715 | } |
716 | |
717 | void RegBankLegalizeHelper::applyMappingDst( |
718 | MachineInstr &MI, unsigned &OpIdx, |
719 | const SmallVectorImpl<RegBankLLTMappingApplyID> &MethodIDs) { |
720 | // Defs start from operand 0 |
721 | for (; OpIdx < MethodIDs.size(); ++OpIdx) { |
722 | if (MethodIDs[OpIdx] == None) |
723 | continue; |
724 | MachineOperand &Op = MI.getOperand(i: OpIdx); |
725 | Register Reg = Op.getReg(); |
726 | LLT Ty = MRI.getType(Reg); |
727 | [[maybe_unused]] const RegisterBank *RB = MRI.getRegBank(Reg); |
728 | |
729 | switch (MethodIDs[OpIdx]) { |
730 | // vcc, sgpr and vgpr scalars, pointers and vectors |
731 | case Vcc: |
732 | case Sgpr16: |
733 | case Sgpr32: |
734 | case Sgpr64: |
735 | case Sgpr128: |
736 | case SgprP1: |
737 | case SgprP3: |
738 | case SgprP4: |
739 | case SgprP5: |
740 | case SgprV2S16: |
741 | case SgprV2S32: |
742 | case SgprV4S32: |
743 | case Vgpr16: |
744 | case Vgpr32: |
745 | case Vgpr64: |
746 | case Vgpr128: |
747 | case VgprP0: |
748 | case VgprP1: |
749 | case VgprP3: |
750 | case VgprP4: |
751 | case VgprP5: |
752 | case VgprV2S16: |
753 | case VgprV2S32: |
754 | case VgprV4S32: { |
755 | assert(Ty == getTyFromID(MethodIDs[OpIdx])); |
756 | assert(RB == getRegBankFromID(MethodIDs[OpIdx])); |
757 | break; |
758 | } |
759 | // sgpr and vgpr B-types |
760 | case SgprB32: |
761 | case SgprB64: |
762 | case SgprB96: |
763 | case SgprB128: |
764 | case SgprB256: |
765 | case SgprB512: |
766 | case SgprPtr32: |
767 | case SgprPtr64: |
768 | case SgprPtr128: |
769 | case VgprB32: |
770 | case VgprB64: |
771 | case VgprB96: |
772 | case VgprB128: |
773 | case VgprB256: |
774 | case VgprB512: |
775 | case VgprPtr32: |
776 | case VgprPtr64: |
777 | case VgprPtr128: { |
778 | assert(Ty == getBTyFromID(MethodIDs[OpIdx], Ty)); |
779 | assert(RB == getRegBankFromID(MethodIDs[OpIdx])); |
780 | break; |
781 | } |
782 | // uniform in vcc/vgpr: scalars, vectors and B-types |
783 | case UniInVcc: { |
784 | assert(Ty == S1); |
785 | assert(RB == SgprRB); |
786 | Register NewDst = MRI.createVirtualRegister(RegAttr: VccRB_S1); |
787 | Op.setReg(NewDst); |
788 | auto CopyS32_Vcc = |
789 | B.buildInstr(Opc: AMDGPU::G_AMDGPU_COPY_SCC_VCC, DstOps: {SgprRB_S32}, SrcOps: {NewDst}); |
790 | B.buildTrunc(Res: Reg, Op: CopyS32_Vcc); |
791 | break; |
792 | } |
793 | case UniInVgprS32: |
794 | case UniInVgprV2S16: |
795 | case UniInVgprV4S32: { |
796 | assert(Ty == getTyFromID(MethodIDs[OpIdx])); |
797 | assert(RB == SgprRB); |
798 | Register NewVgprDst = MRI.createVirtualRegister(RegAttr: {.RCOrRB: VgprRB, .Ty: Ty}); |
799 | Op.setReg(NewVgprDst); |
800 | buildReadAnyLane(B, SgprDst: Reg, VgprSrc: NewVgprDst, RBI); |
801 | break; |
802 | } |
803 | case UniInVgprB32: |
804 | case UniInVgprB64: |
805 | case UniInVgprB96: |
806 | case UniInVgprB128: |
807 | case UniInVgprB256: |
808 | case UniInVgprB512: { |
809 | assert(Ty == getBTyFromID(MethodIDs[OpIdx], Ty)); |
810 | assert(RB == SgprRB); |
811 | Register NewVgprDst = MRI.createVirtualRegister(RegAttr: {.RCOrRB: VgprRB, .Ty: Ty}); |
812 | Op.setReg(NewVgprDst); |
813 | AMDGPU::buildReadAnyLane(B, SgprDst: Reg, VgprSrc: NewVgprDst, RBI); |
814 | break; |
815 | } |
816 | // sgpr trunc |
817 | case Sgpr32Trunc: { |
818 | assert(Ty.getSizeInBits() < 32); |
819 | assert(RB == SgprRB); |
820 | Register NewDst = MRI.createVirtualRegister(RegAttr: SgprRB_S32); |
821 | Op.setReg(NewDst); |
822 | B.buildTrunc(Res: Reg, Op: NewDst); |
823 | break; |
824 | } |
825 | case InvalidMapping: { |
826 | LLVM_DEBUG(dbgs() << "Instruction with Invalid mapping: " ; MI.dump();); |
827 | llvm_unreachable("missing fast rule for MI" ); |
828 | } |
829 | default: |
830 | llvm_unreachable("ID not supported" ); |
831 | } |
832 | } |
833 | } |
834 | |
835 | void RegBankLegalizeHelper::applyMappingSrc( |
836 | MachineInstr &MI, unsigned &OpIdx, |
837 | const SmallVectorImpl<RegBankLLTMappingApplyID> &MethodIDs, |
838 | SmallSet<Register, 4> &SgprWaterfallOperandRegs) { |
839 | for (unsigned i = 0; i < MethodIDs.size(); ++OpIdx, ++i) { |
840 | if (MethodIDs[i] == None || MethodIDs[i] == IntrId || MethodIDs[i] == Imm) |
841 | continue; |
842 | |
843 | MachineOperand &Op = MI.getOperand(i: OpIdx); |
844 | Register Reg = Op.getReg(); |
845 | LLT Ty = MRI.getType(Reg); |
846 | const RegisterBank *RB = MRI.getRegBank(Reg); |
847 | |
848 | switch (MethodIDs[i]) { |
849 | case Vcc: { |
850 | assert(Ty == S1); |
851 | assert(RB == VccRB || RB == SgprRB); |
852 | if (RB == SgprRB) { |
853 | auto Aext = B.buildAnyExt(Res: SgprRB_S32, Op: Reg); |
854 | auto CopyVcc_Scc = |
855 | B.buildInstr(Opc: AMDGPU::G_AMDGPU_COPY_VCC_SCC, DstOps: {VccRB_S1}, SrcOps: {Aext}); |
856 | Op.setReg(CopyVcc_Scc.getReg(Idx: 0)); |
857 | } |
858 | break; |
859 | } |
860 | // sgpr scalars, pointers and vectors |
861 | case Sgpr16: |
862 | case Sgpr32: |
863 | case Sgpr64: |
864 | case Sgpr128: |
865 | case SgprP1: |
866 | case SgprP3: |
867 | case SgprP4: |
868 | case SgprP5: |
869 | case SgprV2S16: |
870 | case SgprV2S32: |
871 | case SgprV4S32: { |
872 | assert(Ty == getTyFromID(MethodIDs[i])); |
873 | assert(RB == getRegBankFromID(MethodIDs[i])); |
874 | break; |
875 | } |
876 | // sgpr B-types |
877 | case SgprB32: |
878 | case SgprB64: |
879 | case SgprB96: |
880 | case SgprB128: |
881 | case SgprB256: |
882 | case SgprB512: |
883 | case SgprPtr32: |
884 | case SgprPtr64: |
885 | case SgprPtr128: { |
886 | assert(Ty == getBTyFromID(MethodIDs[i], Ty)); |
887 | assert(RB == getRegBankFromID(MethodIDs[i])); |
888 | break; |
889 | } |
890 | // vgpr scalars, pointers and vectors |
891 | case Vgpr16: |
892 | case Vgpr32: |
893 | case Vgpr64: |
894 | case Vgpr128: |
895 | case VgprP0: |
896 | case VgprP1: |
897 | case VgprP3: |
898 | case VgprP4: |
899 | case VgprP5: |
900 | case VgprV2S16: |
901 | case VgprV2S32: |
902 | case VgprV4S32: { |
903 | assert(Ty == getTyFromID(MethodIDs[i])); |
904 | if (RB != VgprRB) { |
905 | auto CopyToVgpr = B.buildCopy(Res: {VgprRB, Ty}, Op: Reg); |
906 | Op.setReg(CopyToVgpr.getReg(Idx: 0)); |
907 | } |
908 | break; |
909 | } |
910 | // vgpr B-types |
911 | case VgprB32: |
912 | case VgprB64: |
913 | case VgprB96: |
914 | case VgprB128: |
915 | case VgprB256: |
916 | case VgprB512: |
917 | case VgprPtr32: |
918 | case VgprPtr64: |
919 | case VgprPtr128: { |
920 | assert(Ty == getBTyFromID(MethodIDs[i], Ty)); |
921 | if (RB != VgprRB) { |
922 | auto CopyToVgpr = B.buildCopy(Res: {VgprRB, Ty}, Op: Reg); |
923 | Op.setReg(CopyToVgpr.getReg(Idx: 0)); |
924 | } |
925 | break; |
926 | } |
927 | // sgpr and vgpr scalars with extend |
928 | case Sgpr32AExt: { |
929 | // Note: this ext allows S1, and it is meant to be combined away. |
930 | assert(Ty.getSizeInBits() < 32); |
931 | assert(RB == SgprRB); |
932 | auto Aext = B.buildAnyExt(Res: SgprRB_S32, Op: Reg); |
933 | Op.setReg(Aext.getReg(Idx: 0)); |
934 | break; |
935 | } |
936 | case Sgpr32AExtBoolInReg: { |
937 | // Note: this ext allows S1, and it is meant to be combined away. |
938 | assert(Ty.getSizeInBits() == 1); |
939 | assert(RB == SgprRB); |
940 | auto Aext = B.buildAnyExt(Res: SgprRB_S32, Op: Reg); |
941 | // Zext SgprS1 is not legal, make AND with 1 instead. This instruction is |
942 | // most of times meant to be combined away in AMDGPURegBankCombiner. |
943 | auto Cst1 = B.buildConstant(Res: SgprRB_S32, Val: 1); |
944 | auto BoolInReg = B.buildAnd(Dst: SgprRB_S32, Src0: Aext, Src1: Cst1); |
945 | Op.setReg(BoolInReg.getReg(Idx: 0)); |
946 | break; |
947 | } |
948 | case Sgpr32SExt: { |
949 | assert(1 < Ty.getSizeInBits() && Ty.getSizeInBits() < 32); |
950 | assert(RB == SgprRB); |
951 | auto Sext = B.buildSExt(Res: SgprRB_S32, Op: Reg); |
952 | Op.setReg(Sext.getReg(Idx: 0)); |
953 | break; |
954 | } |
955 | case Sgpr32ZExt: { |
956 | assert(1 < Ty.getSizeInBits() && Ty.getSizeInBits() < 32); |
957 | assert(RB == SgprRB); |
958 | auto Zext = B.buildZExt(Res: {SgprRB, S32}, Op: Reg); |
959 | Op.setReg(Zext.getReg(Idx: 0)); |
960 | break; |
961 | } |
962 | case Vgpr32SExt: { |
963 | // Note this ext allows S1, and it is meant to be combined away. |
964 | assert(Ty.getSizeInBits() < 32); |
965 | assert(RB == VgprRB); |
966 | auto Sext = B.buildSExt(Res: {VgprRB, S32}, Op: Reg); |
967 | Op.setReg(Sext.getReg(Idx: 0)); |
968 | break; |
969 | } |
970 | case Vgpr32ZExt: { |
971 | // Note this ext allows S1, and it is meant to be combined away. |
972 | assert(Ty.getSizeInBits() < 32); |
973 | assert(RB == VgprRB); |
974 | auto Zext = B.buildZExt(Res: {VgprRB, S32}, Op: Reg); |
975 | Op.setReg(Zext.getReg(Idx: 0)); |
976 | break; |
977 | } |
978 | default: |
979 | llvm_unreachable("ID not supported" ); |
980 | } |
981 | } |
982 | } |
983 | |
984 | void RegBankLegalizeHelper::applyMappingPHI(MachineInstr &MI) { |
985 | Register Dst = MI.getOperand(i: 0).getReg(); |
986 | LLT Ty = MRI.getType(Reg: Dst); |
987 | |
988 | if (Ty == LLT::scalar(SizeInBits: 1) && MUI.isUniform(V: Dst)) { |
989 | B.setInsertPt(MBB&: *MI.getParent(), II: MI.getParent()->getFirstNonPHI()); |
990 | |
991 | Register NewDst = MRI.createVirtualRegister(RegAttr: SgprRB_S32); |
992 | MI.getOperand(i: 0).setReg(NewDst); |
993 | B.buildTrunc(Res: Dst, Op: NewDst); |
994 | |
995 | for (unsigned i = 1; i < MI.getNumOperands(); i += 2) { |
996 | Register UseReg = MI.getOperand(i).getReg(); |
997 | |
998 | auto DefMI = MRI.getVRegDef(Reg: UseReg)->getIterator(); |
999 | MachineBasicBlock *DefMBB = DefMI->getParent(); |
1000 | |
1001 | B.setInsertPt(MBB&: *DefMBB, II: DefMBB->SkipPHIsAndLabels(I: std::next(x: DefMI))); |
1002 | |
1003 | auto NewUse = B.buildAnyExt(Res: SgprRB_S32, Op: UseReg); |
1004 | MI.getOperand(i).setReg(NewUse.getReg(Idx: 0)); |
1005 | } |
1006 | |
1007 | return; |
1008 | } |
1009 | |
1010 | // ALL divergent i1 phis should be already lowered and inst-selected into PHI |
1011 | // with sgpr reg class and S1 LLT. |
1012 | // Note: this includes divergent phis that don't require lowering. |
1013 | if (Ty == LLT::scalar(SizeInBits: 1) && MUI.isDivergent(V: Dst)) { |
1014 | LLVM_DEBUG(dbgs() << "Divergent S1 G_PHI: " ; MI.dump();); |
1015 | llvm_unreachable("Make sure to run AMDGPUGlobalISelDivergenceLowering " |
1016 | "before RegBankLegalize to lower lane mask(vcc) phis" ); |
1017 | } |
1018 | |
1019 | // We accept all types that can fit in some register class. |
1020 | // Uniform G_PHIs have all sgpr registers. |
1021 | // Divergent G_PHIs have vgpr dst but inputs can be sgpr or vgpr. |
1022 | if (Ty == LLT::scalar(SizeInBits: 32) || Ty == LLT::pointer(AddressSpace: 1, SizeInBits: 64) || |
1023 | Ty == LLT::pointer(AddressSpace: 4, SizeInBits: 64)) { |
1024 | return; |
1025 | } |
1026 | |
1027 | LLVM_DEBUG(dbgs() << "G_PHI not handled: " ; MI.dump();); |
1028 | llvm_unreachable("type not supported" ); |
1029 | } |
1030 | |
1031 | [[maybe_unused]] static bool verifyRegBankOnOperands(MachineInstr &MI, |
1032 | const RegisterBank *RB, |
1033 | MachineRegisterInfo &MRI, |
1034 | unsigned StartOpIdx, |
1035 | unsigned EndOpIdx) { |
1036 | for (unsigned i = StartOpIdx; i <= EndOpIdx; ++i) { |
1037 | if (MRI.getRegBankOrNull(Reg: MI.getOperand(i).getReg()) != RB) |
1038 | return false; |
1039 | } |
1040 | return true; |
1041 | } |
1042 | |
1043 | void RegBankLegalizeHelper::applyMappingTrivial(MachineInstr &MI) { |
1044 | const RegisterBank *RB = MRI.getRegBank(Reg: MI.getOperand(i: 0).getReg()); |
1045 | // Put RB on all registers |
1046 | unsigned NumDefs = MI.getNumDefs(); |
1047 | unsigned NumOperands = MI.getNumOperands(); |
1048 | |
1049 | assert(verifyRegBankOnOperands(MI, RB, MRI, 0, NumDefs - 1)); |
1050 | if (RB == SgprRB) |
1051 | assert(verifyRegBankOnOperands(MI, RB, MRI, NumDefs, NumOperands - 1)); |
1052 | |
1053 | if (RB == VgprRB) { |
1054 | B.setInstr(MI); |
1055 | for (unsigned i = NumDefs; i < NumOperands; ++i) { |
1056 | Register Reg = MI.getOperand(i).getReg(); |
1057 | if (MRI.getRegBank(Reg) != RB) { |
1058 | auto Copy = B.buildCopy(Res: {VgprRB, MRI.getType(Reg)}, Op: Reg); |
1059 | MI.getOperand(i).setReg(Copy.getReg(Idx: 0)); |
1060 | } |
1061 | } |
1062 | } |
1063 | } |
1064 | |