1 | //===- SPIRVBuiltins.cpp - SPIR-V Built-in Functions ------------*- C++ -*-===// |
2 | // |
3 | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
4 | // See https://llvm.org/LICENSE.txt for license information. |
5 | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
6 | // |
7 | //===----------------------------------------------------------------------===// |
8 | // |
9 | // This file implements lowering builtin function calls and types using their |
10 | // demangled names and TableGen records. |
11 | // |
12 | //===----------------------------------------------------------------------===// |
13 | |
14 | #include "SPIRVBuiltins.h" |
15 | #include "SPIRV.h" |
16 | #include "SPIRVSubtarget.h" |
17 | #include "SPIRVUtils.h" |
18 | #include "llvm/ADT/StringExtras.h" |
19 | #include "llvm/Analysis/ValueTracking.h" |
20 | #include "llvm/IR/IntrinsicsSPIRV.h" |
21 | #include <regex> |
22 | #include <string> |
23 | #include <tuple> |
24 | |
25 | #define DEBUG_TYPE "spirv-builtins" |
26 | |
27 | namespace llvm { |
28 | namespace SPIRV { |
29 | #define GET_BuiltinGroup_DECL |
30 | #include "SPIRVGenTables.inc" |
31 | |
32 | struct DemangledBuiltin { |
33 | StringRef Name; |
34 | InstructionSet::InstructionSet Set; |
35 | BuiltinGroup Group; |
36 | uint8_t MinNumArgs; |
37 | uint8_t MaxNumArgs; |
38 | }; |
39 | |
40 | #define GET_DemangledBuiltins_DECL |
41 | #define GET_DemangledBuiltins_IMPL |
42 | |
43 | struct IncomingCall { |
44 | const std::string BuiltinName; |
45 | const DemangledBuiltin *Builtin; |
46 | |
47 | const Register ReturnRegister; |
48 | const SPIRVType *ReturnType; |
49 | const SmallVectorImpl<Register> &Arguments; |
50 | |
51 | IncomingCall(const std::string BuiltinName, const DemangledBuiltin *Builtin, |
52 | const Register ReturnRegister, const SPIRVType *ReturnType, |
53 | const SmallVectorImpl<Register> &Arguments) |
54 | : BuiltinName(BuiltinName), Builtin(Builtin), |
55 | ReturnRegister(ReturnRegister), ReturnType(ReturnType), |
56 | Arguments(Arguments) {} |
57 | |
58 | bool isSpirvOp() const { return BuiltinName.rfind(s: "__spirv_" , pos: 0) == 0; } |
59 | }; |
60 | |
61 | struct NativeBuiltin { |
62 | StringRef Name; |
63 | InstructionSet::InstructionSet Set; |
64 | uint32_t Opcode; |
65 | }; |
66 | |
67 | #define GET_NativeBuiltins_DECL |
68 | #define GET_NativeBuiltins_IMPL |
69 | |
70 | struct GroupBuiltin { |
71 | StringRef Name; |
72 | uint32_t Opcode; |
73 | uint32_t GroupOperation; |
74 | bool IsElect; |
75 | bool IsAllOrAny; |
76 | bool IsAllEqual; |
77 | bool IsBallot; |
78 | bool IsInverseBallot; |
79 | bool ; |
80 | bool IsBallotFindBit; |
81 | bool IsLogical; |
82 | bool NoGroupOperation; |
83 | bool HasBoolArg; |
84 | }; |
85 | |
86 | #define GET_GroupBuiltins_DECL |
87 | #define GET_GroupBuiltins_IMPL |
88 | |
89 | struct IntelSubgroupsBuiltin { |
90 | StringRef Name; |
91 | uint32_t Opcode; |
92 | bool IsBlock; |
93 | bool IsWrite; |
94 | bool IsMedia; |
95 | }; |
96 | |
97 | #define GET_IntelSubgroupsBuiltins_DECL |
98 | #define GET_IntelSubgroupsBuiltins_IMPL |
99 | |
100 | struct AtomicFloatingBuiltin { |
101 | StringRef Name; |
102 | uint32_t Opcode; |
103 | }; |
104 | |
105 | #define GET_AtomicFloatingBuiltins_DECL |
106 | #define GET_AtomicFloatingBuiltins_IMPL |
107 | struct GroupUniformBuiltin { |
108 | StringRef Name; |
109 | uint32_t Opcode; |
110 | bool IsLogical; |
111 | }; |
112 | |
113 | #define GET_GroupUniformBuiltins_DECL |
114 | #define GET_GroupUniformBuiltins_IMPL |
115 | |
116 | struct GetBuiltin { |
117 | StringRef Name; |
118 | InstructionSet::InstructionSet Set; |
119 | BuiltIn::BuiltIn Value; |
120 | }; |
121 | |
122 | using namespace BuiltIn; |
123 | #define GET_GetBuiltins_DECL |
124 | #define GET_GetBuiltins_IMPL |
125 | |
126 | struct ImageQueryBuiltin { |
127 | StringRef Name; |
128 | InstructionSet::InstructionSet Set; |
129 | uint32_t Component; |
130 | }; |
131 | |
132 | #define GET_ImageQueryBuiltins_DECL |
133 | #define GET_ImageQueryBuiltins_IMPL |
134 | |
135 | struct IntegerDotProductBuiltin { |
136 | StringRef Name; |
137 | uint32_t Opcode; |
138 | bool IsSwapReq; |
139 | }; |
140 | |
141 | #define GET_IntegerDotProductBuiltins_DECL |
142 | #define GET_IntegerDotProductBuiltins_IMPL |
143 | |
144 | struct ConvertBuiltin { |
145 | StringRef Name; |
146 | InstructionSet::InstructionSet Set; |
147 | bool IsDestinationSigned; |
148 | bool IsSaturated; |
149 | bool IsRounded; |
150 | bool IsBfloat16; |
151 | FPRoundingMode::FPRoundingMode RoundingMode; |
152 | }; |
153 | |
154 | struct VectorLoadStoreBuiltin { |
155 | StringRef Name; |
156 | InstructionSet::InstructionSet Set; |
157 | uint32_t Number; |
158 | uint32_t ElementCount; |
159 | bool IsRounded; |
160 | FPRoundingMode::FPRoundingMode RoundingMode; |
161 | }; |
162 | |
163 | using namespace FPRoundingMode; |
164 | #define GET_ConvertBuiltins_DECL |
165 | #define GET_ConvertBuiltins_IMPL |
166 | |
167 | using namespace InstructionSet; |
168 | #define GET_VectorLoadStoreBuiltins_DECL |
169 | #define GET_VectorLoadStoreBuiltins_IMPL |
170 | |
171 | #define GET_CLMemoryScope_DECL |
172 | #define GET_CLSamplerAddressingMode_DECL |
173 | #define GET_CLMemoryFenceFlags_DECL |
174 | #define GET_ExtendedBuiltins_DECL |
175 | #include "SPIRVGenTables.inc" |
176 | } // namespace SPIRV |
177 | |
178 | //===----------------------------------------------------------------------===// |
179 | // Misc functions for looking up builtins and veryfying requirements using |
180 | // TableGen records |
181 | //===----------------------------------------------------------------------===// |
182 | |
183 | namespace SPIRV { |
184 | /// Parses the name part of the demangled builtin call. |
185 | std::string lookupBuiltinNameHelper(StringRef DemangledCall, |
186 | FPDecorationId *DecorationId) { |
187 | StringRef PassPrefix = "(anonymous namespace)::" ; |
188 | std::string BuiltinName; |
189 | // Itanium Demangler result may have "(anonymous namespace)::" prefix |
190 | if (DemangledCall.starts_with(Prefix: PassPrefix)) |
191 | BuiltinName = DemangledCall.substr(Start: PassPrefix.size()); |
192 | else |
193 | BuiltinName = DemangledCall; |
194 | // Extract the builtin function name and types of arguments from the call |
195 | // skeleton. |
196 | BuiltinName = BuiltinName.substr(pos: 0, n: BuiltinName.find(c: '(')); |
197 | |
198 | // Account for possible "__spirv_ocl_" prefix in SPIR-V friendly LLVM IR |
199 | if (BuiltinName.rfind(s: "__spirv_ocl_" , pos: 0) == 0) |
200 | BuiltinName = BuiltinName.substr(pos: 12); |
201 | |
202 | // Check if the extracted name contains type information between angle |
203 | // brackets. If so, the builtin is an instantiated template - needs to have |
204 | // the information after angle brackets and return type removed. |
205 | std::size_t Pos1 = BuiltinName.rfind(c: '<'); |
206 | if (Pos1 != std::string::npos && BuiltinName.back() == '>') { |
207 | std::size_t Pos2 = BuiltinName.rfind(c: ' ', pos: Pos1); |
208 | if (Pos2 == std::string::npos) |
209 | Pos2 = 0; |
210 | else |
211 | ++Pos2; |
212 | BuiltinName = BuiltinName.substr(pos: Pos2, n: Pos1 - Pos2); |
213 | BuiltinName = BuiltinName.substr(pos: BuiltinName.find_last_of(c: ' ') + 1); |
214 | } |
215 | |
216 | // Check if the extracted name begins with: |
217 | // - "__spirv_ImageSampleExplicitLod" |
218 | // - "__spirv_ImageRead" |
219 | // - "__spirv_ImageWrite" |
220 | // - "__spirv_ImageQuerySizeLod" |
221 | // - "__spirv_UDotKHR" |
222 | // - "__spirv_SDotKHR" |
223 | // - "__spirv_SUDotKHR" |
224 | // - "__spirv_SDotAccSatKHR" |
225 | // - "__spirv_UDotAccSatKHR" |
226 | // - "__spirv_SUDotAccSatKHR" |
227 | // - "__spirv_ReadClockKHR" |
228 | // - "__spirv_SubgroupBlockReadINTEL" |
229 | // - "__spirv_SubgroupImageBlockReadINTEL" |
230 | // - "__spirv_SubgroupImageMediaBlockReadINTEL" |
231 | // - "__spirv_SubgroupImageMediaBlockWriteINTEL" |
232 | // - "__spirv_Convert" |
233 | // - "__spirv_UConvert" |
234 | // - "__spirv_SConvert" |
235 | // - "__spirv_FConvert" |
236 | // - "__spirv_SatConvert" |
237 | // and maybe contains return type information at the end "_R<type>". |
238 | // If so, extract the plain builtin name without the type information. |
239 | static const std::regex SpvWithR( |
240 | "(__spirv_(ImageSampleExplicitLod|ImageRead|ImageWrite|ImageQuerySizeLod|" |
241 | "UDotKHR|" |
242 | "SDotKHR|SUDotKHR|SDotAccSatKHR|UDotAccSatKHR|SUDotAccSatKHR|" |
243 | "ReadClockKHR|SubgroupBlockReadINTEL|SubgroupImageBlockReadINTEL|" |
244 | "SubgroupImageMediaBlockReadINTEL|SubgroupImageMediaBlockWriteINTEL|" |
245 | "Convert|" |
246 | "UConvert|SConvert|FConvert|SatConvert)[^_]*)(_R[^_]*_?(\\w+)?.*)?" ); |
247 | std::smatch Match; |
248 | if (std::regex_match(s: BuiltinName, m&: Match, re: SpvWithR) && Match.size() > 1) { |
249 | std::ssub_match SubMatch; |
250 | if (DecorationId && Match.size() > 3) { |
251 | SubMatch = Match[4]; |
252 | *DecorationId = demangledPostfixToDecorationId(S: SubMatch.str()); |
253 | } |
254 | SubMatch = Match[1]; |
255 | BuiltinName = SubMatch.str(); |
256 | } |
257 | |
258 | return BuiltinName; |
259 | } |
260 | } // namespace SPIRV |
261 | |
262 | /// Looks up the demangled builtin call in the SPIRVBuiltins.td records using |
263 | /// the provided \p DemangledCall and specified \p Set. |
264 | /// |
265 | /// The lookup follows the following algorithm, returning the first successful |
266 | /// match: |
267 | /// 1. Search with the plain demangled name (expecting a 1:1 match). |
268 | /// 2. Search with the prefix before or suffix after the demangled name |
269 | /// signyfying the type of the first argument. |
270 | /// |
271 | /// \returns Wrapper around the demangled call and found builtin definition. |
272 | static std::unique_ptr<const SPIRV::IncomingCall> |
273 | lookupBuiltin(StringRef DemangledCall, |
274 | SPIRV::InstructionSet::InstructionSet Set, |
275 | Register ReturnRegister, const SPIRVType *ReturnType, |
276 | const SmallVectorImpl<Register> &Arguments) { |
277 | std::string BuiltinName = SPIRV::lookupBuiltinNameHelper(DemangledCall); |
278 | |
279 | SmallVector<StringRef, 10> BuiltinArgumentTypes; |
280 | StringRef BuiltinArgs = |
281 | DemangledCall.slice(Start: DemangledCall.find(C: '(') + 1, End: DemangledCall.find(C: ')')); |
282 | BuiltinArgs.split(A&: BuiltinArgumentTypes, Separator: ',', MaxSplit: -1, KeepEmpty: false); |
283 | |
284 | // Look up the builtin in the defined set. Start with the plain demangled |
285 | // name, expecting a 1:1 match in the defined builtin set. |
286 | const SPIRV::DemangledBuiltin *Builtin; |
287 | if ((Builtin = SPIRV::lookupBuiltin(Name: BuiltinName, Set))) |
288 | return std::make_unique<SPIRV::IncomingCall>( |
289 | args&: BuiltinName, args&: Builtin, args&: ReturnRegister, args&: ReturnType, args: Arguments); |
290 | |
291 | // If the initial look up was unsuccessful and the demangled call takes at |
292 | // least 1 argument, add a prefix or suffix signifying the type of the first |
293 | // argument and repeat the search. |
294 | if (BuiltinArgumentTypes.size() >= 1) { |
295 | char FirstArgumentType = BuiltinArgumentTypes[0][0]; |
296 | // Prefix to be added to the builtin's name for lookup. |
297 | // For example, OpenCL "abs" taking an unsigned value has a prefix "u_". |
298 | std::string Prefix; |
299 | |
300 | switch (FirstArgumentType) { |
301 | // Unsigned: |
302 | case 'u': |
303 | if (Set == SPIRV::InstructionSet::OpenCL_std) |
304 | Prefix = "u_" ; |
305 | else if (Set == SPIRV::InstructionSet::GLSL_std_450) |
306 | Prefix = "u" ; |
307 | break; |
308 | // Signed: |
309 | case 'c': |
310 | case 's': |
311 | case 'i': |
312 | case 'l': |
313 | if (Set == SPIRV::InstructionSet::OpenCL_std) |
314 | Prefix = "s_" ; |
315 | else if (Set == SPIRV::InstructionSet::GLSL_std_450) |
316 | Prefix = "s" ; |
317 | break; |
318 | // Floating-point: |
319 | case 'f': |
320 | case 'd': |
321 | case 'h': |
322 | if (Set == SPIRV::InstructionSet::OpenCL_std || |
323 | Set == SPIRV::InstructionSet::GLSL_std_450) |
324 | Prefix = "f" ; |
325 | break; |
326 | } |
327 | |
328 | // If argument-type name prefix was added, look up the builtin again. |
329 | if (!Prefix.empty() && |
330 | (Builtin = SPIRV::lookupBuiltin(Name: Prefix + BuiltinName, Set))) |
331 | return std::make_unique<SPIRV::IncomingCall>( |
332 | args&: BuiltinName, args&: Builtin, args&: ReturnRegister, args&: ReturnType, args: Arguments); |
333 | |
334 | // If lookup with a prefix failed, find a suffix to be added to the |
335 | // builtin's name for lookup. For example, OpenCL "group_reduce_max" taking |
336 | // an unsigned value has a suffix "u". |
337 | std::string Suffix; |
338 | |
339 | switch (FirstArgumentType) { |
340 | // Unsigned: |
341 | case 'u': |
342 | Suffix = "u" ; |
343 | break; |
344 | // Signed: |
345 | case 'c': |
346 | case 's': |
347 | case 'i': |
348 | case 'l': |
349 | Suffix = "s" ; |
350 | break; |
351 | // Floating-point: |
352 | case 'f': |
353 | case 'd': |
354 | case 'h': |
355 | Suffix = "f" ; |
356 | break; |
357 | } |
358 | |
359 | // If argument-type name suffix was added, look up the builtin again. |
360 | if (!Suffix.empty() && |
361 | (Builtin = SPIRV::lookupBuiltin(Name: BuiltinName + Suffix, Set))) |
362 | return std::make_unique<SPIRV::IncomingCall>( |
363 | args&: BuiltinName, args&: Builtin, args&: ReturnRegister, args&: ReturnType, args: Arguments); |
364 | } |
365 | |
366 | // No builtin with such name was found in the set. |
367 | return nullptr; |
368 | } |
369 | |
370 | static MachineInstr *getBlockStructInstr(Register ParamReg, |
371 | MachineRegisterInfo *MRI) { |
372 | // We expect the following sequence of instructions: |
373 | // %0:_(pN) = G_INTRINSIC_W_SIDE_EFFECTS intrinsic(@llvm.spv.alloca) |
374 | // or = G_GLOBAL_VALUE @block_literal_global |
375 | // %1:_(pN) = G_INTRINSIC_W_SIDE_EFFECTS intrinsic(@llvm.spv.bitcast), %0 |
376 | // %2:_(p4) = G_ADDRSPACE_CAST %1:_(pN) |
377 | MachineInstr *MI = MRI->getUniqueVRegDef(Reg: ParamReg); |
378 | assert(MI->getOpcode() == TargetOpcode::G_ADDRSPACE_CAST && |
379 | MI->getOperand(1).isReg()); |
380 | Register BitcastReg = MI->getOperand(i: 1).getReg(); |
381 | MachineInstr *BitcastMI = MRI->getUniqueVRegDef(Reg: BitcastReg); |
382 | assert(isSpvIntrinsic(*BitcastMI, Intrinsic::spv_bitcast) && |
383 | BitcastMI->getOperand(2).isReg()); |
384 | Register ValueReg = BitcastMI->getOperand(i: 2).getReg(); |
385 | MachineInstr *ValueMI = MRI->getUniqueVRegDef(Reg: ValueReg); |
386 | return ValueMI; |
387 | } |
388 | |
389 | // Return an integer constant corresponding to the given register and |
390 | // defined in spv_track_constant. |
391 | // TODO: maybe unify with prelegalizer pass. |
392 | static unsigned getConstFromIntrinsic(Register Reg, MachineRegisterInfo *MRI) { |
393 | MachineInstr *DefMI = MRI->getUniqueVRegDef(Reg); |
394 | assert(DefMI->getOpcode() == TargetOpcode::G_CONSTANT && |
395 | DefMI->getOperand(1).isCImm()); |
396 | return DefMI->getOperand(i: 1).getCImm()->getValue().getZExtValue(); |
397 | } |
398 | |
399 | // Return type of the instruction result from spv_assign_type intrinsic. |
400 | // TODO: maybe unify with prelegalizer pass. |
401 | static const Type *getMachineInstrType(MachineInstr *MI) { |
402 | MachineInstr *NextMI = MI->getNextNode(); |
403 | if (!NextMI) |
404 | return nullptr; |
405 | if (isSpvIntrinsic(MI: *NextMI, IntrinsicID: Intrinsic::spv_assign_name)) |
406 | if ((NextMI = NextMI->getNextNode()) == nullptr) |
407 | return nullptr; |
408 | Register ValueReg = MI->getOperand(i: 0).getReg(); |
409 | if ((!isSpvIntrinsic(MI: *NextMI, IntrinsicID: Intrinsic::spv_assign_type) && |
410 | !isSpvIntrinsic(MI: *NextMI, IntrinsicID: Intrinsic::spv_assign_ptr_type)) || |
411 | NextMI->getOperand(i: 1).getReg() != ValueReg) |
412 | return nullptr; |
413 | Type *Ty = getMDOperandAsType(N: NextMI->getOperand(i: 2).getMetadata(), I: 0); |
414 | assert(Ty && "Type is expected" ); |
415 | return Ty; |
416 | } |
417 | |
418 | static const Type *getBlockStructType(Register ParamReg, |
419 | MachineRegisterInfo *MRI) { |
420 | // In principle, this information should be passed to us from Clang via |
421 | // an elementtype attribute. However, said attribute requires that |
422 | // the function call be an intrinsic, which is not. Instead, we rely on being |
423 | // able to trace this to the declaration of a variable: OpenCL C specification |
424 | // section 6.12.5 should guarantee that we can do this. |
425 | MachineInstr *MI = getBlockStructInstr(ParamReg, MRI); |
426 | if (MI->getOpcode() == TargetOpcode::G_GLOBAL_VALUE) |
427 | return MI->getOperand(i: 1).getGlobal()->getType(); |
428 | assert(isSpvIntrinsic(*MI, Intrinsic::spv_alloca) && |
429 | "Blocks in OpenCL C must be traceable to allocation site" ); |
430 | return getMachineInstrType(MI); |
431 | } |
432 | |
433 | //===----------------------------------------------------------------------===// |
434 | // Helper functions for building misc instructions |
435 | //===----------------------------------------------------------------------===// |
436 | |
437 | /// Helper function building either a resulting scalar or vector bool register |
438 | /// depending on the expected \p ResultType. |
439 | /// |
440 | /// \returns Tuple of the resulting register and its type. |
441 | static std::tuple<Register, SPIRVType *> |
442 | buildBoolRegister(MachineIRBuilder &MIRBuilder, const SPIRVType *ResultType, |
443 | SPIRVGlobalRegistry *GR) { |
444 | LLT Type; |
445 | SPIRVType *BoolType = GR->getOrCreateSPIRVBoolType(MIRBuilder, EmitIR: true); |
446 | |
447 | if (ResultType->getOpcode() == SPIRV::OpTypeVector) { |
448 | unsigned VectorElements = ResultType->getOperand(i: 2).getImm(); |
449 | BoolType = GR->getOrCreateSPIRVVectorType(BaseType: BoolType, NumElements: VectorElements, |
450 | MIRBuilder, EmitIR: true); |
451 | const FixedVectorType *LLVMVectorType = |
452 | cast<FixedVectorType>(Val: GR->getTypeForSPIRVType(Ty: BoolType)); |
453 | Type = LLT::vector(EC: LLVMVectorType->getElementCount(), ScalarSizeInBits: 1); |
454 | } else { |
455 | Type = LLT::scalar(SizeInBits: 1); |
456 | } |
457 | |
458 | Register ResultRegister = |
459 | MIRBuilder.getMRI()->createGenericVirtualRegister(Ty: Type); |
460 | MIRBuilder.getMRI()->setRegClass(Reg: ResultRegister, RC: GR->getRegClass(SpvType: ResultType)); |
461 | GR->assignSPIRVTypeToVReg(Type: BoolType, VReg: ResultRegister, MF: MIRBuilder.getMF()); |
462 | return std::make_tuple(args&: ResultRegister, args&: BoolType); |
463 | } |
464 | |
465 | /// Helper function for building either a vector or scalar select instruction |
466 | /// depending on the expected \p ResultType. |
467 | static bool buildSelectInst(MachineIRBuilder &MIRBuilder, |
468 | Register ReturnRegister, Register SourceRegister, |
469 | const SPIRVType *ReturnType, |
470 | SPIRVGlobalRegistry *GR) { |
471 | Register TrueConst, FalseConst; |
472 | |
473 | if (ReturnType->getOpcode() == SPIRV::OpTypeVector) { |
474 | unsigned Bits = GR->getScalarOrVectorBitWidth(Type: ReturnType); |
475 | uint64_t AllOnes = APInt::getAllOnes(numBits: Bits).getZExtValue(); |
476 | TrueConst = |
477 | GR->getOrCreateConsIntVector(Val: AllOnes, MIRBuilder, SpvType: ReturnType, EmitIR: true); |
478 | FalseConst = GR->getOrCreateConsIntVector(Val: 0, MIRBuilder, SpvType: ReturnType, EmitIR: true); |
479 | } else { |
480 | TrueConst = GR->buildConstantInt(Val: 1, MIRBuilder, SpvType: ReturnType, EmitIR: true); |
481 | FalseConst = GR->buildConstantInt(Val: 0, MIRBuilder, SpvType: ReturnType, EmitIR: true); |
482 | } |
483 | |
484 | return MIRBuilder.buildSelect(Res: ReturnRegister, Tst: SourceRegister, Op0: TrueConst, |
485 | Op1: FalseConst); |
486 | } |
487 | |
488 | /// Helper function for building a load instruction loading into the |
489 | /// \p DestinationReg. |
490 | static Register buildLoadInst(SPIRVType *BaseType, Register PtrRegister, |
491 | MachineIRBuilder &MIRBuilder, |
492 | SPIRVGlobalRegistry *GR, LLT LowLevelType, |
493 | Register DestinationReg = Register(0)) { |
494 | if (!DestinationReg.isValid()) |
495 | DestinationReg = createVirtualRegister(SpvType: BaseType, GR, MIRBuilder); |
496 | // TODO: consider using correct address space and alignment (p0 is canonical |
497 | // type for selection though). |
498 | MachinePointerInfo PtrInfo = MachinePointerInfo(); |
499 | MIRBuilder.buildLoad(Res: DestinationReg, Addr: PtrRegister, PtrInfo, Alignment: Align()); |
500 | return DestinationReg; |
501 | } |
502 | |
503 | /// Helper function for building a load instruction for loading a builtin global |
504 | /// variable of \p BuiltinValue value. |
505 | static Register buildBuiltinVariableLoad( |
506 | MachineIRBuilder &MIRBuilder, SPIRVType *VariableType, |
507 | SPIRVGlobalRegistry *GR, SPIRV::BuiltIn::BuiltIn BuiltinValue, LLT LLType, |
508 | Register Reg = Register(0), bool isConst = true, bool hasLinkageTy = true) { |
509 | Register NewRegister = |
510 | MIRBuilder.getMRI()->createVirtualRegister(RegClass: &SPIRV::pIDRegClass); |
511 | MIRBuilder.getMRI()->setType( |
512 | VReg: NewRegister, |
513 | Ty: LLT::pointer(AddressSpace: storageClassToAddressSpace(SC: SPIRV::StorageClass::Function), |
514 | SizeInBits: GR->getPointerSize())); |
515 | SPIRVType *PtrType = GR->getOrCreateSPIRVPointerType( |
516 | BaseType: VariableType, MIRBuilder, SC: SPIRV::StorageClass::Input); |
517 | GR->assignSPIRVTypeToVReg(Type: PtrType, VReg: NewRegister, MF: MIRBuilder.getMF()); |
518 | |
519 | // Set up the global OpVariable with the necessary builtin decorations. |
520 | Register Variable = GR->buildGlobalVariable( |
521 | Reg: NewRegister, BaseType: PtrType, Name: getLinkStringForBuiltIn(BuiltInValue: BuiltinValue), GV: nullptr, |
522 | Storage: SPIRV::StorageClass::Input, Init: nullptr, /* isConst= */ IsConst: isConst, |
523 | /* HasLinkageTy */ hasLinkageTy, LinkageType: SPIRV::LinkageType::Import, MIRBuilder, |
524 | IsInstSelector: false); |
525 | |
526 | // Load the value from the global variable. |
527 | Register LoadedRegister = |
528 | buildLoadInst(BaseType: VariableType, PtrRegister: Variable, MIRBuilder, GR, LowLevelType: LLType, DestinationReg: Reg); |
529 | MIRBuilder.getMRI()->setType(VReg: LoadedRegister, Ty: LLType); |
530 | return LoadedRegister; |
531 | } |
532 | |
533 | /// Helper external function for inserting ASSIGN_TYPE instuction between \p Reg |
534 | /// and its definition, set the new register as a destination of the definition, |
535 | /// assign SPIRVType to both registers. If SpirvTy is provided, use it as |
536 | /// SPIRVType in ASSIGN_TYPE, otherwise create it from \p Ty. Defined in |
537 | /// SPIRVPreLegalizer.cpp. |
538 | extern void insertAssignInstr(Register Reg, Type *Ty, SPIRVType *SpirvTy, |
539 | SPIRVGlobalRegistry *GR, MachineIRBuilder &MIB, |
540 | MachineRegisterInfo &MRI); |
541 | |
542 | // TODO: Move to TableGen. |
543 | static SPIRV::MemorySemantics::MemorySemantics |
544 | getSPIRVMemSemantics(std::memory_order MemOrder) { |
545 | switch (MemOrder) { |
546 | case std::memory_order_relaxed: |
547 | return SPIRV::MemorySemantics::None; |
548 | case std::memory_order_acquire: |
549 | return SPIRV::MemorySemantics::Acquire; |
550 | case std::memory_order_release: |
551 | return SPIRV::MemorySemantics::Release; |
552 | case std::memory_order_acq_rel: |
553 | return SPIRV::MemorySemantics::AcquireRelease; |
554 | case std::memory_order_seq_cst: |
555 | return SPIRV::MemorySemantics::SequentiallyConsistent; |
556 | default: |
557 | report_fatal_error(reason: "Unknown CL memory scope" ); |
558 | } |
559 | } |
560 | |
561 | static SPIRV::Scope::Scope getSPIRVScope(SPIRV::CLMemoryScope ClScope) { |
562 | switch (ClScope) { |
563 | case SPIRV::CLMemoryScope::memory_scope_work_item: |
564 | return SPIRV::Scope::Invocation; |
565 | case SPIRV::CLMemoryScope::memory_scope_work_group: |
566 | return SPIRV::Scope::Workgroup; |
567 | case SPIRV::CLMemoryScope::memory_scope_device: |
568 | return SPIRV::Scope::Device; |
569 | case SPIRV::CLMemoryScope::memory_scope_all_svm_devices: |
570 | return SPIRV::Scope::CrossDevice; |
571 | case SPIRV::CLMemoryScope::memory_scope_sub_group: |
572 | return SPIRV::Scope::Subgroup; |
573 | } |
574 | report_fatal_error(reason: "Unknown CL memory scope" ); |
575 | } |
576 | |
577 | static Register buildConstantIntReg32(uint64_t Val, |
578 | MachineIRBuilder &MIRBuilder, |
579 | SPIRVGlobalRegistry *GR) { |
580 | return GR->buildConstantInt( |
581 | Val, MIRBuilder, SpvType: GR->getOrCreateSPIRVIntegerType(BitWidth: 32, MIRBuilder), EmitIR: true); |
582 | } |
583 | |
584 | static Register buildScopeReg(Register CLScopeRegister, |
585 | SPIRV::Scope::Scope Scope, |
586 | MachineIRBuilder &MIRBuilder, |
587 | SPIRVGlobalRegistry *GR, |
588 | MachineRegisterInfo *MRI) { |
589 | if (CLScopeRegister.isValid()) { |
590 | auto CLScope = |
591 | static_cast<SPIRV::CLMemoryScope>(getIConstVal(ConstReg: CLScopeRegister, MRI)); |
592 | Scope = getSPIRVScope(ClScope: CLScope); |
593 | |
594 | if (CLScope == static_cast<unsigned>(Scope)) { |
595 | MRI->setRegClass(Reg: CLScopeRegister, RC: &SPIRV::iIDRegClass); |
596 | return CLScopeRegister; |
597 | } |
598 | } |
599 | return buildConstantIntReg32(Val: Scope, MIRBuilder, GR); |
600 | } |
601 | |
602 | static void setRegClassIfNull(Register Reg, MachineRegisterInfo *MRI, |
603 | SPIRVGlobalRegistry *GR) { |
604 | if (MRI->getRegClassOrNull(Reg)) |
605 | return; |
606 | SPIRVType *SpvType = GR->getSPIRVTypeForVReg(VReg: Reg); |
607 | MRI->setRegClass(Reg, |
608 | RC: SpvType ? GR->getRegClass(SpvType) : &SPIRV::iIDRegClass); |
609 | } |
610 | |
611 | static Register buildMemSemanticsReg(Register SemanticsRegister, |
612 | Register PtrRegister, unsigned &Semantics, |
613 | MachineIRBuilder &MIRBuilder, |
614 | SPIRVGlobalRegistry *GR) { |
615 | if (SemanticsRegister.isValid()) { |
616 | MachineRegisterInfo *MRI = MIRBuilder.getMRI(); |
617 | std::memory_order Order = |
618 | static_cast<std::memory_order>(getIConstVal(ConstReg: SemanticsRegister, MRI)); |
619 | Semantics = |
620 | getSPIRVMemSemantics(MemOrder: Order) | |
621 | getMemSemanticsForStorageClass(SC: GR->getPointerStorageClass(VReg: PtrRegister)); |
622 | if (static_cast<unsigned>(Order) == Semantics) { |
623 | MRI->setRegClass(Reg: SemanticsRegister, RC: &SPIRV::iIDRegClass); |
624 | return SemanticsRegister; |
625 | } |
626 | } |
627 | return buildConstantIntReg32(Val: Semantics, MIRBuilder, GR); |
628 | } |
629 | |
630 | static bool buildOpFromWrapper(MachineIRBuilder &MIRBuilder, unsigned Opcode, |
631 | const SPIRV::IncomingCall *Call, |
632 | Register TypeReg, |
633 | ArrayRef<uint32_t> ImmArgs = {}) { |
634 | auto MIB = MIRBuilder.buildInstr(Opcode); |
635 | if (TypeReg.isValid()) |
636 | MIB.addDef(RegNo: Call->ReturnRegister).addUse(RegNo: TypeReg); |
637 | unsigned Sz = Call->Arguments.size() - ImmArgs.size(); |
638 | for (unsigned i = 0; i < Sz; ++i) |
639 | MIB.addUse(RegNo: Call->Arguments[i]); |
640 | for (uint32_t ImmArg : ImmArgs) |
641 | MIB.addImm(Val: ImmArg); |
642 | return true; |
643 | } |
644 | |
645 | /// Helper function for translating atomic init to OpStore. |
646 | static bool buildAtomicInitInst(const SPIRV::IncomingCall *Call, |
647 | MachineIRBuilder &MIRBuilder) { |
648 | if (Call->isSpirvOp()) |
649 | return buildOpFromWrapper(MIRBuilder, Opcode: SPIRV::OpStore, Call, TypeReg: Register(0)); |
650 | |
651 | assert(Call->Arguments.size() == 2 && |
652 | "Need 2 arguments for atomic init translation" ); |
653 | MIRBuilder.buildInstr(Opcode: SPIRV::OpStore) |
654 | .addUse(RegNo: Call->Arguments[0]) |
655 | .addUse(RegNo: Call->Arguments[1]); |
656 | return true; |
657 | } |
658 | |
659 | /// Helper function for building an atomic load instruction. |
660 | static bool buildAtomicLoadInst(const SPIRV::IncomingCall *Call, |
661 | MachineIRBuilder &MIRBuilder, |
662 | SPIRVGlobalRegistry *GR) { |
663 | Register TypeReg = GR->getSPIRVTypeID(SpirvType: Call->ReturnType); |
664 | if (Call->isSpirvOp()) |
665 | return buildOpFromWrapper(MIRBuilder, Opcode: SPIRV::OpAtomicLoad, Call, TypeReg); |
666 | |
667 | Register PtrRegister = Call->Arguments[0]; |
668 | // TODO: if true insert call to __translate_ocl_memory_sccope before |
669 | // OpAtomicLoad and the function implementation. We can use Translator's |
670 | // output for transcoding/atomic_explicit_arguments.cl as an example. |
671 | Register ScopeRegister = |
672 | Call->Arguments.size() > 1 |
673 | ? Call->Arguments[1] |
674 | : buildConstantIntReg32(Val: SPIRV::Scope::Device, MIRBuilder, GR); |
675 | Register MemSemanticsReg; |
676 | if (Call->Arguments.size() > 2) { |
677 | // TODO: Insert call to __translate_ocl_memory_order before OpAtomicLoad. |
678 | MemSemanticsReg = Call->Arguments[2]; |
679 | } else { |
680 | int Semantics = |
681 | SPIRV::MemorySemantics::SequentiallyConsistent | |
682 | getMemSemanticsForStorageClass(SC: GR->getPointerStorageClass(VReg: PtrRegister)); |
683 | MemSemanticsReg = buildConstantIntReg32(Val: Semantics, MIRBuilder, GR); |
684 | } |
685 | |
686 | MIRBuilder.buildInstr(Opcode: SPIRV::OpAtomicLoad) |
687 | .addDef(RegNo: Call->ReturnRegister) |
688 | .addUse(RegNo: TypeReg) |
689 | .addUse(RegNo: PtrRegister) |
690 | .addUse(RegNo: ScopeRegister) |
691 | .addUse(RegNo: MemSemanticsReg); |
692 | return true; |
693 | } |
694 | |
695 | /// Helper function for building an atomic store instruction. |
696 | static bool buildAtomicStoreInst(const SPIRV::IncomingCall *Call, |
697 | MachineIRBuilder &MIRBuilder, |
698 | SPIRVGlobalRegistry *GR) { |
699 | if (Call->isSpirvOp()) |
700 | return buildOpFromWrapper(MIRBuilder, Opcode: SPIRV::OpAtomicStore, Call, TypeReg: Register(0)); |
701 | |
702 | Register ScopeRegister = |
703 | buildConstantIntReg32(Val: SPIRV::Scope::Device, MIRBuilder, GR); |
704 | Register PtrRegister = Call->Arguments[0]; |
705 | int Semantics = |
706 | SPIRV::MemorySemantics::SequentiallyConsistent | |
707 | getMemSemanticsForStorageClass(SC: GR->getPointerStorageClass(VReg: PtrRegister)); |
708 | Register MemSemanticsReg = buildConstantIntReg32(Val: Semantics, MIRBuilder, GR); |
709 | MIRBuilder.buildInstr(Opcode: SPIRV::OpAtomicStore) |
710 | .addUse(RegNo: PtrRegister) |
711 | .addUse(RegNo: ScopeRegister) |
712 | .addUse(RegNo: MemSemanticsReg) |
713 | .addUse(RegNo: Call->Arguments[1]); |
714 | return true; |
715 | } |
716 | |
717 | /// Helper function for building an atomic compare-exchange instruction. |
718 | static bool buildAtomicCompareExchangeInst( |
719 | const SPIRV::IncomingCall *Call, const SPIRV::DemangledBuiltin *Builtin, |
720 | unsigned Opcode, MachineIRBuilder &MIRBuilder, SPIRVGlobalRegistry *GR) { |
721 | if (Call->isSpirvOp()) |
722 | return buildOpFromWrapper(MIRBuilder, Opcode, Call, |
723 | TypeReg: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
724 | |
725 | bool IsCmpxchg = Call->Builtin->Name.contains(Other: "cmpxchg" ); |
726 | MachineRegisterInfo *MRI = MIRBuilder.getMRI(); |
727 | |
728 | Register ObjectPtr = Call->Arguments[0]; // Pointer (volatile A *object.) |
729 | Register ExpectedArg = Call->Arguments[1]; // Comparator (C* expected). |
730 | Register Desired = Call->Arguments[2]; // Value (C Desired). |
731 | SPIRVType *SpvDesiredTy = GR->getSPIRVTypeForVReg(VReg: Desired); |
732 | LLT DesiredLLT = MRI->getType(Reg: Desired); |
733 | |
734 | assert(GR->getSPIRVTypeForVReg(ObjectPtr)->getOpcode() == |
735 | SPIRV::OpTypePointer); |
736 | unsigned ExpectedType = GR->getSPIRVTypeForVReg(VReg: ExpectedArg)->getOpcode(); |
737 | (void)ExpectedType; |
738 | assert(IsCmpxchg ? ExpectedType == SPIRV::OpTypeInt |
739 | : ExpectedType == SPIRV::OpTypePointer); |
740 | assert(GR->isScalarOfType(Desired, SPIRV::OpTypeInt)); |
741 | |
742 | SPIRVType *SpvObjectPtrTy = GR->getSPIRVTypeForVReg(VReg: ObjectPtr); |
743 | assert(SpvObjectPtrTy->getOperand(2).isReg() && "SPIRV type is expected" ); |
744 | auto StorageClass = static_cast<SPIRV::StorageClass::StorageClass>( |
745 | SpvObjectPtrTy->getOperand(i: 1).getImm()); |
746 | auto MemSemStorage = getMemSemanticsForStorageClass(SC: StorageClass); |
747 | |
748 | Register MemSemEqualReg; |
749 | Register MemSemUnequalReg; |
750 | uint64_t MemSemEqual = |
751 | IsCmpxchg |
752 | ? SPIRV::MemorySemantics::None |
753 | : SPIRV::MemorySemantics::SequentiallyConsistent | MemSemStorage; |
754 | uint64_t MemSemUnequal = |
755 | IsCmpxchg |
756 | ? SPIRV::MemorySemantics::None |
757 | : SPIRV::MemorySemantics::SequentiallyConsistent | MemSemStorage; |
758 | if (Call->Arguments.size() >= 4) { |
759 | assert(Call->Arguments.size() >= 5 && |
760 | "Need 5+ args for explicit atomic cmpxchg" ); |
761 | auto MemOrdEq = |
762 | static_cast<std::memory_order>(getIConstVal(ConstReg: Call->Arguments[3], MRI)); |
763 | auto MemOrdNeq = |
764 | static_cast<std::memory_order>(getIConstVal(ConstReg: Call->Arguments[4], MRI)); |
765 | MemSemEqual = getSPIRVMemSemantics(MemOrder: MemOrdEq) | MemSemStorage; |
766 | MemSemUnequal = getSPIRVMemSemantics(MemOrder: MemOrdNeq) | MemSemStorage; |
767 | if (static_cast<unsigned>(MemOrdEq) == MemSemEqual) |
768 | MemSemEqualReg = Call->Arguments[3]; |
769 | if (static_cast<unsigned>(MemOrdNeq) == MemSemEqual) |
770 | MemSemUnequalReg = Call->Arguments[4]; |
771 | } |
772 | if (!MemSemEqualReg.isValid()) |
773 | MemSemEqualReg = buildConstantIntReg32(Val: MemSemEqual, MIRBuilder, GR); |
774 | if (!MemSemUnequalReg.isValid()) |
775 | MemSemUnequalReg = buildConstantIntReg32(Val: MemSemUnequal, MIRBuilder, GR); |
776 | |
777 | Register ScopeReg; |
778 | auto Scope = IsCmpxchg ? SPIRV::Scope::Workgroup : SPIRV::Scope::Device; |
779 | if (Call->Arguments.size() >= 6) { |
780 | assert(Call->Arguments.size() == 6 && |
781 | "Extra args for explicit atomic cmpxchg" ); |
782 | auto ClScope = static_cast<SPIRV::CLMemoryScope>( |
783 | getIConstVal(ConstReg: Call->Arguments[5], MRI)); |
784 | Scope = getSPIRVScope(ClScope); |
785 | if (ClScope == static_cast<unsigned>(Scope)) |
786 | ScopeReg = Call->Arguments[5]; |
787 | } |
788 | if (!ScopeReg.isValid()) |
789 | ScopeReg = buildConstantIntReg32(Val: Scope, MIRBuilder, GR); |
790 | |
791 | Register Expected = IsCmpxchg |
792 | ? ExpectedArg |
793 | : buildLoadInst(BaseType: SpvDesiredTy, PtrRegister: ExpectedArg, MIRBuilder, |
794 | GR, LowLevelType: LLT::scalar(SizeInBits: 64)); |
795 | MRI->setType(VReg: Expected, Ty: DesiredLLT); |
796 | Register Tmp = !IsCmpxchg ? MRI->createGenericVirtualRegister(Ty: DesiredLLT) |
797 | : Call->ReturnRegister; |
798 | if (!MRI->getRegClassOrNull(Reg: Tmp)) |
799 | MRI->setRegClass(Reg: Tmp, RC: GR->getRegClass(SpvType: SpvDesiredTy)); |
800 | GR->assignSPIRVTypeToVReg(Type: SpvDesiredTy, VReg: Tmp, MF: MIRBuilder.getMF()); |
801 | |
802 | SPIRVType *IntTy = GR->getOrCreateSPIRVIntegerType(BitWidth: 32, MIRBuilder); |
803 | MIRBuilder.buildInstr(Opcode) |
804 | .addDef(RegNo: Tmp) |
805 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: IntTy)) |
806 | .addUse(RegNo: ObjectPtr) |
807 | .addUse(RegNo: ScopeReg) |
808 | .addUse(RegNo: MemSemEqualReg) |
809 | .addUse(RegNo: MemSemUnequalReg) |
810 | .addUse(RegNo: Desired) |
811 | .addUse(RegNo: Expected); |
812 | if (!IsCmpxchg) { |
813 | MIRBuilder.buildInstr(Opcode: SPIRV::OpStore).addUse(RegNo: ExpectedArg).addUse(RegNo: Tmp); |
814 | MIRBuilder.buildICmp(Pred: CmpInst::ICMP_EQ, Res: Call->ReturnRegister, Op0: Tmp, Op1: Expected); |
815 | } |
816 | return true; |
817 | } |
818 | |
819 | /// Helper function for building atomic instructions. |
820 | static bool buildAtomicRMWInst(const SPIRV::IncomingCall *Call, unsigned Opcode, |
821 | MachineIRBuilder &MIRBuilder, |
822 | SPIRVGlobalRegistry *GR) { |
823 | if (Call->isSpirvOp()) |
824 | return buildOpFromWrapper(MIRBuilder, Opcode, Call, |
825 | TypeReg: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
826 | |
827 | MachineRegisterInfo *MRI = MIRBuilder.getMRI(); |
828 | Register ScopeRegister = |
829 | Call->Arguments.size() >= 4 ? Call->Arguments[3] : Register(); |
830 | |
831 | assert(Call->Arguments.size() <= 4 && |
832 | "Too many args for explicit atomic RMW" ); |
833 | ScopeRegister = buildScopeReg(CLScopeRegister: ScopeRegister, Scope: SPIRV::Scope::Workgroup, |
834 | MIRBuilder, GR, MRI); |
835 | |
836 | Register PtrRegister = Call->Arguments[0]; |
837 | unsigned Semantics = SPIRV::MemorySemantics::None; |
838 | Register MemSemanticsReg = |
839 | Call->Arguments.size() >= 3 ? Call->Arguments[2] : Register(); |
840 | MemSemanticsReg = buildMemSemanticsReg(SemanticsRegister: MemSemanticsReg, PtrRegister, |
841 | Semantics, MIRBuilder, GR); |
842 | Register ValueReg = Call->Arguments[1]; |
843 | Register ValueTypeReg = GR->getSPIRVTypeID(SpirvType: Call->ReturnType); |
844 | // support cl_ext_float_atomics |
845 | if (Call->ReturnType->getOpcode() == SPIRV::OpTypeFloat) { |
846 | if (Opcode == SPIRV::OpAtomicIAdd) { |
847 | Opcode = SPIRV::OpAtomicFAddEXT; |
848 | } else if (Opcode == SPIRV::OpAtomicISub) { |
849 | // Translate OpAtomicISub applied to a floating type argument to |
850 | // OpAtomicFAddEXT with the negative value operand |
851 | Opcode = SPIRV::OpAtomicFAddEXT; |
852 | Register NegValueReg = |
853 | MRI->createGenericVirtualRegister(Ty: MRI->getType(Reg: ValueReg)); |
854 | MRI->setRegClass(Reg: NegValueReg, RC: GR->getRegClass(SpvType: Call->ReturnType)); |
855 | GR->assignSPIRVTypeToVReg(Type: Call->ReturnType, VReg: NegValueReg, |
856 | MF: MIRBuilder.getMF()); |
857 | MIRBuilder.buildInstr(Opcode: TargetOpcode::G_FNEG) |
858 | .addDef(RegNo: NegValueReg) |
859 | .addUse(RegNo: ValueReg); |
860 | insertAssignInstr(Reg: NegValueReg, Ty: nullptr, SpirvTy: Call->ReturnType, GR, MIB&: MIRBuilder, |
861 | MRI&: MIRBuilder.getMF().getRegInfo()); |
862 | ValueReg = NegValueReg; |
863 | } |
864 | } |
865 | MIRBuilder.buildInstr(Opcode) |
866 | .addDef(RegNo: Call->ReturnRegister) |
867 | .addUse(RegNo: ValueTypeReg) |
868 | .addUse(RegNo: PtrRegister) |
869 | .addUse(RegNo: ScopeRegister) |
870 | .addUse(RegNo: MemSemanticsReg) |
871 | .addUse(RegNo: ValueReg); |
872 | return true; |
873 | } |
874 | |
875 | /// Helper function for building an atomic floating-type instruction. |
876 | static bool buildAtomicFloatingRMWInst(const SPIRV::IncomingCall *Call, |
877 | unsigned Opcode, |
878 | MachineIRBuilder &MIRBuilder, |
879 | SPIRVGlobalRegistry *GR) { |
880 | assert(Call->Arguments.size() == 4 && |
881 | "Wrong number of atomic floating-type builtin" ); |
882 | Register PtrReg = Call->Arguments[0]; |
883 | Register ScopeReg = Call->Arguments[1]; |
884 | Register MemSemanticsReg = Call->Arguments[2]; |
885 | Register ValueReg = Call->Arguments[3]; |
886 | MIRBuilder.buildInstr(Opcode) |
887 | .addDef(RegNo: Call->ReturnRegister) |
888 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)) |
889 | .addUse(RegNo: PtrReg) |
890 | .addUse(RegNo: ScopeReg) |
891 | .addUse(RegNo: MemSemanticsReg) |
892 | .addUse(RegNo: ValueReg); |
893 | return true; |
894 | } |
895 | |
896 | /// Helper function for building atomic flag instructions (e.g. |
897 | /// OpAtomicFlagTestAndSet). |
898 | static bool buildAtomicFlagInst(const SPIRV::IncomingCall *Call, |
899 | unsigned Opcode, MachineIRBuilder &MIRBuilder, |
900 | SPIRVGlobalRegistry *GR) { |
901 | bool IsSet = Opcode == SPIRV::OpAtomicFlagTestAndSet; |
902 | Register TypeReg = GR->getSPIRVTypeID(SpirvType: Call->ReturnType); |
903 | if (Call->isSpirvOp()) |
904 | return buildOpFromWrapper(MIRBuilder, Opcode, Call, |
905 | TypeReg: IsSet ? TypeReg : Register(0)); |
906 | |
907 | MachineRegisterInfo *MRI = MIRBuilder.getMRI(); |
908 | Register PtrRegister = Call->Arguments[0]; |
909 | unsigned Semantics = SPIRV::MemorySemantics::SequentiallyConsistent; |
910 | Register MemSemanticsReg = |
911 | Call->Arguments.size() >= 2 ? Call->Arguments[1] : Register(); |
912 | MemSemanticsReg = buildMemSemanticsReg(SemanticsRegister: MemSemanticsReg, PtrRegister, |
913 | Semantics, MIRBuilder, GR); |
914 | |
915 | assert((Opcode != SPIRV::OpAtomicFlagClear || |
916 | (Semantics != SPIRV::MemorySemantics::Acquire && |
917 | Semantics != SPIRV::MemorySemantics::AcquireRelease)) && |
918 | "Invalid memory order argument!" ); |
919 | |
920 | Register ScopeRegister = |
921 | Call->Arguments.size() >= 3 ? Call->Arguments[2] : Register(); |
922 | ScopeRegister = |
923 | buildScopeReg(CLScopeRegister: ScopeRegister, Scope: SPIRV::Scope::Device, MIRBuilder, GR, MRI); |
924 | |
925 | auto MIB = MIRBuilder.buildInstr(Opcode); |
926 | if (IsSet) |
927 | MIB.addDef(RegNo: Call->ReturnRegister).addUse(RegNo: TypeReg); |
928 | |
929 | MIB.addUse(RegNo: PtrRegister).addUse(RegNo: ScopeRegister).addUse(RegNo: MemSemanticsReg); |
930 | return true; |
931 | } |
932 | |
933 | /// Helper function for building barriers, i.e., memory/control ordering |
934 | /// operations. |
935 | static bool buildBarrierInst(const SPIRV::IncomingCall *Call, unsigned Opcode, |
936 | MachineIRBuilder &MIRBuilder, |
937 | SPIRVGlobalRegistry *GR) { |
938 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
939 | const auto *ST = |
940 | static_cast<const SPIRVSubtarget *>(&MIRBuilder.getMF().getSubtarget()); |
941 | if ((Opcode == SPIRV::OpControlBarrierArriveINTEL || |
942 | Opcode == SPIRV::OpControlBarrierWaitINTEL) && |
943 | !ST->canUseExtension(E: SPIRV::Extension::SPV_INTEL_split_barrier)) { |
944 | std::string DiagMsg = std::string(Builtin->Name) + |
945 | ": the builtin requires the following SPIR-V " |
946 | "extension: SPV_INTEL_split_barrier" ; |
947 | report_fatal_error(reason: DiagMsg.c_str(), gen_crash_diag: false); |
948 | } |
949 | |
950 | if (Call->isSpirvOp()) |
951 | return buildOpFromWrapper(MIRBuilder, Opcode, Call, TypeReg: Register(0)); |
952 | |
953 | MachineRegisterInfo *MRI = MIRBuilder.getMRI(); |
954 | unsigned MemFlags = getIConstVal(ConstReg: Call->Arguments[0], MRI); |
955 | unsigned MemSemantics = SPIRV::MemorySemantics::None; |
956 | |
957 | if (MemFlags & SPIRV::CLK_LOCAL_MEM_FENCE) |
958 | MemSemantics |= SPIRV::MemorySemantics::WorkgroupMemory; |
959 | |
960 | if (MemFlags & SPIRV::CLK_GLOBAL_MEM_FENCE) |
961 | MemSemantics |= SPIRV::MemorySemantics::CrossWorkgroupMemory; |
962 | |
963 | if (MemFlags & SPIRV::CLK_IMAGE_MEM_FENCE) |
964 | MemSemantics |= SPIRV::MemorySemantics::ImageMemory; |
965 | |
966 | if (Opcode == SPIRV::OpMemoryBarrier) |
967 | MemSemantics = getSPIRVMemSemantics(MemOrder: static_cast<std::memory_order>( |
968 | getIConstVal(ConstReg: Call->Arguments[1], MRI))) | |
969 | MemSemantics; |
970 | else if (Opcode == SPIRV::OpControlBarrierArriveINTEL) |
971 | MemSemantics |= SPIRV::MemorySemantics::Release; |
972 | else if (Opcode == SPIRV::OpControlBarrierWaitINTEL) |
973 | MemSemantics |= SPIRV::MemorySemantics::Acquire; |
974 | else |
975 | MemSemantics |= SPIRV::MemorySemantics::SequentiallyConsistent; |
976 | |
977 | Register MemSemanticsReg = |
978 | MemFlags == MemSemantics |
979 | ? Call->Arguments[0] |
980 | : buildConstantIntReg32(Val: MemSemantics, MIRBuilder, GR); |
981 | Register ScopeReg; |
982 | SPIRV::Scope::Scope Scope = SPIRV::Scope::Workgroup; |
983 | SPIRV::Scope::Scope MemScope = Scope; |
984 | if (Call->Arguments.size() >= 2) { |
985 | assert( |
986 | ((Opcode != SPIRV::OpMemoryBarrier && Call->Arguments.size() == 2) || |
987 | (Opcode == SPIRV::OpMemoryBarrier && Call->Arguments.size() == 3)) && |
988 | "Extra args for explicitly scoped barrier" ); |
989 | Register ScopeArg = (Opcode == SPIRV::OpMemoryBarrier) ? Call->Arguments[2] |
990 | : Call->Arguments[1]; |
991 | SPIRV::CLMemoryScope CLScope = |
992 | static_cast<SPIRV::CLMemoryScope>(getIConstVal(ConstReg: ScopeArg, MRI)); |
993 | MemScope = getSPIRVScope(ClScope: CLScope); |
994 | if (!(MemFlags & SPIRV::CLK_LOCAL_MEM_FENCE) || |
995 | (Opcode == SPIRV::OpMemoryBarrier)) |
996 | Scope = MemScope; |
997 | if (CLScope == static_cast<unsigned>(Scope)) |
998 | ScopeReg = Call->Arguments[1]; |
999 | } |
1000 | |
1001 | if (!ScopeReg.isValid()) |
1002 | ScopeReg = buildConstantIntReg32(Val: Scope, MIRBuilder, GR); |
1003 | |
1004 | auto MIB = MIRBuilder.buildInstr(Opcode).addUse(RegNo: ScopeReg); |
1005 | if (Opcode != SPIRV::OpMemoryBarrier) |
1006 | MIB.addUse(RegNo: buildConstantIntReg32(Val: MemScope, MIRBuilder, GR)); |
1007 | MIB.addUse(RegNo: MemSemanticsReg); |
1008 | return true; |
1009 | } |
1010 | |
1011 | /// Helper function for building extended bit operations. |
1012 | static bool buildExtendedBitOpsInst(const SPIRV::IncomingCall *Call, |
1013 | unsigned Opcode, |
1014 | MachineIRBuilder &MIRBuilder, |
1015 | SPIRVGlobalRegistry *GR) { |
1016 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1017 | const auto *ST = |
1018 | static_cast<const SPIRVSubtarget *>(&MIRBuilder.getMF().getSubtarget()); |
1019 | if ((Opcode == SPIRV::OpBitFieldInsert || |
1020 | Opcode == SPIRV::OpBitFieldSExtract || |
1021 | Opcode == SPIRV::OpBitFieldUExtract || Opcode == SPIRV::OpBitReverse) && |
1022 | !ST->canUseExtension(E: SPIRV::Extension::SPV_KHR_bit_instructions)) { |
1023 | std::string DiagMsg = std::string(Builtin->Name) + |
1024 | ": the builtin requires the following SPIR-V " |
1025 | "extension: SPV_KHR_bit_instructions" ; |
1026 | report_fatal_error(reason: DiagMsg.c_str(), gen_crash_diag: false); |
1027 | } |
1028 | |
1029 | // Generate SPIRV instruction accordingly. |
1030 | if (Call->isSpirvOp()) |
1031 | return buildOpFromWrapper(MIRBuilder, Opcode, Call, |
1032 | TypeReg: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
1033 | |
1034 | auto MIB = MIRBuilder.buildInstr(Opcode) |
1035 | .addDef(RegNo: Call->ReturnRegister) |
1036 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
1037 | for (unsigned i = 0; i < Call->Arguments.size(); ++i) |
1038 | MIB.addUse(RegNo: Call->Arguments[i]); |
1039 | |
1040 | return true; |
1041 | } |
1042 | |
1043 | /// Helper function for building Intel's bindless image instructions. |
1044 | static bool buildBindlessImageINTELInst(const SPIRV::IncomingCall *Call, |
1045 | unsigned Opcode, |
1046 | MachineIRBuilder &MIRBuilder, |
1047 | SPIRVGlobalRegistry *GR) { |
1048 | // Generate SPIRV instruction accordingly. |
1049 | if (Call->isSpirvOp()) |
1050 | return buildOpFromWrapper(MIRBuilder, Opcode, Call, |
1051 | TypeReg: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
1052 | |
1053 | MIRBuilder.buildInstr(Opcode) |
1054 | .addDef(RegNo: Call->ReturnRegister) |
1055 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)) |
1056 | .addUse(RegNo: Call->Arguments[0]); |
1057 | |
1058 | return true; |
1059 | } |
1060 | |
1061 | /// Helper function for building Intel's OpBitwiseFunctionINTEL instruction. |
1062 | static bool buildTernaryBitwiseFunctionINTELInst( |
1063 | const SPIRV::IncomingCall *Call, unsigned Opcode, |
1064 | MachineIRBuilder &MIRBuilder, SPIRVGlobalRegistry *GR) { |
1065 | // Generate SPIRV instruction accordingly. |
1066 | if (Call->isSpirvOp()) |
1067 | return buildOpFromWrapper(MIRBuilder, Opcode, Call, |
1068 | TypeReg: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
1069 | |
1070 | auto MIB = MIRBuilder.buildInstr(Opcode) |
1071 | .addDef(RegNo: Call->ReturnRegister) |
1072 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
1073 | for (unsigned i = 0; i < Call->Arguments.size(); ++i) |
1074 | MIB.addUse(RegNo: Call->Arguments[i]); |
1075 | |
1076 | return true; |
1077 | } |
1078 | |
1079 | /// Helper function for building Intel's 2d block io instructions. |
1080 | static bool build2DBlockIOINTELInst(const SPIRV::IncomingCall *Call, |
1081 | unsigned Opcode, |
1082 | MachineIRBuilder &MIRBuilder, |
1083 | SPIRVGlobalRegistry *GR) { |
1084 | // Generate SPIRV instruction accordingly. |
1085 | if (Call->isSpirvOp()) |
1086 | return buildOpFromWrapper(MIRBuilder, Opcode, Call, TypeReg: Register(0)); |
1087 | |
1088 | auto MIB = MIRBuilder.buildInstr(Opcode) |
1089 | .addDef(RegNo: Call->ReturnRegister) |
1090 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
1091 | for (unsigned i = 0; i < Call->Arguments.size(); ++i) |
1092 | MIB.addUse(RegNo: Call->Arguments[i]); |
1093 | |
1094 | return true; |
1095 | } |
1096 | |
1097 | static unsigned getNumComponentsForDim(SPIRV::Dim::Dim dim) { |
1098 | switch (dim) { |
1099 | case SPIRV::Dim::DIM_1D: |
1100 | case SPIRV::Dim::DIM_Buffer: |
1101 | return 1; |
1102 | case SPIRV::Dim::DIM_2D: |
1103 | case SPIRV::Dim::DIM_Cube: |
1104 | case SPIRV::Dim::DIM_Rect: |
1105 | return 2; |
1106 | case SPIRV::Dim::DIM_3D: |
1107 | return 3; |
1108 | default: |
1109 | report_fatal_error(reason: "Cannot get num components for given Dim" ); |
1110 | } |
1111 | } |
1112 | |
1113 | /// Helper function for obtaining the number of size components. |
1114 | static unsigned getNumSizeComponents(SPIRVType *imgType) { |
1115 | assert(imgType->getOpcode() == SPIRV::OpTypeImage); |
1116 | auto dim = static_cast<SPIRV::Dim::Dim>(imgType->getOperand(i: 2).getImm()); |
1117 | unsigned numComps = getNumComponentsForDim(dim); |
1118 | bool arrayed = imgType->getOperand(i: 4).getImm() == 1; |
1119 | return arrayed ? numComps + 1 : numComps; |
1120 | } |
1121 | |
1122 | //===----------------------------------------------------------------------===// |
1123 | // Implementation functions for each builtin group |
1124 | //===----------------------------------------------------------------------===// |
1125 | |
1126 | static bool generateExtInst(const SPIRV::IncomingCall *Call, |
1127 | MachineIRBuilder &MIRBuilder, |
1128 | SPIRVGlobalRegistry *GR) { |
1129 | // Lookup the extended instruction number in the TableGen records. |
1130 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1131 | uint32_t Number = |
1132 | SPIRV::lookupExtendedBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Number; |
1133 | |
1134 | // Build extended instruction. |
1135 | auto MIB = |
1136 | MIRBuilder.buildInstr(Opcode: SPIRV::OpExtInst) |
1137 | .addDef(RegNo: Call->ReturnRegister) |
1138 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)) |
1139 | .addImm(Val: static_cast<uint32_t>(SPIRV::InstructionSet::OpenCL_std)) |
1140 | .addImm(Val: Number); |
1141 | |
1142 | for (auto Argument : Call->Arguments) |
1143 | MIB.addUse(RegNo: Argument); |
1144 | return true; |
1145 | } |
1146 | |
1147 | static bool generateRelationalInst(const SPIRV::IncomingCall *Call, |
1148 | MachineIRBuilder &MIRBuilder, |
1149 | SPIRVGlobalRegistry *GR) { |
1150 | // Lookup the instruction opcode in the TableGen records. |
1151 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1152 | unsigned Opcode = |
1153 | SPIRV::lookupNativeBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Opcode; |
1154 | |
1155 | Register CompareRegister; |
1156 | SPIRVType *RelationType; |
1157 | std::tie(args&: CompareRegister, args&: RelationType) = |
1158 | buildBoolRegister(MIRBuilder, ResultType: Call->ReturnType, GR); |
1159 | |
1160 | // Build relational instruction. |
1161 | auto MIB = MIRBuilder.buildInstr(Opcode) |
1162 | .addDef(RegNo: CompareRegister) |
1163 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: RelationType)); |
1164 | |
1165 | for (auto Argument : Call->Arguments) |
1166 | MIB.addUse(RegNo: Argument); |
1167 | |
1168 | // Build select instruction. |
1169 | return buildSelectInst(MIRBuilder, ReturnRegister: Call->ReturnRegister, SourceRegister: CompareRegister, |
1170 | ReturnType: Call->ReturnType, GR); |
1171 | } |
1172 | |
1173 | static bool generateGroupInst(const SPIRV::IncomingCall *Call, |
1174 | MachineIRBuilder &MIRBuilder, |
1175 | SPIRVGlobalRegistry *GR) { |
1176 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1177 | const SPIRV::GroupBuiltin *GroupBuiltin = |
1178 | SPIRV::lookupGroupBuiltin(Name: Builtin->Name); |
1179 | |
1180 | MachineRegisterInfo *MRI = MIRBuilder.getMRI(); |
1181 | if (Call->isSpirvOp()) { |
1182 | if (GroupBuiltin->NoGroupOperation) { |
1183 | SmallVector<uint32_t, 1> ImmArgs; |
1184 | if (GroupBuiltin->Opcode == |
1185 | SPIRV::OpSubgroupMatrixMultiplyAccumulateINTEL && |
1186 | Call->Arguments.size() > 4) |
1187 | ImmArgs.push_back(Elt: getConstFromIntrinsic(Reg: Call->Arguments[4], MRI)); |
1188 | return buildOpFromWrapper(MIRBuilder, Opcode: GroupBuiltin->Opcode, Call, |
1189 | TypeReg: GR->getSPIRVTypeID(SpirvType: Call->ReturnType), ImmArgs); |
1190 | } |
1191 | |
1192 | // Group Operation is a literal |
1193 | Register GroupOpReg = Call->Arguments[1]; |
1194 | const MachineInstr *MI = getDefInstrMaybeConstant(ConstReg&: GroupOpReg, MRI); |
1195 | if (!MI || MI->getOpcode() != TargetOpcode::G_CONSTANT) |
1196 | report_fatal_error( |
1197 | reason: "Group Operation parameter must be an integer constant" ); |
1198 | uint64_t GrpOp = MI->getOperand(i: 1).getCImm()->getValue().getZExtValue(); |
1199 | Register ScopeReg = Call->Arguments[0]; |
1200 | auto MIB = MIRBuilder.buildInstr(Opcode: GroupBuiltin->Opcode) |
1201 | .addDef(RegNo: Call->ReturnRegister) |
1202 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)) |
1203 | .addUse(RegNo: ScopeReg) |
1204 | .addImm(Val: GrpOp); |
1205 | for (unsigned i = 2; i < Call->Arguments.size(); ++i) |
1206 | MIB.addUse(RegNo: Call->Arguments[i]); |
1207 | return true; |
1208 | } |
1209 | |
1210 | Register Arg0; |
1211 | if (GroupBuiltin->HasBoolArg) { |
1212 | SPIRVType *BoolType = GR->getOrCreateSPIRVBoolType(MIRBuilder, EmitIR: true); |
1213 | Register BoolReg = Call->Arguments[0]; |
1214 | SPIRVType *BoolRegType = GR->getSPIRVTypeForVReg(VReg: BoolReg); |
1215 | if (!BoolRegType) |
1216 | report_fatal_error(reason: "Can't find a register's type definition" ); |
1217 | MachineInstr *ArgInstruction = getDefInstrMaybeConstant(ConstReg&: BoolReg, MRI); |
1218 | if (ArgInstruction->getOpcode() == TargetOpcode::G_CONSTANT) { |
1219 | if (BoolRegType->getOpcode() != SPIRV::OpTypeBool) |
1220 | Arg0 = GR->buildConstantInt(Val: getIConstVal(ConstReg: BoolReg, MRI), MIRBuilder, |
1221 | SpvType: BoolType, EmitIR: true); |
1222 | } else { |
1223 | if (BoolRegType->getOpcode() == SPIRV::OpTypeInt) { |
1224 | Arg0 = MRI->createGenericVirtualRegister(Ty: LLT::scalar(SizeInBits: 1)); |
1225 | MRI->setRegClass(Reg: Arg0, RC: &SPIRV::iIDRegClass); |
1226 | GR->assignSPIRVTypeToVReg(Type: BoolType, VReg: Arg0, MF: MIRBuilder.getMF()); |
1227 | MIRBuilder.buildICmp( |
1228 | Pred: CmpInst::ICMP_NE, Res: Arg0, Op0: BoolReg, |
1229 | Op1: GR->buildConstantInt(Val: 0, MIRBuilder, SpvType: BoolRegType, EmitIR: true)); |
1230 | insertAssignInstr(Reg: Arg0, Ty: nullptr, SpirvTy: BoolType, GR, MIB&: MIRBuilder, |
1231 | MRI&: MIRBuilder.getMF().getRegInfo()); |
1232 | } else if (BoolRegType->getOpcode() != SPIRV::OpTypeBool) { |
1233 | report_fatal_error(reason: "Expect a boolean argument" ); |
1234 | } |
1235 | // if BoolReg is a boolean register, we don't need to do anything |
1236 | } |
1237 | } |
1238 | |
1239 | Register GroupResultRegister = Call->ReturnRegister; |
1240 | SPIRVType *GroupResultType = Call->ReturnType; |
1241 | |
1242 | // TODO: maybe we need to check whether the result type is already boolean |
1243 | // and in this case do not insert select instruction. |
1244 | const bool HasBoolReturnTy = |
1245 | GroupBuiltin->IsElect || GroupBuiltin->IsAllOrAny || |
1246 | GroupBuiltin->IsAllEqual || GroupBuiltin->IsLogical || |
1247 | GroupBuiltin->IsInverseBallot || GroupBuiltin->IsBallotBitExtract; |
1248 | |
1249 | if (HasBoolReturnTy) |
1250 | std::tie(args&: GroupResultRegister, args&: GroupResultType) = |
1251 | buildBoolRegister(MIRBuilder, ResultType: Call->ReturnType, GR); |
1252 | |
1253 | auto Scope = Builtin->Name.starts_with(Prefix: "sub_group" ) ? SPIRV::Scope::Subgroup |
1254 | : SPIRV::Scope::Workgroup; |
1255 | Register ScopeRegister = buildConstantIntReg32(Val: Scope, MIRBuilder, GR); |
1256 | |
1257 | Register VecReg; |
1258 | if (GroupBuiltin->Opcode == SPIRV::OpGroupBroadcast && |
1259 | Call->Arguments.size() > 2) { |
1260 | // For OpGroupBroadcast "LocalId must be an integer datatype. It must be a |
1261 | // scalar, a vector with 2 components, or a vector with 3 components.", |
1262 | // meaning that we must create a vector from the function arguments if |
1263 | // it's a work_group_broadcast(val, local_id_x, local_id_y) or |
1264 | // work_group_broadcast(val, local_id_x, local_id_y, local_id_z) call. |
1265 | Register ElemReg = Call->Arguments[1]; |
1266 | SPIRVType *ElemType = GR->getSPIRVTypeForVReg(VReg: ElemReg); |
1267 | if (!ElemType || ElemType->getOpcode() != SPIRV::OpTypeInt) |
1268 | report_fatal_error(reason: "Expect an integer <LocalId> argument" ); |
1269 | unsigned VecLen = Call->Arguments.size() - 1; |
1270 | VecReg = MRI->createGenericVirtualRegister( |
1271 | Ty: LLT::fixed_vector(NumElements: VecLen, ScalarTy: MRI->getType(Reg: ElemReg))); |
1272 | MRI->setRegClass(Reg: VecReg, RC: &SPIRV::vIDRegClass); |
1273 | SPIRVType *VecType = |
1274 | GR->getOrCreateSPIRVVectorType(BaseType: ElemType, NumElements: VecLen, MIRBuilder, EmitIR: true); |
1275 | GR->assignSPIRVTypeToVReg(Type: VecType, VReg: VecReg, MF: MIRBuilder.getMF()); |
1276 | auto MIB = |
1277 | MIRBuilder.buildInstr(Opcode: TargetOpcode::G_BUILD_VECTOR).addDef(RegNo: VecReg); |
1278 | for (unsigned i = 1; i < Call->Arguments.size(); i++) { |
1279 | MIB.addUse(RegNo: Call->Arguments[i]); |
1280 | setRegClassIfNull(Reg: Call->Arguments[i], MRI, GR); |
1281 | } |
1282 | insertAssignInstr(Reg: VecReg, Ty: nullptr, SpirvTy: VecType, GR, MIB&: MIRBuilder, |
1283 | MRI&: MIRBuilder.getMF().getRegInfo()); |
1284 | } |
1285 | |
1286 | // Build work/sub group instruction. |
1287 | auto MIB = MIRBuilder.buildInstr(Opcode: GroupBuiltin->Opcode) |
1288 | .addDef(RegNo: GroupResultRegister) |
1289 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: GroupResultType)) |
1290 | .addUse(RegNo: ScopeRegister); |
1291 | |
1292 | if (!GroupBuiltin->NoGroupOperation) |
1293 | MIB.addImm(Val: GroupBuiltin->GroupOperation); |
1294 | if (Call->Arguments.size() > 0) { |
1295 | MIB.addUse(RegNo: Arg0.isValid() ? Arg0 : Call->Arguments[0]); |
1296 | setRegClassIfNull(Reg: Call->Arguments[0], MRI, GR); |
1297 | if (VecReg.isValid()) |
1298 | MIB.addUse(RegNo: VecReg); |
1299 | else |
1300 | for (unsigned i = 1; i < Call->Arguments.size(); i++) |
1301 | MIB.addUse(RegNo: Call->Arguments[i]); |
1302 | } |
1303 | |
1304 | // Build select instruction. |
1305 | if (HasBoolReturnTy) |
1306 | buildSelectInst(MIRBuilder, ReturnRegister: Call->ReturnRegister, SourceRegister: GroupResultRegister, |
1307 | ReturnType: Call->ReturnType, GR); |
1308 | return true; |
1309 | } |
1310 | |
1311 | static bool generateIntelSubgroupsInst(const SPIRV::IncomingCall *Call, |
1312 | MachineIRBuilder &MIRBuilder, |
1313 | SPIRVGlobalRegistry *GR) { |
1314 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1315 | MachineFunction &MF = MIRBuilder.getMF(); |
1316 | const auto *ST = static_cast<const SPIRVSubtarget *>(&MF.getSubtarget()); |
1317 | const SPIRV::IntelSubgroupsBuiltin *IntelSubgroups = |
1318 | SPIRV::lookupIntelSubgroupsBuiltin(Name: Builtin->Name); |
1319 | |
1320 | if (IntelSubgroups->IsMedia && |
1321 | !ST->canUseExtension(E: SPIRV::Extension::SPV_INTEL_media_block_io)) { |
1322 | std::string DiagMsg = std::string(Builtin->Name) + |
1323 | ": the builtin requires the following SPIR-V " |
1324 | "extension: SPV_INTEL_media_block_io" ; |
1325 | report_fatal_error(reason: DiagMsg.c_str(), gen_crash_diag: false); |
1326 | } else if (!IntelSubgroups->IsMedia && |
1327 | !ST->canUseExtension(E: SPIRV::Extension::SPV_INTEL_subgroups)) { |
1328 | std::string DiagMsg = std::string(Builtin->Name) + |
1329 | ": the builtin requires the following SPIR-V " |
1330 | "extension: SPV_INTEL_subgroups" ; |
1331 | report_fatal_error(reason: DiagMsg.c_str(), gen_crash_diag: false); |
1332 | } |
1333 | |
1334 | uint32_t OpCode = IntelSubgroups->Opcode; |
1335 | if (Call->isSpirvOp()) { |
1336 | bool IsSet = OpCode != SPIRV::OpSubgroupBlockWriteINTEL && |
1337 | OpCode != SPIRV::OpSubgroupImageBlockWriteINTEL && |
1338 | OpCode != SPIRV::OpSubgroupImageMediaBlockWriteINTEL; |
1339 | return buildOpFromWrapper(MIRBuilder, Opcode: OpCode, Call, |
1340 | TypeReg: IsSet ? GR->getSPIRVTypeID(SpirvType: Call->ReturnType) |
1341 | : Register(0)); |
1342 | } |
1343 | |
1344 | if (IntelSubgroups->IsBlock) { |
1345 | // Minimal number or arguments set in TableGen records is 1 |
1346 | if (SPIRVType *Arg0Type = GR->getSPIRVTypeForVReg(VReg: Call->Arguments[0])) { |
1347 | if (Arg0Type->getOpcode() == SPIRV::OpTypeImage) { |
1348 | // TODO: add required validation from the specification: |
1349 | // "'Image' must be an object whose type is OpTypeImage with a 'Sampled' |
1350 | // operand of 0 or 2. If the 'Sampled' operand is 2, then some |
1351 | // dimensions require a capability." |
1352 | switch (OpCode) { |
1353 | case SPIRV::OpSubgroupBlockReadINTEL: |
1354 | OpCode = SPIRV::OpSubgroupImageBlockReadINTEL; |
1355 | break; |
1356 | case SPIRV::OpSubgroupBlockWriteINTEL: |
1357 | OpCode = SPIRV::OpSubgroupImageBlockWriteINTEL; |
1358 | break; |
1359 | } |
1360 | } |
1361 | } |
1362 | } |
1363 | |
1364 | // TODO: opaque pointers types should be eventually resolved in such a way |
1365 | // that validation of block read is enabled with respect to the following |
1366 | // specification requirement: |
1367 | // "'Result Type' may be a scalar or vector type, and its component type must |
1368 | // be equal to the type pointed to by 'Ptr'." |
1369 | // For example, function parameter type should not be default i8 pointer, but |
1370 | // depend on the result type of the instruction where it is used as a pointer |
1371 | // argument of OpSubgroupBlockReadINTEL |
1372 | |
1373 | // Build Intel subgroups instruction |
1374 | MachineInstrBuilder MIB = |
1375 | IntelSubgroups->IsWrite |
1376 | ? MIRBuilder.buildInstr(Opcode: OpCode) |
1377 | : MIRBuilder.buildInstr(Opcode: OpCode) |
1378 | .addDef(RegNo: Call->ReturnRegister) |
1379 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
1380 | for (size_t i = 0; i < Call->Arguments.size(); ++i) |
1381 | MIB.addUse(RegNo: Call->Arguments[i]); |
1382 | return true; |
1383 | } |
1384 | |
1385 | static bool generateGroupUniformInst(const SPIRV::IncomingCall *Call, |
1386 | MachineIRBuilder &MIRBuilder, |
1387 | SPIRVGlobalRegistry *GR) { |
1388 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1389 | MachineFunction &MF = MIRBuilder.getMF(); |
1390 | const auto *ST = static_cast<const SPIRVSubtarget *>(&MF.getSubtarget()); |
1391 | if (!ST->canUseExtension( |
1392 | E: SPIRV::Extension::SPV_KHR_uniform_group_instructions)) { |
1393 | std::string DiagMsg = std::string(Builtin->Name) + |
1394 | ": the builtin requires the following SPIR-V " |
1395 | "extension: SPV_KHR_uniform_group_instructions" ; |
1396 | report_fatal_error(reason: DiagMsg.c_str(), gen_crash_diag: false); |
1397 | } |
1398 | const SPIRV::GroupUniformBuiltin *GroupUniform = |
1399 | SPIRV::lookupGroupUniformBuiltin(Name: Builtin->Name); |
1400 | MachineRegisterInfo *MRI = MIRBuilder.getMRI(); |
1401 | |
1402 | Register GroupResultReg = Call->ReturnRegister; |
1403 | Register ScopeReg = Call->Arguments[0]; |
1404 | Register ValueReg = Call->Arguments[2]; |
1405 | |
1406 | // Group Operation |
1407 | Register ConstGroupOpReg = Call->Arguments[1]; |
1408 | const MachineInstr *Const = getDefInstrMaybeConstant(ConstReg&: ConstGroupOpReg, MRI); |
1409 | if (!Const || Const->getOpcode() != TargetOpcode::G_CONSTANT) |
1410 | report_fatal_error( |
1411 | reason: "expect a constant group operation for a uniform group instruction" , |
1412 | gen_crash_diag: false); |
1413 | const MachineOperand &ConstOperand = Const->getOperand(i: 1); |
1414 | if (!ConstOperand.isCImm()) |
1415 | report_fatal_error(reason: "uniform group instructions: group operation must be an " |
1416 | "integer constant" , |
1417 | gen_crash_diag: false); |
1418 | |
1419 | auto MIB = MIRBuilder.buildInstr(Opcode: GroupUniform->Opcode) |
1420 | .addDef(RegNo: GroupResultReg) |
1421 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)) |
1422 | .addUse(RegNo: ScopeReg); |
1423 | addNumImm(Imm: ConstOperand.getCImm()->getValue(), MIB); |
1424 | MIB.addUse(RegNo: ValueReg); |
1425 | |
1426 | return true; |
1427 | } |
1428 | |
1429 | static bool generateKernelClockInst(const SPIRV::IncomingCall *Call, |
1430 | MachineIRBuilder &MIRBuilder, |
1431 | SPIRVGlobalRegistry *GR) { |
1432 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1433 | MachineFunction &MF = MIRBuilder.getMF(); |
1434 | const auto *ST = static_cast<const SPIRVSubtarget *>(&MF.getSubtarget()); |
1435 | if (!ST->canUseExtension(E: SPIRV::Extension::SPV_KHR_shader_clock)) { |
1436 | std::string DiagMsg = std::string(Builtin->Name) + |
1437 | ": the builtin requires the following SPIR-V " |
1438 | "extension: SPV_KHR_shader_clock" ; |
1439 | report_fatal_error(reason: DiagMsg.c_str(), gen_crash_diag: false); |
1440 | } |
1441 | |
1442 | Register ResultReg = Call->ReturnRegister; |
1443 | |
1444 | // Deduce the `Scope` operand from the builtin function name. |
1445 | SPIRV::Scope::Scope ScopeArg = |
1446 | StringSwitch<SPIRV::Scope::Scope>(Builtin->Name) |
1447 | .EndsWith(S: "device" , Value: SPIRV::Scope::Scope::Device) |
1448 | .EndsWith(S: "work_group" , Value: SPIRV::Scope::Scope::Workgroup) |
1449 | .EndsWith(S: "sub_group" , Value: SPIRV::Scope::Scope::Subgroup); |
1450 | Register ScopeReg = buildConstantIntReg32(Val: ScopeArg, MIRBuilder, GR); |
1451 | |
1452 | MIRBuilder.buildInstr(Opcode: SPIRV::OpReadClockKHR) |
1453 | .addDef(RegNo: ResultReg) |
1454 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)) |
1455 | .addUse(RegNo: ScopeReg); |
1456 | |
1457 | return true; |
1458 | } |
1459 | |
1460 | // These queries ask for a single size_t result for a given dimension index, |
1461 | // e.g. size_t get_global_id(uint dimindex). In SPIR-V, the builtins |
1462 | // corresponding to these values are all vec3 types, so we need to extract the |
1463 | // correct index or return DefaultValue (0 or 1 depending on the query). We also |
1464 | // handle extending or truncating in case size_t does not match the expected |
1465 | // result type's bitwidth. |
1466 | // |
1467 | // For a constant index >= 3 we generate: |
1468 | // %res = OpConstant %SizeT DefaultValue |
1469 | // |
1470 | // For other indices we generate: |
1471 | // %g = OpVariable %ptr_V3_SizeT Input |
1472 | // OpDecorate %g BuiltIn XXX |
1473 | // OpDecorate %g LinkageAttributes "__spirv_BuiltInXXX" |
1474 | // OpDecorate %g Constant |
1475 | // %loadedVec = OpLoad %V3_SizeT %g |
1476 | // |
1477 | // Then, if the index is constant < 3, we generate: |
1478 | // %res = OpCompositeExtract %SizeT %loadedVec idx |
1479 | // If the index is dynamic, we generate: |
1480 | // %tmp = OpVectorExtractDynamic %SizeT %loadedVec %idx |
1481 | // %cmp = OpULessThan %bool %idx %const_3 |
1482 | // %res = OpSelect %SizeT %cmp %tmp %const_<DefaultValue> |
1483 | // |
1484 | // If the bitwidth of %res does not match the expected return type, we add an |
1485 | // extend or truncate. |
1486 | static bool genWorkgroupQuery(const SPIRV::IncomingCall *Call, |
1487 | MachineIRBuilder &MIRBuilder, |
1488 | SPIRVGlobalRegistry *GR, |
1489 | SPIRV::BuiltIn::BuiltIn BuiltinValue, |
1490 | uint64_t DefaultValue) { |
1491 | Register IndexRegister = Call->Arguments[0]; |
1492 | const unsigned ResultWidth = Call->ReturnType->getOperand(i: 1).getImm(); |
1493 | const unsigned PointerSize = GR->getPointerSize(); |
1494 | const SPIRVType *PointerSizeType = |
1495 | GR->getOrCreateSPIRVIntegerType(BitWidth: PointerSize, MIRBuilder); |
1496 | MachineRegisterInfo *MRI = MIRBuilder.getMRI(); |
1497 | auto IndexInstruction = getDefInstrMaybeConstant(ConstReg&: IndexRegister, MRI); |
1498 | |
1499 | // Set up the final register to do truncation or extension on at the end. |
1500 | Register ToTruncate = Call->ReturnRegister; |
1501 | |
1502 | // If the index is constant, we can statically determine if it is in range. |
1503 | bool IsConstantIndex = |
1504 | IndexInstruction->getOpcode() == TargetOpcode::G_CONSTANT; |
1505 | |
1506 | // If it's out of range (max dimension is 3), we can just return the constant |
1507 | // default value (0 or 1 depending on which query function). |
1508 | if (IsConstantIndex && getIConstVal(ConstReg: IndexRegister, MRI) >= 3) { |
1509 | Register DefaultReg = Call->ReturnRegister; |
1510 | if (PointerSize != ResultWidth) { |
1511 | DefaultReg = MRI->createGenericVirtualRegister(Ty: LLT::scalar(SizeInBits: PointerSize)); |
1512 | MRI->setRegClass(Reg: DefaultReg, RC: &SPIRV::iIDRegClass); |
1513 | GR->assignSPIRVTypeToVReg(Type: PointerSizeType, VReg: DefaultReg, |
1514 | MF: MIRBuilder.getMF()); |
1515 | ToTruncate = DefaultReg; |
1516 | } |
1517 | auto NewRegister = |
1518 | GR->buildConstantInt(Val: DefaultValue, MIRBuilder, SpvType: PointerSizeType, EmitIR: true); |
1519 | MIRBuilder.buildCopy(Res: DefaultReg, Op: NewRegister); |
1520 | } else { // If it could be in range, we need to load from the given builtin. |
1521 | auto Vec3Ty = |
1522 | GR->getOrCreateSPIRVVectorType(BaseType: PointerSizeType, NumElements: 3, MIRBuilder, EmitIR: true); |
1523 | Register LoadedVector = |
1524 | buildBuiltinVariableLoad(MIRBuilder, VariableType: Vec3Ty, GR, BuiltinValue, |
1525 | LLType: LLT::fixed_vector(NumElements: 3, ScalarSizeInBits: PointerSize)); |
1526 | // Set up the vreg to extract the result to (possibly a new temporary one). |
1527 | Register = Call->ReturnRegister; |
1528 | if (!IsConstantIndex || PointerSize != ResultWidth) { |
1529 | Extracted = MRI->createGenericVirtualRegister(Ty: LLT::scalar(SizeInBits: PointerSize)); |
1530 | MRI->setRegClass(Reg: Extracted, RC: &SPIRV::iIDRegClass); |
1531 | GR->assignSPIRVTypeToVReg(Type: PointerSizeType, VReg: Extracted, MF: MIRBuilder.getMF()); |
1532 | } |
1533 | // Use Intrinsic::spv_extractelt so dynamic vs static extraction is |
1534 | // handled later: extr = spv_extractelt LoadedVector, IndexRegister. |
1535 | MachineInstrBuilder = MIRBuilder.buildIntrinsic( |
1536 | ID: Intrinsic::spv_extractelt, Res: ArrayRef<Register>{Extracted}, HasSideEffects: true, isConvergent: false); |
1537 | ExtractInst.addUse(RegNo: LoadedVector).addUse(RegNo: IndexRegister); |
1538 | |
1539 | // If the index is dynamic, need check if it's < 3, and then use a select. |
1540 | if (!IsConstantIndex) { |
1541 | insertAssignInstr(Reg: Extracted, Ty: nullptr, SpirvTy: PointerSizeType, GR, MIB&: MIRBuilder, |
1542 | MRI&: *MRI); |
1543 | |
1544 | auto IndexType = GR->getSPIRVTypeForVReg(VReg: IndexRegister); |
1545 | auto BoolType = GR->getOrCreateSPIRVBoolType(MIRBuilder, EmitIR: true); |
1546 | |
1547 | Register CompareRegister = |
1548 | MRI->createGenericVirtualRegister(Ty: LLT::scalar(SizeInBits: 1)); |
1549 | MRI->setRegClass(Reg: CompareRegister, RC: &SPIRV::iIDRegClass); |
1550 | GR->assignSPIRVTypeToVReg(Type: BoolType, VReg: CompareRegister, MF: MIRBuilder.getMF()); |
1551 | |
1552 | // Use G_ICMP to check if idxVReg < 3. |
1553 | MIRBuilder.buildICmp( |
1554 | Pred: CmpInst::ICMP_ULT, Res: CompareRegister, Op0: IndexRegister, |
1555 | Op1: GR->buildConstantInt(Val: 3, MIRBuilder, SpvType: IndexType, EmitIR: true)); |
1556 | |
1557 | // Get constant for the default value (0 or 1 depending on which |
1558 | // function). |
1559 | Register DefaultRegister = |
1560 | GR->buildConstantInt(Val: DefaultValue, MIRBuilder, SpvType: PointerSizeType, EmitIR: true); |
1561 | |
1562 | // Get a register for the selection result (possibly a new temporary one). |
1563 | Register SelectionResult = Call->ReturnRegister; |
1564 | if (PointerSize != ResultWidth) { |
1565 | SelectionResult = |
1566 | MRI->createGenericVirtualRegister(Ty: LLT::scalar(SizeInBits: PointerSize)); |
1567 | MRI->setRegClass(Reg: SelectionResult, RC: &SPIRV::iIDRegClass); |
1568 | GR->assignSPIRVTypeToVReg(Type: PointerSizeType, VReg: SelectionResult, |
1569 | MF: MIRBuilder.getMF()); |
1570 | } |
1571 | // Create the final G_SELECT to return the extracted value or the default. |
1572 | MIRBuilder.buildSelect(Res: SelectionResult, Tst: CompareRegister, Op0: Extracted, |
1573 | Op1: DefaultRegister); |
1574 | ToTruncate = SelectionResult; |
1575 | } else { |
1576 | ToTruncate = Extracted; |
1577 | } |
1578 | } |
1579 | // Alter the result's bitwidth if it does not match the SizeT value extracted. |
1580 | if (PointerSize != ResultWidth) |
1581 | MIRBuilder.buildZExtOrTrunc(Res: Call->ReturnRegister, Op: ToTruncate); |
1582 | return true; |
1583 | } |
1584 | |
1585 | static bool generateBuiltinVar(const SPIRV::IncomingCall *Call, |
1586 | MachineIRBuilder &MIRBuilder, |
1587 | SPIRVGlobalRegistry *GR) { |
1588 | // Lookup the builtin variable record. |
1589 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1590 | SPIRV::BuiltIn::BuiltIn Value = |
1591 | SPIRV::lookupGetBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Value; |
1592 | |
1593 | if (Value == SPIRV::BuiltIn::GlobalInvocationId) |
1594 | return genWorkgroupQuery(Call, MIRBuilder, GR, BuiltinValue: Value, DefaultValue: 0); |
1595 | |
1596 | // Build a load instruction for the builtin variable. |
1597 | unsigned BitWidth = GR->getScalarOrVectorBitWidth(Type: Call->ReturnType); |
1598 | LLT LLType; |
1599 | if (Call->ReturnType->getOpcode() == SPIRV::OpTypeVector) |
1600 | LLType = |
1601 | LLT::fixed_vector(NumElements: Call->ReturnType->getOperand(i: 2).getImm(), ScalarSizeInBits: BitWidth); |
1602 | else |
1603 | LLType = LLT::scalar(SizeInBits: BitWidth); |
1604 | |
1605 | return buildBuiltinVariableLoad(MIRBuilder, VariableType: Call->ReturnType, GR, BuiltinValue: Value, |
1606 | LLType, Reg: Call->ReturnRegister); |
1607 | } |
1608 | |
1609 | static bool generateAtomicInst(const SPIRV::IncomingCall *Call, |
1610 | MachineIRBuilder &MIRBuilder, |
1611 | SPIRVGlobalRegistry *GR) { |
1612 | // Lookup the instruction opcode in the TableGen records. |
1613 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1614 | unsigned Opcode = |
1615 | SPIRV::lookupNativeBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Opcode; |
1616 | |
1617 | switch (Opcode) { |
1618 | case SPIRV::OpStore: |
1619 | return buildAtomicInitInst(Call, MIRBuilder); |
1620 | case SPIRV::OpAtomicLoad: |
1621 | return buildAtomicLoadInst(Call, MIRBuilder, GR); |
1622 | case SPIRV::OpAtomicStore: |
1623 | return buildAtomicStoreInst(Call, MIRBuilder, GR); |
1624 | case SPIRV::OpAtomicCompareExchange: |
1625 | case SPIRV::OpAtomicCompareExchangeWeak: |
1626 | return buildAtomicCompareExchangeInst(Call, Builtin, Opcode, MIRBuilder, |
1627 | GR); |
1628 | case SPIRV::OpAtomicIAdd: |
1629 | case SPIRV::OpAtomicISub: |
1630 | case SPIRV::OpAtomicOr: |
1631 | case SPIRV::OpAtomicXor: |
1632 | case SPIRV::OpAtomicAnd: |
1633 | case SPIRV::OpAtomicExchange: |
1634 | return buildAtomicRMWInst(Call, Opcode, MIRBuilder, GR); |
1635 | case SPIRV::OpMemoryBarrier: |
1636 | return buildBarrierInst(Call, Opcode: SPIRV::OpMemoryBarrier, MIRBuilder, GR); |
1637 | case SPIRV::OpAtomicFlagTestAndSet: |
1638 | case SPIRV::OpAtomicFlagClear: |
1639 | return buildAtomicFlagInst(Call, Opcode, MIRBuilder, GR); |
1640 | default: |
1641 | if (Call->isSpirvOp()) |
1642 | return buildOpFromWrapper(MIRBuilder, Opcode, Call, |
1643 | TypeReg: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
1644 | return false; |
1645 | } |
1646 | } |
1647 | |
1648 | static bool generateAtomicFloatingInst(const SPIRV::IncomingCall *Call, |
1649 | MachineIRBuilder &MIRBuilder, |
1650 | SPIRVGlobalRegistry *GR) { |
1651 | // Lookup the instruction opcode in the TableGen records. |
1652 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1653 | unsigned Opcode = SPIRV::lookupAtomicFloatingBuiltin(Name: Builtin->Name)->Opcode; |
1654 | |
1655 | switch (Opcode) { |
1656 | case SPIRV::OpAtomicFAddEXT: |
1657 | case SPIRV::OpAtomicFMinEXT: |
1658 | case SPIRV::OpAtomicFMaxEXT: |
1659 | return buildAtomicFloatingRMWInst(Call, Opcode, MIRBuilder, GR); |
1660 | default: |
1661 | return false; |
1662 | } |
1663 | } |
1664 | |
1665 | static bool generateBarrierInst(const SPIRV::IncomingCall *Call, |
1666 | MachineIRBuilder &MIRBuilder, |
1667 | SPIRVGlobalRegistry *GR) { |
1668 | // Lookup the instruction opcode in the TableGen records. |
1669 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1670 | unsigned Opcode = |
1671 | SPIRV::lookupNativeBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Opcode; |
1672 | |
1673 | return buildBarrierInst(Call, Opcode, MIRBuilder, GR); |
1674 | } |
1675 | |
1676 | static bool generateCastToPtrInst(const SPIRV::IncomingCall *Call, |
1677 | MachineIRBuilder &MIRBuilder, |
1678 | SPIRVGlobalRegistry *GR) { |
1679 | // Lookup the instruction opcode in the TableGen records. |
1680 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1681 | unsigned Opcode = |
1682 | SPIRV::lookupNativeBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Opcode; |
1683 | |
1684 | if (Opcode == SPIRV::OpGenericCastToPtrExplicit) { |
1685 | SPIRV::StorageClass::StorageClass ResSC = |
1686 | GR->getPointerStorageClass(VReg: Call->ReturnRegister); |
1687 | if (!isGenericCastablePtr(SC: ResSC)) |
1688 | return false; |
1689 | |
1690 | MIRBuilder.buildInstr(Opcode) |
1691 | .addDef(RegNo: Call->ReturnRegister) |
1692 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)) |
1693 | .addUse(RegNo: Call->Arguments[0]) |
1694 | .addImm(Val: ResSC); |
1695 | } else { |
1696 | MIRBuilder.buildInstr(Opcode: TargetOpcode::G_ADDRSPACE_CAST) |
1697 | .addDef(RegNo: Call->ReturnRegister) |
1698 | .addUse(RegNo: Call->Arguments[0]); |
1699 | } |
1700 | return true; |
1701 | } |
1702 | |
1703 | static bool generateDotOrFMulInst(const StringRef DemangledCall, |
1704 | const SPIRV::IncomingCall *Call, |
1705 | MachineIRBuilder &MIRBuilder, |
1706 | SPIRVGlobalRegistry *GR) { |
1707 | if (Call->isSpirvOp()) |
1708 | return buildOpFromWrapper(MIRBuilder, Opcode: SPIRV::OpDot, Call, |
1709 | TypeReg: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
1710 | |
1711 | bool IsVec = GR->getSPIRVTypeForVReg(VReg: Call->Arguments[0])->getOpcode() == |
1712 | SPIRV::OpTypeVector; |
1713 | // Use OpDot only in case of vector args and OpFMul in case of scalar args. |
1714 | uint32_t OC = IsVec ? SPIRV::OpDot : SPIRV::OpFMulS; |
1715 | bool IsSwapReq = false; |
1716 | |
1717 | const auto *ST = |
1718 | static_cast<const SPIRVSubtarget *>(&MIRBuilder.getMF().getSubtarget()); |
1719 | if (GR->isScalarOrVectorOfType(VReg: Call->ReturnRegister, TypeOpcode: SPIRV::OpTypeInt) && |
1720 | (ST->canUseExtension(E: SPIRV::Extension::SPV_KHR_integer_dot_product) || |
1721 | ST->isAtLeastSPIRVVer(VerToCompareTo: VersionTuple(1, 6)))) { |
1722 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1723 | const SPIRV::IntegerDotProductBuiltin *IntDot = |
1724 | SPIRV::lookupIntegerDotProductBuiltin(Name: Builtin->Name); |
1725 | if (IntDot) { |
1726 | OC = IntDot->Opcode; |
1727 | IsSwapReq = IntDot->IsSwapReq; |
1728 | } else if (IsVec) { |
1729 | // Handling "dot" and "dot_acc_sat" builtins which use vectors of |
1730 | // integers. |
1731 | LLVMContext &Ctx = MIRBuilder.getContext(); |
1732 | SmallVector<StringRef, 10> TypeStrs; |
1733 | SPIRV::parseBuiltinTypeStr(BuiltinArgsTypeStrs&: TypeStrs, DemangledCall, Ctx); |
1734 | bool IsFirstSigned = TypeStrs[0].trim()[0] != 'u'; |
1735 | bool IsSecondSigned = TypeStrs[1].trim()[0] != 'u'; |
1736 | |
1737 | if (Call->BuiltinName == "dot" ) { |
1738 | if (IsFirstSigned && IsSecondSigned) |
1739 | OC = SPIRV::OpSDot; |
1740 | else if (!IsFirstSigned && !IsSecondSigned) |
1741 | OC = SPIRV::OpUDot; |
1742 | else { |
1743 | OC = SPIRV::OpSUDot; |
1744 | if (!IsFirstSigned) |
1745 | IsSwapReq = true; |
1746 | } |
1747 | } else if (Call->BuiltinName == "dot_acc_sat" ) { |
1748 | if (IsFirstSigned && IsSecondSigned) |
1749 | OC = SPIRV::OpSDotAccSat; |
1750 | else if (!IsFirstSigned && !IsSecondSigned) |
1751 | OC = SPIRV::OpUDotAccSat; |
1752 | else { |
1753 | OC = SPIRV::OpSUDotAccSat; |
1754 | if (!IsFirstSigned) |
1755 | IsSwapReq = true; |
1756 | } |
1757 | } |
1758 | } |
1759 | } |
1760 | |
1761 | MachineInstrBuilder MIB = MIRBuilder.buildInstr(Opcode: OC) |
1762 | .addDef(RegNo: Call->ReturnRegister) |
1763 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
1764 | |
1765 | if (IsSwapReq) { |
1766 | MIB.addUse(RegNo: Call->Arguments[1]); |
1767 | MIB.addUse(RegNo: Call->Arguments[0]); |
1768 | // needed for dot_acc_sat* builtins |
1769 | for (size_t i = 2; i < Call->Arguments.size(); ++i) |
1770 | MIB.addUse(RegNo: Call->Arguments[i]); |
1771 | } else { |
1772 | for (size_t i = 0; i < Call->Arguments.size(); ++i) |
1773 | MIB.addUse(RegNo: Call->Arguments[i]); |
1774 | } |
1775 | |
1776 | // Add Packed Vector Format for Integer dot product builtins if arguments are |
1777 | // scalar |
1778 | if (!IsVec && OC != SPIRV::OpFMulS) |
1779 | MIB.addImm(Val: 0); |
1780 | |
1781 | return true; |
1782 | } |
1783 | |
1784 | static bool generateWaveInst(const SPIRV::IncomingCall *Call, |
1785 | MachineIRBuilder &MIRBuilder, |
1786 | SPIRVGlobalRegistry *GR) { |
1787 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1788 | SPIRV::BuiltIn::BuiltIn Value = |
1789 | SPIRV::lookupGetBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Value; |
1790 | |
1791 | // For now, we only support a single Wave intrinsic with a single return type. |
1792 | assert(Call->ReturnType->getOpcode() == SPIRV::OpTypeInt); |
1793 | LLT LLType = LLT::scalar(SizeInBits: GR->getScalarOrVectorBitWidth(Type: Call->ReturnType)); |
1794 | |
1795 | return buildBuiltinVariableLoad( |
1796 | MIRBuilder, VariableType: Call->ReturnType, GR, BuiltinValue: Value, LLType, Reg: Call->ReturnRegister, |
1797 | /* isConst= */ false, /* hasLinkageTy= */ false); |
1798 | } |
1799 | |
1800 | // We expect a builtin |
1801 | // Name(ptr sret([RetType]) %result, Type %operand1, Type %operand1) |
1802 | // where %result is a pointer to where the result of the builtin execution |
1803 | // is to be stored, and generate the following instructions: |
1804 | // Res = Opcode RetType Operand1 Operand1 |
1805 | // OpStore RetVariable Res |
1806 | static bool generateICarryBorrowInst(const SPIRV::IncomingCall *Call, |
1807 | MachineIRBuilder &MIRBuilder, |
1808 | SPIRVGlobalRegistry *GR) { |
1809 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1810 | unsigned Opcode = |
1811 | SPIRV::lookupNativeBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Opcode; |
1812 | |
1813 | Register SRetReg = Call->Arguments[0]; |
1814 | SPIRVType *PtrRetType = GR->getSPIRVTypeForVReg(VReg: SRetReg); |
1815 | SPIRVType *RetType = GR->getPointeeType(PtrType: PtrRetType); |
1816 | if (!RetType) |
1817 | report_fatal_error(reason: "The first parameter must be a pointer" ); |
1818 | if (RetType->getOpcode() != SPIRV::OpTypeStruct) |
1819 | report_fatal_error(reason: "Expected struct type result for the arithmetic with " |
1820 | "overflow builtins" ); |
1821 | |
1822 | SPIRVType *OpType1 = GR->getSPIRVTypeForVReg(VReg: Call->Arguments[1]); |
1823 | SPIRVType *OpType2 = GR->getSPIRVTypeForVReg(VReg: Call->Arguments[2]); |
1824 | if (!OpType1 || !OpType2 || OpType1 != OpType2) |
1825 | report_fatal_error(reason: "Operands must have the same type" ); |
1826 | if (OpType1->getOpcode() == SPIRV::OpTypeVector) |
1827 | switch (Opcode) { |
1828 | case SPIRV::OpIAddCarryS: |
1829 | Opcode = SPIRV::OpIAddCarryV; |
1830 | break; |
1831 | case SPIRV::OpISubBorrowS: |
1832 | Opcode = SPIRV::OpISubBorrowV; |
1833 | break; |
1834 | } |
1835 | |
1836 | MachineRegisterInfo *MRI = MIRBuilder.getMRI(); |
1837 | Register ResReg = MRI->createVirtualRegister(RegClass: &SPIRV::iIDRegClass); |
1838 | if (const TargetRegisterClass *DstRC = |
1839 | MRI->getRegClassOrNull(Reg: Call->Arguments[1])) { |
1840 | MRI->setRegClass(Reg: ResReg, RC: DstRC); |
1841 | MRI->setType(VReg: ResReg, Ty: MRI->getType(Reg: Call->Arguments[1])); |
1842 | } else { |
1843 | MRI->setType(VReg: ResReg, Ty: LLT::scalar(SizeInBits: 64)); |
1844 | } |
1845 | GR->assignSPIRVTypeToVReg(Type: RetType, VReg: ResReg, MF: MIRBuilder.getMF()); |
1846 | MIRBuilder.buildInstr(Opcode) |
1847 | .addDef(RegNo: ResReg) |
1848 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: RetType)) |
1849 | .addUse(RegNo: Call->Arguments[1]) |
1850 | .addUse(RegNo: Call->Arguments[2]); |
1851 | MIRBuilder.buildInstr(Opcode: SPIRV::OpStore).addUse(RegNo: SRetReg).addUse(RegNo: ResReg); |
1852 | return true; |
1853 | } |
1854 | |
1855 | static bool generateGetQueryInst(const SPIRV::IncomingCall *Call, |
1856 | MachineIRBuilder &MIRBuilder, |
1857 | SPIRVGlobalRegistry *GR) { |
1858 | // Lookup the builtin record. |
1859 | SPIRV::BuiltIn::BuiltIn Value = |
1860 | SPIRV::lookupGetBuiltin(Name: Call->Builtin->Name, Set: Call->Builtin->Set)->Value; |
1861 | const bool IsDefaultOne = (Value == SPIRV::BuiltIn::GlobalSize || |
1862 | Value == SPIRV::BuiltIn::NumWorkgroups || |
1863 | Value == SPIRV::BuiltIn::WorkgroupSize || |
1864 | Value == SPIRV::BuiltIn::EnqueuedWorkgroupSize); |
1865 | return genWorkgroupQuery(Call, MIRBuilder, GR, BuiltinValue: Value, DefaultValue: IsDefaultOne ? 1 : 0); |
1866 | } |
1867 | |
1868 | static bool generateImageSizeQueryInst(const SPIRV::IncomingCall *Call, |
1869 | MachineIRBuilder &MIRBuilder, |
1870 | SPIRVGlobalRegistry *GR) { |
1871 | // Lookup the image size query component number in the TableGen records. |
1872 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1873 | uint32_t Component = |
1874 | SPIRV::lookupImageQueryBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Component; |
1875 | // Query result may either be a vector or a scalar. If return type is not a |
1876 | // vector, expect only a single size component. Otherwise get the number of |
1877 | // expected components. |
1878 | unsigned NumExpectedRetComponents = |
1879 | Call->ReturnType->getOpcode() == SPIRV::OpTypeVector |
1880 | ? Call->ReturnType->getOperand(i: 2).getImm() |
1881 | : 1; |
1882 | // Get the actual number of query result/size components. |
1883 | SPIRVType *ImgType = GR->getSPIRVTypeForVReg(VReg: Call->Arguments[0]); |
1884 | unsigned NumActualRetComponents = getNumSizeComponents(imgType: ImgType); |
1885 | Register QueryResult = Call->ReturnRegister; |
1886 | SPIRVType *QueryResultType = Call->ReturnType; |
1887 | if (NumExpectedRetComponents != NumActualRetComponents) { |
1888 | unsigned Bitwidth = Call->ReturnType->getOpcode() == SPIRV::OpTypeInt |
1889 | ? Call->ReturnType->getOperand(i: 1).getImm() |
1890 | : 32; |
1891 | QueryResult = MIRBuilder.getMRI()->createGenericVirtualRegister( |
1892 | Ty: LLT::fixed_vector(NumElements: NumActualRetComponents, ScalarSizeInBits: Bitwidth)); |
1893 | MIRBuilder.getMRI()->setRegClass(Reg: QueryResult, RC: &SPIRV::vIDRegClass); |
1894 | SPIRVType *IntTy = GR->getOrCreateSPIRVIntegerType(BitWidth: Bitwidth, MIRBuilder); |
1895 | QueryResultType = GR->getOrCreateSPIRVVectorType( |
1896 | BaseType: IntTy, NumElements: NumActualRetComponents, MIRBuilder, EmitIR: true); |
1897 | GR->assignSPIRVTypeToVReg(Type: QueryResultType, VReg: QueryResult, MF: MIRBuilder.getMF()); |
1898 | } |
1899 | bool IsDimBuf = ImgType->getOperand(i: 2).getImm() == SPIRV::Dim::DIM_Buffer; |
1900 | unsigned Opcode = |
1901 | IsDimBuf ? SPIRV::OpImageQuerySize : SPIRV::OpImageQuerySizeLod; |
1902 | auto MIB = MIRBuilder.buildInstr(Opcode) |
1903 | .addDef(RegNo: QueryResult) |
1904 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: QueryResultType)) |
1905 | .addUse(RegNo: Call->Arguments[0]); |
1906 | if (!IsDimBuf) |
1907 | MIB.addUse(RegNo: buildConstantIntReg32(Val: 0, MIRBuilder, GR)); // Lod id. |
1908 | if (NumExpectedRetComponents == NumActualRetComponents) |
1909 | return true; |
1910 | if (NumExpectedRetComponents == 1) { |
1911 | // Only 1 component is expected, build OpCompositeExtract instruction. |
1912 | unsigned = |
1913 | Component == 3 ? NumActualRetComponents - 1 : Component; |
1914 | assert(ExtractedComposite < NumActualRetComponents && |
1915 | "Invalid composite index!" ); |
1916 | Register TypeReg = GR->getSPIRVTypeID(SpirvType: Call->ReturnType); |
1917 | SPIRVType *NewType = nullptr; |
1918 | if (QueryResultType->getOpcode() == SPIRV::OpTypeVector) { |
1919 | Register NewTypeReg = QueryResultType->getOperand(i: 1).getReg(); |
1920 | if (TypeReg != NewTypeReg && |
1921 | (NewType = GR->getSPIRVTypeForVReg(VReg: NewTypeReg)) != nullptr) |
1922 | TypeReg = NewTypeReg; |
1923 | } |
1924 | MIRBuilder.buildInstr(Opcode: SPIRV::OpCompositeExtract) |
1925 | .addDef(RegNo: Call->ReturnRegister) |
1926 | .addUse(RegNo: TypeReg) |
1927 | .addUse(RegNo: QueryResult) |
1928 | .addImm(Val: ExtractedComposite); |
1929 | if (NewType != nullptr) |
1930 | insertAssignInstr(Reg: Call->ReturnRegister, Ty: nullptr, SpirvTy: NewType, GR, MIB&: MIRBuilder, |
1931 | MRI&: MIRBuilder.getMF().getRegInfo()); |
1932 | } else { |
1933 | // More than 1 component is expected, fill a new vector. |
1934 | auto MIB = MIRBuilder.buildInstr(Opcode: SPIRV::OpVectorShuffle) |
1935 | .addDef(RegNo: Call->ReturnRegister) |
1936 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)) |
1937 | .addUse(RegNo: QueryResult) |
1938 | .addUse(RegNo: QueryResult); |
1939 | for (unsigned i = 0; i < NumExpectedRetComponents; ++i) |
1940 | MIB.addImm(Val: i < NumActualRetComponents ? i : 0xffffffff); |
1941 | } |
1942 | return true; |
1943 | } |
1944 | |
1945 | static bool generateImageMiscQueryInst(const SPIRV::IncomingCall *Call, |
1946 | MachineIRBuilder &MIRBuilder, |
1947 | SPIRVGlobalRegistry *GR) { |
1948 | assert(Call->ReturnType->getOpcode() == SPIRV::OpTypeInt && |
1949 | "Image samples query result must be of int type!" ); |
1950 | |
1951 | // Lookup the instruction opcode in the TableGen records. |
1952 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
1953 | unsigned Opcode = |
1954 | SPIRV::lookupNativeBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Opcode; |
1955 | |
1956 | Register Image = Call->Arguments[0]; |
1957 | SPIRV::Dim::Dim ImageDimensionality = static_cast<SPIRV::Dim::Dim>( |
1958 | GR->getSPIRVTypeForVReg(VReg: Image)->getOperand(i: 2).getImm()); |
1959 | (void)ImageDimensionality; |
1960 | |
1961 | switch (Opcode) { |
1962 | case SPIRV::OpImageQuerySamples: |
1963 | assert(ImageDimensionality == SPIRV::Dim::DIM_2D && |
1964 | "Image must be of 2D dimensionality" ); |
1965 | break; |
1966 | case SPIRV::OpImageQueryLevels: |
1967 | assert((ImageDimensionality == SPIRV::Dim::DIM_1D || |
1968 | ImageDimensionality == SPIRV::Dim::DIM_2D || |
1969 | ImageDimensionality == SPIRV::Dim::DIM_3D || |
1970 | ImageDimensionality == SPIRV::Dim::DIM_Cube) && |
1971 | "Image must be of 1D/2D/3D/Cube dimensionality" ); |
1972 | break; |
1973 | } |
1974 | |
1975 | MIRBuilder.buildInstr(Opcode) |
1976 | .addDef(RegNo: Call->ReturnRegister) |
1977 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)) |
1978 | .addUse(RegNo: Image); |
1979 | return true; |
1980 | } |
1981 | |
1982 | // TODO: Move to TableGen. |
1983 | static SPIRV::SamplerAddressingMode::SamplerAddressingMode |
1984 | getSamplerAddressingModeFromBitmask(unsigned Bitmask) { |
1985 | switch (Bitmask & SPIRV::CLK_ADDRESS_MODE_MASK) { |
1986 | case SPIRV::CLK_ADDRESS_CLAMP: |
1987 | return SPIRV::SamplerAddressingMode::Clamp; |
1988 | case SPIRV::CLK_ADDRESS_CLAMP_TO_EDGE: |
1989 | return SPIRV::SamplerAddressingMode::ClampToEdge; |
1990 | case SPIRV::CLK_ADDRESS_REPEAT: |
1991 | return SPIRV::SamplerAddressingMode::Repeat; |
1992 | case SPIRV::CLK_ADDRESS_MIRRORED_REPEAT: |
1993 | return SPIRV::SamplerAddressingMode::RepeatMirrored; |
1994 | case SPIRV::CLK_ADDRESS_NONE: |
1995 | return SPIRV::SamplerAddressingMode::None; |
1996 | default: |
1997 | report_fatal_error(reason: "Unknown CL address mode" ); |
1998 | } |
1999 | } |
2000 | |
2001 | static unsigned getSamplerParamFromBitmask(unsigned Bitmask) { |
2002 | return (Bitmask & SPIRV::CLK_NORMALIZED_COORDS_TRUE) ? 1 : 0; |
2003 | } |
2004 | |
2005 | static SPIRV::SamplerFilterMode::SamplerFilterMode |
2006 | getSamplerFilterModeFromBitmask(unsigned Bitmask) { |
2007 | if (Bitmask & SPIRV::CLK_FILTER_LINEAR) |
2008 | return SPIRV::SamplerFilterMode::Linear; |
2009 | if (Bitmask & SPIRV::CLK_FILTER_NEAREST) |
2010 | return SPIRV::SamplerFilterMode::Nearest; |
2011 | return SPIRV::SamplerFilterMode::Nearest; |
2012 | } |
2013 | |
2014 | static bool generateReadImageInst(const StringRef DemangledCall, |
2015 | const SPIRV::IncomingCall *Call, |
2016 | MachineIRBuilder &MIRBuilder, |
2017 | SPIRVGlobalRegistry *GR) { |
2018 | if (Call->isSpirvOp()) |
2019 | return buildOpFromWrapper(MIRBuilder, Opcode: SPIRV::OpImageRead, Call, |
2020 | TypeReg: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
2021 | Register Image = Call->Arguments[0]; |
2022 | MachineRegisterInfo *MRI = MIRBuilder.getMRI(); |
2023 | bool HasOclSampler = DemangledCall.contains_insensitive(Other: "ocl_sampler" ); |
2024 | bool HasMsaa = DemangledCall.contains_insensitive(Other: "msaa" ); |
2025 | if (HasOclSampler) { |
2026 | Register Sampler = Call->Arguments[1]; |
2027 | |
2028 | if (!GR->isScalarOfType(VReg: Sampler, TypeOpcode: SPIRV::OpTypeSampler) && |
2029 | getDefInstrMaybeConstant(ConstReg&: Sampler, MRI)->getOperand(i: 1).isCImm()) { |
2030 | uint64_t SamplerMask = getIConstVal(ConstReg: Sampler, MRI); |
2031 | Sampler = GR->buildConstantSampler( |
2032 | Res: Register(), AddrMode: getSamplerAddressingModeFromBitmask(Bitmask: SamplerMask), |
2033 | Param: getSamplerParamFromBitmask(Bitmask: SamplerMask), |
2034 | FilerMode: getSamplerFilterModeFromBitmask(Bitmask: SamplerMask), MIRBuilder); |
2035 | } |
2036 | SPIRVType *ImageType = GR->getSPIRVTypeForVReg(VReg: Image); |
2037 | SPIRVType *SampledImageType = |
2038 | GR->getOrCreateOpTypeSampledImage(ImageType, MIRBuilder); |
2039 | Register SampledImage = MRI->createVirtualRegister(RegClass: &SPIRV::iIDRegClass); |
2040 | |
2041 | MIRBuilder.buildInstr(Opcode: SPIRV::OpSampledImage) |
2042 | .addDef(RegNo: SampledImage) |
2043 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: SampledImageType)) |
2044 | .addUse(RegNo: Image) |
2045 | .addUse(RegNo: Sampler); |
2046 | |
2047 | Register Lod = GR->buildConstantFP(Val: APFloat::getZero(Sem: APFloat::IEEEsingle()), |
2048 | MIRBuilder); |
2049 | |
2050 | if (Call->ReturnType->getOpcode() != SPIRV::OpTypeVector) { |
2051 | SPIRVType *TempType = |
2052 | GR->getOrCreateSPIRVVectorType(BaseType: Call->ReturnType, NumElements: 4, MIRBuilder, EmitIR: true); |
2053 | Register TempRegister = |
2054 | MRI->createGenericVirtualRegister(Ty: GR->getRegType(SpvType: TempType)); |
2055 | MRI->setRegClass(Reg: TempRegister, RC: GR->getRegClass(SpvType: TempType)); |
2056 | GR->assignSPIRVTypeToVReg(Type: TempType, VReg: TempRegister, MF: MIRBuilder.getMF()); |
2057 | MIRBuilder.buildInstr(Opcode: SPIRV::OpImageSampleExplicitLod) |
2058 | .addDef(RegNo: TempRegister) |
2059 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: TempType)) |
2060 | .addUse(RegNo: SampledImage) |
2061 | .addUse(RegNo: Call->Arguments[2]) // Coordinate. |
2062 | .addImm(Val: SPIRV::ImageOperand::Lod) |
2063 | .addUse(RegNo: Lod); |
2064 | MIRBuilder.buildInstr(Opcode: SPIRV::OpCompositeExtract) |
2065 | .addDef(RegNo: Call->ReturnRegister) |
2066 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)) |
2067 | .addUse(RegNo: TempRegister) |
2068 | .addImm(Val: 0); |
2069 | } else { |
2070 | MIRBuilder.buildInstr(Opcode: SPIRV::OpImageSampleExplicitLod) |
2071 | .addDef(RegNo: Call->ReturnRegister) |
2072 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)) |
2073 | .addUse(RegNo: SampledImage) |
2074 | .addUse(RegNo: Call->Arguments[2]) // Coordinate. |
2075 | .addImm(Val: SPIRV::ImageOperand::Lod) |
2076 | .addUse(RegNo: Lod); |
2077 | } |
2078 | } else if (HasMsaa) { |
2079 | MIRBuilder.buildInstr(Opcode: SPIRV::OpImageRead) |
2080 | .addDef(RegNo: Call->ReturnRegister) |
2081 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)) |
2082 | .addUse(RegNo: Image) |
2083 | .addUse(RegNo: Call->Arguments[1]) // Coordinate. |
2084 | .addImm(Val: SPIRV::ImageOperand::Sample) |
2085 | .addUse(RegNo: Call->Arguments[2]); |
2086 | } else { |
2087 | MIRBuilder.buildInstr(Opcode: SPIRV::OpImageRead) |
2088 | .addDef(RegNo: Call->ReturnRegister) |
2089 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)) |
2090 | .addUse(RegNo: Image) |
2091 | .addUse(RegNo: Call->Arguments[1]); // Coordinate. |
2092 | } |
2093 | return true; |
2094 | } |
2095 | |
2096 | static bool generateWriteImageInst(const SPIRV::IncomingCall *Call, |
2097 | MachineIRBuilder &MIRBuilder, |
2098 | SPIRVGlobalRegistry *GR) { |
2099 | if (Call->isSpirvOp()) |
2100 | return buildOpFromWrapper(MIRBuilder, Opcode: SPIRV::OpImageWrite, Call, |
2101 | TypeReg: Register(0)); |
2102 | MIRBuilder.buildInstr(Opcode: SPIRV::OpImageWrite) |
2103 | .addUse(RegNo: Call->Arguments[0]) // Image. |
2104 | .addUse(RegNo: Call->Arguments[1]) // Coordinate. |
2105 | .addUse(RegNo: Call->Arguments[2]); // Texel. |
2106 | return true; |
2107 | } |
2108 | |
2109 | static bool generateSampleImageInst(const StringRef DemangledCall, |
2110 | const SPIRV::IncomingCall *Call, |
2111 | MachineIRBuilder &MIRBuilder, |
2112 | SPIRVGlobalRegistry *GR) { |
2113 | MachineRegisterInfo *MRI = MIRBuilder.getMRI(); |
2114 | if (Call->Builtin->Name.contains_insensitive( |
2115 | Other: "__translate_sampler_initializer" )) { |
2116 | // Build sampler literal. |
2117 | uint64_t Bitmask = getIConstVal(ConstReg: Call->Arguments[0], MRI); |
2118 | Register Sampler = GR->buildConstantSampler( |
2119 | Res: Call->ReturnRegister, AddrMode: getSamplerAddressingModeFromBitmask(Bitmask), |
2120 | Param: getSamplerParamFromBitmask(Bitmask), |
2121 | FilerMode: getSamplerFilterModeFromBitmask(Bitmask), MIRBuilder); |
2122 | return Sampler.isValid(); |
2123 | } else if (Call->Builtin->Name.contains_insensitive(Other: "__spirv_SampledImage" )) { |
2124 | // Create OpSampledImage. |
2125 | Register Image = Call->Arguments[0]; |
2126 | SPIRVType *ImageType = GR->getSPIRVTypeForVReg(VReg: Image); |
2127 | SPIRVType *SampledImageType = |
2128 | GR->getOrCreateOpTypeSampledImage(ImageType, MIRBuilder); |
2129 | Register SampledImage = |
2130 | Call->ReturnRegister.isValid() |
2131 | ? Call->ReturnRegister |
2132 | : MRI->createVirtualRegister(RegClass: &SPIRV::iIDRegClass); |
2133 | MIRBuilder.buildInstr(Opcode: SPIRV::OpSampledImage) |
2134 | .addDef(RegNo: SampledImage) |
2135 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: SampledImageType)) |
2136 | .addUse(RegNo: Image) |
2137 | .addUse(RegNo: Call->Arguments[1]); // Sampler. |
2138 | return true; |
2139 | } else if (Call->Builtin->Name.contains_insensitive( |
2140 | Other: "__spirv_ImageSampleExplicitLod" )) { |
2141 | // Sample an image using an explicit level of detail. |
2142 | std::string ReturnType = DemangledCall.str(); |
2143 | if (DemangledCall.contains(Other: "_R" )) { |
2144 | ReturnType = ReturnType.substr(pos: ReturnType.find(s: "_R" ) + 2); |
2145 | ReturnType = ReturnType.substr(pos: 0, n: ReturnType.find(c: '(')); |
2146 | } |
2147 | SPIRVType *Type = |
2148 | Call->ReturnType |
2149 | ? Call->ReturnType |
2150 | : GR->getOrCreateSPIRVTypeByName(TypeStr: ReturnType, MIRBuilder, EmitIR: true); |
2151 | if (!Type) { |
2152 | std::string DiagMsg = |
2153 | "Unable to recognize SPIRV type name: " + ReturnType; |
2154 | report_fatal_error(reason: DiagMsg.c_str()); |
2155 | } |
2156 | MIRBuilder.buildInstr(Opcode: SPIRV::OpImageSampleExplicitLod) |
2157 | .addDef(RegNo: Call->ReturnRegister) |
2158 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Type)) |
2159 | .addUse(RegNo: Call->Arguments[0]) // Image. |
2160 | .addUse(RegNo: Call->Arguments[1]) // Coordinate. |
2161 | .addImm(Val: SPIRV::ImageOperand::Lod) |
2162 | .addUse(RegNo: Call->Arguments[3]); |
2163 | return true; |
2164 | } |
2165 | return false; |
2166 | } |
2167 | |
2168 | static bool generateSelectInst(const SPIRV::IncomingCall *Call, |
2169 | MachineIRBuilder &MIRBuilder) { |
2170 | MIRBuilder.buildSelect(Res: Call->ReturnRegister, Tst: Call->Arguments[0], |
2171 | Op0: Call->Arguments[1], Op1: Call->Arguments[2]); |
2172 | return true; |
2173 | } |
2174 | |
2175 | static bool generateConstructInst(const SPIRV::IncomingCall *Call, |
2176 | MachineIRBuilder &MIRBuilder, |
2177 | SPIRVGlobalRegistry *GR) { |
2178 | createContinuedInstructions(MIRBuilder, Opcode: SPIRV::OpCompositeConstruct, MinWC: 3, |
2179 | ContinuedOpcode: SPIRV::OpCompositeConstructContinuedINTEL, |
2180 | Args: Call->Arguments, ReturnRegister: Call->ReturnRegister, |
2181 | TypeID: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
2182 | return true; |
2183 | } |
2184 | |
2185 | static bool generateCoopMatrInst(const SPIRV::IncomingCall *Call, |
2186 | MachineIRBuilder &MIRBuilder, |
2187 | SPIRVGlobalRegistry *GR) { |
2188 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
2189 | unsigned Opcode = |
2190 | SPIRV::lookupNativeBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Opcode; |
2191 | bool IsSet = Opcode != SPIRV::OpCooperativeMatrixStoreKHR && |
2192 | Opcode != SPIRV::OpCooperativeMatrixStoreCheckedINTEL && |
2193 | Opcode != SPIRV::OpCooperativeMatrixPrefetchINTEL; |
2194 | unsigned ArgSz = Call->Arguments.size(); |
2195 | unsigned LiteralIdx = 0; |
2196 | switch (Opcode) { |
2197 | // Memory operand is optional and is literal. |
2198 | case SPIRV::OpCooperativeMatrixLoadKHR: |
2199 | LiteralIdx = ArgSz > 3 ? 3 : 0; |
2200 | break; |
2201 | case SPIRV::OpCooperativeMatrixStoreKHR: |
2202 | LiteralIdx = ArgSz > 4 ? 4 : 0; |
2203 | break; |
2204 | case SPIRV::OpCooperativeMatrixLoadCheckedINTEL: |
2205 | LiteralIdx = ArgSz > 7 ? 7 : 0; |
2206 | break; |
2207 | case SPIRV::OpCooperativeMatrixStoreCheckedINTEL: |
2208 | LiteralIdx = ArgSz > 8 ? 8 : 0; |
2209 | break; |
2210 | // Cooperative Matrix Operands operand is optional and is literal. |
2211 | case SPIRV::OpCooperativeMatrixMulAddKHR: |
2212 | LiteralIdx = ArgSz > 3 ? 3 : 0; |
2213 | break; |
2214 | }; |
2215 | |
2216 | SmallVector<uint32_t, 1> ImmArgs; |
2217 | MachineRegisterInfo *MRI = MIRBuilder.getMRI(); |
2218 | if (Opcode == SPIRV::OpCooperativeMatrixPrefetchINTEL) { |
2219 | const uint32_t CacheLevel = getConstFromIntrinsic(Reg: Call->Arguments[3], MRI); |
2220 | auto MIB = MIRBuilder.buildInstr(Opcode: SPIRV::OpCooperativeMatrixPrefetchINTEL) |
2221 | .addUse(RegNo: Call->Arguments[0]) // pointer |
2222 | .addUse(RegNo: Call->Arguments[1]) // rows |
2223 | .addUse(RegNo: Call->Arguments[2]) // columns |
2224 | .addImm(Val: CacheLevel) // cache level |
2225 | .addUse(RegNo: Call->Arguments[4]); // memory layout |
2226 | if (ArgSz > 5) |
2227 | MIB.addUse(RegNo: Call->Arguments[5]); // stride |
2228 | if (ArgSz > 6) { |
2229 | const uint32_t MemOp = getConstFromIntrinsic(Reg: Call->Arguments[6], MRI); |
2230 | MIB.addImm(Val: MemOp); // memory operand |
2231 | } |
2232 | return true; |
2233 | } |
2234 | if (LiteralIdx > 0) |
2235 | ImmArgs.push_back(Elt: getConstFromIntrinsic(Reg: Call->Arguments[LiteralIdx], MRI)); |
2236 | Register TypeReg = GR->getSPIRVTypeID(SpirvType: Call->ReturnType); |
2237 | if (Opcode == SPIRV::OpCooperativeMatrixLengthKHR) { |
2238 | SPIRVType *CoopMatrType = GR->getSPIRVTypeForVReg(VReg: Call->Arguments[0]); |
2239 | if (!CoopMatrType) |
2240 | report_fatal_error(reason: "Can't find a register's type definition" ); |
2241 | MIRBuilder.buildInstr(Opcode) |
2242 | .addDef(RegNo: Call->ReturnRegister) |
2243 | .addUse(RegNo: TypeReg) |
2244 | .addUse(RegNo: CoopMatrType->getOperand(i: 0).getReg()); |
2245 | return true; |
2246 | } |
2247 | return buildOpFromWrapper(MIRBuilder, Opcode, Call, |
2248 | TypeReg: IsSet ? TypeReg : Register(0), ImmArgs); |
2249 | } |
2250 | |
2251 | static bool generateSpecConstantInst(const SPIRV::IncomingCall *Call, |
2252 | MachineIRBuilder &MIRBuilder, |
2253 | SPIRVGlobalRegistry *GR) { |
2254 | // Lookup the instruction opcode in the TableGen records. |
2255 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
2256 | unsigned Opcode = |
2257 | SPIRV::lookupNativeBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Opcode; |
2258 | const MachineRegisterInfo *MRI = MIRBuilder.getMRI(); |
2259 | |
2260 | switch (Opcode) { |
2261 | case SPIRV::OpSpecConstant: { |
2262 | // Build the SpecID decoration. |
2263 | unsigned SpecId = |
2264 | static_cast<unsigned>(getIConstVal(ConstReg: Call->Arguments[0], MRI)); |
2265 | buildOpDecorate(Reg: Call->ReturnRegister, MIRBuilder, Dec: SPIRV::Decoration::SpecId, |
2266 | DecArgs: {SpecId}); |
2267 | // Determine the constant MI. |
2268 | Register ConstRegister = Call->Arguments[1]; |
2269 | const MachineInstr *Const = getDefInstrMaybeConstant(ConstReg&: ConstRegister, MRI); |
2270 | assert(Const && |
2271 | (Const->getOpcode() == TargetOpcode::G_CONSTANT || |
2272 | Const->getOpcode() == TargetOpcode::G_FCONSTANT) && |
2273 | "Argument should be either an int or floating-point constant" ); |
2274 | // Determine the opcode and built the OpSpec MI. |
2275 | const MachineOperand &ConstOperand = Const->getOperand(i: 1); |
2276 | if (Call->ReturnType->getOpcode() == SPIRV::OpTypeBool) { |
2277 | assert(ConstOperand.isCImm() && "Int constant operand is expected" ); |
2278 | Opcode = ConstOperand.getCImm()->getValue().getZExtValue() |
2279 | ? SPIRV::OpSpecConstantTrue |
2280 | : SPIRV::OpSpecConstantFalse; |
2281 | } |
2282 | auto MIB = MIRBuilder.buildInstr(Opcode) |
2283 | .addDef(RegNo: Call->ReturnRegister) |
2284 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
2285 | |
2286 | if (Call->ReturnType->getOpcode() != SPIRV::OpTypeBool) { |
2287 | if (Const->getOpcode() == TargetOpcode::G_CONSTANT) |
2288 | addNumImm(Imm: ConstOperand.getCImm()->getValue(), MIB); |
2289 | else |
2290 | addNumImm(Imm: ConstOperand.getFPImm()->getValueAPF().bitcastToAPInt(), MIB); |
2291 | } |
2292 | return true; |
2293 | } |
2294 | case SPIRV::OpSpecConstantComposite: { |
2295 | createContinuedInstructions(MIRBuilder, Opcode, MinWC: 3, |
2296 | ContinuedOpcode: SPIRV::OpSpecConstantCompositeContinuedINTEL, |
2297 | Args: Call->Arguments, ReturnRegister: Call->ReturnRegister, |
2298 | TypeID: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
2299 | return true; |
2300 | } |
2301 | default: |
2302 | return false; |
2303 | } |
2304 | } |
2305 | |
2306 | static bool generateExtendedBitOpsInst(const SPIRV::IncomingCall *Call, |
2307 | MachineIRBuilder &MIRBuilder, |
2308 | SPIRVGlobalRegistry *GR) { |
2309 | // Lookup the instruction opcode in the TableGen records. |
2310 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
2311 | unsigned Opcode = |
2312 | SPIRV::lookupNativeBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Opcode; |
2313 | |
2314 | return buildExtendedBitOpsInst(Call, Opcode, MIRBuilder, GR); |
2315 | } |
2316 | |
2317 | static bool generateBindlessImageINTELInst(const SPIRV::IncomingCall *Call, |
2318 | MachineIRBuilder &MIRBuilder, |
2319 | SPIRVGlobalRegistry *GR) { |
2320 | // Lookup the instruction opcode in the TableGen records. |
2321 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
2322 | unsigned Opcode = |
2323 | SPIRV::lookupNativeBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Opcode; |
2324 | |
2325 | return buildBindlessImageINTELInst(Call, Opcode, MIRBuilder, GR); |
2326 | } |
2327 | |
2328 | static bool |
2329 | generateTernaryBitwiseFunctionINTELInst(const SPIRV::IncomingCall *Call, |
2330 | MachineIRBuilder &MIRBuilder, |
2331 | SPIRVGlobalRegistry *GR) { |
2332 | // Lookup the instruction opcode in the TableGen records. |
2333 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
2334 | unsigned Opcode = |
2335 | SPIRV::lookupNativeBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Opcode; |
2336 | |
2337 | return buildTernaryBitwiseFunctionINTELInst(Call, Opcode, MIRBuilder, GR); |
2338 | } |
2339 | |
2340 | static bool generate2DBlockIOINTELInst(const SPIRV::IncomingCall *Call, |
2341 | MachineIRBuilder &MIRBuilder, |
2342 | SPIRVGlobalRegistry *GR) { |
2343 | // Lookup the instruction opcode in the TableGen records. |
2344 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
2345 | unsigned Opcode = |
2346 | SPIRV::lookupNativeBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Opcode; |
2347 | |
2348 | return build2DBlockIOINTELInst(Call, Opcode, MIRBuilder, GR); |
2349 | } |
2350 | |
2351 | static bool buildNDRange(const SPIRV::IncomingCall *Call, |
2352 | MachineIRBuilder &MIRBuilder, |
2353 | SPIRVGlobalRegistry *GR) { |
2354 | MachineRegisterInfo *MRI = MIRBuilder.getMRI(); |
2355 | SPIRVType *PtrType = GR->getSPIRVTypeForVReg(VReg: Call->Arguments[0]); |
2356 | assert(PtrType->getOpcode() == SPIRV::OpTypePointer && |
2357 | PtrType->getOperand(2).isReg()); |
2358 | Register TypeReg = PtrType->getOperand(i: 2).getReg(); |
2359 | SPIRVType *StructType = GR->getSPIRVTypeForVReg(VReg: TypeReg); |
2360 | MachineFunction &MF = MIRBuilder.getMF(); |
2361 | Register TmpReg = MRI->createVirtualRegister(RegClass: &SPIRV::iIDRegClass); |
2362 | GR->assignSPIRVTypeToVReg(Type: StructType, VReg: TmpReg, MF); |
2363 | // Skip the first arg, it's the destination pointer. OpBuildNDRange takes |
2364 | // three other arguments, so pass zero constant on absence. |
2365 | unsigned NumArgs = Call->Arguments.size(); |
2366 | assert(NumArgs >= 2); |
2367 | Register GlobalWorkSize = Call->Arguments[NumArgs < 4 ? 1 : 2]; |
2368 | Register LocalWorkSize = |
2369 | NumArgs == 2 ? Register(0) : Call->Arguments[NumArgs < 4 ? 2 : 3]; |
2370 | Register GlobalWorkOffset = NumArgs <= 3 ? Register(0) : Call->Arguments[1]; |
2371 | if (NumArgs < 4) { |
2372 | Register Const; |
2373 | SPIRVType *SpvTy = GR->getSPIRVTypeForVReg(VReg: GlobalWorkSize); |
2374 | if (SpvTy->getOpcode() == SPIRV::OpTypePointer) { |
2375 | MachineInstr *DefInstr = MRI->getUniqueVRegDef(Reg: GlobalWorkSize); |
2376 | assert(DefInstr && isSpvIntrinsic(*DefInstr, Intrinsic::spv_gep) && |
2377 | DefInstr->getOperand(3).isReg()); |
2378 | Register GWSPtr = DefInstr->getOperand(i: 3).getReg(); |
2379 | // TODO: Maybe simplify generation of the type of the fields. |
2380 | unsigned Size = Call->Builtin->Name == "ndrange_3D" ? 3 : 2; |
2381 | unsigned BitWidth = GR->getPointerSize() == 64 ? 64 : 32; |
2382 | Type *BaseTy = IntegerType::get(C&: MF.getFunction().getContext(), NumBits: BitWidth); |
2383 | Type *FieldTy = ArrayType::get(ElementType: BaseTy, NumElements: Size); |
2384 | SPIRVType *SpvFieldTy = GR->getOrCreateSPIRVType( |
2385 | Type: FieldTy, MIRBuilder, AQ: SPIRV::AccessQualifier::ReadWrite, EmitIR: true); |
2386 | GlobalWorkSize = MRI->createVirtualRegister(RegClass: &SPIRV::iIDRegClass); |
2387 | GR->assignSPIRVTypeToVReg(Type: SpvFieldTy, VReg: GlobalWorkSize, MF); |
2388 | MIRBuilder.buildInstr(Opcode: SPIRV::OpLoad) |
2389 | .addDef(RegNo: GlobalWorkSize) |
2390 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: SpvFieldTy)) |
2391 | .addUse(RegNo: GWSPtr); |
2392 | const SPIRVSubtarget &ST = |
2393 | cast<SPIRVSubtarget>(Val: MIRBuilder.getMF().getSubtarget()); |
2394 | Const = GR->getOrCreateConstIntArray(Val: 0, Num: Size, I&: *MIRBuilder.getInsertPt(), |
2395 | SpvType: SpvFieldTy, TII: *ST.getInstrInfo()); |
2396 | } else { |
2397 | Const = GR->buildConstantInt(Val: 0, MIRBuilder, SpvType: SpvTy, EmitIR: true); |
2398 | } |
2399 | if (!LocalWorkSize.isValid()) |
2400 | LocalWorkSize = Const; |
2401 | if (!GlobalWorkOffset.isValid()) |
2402 | GlobalWorkOffset = Const; |
2403 | } |
2404 | assert(LocalWorkSize.isValid() && GlobalWorkOffset.isValid()); |
2405 | MIRBuilder.buildInstr(Opcode: SPIRV::OpBuildNDRange) |
2406 | .addDef(RegNo: TmpReg) |
2407 | .addUse(RegNo: TypeReg) |
2408 | .addUse(RegNo: GlobalWorkSize) |
2409 | .addUse(RegNo: LocalWorkSize) |
2410 | .addUse(RegNo: GlobalWorkOffset); |
2411 | return MIRBuilder.buildInstr(Opcode: SPIRV::OpStore) |
2412 | .addUse(RegNo: Call->Arguments[0]) |
2413 | .addUse(RegNo: TmpReg); |
2414 | } |
2415 | |
2416 | // TODO: maybe move to the global register. |
2417 | static SPIRVType * |
2418 | getOrCreateSPIRVDeviceEventPointer(MachineIRBuilder &MIRBuilder, |
2419 | SPIRVGlobalRegistry *GR) { |
2420 | LLVMContext &Context = MIRBuilder.getMF().getFunction().getContext(); |
2421 | unsigned SC1 = storageClassToAddressSpace(SC: SPIRV::StorageClass::Generic); |
2422 | Type *PtrType = PointerType::get(C&: Context, AddressSpace: SC1); |
2423 | return GR->getOrCreateSPIRVType(Type: PtrType, MIRBuilder, |
2424 | AQ: SPIRV::AccessQualifier::ReadWrite, EmitIR: true); |
2425 | } |
2426 | |
2427 | static bool buildEnqueueKernel(const SPIRV::IncomingCall *Call, |
2428 | MachineIRBuilder &MIRBuilder, |
2429 | SPIRVGlobalRegistry *GR) { |
2430 | MachineRegisterInfo *MRI = MIRBuilder.getMRI(); |
2431 | const DataLayout &DL = MIRBuilder.getDataLayout(); |
2432 | bool IsSpirvOp = Call->isSpirvOp(); |
2433 | bool HasEvents = Call->Builtin->Name.contains(Other: "events" ) || IsSpirvOp; |
2434 | const SPIRVType *Int32Ty = GR->getOrCreateSPIRVIntegerType(BitWidth: 32, MIRBuilder); |
2435 | |
2436 | // Make vararg instructions before OpEnqueueKernel. |
2437 | // Local sizes arguments: Sizes of block invoke arguments. Clang generates |
2438 | // local size operands as an array, so we need to unpack them. |
2439 | SmallVector<Register, 16> LocalSizes; |
2440 | if (Call->Builtin->Name.contains(Other: "_varargs" ) || IsSpirvOp) { |
2441 | const unsigned LocalSizeArrayIdx = HasEvents ? 9 : 6; |
2442 | Register GepReg = Call->Arguments[LocalSizeArrayIdx]; |
2443 | MachineInstr *GepMI = MRI->getUniqueVRegDef(Reg: GepReg); |
2444 | assert(isSpvIntrinsic(*GepMI, Intrinsic::spv_gep) && |
2445 | GepMI->getOperand(3).isReg()); |
2446 | Register ArrayReg = GepMI->getOperand(i: 3).getReg(); |
2447 | MachineInstr *ArrayMI = MRI->getUniqueVRegDef(Reg: ArrayReg); |
2448 | const Type *LocalSizeTy = getMachineInstrType(MI: ArrayMI); |
2449 | assert(LocalSizeTy && "Local size type is expected" ); |
2450 | const uint64_t LocalSizeNum = |
2451 | cast<ArrayType>(Val: LocalSizeTy)->getNumElements(); |
2452 | unsigned SC = storageClassToAddressSpace(SC: SPIRV::StorageClass::Generic); |
2453 | const LLT LLType = LLT::pointer(AddressSpace: SC, SizeInBits: GR->getPointerSize()); |
2454 | const SPIRVType *PointerSizeTy = GR->getOrCreateSPIRVPointerType( |
2455 | BaseType: Int32Ty, MIRBuilder, SC: SPIRV::StorageClass::Function); |
2456 | for (unsigned I = 0; I < LocalSizeNum; ++I) { |
2457 | Register Reg = MRI->createVirtualRegister(RegClass: &SPIRV::pIDRegClass); |
2458 | MRI->setType(VReg: Reg, Ty: LLType); |
2459 | GR->assignSPIRVTypeToVReg(Type: PointerSizeTy, VReg: Reg, MF: MIRBuilder.getMF()); |
2460 | auto GEPInst = MIRBuilder.buildIntrinsic( |
2461 | ID: Intrinsic::spv_gep, Res: ArrayRef<Register>{Reg}, HasSideEffects: true, isConvergent: false); |
2462 | GEPInst |
2463 | .addImm(Val: GepMI->getOperand(i: 2).getImm()) // In bound. |
2464 | .addUse(RegNo: ArrayMI->getOperand(i: 0).getReg()) // Alloca. |
2465 | .addUse(RegNo: buildConstantIntReg32(Val: 0, MIRBuilder, GR)) // Indices. |
2466 | .addUse(RegNo: buildConstantIntReg32(Val: I, MIRBuilder, GR)); |
2467 | LocalSizes.push_back(Elt: Reg); |
2468 | } |
2469 | } |
2470 | |
2471 | // SPIRV OpEnqueueKernel instruction has 10+ arguments. |
2472 | auto MIB = MIRBuilder.buildInstr(Opcode: SPIRV::OpEnqueueKernel) |
2473 | .addDef(RegNo: Call->ReturnRegister) |
2474 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Int32Ty)); |
2475 | |
2476 | // Copy all arguments before block invoke function pointer. |
2477 | const unsigned BlockFIdx = HasEvents ? 6 : 3; |
2478 | for (unsigned i = 0; i < BlockFIdx; i++) |
2479 | MIB.addUse(RegNo: Call->Arguments[i]); |
2480 | |
2481 | // If there are no event arguments in the original call, add dummy ones. |
2482 | if (!HasEvents) { |
2483 | MIB.addUse(RegNo: buildConstantIntReg32(Val: 0, MIRBuilder, GR)); // Dummy num events. |
2484 | Register NullPtr = GR->getOrCreateConstNullPtr( |
2485 | MIRBuilder, SpvType: getOrCreateSPIRVDeviceEventPointer(MIRBuilder, GR)); |
2486 | MIB.addUse(RegNo: NullPtr); // Dummy wait events. |
2487 | MIB.addUse(RegNo: NullPtr); // Dummy ret event. |
2488 | } |
2489 | |
2490 | MachineInstr *BlockMI = getBlockStructInstr(ParamReg: Call->Arguments[BlockFIdx], MRI); |
2491 | assert(BlockMI->getOpcode() == TargetOpcode::G_GLOBAL_VALUE); |
2492 | // Invoke: Pointer to invoke function. |
2493 | MIB.addGlobalAddress(GV: BlockMI->getOperand(i: 1).getGlobal()); |
2494 | |
2495 | Register BlockLiteralReg = Call->Arguments[BlockFIdx + 1]; |
2496 | // Param: Pointer to block literal. |
2497 | MIB.addUse(RegNo: BlockLiteralReg); |
2498 | |
2499 | Type *PType = const_cast<Type *>(getBlockStructType(ParamReg: BlockLiteralReg, MRI)); |
2500 | // TODO: these numbers should be obtained from block literal structure. |
2501 | // Param Size: Size of block literal structure. |
2502 | MIB.addUse(RegNo: buildConstantIntReg32(Val: DL.getTypeStoreSize(Ty: PType), MIRBuilder, GR)); |
2503 | // Param Aligment: Aligment of block literal structure. |
2504 | MIB.addUse(RegNo: buildConstantIntReg32(Val: DL.getPrefTypeAlign(Ty: PType).value(), |
2505 | MIRBuilder, GR)); |
2506 | |
2507 | for (unsigned i = 0; i < LocalSizes.size(); i++) |
2508 | MIB.addUse(RegNo: LocalSizes[i]); |
2509 | return true; |
2510 | } |
2511 | |
2512 | static bool generateEnqueueInst(const SPIRV::IncomingCall *Call, |
2513 | MachineIRBuilder &MIRBuilder, |
2514 | SPIRVGlobalRegistry *GR) { |
2515 | // Lookup the instruction opcode in the TableGen records. |
2516 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
2517 | unsigned Opcode = |
2518 | SPIRV::lookupNativeBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Opcode; |
2519 | |
2520 | switch (Opcode) { |
2521 | case SPIRV::OpRetainEvent: |
2522 | case SPIRV::OpReleaseEvent: |
2523 | return MIRBuilder.buildInstr(Opcode).addUse(RegNo: Call->Arguments[0]); |
2524 | case SPIRV::OpCreateUserEvent: |
2525 | case SPIRV::OpGetDefaultQueue: |
2526 | return MIRBuilder.buildInstr(Opcode) |
2527 | .addDef(RegNo: Call->ReturnRegister) |
2528 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
2529 | case SPIRV::OpIsValidEvent: |
2530 | return MIRBuilder.buildInstr(Opcode) |
2531 | .addDef(RegNo: Call->ReturnRegister) |
2532 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)) |
2533 | .addUse(RegNo: Call->Arguments[0]); |
2534 | case SPIRV::OpSetUserEventStatus: |
2535 | return MIRBuilder.buildInstr(Opcode) |
2536 | .addUse(RegNo: Call->Arguments[0]) |
2537 | .addUse(RegNo: Call->Arguments[1]); |
2538 | case SPIRV::OpCaptureEventProfilingInfo: |
2539 | return MIRBuilder.buildInstr(Opcode) |
2540 | .addUse(RegNo: Call->Arguments[0]) |
2541 | .addUse(RegNo: Call->Arguments[1]) |
2542 | .addUse(RegNo: Call->Arguments[2]); |
2543 | case SPIRV::OpBuildNDRange: |
2544 | return buildNDRange(Call, MIRBuilder, GR); |
2545 | case SPIRV::OpEnqueueKernel: |
2546 | return buildEnqueueKernel(Call, MIRBuilder, GR); |
2547 | default: |
2548 | return false; |
2549 | } |
2550 | } |
2551 | |
2552 | static bool generateAsyncCopy(const SPIRV::IncomingCall *Call, |
2553 | MachineIRBuilder &MIRBuilder, |
2554 | SPIRVGlobalRegistry *GR) { |
2555 | // Lookup the instruction opcode in the TableGen records. |
2556 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
2557 | unsigned Opcode = |
2558 | SPIRV::lookupNativeBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Opcode; |
2559 | |
2560 | bool IsSet = Opcode == SPIRV::OpGroupAsyncCopy; |
2561 | Register TypeReg = GR->getSPIRVTypeID(SpirvType: Call->ReturnType); |
2562 | if (Call->isSpirvOp()) |
2563 | return buildOpFromWrapper(MIRBuilder, Opcode, Call, |
2564 | TypeReg: IsSet ? TypeReg : Register(0)); |
2565 | |
2566 | auto Scope = buildConstantIntReg32(Val: SPIRV::Scope::Workgroup, MIRBuilder, GR); |
2567 | |
2568 | switch (Opcode) { |
2569 | case SPIRV::OpGroupAsyncCopy: { |
2570 | SPIRVType *NewType = |
2571 | Call->ReturnType->getOpcode() == SPIRV::OpTypeEvent |
2572 | ? nullptr |
2573 | : GR->getOrCreateSPIRVTypeByName(TypeStr: "spirv.Event" , MIRBuilder, EmitIR: true); |
2574 | Register TypeReg = GR->getSPIRVTypeID(SpirvType: NewType ? NewType : Call->ReturnType); |
2575 | unsigned NumArgs = Call->Arguments.size(); |
2576 | Register EventReg = Call->Arguments[NumArgs - 1]; |
2577 | bool Res = MIRBuilder.buildInstr(Opcode) |
2578 | .addDef(RegNo: Call->ReturnRegister) |
2579 | .addUse(RegNo: TypeReg) |
2580 | .addUse(RegNo: Scope) |
2581 | .addUse(RegNo: Call->Arguments[0]) |
2582 | .addUse(RegNo: Call->Arguments[1]) |
2583 | .addUse(RegNo: Call->Arguments[2]) |
2584 | .addUse(RegNo: Call->Arguments.size() > 4 |
2585 | ? Call->Arguments[3] |
2586 | : buildConstantIntReg32(Val: 1, MIRBuilder, GR)) |
2587 | .addUse(RegNo: EventReg); |
2588 | if (NewType != nullptr) |
2589 | insertAssignInstr(Reg: Call->ReturnRegister, Ty: nullptr, SpirvTy: NewType, GR, MIB&: MIRBuilder, |
2590 | MRI&: MIRBuilder.getMF().getRegInfo()); |
2591 | return Res; |
2592 | } |
2593 | case SPIRV::OpGroupWaitEvents: |
2594 | return MIRBuilder.buildInstr(Opcode) |
2595 | .addUse(RegNo: Scope) |
2596 | .addUse(RegNo: Call->Arguments[0]) |
2597 | .addUse(RegNo: Call->Arguments[1]); |
2598 | default: |
2599 | return false; |
2600 | } |
2601 | } |
2602 | |
2603 | static bool generateConvertInst(const StringRef DemangledCall, |
2604 | const SPIRV::IncomingCall *Call, |
2605 | MachineIRBuilder &MIRBuilder, |
2606 | SPIRVGlobalRegistry *GR) { |
2607 | // Lookup the conversion builtin in the TableGen records. |
2608 | const SPIRV::ConvertBuiltin *Builtin = |
2609 | SPIRV::lookupConvertBuiltin(Name: Call->Builtin->Name, Set: Call->Builtin->Set); |
2610 | |
2611 | if (!Builtin && Call->isSpirvOp()) { |
2612 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
2613 | unsigned Opcode = |
2614 | SPIRV::lookupNativeBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Opcode; |
2615 | return buildOpFromWrapper(MIRBuilder, Opcode, Call, |
2616 | TypeReg: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
2617 | } |
2618 | |
2619 | if (Builtin->IsSaturated) |
2620 | buildOpDecorate(Reg: Call->ReturnRegister, MIRBuilder, |
2621 | Dec: SPIRV::Decoration::SaturatedConversion, DecArgs: {}); |
2622 | if (Builtin->IsRounded) |
2623 | buildOpDecorate(Reg: Call->ReturnRegister, MIRBuilder, |
2624 | Dec: SPIRV::Decoration::FPRoundingMode, |
2625 | DecArgs: {(unsigned)Builtin->RoundingMode}); |
2626 | |
2627 | std::string NeedExtMsg; // no errors if empty |
2628 | bool IsRightComponentsNumber = true; // check if input/output accepts vectors |
2629 | unsigned Opcode = SPIRV::OpNop; |
2630 | if (GR->isScalarOrVectorOfType(VReg: Call->Arguments[0], TypeOpcode: SPIRV::OpTypeInt)) { |
2631 | // Int -> ... |
2632 | if (GR->isScalarOrVectorOfType(VReg: Call->ReturnRegister, TypeOpcode: SPIRV::OpTypeInt)) { |
2633 | // Int -> Int |
2634 | if (Builtin->IsSaturated) |
2635 | Opcode = Builtin->IsDestinationSigned ? SPIRV::OpSatConvertUToS |
2636 | : SPIRV::OpSatConvertSToU; |
2637 | else |
2638 | Opcode = Builtin->IsDestinationSigned ? SPIRV::OpUConvert |
2639 | : SPIRV::OpSConvert; |
2640 | } else if (GR->isScalarOrVectorOfType(VReg: Call->ReturnRegister, |
2641 | TypeOpcode: SPIRV::OpTypeFloat)) { |
2642 | // Int -> Float |
2643 | if (Builtin->IsBfloat16) { |
2644 | const auto *ST = static_cast<const SPIRVSubtarget *>( |
2645 | &MIRBuilder.getMF().getSubtarget()); |
2646 | if (!ST->canUseExtension( |
2647 | E: SPIRV::Extension::SPV_INTEL_bfloat16_conversion)) |
2648 | NeedExtMsg = "SPV_INTEL_bfloat16_conversion" ; |
2649 | IsRightComponentsNumber = |
2650 | GR->getScalarOrVectorComponentCount(VReg: Call->Arguments[0]) == |
2651 | GR->getScalarOrVectorComponentCount(VReg: Call->ReturnRegister); |
2652 | Opcode = SPIRV::OpConvertBF16ToFINTEL; |
2653 | } else { |
2654 | bool IsSourceSigned = |
2655 | DemangledCall[DemangledCall.find_first_of(C: '(') + 1] != 'u'; |
2656 | Opcode = IsSourceSigned ? SPIRV::OpConvertSToF : SPIRV::OpConvertUToF; |
2657 | } |
2658 | } |
2659 | } else if (GR->isScalarOrVectorOfType(VReg: Call->Arguments[0], |
2660 | TypeOpcode: SPIRV::OpTypeFloat)) { |
2661 | // Float -> ... |
2662 | if (GR->isScalarOrVectorOfType(VReg: Call->ReturnRegister, TypeOpcode: SPIRV::OpTypeInt)) { |
2663 | // Float -> Int |
2664 | if (Builtin->IsBfloat16) { |
2665 | const auto *ST = static_cast<const SPIRVSubtarget *>( |
2666 | &MIRBuilder.getMF().getSubtarget()); |
2667 | if (!ST->canUseExtension( |
2668 | E: SPIRV::Extension::SPV_INTEL_bfloat16_conversion)) |
2669 | NeedExtMsg = "SPV_INTEL_bfloat16_conversion" ; |
2670 | IsRightComponentsNumber = |
2671 | GR->getScalarOrVectorComponentCount(VReg: Call->Arguments[0]) == |
2672 | GR->getScalarOrVectorComponentCount(VReg: Call->ReturnRegister); |
2673 | Opcode = SPIRV::OpConvertFToBF16INTEL; |
2674 | } else { |
2675 | Opcode = Builtin->IsDestinationSigned ? SPIRV::OpConvertFToS |
2676 | : SPIRV::OpConvertFToU; |
2677 | } |
2678 | } else if (GR->isScalarOrVectorOfType(VReg: Call->ReturnRegister, |
2679 | TypeOpcode: SPIRV::OpTypeFloat)) { |
2680 | // Float -> Float |
2681 | Opcode = SPIRV::OpFConvert; |
2682 | } |
2683 | } |
2684 | |
2685 | if (!NeedExtMsg.empty()) { |
2686 | std::string DiagMsg = std::string(Builtin->Name) + |
2687 | ": the builtin requires the following SPIR-V " |
2688 | "extension: " + |
2689 | NeedExtMsg; |
2690 | report_fatal_error(reason: DiagMsg.c_str(), gen_crash_diag: false); |
2691 | } |
2692 | if (!IsRightComponentsNumber) { |
2693 | std::string DiagMsg = |
2694 | std::string(Builtin->Name) + |
2695 | ": result and argument must have the same number of components" ; |
2696 | report_fatal_error(reason: DiagMsg.c_str(), gen_crash_diag: false); |
2697 | } |
2698 | assert(Opcode != SPIRV::OpNop && |
2699 | "Conversion between the types not implemented!" ); |
2700 | |
2701 | MIRBuilder.buildInstr(Opcode) |
2702 | .addDef(RegNo: Call->ReturnRegister) |
2703 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)) |
2704 | .addUse(RegNo: Call->Arguments[0]); |
2705 | return true; |
2706 | } |
2707 | |
2708 | static bool generateVectorLoadStoreInst(const SPIRV::IncomingCall *Call, |
2709 | MachineIRBuilder &MIRBuilder, |
2710 | SPIRVGlobalRegistry *GR) { |
2711 | // Lookup the vector load/store builtin in the TableGen records. |
2712 | const SPIRV::VectorLoadStoreBuiltin *Builtin = |
2713 | SPIRV::lookupVectorLoadStoreBuiltin(Name: Call->Builtin->Name, |
2714 | Set: Call->Builtin->Set); |
2715 | // Build extended instruction. |
2716 | auto MIB = |
2717 | MIRBuilder.buildInstr(Opcode: SPIRV::OpExtInst) |
2718 | .addDef(RegNo: Call->ReturnRegister) |
2719 | .addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)) |
2720 | .addImm(Val: static_cast<uint32_t>(SPIRV::InstructionSet::OpenCL_std)) |
2721 | .addImm(Val: Builtin->Number); |
2722 | for (auto Argument : Call->Arguments) |
2723 | MIB.addUse(RegNo: Argument); |
2724 | if (Builtin->Name.contains(Other: "load" ) && Builtin->ElementCount > 1) |
2725 | MIB.addImm(Val: Builtin->ElementCount); |
2726 | |
2727 | // Rounding mode should be passed as a last argument in the MI for builtins |
2728 | // like "vstorea_halfn_r". |
2729 | if (Builtin->IsRounded) |
2730 | MIB.addImm(Val: static_cast<uint32_t>(Builtin->RoundingMode)); |
2731 | return true; |
2732 | } |
2733 | |
2734 | static bool generateLoadStoreInst(const SPIRV::IncomingCall *Call, |
2735 | MachineIRBuilder &MIRBuilder, |
2736 | SPIRVGlobalRegistry *GR) { |
2737 | // Lookup the instruction opcode in the TableGen records. |
2738 | const SPIRV::DemangledBuiltin *Builtin = Call->Builtin; |
2739 | unsigned Opcode = |
2740 | SPIRV::lookupNativeBuiltin(Name: Builtin->Name, Set: Builtin->Set)->Opcode; |
2741 | bool IsLoad = Opcode == SPIRV::OpLoad; |
2742 | // Build the instruction. |
2743 | auto MIB = MIRBuilder.buildInstr(Opcode); |
2744 | if (IsLoad) { |
2745 | MIB.addDef(RegNo: Call->ReturnRegister); |
2746 | MIB.addUse(RegNo: GR->getSPIRVTypeID(SpirvType: Call->ReturnType)); |
2747 | } |
2748 | // Add a pointer to the value to load/store. |
2749 | MIB.addUse(RegNo: Call->Arguments[0]); |
2750 | MachineRegisterInfo *MRI = MIRBuilder.getMRI(); |
2751 | // Add a value to store. |
2752 | if (!IsLoad) |
2753 | MIB.addUse(RegNo: Call->Arguments[1]); |
2754 | // Add optional memory attributes and an alignment. |
2755 | unsigned NumArgs = Call->Arguments.size(); |
2756 | if ((IsLoad && NumArgs >= 2) || NumArgs >= 3) |
2757 | MIB.addImm(Val: getConstFromIntrinsic(Reg: Call->Arguments[IsLoad ? 1 : 2], MRI)); |
2758 | if ((IsLoad && NumArgs >= 3) || NumArgs >= 4) |
2759 | MIB.addImm(Val: getConstFromIntrinsic(Reg: Call->Arguments[IsLoad ? 2 : 3], MRI)); |
2760 | return true; |
2761 | } |
2762 | |
2763 | namespace SPIRV { |
2764 | // Try to find a builtin function attributes by a demangled function name and |
2765 | // return a tuple <builtin group, op code, ext instruction number>, or a special |
2766 | // tuple value <-1, 0, 0> if the builtin function is not found. |
2767 | // Not all builtin functions are supported, only those with a ready-to-use op |
2768 | // code or instruction number defined in TableGen. |
2769 | // TODO: consider a major rework of mapping demangled calls into a builtin |
2770 | // functions to unify search and decrease number of individual cases. |
2771 | std::tuple<int, unsigned, unsigned> |
2772 | mapBuiltinToOpcode(const StringRef DemangledCall, |
2773 | SPIRV::InstructionSet::InstructionSet Set) { |
2774 | Register Reg; |
2775 | SmallVector<Register> Args; |
2776 | std::unique_ptr<const IncomingCall> Call = |
2777 | lookupBuiltin(DemangledCall, Set, ReturnRegister: Reg, ReturnType: nullptr, Arguments: Args); |
2778 | if (!Call) |
2779 | return std::make_tuple(args: -1, args: 0, args: 0); |
2780 | |
2781 | switch (Call->Builtin->Group) { |
2782 | case SPIRV::Relational: |
2783 | case SPIRV::Atomic: |
2784 | case SPIRV::Barrier: |
2785 | case SPIRV::CastToPtr: |
2786 | case SPIRV::ImageMiscQuery: |
2787 | case SPIRV::SpecConstant: |
2788 | case SPIRV::Enqueue: |
2789 | case SPIRV::AsyncCopy: |
2790 | case SPIRV::LoadStore: |
2791 | case SPIRV::CoopMatr: |
2792 | if (const auto *R = |
2793 | SPIRV::lookupNativeBuiltin(Name: Call->Builtin->Name, Set: Call->Builtin->Set)) |
2794 | return std::make_tuple(args: Call->Builtin->Group, args: R->Opcode, args: 0); |
2795 | break; |
2796 | case SPIRV::Extended: |
2797 | if (const auto *R = SPIRV::lookupExtendedBuiltin(Name: Call->Builtin->Name, |
2798 | Set: Call->Builtin->Set)) |
2799 | return std::make_tuple(args: Call->Builtin->Group, args: 0, args: R->Number); |
2800 | break; |
2801 | case SPIRV::VectorLoadStore: |
2802 | if (const auto *R = SPIRV::lookupVectorLoadStoreBuiltin(Name: Call->Builtin->Name, |
2803 | Set: Call->Builtin->Set)) |
2804 | return std::make_tuple(args: SPIRV::Extended, args: 0, args: R->Number); |
2805 | break; |
2806 | case SPIRV::Group: |
2807 | if (const auto *R = SPIRV::lookupGroupBuiltin(Name: Call->Builtin->Name)) |
2808 | return std::make_tuple(args: Call->Builtin->Group, args: R->Opcode, args: 0); |
2809 | break; |
2810 | case SPIRV::AtomicFloating: |
2811 | if (const auto *R = SPIRV::lookupAtomicFloatingBuiltin(Name: Call->Builtin->Name)) |
2812 | return std::make_tuple(args: Call->Builtin->Group, args: R->Opcode, args: 0); |
2813 | break; |
2814 | case SPIRV::IntelSubgroups: |
2815 | if (const auto *R = SPIRV::lookupIntelSubgroupsBuiltin(Name: Call->Builtin->Name)) |
2816 | return std::make_tuple(args: Call->Builtin->Group, args: R->Opcode, args: 0); |
2817 | break; |
2818 | case SPIRV::GroupUniform: |
2819 | if (const auto *R = SPIRV::lookupGroupUniformBuiltin(Name: Call->Builtin->Name)) |
2820 | return std::make_tuple(args: Call->Builtin->Group, args: R->Opcode, args: 0); |
2821 | break; |
2822 | case SPIRV::IntegerDot: |
2823 | if (const auto *R = |
2824 | SPIRV::lookupIntegerDotProductBuiltin(Name: Call->Builtin->Name)) |
2825 | return std::make_tuple(args: Call->Builtin->Group, args: R->Opcode, args: 0); |
2826 | break; |
2827 | case SPIRV::WriteImage: |
2828 | return std::make_tuple(args: Call->Builtin->Group, args: SPIRV::OpImageWrite, args: 0); |
2829 | case SPIRV::Select: |
2830 | return std::make_tuple(args: Call->Builtin->Group, args: TargetOpcode::G_SELECT, args: 0); |
2831 | case SPIRV::Construct: |
2832 | return std::make_tuple(args: Call->Builtin->Group, args: SPIRV::OpCompositeConstruct, |
2833 | args: 0); |
2834 | case SPIRV::KernelClock: |
2835 | return std::make_tuple(args: Call->Builtin->Group, args: SPIRV::OpReadClockKHR, args: 0); |
2836 | default: |
2837 | return std::make_tuple(args: -1, args: 0, args: 0); |
2838 | } |
2839 | return std::make_tuple(args: -1, args: 0, args: 0); |
2840 | } |
2841 | |
2842 | std::optional<bool> lowerBuiltin(const StringRef DemangledCall, |
2843 | SPIRV::InstructionSet::InstructionSet Set, |
2844 | MachineIRBuilder &MIRBuilder, |
2845 | const Register OrigRet, const Type *OrigRetTy, |
2846 | const SmallVectorImpl<Register> &Args, |
2847 | SPIRVGlobalRegistry *GR) { |
2848 | LLVM_DEBUG(dbgs() << "Lowering builtin call: " << DemangledCall << "\n" ); |
2849 | |
2850 | // Lookup the builtin in the TableGen records. |
2851 | SPIRVType *SpvType = GR->getSPIRVTypeForVReg(VReg: OrigRet); |
2852 | assert(SpvType && "Inconsistent return register: expected valid type info" ); |
2853 | std::unique_ptr<const IncomingCall> Call = |
2854 | lookupBuiltin(DemangledCall, Set, ReturnRegister: OrigRet, ReturnType: SpvType, Arguments: Args); |
2855 | |
2856 | if (!Call) { |
2857 | LLVM_DEBUG(dbgs() << "Builtin record was not found!\n" ); |
2858 | return std::nullopt; |
2859 | } |
2860 | |
2861 | // TODO: check if the provided args meet the builtin requirments. |
2862 | assert(Args.size() >= Call->Builtin->MinNumArgs && |
2863 | "Too few arguments to generate the builtin" ); |
2864 | if (Call->Builtin->MaxNumArgs && Args.size() > Call->Builtin->MaxNumArgs) |
2865 | LLVM_DEBUG(dbgs() << "More arguments provided than required!\n" ); |
2866 | |
2867 | // Match the builtin with implementation based on the grouping. |
2868 | switch (Call->Builtin->Group) { |
2869 | case SPIRV::Extended: |
2870 | return generateExtInst(Call: Call.get(), MIRBuilder, GR); |
2871 | case SPIRV::Relational: |
2872 | return generateRelationalInst(Call: Call.get(), MIRBuilder, GR); |
2873 | case SPIRV::Group: |
2874 | return generateGroupInst(Call: Call.get(), MIRBuilder, GR); |
2875 | case SPIRV::Variable: |
2876 | return generateBuiltinVar(Call: Call.get(), MIRBuilder, GR); |
2877 | case SPIRV::Atomic: |
2878 | return generateAtomicInst(Call: Call.get(), MIRBuilder, GR); |
2879 | case SPIRV::AtomicFloating: |
2880 | return generateAtomicFloatingInst(Call: Call.get(), MIRBuilder, GR); |
2881 | case SPIRV::Barrier: |
2882 | return generateBarrierInst(Call: Call.get(), MIRBuilder, GR); |
2883 | case SPIRV::CastToPtr: |
2884 | return generateCastToPtrInst(Call: Call.get(), MIRBuilder, GR); |
2885 | case SPIRV::Dot: |
2886 | case SPIRV::IntegerDot: |
2887 | return generateDotOrFMulInst(DemangledCall, Call: Call.get(), MIRBuilder, GR); |
2888 | case SPIRV::Wave: |
2889 | return generateWaveInst(Call: Call.get(), MIRBuilder, GR); |
2890 | case SPIRV::ICarryBorrow: |
2891 | return generateICarryBorrowInst(Call: Call.get(), MIRBuilder, GR); |
2892 | case SPIRV::GetQuery: |
2893 | return generateGetQueryInst(Call: Call.get(), MIRBuilder, GR); |
2894 | case SPIRV::ImageSizeQuery: |
2895 | return generateImageSizeQueryInst(Call: Call.get(), MIRBuilder, GR); |
2896 | case SPIRV::ImageMiscQuery: |
2897 | return generateImageMiscQueryInst(Call: Call.get(), MIRBuilder, GR); |
2898 | case SPIRV::ReadImage: |
2899 | return generateReadImageInst(DemangledCall, Call: Call.get(), MIRBuilder, GR); |
2900 | case SPIRV::WriteImage: |
2901 | return generateWriteImageInst(Call: Call.get(), MIRBuilder, GR); |
2902 | case SPIRV::SampleImage: |
2903 | return generateSampleImageInst(DemangledCall, Call: Call.get(), MIRBuilder, GR); |
2904 | case SPIRV::Select: |
2905 | return generateSelectInst(Call: Call.get(), MIRBuilder); |
2906 | case SPIRV::Construct: |
2907 | return generateConstructInst(Call: Call.get(), MIRBuilder, GR); |
2908 | case SPIRV::SpecConstant: |
2909 | return generateSpecConstantInst(Call: Call.get(), MIRBuilder, GR); |
2910 | case SPIRV::Enqueue: |
2911 | return generateEnqueueInst(Call: Call.get(), MIRBuilder, GR); |
2912 | case SPIRV::AsyncCopy: |
2913 | return generateAsyncCopy(Call: Call.get(), MIRBuilder, GR); |
2914 | case SPIRV::Convert: |
2915 | return generateConvertInst(DemangledCall, Call: Call.get(), MIRBuilder, GR); |
2916 | case SPIRV::VectorLoadStore: |
2917 | return generateVectorLoadStoreInst(Call: Call.get(), MIRBuilder, GR); |
2918 | case SPIRV::LoadStore: |
2919 | return generateLoadStoreInst(Call: Call.get(), MIRBuilder, GR); |
2920 | case SPIRV::IntelSubgroups: |
2921 | return generateIntelSubgroupsInst(Call: Call.get(), MIRBuilder, GR); |
2922 | case SPIRV::GroupUniform: |
2923 | return generateGroupUniformInst(Call: Call.get(), MIRBuilder, GR); |
2924 | case SPIRV::KernelClock: |
2925 | return generateKernelClockInst(Call: Call.get(), MIRBuilder, GR); |
2926 | case SPIRV::CoopMatr: |
2927 | return generateCoopMatrInst(Call: Call.get(), MIRBuilder, GR); |
2928 | case SPIRV::ExtendedBitOps: |
2929 | return generateExtendedBitOpsInst(Call: Call.get(), MIRBuilder, GR); |
2930 | case SPIRV::BindlessINTEL: |
2931 | return generateBindlessImageINTELInst(Call: Call.get(), MIRBuilder, GR); |
2932 | case SPIRV::TernaryBitwiseINTEL: |
2933 | return generateTernaryBitwiseFunctionINTELInst(Call: Call.get(), MIRBuilder, GR); |
2934 | case SPIRV::Block2DLoadStore: |
2935 | return generate2DBlockIOINTELInst(Call: Call.get(), MIRBuilder, GR); |
2936 | } |
2937 | return false; |
2938 | } |
2939 | |
2940 | Type *parseBuiltinCallArgumentType(StringRef TypeStr, LLVMContext &Ctx) { |
2941 | // Parse strings representing OpenCL builtin types. |
2942 | if (hasBuiltinTypePrefix(Name: TypeStr)) { |
2943 | // OpenCL builtin types in demangled call strings have the following format: |
2944 | // e.g. ocl_image2d_ro |
2945 | [[maybe_unused]] bool IsOCLBuiltinType = TypeStr.consume_front(Prefix: "ocl_" ); |
2946 | assert(IsOCLBuiltinType && "Invalid OpenCL builtin prefix" ); |
2947 | |
2948 | // Check if this is pointer to a builtin type and not just pointer |
2949 | // representing a builtin type. In case it is a pointer to builtin type, |
2950 | // this will require additional handling in the method calling |
2951 | // parseBuiltinCallArgumentBaseType(...) as this function only retrieves the |
2952 | // base types. |
2953 | if (TypeStr.ends_with(Suffix: "*" )) |
2954 | TypeStr = TypeStr.slice(Start: 0, End: TypeStr.find_first_of(Chars: " *" )); |
2955 | |
2956 | return parseBuiltinTypeNameToTargetExtType(TypeName: "opencl." + TypeStr.str() + "_t" , |
2957 | Context&: Ctx); |
2958 | } |
2959 | |
2960 | // Parse type name in either "typeN" or "type vector[N]" format, where |
2961 | // N is the number of elements of the vector. |
2962 | Type *BaseType; |
2963 | unsigned VecElts = 0; |
2964 | |
2965 | BaseType = parseBasicTypeName(TypeName&: TypeStr, Ctx); |
2966 | if (!BaseType) |
2967 | // Unable to recognize SPIRV type name. |
2968 | return nullptr; |
2969 | |
2970 | // Handle "typeN*" or "type vector[N]*". |
2971 | TypeStr.consume_back(Suffix: "*" ); |
2972 | |
2973 | if (TypeStr.consume_front(Prefix: " vector[" )) |
2974 | TypeStr = TypeStr.substr(Start: 0, N: TypeStr.find(C: ']')); |
2975 | |
2976 | TypeStr.getAsInteger(Radix: 10, Result&: VecElts); |
2977 | if (VecElts > 0) |
2978 | BaseType = VectorType::get( |
2979 | ElementType: BaseType->isVoidTy() ? Type::getInt8Ty(C&: Ctx) : BaseType, NumElements: VecElts, Scalable: false); |
2980 | |
2981 | return BaseType; |
2982 | } |
2983 | |
2984 | bool parseBuiltinTypeStr(SmallVector<StringRef, 10> &BuiltinArgsTypeStrs, |
2985 | const StringRef DemangledCall, LLVMContext &Ctx) { |
2986 | auto Pos1 = DemangledCall.find(C: '('); |
2987 | if (Pos1 == StringRef::npos) |
2988 | return false; |
2989 | auto Pos2 = DemangledCall.find(C: ')'); |
2990 | if (Pos2 == StringRef::npos || Pos1 > Pos2) |
2991 | return false; |
2992 | DemangledCall.slice(Start: Pos1 + 1, End: Pos2) |
2993 | .split(A&: BuiltinArgsTypeStrs, Separator: ',', MaxSplit: -1, KeepEmpty: false); |
2994 | return true; |
2995 | } |
2996 | |
2997 | Type *parseBuiltinCallArgumentBaseType(const StringRef DemangledCall, |
2998 | unsigned ArgIdx, LLVMContext &Ctx) { |
2999 | SmallVector<StringRef, 10> BuiltinArgsTypeStrs; |
3000 | parseBuiltinTypeStr(BuiltinArgsTypeStrs, DemangledCall, Ctx); |
3001 | if (ArgIdx >= BuiltinArgsTypeStrs.size()) |
3002 | return nullptr; |
3003 | StringRef TypeStr = BuiltinArgsTypeStrs[ArgIdx].trim(); |
3004 | return parseBuiltinCallArgumentType(TypeStr, Ctx); |
3005 | } |
3006 | |
3007 | struct BuiltinType { |
3008 | StringRef Name; |
3009 | uint32_t Opcode; |
3010 | }; |
3011 | |
3012 | #define GET_BuiltinTypes_DECL |
3013 | #define GET_BuiltinTypes_IMPL |
3014 | |
3015 | struct OpenCLType { |
3016 | StringRef Name; |
3017 | StringRef SpirvTypeLiteral; |
3018 | }; |
3019 | |
3020 | #define GET_OpenCLTypes_DECL |
3021 | #define GET_OpenCLTypes_IMPL |
3022 | |
3023 | #include "SPIRVGenTables.inc" |
3024 | } // namespace SPIRV |
3025 | |
3026 | //===----------------------------------------------------------------------===// |
3027 | // Misc functions for parsing builtin types. |
3028 | //===----------------------------------------------------------------------===// |
3029 | |
3030 | static Type *parseTypeString(const StringRef Name, LLVMContext &Context) { |
3031 | if (Name.starts_with(Prefix: "void" )) |
3032 | return Type::getVoidTy(C&: Context); |
3033 | else if (Name.starts_with(Prefix: "int" ) || Name.starts_with(Prefix: "uint" )) |
3034 | return Type::getInt32Ty(C&: Context); |
3035 | else if (Name.starts_with(Prefix: "float" )) |
3036 | return Type::getFloatTy(C&: Context); |
3037 | else if (Name.starts_with(Prefix: "half" )) |
3038 | return Type::getHalfTy(C&: Context); |
3039 | report_fatal_error(reason: "Unable to recognize type!" ); |
3040 | } |
3041 | |
3042 | //===----------------------------------------------------------------------===// |
3043 | // Implementation functions for builtin types. |
3044 | //===----------------------------------------------------------------------===// |
3045 | |
3046 | static SPIRVType *getNonParameterizedType(const TargetExtType *ExtensionType, |
3047 | const SPIRV::BuiltinType *TypeRecord, |
3048 | MachineIRBuilder &MIRBuilder, |
3049 | SPIRVGlobalRegistry *GR) { |
3050 | unsigned Opcode = TypeRecord->Opcode; |
3051 | // Create or get an existing type from GlobalRegistry. |
3052 | return GR->getOrCreateOpTypeByOpcode(Ty: ExtensionType, MIRBuilder, Opcode); |
3053 | } |
3054 | |
3055 | static SPIRVType *getSamplerType(MachineIRBuilder &MIRBuilder, |
3056 | SPIRVGlobalRegistry *GR) { |
3057 | // Create or get an existing type from GlobalRegistry. |
3058 | return GR->getOrCreateOpTypeSampler(MIRBuilder); |
3059 | } |
3060 | |
3061 | static SPIRVType *getPipeType(const TargetExtType *ExtensionType, |
3062 | MachineIRBuilder &MIRBuilder, |
3063 | SPIRVGlobalRegistry *GR) { |
3064 | assert(ExtensionType->getNumIntParameters() == 1 && |
3065 | "Invalid number of parameters for SPIR-V pipe builtin!" ); |
3066 | // Create or get an existing type from GlobalRegistry. |
3067 | return GR->getOrCreateOpTypePipe(MIRBuilder, |
3068 | AccQual: SPIRV::AccessQualifier::AccessQualifier( |
3069 | ExtensionType->getIntParameter(i: 0))); |
3070 | } |
3071 | |
3072 | static SPIRVType *getCoopMatrType(const TargetExtType *ExtensionType, |
3073 | MachineIRBuilder &MIRBuilder, |
3074 | SPIRVGlobalRegistry *GR) { |
3075 | assert(ExtensionType->getNumIntParameters() == 4 && |
3076 | "Invalid number of parameters for SPIR-V coop matrices builtin!" ); |
3077 | assert(ExtensionType->getNumTypeParameters() == 1 && |
3078 | "SPIR-V coop matrices builtin type must have a type parameter!" ); |
3079 | const SPIRVType *ElemType = |
3080 | GR->getOrCreateSPIRVType(Type: ExtensionType->getTypeParameter(i: 0), MIRBuilder, |
3081 | AQ: SPIRV::AccessQualifier::ReadWrite, EmitIR: true); |
3082 | // Create or get an existing type from GlobalRegistry. |
3083 | return GR->getOrCreateOpTypeCoopMatr( |
3084 | MIRBuilder, ExtensionType, ElemType, Scope: ExtensionType->getIntParameter(i: 0), |
3085 | Rows: ExtensionType->getIntParameter(i: 1), Columns: ExtensionType->getIntParameter(i: 2), |
3086 | Use: ExtensionType->getIntParameter(i: 3), EmitIR: true); |
3087 | } |
3088 | |
3089 | static SPIRVType *getSampledImageType(const TargetExtType *OpaqueType, |
3090 | MachineIRBuilder &MIRBuilder, |
3091 | SPIRVGlobalRegistry *GR) { |
3092 | SPIRVType *OpaqueImageType = GR->getImageType( |
3093 | ExtensionType: OpaqueType, Qualifier: SPIRV::AccessQualifier::ReadOnly, MIRBuilder); |
3094 | // Create or get an existing type from GlobalRegistry. |
3095 | return GR->getOrCreateOpTypeSampledImage(ImageType: OpaqueImageType, MIRBuilder); |
3096 | } |
3097 | |
3098 | static SPIRVType *getInlineSpirvType(const TargetExtType *ExtensionType, |
3099 | MachineIRBuilder &MIRBuilder, |
3100 | SPIRVGlobalRegistry *GR) { |
3101 | assert(ExtensionType->getNumIntParameters() == 3 && |
3102 | "Inline SPIR-V type builtin takes an opcode, size, and alignment " |
3103 | "parameter" ); |
3104 | auto Opcode = ExtensionType->getIntParameter(i: 0); |
3105 | |
3106 | SmallVector<MCOperand> Operands; |
3107 | for (Type *Param : ExtensionType->type_params()) { |
3108 | if (const TargetExtType *ParamEType = dyn_cast<TargetExtType>(Val: Param)) { |
3109 | if (ParamEType->getName() == "spirv.IntegralConstant" ) { |
3110 | assert(ParamEType->getNumTypeParameters() == 1 && |
3111 | "Inline SPIR-V integral constant builtin must have a type " |
3112 | "parameter" ); |
3113 | assert(ParamEType->getNumIntParameters() == 1 && |
3114 | "Inline SPIR-V integral constant builtin must have a " |
3115 | "value parameter" ); |
3116 | |
3117 | auto OperandValue = ParamEType->getIntParameter(i: 0); |
3118 | auto *OperandType = ParamEType->getTypeParameter(i: 0); |
3119 | |
3120 | const SPIRVType *OperandSPIRVType = GR->getOrCreateSPIRVType( |
3121 | Type: OperandType, MIRBuilder, AQ: SPIRV::AccessQualifier::ReadWrite, EmitIR: true); |
3122 | |
3123 | Operands.push_back(Elt: MCOperand::createReg(Reg: GR->buildConstantInt( |
3124 | Val: OperandValue, MIRBuilder, SpvType: OperandSPIRVType, EmitIR: true))); |
3125 | continue; |
3126 | } else if (ParamEType->getName() == "spirv.Literal" ) { |
3127 | assert(ParamEType->getNumTypeParameters() == 0 && |
3128 | "Inline SPIR-V literal builtin does not take type " |
3129 | "parameters" ); |
3130 | assert(ParamEType->getNumIntParameters() == 1 && |
3131 | "Inline SPIR-V literal builtin must have an integer " |
3132 | "parameter" ); |
3133 | |
3134 | auto OperandValue = ParamEType->getIntParameter(i: 0); |
3135 | |
3136 | Operands.push_back(Elt: MCOperand::createImm(Val: OperandValue)); |
3137 | continue; |
3138 | } |
3139 | } |
3140 | const SPIRVType *TypeOperand = GR->getOrCreateSPIRVType( |
3141 | Type: Param, MIRBuilder, AQ: SPIRV::AccessQualifier::ReadWrite, EmitIR: true); |
3142 | Operands.push_back(Elt: MCOperand::createReg(Reg: GR->getSPIRVTypeID(SpirvType: TypeOperand))); |
3143 | } |
3144 | |
3145 | return GR->getOrCreateUnknownType(Ty: ExtensionType, MIRBuilder, Opcode, |
3146 | Operands); |
3147 | } |
3148 | |
3149 | static SPIRVType *getVulkanBufferType(const TargetExtType *ExtensionType, |
3150 | MachineIRBuilder &MIRBuilder, |
3151 | SPIRVGlobalRegistry *GR) { |
3152 | assert(ExtensionType->getNumTypeParameters() == 1 && |
3153 | "Vulkan buffers have exactly one type for the type of the buffer." ); |
3154 | assert(ExtensionType->getNumIntParameters() == 2 && |
3155 | "Vulkan buffer have 2 integer parameters: storage class and is " |
3156 | "writable." ); |
3157 | |
3158 | auto *T = ExtensionType->getTypeParameter(i: 0); |
3159 | auto SC = static_cast<SPIRV::StorageClass::StorageClass>( |
3160 | ExtensionType->getIntParameter(i: 0)); |
3161 | bool IsWritable = ExtensionType->getIntParameter(i: 1); |
3162 | return GR->getOrCreateVulkanBufferType(MIRBuilder, ElemType: T, SC, IsWritable); |
3163 | } |
3164 | |
3165 | static SPIRVType *getLayoutType(const TargetExtType *ExtensionType, |
3166 | MachineIRBuilder &MIRBuilder, |
3167 | SPIRVGlobalRegistry *GR) { |
3168 | return GR->getOrCreateLayoutType(MIRBuilder, T: ExtensionType); |
3169 | } |
3170 | |
3171 | namespace SPIRV { |
3172 | TargetExtType *parseBuiltinTypeNameToTargetExtType(std::string TypeName, |
3173 | LLVMContext &Context) { |
3174 | StringRef NameWithParameters = TypeName; |
3175 | |
3176 | // Pointers-to-opaque-structs representing OpenCL types are first translated |
3177 | // to equivalent SPIR-V types. OpenCL builtin type names should have the |
3178 | // following format: e.g. %opencl.event_t |
3179 | if (NameWithParameters.starts_with(Prefix: "opencl." )) { |
3180 | const SPIRV::OpenCLType *OCLTypeRecord = |
3181 | SPIRV::lookupOpenCLType(Name: NameWithParameters); |
3182 | if (!OCLTypeRecord) |
3183 | report_fatal_error(reason: "Missing TableGen record for OpenCL type: " + |
3184 | NameWithParameters); |
3185 | NameWithParameters = OCLTypeRecord->SpirvTypeLiteral; |
3186 | // Continue with the SPIR-V builtin type... |
3187 | } |
3188 | |
3189 | // Names of the opaque structs representing a SPIR-V builtins without |
3190 | // parameters should have the following format: e.g. %spirv.Event |
3191 | assert(NameWithParameters.starts_with("spirv." ) && |
3192 | "Unknown builtin opaque type!" ); |
3193 | |
3194 | // Parameterized SPIR-V builtins names follow this format: |
3195 | // e.g. %spirv.Image._void_1_0_0_0_0_0_0, %spirv.Pipe._0 |
3196 | if (!NameWithParameters.contains(C: '_')) |
3197 | return TargetExtType::get(Context, Name: NameWithParameters); |
3198 | |
3199 | SmallVector<StringRef> Parameters; |
3200 | unsigned BaseNameLength = NameWithParameters.find(C: '_') - 1; |
3201 | SplitString(Source: NameWithParameters.substr(Start: BaseNameLength + 1), OutFragments&: Parameters, Delimiters: "_" ); |
3202 | |
3203 | SmallVector<Type *, 1> TypeParameters; |
3204 | bool HasTypeParameter = !isDigit(C: Parameters[0][0]); |
3205 | if (HasTypeParameter) |
3206 | TypeParameters.push_back(Elt: parseTypeString(Name: Parameters[0], Context)); |
3207 | SmallVector<unsigned> IntParameters; |
3208 | for (unsigned i = HasTypeParameter ? 1 : 0; i < Parameters.size(); i++) { |
3209 | unsigned IntParameter = 0; |
3210 | bool ValidLiteral = !Parameters[i].getAsInteger(Radix: 10, Result&: IntParameter); |
3211 | (void)ValidLiteral; |
3212 | assert(ValidLiteral && |
3213 | "Invalid format of SPIR-V builtin parameter literal!" ); |
3214 | IntParameters.push_back(Elt: IntParameter); |
3215 | } |
3216 | return TargetExtType::get(Context, |
3217 | Name: NameWithParameters.substr(Start: 0, N: BaseNameLength), |
3218 | Types: TypeParameters, Ints: IntParameters); |
3219 | } |
3220 | |
3221 | SPIRVType *lowerBuiltinType(const Type *OpaqueType, |
3222 | SPIRV::AccessQualifier::AccessQualifier AccessQual, |
3223 | MachineIRBuilder &MIRBuilder, |
3224 | SPIRVGlobalRegistry *GR) { |
3225 | // In LLVM IR, SPIR-V and OpenCL builtin types are represented as either |
3226 | // target(...) target extension types or pointers-to-opaque-structs. The |
3227 | // approach relying on structs is deprecated and works only in the non-opaque |
3228 | // pointer mode (-opaque-pointers=0). |
3229 | // In order to maintain compatibility with LLVM IR generated by older versions |
3230 | // of Clang and LLVM/SPIR-V Translator, the pointers-to-opaque-structs are |
3231 | // "translated" to target extension types. This translation is temporary and |
3232 | // will be removed in the future release of LLVM. |
3233 | const TargetExtType *BuiltinType = dyn_cast<TargetExtType>(Val: OpaqueType); |
3234 | if (!BuiltinType) |
3235 | BuiltinType = parseBuiltinTypeNameToTargetExtType( |
3236 | TypeName: OpaqueType->getStructName().str(), Context&: MIRBuilder.getContext()); |
3237 | |
3238 | unsigned NumStartingVRegs = MIRBuilder.getMRI()->getNumVirtRegs(); |
3239 | |
3240 | const StringRef Name = BuiltinType->getName(); |
3241 | LLVM_DEBUG(dbgs() << "Lowering builtin type: " << Name << "\n" ); |
3242 | |
3243 | SPIRVType *TargetType; |
3244 | if (Name == "spirv.Type" ) { |
3245 | TargetType = getInlineSpirvType(ExtensionType: BuiltinType, MIRBuilder, GR); |
3246 | } else if (Name == "spirv.VulkanBuffer" ) { |
3247 | TargetType = getVulkanBufferType(ExtensionType: BuiltinType, MIRBuilder, GR); |
3248 | } else if (Name == "spirv.Layout" ) { |
3249 | TargetType = getLayoutType(ExtensionType: BuiltinType, MIRBuilder, GR); |
3250 | } else { |
3251 | // Lookup the demangled builtin type in the TableGen records. |
3252 | const SPIRV::BuiltinType *TypeRecord = SPIRV::lookupBuiltinType(Name); |
3253 | if (!TypeRecord) |
3254 | report_fatal_error(reason: "Missing TableGen record for builtin type: " + Name); |
3255 | |
3256 | // "Lower" the BuiltinType into TargetType. The following get<...>Type |
3257 | // methods use the implementation details from TableGen records or |
3258 | // TargetExtType parameters to either create a new OpType<...> machine |
3259 | // instruction or get an existing equivalent SPIRVType from |
3260 | // GlobalRegistry. |
3261 | |
3262 | switch (TypeRecord->Opcode) { |
3263 | case SPIRV::OpTypeImage: |
3264 | TargetType = GR->getImageType(ExtensionType: BuiltinType, Qualifier: AccessQual, MIRBuilder); |
3265 | break; |
3266 | case SPIRV::OpTypePipe: |
3267 | TargetType = getPipeType(ExtensionType: BuiltinType, MIRBuilder, GR); |
3268 | break; |
3269 | case SPIRV::OpTypeDeviceEvent: |
3270 | TargetType = GR->getOrCreateOpTypeDeviceEvent(MIRBuilder); |
3271 | break; |
3272 | case SPIRV::OpTypeSampler: |
3273 | TargetType = getSamplerType(MIRBuilder, GR); |
3274 | break; |
3275 | case SPIRV::OpTypeSampledImage: |
3276 | TargetType = getSampledImageType(OpaqueType: BuiltinType, MIRBuilder, GR); |
3277 | break; |
3278 | case SPIRV::OpTypeCooperativeMatrixKHR: |
3279 | TargetType = getCoopMatrType(ExtensionType: BuiltinType, MIRBuilder, GR); |
3280 | break; |
3281 | default: |
3282 | TargetType = |
3283 | getNonParameterizedType(ExtensionType: BuiltinType, TypeRecord, MIRBuilder, GR); |
3284 | break; |
3285 | } |
3286 | } |
3287 | |
3288 | // Emit OpName instruction if a new OpType<...> instruction was added |
3289 | // (equivalent type was not found in GlobalRegistry). |
3290 | if (NumStartingVRegs < MIRBuilder.getMRI()->getNumVirtRegs()) |
3291 | buildOpName(Target: GR->getSPIRVTypeID(SpirvType: TargetType), Name, MIRBuilder); |
3292 | |
3293 | return TargetType; |
3294 | } |
3295 | } // namespace SPIRV |
3296 | } // namespace llvm |
3297 | |