1 | //===- AArch64.cpp --------------------------------------------------------===// |
2 | // |
3 | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
4 | // See https://llvm.org/LICENSE.txt for license information. |
5 | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
6 | // |
7 | //===----------------------------------------------------------------------===// |
8 | |
9 | #include "ABIInfoImpl.h" |
10 | #include "TargetInfo.h" |
11 | #include "clang/AST/Decl.h" |
12 | #include "clang/Basic/DiagnosticFrontend.h" |
13 | #include "llvm/TargetParser/AArch64TargetParser.h" |
14 | |
15 | using namespace clang; |
16 | using namespace clang::CodeGen; |
17 | |
18 | //===----------------------------------------------------------------------===// |
19 | // AArch64 ABI Implementation |
20 | //===----------------------------------------------------------------------===// |
21 | |
22 | namespace { |
23 | |
24 | class AArch64ABIInfo : public ABIInfo { |
25 | AArch64ABIKind Kind; |
26 | |
27 | public: |
28 | AArch64ABIInfo(CodeGenTypes &CGT, AArch64ABIKind Kind) |
29 | : ABIInfo(CGT), Kind(Kind) {} |
30 | |
31 | bool isSoftFloat() const { return Kind == AArch64ABIKind::AAPCSSoft; } |
32 | |
33 | private: |
34 | AArch64ABIKind getABIKind() const { return Kind; } |
35 | bool isDarwinPCS() const { return Kind == AArch64ABIKind::DarwinPCS; } |
36 | |
37 | ABIArgInfo classifyReturnType(QualType RetTy, bool IsVariadicFn) const; |
38 | ABIArgInfo classifyArgumentType(QualType RetTy, bool IsVariadicFn, |
39 | bool IsNamedArg, unsigned CallingConvention, |
40 | unsigned &NSRN, unsigned &NPRN) const; |
41 | llvm::Type *convertFixedToScalableVectorType(const VectorType *VT) const; |
42 | ABIArgInfo coerceIllegalVector(QualType Ty, unsigned &NSRN, |
43 | unsigned &NPRN) const; |
44 | ABIArgInfo coerceAndExpandPureScalableAggregate( |
45 | QualType Ty, bool IsNamedArg, unsigned NVec, unsigned NPred, |
46 | const SmallVectorImpl<llvm::Type *> &UnpaddedCoerceToSeq, unsigned &NSRN, |
47 | unsigned &NPRN) const; |
48 | bool isHomogeneousAggregateBaseType(QualType Ty) const override; |
49 | bool isHomogeneousAggregateSmallEnough(const Type *Ty, |
50 | uint64_t Members) const override; |
51 | bool isZeroLengthBitfieldPermittedInHomogeneousAggregate() const override; |
52 | |
53 | bool isIllegalVectorType(QualType Ty) const; |
54 | |
55 | bool passAsAggregateType(QualType Ty) const; |
56 | bool passAsPureScalableType(QualType Ty, unsigned &NV, unsigned &NP, |
57 | SmallVectorImpl<llvm::Type *> &CoerceToSeq) const; |
58 | |
59 | void flattenType(llvm::Type *Ty, |
60 | SmallVectorImpl<llvm::Type *> &Flattened) const; |
61 | |
62 | void computeInfo(CGFunctionInfo &FI) const override { |
63 | if (!::classifyReturnType(CXXABI: getCXXABI(), FI, Info: *this)) |
64 | FI.getReturnInfo() = |
65 | classifyReturnType(RetTy: FI.getReturnType(), IsVariadicFn: FI.isVariadic()); |
66 | |
67 | unsigned ArgNo = 0; |
68 | unsigned NSRN = 0, NPRN = 0; |
69 | for (auto &it : FI.arguments()) { |
70 | const bool IsNamedArg = |
71 | !FI.isVariadic() || ArgNo < FI.getRequiredArgs().getNumRequiredArgs(); |
72 | ++ArgNo; |
73 | it.info = classifyArgumentType(RetTy: it.type, IsVariadicFn: FI.isVariadic(), IsNamedArg, |
74 | CallingConvention: FI.getCallingConvention(), NSRN, NPRN); |
75 | } |
76 | } |
77 | |
78 | RValue EmitDarwinVAArg(Address VAListAddr, QualType Ty, CodeGenFunction &CGF, |
79 | AggValueSlot Slot) const; |
80 | |
81 | RValue EmitAAPCSVAArg(Address VAListAddr, QualType Ty, CodeGenFunction &CGF, |
82 | AArch64ABIKind Kind, AggValueSlot Slot) const; |
83 | |
84 | RValue EmitVAArg(CodeGenFunction &CGF, Address VAListAddr, QualType Ty, |
85 | AggValueSlot Slot) const override { |
86 | llvm::Type *BaseTy = CGF.ConvertType(T: Ty); |
87 | if (isa<llvm::ScalableVectorType>(Val: BaseTy)) |
88 | llvm::report_fatal_error(reason: "Passing SVE types to variadic functions is " |
89 | "currently not supported" ); |
90 | |
91 | return Kind == AArch64ABIKind::Win64 |
92 | ? EmitMSVAArg(CGF, VAListAddr, Ty, Slot) |
93 | : isDarwinPCS() ? EmitDarwinVAArg(VAListAddr, Ty, CGF, Slot) |
94 | : EmitAAPCSVAArg(VAListAddr, Ty, CGF, Kind, Slot); |
95 | } |
96 | |
97 | RValue EmitMSVAArg(CodeGenFunction &CGF, Address VAListAddr, QualType Ty, |
98 | AggValueSlot Slot) const override; |
99 | |
100 | bool allowBFloatArgsAndRet() const override { |
101 | return getTarget().hasBFloat16Type(); |
102 | } |
103 | |
104 | using ABIInfo::appendAttributeMangling; |
105 | void appendAttributeMangling(TargetClonesAttr *Attr, unsigned Index, |
106 | raw_ostream &Out) const override; |
107 | void appendAttributeMangling(StringRef AttrStr, |
108 | raw_ostream &Out) const override; |
109 | }; |
110 | |
111 | class AArch64SwiftABIInfo : public SwiftABIInfo { |
112 | public: |
113 | explicit AArch64SwiftABIInfo(CodeGenTypes &CGT) |
114 | : SwiftABIInfo(CGT, /*SwiftErrorInRegister=*/true) {} |
115 | |
116 | bool isLegalVectorType(CharUnits VectorSize, llvm::Type *EltTy, |
117 | unsigned NumElts) const override; |
118 | }; |
119 | |
120 | class AArch64TargetCodeGenInfo : public TargetCodeGenInfo { |
121 | public: |
122 | AArch64TargetCodeGenInfo(CodeGenTypes &CGT, AArch64ABIKind Kind) |
123 | : TargetCodeGenInfo(std::make_unique<AArch64ABIInfo>(args&: CGT, args&: Kind)) { |
124 | SwiftInfo = std::make_unique<AArch64SwiftABIInfo>(args&: CGT); |
125 | } |
126 | |
127 | StringRef getARCRetainAutoreleasedReturnValueMarker() const override { |
128 | return "mov\tfp, fp\t\t// marker for objc_retainAutoreleaseReturnValue" ; |
129 | } |
130 | |
131 | int getDwarfEHStackPointer(CodeGen::CodeGenModule &M) const override { |
132 | return 31; |
133 | } |
134 | |
135 | bool doesReturnSlotInterfereWithArgs() const override { return false; } |
136 | |
137 | void setTargetAttributes(const Decl *D, llvm::GlobalValue *GV, |
138 | CodeGen::CodeGenModule &CGM) const override { |
139 | auto *Fn = dyn_cast<llvm::Function>(Val: GV); |
140 | if (!Fn) |
141 | return; |
142 | |
143 | const auto *FD = dyn_cast_or_null<FunctionDecl>(Val: D); |
144 | TargetInfo::BranchProtectionInfo BPI(CGM.getLangOpts()); |
145 | |
146 | if (FD && FD->hasAttr<TargetAttr>()) { |
147 | const auto *TA = FD->getAttr<TargetAttr>(); |
148 | ParsedTargetAttr Attr = |
149 | CGM.getTarget().parseTargetAttr(Str: TA->getFeaturesStr()); |
150 | if (!Attr.BranchProtection.empty()) { |
151 | StringRef Error; |
152 | (void)CGM.getTarget().validateBranchProtection( |
153 | Spec: Attr.BranchProtection, Arch: Attr.CPU, BPI, LO: CGM.getLangOpts(), Err&: Error); |
154 | assert(Error.empty()); |
155 | } |
156 | } |
157 | setBranchProtectionFnAttributes(BPI, F&: *Fn); |
158 | setPointerAuthFnAttributes(Opts: CGM.getCodeGenOpts().PointerAuth, F&: *Fn); |
159 | } |
160 | |
161 | bool isScalarizableAsmOperand(CodeGen::CodeGenFunction &CGF, |
162 | llvm::Type *Ty) const override { |
163 | if (CGF.getTarget().hasFeature(Feature: "ls64" )) { |
164 | auto *ST = dyn_cast<llvm::StructType>(Val: Ty); |
165 | if (ST && ST->getNumElements() == 1) { |
166 | auto *AT = dyn_cast<llvm::ArrayType>(Val: ST->getElementType(N: 0)); |
167 | if (AT && AT->getNumElements() == 8 && |
168 | AT->getElementType()->isIntegerTy(Bitwidth: 64)) |
169 | return true; |
170 | } |
171 | } |
172 | return TargetCodeGenInfo::isScalarizableAsmOperand(CGF, Ty); |
173 | } |
174 | |
175 | void checkFunctionABI(CodeGenModule &CGM, |
176 | const FunctionDecl *Decl) const override; |
177 | |
178 | void checkFunctionCallABI(CodeGenModule &CGM, SourceLocation CallLoc, |
179 | const FunctionDecl *Caller, |
180 | const FunctionDecl *Callee, const CallArgList &Args, |
181 | QualType ReturnType) const override; |
182 | |
183 | bool wouldInliningViolateFunctionCallABI( |
184 | const FunctionDecl *Caller, const FunctionDecl *Callee) const override; |
185 | |
186 | private: |
187 | // Diagnose calls between functions with incompatible Streaming SVE |
188 | // attributes. |
189 | void checkFunctionCallABIStreaming(CodeGenModule &CGM, SourceLocation CallLoc, |
190 | const FunctionDecl *Caller, |
191 | const FunctionDecl *Callee) const; |
192 | // Diagnose calls which must pass arguments in floating-point registers when |
193 | // the selected target does not have floating-point registers. |
194 | void checkFunctionCallABISoftFloat(CodeGenModule &CGM, SourceLocation CallLoc, |
195 | const FunctionDecl *Caller, |
196 | const FunctionDecl *Callee, |
197 | const CallArgList &Args, |
198 | QualType ReturnType) const; |
199 | }; |
200 | |
201 | class WindowsAArch64TargetCodeGenInfo : public AArch64TargetCodeGenInfo { |
202 | public: |
203 | WindowsAArch64TargetCodeGenInfo(CodeGenTypes &CGT, AArch64ABIKind K) |
204 | : AArch64TargetCodeGenInfo(CGT, K) {} |
205 | |
206 | void setTargetAttributes(const Decl *D, llvm::GlobalValue *GV, |
207 | CodeGen::CodeGenModule &CGM) const override; |
208 | |
209 | void getDependentLibraryOption(llvm::StringRef Lib, |
210 | llvm::SmallString<24> &Opt) const override { |
211 | Opt = "/DEFAULTLIB:" + qualifyWindowsLibrary(Lib); |
212 | } |
213 | |
214 | void getDetectMismatchOption(llvm::StringRef Name, llvm::StringRef Value, |
215 | llvm::SmallString<32> &Opt) const override { |
216 | Opt = "/FAILIFMISMATCH:\"" + Name.str() + "=" + Value.str() + "\"" ; |
217 | } |
218 | }; |
219 | |
220 | void WindowsAArch64TargetCodeGenInfo::setTargetAttributes( |
221 | const Decl *D, llvm::GlobalValue *GV, CodeGen::CodeGenModule &CGM) const { |
222 | AArch64TargetCodeGenInfo::setTargetAttributes(D, GV, CGM); |
223 | if (GV->isDeclaration()) |
224 | return; |
225 | addStackProbeTargetAttributes(D, GV, CGM); |
226 | } |
227 | } |
228 | |
229 | llvm::Type * |
230 | AArch64ABIInfo::convertFixedToScalableVectorType(const VectorType *VT) const { |
231 | assert(VT->getElementType()->isBuiltinType() && "expected builtin type!" ); |
232 | |
233 | if (VT->getVectorKind() == VectorKind::SveFixedLengthPredicate) { |
234 | assert(VT->getElementType()->castAs<BuiltinType>()->getKind() == |
235 | BuiltinType::UChar && |
236 | "unexpected builtin type for SVE predicate!" ); |
237 | return llvm::ScalableVectorType::get(ElementType: llvm::Type::getInt1Ty(C&: getVMContext()), |
238 | MinNumElts: 16); |
239 | } |
240 | |
241 | if (VT->getVectorKind() == VectorKind::SveFixedLengthData) { |
242 | const auto *BT = VT->getElementType()->castAs<BuiltinType>(); |
243 | switch (BT->getKind()) { |
244 | default: |
245 | llvm_unreachable("unexpected builtin type for SVE vector!" ); |
246 | |
247 | case BuiltinType::SChar: |
248 | case BuiltinType::UChar: |
249 | case BuiltinType::MFloat8: |
250 | return llvm::ScalableVectorType::get( |
251 | ElementType: llvm::Type::getInt8Ty(C&: getVMContext()), MinNumElts: 16); |
252 | |
253 | case BuiltinType::Short: |
254 | case BuiltinType::UShort: |
255 | return llvm::ScalableVectorType::get( |
256 | ElementType: llvm::Type::getInt16Ty(C&: getVMContext()), MinNumElts: 8); |
257 | |
258 | case BuiltinType::Int: |
259 | case BuiltinType::UInt: |
260 | return llvm::ScalableVectorType::get( |
261 | ElementType: llvm::Type::getInt32Ty(C&: getVMContext()), MinNumElts: 4); |
262 | |
263 | case BuiltinType::Long: |
264 | case BuiltinType::ULong: |
265 | return llvm::ScalableVectorType::get( |
266 | ElementType: llvm::Type::getInt64Ty(C&: getVMContext()), MinNumElts: 2); |
267 | |
268 | case BuiltinType::Half: |
269 | return llvm::ScalableVectorType::get( |
270 | ElementType: llvm::Type::getHalfTy(C&: getVMContext()), MinNumElts: 8); |
271 | |
272 | case BuiltinType::Float: |
273 | return llvm::ScalableVectorType::get( |
274 | ElementType: llvm::Type::getFloatTy(C&: getVMContext()), MinNumElts: 4); |
275 | |
276 | case BuiltinType::Double: |
277 | return llvm::ScalableVectorType::get( |
278 | ElementType: llvm::Type::getDoubleTy(C&: getVMContext()), MinNumElts: 2); |
279 | |
280 | case BuiltinType::BFloat16: |
281 | return llvm::ScalableVectorType::get( |
282 | ElementType: llvm::Type::getBFloatTy(C&: getVMContext()), MinNumElts: 8); |
283 | } |
284 | } |
285 | |
286 | llvm_unreachable("expected fixed-length SVE vector" ); |
287 | } |
288 | |
289 | ABIArgInfo AArch64ABIInfo::coerceIllegalVector(QualType Ty, unsigned &NSRN, |
290 | unsigned &NPRN) const { |
291 | assert(Ty->isVectorType() && "expected vector type!" ); |
292 | |
293 | const auto *VT = Ty->castAs<VectorType>(); |
294 | if (VT->getVectorKind() == VectorKind::SveFixedLengthPredicate) { |
295 | assert(VT->getElementType()->isBuiltinType() && "expected builtin type!" ); |
296 | assert(VT->getElementType()->castAs<BuiltinType>()->getKind() == |
297 | BuiltinType::UChar && |
298 | "unexpected builtin type for SVE predicate!" ); |
299 | NPRN = std::min(a: NPRN + 1, b: 4u); |
300 | return ABIArgInfo::getDirect(T: llvm::ScalableVectorType::get( |
301 | ElementType: llvm::Type::getInt1Ty(C&: getVMContext()), MinNumElts: 16)); |
302 | } |
303 | |
304 | if (VT->getVectorKind() == VectorKind::SveFixedLengthData) { |
305 | NSRN = std::min(a: NSRN + 1, b: 8u); |
306 | return ABIArgInfo::getDirect(T: convertFixedToScalableVectorType(VT)); |
307 | } |
308 | |
309 | uint64_t Size = getContext().getTypeSize(T: Ty); |
310 | // Android promotes <2 x i8> to i16, not i32 |
311 | if ((isAndroid() || isOHOSFamily()) && (Size <= 16)) { |
312 | llvm::Type *ResType = llvm::Type::getInt16Ty(C&: getVMContext()); |
313 | return ABIArgInfo::getDirect(T: ResType); |
314 | } |
315 | if (Size <= 32) { |
316 | llvm::Type *ResType = llvm::Type::getInt32Ty(C&: getVMContext()); |
317 | return ABIArgInfo::getDirect(T: ResType); |
318 | } |
319 | if (Size == 64) { |
320 | NSRN = std::min(a: NSRN + 1, b: 8u); |
321 | auto *ResType = |
322 | llvm::FixedVectorType::get(ElementType: llvm::Type::getInt32Ty(C&: getVMContext()), NumElts: 2); |
323 | return ABIArgInfo::getDirect(T: ResType); |
324 | } |
325 | if (Size == 128) { |
326 | NSRN = std::min(a: NSRN + 1, b: 8u); |
327 | auto *ResType = |
328 | llvm::FixedVectorType::get(ElementType: llvm::Type::getInt32Ty(C&: getVMContext()), NumElts: 4); |
329 | return ABIArgInfo::getDirect(T: ResType); |
330 | } |
331 | |
332 | return getNaturalAlignIndirect(Ty, AddrSpace: getDataLayout().getAllocaAddrSpace(), |
333 | /*ByVal=*/false); |
334 | } |
335 | |
336 | ABIArgInfo AArch64ABIInfo::coerceAndExpandPureScalableAggregate( |
337 | QualType Ty, bool IsNamedArg, unsigned NVec, unsigned NPred, |
338 | const SmallVectorImpl<llvm::Type *> &UnpaddedCoerceToSeq, unsigned &NSRN, |
339 | unsigned &NPRN) const { |
340 | if (!IsNamedArg || NSRN + NVec > 8 || NPRN + NPred > 4) |
341 | return getNaturalAlignIndirect(Ty, AddrSpace: getDataLayout().getAllocaAddrSpace(), |
342 | /*ByVal=*/false); |
343 | NSRN += NVec; |
344 | NPRN += NPred; |
345 | |
346 | // Handle SVE vector tuples. |
347 | if (Ty->isSVESizelessBuiltinType()) |
348 | return ABIArgInfo::getDirect(); |
349 | |
350 | llvm::Type *UnpaddedCoerceToType = |
351 | UnpaddedCoerceToSeq.size() == 1 |
352 | ? UnpaddedCoerceToSeq[0] |
353 | : llvm::StructType::get(Context&: CGT.getLLVMContext(), Elements: UnpaddedCoerceToSeq, |
354 | isPacked: true); |
355 | |
356 | SmallVector<llvm::Type *> CoerceToSeq; |
357 | flattenType(Ty: CGT.ConvertType(T: Ty), Flattened&: CoerceToSeq); |
358 | auto *CoerceToType = |
359 | llvm::StructType::get(Context&: CGT.getLLVMContext(), Elements: CoerceToSeq, isPacked: false); |
360 | |
361 | return ABIArgInfo::getCoerceAndExpand(coerceToType: CoerceToType, unpaddedCoerceToType: UnpaddedCoerceToType); |
362 | } |
363 | |
364 | ABIArgInfo AArch64ABIInfo::classifyArgumentType(QualType Ty, bool IsVariadicFn, |
365 | bool IsNamedArg, |
366 | unsigned CallingConvention, |
367 | unsigned &NSRN, |
368 | unsigned &NPRN) const { |
369 | Ty = useFirstFieldIfTransparentUnion(Ty); |
370 | |
371 | // Handle illegal vector types here. |
372 | if (isIllegalVectorType(Ty)) |
373 | return coerceIllegalVector(Ty, NSRN, NPRN); |
374 | |
375 | if (!passAsAggregateType(Ty)) { |
376 | // Treat an enum type as its underlying type. |
377 | if (const EnumType *EnumTy = Ty->getAs<EnumType>()) |
378 | Ty = EnumTy->getDecl()->getIntegerType(); |
379 | |
380 | if (const auto *EIT = Ty->getAs<BitIntType>()) |
381 | if (EIT->getNumBits() > 128) |
382 | return getNaturalAlignIndirect(Ty, AddrSpace: getDataLayout().getAllocaAddrSpace(), |
383 | ByVal: false); |
384 | |
385 | if (Ty->isVectorType()) |
386 | NSRN = std::min(a: NSRN + 1, b: 8u); |
387 | else if (const auto *BT = Ty->getAs<BuiltinType>()) { |
388 | if (BT->isFloatingPoint()) |
389 | NSRN = std::min(a: NSRN + 1, b: 8u); |
390 | else { |
391 | switch (BT->getKind()) { |
392 | case BuiltinType::SveBool: |
393 | case BuiltinType::SveCount: |
394 | NPRN = std::min(a: NPRN + 1, b: 4u); |
395 | break; |
396 | case BuiltinType::SveBoolx2: |
397 | NPRN = std::min(a: NPRN + 2, b: 4u); |
398 | break; |
399 | case BuiltinType::SveBoolx4: |
400 | NPRN = std::min(a: NPRN + 4, b: 4u); |
401 | break; |
402 | default: |
403 | if (BT->isSVESizelessBuiltinType()) |
404 | NSRN = std::min( |
405 | a: NSRN + getContext().getBuiltinVectorTypeInfo(VecTy: BT).NumVectors, |
406 | b: 8u); |
407 | } |
408 | } |
409 | } |
410 | |
411 | return (isPromotableIntegerTypeForABI(Ty) && isDarwinPCS() |
412 | ? ABIArgInfo::getExtend(Ty, T: CGT.ConvertType(T: Ty)) |
413 | : ABIArgInfo::getDirect()); |
414 | } |
415 | |
416 | // Structures with either a non-trivial destructor or a non-trivial |
417 | // copy constructor are always indirect. |
418 | if (CGCXXABI::RecordArgABI RAA = getRecordArgABI(T: Ty, CXXABI&: getCXXABI())) { |
419 | return getNaturalAlignIndirect( |
420 | Ty, /*AddrSpace=*/getDataLayout().getAllocaAddrSpace(), |
421 | /*ByVal=*/RAA == CGCXXABI::RAA_DirectInMemory); |
422 | } |
423 | |
424 | // Empty records: |
425 | uint64_t Size = getContext().getTypeSize(T: Ty); |
426 | bool IsEmpty = isEmptyRecord(Context&: getContext(), T: Ty, AllowArrays: true); |
427 | if (!Ty->isSVESizelessBuiltinType() && (IsEmpty || Size == 0)) { |
428 | // Empty records are ignored in C mode, and in C++ on Darwin. |
429 | if (!getContext().getLangOpts().CPlusPlus || isDarwinPCS()) |
430 | return ABIArgInfo::getIgnore(); |
431 | |
432 | // In C++ mode, arguments which have sizeof() == 0 (which are non-standard |
433 | // C++) are ignored. This isn't defined by any standard, so we copy GCC's |
434 | // behaviour here. |
435 | if (Size == 0) |
436 | return ABIArgInfo::getIgnore(); |
437 | |
438 | // Otherwise, they are passed as if they have a size of 1 byte. |
439 | return ABIArgInfo::getDirect(T: llvm::Type::getInt8Ty(C&: getVMContext())); |
440 | } |
441 | |
442 | // Homogeneous Floating-point Aggregates (HFAs) need to be expanded. |
443 | const Type *Base = nullptr; |
444 | uint64_t Members = 0; |
445 | bool IsWin64 = Kind == AArch64ABIKind::Win64 || |
446 | CallingConvention == llvm::CallingConv::Win64; |
447 | bool IsWinVariadic = IsWin64 && IsVariadicFn; |
448 | // In variadic functions on Windows, all composite types are treated alike, |
449 | // no special handling of HFAs/HVAs. |
450 | if (!IsWinVariadic && isHomogeneousAggregate(Ty, Base, Members)) { |
451 | NSRN = std::min(a: NSRN + Members, b: uint64_t(8)); |
452 | if (Kind != AArch64ABIKind::AAPCS) |
453 | return ABIArgInfo::getDirect( |
454 | T: llvm::ArrayType::get(ElementType: CGT.ConvertType(T: QualType(Base, 0)), NumElements: Members)); |
455 | |
456 | // For HFAs/HVAs, cap the argument alignment to 16, otherwise |
457 | // set it to 8 according to the AAPCS64 document. |
458 | unsigned Align = |
459 | getContext().getTypeUnadjustedAlignInChars(T: Ty).getQuantity(); |
460 | Align = (Align >= 16) ? 16 : 8; |
461 | return ABIArgInfo::getDirect( |
462 | T: llvm::ArrayType::get(ElementType: CGT.ConvertType(T: QualType(Base, 0)), NumElements: Members), Offset: 0, |
463 | Padding: nullptr, CanBeFlattened: true, Align); |
464 | } |
465 | |
466 | // In AAPCS named arguments of a Pure Scalable Type are passed expanded in |
467 | // registers, or indirectly if there are not enough registers. |
468 | if (Kind == AArch64ABIKind::AAPCS) { |
469 | unsigned NVec = 0, NPred = 0; |
470 | SmallVector<llvm::Type *> UnpaddedCoerceToSeq; |
471 | if (passAsPureScalableType(Ty, NV&: NVec, NP&: NPred, CoerceToSeq&: UnpaddedCoerceToSeq) && |
472 | (NVec + NPred) > 0) |
473 | return coerceAndExpandPureScalableAggregate( |
474 | Ty, IsNamedArg, NVec, NPred, UnpaddedCoerceToSeq, NSRN, NPRN); |
475 | } |
476 | |
477 | // Aggregates <= 16 bytes are passed directly in registers or on the stack. |
478 | if (Size <= 128) { |
479 | unsigned Alignment; |
480 | if (Kind == AArch64ABIKind::AAPCS) { |
481 | Alignment = getContext().getTypeUnadjustedAlign(T: Ty); |
482 | Alignment = Alignment < 128 ? 64 : 128; |
483 | } else { |
484 | Alignment = |
485 | std::max(a: getContext().getTypeAlign(T: Ty), |
486 | b: (unsigned)getTarget().getPointerWidth(AddrSpace: LangAS::Default)); |
487 | } |
488 | Size = llvm::alignTo(Value: Size, Align: Alignment); |
489 | |
490 | // If the Aggregate is made up of pointers, use an array of pointers for the |
491 | // coerced type. This prevents having to convert ptr2int->int2ptr through |
492 | // the call, allowing alias analysis to produce better code. |
493 | auto ContainsOnlyPointers = [&](const auto &Self, QualType Ty) { |
494 | if (isEmptyRecord(Context&: getContext(), T: Ty, AllowArrays: true)) |
495 | return false; |
496 | const RecordType *RT = Ty->getAs<RecordType>(); |
497 | if (!RT) |
498 | return false; |
499 | const RecordDecl *RD = RT->getDecl(); |
500 | if (const CXXRecordDecl *CXXRD = dyn_cast<CXXRecordDecl>(Val: RD)) { |
501 | for (const auto &I : CXXRD->bases()) |
502 | if (!Self(Self, I.getType())) |
503 | return false; |
504 | } |
505 | return all_of(RD->fields(), [&](FieldDecl *FD) { |
506 | QualType FDTy = FD->getType(); |
507 | if (FDTy->isArrayType()) |
508 | FDTy = getContext().getBaseElementType(QT: FDTy); |
509 | return (FDTy->isPointerOrReferenceType() && |
510 | getContext().getTypeSize(T: FDTy) == 64 && |
511 | !FDTy->getPointeeType().hasAddressSpace()) || |
512 | Self(Self, FDTy); |
513 | }); |
514 | }; |
515 | |
516 | // We use a pair of i64 for 16-byte aggregate with 8-byte alignment. |
517 | // For aggregates with 16-byte alignment, we use i128. |
518 | llvm::Type *BaseTy = llvm::Type::getIntNTy(C&: getVMContext(), N: Alignment); |
519 | if ((Size == 64 || Size == 128) && Alignment == 64 && |
520 | ContainsOnlyPointers(ContainsOnlyPointers, Ty)) |
521 | BaseTy = llvm::PointerType::getUnqual(C&: getVMContext()); |
522 | return ABIArgInfo::getDirect( |
523 | T: Size == Alignment ? BaseTy |
524 | : llvm::ArrayType::get(ElementType: BaseTy, NumElements: Size / Alignment)); |
525 | } |
526 | |
527 | return getNaturalAlignIndirect(Ty, AddrSpace: getDataLayout().getAllocaAddrSpace(), |
528 | /*ByVal=*/false); |
529 | } |
530 | |
531 | ABIArgInfo AArch64ABIInfo::classifyReturnType(QualType RetTy, |
532 | bool IsVariadicFn) const { |
533 | if (RetTy->isVoidType()) |
534 | return ABIArgInfo::getIgnore(); |
535 | |
536 | if (const auto *VT = RetTy->getAs<VectorType>()) { |
537 | if (VT->getVectorKind() == VectorKind::SveFixedLengthData || |
538 | VT->getVectorKind() == VectorKind::SveFixedLengthPredicate) { |
539 | unsigned NSRN = 0, NPRN = 0; |
540 | return coerceIllegalVector(Ty: RetTy, NSRN, NPRN); |
541 | } |
542 | } |
543 | |
544 | // Large vector types should be returned via memory. |
545 | if (RetTy->isVectorType() && getContext().getTypeSize(T: RetTy) > 128) |
546 | return getNaturalAlignIndirect(Ty: RetTy, AddrSpace: getDataLayout().getAllocaAddrSpace()); |
547 | |
548 | if (!passAsAggregateType(Ty: RetTy)) { |
549 | // Treat an enum type as its underlying type. |
550 | if (const EnumType *EnumTy = RetTy->getAs<EnumType>()) |
551 | RetTy = EnumTy->getDecl()->getIntegerType(); |
552 | |
553 | if (const auto *EIT = RetTy->getAs<BitIntType>()) |
554 | if (EIT->getNumBits() > 128) |
555 | return getNaturalAlignIndirect(Ty: RetTy, |
556 | AddrSpace: getDataLayout().getAllocaAddrSpace()); |
557 | |
558 | return (isPromotableIntegerTypeForABI(Ty: RetTy) && isDarwinPCS() |
559 | ? ABIArgInfo::getExtend(Ty: RetTy) |
560 | : ABIArgInfo::getDirect()); |
561 | } |
562 | |
563 | uint64_t Size = getContext().getTypeSize(T: RetTy); |
564 | if (!RetTy->isSVESizelessBuiltinType() && |
565 | (isEmptyRecord(Context&: getContext(), T: RetTy, AllowArrays: true) || Size == 0)) |
566 | return ABIArgInfo::getIgnore(); |
567 | |
568 | const Type *Base = nullptr; |
569 | uint64_t Members = 0; |
570 | if (isHomogeneousAggregate(Ty: RetTy, Base, Members) && |
571 | !(getTarget().getTriple().getArch() == llvm::Triple::aarch64_32 && |
572 | IsVariadicFn)) |
573 | // Homogeneous Floating-point Aggregates (HFAs) are returned directly. |
574 | return ABIArgInfo::getDirect(); |
575 | |
576 | // In AAPCS return values of a Pure Scalable type are treated as a single |
577 | // named argument and passed expanded in registers, or indirectly if there are |
578 | // not enough registers. |
579 | if (Kind == AArch64ABIKind::AAPCS) { |
580 | unsigned NSRN = 0, NPRN = 0; |
581 | unsigned NVec = 0, NPred = 0; |
582 | SmallVector<llvm::Type *> UnpaddedCoerceToSeq; |
583 | if (passAsPureScalableType(Ty: RetTy, NV&: NVec, NP&: NPred, CoerceToSeq&: UnpaddedCoerceToSeq) && |
584 | (NVec + NPred) > 0) |
585 | return coerceAndExpandPureScalableAggregate( |
586 | Ty: RetTy, /* IsNamedArg */ true, NVec, NPred, UnpaddedCoerceToSeq, NSRN, |
587 | NPRN); |
588 | } |
589 | |
590 | // Aggregates <= 16 bytes are returned directly in registers or on the stack. |
591 | if (Size <= 128) { |
592 | if (Size <= 64 && getDataLayout().isLittleEndian()) { |
593 | // Composite types are returned in lower bits of a 64-bit register for LE, |
594 | // and in higher bits for BE. However, integer types are always returned |
595 | // in lower bits for both LE and BE, and they are not rounded up to |
596 | // 64-bits. We can skip rounding up of composite types for LE, but not for |
597 | // BE, otherwise composite types will be indistinguishable from integer |
598 | // types. |
599 | return ABIArgInfo::getDirect( |
600 | T: llvm::IntegerType::get(C&: getVMContext(), NumBits: Size)); |
601 | } |
602 | |
603 | unsigned Alignment = getContext().getTypeAlign(T: RetTy); |
604 | Size = llvm::alignTo(Value: Size, Align: 64); // round up to multiple of 8 bytes |
605 | |
606 | // We use a pair of i64 for 16-byte aggregate with 8-byte alignment. |
607 | // For aggregates with 16-byte alignment, we use i128. |
608 | if (Alignment < 128 && Size == 128) { |
609 | llvm::Type *BaseTy = llvm::Type::getInt64Ty(C&: getVMContext()); |
610 | return ABIArgInfo::getDirect(T: llvm::ArrayType::get(ElementType: BaseTy, NumElements: Size / 64)); |
611 | } |
612 | return ABIArgInfo::getDirect(T: llvm::IntegerType::get(C&: getVMContext(), NumBits: Size)); |
613 | } |
614 | |
615 | return getNaturalAlignIndirect(Ty: RetTy, AddrSpace: getDataLayout().getAllocaAddrSpace()); |
616 | } |
617 | |
618 | /// isIllegalVectorType - check whether the vector type is legal for AArch64. |
619 | bool AArch64ABIInfo::isIllegalVectorType(QualType Ty) const { |
620 | if (const VectorType *VT = Ty->getAs<VectorType>()) { |
621 | // Check whether VT is a fixed-length SVE vector. These types are |
622 | // represented as scalable vectors in function args/return and must be |
623 | // coerced from fixed vectors. |
624 | if (VT->getVectorKind() == VectorKind::SveFixedLengthData || |
625 | VT->getVectorKind() == VectorKind::SveFixedLengthPredicate) |
626 | return true; |
627 | |
628 | // Check whether VT is legal. |
629 | unsigned NumElements = VT->getNumElements(); |
630 | uint64_t Size = getContext().getTypeSize(T: VT); |
631 | // NumElements should be power of 2. |
632 | if (!llvm::isPowerOf2_32(Value: NumElements)) |
633 | return true; |
634 | |
635 | // arm64_32 has to be compatible with the ARM logic here, which allows huge |
636 | // vectors for some reason. |
637 | llvm::Triple Triple = getTarget().getTriple(); |
638 | if (Triple.getArch() == llvm::Triple::aarch64_32 && |
639 | Triple.isOSBinFormatMachO()) |
640 | return Size <= 32; |
641 | |
642 | return Size != 64 && (Size != 128 || NumElements == 1); |
643 | } |
644 | return false; |
645 | } |
646 | |
647 | bool AArch64SwiftABIInfo::isLegalVectorType(CharUnits VectorSize, |
648 | llvm::Type *EltTy, |
649 | unsigned NumElts) const { |
650 | if (!llvm::isPowerOf2_32(Value: NumElts)) |
651 | return false; |
652 | if (VectorSize.getQuantity() != 8 && |
653 | (VectorSize.getQuantity() != 16 || NumElts == 1)) |
654 | return false; |
655 | return true; |
656 | } |
657 | |
658 | bool AArch64ABIInfo::isHomogeneousAggregateBaseType(QualType Ty) const { |
659 | // For the soft-float ABI variant, no types are considered to be homogeneous |
660 | // aggregates. |
661 | if (isSoftFloat()) |
662 | return false; |
663 | |
664 | // Homogeneous aggregates for AAPCS64 must have base types of a floating |
665 | // point type or a short-vector type. This is the same as the 32-bit ABI, |
666 | // but with the difference that any floating-point type is allowed, |
667 | // including __fp16. |
668 | if (const BuiltinType *BT = Ty->getAs<BuiltinType>()) { |
669 | if (BT->isFloatingPoint()) |
670 | return true; |
671 | } else if (const VectorType *VT = Ty->getAs<VectorType>()) { |
672 | if (auto Kind = VT->getVectorKind(); |
673 | Kind == VectorKind::SveFixedLengthData || |
674 | Kind == VectorKind::SveFixedLengthPredicate) |
675 | return false; |
676 | |
677 | unsigned VecSize = getContext().getTypeSize(T: VT); |
678 | if (VecSize == 64 || VecSize == 128) |
679 | return true; |
680 | } |
681 | return false; |
682 | } |
683 | |
684 | bool AArch64ABIInfo::isHomogeneousAggregateSmallEnough(const Type *Base, |
685 | uint64_t Members) const { |
686 | return Members <= 4; |
687 | } |
688 | |
689 | bool AArch64ABIInfo::isZeroLengthBitfieldPermittedInHomogeneousAggregate() |
690 | const { |
691 | // AAPCS64 says that the rule for whether something is a homogeneous |
692 | // aggregate is applied to the output of the data layout decision. So |
693 | // anything that doesn't affect the data layout also does not affect |
694 | // homogeneity. In particular, zero-length bitfields don't stop a struct |
695 | // being homogeneous. |
696 | return true; |
697 | } |
698 | |
699 | bool AArch64ABIInfo::passAsAggregateType(QualType Ty) const { |
700 | if (Kind == AArch64ABIKind::AAPCS && Ty->isSVESizelessBuiltinType()) { |
701 | const auto *BT = Ty->castAs<BuiltinType>(); |
702 | return !BT->isSVECount() && |
703 | getContext().getBuiltinVectorTypeInfo(VecTy: BT).NumVectors > 1; |
704 | } |
705 | return isAggregateTypeForABI(T: Ty); |
706 | } |
707 | |
708 | // Check if a type needs to be passed in registers as a Pure Scalable Type (as |
709 | // defined by AAPCS64). Return the number of data vectors and the number of |
710 | // predicate vectors in the type, into `NVec` and `NPred`, respectively. Upon |
711 | // return `CoerceToSeq` contains an expanded sequence of LLVM IR types, one |
712 | // element for each non-composite member. For practical purposes, limit the |
713 | // length of `CoerceToSeq` to about 12 (the maximum that could possibly fit |
714 | // in registers) and return false, the effect of which will be to pass the |
715 | // argument under the rules for a large (> 128 bytes) composite. |
716 | bool AArch64ABIInfo::passAsPureScalableType( |
717 | QualType Ty, unsigned &NVec, unsigned &NPred, |
718 | SmallVectorImpl<llvm::Type *> &CoerceToSeq) const { |
719 | if (const ConstantArrayType *AT = getContext().getAsConstantArrayType(T: Ty)) { |
720 | uint64_t NElt = AT->getZExtSize(); |
721 | if (NElt == 0) |
722 | return false; |
723 | |
724 | unsigned NV = 0, NP = 0; |
725 | SmallVector<llvm::Type *> EltCoerceToSeq; |
726 | if (!passAsPureScalableType(Ty: AT->getElementType(), NVec&: NV, NPred&: NP, CoerceToSeq&: EltCoerceToSeq)) |
727 | return false; |
728 | |
729 | if (CoerceToSeq.size() + NElt * EltCoerceToSeq.size() > 12) |
730 | return false; |
731 | |
732 | for (uint64_t I = 0; I < NElt; ++I) |
733 | llvm::append_range(C&: CoerceToSeq, R&: EltCoerceToSeq); |
734 | |
735 | NVec += NElt * NV; |
736 | NPred += NElt * NP; |
737 | return true; |
738 | } |
739 | |
740 | if (const RecordType *RT = Ty->getAs<RecordType>()) { |
741 | // If the record cannot be passed in registers, then it's not a PST. |
742 | if (CGCXXABI::RecordArgABI RAA = getRecordArgABI(RT, CXXABI&: getCXXABI()); |
743 | RAA != CGCXXABI::RAA_Default) |
744 | return false; |
745 | |
746 | // Pure scalable types are never unions and never contain unions. |
747 | const RecordDecl *RD = RT->getDecl(); |
748 | if (RD->isUnion()) |
749 | return false; |
750 | |
751 | // If this is a C++ record, check the bases. |
752 | if (const CXXRecordDecl *CXXRD = dyn_cast<CXXRecordDecl>(Val: RD)) { |
753 | for (const auto &I : CXXRD->bases()) { |
754 | if (isEmptyRecord(Context&: getContext(), T: I.getType(), AllowArrays: true)) |
755 | continue; |
756 | if (!passAsPureScalableType(Ty: I.getType(), NVec, NPred, CoerceToSeq)) |
757 | return false; |
758 | } |
759 | } |
760 | |
761 | // Check members. |
762 | for (const auto *FD : RD->fields()) { |
763 | QualType FT = FD->getType(); |
764 | if (isEmptyField(Context&: getContext(), FD, /* AllowArrays */ true)) |
765 | continue; |
766 | if (!passAsPureScalableType(Ty: FT, NVec, NPred, CoerceToSeq)) |
767 | return false; |
768 | } |
769 | |
770 | return true; |
771 | } |
772 | |
773 | if (const auto *VT = Ty->getAs<VectorType>()) { |
774 | if (VT->getVectorKind() == VectorKind::SveFixedLengthPredicate) { |
775 | ++NPred; |
776 | if (CoerceToSeq.size() + 1 > 12) |
777 | return false; |
778 | CoerceToSeq.push_back(Elt: convertFixedToScalableVectorType(VT)); |
779 | return true; |
780 | } |
781 | |
782 | if (VT->getVectorKind() == VectorKind::SveFixedLengthData) { |
783 | ++NVec; |
784 | if (CoerceToSeq.size() + 1 > 12) |
785 | return false; |
786 | CoerceToSeq.push_back(Elt: convertFixedToScalableVectorType(VT)); |
787 | return true; |
788 | } |
789 | |
790 | return false; |
791 | } |
792 | |
793 | if (!Ty->isBuiltinType()) |
794 | return false; |
795 | |
796 | bool isPredicate; |
797 | switch (Ty->castAs<BuiltinType>()->getKind()) { |
798 | #define SVE_VECTOR_TYPE(Name, MangledName, Id, SingletonId) \ |
799 | case BuiltinType::Id: \ |
800 | isPredicate = false; \ |
801 | break; |
802 | #define SVE_PREDICATE_TYPE(Name, MangledName, Id, SingletonId) \ |
803 | case BuiltinType::Id: \ |
804 | isPredicate = true; \ |
805 | break; |
806 | #include "clang/Basic/AArch64ACLETypes.def" |
807 | default: |
808 | return false; |
809 | } |
810 | |
811 | ASTContext::BuiltinVectorTypeInfo Info = |
812 | getContext().getBuiltinVectorTypeInfo(VecTy: cast<BuiltinType>(Val&: Ty)); |
813 | assert(Info.NumVectors > 0 && Info.NumVectors <= 4 && |
814 | "Expected 1, 2, 3 or 4 vectors!" ); |
815 | if (isPredicate) |
816 | NPred += Info.NumVectors; |
817 | else |
818 | NVec += Info.NumVectors; |
819 | llvm::Type *EltTy = Info.ElementType->isMFloat8Type() |
820 | ? llvm::Type::getInt8Ty(C&: getVMContext()) |
821 | : CGT.ConvertType(T: Info.ElementType); |
822 | auto *VTy = llvm::ScalableVectorType::get(ElementType: EltTy, MinNumElts: Info.EC.getKnownMinValue()); |
823 | |
824 | if (CoerceToSeq.size() + Info.NumVectors > 12) |
825 | return false; |
826 | std::fill_n(first: std::back_inserter(x&: CoerceToSeq), n: Info.NumVectors, value: VTy); |
827 | |
828 | return true; |
829 | } |
830 | |
831 | // Expand an LLVM IR type into a sequence with a element for each non-struct, |
832 | // non-array member of the type, with the exception of the padding types, which |
833 | // are retained. |
834 | void AArch64ABIInfo::flattenType( |
835 | llvm::Type *Ty, SmallVectorImpl<llvm::Type *> &Flattened) const { |
836 | |
837 | if (ABIArgInfo::isPaddingForCoerceAndExpand(eltType: Ty)) { |
838 | Flattened.push_back(Elt: Ty); |
839 | return; |
840 | } |
841 | |
842 | if (const auto *AT = dyn_cast<llvm::ArrayType>(Val: Ty)) { |
843 | uint64_t NElt = AT->getNumElements(); |
844 | if (NElt == 0) |
845 | return; |
846 | |
847 | SmallVector<llvm::Type *> EltFlattened; |
848 | flattenType(Ty: AT->getElementType(), Flattened&: EltFlattened); |
849 | |
850 | for (uint64_t I = 0; I < NElt; ++I) |
851 | llvm::append_range(C&: Flattened, R&: EltFlattened); |
852 | return; |
853 | } |
854 | |
855 | if (const auto *ST = dyn_cast<llvm::StructType>(Val: Ty)) { |
856 | for (auto *ET : ST->elements()) |
857 | flattenType(Ty: ET, Flattened); |
858 | return; |
859 | } |
860 | |
861 | Flattened.push_back(Elt: Ty); |
862 | } |
863 | |
864 | RValue AArch64ABIInfo::EmitAAPCSVAArg(Address VAListAddr, QualType Ty, |
865 | CodeGenFunction &CGF, AArch64ABIKind Kind, |
866 | AggValueSlot Slot) const { |
867 | // These numbers are not used for variadic arguments, hence it doesn't matter |
868 | // they don't retain their values across multiple calls to |
869 | // `classifyArgumentType` here. |
870 | unsigned NSRN = 0, NPRN = 0; |
871 | ABIArgInfo AI = |
872 | classifyArgumentType(Ty, /*IsVariadicFn=*/true, /* IsNamedArg */ false, |
873 | CallingConvention: CGF.CurFnInfo->getCallingConvention(), NSRN, NPRN); |
874 | // Empty records are ignored for parameter passing purposes. |
875 | if (AI.isIgnore()) |
876 | return Slot.asRValue(); |
877 | |
878 | bool IsIndirect = AI.isIndirect(); |
879 | |
880 | llvm::Type *BaseTy = CGF.ConvertType(T: Ty); |
881 | if (IsIndirect) |
882 | BaseTy = llvm::PointerType::getUnqual(C&: BaseTy->getContext()); |
883 | else if (AI.getCoerceToType()) |
884 | BaseTy = AI.getCoerceToType(); |
885 | |
886 | unsigned NumRegs = 1; |
887 | if (llvm::ArrayType *ArrTy = dyn_cast<llvm::ArrayType>(Val: BaseTy)) { |
888 | BaseTy = ArrTy->getElementType(); |
889 | NumRegs = ArrTy->getNumElements(); |
890 | } |
891 | bool IsFPR = |
892 | !isSoftFloat() && (BaseTy->isFloatingPointTy() || BaseTy->isVectorTy()); |
893 | |
894 | // The AArch64 va_list type and handling is specified in the Procedure Call |
895 | // Standard, section B.4: |
896 | // |
897 | // struct { |
898 | // void *__stack; |
899 | // void *__gr_top; |
900 | // void *__vr_top; |
901 | // int __gr_offs; |
902 | // int __vr_offs; |
903 | // }; |
904 | |
905 | llvm::BasicBlock *MaybeRegBlock = CGF.createBasicBlock(name: "vaarg.maybe_reg" ); |
906 | llvm::BasicBlock *InRegBlock = CGF.createBasicBlock(name: "vaarg.in_reg" ); |
907 | llvm::BasicBlock *OnStackBlock = CGF.createBasicBlock(name: "vaarg.on_stack" ); |
908 | llvm::BasicBlock *ContBlock = CGF.createBasicBlock(name: "vaarg.end" ); |
909 | |
910 | CharUnits TySize = getContext().getTypeSizeInChars(T: Ty); |
911 | CharUnits TyAlign = getContext().getTypeUnadjustedAlignInChars(T: Ty); |
912 | |
913 | Address reg_offs_p = Address::invalid(); |
914 | llvm::Value *reg_offs = nullptr; |
915 | int reg_top_index; |
916 | int RegSize = IsIndirect ? 8 : TySize.getQuantity(); |
917 | if (!IsFPR) { |
918 | // 3 is the field number of __gr_offs |
919 | reg_offs_p = CGF.Builder.CreateStructGEP(Addr: VAListAddr, Index: 3, Name: "gr_offs_p" ); |
920 | reg_offs = CGF.Builder.CreateLoad(Addr: reg_offs_p, Name: "gr_offs" ); |
921 | reg_top_index = 1; // field number for __gr_top |
922 | RegSize = llvm::alignTo(Value: RegSize, Align: 8); |
923 | } else { |
924 | // 4 is the field number of __vr_offs. |
925 | reg_offs_p = CGF.Builder.CreateStructGEP(Addr: VAListAddr, Index: 4, Name: "vr_offs_p" ); |
926 | reg_offs = CGF.Builder.CreateLoad(Addr: reg_offs_p, Name: "vr_offs" ); |
927 | reg_top_index = 2; // field number for __vr_top |
928 | RegSize = 16 * NumRegs; |
929 | } |
930 | |
931 | //======================================= |
932 | // Find out where argument was passed |
933 | //======================================= |
934 | |
935 | // If reg_offs >= 0 we're already using the stack for this type of |
936 | // argument. We don't want to keep updating reg_offs (in case it overflows, |
937 | // though anyone passing 2GB of arguments, each at most 16 bytes, deserves |
938 | // whatever they get). |
939 | llvm::Value *UsingStack = nullptr; |
940 | UsingStack = CGF.Builder.CreateICmpSGE( |
941 | LHS: reg_offs, RHS: llvm::ConstantInt::get(Ty: CGF.Int32Ty, V: 0)); |
942 | |
943 | CGF.Builder.CreateCondBr(Cond: UsingStack, True: OnStackBlock, False: MaybeRegBlock); |
944 | |
945 | // Otherwise, at least some kind of argument could go in these registers, the |
946 | // question is whether this particular type is too big. |
947 | CGF.EmitBlock(BB: MaybeRegBlock); |
948 | |
949 | // Integer arguments may need to correct register alignment (for example a |
950 | // "struct { __int128 a; };" gets passed in x_2N, x_{2N+1}). In this case we |
951 | // align __gr_offs to calculate the potential address. |
952 | if (!IsFPR && !IsIndirect && TyAlign.getQuantity() > 8) { |
953 | int Align = TyAlign.getQuantity(); |
954 | |
955 | reg_offs = CGF.Builder.CreateAdd( |
956 | LHS: reg_offs, RHS: llvm::ConstantInt::get(Ty: CGF.Int32Ty, V: Align - 1), |
957 | Name: "align_regoffs" ); |
958 | reg_offs = CGF.Builder.CreateAnd( |
959 | LHS: reg_offs, RHS: llvm::ConstantInt::get(Ty: CGF.Int32Ty, V: -Align), |
960 | Name: "aligned_regoffs" ); |
961 | } |
962 | |
963 | // Update the gr_offs/vr_offs pointer for next call to va_arg on this va_list. |
964 | // The fact that this is done unconditionally reflects the fact that |
965 | // allocating an argument to the stack also uses up all the remaining |
966 | // registers of the appropriate kind. |
967 | llvm::Value *NewOffset = nullptr; |
968 | NewOffset = CGF.Builder.CreateAdd( |
969 | LHS: reg_offs, RHS: llvm::ConstantInt::get(Ty: CGF.Int32Ty, V: RegSize), Name: "new_reg_offs" ); |
970 | CGF.Builder.CreateStore(Val: NewOffset, Addr: reg_offs_p); |
971 | |
972 | // Now we're in a position to decide whether this argument really was in |
973 | // registers or not. |
974 | llvm::Value *InRegs = nullptr; |
975 | InRegs = CGF.Builder.CreateICmpSLE( |
976 | LHS: NewOffset, RHS: llvm::ConstantInt::get(Ty: CGF.Int32Ty, V: 0), Name: "inreg" ); |
977 | |
978 | CGF.Builder.CreateCondBr(Cond: InRegs, True: InRegBlock, False: OnStackBlock); |
979 | |
980 | //======================================= |
981 | // Argument was in registers |
982 | //======================================= |
983 | |
984 | // Now we emit the code for if the argument was originally passed in |
985 | // registers. First start the appropriate block: |
986 | CGF.EmitBlock(BB: InRegBlock); |
987 | |
988 | llvm::Value *reg_top = nullptr; |
989 | Address reg_top_p = |
990 | CGF.Builder.CreateStructGEP(Addr: VAListAddr, Index: reg_top_index, Name: "reg_top_p" ); |
991 | reg_top = CGF.Builder.CreateLoad(Addr: reg_top_p, Name: "reg_top" ); |
992 | Address BaseAddr(CGF.Builder.CreateInBoundsGEP(Ty: CGF.Int8Ty, Ptr: reg_top, IdxList: reg_offs), |
993 | CGF.Int8Ty, CharUnits::fromQuantity(Quantity: IsFPR ? 16 : 8)); |
994 | Address RegAddr = Address::invalid(); |
995 | llvm::Type *MemTy = CGF.ConvertTypeForMem(T: Ty), *ElementTy = MemTy; |
996 | |
997 | if (IsIndirect) { |
998 | // If it's been passed indirectly (actually a struct), whatever we find from |
999 | // stored registers or on the stack will actually be a struct **. |
1000 | MemTy = llvm::PointerType::getUnqual(C&: MemTy->getContext()); |
1001 | } |
1002 | |
1003 | const Type *Base = nullptr; |
1004 | uint64_t NumMembers = 0; |
1005 | bool IsHFA = isHomogeneousAggregate(Ty, Base, Members&: NumMembers); |
1006 | if (IsHFA && NumMembers > 1) { |
1007 | // Homogeneous aggregates passed in registers will have their elements split |
1008 | // and stored 16-bytes apart regardless of size (they're notionally in qN, |
1009 | // qN+1, ...). We reload and store into a temporary local variable |
1010 | // contiguously. |
1011 | assert(!IsIndirect && "Homogeneous aggregates should be passed directly" ); |
1012 | auto BaseTyInfo = getContext().getTypeInfoInChars(T: QualType(Base, 0)); |
1013 | llvm::Type *BaseTy = CGF.ConvertType(T: QualType(Base, 0)); |
1014 | llvm::Type *HFATy = llvm::ArrayType::get(ElementType: BaseTy, NumElements: NumMembers); |
1015 | Address Tmp = CGF.CreateTempAlloca(Ty: HFATy, |
1016 | align: std::max(a: TyAlign, b: BaseTyInfo.Align)); |
1017 | |
1018 | // On big-endian platforms, the value will be right-aligned in its slot. |
1019 | int Offset = 0; |
1020 | if (CGF.CGM.getDataLayout().isBigEndian() && |
1021 | BaseTyInfo.Width.getQuantity() < 16) |
1022 | Offset = 16 - BaseTyInfo.Width.getQuantity(); |
1023 | |
1024 | for (unsigned i = 0; i < NumMembers; ++i) { |
1025 | CharUnits BaseOffset = CharUnits::fromQuantity(Quantity: 16 * i + Offset); |
1026 | Address LoadAddr = |
1027 | CGF.Builder.CreateConstInBoundsByteGEP(Addr: BaseAddr, Offset: BaseOffset); |
1028 | LoadAddr = LoadAddr.withElementType(ElemTy: BaseTy); |
1029 | |
1030 | Address StoreAddr = CGF.Builder.CreateConstArrayGEP(Addr: Tmp, Index: i); |
1031 | |
1032 | llvm::Value *Elem = CGF.Builder.CreateLoad(Addr: LoadAddr); |
1033 | CGF.Builder.CreateStore(Val: Elem, Addr: StoreAddr); |
1034 | } |
1035 | |
1036 | RegAddr = Tmp.withElementType(ElemTy: MemTy); |
1037 | } else { |
1038 | // Otherwise the object is contiguous in memory. |
1039 | |
1040 | // It might be right-aligned in its slot. |
1041 | CharUnits SlotSize = BaseAddr.getAlignment(); |
1042 | if (CGF.CGM.getDataLayout().isBigEndian() && !IsIndirect && |
1043 | (IsHFA || !isAggregateTypeForABI(T: Ty)) && |
1044 | TySize < SlotSize) { |
1045 | CharUnits Offset = SlotSize - TySize; |
1046 | BaseAddr = CGF.Builder.CreateConstInBoundsByteGEP(Addr: BaseAddr, Offset); |
1047 | } |
1048 | |
1049 | RegAddr = BaseAddr.withElementType(ElemTy: MemTy); |
1050 | } |
1051 | |
1052 | CGF.EmitBranch(Block: ContBlock); |
1053 | |
1054 | //======================================= |
1055 | // Argument was on the stack |
1056 | //======================================= |
1057 | CGF.EmitBlock(BB: OnStackBlock); |
1058 | |
1059 | Address stack_p = CGF.Builder.CreateStructGEP(Addr: VAListAddr, Index: 0, Name: "stack_p" ); |
1060 | llvm::Value *OnStackPtr = CGF.Builder.CreateLoad(Addr: stack_p, Name: "stack" ); |
1061 | |
1062 | // Again, stack arguments may need realignment. In this case both integer and |
1063 | // floating-point ones might be affected. |
1064 | if (!IsIndirect && TyAlign.getQuantity() > 8) { |
1065 | OnStackPtr = emitRoundPointerUpToAlignment(CGF, Ptr: OnStackPtr, Align: TyAlign); |
1066 | } |
1067 | Address OnStackAddr = Address(OnStackPtr, CGF.Int8Ty, |
1068 | std::max(a: CharUnits::fromQuantity(Quantity: 8), b: TyAlign)); |
1069 | |
1070 | // All stack slots are multiples of 8 bytes. |
1071 | CharUnits StackSlotSize = CharUnits::fromQuantity(Quantity: 8); |
1072 | CharUnits StackSize; |
1073 | if (IsIndirect) |
1074 | StackSize = StackSlotSize; |
1075 | else |
1076 | StackSize = TySize.alignTo(Align: StackSlotSize); |
1077 | |
1078 | llvm::Value *StackSizeC = CGF.Builder.getSize(N: StackSize); |
1079 | llvm::Value *NewStack = CGF.Builder.CreateInBoundsGEP( |
1080 | Ty: CGF.Int8Ty, Ptr: OnStackPtr, IdxList: StackSizeC, Name: "new_stack" ); |
1081 | |
1082 | // Write the new value of __stack for the next call to va_arg |
1083 | CGF.Builder.CreateStore(Val: NewStack, Addr: stack_p); |
1084 | |
1085 | if (CGF.CGM.getDataLayout().isBigEndian() && !isAggregateTypeForABI(T: Ty) && |
1086 | TySize < StackSlotSize) { |
1087 | CharUnits Offset = StackSlotSize - TySize; |
1088 | OnStackAddr = CGF.Builder.CreateConstInBoundsByteGEP(Addr: OnStackAddr, Offset); |
1089 | } |
1090 | |
1091 | OnStackAddr = OnStackAddr.withElementType(ElemTy: MemTy); |
1092 | |
1093 | CGF.EmitBranch(Block: ContBlock); |
1094 | |
1095 | //======================================= |
1096 | // Tidy up |
1097 | //======================================= |
1098 | CGF.EmitBlock(BB: ContBlock); |
1099 | |
1100 | Address ResAddr = emitMergePHI(CGF, Addr1: RegAddr, Block1: InRegBlock, Addr2: OnStackAddr, |
1101 | Block2: OnStackBlock, Name: "vaargs.addr" ); |
1102 | |
1103 | if (IsIndirect) |
1104 | return CGF.EmitLoadOfAnyValue( |
1105 | V: CGF.MakeAddrLValue( |
1106 | Addr: Address(CGF.Builder.CreateLoad(Addr: ResAddr, Name: "vaarg.addr" ), ElementTy, |
1107 | TyAlign), |
1108 | T: Ty), |
1109 | Slot); |
1110 | |
1111 | return CGF.EmitLoadOfAnyValue(V: CGF.MakeAddrLValue(Addr: ResAddr, T: Ty), Slot); |
1112 | } |
1113 | |
1114 | RValue AArch64ABIInfo::EmitDarwinVAArg(Address VAListAddr, QualType Ty, |
1115 | CodeGenFunction &CGF, |
1116 | AggValueSlot Slot) const { |
1117 | // The backend's lowering doesn't support va_arg for aggregates or |
1118 | // illegal vector types. Lower VAArg here for these cases and use |
1119 | // the LLVM va_arg instruction for everything else. |
1120 | if (!isAggregateTypeForABI(T: Ty) && !isIllegalVectorType(Ty)) |
1121 | return CGF.EmitLoadOfAnyValue( |
1122 | V: CGF.MakeAddrLValue( |
1123 | Addr: EmitVAArgInstr(CGF, VAListAddr, Ty, AI: ABIArgInfo::getDirect()), T: Ty), |
1124 | Slot); |
1125 | |
1126 | uint64_t PointerSize = getTarget().getPointerWidth(AddrSpace: LangAS::Default) / 8; |
1127 | CharUnits SlotSize = CharUnits::fromQuantity(Quantity: PointerSize); |
1128 | |
1129 | // Empty records are ignored for parameter passing purposes. |
1130 | if (isEmptyRecord(Context&: getContext(), T: Ty, AllowArrays: true)) |
1131 | return Slot.asRValue(); |
1132 | |
1133 | // The size of the actual thing passed, which might end up just |
1134 | // being a pointer for indirect types. |
1135 | auto TyInfo = getContext().getTypeInfoInChars(T: Ty); |
1136 | |
1137 | // Arguments bigger than 16 bytes which aren't homogeneous |
1138 | // aggregates should be passed indirectly. |
1139 | bool IsIndirect = false; |
1140 | if (TyInfo.Width.getQuantity() > 16) { |
1141 | const Type *Base = nullptr; |
1142 | uint64_t Members = 0; |
1143 | IsIndirect = !isHomogeneousAggregate(Ty, Base, Members); |
1144 | } |
1145 | |
1146 | return emitVoidPtrVAArg(CGF, VAListAddr, ValueTy: Ty, IsIndirect, ValueInfo: TyInfo, SlotSizeAndAlign: SlotSize, |
1147 | /*AllowHigherAlign*/ true, Slot); |
1148 | } |
1149 | |
1150 | RValue AArch64ABIInfo::EmitMSVAArg(CodeGenFunction &CGF, Address VAListAddr, |
1151 | QualType Ty, AggValueSlot Slot) const { |
1152 | bool IsIndirect = false; |
1153 | |
1154 | // Composites larger than 16 bytes are passed by reference. |
1155 | if (isAggregateTypeForABI(T: Ty) && getContext().getTypeSize(T: Ty) > 128) |
1156 | IsIndirect = true; |
1157 | |
1158 | return emitVoidPtrVAArg(CGF, VAListAddr, ValueTy: Ty, IsIndirect, |
1159 | ValueInfo: CGF.getContext().getTypeInfoInChars(T: Ty), |
1160 | SlotSizeAndAlign: CharUnits::fromQuantity(Quantity: 8), |
1161 | /*allowHigherAlign*/ AllowHigherAlign: false, Slot); |
1162 | } |
1163 | |
1164 | static bool isStreamingCompatible(const FunctionDecl *F) { |
1165 | if (const auto *T = F->getType()->getAs<FunctionProtoType>()) |
1166 | return T->getAArch64SMEAttributes() & |
1167 | FunctionType::SME_PStateSMCompatibleMask; |
1168 | return false; |
1169 | } |
1170 | |
1171 | // Report an error if an argument or return value of type Ty would need to be |
1172 | // passed in a floating-point register. |
1173 | static void diagnoseIfNeedsFPReg(DiagnosticsEngine &Diags, |
1174 | const StringRef ABIName, |
1175 | const AArch64ABIInfo &ABIInfo, |
1176 | const QualType &Ty, const NamedDecl *D, |
1177 | SourceLocation loc) { |
1178 | const Type *HABase = nullptr; |
1179 | uint64_t HAMembers = 0; |
1180 | if (Ty->isFloatingType() || Ty->isVectorType() || |
1181 | ABIInfo.isHomogeneousAggregate(Ty, Base&: HABase, Members&: HAMembers)) { |
1182 | Diags.Report(Loc: loc, DiagID: diag::err_target_unsupported_type_for_abi) |
1183 | << D->getDeclName() << Ty << ABIName; |
1184 | } |
1185 | } |
1186 | |
1187 | // If we are using a hard-float ABI, but do not have floating point registers, |
1188 | // then report an error for any function arguments or returns which would be |
1189 | // passed in floating-pint registers. |
1190 | void AArch64TargetCodeGenInfo::checkFunctionABI( |
1191 | CodeGenModule &CGM, const FunctionDecl *FuncDecl) const { |
1192 | const AArch64ABIInfo &ABIInfo = getABIInfo<AArch64ABIInfo>(); |
1193 | const TargetInfo &TI = ABIInfo.getContext().getTargetInfo(); |
1194 | |
1195 | if (!TI.hasFeature(Feature: "fp" ) && !ABIInfo.isSoftFloat()) { |
1196 | diagnoseIfNeedsFPReg(Diags&: CGM.getDiags(), ABIName: TI.getABI(), ABIInfo, |
1197 | Ty: FuncDecl->getReturnType(), D: FuncDecl, |
1198 | loc: FuncDecl->getLocation()); |
1199 | for (ParmVarDecl *PVD : FuncDecl->parameters()) { |
1200 | diagnoseIfNeedsFPReg(Diags&: CGM.getDiags(), ABIName: TI.getABI(), ABIInfo, Ty: PVD->getType(), |
1201 | D: PVD, loc: FuncDecl->getLocation()); |
1202 | } |
1203 | } |
1204 | } |
1205 | |
1206 | enum class ArmSMEInlinability : uint8_t { |
1207 | Ok = 0, |
1208 | ErrorCalleeRequiresNewZA = 1 << 0, |
1209 | ErrorCalleeRequiresNewZT0 = 1 << 1, |
1210 | WarnIncompatibleStreamingModes = 1 << 2, |
1211 | ErrorIncompatibleStreamingModes = 1 << 3, |
1212 | |
1213 | IncompatibleStreamingModes = |
1214 | WarnIncompatibleStreamingModes | ErrorIncompatibleStreamingModes, |
1215 | |
1216 | LLVM_MARK_AS_BITMASK_ENUM(/*LargestValue=*/ErrorIncompatibleStreamingModes), |
1217 | }; |
1218 | |
1219 | /// Determines if there are any Arm SME ABI issues with inlining \p Callee into |
1220 | /// \p Caller. Returns the issue (if any) in the ArmSMEInlinability bit enum. |
1221 | static ArmSMEInlinability GetArmSMEInlinability(const FunctionDecl *Caller, |
1222 | const FunctionDecl *Callee) { |
1223 | bool CallerIsStreaming = |
1224 | IsArmStreamingFunction(FD: Caller, /*IncludeLocallyStreaming=*/true); |
1225 | bool CalleeIsStreaming = |
1226 | IsArmStreamingFunction(FD: Callee, /*IncludeLocallyStreaming=*/true); |
1227 | bool CallerIsStreamingCompatible = isStreamingCompatible(F: Caller); |
1228 | bool CalleeIsStreamingCompatible = isStreamingCompatible(F: Callee); |
1229 | |
1230 | ArmSMEInlinability Inlinability = ArmSMEInlinability::Ok; |
1231 | |
1232 | if (!CalleeIsStreamingCompatible && |
1233 | (CallerIsStreaming != CalleeIsStreaming || CallerIsStreamingCompatible)) { |
1234 | if (CalleeIsStreaming) |
1235 | Inlinability |= ArmSMEInlinability::ErrorIncompatibleStreamingModes; |
1236 | else |
1237 | Inlinability |= ArmSMEInlinability::WarnIncompatibleStreamingModes; |
1238 | } |
1239 | if (auto *NewAttr = Callee->getAttr<ArmNewAttr>()) { |
1240 | if (NewAttr->isNewZA()) |
1241 | Inlinability |= ArmSMEInlinability::ErrorCalleeRequiresNewZA; |
1242 | if (NewAttr->isNewZT0()) |
1243 | Inlinability |= ArmSMEInlinability::ErrorCalleeRequiresNewZT0; |
1244 | } |
1245 | |
1246 | return Inlinability; |
1247 | } |
1248 | |
1249 | void AArch64TargetCodeGenInfo::checkFunctionCallABIStreaming( |
1250 | CodeGenModule &CGM, SourceLocation CallLoc, const FunctionDecl *Caller, |
1251 | const FunctionDecl *Callee) const { |
1252 | if (!Caller || !Callee || !Callee->hasAttr<AlwaysInlineAttr>()) |
1253 | return; |
1254 | |
1255 | ArmSMEInlinability Inlinability = GetArmSMEInlinability(Caller, Callee); |
1256 | |
1257 | if ((Inlinability & ArmSMEInlinability::IncompatibleStreamingModes) != |
1258 | ArmSMEInlinability::Ok) |
1259 | CGM.getDiags().Report( |
1260 | Loc: CallLoc, |
1261 | DiagID: (Inlinability & ArmSMEInlinability::ErrorIncompatibleStreamingModes) == |
1262 | ArmSMEInlinability::ErrorIncompatibleStreamingModes |
1263 | ? diag::err_function_always_inline_attribute_mismatch |
1264 | : diag::warn_function_always_inline_attribute_mismatch) |
1265 | << Caller->getDeclName() << Callee->getDeclName() << "streaming" ; |
1266 | |
1267 | if ((Inlinability & ArmSMEInlinability::ErrorCalleeRequiresNewZA) == |
1268 | ArmSMEInlinability::ErrorCalleeRequiresNewZA) |
1269 | CGM.getDiags().Report(Loc: CallLoc, DiagID: diag::err_function_always_inline_new_za) |
1270 | << Callee->getDeclName(); |
1271 | |
1272 | if ((Inlinability & ArmSMEInlinability::ErrorCalleeRequiresNewZT0) == |
1273 | ArmSMEInlinability::ErrorCalleeRequiresNewZT0) |
1274 | CGM.getDiags().Report(Loc: CallLoc, DiagID: diag::err_function_always_inline_new_zt0) |
1275 | << Callee->getDeclName(); |
1276 | } |
1277 | |
1278 | // If the target does not have floating-point registers, but we are using a |
1279 | // hard-float ABI, there is no way to pass floating-point, vector or HFA values |
1280 | // to functions, so we report an error. |
1281 | void AArch64TargetCodeGenInfo::checkFunctionCallABISoftFloat( |
1282 | CodeGenModule &CGM, SourceLocation CallLoc, const FunctionDecl *Caller, |
1283 | const FunctionDecl *Callee, const CallArgList &Args, |
1284 | QualType ReturnType) const { |
1285 | const AArch64ABIInfo &ABIInfo = getABIInfo<AArch64ABIInfo>(); |
1286 | const TargetInfo &TI = ABIInfo.getContext().getTargetInfo(); |
1287 | |
1288 | if (!Caller || TI.hasFeature(Feature: "fp" ) || ABIInfo.isSoftFloat()) |
1289 | return; |
1290 | |
1291 | diagnoseIfNeedsFPReg(Diags&: CGM.getDiags(), ABIName: TI.getABI(), ABIInfo, Ty: ReturnType, |
1292 | D: Callee ? Callee : Caller, loc: CallLoc); |
1293 | |
1294 | for (const CallArg &Arg : Args) |
1295 | diagnoseIfNeedsFPReg(Diags&: CGM.getDiags(), ABIName: TI.getABI(), ABIInfo, Ty: Arg.getType(), |
1296 | D: Callee ? Callee : Caller, loc: CallLoc); |
1297 | } |
1298 | |
1299 | void AArch64TargetCodeGenInfo::checkFunctionCallABI(CodeGenModule &CGM, |
1300 | SourceLocation CallLoc, |
1301 | const FunctionDecl *Caller, |
1302 | const FunctionDecl *Callee, |
1303 | const CallArgList &Args, |
1304 | QualType ReturnType) const { |
1305 | checkFunctionCallABIStreaming(CGM, CallLoc, Caller, Callee); |
1306 | checkFunctionCallABISoftFloat(CGM, CallLoc, Caller, Callee, Args, ReturnType); |
1307 | } |
1308 | |
1309 | bool AArch64TargetCodeGenInfo::wouldInliningViolateFunctionCallABI( |
1310 | const FunctionDecl *Caller, const FunctionDecl *Callee) const { |
1311 | return Caller && Callee && |
1312 | GetArmSMEInlinability(Caller, Callee) != ArmSMEInlinability::Ok; |
1313 | } |
1314 | |
1315 | void AArch64ABIInfo::appendAttributeMangling(TargetClonesAttr *Attr, |
1316 | unsigned Index, |
1317 | raw_ostream &Out) const { |
1318 | appendAttributeMangling(AttrStr: Attr->getFeatureStr(Index), Out); |
1319 | } |
1320 | |
1321 | void AArch64ABIInfo::appendAttributeMangling(StringRef AttrStr, |
1322 | raw_ostream &Out) const { |
1323 | if (AttrStr == "default" ) { |
1324 | Out << ".default" ; |
1325 | return; |
1326 | } |
1327 | |
1328 | Out << "._" ; |
1329 | SmallVector<StringRef, 8> Features; |
1330 | AttrStr.split(A&: Features, Separator: "+" ); |
1331 | for (auto &Feat : Features) |
1332 | Feat = Feat.trim(); |
1333 | |
1334 | llvm::sort(C&: Features, Comp: [](const StringRef LHS, const StringRef RHS) { |
1335 | return LHS.compare(RHS) < 0; |
1336 | }); |
1337 | |
1338 | llvm::SmallDenseSet<StringRef, 8> UniqueFeats; |
1339 | for (auto &Feat : Features) |
1340 | if (auto Ext = llvm::AArch64::parseFMVExtension(Extension: Feat)) |
1341 | if (UniqueFeats.insert(V: Ext->Name).second) |
1342 | Out << 'M' << Ext->Name; |
1343 | } |
1344 | |
1345 | std::unique_ptr<TargetCodeGenInfo> |
1346 | CodeGen::createAArch64TargetCodeGenInfo(CodeGenModule &CGM, |
1347 | AArch64ABIKind Kind) { |
1348 | return std::make_unique<AArch64TargetCodeGenInfo>(args&: CGM.getTypes(), args&: Kind); |
1349 | } |
1350 | |
1351 | std::unique_ptr<TargetCodeGenInfo> |
1352 | CodeGen::createWindowsAArch64TargetCodeGenInfo(CodeGenModule &CGM, |
1353 | AArch64ABIKind K) { |
1354 | return std::make_unique<WindowsAArch64TargetCodeGenInfo>(args&: CGM.getTypes(), args&: K); |
1355 | } |
1356 | |