| //===- Intrinsics.td - Defines all LLVM intrinsics ---------*- tablegen -*-===// |
| // |
| // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
| // See https://llvm.org/LICENSE.txt for license information. |
| // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
| // |
| //===----------------------------------------------------------------------===// |
| // |
| // This file defines properties of all LLVM intrinsics. |
| // |
| //===----------------------------------------------------------------------===// |
| |
| include "llvm/CodeGen/ValueTypes.td" |
| include "llvm/CodeGen/SDNodeProperties.td" |
| |
| //===----------------------------------------------------------------------===// |
| // Properties we keep track of for intrinsics. |
| //===----------------------------------------------------------------------===// |
| |
| class IntrinsicProperty<bit is_default = false> { |
| bit IsDefault = is_default; |
| } |
| |
| // Intr*Mem - Memory properties. If no property is set, the worst case |
| // is assumed (it may read and write any memory it can get access to and it may |
| // have other side effects). |
| |
| // IntrNoMem - The intrinsic does not access memory or have any other side |
| // effects. It may be CSE'd deleted if dead, etc. |
| def IntrNoMem : IntrinsicProperty; |
| |
| // IntrReadMem - This intrinsic only reads from memory. It does not write to |
| // memory and has no other side effects. Therefore, it cannot be moved across |
| // potentially aliasing stores. However, it can be reordered otherwise and can |
| // be deleted if dead. |
| def IntrReadMem : IntrinsicProperty; |
| |
| // IntrWriteMem - This intrinsic only writes to memory, but does not read from |
| // memory, and has no other side effects. This means dead stores before calls |
| // to this intrinsics may be removed. |
| def IntrWriteMem : IntrinsicProperty; |
| |
| // IntrArgMemOnly - This intrinsic only accesses memory that its pointer-typed |
| // argument(s) points to, but may access an unspecified amount. Other than |
| // reads from and (possibly volatile) writes to memory, it has no side effects. |
| def IntrArgMemOnly : IntrinsicProperty; |
| |
| // IntrInaccessibleMemOnly -- This intrinsic only accesses memory that is not |
| // accessible by the module being compiled. This is a weaker form of IntrNoMem. |
| def IntrInaccessibleMemOnly : IntrinsicProperty; |
| |
| // IntrInaccessibleMemOrArgMemOnly -- This intrinsic only accesses memory that |
| // its pointer-typed arguments point to or memory that is not accessible |
| // by the module being compiled. This is a weaker form of IntrArgMemOnly. |
| def IntrInaccessibleMemOrArgMemOnly : IntrinsicProperty; |
| |
| // Commutative - This intrinsic is commutative: X op Y == Y op X. |
| def Commutative : IntrinsicProperty; |
| |
| // Throws - This intrinsic can throw. |
| def Throws : IntrinsicProperty; |
| |
| // Attribute index needs to match `AttrIndex` defined `Attributes.h`. |
| class AttrIndex<int idx> { |
| int Value = idx; |
| } |
| def FuncIndex : AttrIndex<-1>; |
| def RetIndex : AttrIndex<0>; |
| class ArgIndex<int argNo> : AttrIndex<!add(argNo, 1)>; |
| |
| // NoCapture - The specified argument pointer is not captured by the intrinsic. |
| class NoCapture<AttrIndex idx> : IntrinsicProperty { |
| int ArgNo = idx.Value; |
| } |
| |
| // NoAlias - The specified argument pointer is not aliasing other "noalias" pointer |
| // arguments of the intrinsic wrt. the intrinsic scope. |
| class NoAlias<AttrIndex idx> : IntrinsicProperty { |
| int ArgNo = idx.Value; |
| } |
| |
| // NoUndef - The specified argument is neither undef nor poison. |
| class NoUndef<AttrIndex idx> : IntrinsicProperty { |
| int ArgNo = idx.Value; |
| } |
| |
| class Align<AttrIndex idx, int align> : IntrinsicProperty { |
| int ArgNo = idx.Value; |
| int Align = align; |
| } |
| |
| // Returned - The specified argument is always the return value of the |
| // intrinsic. |
| class Returned<AttrIndex idx> : IntrinsicProperty { |
| int ArgNo = idx.Value; |
| } |
| |
| // ImmArg - The specified argument must be an immediate. |
| class ImmArg<AttrIndex idx> : IntrinsicProperty { |
| int ArgNo = idx.Value; |
| } |
| |
| // ReadOnly - The specified argument pointer is not written to through the |
| // pointer by the intrinsic. |
| class ReadOnly<AttrIndex idx> : IntrinsicProperty { |
| int ArgNo = idx.Value; |
| } |
| |
| // WriteOnly - The intrinsic does not read memory through the specified |
| // argument pointer. |
| class WriteOnly<AttrIndex idx> : IntrinsicProperty { |
| int ArgNo = idx.Value; |
| } |
| |
| // ReadNone - The specified argument pointer is not dereferenced by the |
| // intrinsic. |
| class ReadNone<AttrIndex idx> : IntrinsicProperty { |
| int ArgNo = idx.Value; |
| } |
| |
| def IntrNoReturn : IntrinsicProperty; |
| |
| // IntrNoSync - Threads executing the intrinsic will not synchronize using |
| // memory or other means. Applied by default. |
| def IntrNoSync : IntrinsicProperty<1>; |
| |
| // Applied by default. |
| def IntrNoFree : IntrinsicProperty<1>; |
| |
| // Applied by default. |
| def IntrWillReturn : IntrinsicProperty<1>; |
| |
| // IntrCold - Calls to this intrinsic are cold. |
| // Parallels the cold attribute on LLVM IR functions. |
| def IntrCold : IntrinsicProperty; |
| |
| // IntrNoDuplicate - Calls to this intrinsic cannot be duplicated. |
| // Parallels the noduplicate attribute on LLVM IR functions. |
| def IntrNoDuplicate : IntrinsicProperty; |
| |
| // IntrNoMerge - Calls to this intrinsic cannot be merged |
| // Parallels the nomerge attribute on LLVM IR functions. |
| def IntrNoMerge : IntrinsicProperty; |
| |
| // IntrConvergent - Calls to this intrinsic are convergent and may not be made |
| // control-dependent on any additional values. |
| // Parallels the convergent attribute on LLVM IR functions. |
| def IntrConvergent : IntrinsicProperty; |
| |
| // This property indicates that the intrinsic is safe to speculate. |
| def IntrSpeculatable : IntrinsicProperty; |
| |
| // This property can be used to override the 'has no other side effects' |
| // language of the IntrNoMem, IntrReadMem, IntrWriteMem, and IntrArgMemOnly |
| // intrinsic properties. By default, intrinsics are assumed to have side |
| // effects, so this property is only necessary if you have defined one of |
| // the memory properties listed above. |
| // For this property, 'side effects' has the same meaning as 'side effects' |
| // defined by the hasSideEffects property of the TableGen Instruction class. |
| def IntrHasSideEffects : IntrinsicProperty; |
| |
| //===----------------------------------------------------------------------===// |
| // Types used by intrinsics. |
| //===----------------------------------------------------------------------===// |
| |
| class LLVMType<ValueType vt> { |
| ValueType VT = vt; |
| int isAny = false; |
| } |
| |
| class LLVMQualPointerType<LLVMType elty, int addrspace> |
| : LLVMType<iPTR>{ |
| LLVMType ElTy = elty; |
| int AddrSpace = addrspace; |
| } |
| |
| class LLVMPointerType<LLVMType elty> |
| : LLVMQualPointerType<elty, 0>; |
| |
| class LLVMAnyPointerType<LLVMType elty> |
| : LLVMType<iPTRAny>{ |
| LLVMType ElTy = elty; |
| |
| let isAny = true; |
| } |
| |
| // Match the type of another intrinsic parameter. Number is an index into the |
| // list of overloaded types for the intrinsic, excluding all the fixed types. |
| // The Number value must refer to a previously listed type. For example: |
| // Intrinsic<[llvm_i32_ty], [llvm_i32_ty, llvm_anyfloat_ty, LLVMMatchType<0>]> |
| // has two overloaded types, the 2nd and 3rd arguments. LLVMMatchType<0> |
| // refers to the first overloaded type, which is the 2nd argument. |
| class LLVMMatchType<int num> |
| : LLVMType<OtherVT>{ |
| int Number = num; |
| } |
| |
| // Match the type of another intrinsic parameter that is expected to be based on |
| // an integral type (i.e. either iN or <N x iM>), but change the scalar size to |
| // be twice as wide or half as wide as the other type. This is only useful when |
| // the intrinsic is overloaded, so the matched type should be declared as iAny. |
| class LLVMExtendedType<int num> : LLVMMatchType<num>; |
| class LLVMTruncatedType<int num> : LLVMMatchType<num>; |
| |
| // Match the scalar/vector of another intrinsic parameter but with a different |
| // element type. Either both are scalars or both are vectors with the same |
| // number of elements. |
| class LLVMScalarOrSameVectorWidth<int idx, LLVMType elty> |
| : LLVMMatchType<idx> { |
| ValueType ElTy = elty.VT; |
| } |
| |
| class LLVMPointerTo<int num> : LLVMMatchType<num>; |
| class LLVMPointerToElt<int num> : LLVMMatchType<num>; |
| class LLVMVectorOfAnyPointersToElt<int num> : LLVMMatchType<num>; |
| class LLVMVectorElementType<int num> : LLVMMatchType<num>; |
| |
| // Match the type of another intrinsic parameter that is expected to be a |
| // vector type, but change the element count to be half as many. |
| class LLVMHalfElementsVectorType<int num> : LLVMMatchType<num>; |
| |
| // Match the type of another intrinsic parameter that is expected to be a |
| // vector type (i.e. <N x iM>) but with each element subdivided to |
| // form a vector with more elements that are smaller than the original. |
| class LLVMSubdivide2VectorType<int num> : LLVMMatchType<num>; |
| class LLVMSubdivide4VectorType<int num> : LLVMMatchType<num>; |
| |
| // Match the element count and bit width of another intrinsic parameter, but |
| // change the element type to an integer. |
| class LLVMVectorOfBitcastsToInt<int num> : LLVMMatchType<num>; |
| |
| def llvm_void_ty : LLVMType<isVoid>; |
| let isAny = true in { |
| def llvm_any_ty : LLVMType<Any>; |
| def llvm_anyint_ty : LLVMType<iAny>; |
| def llvm_anyfloat_ty : LLVMType<fAny>; |
| def llvm_anyvector_ty : LLVMType<vAny>; |
| } |
| def llvm_i1_ty : LLVMType<i1>; |
| def llvm_i8_ty : LLVMType<i8>; |
| def llvm_i16_ty : LLVMType<i16>; |
| def llvm_i32_ty : LLVMType<i32>; |
| def llvm_i64_ty : LLVMType<i64>; |
| def llvm_half_ty : LLVMType<f16>; |
| def llvm_bfloat_ty : LLVMType<bf16>; |
| def llvm_float_ty : LLVMType<f32>; |
| def llvm_double_ty : LLVMType<f64>; |
| def llvm_f80_ty : LLVMType<f80>; |
| def llvm_f128_ty : LLVMType<f128>; |
| def llvm_ppcf128_ty : LLVMType<ppcf128>; |
| def llvm_ptr_ty : LLVMPointerType<llvm_i8_ty>; // i8* |
| def llvm_ptrptr_ty : LLVMPointerType<llvm_ptr_ty>; // i8** |
| def llvm_anyptr_ty : LLVMAnyPointerType<llvm_i8_ty>; // (space)i8* |
| def llvm_empty_ty : LLVMType<OtherVT>; // { } |
| def llvm_descriptor_ty : LLVMPointerType<llvm_empty_ty>; // { }* |
| def llvm_metadata_ty : LLVMType<MetadataVT>; // !{...} |
| def llvm_token_ty : LLVMType<token>; // token |
| |
| def llvm_x86mmx_ty : LLVMType<x86mmx>; |
| def llvm_ptrx86mmx_ty : LLVMPointerType<llvm_x86mmx_ty>; // <1 x i64>* |
| |
| def llvm_x86amx_ty : LLVMType<x86amx>; |
| |
| def llvm_v2i1_ty : LLVMType<v2i1>; // 2 x i1 |
| def llvm_v4i1_ty : LLVMType<v4i1>; // 4 x i1 |
| def llvm_v8i1_ty : LLVMType<v8i1>; // 8 x i1 |
| def llvm_v16i1_ty : LLVMType<v16i1>; // 16 x i1 |
| def llvm_v32i1_ty : LLVMType<v32i1>; // 32 x i1 |
| def llvm_v64i1_ty : LLVMType<v64i1>; // 64 x i1 |
| def llvm_v128i1_ty : LLVMType<v128i1>; // 128 x i1 |
| def llvm_v256i1_ty : LLVMType<v256i1>; // 256 x i1 |
| def llvm_v512i1_ty : LLVMType<v512i1>; // 512 x i1 |
| def llvm_v1024i1_ty : LLVMType<v1024i1>; //1024 x i1 |
| |
| def llvm_v1i8_ty : LLVMType<v1i8>; // 1 x i8 |
| def llvm_v2i8_ty : LLVMType<v2i8>; // 2 x i8 |
| def llvm_v4i8_ty : LLVMType<v4i8>; // 4 x i8 |
| def llvm_v8i8_ty : LLVMType<v8i8>; // 8 x i8 |
| def llvm_v16i8_ty : LLVMType<v16i8>; // 16 x i8 |
| def llvm_v32i8_ty : LLVMType<v32i8>; // 32 x i8 |
| def llvm_v64i8_ty : LLVMType<v64i8>; // 64 x i8 |
| def llvm_v128i8_ty : LLVMType<v128i8>; //128 x i8 |
| def llvm_v256i8_ty : LLVMType<v256i8>; //256 x i8 |
| |
| def llvm_v1i16_ty : LLVMType<v1i16>; // 1 x i16 |
| def llvm_v2i16_ty : LLVMType<v2i16>; // 2 x i16 |
| def llvm_v4i16_ty : LLVMType<v4i16>; // 4 x i16 |
| def llvm_v8i16_ty : LLVMType<v8i16>; // 8 x i16 |
| def llvm_v16i16_ty : LLVMType<v16i16>; // 16 x i16 |
| def llvm_v32i16_ty : LLVMType<v32i16>; // 32 x i16 |
| def llvm_v64i16_ty : LLVMType<v64i16>; // 64 x i16 |
| def llvm_v128i16_ty : LLVMType<v128i16>; //128 x i16 |
| |
| def llvm_v1i32_ty : LLVMType<v1i32>; // 1 x i32 |
| def llvm_v2i32_ty : LLVMType<v2i32>; // 2 x i32 |
| def llvm_v4i32_ty : LLVMType<v4i32>; // 4 x i32 |
| def llvm_v8i32_ty : LLVMType<v8i32>; // 8 x i32 |
| def llvm_v16i32_ty : LLVMType<v16i32>; // 16 x i32 |
| def llvm_v32i32_ty : LLVMType<v32i32>; // 32 x i32 |
| def llvm_v64i32_ty : LLVMType<v64i32>; // 64 x i32 |
| def llvm_v256i32_ty : LLVMType<v256i32>; //256 x i32 |
| |
| def llvm_v1i64_ty : LLVMType<v1i64>; // 1 x i64 |
| def llvm_v2i64_ty : LLVMType<v2i64>; // 2 x i64 |
| def llvm_v4i64_ty : LLVMType<v4i64>; // 4 x i64 |
| def llvm_v8i64_ty : LLVMType<v8i64>; // 8 x i64 |
| def llvm_v16i64_ty : LLVMType<v16i64>; // 16 x i64 |
| def llvm_v32i64_ty : LLVMType<v32i64>; // 32 x i64 |
| |
| def llvm_v1i128_ty : LLVMType<v1i128>; // 1 x i128 |
| |
| def llvm_v2f16_ty : LLVMType<v2f16>; // 2 x half (__fp16) |
| def llvm_v4f16_ty : LLVMType<v4f16>; // 4 x half (__fp16) |
| def llvm_v8f16_ty : LLVMType<v8f16>; // 8 x half (__fp16) |
| def llvm_v16f16_ty : LLVMType<v16f16>; // 16 x half (__fp16) |
| def llvm_v32f16_ty : LLVMType<v32f16>; // 32 x half (__fp16) |
| def llvm_v2bf16_ty : LLVMType<v2bf16>; // 2 x bfloat (__bf16) |
| def llvm_v4bf16_ty : LLVMType<v4bf16>; // 4 x bfloat (__bf16) |
| def llvm_v8bf16_ty : LLVMType<v8bf16>; // 8 x bfloat (__bf16) |
| def llvm_v1f32_ty : LLVMType<v1f32>; // 1 x float |
| def llvm_v2f32_ty : LLVMType<v2f32>; // 2 x float |
| def llvm_v3f32_ty : LLVMType<v3f32>; // 3 x float |
| def llvm_v4f32_ty : LLVMType<v4f32>; // 4 x float |
| def llvm_v8f32_ty : LLVMType<v8f32>; // 8 x float |
| def llvm_v16f32_ty : LLVMType<v16f32>; // 16 x float |
| def llvm_v32f32_ty : LLVMType<v32f32>; // 32 x float |
| def llvm_v1f64_ty : LLVMType<v1f64>; // 1 x double |
| def llvm_v2f64_ty : LLVMType<v2f64>; // 2 x double |
| def llvm_v4f64_ty : LLVMType<v4f64>; // 4 x double |
| def llvm_v8f64_ty : LLVMType<v8f64>; // 8 x double |
| def llvm_v16f64_ty : LLVMType<v16f64>; // 16 x double |
| |
| def llvm_vararg_ty : LLVMType<isVoid>; // this means vararg here |
| |
| def llvm_externref_ty : LLVMType<externref>; |
| def llvm_funcref_ty : LLVMType<funcref>; |
| |
| //===----------------------------------------------------------------------===// |
| // Intrinsic Definitions. |
| //===----------------------------------------------------------------------===// |
| |
| // Intrinsic class - This is used to define one LLVM intrinsic. The name of the |
| // intrinsic definition should start with "int_", then match the LLVM intrinsic |
| // name with the "llvm." prefix removed, and all "."s turned into "_"s. For |
| // example, llvm.bswap.i16 -> int_bswap_i16. |
| // |
| // * RetTypes is a list containing the return types expected for the |
| // intrinsic. |
| // * ParamTypes is a list containing the parameter types expected for the |
| // intrinsic. |
| // * Properties can be set to describe the behavior of the intrinsic. |
| // |
| class Intrinsic<list<LLVMType> ret_types, |
| list<LLVMType> param_types = [], |
| list<IntrinsicProperty> intr_properties = [], |
| string name = "", |
| list<SDNodeProperty> sd_properties = [], |
| bit disable_default_attributes = true> : SDPatternOperator { |
| string LLVMName = name; |
| string TargetPrefix = ""; // Set to a prefix for target-specific intrinsics. |
| list<LLVMType> RetTypes = ret_types; |
| list<LLVMType> ParamTypes = param_types; |
| list<IntrinsicProperty> IntrProperties = intr_properties; |
| let Properties = sd_properties; |
| |
| // Disable applying IntrinsicProperties that are marked default with |
| // IntrinsicProperty<1> |
| bit DisableDefaultAttributes = disable_default_attributes; |
| |
| bit isTarget = false; |
| } |
| |
| // Intrinsic with default attributes (disable_default_attributes = false). |
| class DefaultAttrsIntrinsic<list<LLVMType> ret_types, |
| list<LLVMType> param_types = [], |
| list<IntrinsicProperty> intr_properties = [], |
| string name = "", |
| list<SDNodeProperty> sd_properties = []> |
| : Intrinsic<ret_types, param_types, |
| intr_properties, name, |
| sd_properties, /*disable_default_attributes*/ 0> {} |
| |
| /// GCCBuiltin - If this intrinsic exactly corresponds to a GCC builtin, this |
| /// specifies the name of the builtin. This provides automatic CBE and CFE |
| /// support. |
| class GCCBuiltin<string name> { |
| string GCCBuiltinName = name; |
| } |
| |
| class MSBuiltin<string name> { |
| string MSBuiltinName = name; |
| } |
| |
| |
| //===--------------- Variable Argument Handling Intrinsics ----------------===// |
| // |
| |
| def int_vastart : DefaultAttrsIntrinsic<[], [llvm_ptr_ty], [], "llvm.va_start">; |
| def int_vacopy : DefaultAttrsIntrinsic<[], [llvm_ptr_ty, llvm_ptr_ty], [], |
| "llvm.va_copy">; |
| def int_vaend : DefaultAttrsIntrinsic<[], [llvm_ptr_ty], [], "llvm.va_end">; |
| |
| //===------------------- Garbage Collection Intrinsics --------------------===// |
| // |
| def int_gcroot : Intrinsic<[], |
| [llvm_ptrptr_ty, llvm_ptr_ty]>; |
| def int_gcread : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty, llvm_ptrptr_ty], |
| [IntrReadMem, IntrArgMemOnly]>; |
| def int_gcwrite : Intrinsic<[], |
| [llvm_ptr_ty, llvm_ptr_ty, llvm_ptrptr_ty], |
| [IntrArgMemOnly, NoCapture<ArgIndex<1>>, |
| NoCapture<ArgIndex<2>>]>; |
| |
| //===------------------- ObjC ARC runtime Intrinsics --------------------===// |
| // |
| // Note these are to support the Objective-C ARC optimizer which wants to |
| // eliminate retain and releases where possible. |
| |
| def int_objc_autorelease : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty]>; |
| def int_objc_autoreleasePoolPop : Intrinsic<[], [llvm_ptr_ty]>; |
| def int_objc_autoreleasePoolPush : Intrinsic<[llvm_ptr_ty], []>; |
| def int_objc_autoreleaseReturnValue : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty]>; |
| def int_objc_copyWeak : Intrinsic<[], |
| [llvm_ptrptr_ty, |
| llvm_ptrptr_ty]>; |
| def int_objc_destroyWeak : Intrinsic<[], [llvm_ptrptr_ty]>; |
| def int_objc_initWeak : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptrptr_ty, |
| llvm_ptr_ty]>; |
| def int_objc_loadWeak : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptrptr_ty]>; |
| def int_objc_loadWeakRetained : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptrptr_ty]>; |
| def int_objc_moveWeak : Intrinsic<[], |
| [llvm_ptrptr_ty, |
| llvm_ptrptr_ty]>; |
| def int_objc_release : Intrinsic<[], [llvm_ptr_ty]>; |
| def int_objc_retain : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty]>; |
| def int_objc_retainAutorelease : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty]>; |
| def int_objc_retainAutoreleaseReturnValue : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty]>; |
| def int_objc_retainAutoreleasedReturnValue : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty]>; |
| def int_objc_retainBlock : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty]>; |
| def int_objc_storeStrong : Intrinsic<[], |
| [llvm_ptrptr_ty, |
| llvm_ptr_ty]>; |
| def int_objc_storeWeak : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptrptr_ty, |
| llvm_ptr_ty]>; |
| def int_objc_clang_arc_use : Intrinsic<[], |
| [llvm_vararg_ty]>; |
| def int_objc_clang_arc_noop_use : DefaultAttrsIntrinsic<[], |
| [llvm_vararg_ty], |
| [IntrInaccessibleMemOnly]>; |
| def int_objc_unsafeClaimAutoreleasedReturnValue : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty]>; |
| def int_objc_retainedObject : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty]>; |
| def int_objc_unretainedObject : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty]>; |
| def int_objc_unretainedPointer : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty]>; |
| def int_objc_retain_autorelease : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty]>; |
| def int_objc_sync_enter : Intrinsic<[llvm_i32_ty], |
| [llvm_ptr_ty]>; |
| def int_objc_sync_exit : Intrinsic<[llvm_i32_ty], |
| [llvm_ptr_ty]>; |
| def int_objc_arc_annotation_topdown_bbstart : Intrinsic<[], |
| [llvm_ptrptr_ty, |
| llvm_ptrptr_ty]>; |
| def int_objc_arc_annotation_topdown_bbend : Intrinsic<[], |
| [llvm_ptrptr_ty, |
| llvm_ptrptr_ty]>; |
| def int_objc_arc_annotation_bottomup_bbstart : Intrinsic<[], |
| [llvm_ptrptr_ty, |
| llvm_ptrptr_ty]>; |
| def int_objc_arc_annotation_bottomup_bbend : Intrinsic<[], |
| [llvm_ptrptr_ty, |
| llvm_ptrptr_ty]>; |
| //===--------------- Swift asynchronous context intrinsics ----------------===// |
| |
| // Returns the location of the Swift asynchronous context (usually stored just |
| // before the frame pointer), and triggers the creation of a null context if it |
| // would otherwise be unneeded. |
| def int_swift_async_context_addr : Intrinsic<[llvm_ptrptr_ty], [], [IntrNoMem]>; |
| |
| //===--------------------- Code Generator Intrinsics ----------------------===// |
| // |
| def int_returnaddress : DefaultAttrsIntrinsic<[llvm_ptr_ty], [llvm_i32_ty], |
| [IntrNoMem, ImmArg<ArgIndex<0>>]>; |
| def int_addressofreturnaddress : DefaultAttrsIntrinsic<[llvm_anyptr_ty], [], [IntrNoMem]>; |
| def int_frameaddress : DefaultAttrsIntrinsic<[llvm_anyptr_ty], [llvm_i32_ty], |
| [IntrNoMem, ImmArg<ArgIndex<0>>]>; |
| def int_sponentry : DefaultAttrsIntrinsic<[llvm_anyptr_ty], [], [IntrNoMem]>; |
| def int_read_register : Intrinsic<[llvm_anyint_ty], [llvm_metadata_ty], |
| [IntrReadMem], "llvm.read_register">; |
| def int_write_register : Intrinsic<[], [llvm_metadata_ty, llvm_anyint_ty], |
| [], "llvm.write_register">; |
| def int_read_volatile_register : Intrinsic<[llvm_anyint_ty], [llvm_metadata_ty], |
| [IntrHasSideEffects], |
| "llvm.read_volatile_register">; |
| |
| // Gets the address of the local variable area. This is typically a copy of the |
| // stack, frame, or base pointer depending on the type of prologue. |
| def int_localaddress : DefaultAttrsIntrinsic<[llvm_ptr_ty], [], [IntrNoMem]>; |
| |
| // Escapes local variables to allow access from other functions. |
| def int_localescape : DefaultAttrsIntrinsic<[], [llvm_vararg_ty]>; |
| |
| // Given a function and the localaddress of a parent frame, returns a pointer |
| // to an escaped allocation indicated by the index. |
| def int_localrecover : DefaultAttrsIntrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty, llvm_ptr_ty, llvm_i32_ty], |
| [IntrNoMem, ImmArg<ArgIndex<2>>]>; |
| |
| // Given the frame pointer passed into an SEH filter function, returns a |
| // pointer to the local variable area suitable for use with llvm.localrecover. |
| def int_eh_recoverfp : DefaultAttrsIntrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty, llvm_ptr_ty], |
| [IntrNoMem]>; |
| |
| // To mark the beginning/end of a try-scope for Windows SEH -EHa |
| // calls/invokes to these intrinsics are placed to model control flows |
| // caused by HW exceptions under option -EHa. |
| // calls/invokes to these intrinsics will be discarded during a codegen pass |
| // after EH tables are generated |
| def int_seh_try_begin : Intrinsic<[], [], [IntrWriteMem, IntrWillReturn]>; |
| def int_seh_try_end : Intrinsic<[], [], [IntrWriteMem, IntrWillReturn]>; |
| def int_seh_scope_begin : Intrinsic<[], [], [IntrNoMem]>; |
| def int_seh_scope_end : Intrinsic<[], [], [IntrNoMem]>; |
| |
| // Note: we treat stacksave/stackrestore as writemem because we don't otherwise |
| // model their dependencies on allocas. |
| def int_stacksave : DefaultAttrsIntrinsic<[llvm_ptr_ty]>, |
| GCCBuiltin<"__builtin_stack_save">; |
| def int_stackrestore : DefaultAttrsIntrinsic<[], [llvm_ptr_ty]>, |
| GCCBuiltin<"__builtin_stack_restore">; |
| |
| def int_get_dynamic_area_offset : DefaultAttrsIntrinsic<[llvm_anyint_ty]>; |
| |
| def int_thread_pointer : DefaultAttrsIntrinsic<[llvm_ptr_ty], [], [IntrNoMem]>, |
| GCCBuiltin<"__builtin_thread_pointer">; |
| |
| // IntrInaccessibleMemOrArgMemOnly is a little more pessimistic than strictly |
| // necessary for prefetch, however it does conveniently prevent the prefetch |
| // from being reordered overly much with respect to nearby access to the same |
| // memory while not impeding optimization. |
| def int_prefetch |
| : DefaultAttrsIntrinsic<[], [ llvm_anyptr_ty, llvm_i32_ty, llvm_i32_ty, llvm_i32_ty ], |
| [IntrInaccessibleMemOrArgMemOnly, IntrWillReturn, |
| ReadOnly<ArgIndex<0>>, NoCapture<ArgIndex<0>>, |
| ImmArg<ArgIndex<1>>, ImmArg<ArgIndex<2>>]>; |
| def int_pcmarker : DefaultAttrsIntrinsic<[], [llvm_i32_ty]>; |
| |
| def int_readcyclecounter : DefaultAttrsIntrinsic<[llvm_i64_ty]>; |
| |
| // The assume intrinsic is marked InaccessibleMemOnly so that proper control |
| // dependencies will be maintained. |
| def int_assume : DefaultAttrsIntrinsic< |
| [], [llvm_i1_ty], [IntrInaccessibleMemOnly, NoUndef<ArgIndex<0>>]>; |
| |
| // 'llvm.experimental.noalias.scope.decl' intrinsic: Inserted at the location of |
| // noalias scope declaration. Makes it possible to identify that a noalias scope |
| // is only valid inside the body of a loop. |
| // |
| // Purpose of the different arguments: |
| // - arg0: id.scope: metadata representing the scope declaration. |
| def int_experimental_noalias_scope_decl |
| : DefaultAttrsIntrinsic<[], [llvm_metadata_ty], |
| [IntrInaccessibleMemOnly]>; // blocks LICM and some more |
| |
| // Stack Protector Intrinsic - The stackprotector intrinsic writes the stack |
| // guard to the correct place on the stack frame. |
| def int_stackprotector : DefaultAttrsIntrinsic<[], [llvm_ptr_ty, llvm_ptrptr_ty], []>; |
| def int_stackguard : DefaultAttrsIntrinsic<[llvm_ptr_ty], [], []>; |
| |
| // A cover for instrumentation based profiling. |
| def int_instrprof_cover : Intrinsic<[], [llvm_ptr_ty, llvm_i64_ty, |
| llvm_i32_ty, llvm_i32_ty]>; |
| |
| // A counter increment for instrumentation based profiling. |
| def int_instrprof_increment : Intrinsic<[], |
| [llvm_ptr_ty, llvm_i64_ty, |
| llvm_i32_ty, llvm_i32_ty]>; |
| |
| // A counter increment with step for instrumentation based profiling. |
| def int_instrprof_increment_step : Intrinsic<[], |
| [llvm_ptr_ty, llvm_i64_ty, |
| llvm_i32_ty, llvm_i32_ty, llvm_i64_ty]>; |
| |
| // A call to profile runtime for value profiling of target expressions |
| // through instrumentation based profiling. |
| def int_instrprof_value_profile : Intrinsic<[], |
| [llvm_ptr_ty, llvm_i64_ty, |
| llvm_i64_ty, llvm_i32_ty, |
| llvm_i32_ty]>; |
| |
| def int_call_preallocated_setup : DefaultAttrsIntrinsic<[llvm_token_ty], [llvm_i32_ty]>; |
| def int_call_preallocated_arg : DefaultAttrsIntrinsic<[llvm_ptr_ty], [llvm_token_ty, llvm_i32_ty]>; |
| def int_call_preallocated_teardown : DefaultAttrsIntrinsic<[], [llvm_token_ty]>; |
| |
| //===------------------- Standard C Library Intrinsics --------------------===// |
| // |
| |
| def int_memcpy : Intrinsic<[], |
| [llvm_anyptr_ty, llvm_anyptr_ty, llvm_anyint_ty, |
| llvm_i1_ty], |
| [IntrArgMemOnly, IntrWillReturn, IntrNoFree, |
| NoCapture<ArgIndex<0>>, NoCapture<ArgIndex<1>>, |
| NoAlias<ArgIndex<0>>, NoAlias<ArgIndex<1>>, |
| WriteOnly<ArgIndex<0>>, ReadOnly<ArgIndex<1>>, |
| ImmArg<ArgIndex<3>>]>; |
| |
| // Memcpy semantic that is guaranteed to be inlined. |
| // In particular this means that the generated code is not allowed to call any |
| // external function. |
| // The third argument (specifying the size) must be a constant. |
| def int_memcpy_inline |
| : Intrinsic<[], |
| [llvm_anyptr_ty, llvm_anyptr_ty, llvm_anyint_ty, llvm_i1_ty], |
| [IntrArgMemOnly, IntrWillReturn, IntrNoFree, |
| NoCapture<ArgIndex<0>>, NoCapture<ArgIndex<1>>, |
| NoAlias<ArgIndex<0>>, NoAlias<ArgIndex<1>>, |
| WriteOnly<ArgIndex<0>>, ReadOnly<ArgIndex<1>>, |
| ImmArg<ArgIndex<2>>, ImmArg<ArgIndex<3>>]>; |
| |
| def int_memmove : Intrinsic<[], |
| [llvm_anyptr_ty, llvm_anyptr_ty, llvm_anyint_ty, |
| llvm_i1_ty], |
| [IntrArgMemOnly, IntrWillReturn, IntrNoFree, |
| NoCapture<ArgIndex<0>>, NoCapture<ArgIndex<1>>, |
| WriteOnly<ArgIndex<0>>, ReadOnly<ArgIndex<1>>, |
| ImmArg<ArgIndex<3>>]>; |
| def int_memset : Intrinsic<[], |
| [llvm_anyptr_ty, llvm_i8_ty, llvm_anyint_ty, |
| llvm_i1_ty], |
| [IntrWriteMem, IntrArgMemOnly, IntrWillReturn, |
| IntrNoFree, |
| NoCapture<ArgIndex<0>>, WriteOnly<ArgIndex<0>>, |
| ImmArg<ArgIndex<3>>]>; |
| |
| // FIXME: Add version of these floating point intrinsics which allow non-default |
| // rounding modes and FP exception handling. |
| |
| let IntrProperties = [IntrNoMem, IntrSpeculatable, IntrWillReturn] in { |
| def int_fma : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>, |
| LLVMMatchType<0>]>; |
| def int_fmuladd : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>, |
| LLVMMatchType<0>]>; |
| |
| // These functions do not read memory, but are sensitive to the |
| // rounding mode. LLVM purposely does not model changes to the FP |
| // environment so they can be treated as readnone. |
| def int_sqrt : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>]>; |
| def int_powi : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>, llvm_anyint_ty]>; |
| def int_sin : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>]>; |
| def int_cos : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>]>; |
| def int_pow : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>]>; |
| def int_log : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>]>; |
| def int_log10: DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>]>; |
| def int_log2 : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>]>; |
| def int_exp : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>]>; |
| def int_exp2 : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>]>; |
| def int_fabs : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>]>; |
| def int_copysign : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>]>; |
| def int_floor : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>]>; |
| def int_ceil : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>]>; |
| def int_trunc : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>]>; |
| def int_rint : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>]>; |
| def int_nearbyint : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>]>; |
| def int_round : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>]>; |
| def int_roundeven : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>]>; |
| def int_canonicalize : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>], |
| [IntrNoMem]>; |
| |
| def int_lround : DefaultAttrsIntrinsic<[llvm_anyint_ty], [llvm_anyfloat_ty]>; |
| def int_llround : DefaultAttrsIntrinsic<[llvm_anyint_ty], [llvm_anyfloat_ty]>; |
| def int_lrint : DefaultAttrsIntrinsic<[llvm_anyint_ty], [llvm_anyfloat_ty]>; |
| def int_llrint : DefaultAttrsIntrinsic<[llvm_anyint_ty], [llvm_anyfloat_ty]>; |
| } |
| |
| def int_minnum : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn, Commutative] |
| >; |
| def int_maxnum : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn, Commutative] |
| >; |
| def int_minimum : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn, Commutative] |
| >; |
| def int_maximum : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn, Commutative] |
| >; |
| |
| // Internal interface for object size checking |
| def int_objectsize : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [llvm_anyptr_ty, llvm_i1_ty, |
| llvm_i1_ty, llvm_i1_ty], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn, |
| ImmArg<ArgIndex<1>>, ImmArg<ArgIndex<2>>, |
| ImmArg<ArgIndex<3>>]>, |
| GCCBuiltin<"__builtin_object_size">; |
| |
| //===--------------- Access to Floating Point Environment -----------------===// |
| // |
| |
| let IntrProperties = [IntrInaccessibleMemOnly, IntrWillReturn] in { |
| def int_flt_rounds : DefaultAttrsIntrinsic<[llvm_i32_ty], []>; |
| def int_set_rounding : DefaultAttrsIntrinsic<[], [llvm_i32_ty]>; |
| } |
| |
| //===--------------- Constrained Floating Point Intrinsics ----------------===// |
| // |
| |
| let IntrProperties = [IntrInaccessibleMemOnly, IntrWillReturn] in { |
| def int_experimental_constrained_fadd : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_fsub : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_fmul : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_fdiv : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_frem : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| |
| def int_experimental_constrained_fma : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| |
| def int_experimental_constrained_fmuladd : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| |
| def int_experimental_constrained_fptosi : DefaultAttrsIntrinsic<[ llvm_anyint_ty ], |
| [ llvm_anyfloat_ty, |
| llvm_metadata_ty ]>; |
| |
| def int_experimental_constrained_fptoui : DefaultAttrsIntrinsic<[ llvm_anyint_ty ], |
| [ llvm_anyfloat_ty, |
| llvm_metadata_ty ]>; |
| |
| def int_experimental_constrained_sitofp : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ llvm_anyint_ty, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| |
| def int_experimental_constrained_uitofp : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ llvm_anyint_ty, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| |
| def int_experimental_constrained_fptrunc : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ llvm_anyfloat_ty, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| |
| def int_experimental_constrained_fpext : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ llvm_anyfloat_ty, |
| llvm_metadata_ty ]>; |
| |
| // These intrinsics are sensitive to the rounding mode so we need constrained |
| // versions of each of them. When strict rounding and exception control are |
| // not required the non-constrained versions of these intrinsics should be |
| // used. |
| def int_experimental_constrained_sqrt : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_powi : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| llvm_i32_ty, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_sin : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_cos : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_pow : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_log : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_log10: DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_log2 : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_exp : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_exp2 : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_rint : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_nearbyint : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_lrint : DefaultAttrsIntrinsic<[ llvm_anyint_ty ], |
| [ llvm_anyfloat_ty, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_llrint : DefaultAttrsIntrinsic<[ llvm_anyint_ty ], |
| [ llvm_anyfloat_ty, |
| llvm_metadata_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_maxnum : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_minnum : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_maximum : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_minimum : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_ceil : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_floor : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_lround : DefaultAttrsIntrinsic<[ llvm_anyint_ty ], |
| [ llvm_anyfloat_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_llround : DefaultAttrsIntrinsic<[ llvm_anyint_ty ], |
| [ llvm_anyfloat_ty, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_round : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_roundeven : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| llvm_metadata_ty ]>; |
| def int_experimental_constrained_trunc : DefaultAttrsIntrinsic<[ llvm_anyfloat_ty ], |
| [ LLVMMatchType<0>, |
| llvm_metadata_ty ]>; |
| |
| // Constrained floating-point comparison (quiet and signaling variants). |
| // Third operand is the predicate represented as a metadata string. |
| def int_experimental_constrained_fcmp |
| : DefaultAttrsIntrinsic<[ LLVMScalarOrSameVectorWidth<0, llvm_i1_ty> ], |
| [ llvm_anyfloat_ty, LLVMMatchType<0>, |
| llvm_metadata_ty, llvm_metadata_ty ]>; |
| def int_experimental_constrained_fcmps |
| : DefaultAttrsIntrinsic<[ LLVMScalarOrSameVectorWidth<0, llvm_i1_ty> ], |
| [ llvm_anyfloat_ty, LLVMMatchType<0>, |
| llvm_metadata_ty, llvm_metadata_ty ]>; |
| } |
| // FIXME: Consider maybe adding intrinsics for sitofp, uitofp. |
| |
| //===------------------------- Expect Intrinsics --------------------------===// |
| // |
| def int_expect : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>], [IntrNoMem, IntrWillReturn]>; |
| |
| def int_expect_with_probability : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>, llvm_double_ty], |
| [IntrNoMem, IntrWillReturn]>; |
| |
| //===-------------------- Bit Manipulation Intrinsics ---------------------===// |
| // |
| |
| // None of these intrinsics accesses memory at all. |
| let IntrProperties = [IntrNoMem, IntrSpeculatable, IntrWillReturn] in { |
| def int_bswap: DefaultAttrsIntrinsic<[llvm_anyint_ty], [LLVMMatchType<0>]>; |
| def int_ctpop: DefaultAttrsIntrinsic<[llvm_anyint_ty], [LLVMMatchType<0>]>; |
| def int_bitreverse : DefaultAttrsIntrinsic<[llvm_anyint_ty], [LLVMMatchType<0>]>; |
| def int_fshl : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>, LLVMMatchType<0>]>; |
| def int_fshr : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>, LLVMMatchType<0>]>; |
| } |
| |
| let IntrProperties = [IntrNoMem, IntrSpeculatable, IntrWillReturn, |
| ImmArg<ArgIndex<1>>] in { |
| def int_ctlz : DefaultAttrsIntrinsic<[llvm_anyint_ty], [LLVMMatchType<0>, llvm_i1_ty]>; |
| def int_cttz : DefaultAttrsIntrinsic<[llvm_anyint_ty], [LLVMMatchType<0>, llvm_i1_ty]>; |
| } |
| |
| //===------------------------ Debugger Intrinsics -------------------------===// |
| // |
| |
| // None of these intrinsics accesses memory at all...but that doesn't |
| // mean the optimizers can change them aggressively. Special handling |
| // needed in a few places. These synthetic intrinsics have no |
| // side-effects and just mark information about their operands. |
| let IntrProperties = [IntrNoMem, IntrSpeculatable, IntrWillReturn] in { |
| def int_dbg_declare : DefaultAttrsIntrinsic<[], |
| [llvm_metadata_ty, |
| llvm_metadata_ty, |
| llvm_metadata_ty]>; |
| def int_dbg_value : DefaultAttrsIntrinsic<[], |
| [llvm_metadata_ty, |
| llvm_metadata_ty, |
| llvm_metadata_ty]>; |
| def int_dbg_addr : DefaultAttrsIntrinsic<[], |
| [llvm_metadata_ty, |
| llvm_metadata_ty, |
| llvm_metadata_ty]>; |
| def int_dbg_label : DefaultAttrsIntrinsic<[], |
| [llvm_metadata_ty]>; |
| } |
| |
| //===------------------ Exception Handling Intrinsics----------------------===// |
| // |
| |
| // The result of eh.typeid.for depends on the enclosing function, but inside a |
| // given function it is 'const' and may be CSE'd etc. |
| def int_eh_typeid_for : Intrinsic<[llvm_i32_ty], [llvm_ptr_ty], [IntrNoMem]>; |
| |
| def int_eh_return_i32 : Intrinsic<[], [llvm_i32_ty, llvm_ptr_ty]>; |
| def int_eh_return_i64 : Intrinsic<[], [llvm_i64_ty, llvm_ptr_ty]>; |
| |
| // eh.exceptionpointer returns the pointer to the exception caught by |
| // the given `catchpad`. |
| def int_eh_exceptionpointer : Intrinsic<[llvm_anyptr_ty], [llvm_token_ty], |
| [IntrNoMem]>; |
| |
| // Gets the exception code from a catchpad token. Only used on some platforms. |
| def int_eh_exceptioncode : Intrinsic<[llvm_i32_ty], [llvm_token_ty], [IntrNoMem]>; |
| |
| // __builtin_unwind_init is an undocumented GCC intrinsic that causes all |
| // callee-saved registers to be saved and restored (regardless of whether they |
| // are used) in the calling function. It is used by libgcc_eh. |
| def int_eh_unwind_init: Intrinsic<[]>, |
| GCCBuiltin<"__builtin_unwind_init">; |
| |
| def int_eh_dwarf_cfa : Intrinsic<[llvm_ptr_ty], [llvm_i32_ty]>; |
| |
| def int_eh_sjlj_lsda : Intrinsic<[llvm_ptr_ty], [], [IntrNoMem]>; |
| def int_eh_sjlj_callsite : Intrinsic<[], [llvm_i32_ty], [IntrNoMem]>; |
| |
| def int_eh_sjlj_functioncontext : Intrinsic<[], [llvm_ptr_ty]>; |
| def int_eh_sjlj_setjmp : Intrinsic<[llvm_i32_ty], [llvm_ptr_ty]>; |
| def int_eh_sjlj_longjmp : Intrinsic<[], [llvm_ptr_ty], [IntrNoReturn]>; |
| def int_eh_sjlj_setup_dispatch : Intrinsic<[], []>; |
| |
| //===---------------- Generic Variable Attribute Intrinsics----------------===// |
| // |
| def int_var_annotation : DefaultAttrsIntrinsic< |
| [], [llvm_ptr_ty, llvm_ptr_ty, llvm_ptr_ty, llvm_i32_ty, llvm_ptr_ty], |
| [IntrInaccessibleMemOnly], "llvm.var.annotation">; |
| |
| def int_ptr_annotation : DefaultAttrsIntrinsic< |
| [LLVMAnyPointerType<llvm_anyint_ty>], |
| [LLVMMatchType<0>, llvm_ptr_ty, llvm_ptr_ty, llvm_i32_ty, llvm_ptr_ty], |
| [IntrInaccessibleMemOnly], "llvm.ptr.annotation">; |
| |
| def int_annotation : DefaultAttrsIntrinsic< |
| [llvm_anyint_ty], |
| [LLVMMatchType<0>, llvm_ptr_ty, llvm_ptr_ty, llvm_i32_ty], |
| [IntrInaccessibleMemOnly], "llvm.annotation">; |
| |
| // Annotates the current program point with metadata strings which are emitted |
| // as CodeView debug info records. This is expensive, as it disables inlining |
| // and is modelled as having side effects. |
| def int_codeview_annotation : DefaultAttrsIntrinsic<[], [llvm_metadata_ty], |
| [IntrInaccessibleMemOnly, IntrNoDuplicate, IntrWillReturn], |
| "llvm.codeview.annotation">; |
| |
| //===------------------------ Trampoline Intrinsics -----------------------===// |
| // |
| def int_init_trampoline : DefaultAttrsIntrinsic< |
| [], [llvm_ptr_ty, llvm_ptr_ty, llvm_ptr_ty], |
| [IntrArgMemOnly, NoCapture<ArgIndex<0>>, WriteOnly<ArgIndex<0>>, |
| ReadNone<ArgIndex<1>>, ReadNone<ArgIndex<2>>]>, |
| GCCBuiltin<"__builtin_init_trampoline">; |
| |
| def int_adjust_trampoline : DefaultAttrsIntrinsic< |
| [llvm_ptr_ty], [llvm_ptr_ty], [IntrReadMem, IntrArgMemOnly]>, |
| GCCBuiltin<"__builtin_adjust_trampoline">; |
| |
| //===------------------------ Overflow Intrinsics -------------------------===// |
| // |
| |
| // Expose the carry flag from add operations on two integrals. |
| let IntrProperties = [IntrNoMem, IntrSpeculatable, IntrWillReturn] in { |
| def int_sadd_with_overflow : DefaultAttrsIntrinsic<[llvm_anyint_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>], |
| [LLVMMatchType<0>, LLVMMatchType<0>]>; |
| def int_uadd_with_overflow : DefaultAttrsIntrinsic<[llvm_anyint_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>], |
| [LLVMMatchType<0>, LLVMMatchType<0>]>; |
| |
| def int_ssub_with_overflow : DefaultAttrsIntrinsic<[llvm_anyint_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>], |
| [LLVMMatchType<0>, LLVMMatchType<0>]>; |
| def int_usub_with_overflow : DefaultAttrsIntrinsic<[llvm_anyint_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>], |
| [LLVMMatchType<0>, LLVMMatchType<0>]>; |
| |
| def int_smul_with_overflow : DefaultAttrsIntrinsic<[llvm_anyint_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>], |
| [LLVMMatchType<0>, LLVMMatchType<0>]>; |
| def int_umul_with_overflow : DefaultAttrsIntrinsic<[llvm_anyint_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>], |
| [LLVMMatchType<0>, LLVMMatchType<0>]>; |
| } |
| //===------------------------- Saturation Arithmetic Intrinsics ---------------------===// |
| // |
| def int_sadd_sat : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn, Commutative]>; |
| def int_uadd_sat : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn, Commutative]>; |
| def int_ssub_sat : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn]>; |
| def int_usub_sat : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn]>; |
| def int_sshl_sat : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn]>; |
| def int_ushl_sat : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn]>; |
| |
| //===------------------------- Fixed Point Arithmetic Intrinsics ---------------------===// |
| // |
| def int_smul_fix : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>, llvm_i32_ty], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn, |
| Commutative, ImmArg<ArgIndex<2>>]>; |
| |
| def int_umul_fix : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>, llvm_i32_ty], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn, |
| Commutative, ImmArg<ArgIndex<2>>]>; |
| |
| def int_sdiv_fix : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>, llvm_i32_ty], |
| [IntrNoMem, ImmArg<ArgIndex<2>>]>; |
| |
| def int_udiv_fix : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>, llvm_i32_ty], |
| [IntrNoMem, ImmArg<ArgIndex<2>>]>; |
| |
| //===------------------- Fixed Point Saturation Arithmetic Intrinsics ----------------===// |
| // |
| def int_smul_fix_sat : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>, llvm_i32_ty], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn, |
| Commutative, ImmArg<ArgIndex<2>>]>; |
| def int_umul_fix_sat : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>, llvm_i32_ty], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn, |
| Commutative, ImmArg<ArgIndex<2>>]>; |
| |
| def int_sdiv_fix_sat : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>, llvm_i32_ty], |
| [IntrNoMem, ImmArg<ArgIndex<2>>]>; |
| |
| def int_udiv_fix_sat : DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>, llvm_i32_ty], |
| [IntrNoMem, ImmArg<ArgIndex<2>>]>; |
| |
| //===------------------ Integer Min/Max/Abs Intrinsics --------------------===// |
| // |
| def int_abs : DefaultAttrsIntrinsic< |
| [llvm_anyint_ty], [LLVMMatchType<0>, llvm_i1_ty], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn, ImmArg<ArgIndex<1>>]>; |
| |
| def int_smax : DefaultAttrsIntrinsic< |
| [llvm_anyint_ty], [LLVMMatchType<0>, LLVMMatchType<0>], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn]>; |
| def int_smin : DefaultAttrsIntrinsic< |
| [llvm_anyint_ty], [LLVMMatchType<0>, LLVMMatchType<0>], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn]>; |
| def int_umax : DefaultAttrsIntrinsic< |
| [llvm_anyint_ty], [LLVMMatchType<0>, LLVMMatchType<0>], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn]>; |
| def int_umin : DefaultAttrsIntrinsic< |
| [llvm_anyint_ty], [LLVMMatchType<0>, LLVMMatchType<0>], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn]>; |
| |
| //===------------------------- Memory Use Markers -------------------------===// |
| // |
| def int_lifetime_start : DefaultAttrsIntrinsic<[], |
| [llvm_i64_ty, llvm_anyptr_ty], |
| [IntrArgMemOnly, IntrWillReturn, |
| NoCapture<ArgIndex<1>>, |
| ImmArg<ArgIndex<0>>]>; |
| def int_lifetime_end : DefaultAttrsIntrinsic<[], |
| [llvm_i64_ty, llvm_anyptr_ty], |
| [IntrArgMemOnly, IntrWillReturn, |
| NoCapture<ArgIndex<1>>, |
| ImmArg<ArgIndex<0>>]>; |
| def int_invariant_start : DefaultAttrsIntrinsic<[llvm_descriptor_ty], |
| [llvm_i64_ty, llvm_anyptr_ty], |
| [IntrArgMemOnly, IntrWillReturn, |
| NoCapture<ArgIndex<1>>, |
| ImmArg<ArgIndex<0>>]>; |
| def int_invariant_end : DefaultAttrsIntrinsic<[], |
| [llvm_descriptor_ty, llvm_i64_ty, |
| llvm_anyptr_ty], |
| [IntrArgMemOnly, IntrWillReturn, |
| NoCapture<ArgIndex<2>>, |
| ImmArg<ArgIndex<1>>]>; |
| |
| // launder.invariant.group can't be marked with 'readnone' (IntrNoMem), |
| // because it would cause CSE of two barriers with the same argument. |
| // Inaccessiblememonly says that the barrier doesn't read the argument, |
| // but it changes state not accessible to this module. This way |
| // we can DSE through the barrier because it doesn't read the value |
| // after store. Although the barrier doesn't modify any memory it |
| // can't be marked as readonly, because it would be possible to |
| // CSE 2 barriers with store in between. |
| // The argument also can't be marked with 'returned' attribute, because |
| // it would remove barrier. |
| // Note that it is still experimental, which means that its semantics |
| // might change in the future. |
| def int_launder_invariant_group : DefaultAttrsIntrinsic<[llvm_anyptr_ty], |
| [LLVMMatchType<0>], |
| [IntrInaccessibleMemOnly, IntrSpeculatable, IntrWillReturn]>; |
| |
| |
| def int_strip_invariant_group : DefaultAttrsIntrinsic<[llvm_anyptr_ty], |
| [LLVMMatchType<0>], |
| [IntrSpeculatable, IntrNoMem, IntrWillReturn]>; |
| |
| //===------------------------ Stackmap Intrinsics -------------------------===// |
| // |
| def int_experimental_stackmap : DefaultAttrsIntrinsic<[], |
| [llvm_i64_ty, llvm_i32_ty, llvm_vararg_ty], |
| [Throws]>; |
| def int_experimental_patchpoint_void : DefaultAttrsIntrinsic<[], |
| [llvm_i64_ty, llvm_i32_ty, |
| llvm_ptr_ty, llvm_i32_ty, |
| llvm_vararg_ty], |
| [Throws]>; |
| def int_experimental_patchpoint_i64 : DefaultAttrsIntrinsic<[llvm_i64_ty], |
| [llvm_i64_ty, llvm_i32_ty, |
| llvm_ptr_ty, llvm_i32_ty, |
| llvm_vararg_ty], |
| [Throws]>; |
| |
| |
| //===------------------------ Garbage Collection Intrinsics ---------------===// |
| // These are documented in docs/Statepoint.rst |
| |
| def int_experimental_gc_statepoint : Intrinsic<[llvm_token_ty], |
| [llvm_i64_ty, llvm_i32_ty, |
| llvm_anyptr_ty, llvm_i32_ty, |
| llvm_i32_ty, llvm_vararg_ty], |
| [Throws, ImmArg<ArgIndex<0>>, |
| ImmArg<ArgIndex<1>>, ImmArg<ArgIndex<3>>, |
| ImmArg<ArgIndex<4>>]>; |
| |
| def int_experimental_gc_result : Intrinsic<[llvm_any_ty], [llvm_token_ty], |
| [IntrNoMem]>; |
| def int_experimental_gc_relocate : Intrinsic<[llvm_any_ty], |
| [llvm_token_ty, llvm_i32_ty, |
| llvm_i32_ty], |
| [IntrNoMem, ImmArg<ArgIndex<1>>, |
| ImmArg<ArgIndex<2>>]>; |
| |
| def int_experimental_gc_get_pointer_base : Intrinsic<[llvm_anyptr_ty], |
| [llvm_anyptr_ty], [IntrNoMem, IntrWillReturn, |
| ReadNone<ArgIndex<0>>, NoCapture<ArgIndex<0>>]>; |
| |
| def int_experimental_gc_get_pointer_offset : Intrinsic<[llvm_i64_ty], |
| [llvm_anyptr_ty], [IntrNoMem, IntrWillReturn, |
| ReadNone<ArgIndex<0>>, NoCapture<ArgIndex<0>>]>; |
| |
| //===------------------------ Coroutine Intrinsics ---------------===// |
| // These are documented in docs/Coroutines.rst |
| |
| // Coroutine Structure Intrinsics. |
| |
| def int_coro_id : Intrinsic<[llvm_token_ty], [llvm_i32_ty, llvm_ptr_ty, |
| llvm_ptr_ty, llvm_ptr_ty], |
| [IntrArgMemOnly, IntrReadMem, |
| ReadNone<ArgIndex<1>>, ReadOnly<ArgIndex<2>>, |
| NoCapture<ArgIndex<2>>]>; |
| def int_coro_id_retcon : Intrinsic<[llvm_token_ty], |
| [llvm_i32_ty, llvm_i32_ty, llvm_ptr_ty, |
| llvm_ptr_ty, llvm_ptr_ty, llvm_ptr_ty], |
| []>; |
| def int_coro_id_retcon_once : Intrinsic<[llvm_token_ty], |
| [llvm_i32_ty, llvm_i32_ty, llvm_ptr_ty, |
| llvm_ptr_ty, llvm_ptr_ty, llvm_ptr_ty], |
| []>; |
| def int_coro_alloc : Intrinsic<[llvm_i1_ty], [llvm_token_ty], []>; |
| def int_coro_id_async : Intrinsic<[llvm_token_ty], |
| [llvm_i32_ty, llvm_i32_ty, llvm_i32_ty, llvm_ptr_ty], |
| []>; |
| def int_coro_async_context_alloc : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty, llvm_ptr_ty], |
| []>; |
| def int_coro_async_context_dealloc : Intrinsic<[], |
| [llvm_ptr_ty], |
| []>; |
| def int_coro_async_resume : Intrinsic<[llvm_ptr_ty], |
| [], |
| []>; |
| def int_coro_async_size_replace : Intrinsic<[], [llvm_ptr_ty, llvm_ptr_ty], []>; |
| def int_coro_suspend_async |
| : Intrinsic<[llvm_any_ty], |
| [llvm_i32_ty, llvm_ptr_ty, llvm_ptr_ty, llvm_vararg_ty], []>; |
| def int_coro_prepare_async : Intrinsic<[llvm_ptr_ty], [llvm_ptr_ty], |
| [IntrNoMem]>; |
| def int_coro_begin : Intrinsic<[llvm_ptr_ty], [llvm_token_ty, llvm_ptr_ty], |
| [WriteOnly<ArgIndex<1>>]>; |
| |
| def int_coro_free : Intrinsic<[llvm_ptr_ty], [llvm_token_ty, llvm_ptr_ty], |
| [IntrReadMem, IntrArgMemOnly, |
| ReadOnly<ArgIndex<1>>, |
| NoCapture<ArgIndex<1>>]>; |
| def int_coro_end : Intrinsic<[llvm_i1_ty], [llvm_ptr_ty, llvm_i1_ty], []>; |
| def int_coro_end_async |
| : Intrinsic<[llvm_i1_ty], [llvm_ptr_ty, llvm_i1_ty, llvm_vararg_ty], []>; |
| |
| def int_coro_frame : Intrinsic<[llvm_ptr_ty], [], [IntrNoMem]>; |
| def int_coro_noop : Intrinsic<[llvm_ptr_ty], [], [IntrNoMem]>; |
| def int_coro_size : Intrinsic<[llvm_anyint_ty], [], [IntrNoMem]>; |
| def int_coro_align : Intrinsic<[llvm_anyint_ty], [], [IntrNoMem]>; |
| |
| def int_coro_save : Intrinsic<[llvm_token_ty], [llvm_ptr_ty], []>; |
| def int_coro_suspend : Intrinsic<[llvm_i8_ty], [llvm_token_ty, llvm_i1_ty], []>; |
| def int_coro_suspend_retcon : Intrinsic<[llvm_any_ty], [llvm_vararg_ty], []>; |
| def int_coro_prepare_retcon : Intrinsic<[llvm_ptr_ty], [llvm_ptr_ty], |
| [IntrNoMem]>; |
| def int_coro_alloca_alloc : Intrinsic<[llvm_token_ty], |
| [llvm_anyint_ty, llvm_i32_ty], []>; |
| def int_coro_alloca_get : Intrinsic<[llvm_ptr_ty], [llvm_token_ty], []>; |
| def int_coro_alloca_free : Intrinsic<[], [llvm_token_ty], []>; |
| |
| // Coroutine Manipulation Intrinsics. |
| |
| def int_coro_resume : Intrinsic<[], [llvm_ptr_ty], [Throws]>; |
| def int_coro_destroy : Intrinsic<[], [llvm_ptr_ty], [Throws]>; |
| def int_coro_done : Intrinsic<[llvm_i1_ty], [llvm_ptr_ty], |
| [IntrArgMemOnly, ReadOnly<ArgIndex<0>>, |
| NoCapture<ArgIndex<0>>]>; |
| def int_coro_promise : Intrinsic<[llvm_ptr_ty], |
| [llvm_ptr_ty, llvm_i32_ty, llvm_i1_ty], |
| [IntrNoMem, NoCapture<ArgIndex<0>>]>; |
| |
| // Coroutine Lowering Intrinsics. Used internally by coroutine passes. |
| |
| def int_coro_subfn_addr : Intrinsic<[llvm_ptr_ty], [llvm_ptr_ty, llvm_i8_ty], |
| [IntrReadMem, IntrArgMemOnly, |
| ReadOnly<ArgIndex<0>>, |
| NoCapture<ArgIndex<0>>]>; |
| |
| ///===-------------------------- Other Intrinsics --------------------------===// |
| // |
| def int_trap : Intrinsic<[], [], [IntrNoReturn, IntrCold]>, |
| GCCBuiltin<"__builtin_trap">; |
| def int_debugtrap : Intrinsic<[]>, |
| GCCBuiltin<"__builtin_debugtrap">; |
| def int_ubsantrap : Intrinsic<[], [llvm_i8_ty], |
| [IntrNoReturn, IntrCold, ImmArg<ArgIndex<0>>]>; |
| |
| // Support for dynamic deoptimization (or de-specialization) |
| def int_experimental_deoptimize : Intrinsic<[llvm_any_ty], [llvm_vararg_ty], |
| [Throws]>; |
| |
| // Support for speculative runtime guards |
| def int_experimental_guard : DefaultAttrsIntrinsic<[], [llvm_i1_ty, llvm_vararg_ty], |
| [Throws]>; |
| |
| // Supports widenable conditions for guards represented as explicit branches. |
| def int_experimental_widenable_condition : DefaultAttrsIntrinsic<[llvm_i1_ty], [], |
| [IntrInaccessibleMemOnly, IntrWillReturn, IntrSpeculatable]>; |
| |
| // NOP: calls/invokes to this intrinsic are removed by codegen |
| def int_donothing : DefaultAttrsIntrinsic<[], [], [IntrNoMem, IntrWillReturn]>; |
| |
| // This instruction has no actual effect, though it is treated by the optimizer |
| // has having opaque side effects. This may be inserted into loops to ensure |
| // that they are not removed even if they turn out to be empty, for languages |
| // which specify that infinite loops must be preserved. |
| def int_sideeffect : DefaultAttrsIntrinsic<[], [], [IntrInaccessibleMemOnly, IntrWillReturn]>; |
| |
| // The pseudoprobe intrinsic works as a place holder to the block it probes. |
| // Like the sideeffect intrinsic defined above, this intrinsic is treated by the |
| // optimizer as having opaque side effects so that it won't be get rid of or moved |
| // out of the block it probes. |
| def int_pseudoprobe : DefaultAttrsIntrinsic<[], [llvm_i64_ty, llvm_i64_ty, llvm_i32_ty, llvm_i64_ty], |
| [IntrInaccessibleMemOnly, IntrWillReturn]>; |
| |
| // Arithmetic fence intrinsic. |
| def int_arithmetic_fence : Intrinsic<[llvm_anyfloat_ty], [LLVMMatchType<0>], [IntrNoMem]>; |
| |
| // Intrinsics to support half precision floating point format |
| let IntrProperties = [IntrNoMem, IntrWillReturn] in { |
| def int_convert_to_fp16 : DefaultAttrsIntrinsic<[llvm_i16_ty], [llvm_anyfloat_ty]>; |
| def int_convert_from_fp16 : DefaultAttrsIntrinsic<[llvm_anyfloat_ty], [llvm_i16_ty]>; |
| } |
| |
| // Saturating floating point to integer intrinsics |
| let IntrProperties = [IntrNoMem, IntrSpeculatable, IntrWillReturn] in { |
| def int_fptoui_sat : DefaultAttrsIntrinsic<[llvm_anyint_ty], [llvm_anyfloat_ty]>; |
| def int_fptosi_sat : DefaultAttrsIntrinsic<[llvm_anyint_ty], [llvm_anyfloat_ty]>; |
| } |
| |
| // Clear cache intrinsic, default to ignore (ie. emit nothing) |
| // maps to void __clear_cache() on supporting platforms |
| def int_clear_cache : Intrinsic<[], [llvm_ptr_ty, llvm_ptr_ty], |
| [], "llvm.clear_cache">; |
| |
| // Intrinsic to detect whether its argument is a constant. |
| def int_is_constant : DefaultAttrsIntrinsic<[llvm_i1_ty], [llvm_any_ty], |
| [IntrNoMem, IntrWillReturn, IntrConvergent], |
| "llvm.is.constant">; |
| |
| // Intrinsic to mask out bits of a pointer. |
| def int_ptrmask: DefaultAttrsIntrinsic<[llvm_anyptr_ty], [LLVMMatchType<0>, llvm_anyint_ty], |
| [IntrNoMem, IntrSpeculatable, IntrWillReturn]>; |
| |
| def int_experimental_stepvector : DefaultAttrsIntrinsic<[llvm_anyvector_ty], |
| [], [IntrNoMem]>; |
| |
| //===---------------- Vector Predication Intrinsics --------------===// |
| // Memory Intrinsics |
| def int_vp_store : DefaultAttrsIntrinsic<[], |
| [ llvm_anyvector_ty, |
| LLVMAnyPointerType<LLVMMatchType<0>>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty], |
| [ NoCapture<ArgIndex<1>>, IntrNoSync, IntrWriteMem, IntrArgMemOnly, IntrWillReturn ]>; |
| |
| def int_vp_load : DefaultAttrsIntrinsic<[ llvm_anyvector_ty], |
| [ LLVMAnyPointerType<LLVMMatchType<0>>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty], |
| [ NoCapture<ArgIndex<0>>, IntrNoSync, IntrReadMem, IntrWillReturn, IntrArgMemOnly ]>; |
| |
| def int_vp_gather: DefaultAttrsIntrinsic<[ llvm_anyvector_ty], |
| [ LLVMVectorOfAnyPointersToElt<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty], |
| [ IntrReadMem, IntrNoSync, IntrWillReturn, IntrArgMemOnly ]>; |
| |
| def int_vp_scatter: DefaultAttrsIntrinsic<[], |
| [ llvm_anyvector_ty, |
| LLVMVectorOfAnyPointersToElt<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty], |
| [ IntrArgMemOnly, IntrNoSync, IntrWillReturn ]>; // TODO allow IntrNoCapture for vectors of pointers |
| |
| // Speculatable Binary operators |
| let IntrProperties = [IntrSpeculatable, IntrNoMem, IntrNoSync, IntrWillReturn] in { |
| def int_vp_add : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_sub : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_mul : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_ashr : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_lshr : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_shl : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_or : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_and : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_xor : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| } |
| |
| // Non-speculatable binary operators. |
| let IntrProperties = [IntrNoMem, IntrNoSync, IntrWillReturn] in { |
| def int_vp_sdiv : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_udiv : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_srem : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_urem : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| } |
| |
| // Floating-point arithmetic. |
| let IntrProperties = |
| [IntrSpeculatable, IntrNoMem, IntrNoSync, IntrWillReturn] in { |
| def int_vp_fadd : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_fsub : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_fmul : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_fdiv : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_frem : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| } |
| // Shuffles. |
| def int_vp_select : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| llvm_i32_ty]>; |
| |
| def int_vp_merge : DefaultAttrsIntrinsic<[ llvm_anyvector_ty ], |
| [ LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| llvm_i32_ty]>; |
| |
| // Reductions |
| let IntrProperties = [IntrSpeculatable, IntrNoMem, IntrNoSync, IntrWillReturn] in { |
| def int_vp_reduce_fadd : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [LLVMVectorElementType<0>, |
| llvm_anyvector_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_reduce_fmul : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [LLVMVectorElementType<0>, |
| llvm_anyvector_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_reduce_add : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [LLVMVectorElementType<0>, |
| llvm_anyvector_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_reduce_mul : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [LLVMVectorElementType<0>, |
| llvm_anyvector_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_reduce_and : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [LLVMVectorElementType<0>, |
| llvm_anyvector_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_reduce_or : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [LLVMVectorElementType<0>, |
| llvm_anyvector_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_reduce_xor : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [LLVMVectorElementType<0>, |
| llvm_anyvector_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_reduce_smax : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [LLVMVectorElementType<0>, |
| llvm_anyvector_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_reduce_smin : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [LLVMVectorElementType<0>, |
| llvm_anyvector_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_reduce_umax : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [LLVMVectorElementType<0>, |
| llvm_anyvector_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_reduce_umin : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [LLVMVectorElementType<0>, |
| llvm_anyvector_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_reduce_fmax : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [LLVMVectorElementType<0>, |
| llvm_anyvector_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| def int_vp_reduce_fmin : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [LLVMVectorElementType<0>, |
| llvm_anyvector_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty]>; |
| } |
| |
| def int_get_active_lane_mask: |
| DefaultAttrsIntrinsic<[llvm_anyvector_ty], |
| [llvm_anyint_ty, LLVMMatchType<1>], |
| [IntrNoMem, IntrNoSync, IntrWillReturn]>; |
| |
| def int_experimental_vp_splice: |
| DefaultAttrsIntrinsic<[llvm_anyvector_ty], |
| [LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| llvm_i32_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| llvm_i32_ty, llvm_i32_ty], |
| [IntrNoMem, ImmArg<ArgIndex<2>>]>; |
| |
| //===-------------------------- Masked Intrinsics -------------------------===// |
| // |
| def int_masked_load: |
| DefaultAttrsIntrinsic<[llvm_anyvector_ty], |
| [LLVMAnyPointerType<LLVMMatchType<0>>, llvm_i32_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, LLVMMatchType<0>], |
| [IntrReadMem, IntrArgMemOnly, IntrWillReturn, ImmArg<ArgIndex<1>>]>; |
| |
| def int_masked_store: |
| DefaultAttrsIntrinsic<[], |
| [llvm_anyvector_ty, LLVMAnyPointerType<LLVMMatchType<0>>, |
| llvm_i32_ty, LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>], |
| [IntrWriteMem, IntrArgMemOnly, IntrWillReturn, |
| ImmArg<ArgIndex<2>>]>; |
| |
| def int_masked_gather: |
| DefaultAttrsIntrinsic<[llvm_anyvector_ty], |
| [LLVMVectorOfAnyPointersToElt<0>, llvm_i32_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, LLVMMatchType<0>], |
| [IntrReadMem, IntrWillReturn, ImmArg<ArgIndex<1>>]>; |
| |
| def int_masked_scatter: |
| DefaultAttrsIntrinsic<[], |
| [llvm_anyvector_ty, LLVMVectorOfAnyPointersToElt<0>, llvm_i32_ty, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>], |
| [IntrWriteMem, IntrWillReturn, ImmArg<ArgIndex<2>>]>; |
| |
| def int_masked_expandload: |
| DefaultAttrsIntrinsic<[llvm_anyvector_ty], |
| [LLVMPointerToElt<0>, LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>, |
| LLVMMatchType<0>], |
| [IntrReadMem, IntrWillReturn]>; |
| |
| def int_masked_compressstore: |
| DefaultAttrsIntrinsic<[], |
| [llvm_anyvector_ty, LLVMPointerToElt<0>, |
| LLVMScalarOrSameVectorWidth<0, llvm_i1_ty>], |
| [IntrWriteMem, IntrArgMemOnly, IntrWillReturn]>; |
| |
| // Test whether a pointer is associated with a type metadata identifier. |
| def int_type_test : DefaultAttrsIntrinsic<[llvm_i1_ty], [llvm_ptr_ty, llvm_metadata_ty], |
| [IntrNoMem, IntrWillReturn, IntrSpeculatable]>; |
| |
| // Safely loads a function pointer from a virtual table pointer using type metadata. |
| def int_type_checked_load : DefaultAttrsIntrinsic<[llvm_ptr_ty, llvm_i1_ty], |
| [llvm_ptr_ty, llvm_i32_ty, llvm_metadata_ty], |
| [IntrNoMem, IntrWillReturn]>; |
| |
| // Create a branch funnel that implements an indirect call to a limited set of |
| // callees. This needs to be a musttail call. |
| def int_icall_branch_funnel : DefaultAttrsIntrinsic<[], [llvm_vararg_ty], []>; |
| |
| def int_load_relative: DefaultAttrsIntrinsic<[llvm_ptr_ty], [llvm_ptr_ty, llvm_anyint_ty], |
| [IntrReadMem, IntrArgMemOnly]>; |
| |
| def int_asan_check_memaccess : |
| Intrinsic<[],[llvm_ptr_ty, llvm_i32_ty], [ImmArg<ArgIndex<1>>]>; |
| |
| def int_hwasan_check_memaccess : |
| Intrinsic<[], [llvm_ptr_ty, llvm_ptr_ty, llvm_i32_ty], |
| [ImmArg<ArgIndex<2>>]>; |
| def int_hwasan_check_memaccess_shortgranules : |
| Intrinsic<[], [llvm_ptr_ty, llvm_ptr_ty, llvm_i32_ty], |
| [ImmArg<ArgIndex<2>>]>; |
| |
| // Xray intrinsics |
| //===----------------------------------------------------------------------===// |
| // Custom event logging for x-ray. |
| // Takes a pointer to a string and the length of the string. |
| def int_xray_customevent : Intrinsic<[], [llvm_ptr_ty, llvm_i32_ty], |
| [IntrWriteMem, NoCapture<ArgIndex<0>>, |
| ReadOnly<ArgIndex<0>>]>; |
| // Typed event logging for x-ray. |
| // Takes a numeric type tag, a pointer to a string and the length of the string. |
| def int_xray_typedevent : Intrinsic<[], [llvm_i16_ty, llvm_ptr_ty, llvm_i32_ty], |
| [IntrWriteMem, NoCapture<ArgIndex<1>>, |
| ReadOnly<ArgIndex<1>>]>; |
| //===----------------------------------------------------------------------===// |
| |
| //===------ Memory intrinsics with element-wise atomicity guarantees ------===// |
| // |
| |
| // @llvm.memcpy.element.unordered.atomic.*(dest, src, length, elementsize) |
| def int_memcpy_element_unordered_atomic |
| : Intrinsic<[], |
| [llvm_anyptr_ty, llvm_anyptr_ty, llvm_anyint_ty, llvm_i32_ty], |
| [IntrArgMemOnly, IntrWillReturn, IntrNoSync, |
| NoCapture<ArgIndex<0>>, NoCapture<ArgIndex<1>>, |
| WriteOnly<ArgIndex<0>>, ReadOnly<ArgIndex<1>>, |
| ImmArg<ArgIndex<3>>]>; |
| |
| // @llvm.memmove.element.unordered.atomic.*(dest, src, length, elementsize) |
| def int_memmove_element_unordered_atomic |
| : Intrinsic<[], |
| [llvm_anyptr_ty, llvm_anyptr_ty, llvm_anyint_ty, llvm_i32_ty], |
| [IntrArgMemOnly, IntrWillReturn, IntrNoSync, |
| NoCapture<ArgIndex<0>>, NoCapture<ArgIndex<1>>, |
| WriteOnly<ArgIndex<0>>, ReadOnly<ArgIndex<1>>, |
| ImmArg<ArgIndex<3>>]>; |
| |
| // @llvm.memset.element.unordered.atomic.*(dest, value, length, elementsize) |
| def int_memset_element_unordered_atomic |
| : Intrinsic<[], [llvm_anyptr_ty, llvm_i8_ty, llvm_anyint_ty, llvm_i32_ty], |
| [IntrWriteMem, IntrArgMemOnly, IntrWillReturn, IntrNoSync, |
| NoCapture<ArgIndex<0>>, WriteOnly<ArgIndex<0>>, |
| ImmArg<ArgIndex<3>>]>; |
| |
| //===------------------------ Reduction Intrinsics ------------------------===// |
| // |
| let IntrProperties = [IntrNoMem] in { |
| |
| def int_vector_reduce_fadd : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [LLVMVectorElementType<0>, |
| llvm_anyvector_ty]>; |
| def int_vector_reduce_fmul : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [LLVMVectorElementType<0>, |
| llvm_anyvector_ty]>; |
| def int_vector_reduce_add : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [llvm_anyvector_ty]>; |
| def int_vector_reduce_mul : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [llvm_anyvector_ty]>; |
| def int_vector_reduce_and : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [llvm_anyvector_ty]>; |
| def int_vector_reduce_or : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [llvm_anyvector_ty]>; |
| def int_vector_reduce_xor : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [llvm_anyvector_ty]>; |
| def int_vector_reduce_smax : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [llvm_anyvector_ty]>; |
| def int_vector_reduce_smin : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [llvm_anyvector_ty]>; |
| def int_vector_reduce_umax : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [llvm_anyvector_ty]>; |
| def int_vector_reduce_umin : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [llvm_anyvector_ty]>; |
| def int_vector_reduce_fmax : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [llvm_anyvector_ty]>; |
| def int_vector_reduce_fmin : DefaultAttrsIntrinsic<[LLVMVectorElementType<0>], |
| [llvm_anyvector_ty]>; |
| } |
| |
| //===----- Matrix intrinsics ---------------------------------------------===// |
| |
| def int_matrix_transpose |
| : DefaultAttrsIntrinsic<[llvm_anyvector_ty], |
| [LLVMMatchType<0>, llvm_i32_ty, llvm_i32_ty], |
| [ IntrNoSync, IntrWillReturn, IntrNoMem, IntrSpeculatable, ImmArg<ArgIndex<1>>, |
| ImmArg<ArgIndex<2>>]>; |
| |
| def int_matrix_multiply |
| : DefaultAttrsIntrinsic<[llvm_anyvector_ty], |
| [llvm_anyvector_ty, llvm_anyvector_ty, llvm_i32_ty, llvm_i32_ty, |
| llvm_i32_ty], |
| [IntrNoSync, IntrWillReturn, IntrNoMem, IntrSpeculatable, ImmArg<ArgIndex<2>>, |
| ImmArg<ArgIndex<3>>, ImmArg<ArgIndex<4>>]>; |
| |
| def int_matrix_column_major_load |
| : DefaultAttrsIntrinsic<[llvm_anyvector_ty], |
| [LLVMPointerToElt<0>, llvm_anyint_ty, llvm_i1_ty, |
| llvm_i32_ty, llvm_i32_ty], |
| [IntrNoSync, IntrWillReturn, IntrArgMemOnly, IntrReadMem, |
| NoCapture<ArgIndex<0>>, ImmArg<ArgIndex<2>>, ImmArg<ArgIndex<3>>, |
| ImmArg<ArgIndex<4>>]>; |
| |
| def int_matrix_column_major_store |
| : DefaultAttrsIntrinsic<[], |
| [llvm_anyvector_ty, LLVMPointerToElt<0>, |
| llvm_anyint_ty, llvm_i1_ty, llvm_i32_ty, llvm_i32_ty], |
| [IntrNoSync, IntrWillReturn, IntrArgMemOnly, IntrWriteMem, |
| WriteOnly<ArgIndex<1>>, NoCapture<ArgIndex<1>>, |
| ImmArg<ArgIndex<3>>, ImmArg<ArgIndex<4>>, ImmArg<ArgIndex<5>>]>; |
| |
| //===---------- Intrinsics to control hardware supported loops ----------===// |
| |
| // Specify that the value given is the number of iterations that the next loop |
| // will execute. |
| def int_set_loop_iterations : |
| DefaultAttrsIntrinsic<[], [llvm_anyint_ty], [IntrNoDuplicate]>; |
| |
| // Same as the above, but produces a value (the same as the input operand) to |
| // be fed into the loop. |
| def int_start_loop_iterations : |
| DefaultAttrsIntrinsic<[llvm_anyint_ty], [LLVMMatchType<0>], [IntrNoDuplicate]>; |
| |
| // Specify that the value given is the number of iterations that the next loop |
| // will execute. Also test that the given count is not zero, allowing it to |
| // control entry to a 'while' loop. |
| def int_test_set_loop_iterations : |
| DefaultAttrsIntrinsic<[llvm_i1_ty], [llvm_anyint_ty], [IntrNoDuplicate]>; |
| |
| // Same as the above, but produces an extra value (the same as the input |
| // operand) to be fed into the loop. |
| def int_test_start_loop_iterations : |
| DefaultAttrsIntrinsic<[llvm_anyint_ty, llvm_i1_ty], [LLVMMatchType<0>], |
| [IntrNoDuplicate]>; |
| |
| // Decrement loop counter by the given argument. Return false if the loop |
| // should exit. |
| def int_loop_decrement : |
| DefaultAttrsIntrinsic<[llvm_i1_ty], [llvm_anyint_ty], [IntrNoDuplicate]>; |
| |
| // Decrement the first operand (the loop counter) by the second operand (the |
| // maximum number of elements processed in an iteration). Return the remaining |
| // number of iterations still to be executed. This is effectively a sub which |
| // can be used with a phi, icmp and br to control the number of iterations |
| // executed, as usual. Any optimisations are allowed to treat it is a sub, and |
| // it's scevable, so it's the backends responsibility to handle cases where it |
| // may be optimised. |
| def int_loop_decrement_reg : |
| DefaultAttrsIntrinsic<[llvm_anyint_ty], |
| [LLVMMatchType<0>, LLVMMatchType<0>], [IntrNoDuplicate]>; |
| |
| //===----- Intrinsics that are used to provide predicate information -----===// |
| |
| def int_ssa_copy : DefaultAttrsIntrinsic<[llvm_any_ty], [LLVMMatchType<0>], |
| [IntrNoMem, Returned<ArgIndex<0>>]>; |
| |
| //===------- Intrinsics that are used to preserve debug information -------===// |
| |
| def int_preserve_array_access_index : DefaultAttrsIntrinsic<[llvm_anyptr_ty], |
| [llvm_anyptr_ty, llvm_i32_ty, |
| llvm_i32_ty], |
| [IntrNoMem, |
| ImmArg<ArgIndex<1>>, |
| ImmArg<ArgIndex<2>>]>; |
| def int_preserve_union_access_index : DefaultAttrsIntrinsic<[llvm_anyptr_ty], |
| [llvm_anyptr_ty, llvm_i32_ty], |
| [IntrNoMem, |
| ImmArg<ArgIndex<1>>]>; |
| def int_preserve_struct_access_index : DefaultAttrsIntrinsic<[llvm_anyptr_ty], |
| [llvm_anyptr_ty, llvm_i32_ty, |
| llvm_i32_ty], |
| [IntrNoMem, |
| ImmArg<ArgIndex<1>>, |
| ImmArg<ArgIndex<2>>]>; |
| |
| //===------------ Intrinsics to perform common vector shuffles ------------===// |
| |
| def int_experimental_vector_reverse : DefaultAttrsIntrinsic<[llvm_anyvector_ty], |
| [LLVMMatchType<0>], |
| [IntrNoMem]>; |
| |
| //===---------- Intrinsics to query properties of scalable vectors --------===// |
| def int_vscale : DefaultAttrsIntrinsic<[llvm_anyint_ty], [], [IntrNoMem]>; |
| |
| //===---------- Intrinsics to perform subvector insertion/extraction ------===// |
| def int_experimental_vector_insert : DefaultAttrsIntrinsic<[llvm_anyvector_ty], |
| [LLVMMatchType<0>, llvm_anyvector_ty, llvm_i64_ty], |
| [IntrNoMem, ImmArg<ArgIndex<2>>]>; |
| |
| def int_experimental_vector_extract : DefaultAttrsIntrinsic<[llvm_anyvector_ty], |
| [llvm_anyvector_ty, llvm_i64_ty], |
| [IntrNoMem, ImmArg<ArgIndex<1>>]>; |
| |
| //===---------- Named shufflevector intrinsics ------===// |
| def int_experimental_vector_splice : DefaultAttrsIntrinsic<[llvm_anyvector_ty], |
| [LLVMMatchType<0>, |
| LLVMMatchType<0>, |
| llvm_i32_ty], |
| [IntrNoMem, ImmArg<ArgIndex<2>>]>; |
| |
| |
| //===----------------- Pointer Authentication Intrinsics ------------------===// |
| // |
| |
| // Sign an unauthenticated pointer using the specified key and discriminator, |
| // passed in that order. |
| // Returns the first argument, with some known bits replaced with a signature. |
| def int_ptrauth_sign : Intrinsic<[llvm_i64_ty], |
| [llvm_i64_ty, llvm_i32_ty, llvm_i64_ty], |
| [IntrNoMem, ImmArg<ArgIndex<1>>]>; |
| |
| // Authenticate a signed pointer, using the specified key and discriminator. |
| // Returns the first argument, with the signature bits removed. |
| // The signature must be valid. |
| def int_ptrauth_auth : Intrinsic<[llvm_i64_ty], |
| [llvm_i64_ty, llvm_i32_ty, llvm_i64_ty], |
| [IntrNoMem,ImmArg<ArgIndex<1>>]>; |
| |
| // Authenticate a signed pointer and resign it. |
| // The second (key) and third (discriminator) arguments specify the signing |
| // schema used for authenticating. |
| // The fourth and fifth arguments specify the schema used for signing. |
| // The signature must be valid. |
| // This is a combined form of @llvm.ptrauth.sign and @llvm.ptrauth.auth, with |
| // an additional integrity guarantee on the intermediate value. |
| def int_ptrauth_resign : Intrinsic<[llvm_i64_ty], |
| [llvm_i64_ty, llvm_i32_ty, llvm_i64_ty, |
| llvm_i32_ty, llvm_i64_ty], |
| [IntrNoMem, ImmArg<ArgIndex<1>>, |
| ImmArg<ArgIndex<3>>]>; |
| |
| // Strip the embedded signature out of a signed pointer. |
| // The second argument specifies the key. |
| // This behaves like @llvm.ptrauth.auth, but doesn't require the signature to |
| // be valid. |
| def int_ptrauth_strip : Intrinsic<[llvm_i64_ty], |
| [llvm_i64_ty, llvm_i32_ty], |
| [IntrNoMem, ImmArg<ArgIndex<1>>]>; |
| |
| // Blend a small integer discriminator with an address discriminator, producing |
| // a new discriminator value. |
| def int_ptrauth_blend : Intrinsic<[llvm_i64_ty], |
| [llvm_i64_ty, llvm_i64_ty], |
| [IntrNoMem]>; |
| |
| // Compute the signature of a value, using a given discriminator. |
| // This differs from @llvm.ptrauth.sign in that it doesn't embed the computed |
| // signature in the pointer, but instead returns the signature as a value. |
| // That allows it to be used to sign non-pointer data: in that sense, it is |
| // generic. There is no generic @llvm.ptrauth.auth: instead, the signature |
| // can be computed using @llvm.ptrauth.sign_generic, and compared with icmp. |
| def int_ptrauth_sign_generic : Intrinsic<[llvm_i64_ty], |
| [llvm_i64_ty, llvm_i64_ty], |
| [IntrNoMem]>; |
| |
| //===----------------------------------------------------------------------===// |
| |
| //===----------------------------------------------------------------------===// |
| // Target-specific intrinsics |
| //===----------------------------------------------------------------------===// |
| |
| include "llvm/IR/IntrinsicsPowerPC.td" |
| include "llvm/IR/IntrinsicsX86.td" |
| include "llvm/IR/IntrinsicsARM.td" |
| include "llvm/IR/IntrinsicsAArch64.td" |
| include "llvm/IR/IntrinsicsXCore.td" |
| include "llvm/IR/IntrinsicsHexagon.td" |
| include "llvm/IR/IntrinsicsNVVM.td" |
| include "llvm/IR/IntrinsicsMips.td" |
| include "llvm/IR/IntrinsicsAMDGPU.td" |
| include "llvm/IR/IntrinsicsBPF.td" |
| include "llvm/IR/IntrinsicsSystemZ.td" |
| include "llvm/IR/IntrinsicsWebAssembly.td" |
| include "llvm/IR/IntrinsicsRISCV.td" |
| include "llvm/IR/IntrinsicsVE.td" |