/netbsd-current/external/apache2/llvm/dist/llvm/lib/Transforms/Scalar/ |
H A D | AlignmentFromAssumptions.cpp | 206 bool AlignmentFromAssumptionsPass::extractAlignmentInfo(CallInst *I, 229 bool AlignmentFromAssumptionsPass::processAssumption(CallInst *ACall, 336 CallInst *Call = cast<CallInst>(AssumeVH);
|
H A D | ScalarizeMaskedMemIntrin.cpp | 74 static bool optimizeCallInst(CallInst *CI, bool &ModifiedDT, 145 static void scalarizeMaskedLoad(const DataLayout &DL, CallInst *CI, 287 static void scalarizeMaskedStore(const DataLayout &DL, CallInst *CI, 415 static void scalarizeMaskedGather(const DataLayout &DL, CallInst *CI, 545 static void scalarizeMaskedScatter(const DataLayout &DL, CallInst *CI, 635 static void scalarizeMaskedExpandLoad(const DataLayout &DL, CallInst *CI, 762 static void scalarizeMaskedCompressStore(const DataLayout &DL, CallInst *CI, 920 if (CallInst *CI = dyn_cast<CallInst>(&*CurInstIterator++)) 929 static bool optimizeCallInst(CallInst *C [all...] |
H A D | AnnotationRemarks.cpp | 70 if (auto *CI = dyn_cast<CallInst>(I)) {
|
/netbsd-current/external/apache2/llvm/dist/llvm/lib/Analysis/ |
H A D | AssumeBundleQueries.cpp | 132 static CallInst::BundleOpInfo *getBundleFromUse(const Use *U) { 143 CallInst::BundleOpInfo* Bundle = getBundleFromUse(U); 182 CallInst::BundleOpInfo* Bundle = getBundleFromUse(&U);
|
/netbsd-current/external/apache2/llvm/dist/llvm/lib/Transforms/IPO/ |
H A D | PruneEH.cpp | 120 if (const auto *CI = dyn_cast<CallInst>(&I)) { 201 if (CallInst *CI = dyn_cast<CallInst>(I++))
|
/netbsd-current/external/apache2/llvm/dist/llvm/lib/Target/WebAssembly/ |
H A D | WebAssemblyLowerEmscriptenEHSjLj.cpp | 430 CallInst *NewCall = IRB.CreateCall(getInvokeWrapper(CI), Args); 654 SmallVector<CallInst *, 8> ToErase; 661 auto *CI = dyn_cast<CallInst>(U); 826 CallInst *NewCall = 866 auto *CI = dyn_cast<CallInst>(&I); 877 CallInst *NewCI = 907 CallInst *FMCI = IRB.CreateCall(FMCF, FMCArgs, "fmc"); 968 Instruction *SetjmpTable = CallInst::CreateMalloc( 972 // CallInst::CreateMalloc may return a bitcast instruction if the result types 988 auto *CI = dyn_cast<CallInst>( [all...] |
H A D | WebAssemblyLowerGlobalDtors.cpp | 166 CallInst::Create(VoidVoid, Dtor, "", BB); 185 Value *Res = CallInst::Create(AtExit, Args, "call", EntryBB); 193 CallInst::Create(Intrinsic::getDeclaration(&M, Intrinsic::trap), "",
|
H A D | WebAssemblyTargetTransformInfo.cpp | 122 if (isa<CallInst>(I) || isa<InvokeInst>(I))
|
/netbsd-current/external/apache2/llvm/dist/llvm/lib/IR/ |
H A D | BasicBlock.cpp | 154 const CallInst *BasicBlock::getTerminatingMustTailCall() const { 178 if (auto *CI = dyn_cast<CallInst>(Prev)) { 185 const CallInst *BasicBlock::getTerminatingDeoptimizeCall() const { 192 if (auto *CI = dyn_cast_or_null<CallInst>(RI->getPrevNode())) 200 const CallInst *BasicBlock::getPostdominatingDeoptimizeCall() const {
|
H A D | Instruction.cpp | 428 if (const CallInst *CI = dyn_cast<CallInst>(I1)) 429 return CI->isTailCall() == cast<CallInst>(I2)->isTailCall() && 430 CI->getCallingConv() == cast<CallInst>(I2)->getCallingConv() && 431 CI->getAttributes() == cast<CallInst>(I2)->getAttributes() && 432 CI->hasIdenticalOperandBundleSchema(*cast<CallInst>(I2)); 653 if (const CallInst *CI = dyn_cast<CallInst>(this)) 663 return (!isa<CallInst>(this) || !this->mayHaveSideEffects()) &&
|
/netbsd-current/external/apache2/llvm/dist/clang/lib/CodeGen/ |
H A D | CGBuilder.h | 285 llvm::CallInst *CreateMemCpy(Address Dest, Address Src, llvm::Value *Size, 291 llvm::CallInst *CreateMemCpy(Address Dest, Address Src, uint64_t Size, 299 llvm::CallInst *CreateMemCpyInline(Address Dest, Address Src, uint64_t Size) { 306 llvm::CallInst *CreateMemMove(Address Dest, Address Src, llvm::Value *Size, 314 llvm::CallInst *CreateMemSet(Address Dest, llvm::Value *Value,
|
/netbsd-current/external/apache2/llvm/dist/llvm/lib/CodeGen/ |
H A D | IntrinsicLowering.cpp | 30 static CallInst *ReplaceCallWith(const char *NewFn, CallInst *CI, 45 CallInst *NewCI = Builder.CreateCall(FCache, Args); 202 static void ReplaceFPIntrinsicWithCall(CallInst *CI, const char *Fname, 224 void IntrinsicLowering::LowerIntrinsicCall(CallInst *CI) { 454 bool IntrinsicLowering::LowerToByteSwap(CallInst *CI) { 470 Op = CallInst::Create(Int, Op, CI->getName(), CI);
|
H A D | SjLjEHPrepare.cpp | 434 CallInst::Create(CallSiteFn, CallSiteNum, "", Invokes[I]); 451 CallInst *Register = 452 CallInst::Create(RegisterFn, FuncCtx, "", EntryBB->getTerminator()); 461 if (auto *CI = dyn_cast<CallInst>(&I)) { 467 Instruction *StackAddr = CallInst::Create(StackAddrFn, "sp"); 476 CallInst::Create(UnregisterFn, FuncCtx, "", Return);
|
H A D | WasmEHPrepare.cpp | 196 auto *ThrowI = cast<CallInst>(U); 292 if (auto *CI = dyn_cast<CallInst>(U.getUser())) { 345 CallInst *PersCI = IRB.CreateCall(CallPersonalityF, CatchCI,
|
/netbsd-current/external/apache2/llvm/dist/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPUPromoteAlloca.cpp | 212 CallInst *LocalSizeY = Builder.CreateCall(LocalSizeYFn, {}); 213 CallInst *LocalSizeZ = Builder.CreateCall(LocalSizeZFn, {}); 258 CallInst *DispatchPtr = Builder.CreateCall(DispatchPtrFn, {}); 314 CallInst *CI = Builder.CreateCall(WorkitemIdFn); 552 static bool isCallPromotable(CallInst *CI) { 611 if (CallInst *CI = dyn_cast<CallInst>(User)) { 959 CallInst *Call = dyn_cast<CallInst>(V); 1044 CallInst *NewCal [all...] |
H A D | AMDGPULowerIntrinsics.cpp | 134 auto *CI = dyn_cast<CallInst>(U);
|
H A D | AMDGPUOpenCLEnqueuedBlockLowering.cpp | 78 if (auto *CI = dyn_cast<CallInst>(&*U)) {
|
H A D | AMDGPUCallLowering.h | 76 MachineIRBuilder &MIRBuilder, MachineInstrBuilder &CallInst,
|
/netbsd-current/external/apache2/llvm/dist/llvm/lib/Transforms/Instrumentation/ |
H A D | ThreadSanitizer.cpp | 553 else if (isa<CallInst>(Inst) || isa<InvokeInst>(Inst)) { 554 if (CallInst *CI = dyn_cast<CallInst>(&Inst)) 770 CallInst *C = CallInst::Create(TsanAtomicStore[Idx], Args); 787 CallInst *C = CallInst::Create(F, Args); 807 CallInst *C = IRB.CreateCall(TsanAtomicCAS[Idx], Args); 827 CallInst *C = CallInst [all...] |
/netbsd-current/external/apache2/llvm/dist/llvm/lib/Transforms/Utils/ |
H A D | InlineFunction.cpp | 547 CallInst *CI = dyn_cast<CallInst>(I); 1296 CallInst *NewAsmp = 1637 if (isa<CallInst>(Entry.first)) 1638 if (auto *CI = dyn_cast_or_null<CallInst>(Entry.second)) 1649 if (CallInst *CI = dyn_cast<CallInst>(&I)) 1714 } else if (auto *CI = dyn_cast<CallInst>(&*CurI)) { 1873 if (CallSiteEHPad && isa<CallInst>(CB)) { 2107 CallInst [all...] |
H A D | GuardUtils.cpp | 31 CallInst *Guard, bool UseWC) {
|
H A D | AddDiscriminators.cpp | 239 if (!isa<InvokeInst>(I) && (!isa<CallInst>(I) || isa<IntrinsicInst>(I)))
|
/netbsd-current/external/apache2/llvm/dist/llvm/include/llvm/CodeGen/GlobalISel/ |
H A D | IRTranslator.h | 38 class CallInst; 221 bool translateMemFunc(const CallInst &CI, MachineIRBuilder &MIRBuilder, 226 bool translateOverflowIntrinsic(const CallInst &CI, unsigned Op, 228 bool translateFixedPointIntrinsic(unsigned Op, const CallInst &CI, 239 bool translateSimpleIntrinsic(const CallInst &CI, Intrinsic::ID ID, 245 bool translateKnownIntrinsic(const CallInst &CI, Intrinsic::ID ID,
|
/netbsd-current/external/apache2/llvm/dist/llvm/lib/Target/SystemZ/ |
H A D | SystemZTDC.cpp | 224 if (CallInst *CI = dyn_cast<CallInst>(Op0)) { 267 } else if (auto *CI = dyn_cast<CallInst>(Op0)) {
|
/netbsd-current/external/apache2/llvm/dist/llvm/lib/Target/ARC/ |
H A D | ARCISelLowering.h | 114 bool mayBeEmittedAsTailCall(const CallInst *CI) const override;
|