/freebsd-13-stable/contrib/llvm-project/llvm/lib/Analysis/ |
H A D | InlineAdvisor.cpp | 175 /// inlined which is proved to be more beneficial. \p IC is the 180 shouldBeDeferred(Function *Caller, InlineCost IC, int &TotalSecondaryCost, argument 188 if (IC.getCost() <= 0) 208 int CandidateCost = IC.getCost() - 1; 256 // inlining -- IC.getCost() multiplied by the number of callers to Caller. 258 return TotalSecondaryCost < IC.getCost(); 260 int TotalCost = TotalSecondaryCost + IC.getCost() * NumCallerUsers; 261 int Allowance = IC.getCost() * InlineDeferralScale; 272 RemarkT &operator<<(RemarkT &&R, const InlineCost &IC) { argument 274 if (IC 288 inlineCostStr(const InlineCost &IC) argument 290 Remark << IC; local 391 emitInlinedInto(OptimizationRemarkEmitter &ORE, DebugLoc DLoc, const BasicBlock *Block, const Function &Callee, const Function &Caller, const InlineCost &IC, bool ForProfileContext, const char *PassName) argument [all...] |
/freebsd-13-stable/contrib/llvm-project/llvm/lib/Transforms/InstCombine/ |
H A D | InstCombineLoadStoreAlloca.cpp | 169 static Instruction *simplifyAllocaArraySize(InstCombiner &IC, AllocaInst &AI) { argument 177 return IC.replaceOperand(AI, 0, IC.Builder.getInt32(1)); 184 AllocaInst *New = IC.Builder.CreateAlloca(NewTy, nullptr, AI.getName()); 197 Type *IdxTy = IC.getDataLayout().getIntPtrType(AI.getType()); 202 IC.InsertNewInstBefore(GEP, *It); 206 return IC.replaceInstUsesWith(AI, GEP); 211 return IC.replaceInstUsesWith(AI, Constant::getNullValue(AI.getType())); 215 Type *IntPtrTy = IC.getDataLayout().getIntPtrType(AI.getType()); 217 Value *V = IC 237 PointerReplacer(InstCombiner &IC) argument 247 InstCombiner &IC; member in class:__anon4564::PointerReplacer 448 combineStoreToNewValue(InstCombiner &IC, StoreInst &SI, Value *V) argument 540 combineLoadToOperationType(InstCombiner &IC, LoadInst &LI) argument 608 unpackLoadToAggregate(InstCombiner &IC, LoadInst &LI) argument 807 canReplaceGEPIdxWithZero(InstCombiner &IC, GetElementPtrInst *GEPI, Instruction *MemI, unsigned &Idx) argument 877 replaceGEPIdxWithZero(InstCombiner &IC, Value *Ptr, T &MemI) argument 1036 likeBitCastFromVector(InstCombiner &IC, Value *V) argument 1097 combineStoreToValueType(InstCombiner &IC, StoreInst &SI) argument 1129 unpackStoreToAggregate(InstCombiner &IC, StoreInst &SI) argument 1269 removeBitcastsFromLoadStoreOnMinMax(InstCombiner &IC, StoreInst &SI) argument [all...] |
H A D | InstCombineNegator.cpp | 433 InstCombiner &IC) { 441 Negator N(Root->getContext(), IC.getDataLayout(), IC.getAssumptionCache(), 442 IC.getDominatorTree(), LHSIsZero); 457 InstCombiner::BuilderTy::InsertPointGuard Guard(IC.Builder); 458 IC.Builder.ClearInsertionPoint(); 459 IC.Builder.SetCurrentDebugLocation(DebugLoc()); 470 [&](Instruction *I) { IC.Builder.Insert(I, I->getName()); }); 432 Negate(bool LHSIsZero, Value *Root, InstCombiner &IC) argument
|
H A D | InstCombineCasts.cpp | 345 static bool canEvaluateTruncated(Value *V, Type *Ty, InstCombiner &IC, argument 362 return canEvaluateTruncated(I->getOperand(0), Ty, IC, CxtI) && 363 canEvaluateTruncated(I->getOperand(1), Ty, IC, CxtI); 372 if (IC.MaskedValueIsZero(I->getOperand(0), Mask, 0, CxtI) && 373 IC.MaskedValueIsZero(I->getOperand(1), Mask, 0, CxtI)) { 374 return canEvaluateTruncated(I->getOperand(0), Ty, IC, CxtI) && 375 canEvaluateTruncated(I->getOperand(1), Ty, IC, CxtI); 384 llvm::computeKnownBits(I->getOperand(1), IC.getDataLayout()); 386 return canEvaluateTruncated(I->getOperand(0), Ty, IC, CxtI) && 387 canEvaluateTruncated(I->getOperand(1), Ty, IC, Cxt 462 foldVecTruncToExtElt(TruncInst &Trunc, InstCombiner &IC) argument 1033 canEvaluateZExtd(Value *V, Type *Ty, unsigned &BitsToClear, InstCombiner &IC, Instruction *CxtI) argument 1966 optimizeVectorResizeWithIntegerBitCasts(Value *InVal, VectorType *DestTy, InstCombiner &IC) argument 2167 optimizeIntegerToVectorInsertions(BitCastInst &CI, InstCombiner &IC) argument 2196 canonicalizeBitCastExtElt(BitCastInst &BitCast, InstCombiner &IC) argument [all...] |
H A D | InstCombineShifts.cpp | 423 Instruction *InnerShift, InstCombiner &IC, 457 if (IC.MaskedValueIsZero(InnerShift->getOperand(0), Mask, 0, CxtI)) 475 InstCombiner &IC, Instruction *CxtI) { 518 return canEvaluateShifted(I->getOperand(0), NumBits, IsLeftShift, IC, I) && 519 canEvaluateShifted(I->getOperand(1), NumBits, IsLeftShift, IC, I); 523 return canEvaluateShiftedShift(NumBits, IsLeftShift, I, IC, CxtI); 529 return canEvaluateShifted(TrueVal, NumBits, IsLeftShift, IC, SI) && 530 canEvaluateShifted(FalseVal, NumBits, IsLeftShift, IC, SI); 538 if (!canEvaluateShifted(IncValue, NumBits, IsLeftShift, IC, PN)) 611 InstCombiner &IC, cons 422 canEvaluateShiftedShift(unsigned OuterShAmt, bool IsOuterShl, Instruction *InnerShift, InstCombiner &IC, Instruction *CxtI) argument 474 canEvaluateShifted(Value *V, unsigned NumBits, bool IsLeftShift, InstCombiner &IC, Instruction *CxtI) argument 610 getShiftedValue(Value *V, unsigned NumBits, bool isLeftShift, InstCombiner &IC, const DataLayout &DL) argument [all...] |
H A D | InstCombineSelect.cpp | 60 InstCombiner &IC) { 111 return IC.replaceOperand(Sel, IsEq ? 1 : 2, Y); 536 static Value *foldSelectICmpLshrAshr(const ICmpInst *IC, Value *TrueVal, argument 539 ICmpInst::Predicate Pred = IC->getPredicate(); 540 Value *CmpLHS = IC->getOperand(0); 541 Value *CmpRHS = IC->getOperand(1); 565 return Builder.CreateAShr(X, Y, IC->getName(), IsExact); 584 static Value *foldSelectICmpAndOr(const ICmpInst *IC, Value *TrueVal, argument 590 TrueVal->getType()->isVectorTy() != IC->getType()->isVectorTy()) 593 Value *CmpLHS = IC 58 foldSelectBinOpIdentity(SelectInst &Sel, const TargetLibraryInfo &TLI, InstCombiner &IC) argument 1030 canonicalizeMinMaxWithConstant(SelectInst &Sel, ICmpInst &Cmp, InstCombiner &IC) argument 1074 canonicalizeAbsNabs(SelectInst &Sel, ICmpInst &Cmp, InstCombiner &IC) argument 1373 tryToReuseConstantFromSelectInComparison(SelectInst &Sel, ICmpInst &Cmp, InstCombiner &IC) argument 2015 canonicalizeScalarSelectOfVecs( SelectInst &Sel, InstCombiner &IC) argument [all...] |
H A D | InstCombineCalls.cpp | 1209 InstCombiner &IC) { 1219 Result = IC.Builder.CreateLaunderInvariantGroup(StrippedInvariantGroupsArg); 1221 Result = IC.Builder.CreateStripInvariantGroup(StrippedInvariantGroupsArg); 1227 Result = IC.Builder.CreateAddrSpaceCast(Result, II.getType()); 1229 Result = IC.Builder.CreateBitCast(Result, II.getType()); 1234 static Instruction *foldCttzCtlz(IntrinsicInst &II, InstCombiner &IC) { argument 1252 return IC.replaceOperand(II, 0, X); 1259 return IC.replaceOperand(II, 0, X); 1262 KnownBits Known = IC.computeKnownBits(Op0, 0, &II); 1276 return IC 1208 simplifyInvariantGroupIntrinsic(IntrinsicInst &II, InstCombiner &IC) argument 1304 foldCtpop(IntrinsicInst &II, InstCombiner &IC) argument 1362 simplifyX86MaskedLoad(IntrinsicInst &II, InstCombiner &IC) argument 1398 simplifyX86MaskedStore(IntrinsicInst &II, InstCombiner &IC) argument 1541 removeTriviallyEmptyRange( IntrinsicInst &EndI, InstCombiner &IC, std::function<bool(const IntrinsicInst &)> IsStart) argument 1570 SimplifyNVVMIntrinsic(IntrinsicInst *II, InstCombiner &IC) argument [all...] |
H A D | InstCombineMulDivRem.cpp | 49 static Value *simplifyValueKnownNonZero(Value *V, InstCombiner &IC, argument 63 A = IC.Builder.CreateSub(A, B); 64 return IC.Builder.CreateShl(One, A); 71 IC.isKnownToBeAPowerOfTwo(I->getOperand(0), false, 0, &CxtI)) { 74 if (Value *V2 = simplifyValueKnownNonZero(I->getOperand(0), IC, CxtI)) { 75 IC.replaceOperand(*I, 0, V2); 881 InstCombiner &IC); 910 const BinaryOperator &I, InstCombiner &IC) { 923 InstCombiner &IC) { 935 N = IC 909 foldUDivPow2Cst(Value *Op0, Value *Op1, const BinaryOperator &I, InstCombiner &IC) argument 922 foldUDivShl(Value *Op0, Value *Op1, const BinaryOperator &I, InstCombiner &IC) argument [all...] |
/freebsd-13-stable/contrib/llvm-project/llvm/utils/TableGen/ |
H A D | X86DisassemblerShared.h | 24 insnContext = llvm::X86Disassembler::IC;
|
H A D | CodeEmitterGen.cpp | 464 for (std::vector<Record*>::iterator IC = Insts.begin(), EC = Insts.end(); 465 IC != EC; ++IC) { 466 Record *R = *IC;
|
/freebsd-13-stable/contrib/llvm-project/clang/lib/StaticAnalyzer/Checkers/ |
H A D | SmartPtrModeling.cpp | 163 const auto *IC = dyn_cast<CXXInstanceCall>(&Call); local 164 if (!IC) 167 const MemRegion *ThisValRegion = IC->getCXXThisVal().getAsRegion(); 178 const auto *IC = dyn_cast<CXXInstanceCall>(&Call); local 179 if (!IC) 182 const MemRegion *ThisValRegion = IC->getCXXThisVal().getAsRegion();
|
H A D | MoveChecker.cpp | 473 if (const auto *IC = dyn_cast<CXXInstanceCall>(AFC)) 474 if (IC->getCXXThisVal().getAsRegion() == ArgRegion) 623 const auto IC = dyn_cast<CXXInstanceCall>(&Call); local 624 if (!IC) 628 if (isa<CXXDestructorCall>(IC)) 631 const MemRegion *ThisRegion = IC->getCXXThisVal().getAsRegion(); 636 const auto MethodDecl = dyn_cast_or_null<CXXMethodDecl>(IC->getDecl()); 666 const MemRegion *ArgRegion = IC->getArgSVal(0).getAsRegion(); 713 if (const auto *IC = dyn_cast<CXXInstanceCall>(Call)) 714 ThisRegion = IC [all...] |
/freebsd-13-stable/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPUInline.cpp | 215 auto IC = llvm::getInlineCost(CB, Callee, LocalParams, TTI, local 219 if (IC && !IC.isAlways() && !Callee->hasFnAttribute(Attribute::InlineHint)) { 225 return IC;
|
/freebsd-13-stable/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
H A D | LoopSink.cpp | 230 Instruction *IC = I.clone(); local 231 IC->setName(I.getName()); 232 IC->insertBefore(&*N->getFirstInsertionPt()); 233 // Replaces uses of I with IC in N 234 I.replaceUsesWithIf(IC, [N](Use &U) { 237 // Replaces uses of I with IC in blocks dominated by N 238 replaceDominatedUsesWith(&I, IC, DT, N);
|
H A D | GuardWidening.cpp | 595 auto *IC = dyn_cast<ICmpInst>(CheckCond); 596 if (!IC || !IC->getOperand(0)->getType()->isIntegerTy() || 597 (IC->getPredicate() != ICmpInst::ICMP_ULT && 598 IC->getPredicate() != ICmpInst::ICMP_UGT)) 601 const Value *CmpLHS = IC->getOperand(0), *CmpRHS = IC->getOperand(1); 602 if (IC->getPredicate() == ICmpInst::ICMP_UGT) 605 auto &DL = IC->getModule()->getDataLayout(); 609 CmpRHS, IC); [all...] |
/freebsd-13-stable/contrib/llvm-project/llvm/include/llvm/Analysis/ |
H A D | InlineAdvisor.h | 225 const Function &Caller, const InlineCost &IC, 236 std::string inlineCostStr(const InlineCost &IC);
|
/freebsd-13-stable/sys/mips/include/ |
H A D | regnum.h | 107 * IC is valid only on RM7K and RM9K processors. Access to this is 110 #define IC 38 macro
|
/freebsd-13-stable/contrib/llvm-project/clang/lib/StaticAnalyzer/Checkers/UninitializedObject/ |
H A D | UninitializedPointee.cpp | 120 DereferenceInfo(const TypedValueRegion *R, bool NCB, bool IC) argument 121 : R(R), NeedsCastBack(NCB), IsCyclic(IC) {}
|
/freebsd-13-stable/contrib/llvm-project/llvm/lib/Transforms/IPO/ |
H A D | PartialInlining.cpp | 777 InlineCost IC = local 781 if (IC.isAlways()) { 790 if (IC.isNever()) { 800 if (!IC) { 805 << NV("Cost", IC.getCost()) << ", threshold=" 806 << NV("Threshold", IC.getCostDelta() + IC.getCost()) << ")"; 836 << NV("Caller", Caller) << " with cost=" << NV("Cost", IC.getCost()) 838 << NV("Threshold", IC.getCostDelta() + IC [all...] |
/freebsd-13-stable/contrib/llvm-project/llvm/lib/Target/X86/AsmParser/ |
H A D | X86AsmParser.cpp | 356 InfixCalculator IC; member in class:__anon4389::X86AsmParser::IntelExprStateMachine 391 int64_t getImm() { return Imm + IC.execute(); } 408 IC.pushOperator(IC_OR); 423 IC.pushOperator(IC_XOR); 438 IC.pushOperator(IC_AND); 453 IC.pushOperator(IC_LSHIFT); 468 IC.pushOperator(IC_RSHIFT); 484 IC.pushOperator(IC_PLUS); 533 IC.pushOperator(IC_MINUS); 539 IC [all...] |
/freebsd-13-stable/libexec/getty/ |
H A D | main.c | 238 if (IC || AC) { 245 if (IC) { 246 if (getty_chat(IC, CT, DC) > 0) {
|
H A D | gettytab.h | 89 #define IC gettystrs[27].value macro
|
/freebsd-13-stable/contrib/llvm-project/clang/lib/AST/ |
H A D | CommentParser.cpp | 413 InlineCommandComment *IC; local 415 IC = S.actOnInlineCommand(CommandTok.getLocation(), 422 IC = S.actOnInlineCommand(CommandTok.getLocation(), 435 return IC;
|
/freebsd-13-stable/contrib/llvm-project/clang/lib/Sema/ |
H A D | SemaObjCProperty.cpp | 1108 ObjCImplementationDecl *IC = nullptr; local 1110 if ((IC = dyn_cast<ObjCImplementationDecl>(ClassImpDecl))) { 1111 IDecl = IC->getClassInterface(); 1133 Diag(IC->getLocation(), diag::warn_auto_implicit_atomic_property); 1451 OMD = RedeclarePropertyAccessor(Context, IC, getterMethod, AtLoc, 1517 OMD = RedeclarePropertyAccessor(Context, IC, setterMethod, 1568 if (IC) { 1571 IC->FindPropertyImplIvarDecl(PropertyIvar)) { 1579 = IC->FindPropertyImplDecl(PropertyId, QueryKind)) { 1584 IC [all...] |
/freebsd-13-stable/contrib/llvm-project/llvm/lib/Transforms/Vectorize/ |
H A D | LoopVectorize.cpp | 5289 // this currently happens when OptForSize, in which case IC is set to 1 above. 5290 unsigned IC = UINT_MAX; 5317 IC = std::min(IC, TmpIC); 5345 // Clamp the calculated IC to be between the 1 and the max interleave count 5347 if (IC > MaxInterleaveCount) 5348 IC = MaxInterleaveCount; 5349 else if (IC < 1) 5350 IC = 1; 5356 return IC; [all...] |