/freebsd-11-stable/contrib/llvm-project/llvm/lib/Transforms/IPO/ |
H A D | FunctionAttrs.cpp | 48 #include "llvm/IR/Use.h" 364 bool captured(const Use *U) override { 382 std::distance(const_cast<const Use *>(CS.arg_begin()), U); 448 SmallVector<Use *, 32> Worklist; 449 SmallPtrSet<Use *, 32> Visited; 458 for (Use &U : A->uses()) { 464 Use *U = Worklist.pop_back_val(); 474 for (Use &UU : I->uses()) 488 for (Use &UU : I->uses()) 735 for (Argument *Use [all...] |
H A D | IPConstantPropagation.cpp | 60 for (Use &U : F.uses()) { 224 for (Use &U : F.uses()) {
|
/freebsd-11-stable/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
H A D | SIFoldOperands.cpp | 652 Use = MRI->use_begin(DestReg), E = MRI->use_end(); 653 Use != E; Use = NextUse) { 654 NextUse = std::next(Use); 655 FoldCandidate FC = FoldCandidate(Use->getParent(), 656 Use.getOperandNo(), &UseMI->getOperand(1)); 1152 Use = MRI->use_begin(Dst.getReg()), E = MRI->use_end(); 1153 Use != E; Use = NextUse) { 1154 NextUse = std::next(Use); [all...] |
H A D | GCNDPPCombine.cpp | 449 for (auto &Use : MRI->use_nodbg_operands(DPPMovReg)) { 450 Uses.push_back(&Use); 454 MachineOperand *Use = Uses.pop_back_val(); local 457 auto &OrigMI = *Use->getParent(); 510 if (Use == TII->getNamedOperand(OrigMI, AMDGPU::OpName::src0)) { 517 Use == TII->getNamedOperand(OrigMI, AMDGPU::OpName::src1)) {
|
H A D | AMDGPUAnnotateKernelFeatures.cpp | 33 #include "llvm/IR/Use.h" 125 for (const Use &U : C->operands()) { 314 for (const Use &U : I.operands()) {
|
/freebsd-11-stable/contrib/llvm-project/clang/lib/Sema/ |
H A D | AnalysisBasedWarnings.cpp | 810 static void DiagUninitUse(Sema &S, const VarDecl *VD, const UninitUse &Use, argument 814 switch (Use.getKind()) { 816 S.Diag(Use.getUser()->getBeginLoc(), diag::warn_uninit_var) 818 << Use.getUser()->getSourceRange(); 825 << (Use.getKind() == UninitUse::AfterDecl ? 4 : 5) 828 S.Diag(Use.getUser()->getBeginLoc(), diag::note_uninit_var_use) 829 << IsCapturedByBlock << Use.getUser()->getSourceRange(); 840 for (UninitUse::branch_iterator I = Use.branch_begin(), E = Use.branch_end(); 842 assert(Use 982 DiagnoseUninitializedUse(Sema &S, const VarDecl *VD, const UninitUse &Use, bool alwaysReportSelfInit = false) argument 1560 UninitUse Use = hasSelfInit ? UninitUse(U.getUser(), false) : U; local [all...] |
/freebsd-11-stable/contrib/llvm-project/llvm/lib/CodeGen/ |
H A D | DetectDeadLanes.cpp | 82 /// Given a use regiser operand \p Use and a mask of defined lanes, check 85 void transferDefinedLanesStep(const MachineOperand &Use, 274 void DetectDeadLanes::transferDefinedLanesStep(const MachineOperand &Use, argument 276 if (!Use.readsReg()) 280 const MachineInstr &MI = *Use.getParent(); 295 unsigned OpNum = MI.getOperandNo(&Use); 297 TRI->reverseComposeSubRegIndexLaneMask(Use.getSubReg(), DefinedLanes);
|
H A D | ModuloSchedule.cpp | 486 // Use the loop value defined in the kernel, unless the kernel 492 // Use the value defined by the Phi. We add one because we switch 497 // Use the loop value defined in the kernel. 501 // Use the value defined by the Phi, unless we're generating the first 1199 MachineInstr *Use = MRI.getVRegDef(LoopVal); local 1200 if (!Use || Use->isPHI()) 1202 int LoopCycle = Schedule.getCycle(Use); 1203 int LoopStage = Schedule.getStage(Use); 1361 // Producer was not inside the loop. Use th 1686 MachineInstr *Use = MRI.getUniqueVRegDef(MO.getReg()); local 1787 MachineInstr *Use = MRI.getUniqueVRegDef(Reg); local [all...] |
H A D | LiveInterval.cpp | 120 VNInfo *extendInBlock(SlotIndex StartIdx, SlotIndex Use) { argument 124 impl().findInsertPos(Segment(Use.getPrevSlot(), Use, nullptr)); 130 if (I->end < Use) 131 extendSegmentEndTo(I, Use); 136 SlotIndex StartIdx, SlotIndex Use) { 139 SlotIndex BeforeUse = Use.getPrevSlot(); 140 iterator I = impl().findInsertPos(Segment(BeforeUse, Use, nullptr)); 146 if (I->end < Use) { 149 extendSegmentEndTo(I, Use); 135 extendInBlock(ArrayRef<SlotIndex> Undefs, SlotIndex StartIdx, SlotIndex Use) argument [all...] |
/freebsd-11-stable/contrib/llvm-project/llvm/lib/IR/ |
H A D | Dominators.cpp | 231 bool DominatorTree::dominates(const BasicBlockEdge &BBE, const Use &U) const { 249 bool DominatorTree::dominates(const Instruction *Def, const Use &U) const { 300 bool DominatorTree::isReachableFromEntry(const Use &U) const {
|
H A D | TypeFinder.cpp | 23 #include "llvm/IR/Use.h" 53 for (const Use &U : FI.operands())
|
H A D | SafepointIRVerifier.cpp | 73 SetVector<const Use *> DeadEdges; // Contains all dead edges from live blocks. 77 static const Use& getEdge(const_pred_iterator &PredIt) { 105 bool isDeadEdge(const Use *U) const { 119 const Use &U = PU.getUser()->getOperandUse(PU.getOperandNo()); 185 void addDeadEdge(const Use &DeadEdge) { 878 errs() << "Use: " << I << "\n";
|
/freebsd-11-stable/cddl/contrib/opensolaris/cmd/dtrace/test/tst/common/funcs/ |
H A D | err.copyoutbadaddr.ksh | 24 * Use is subject to license terms.
|
H A D | err.copyoutstrbadaddr.ksh | 24 * Use is subject to license terms.
|
/freebsd-11-stable/contrib/xz/src/liblzma/api/lzma/ |
H A D | block.h | 16 # error Never include this file directly. Use <lzma.h> instead. 124 * - Use lzma_block_header_encode() to encode the Block Header.
|
H A D | filter.h | 16 # error Never include this file directly. Use <lzma.h> instead. 47 * Use constants whose name begin with `LZMA_FILTER_' to specify
|
/freebsd-11-stable/contrib/llvm-project/llvm/lib/Target/WebAssembly/ |
H A D | WebAssemblyAddMissingPrototypes.cpp | 92 for (Use &U : F.uses()) {
|
/freebsd-11-stable/contrib/llvm-project/llvm/lib/CodeGen/SelectionDAG/ |
H A D | ScheduleDAGSDNodes.h | 107 virtual void computeOperandLatency(SDNode *Def, SDNode *Use,
|
/freebsd-11-stable/contrib/llvm-project/llvm/lib/Analysis/ |
H A D | IVDescriptors.cpp | 44 for (User::op_iterator Use = I->op_begin(), E = I->op_end(); Use != E; ++Use) 45 if (!Set.count(dyn_cast<Instruction>(*Use))) 120 // Use the demanded bits analysis to determine the bits that are live out 600 for (User::op_iterator Use = I->op_begin(), E = I->op_end(); Use != E; 601 ++Use) { 602 if (Insts.count(dyn_cast<Instruction>(*Use))) 706 return all_of(I->uses(), [DT, DominatedBy](Use [all...] |
H A D | MemorySSA.cpp | 40 #include "llvm/IR/Use.h" 220 /// This does one-way checks to see if Use could theoretically be hoisted above 223 /// This assumes that, for the purposes of MemorySSA, Use comes directly after 226 static bool areLoadsReorderable(const LoadInst *Use, argument 228 bool VolatileUse = Use->isVolatile(); 244 bool SeqCstUse = Use->getOrdering() == AtomicOrdering::SequentiallyConsistent; 1751 bool Def, Use; 1754 Use = dyn_cast_or_null<MemoryUse>(Template) != nullptr; 1760 assert(Def == DefCheck && (Def || Use == UseCheck) && "Invalid template"); 1774 Use [all...] |
/freebsd-11-stable/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
H A D | BDCE.cpp | 112 for (Use &U : I.operands()) {
|
/freebsd-11-stable/contrib/bmake/unit-tests/ |
H A D | modts.mk | 12 # Use print or printf iff they are builtin.
|
/freebsd-11-stable/contrib/llvm-project/llvm/utils/TableGen/ |
H A D | InstrDocsEmitter.cpp | 209 for (Record *Use : II->ImplicitUses) { 212 OS << "``" << Use->getName() << "``";
|
/freebsd-11-stable/contrib/llvm-project/llvm/lib/Target/Hexagon/ |
H A D | RDFDeadCode.cpp | 211 if (KindA == NodeAttrs::Use && KindB == NodeAttrs::Def) 213 if (KindA == NodeAttrs::Def && KindB == NodeAttrs::Use)
|
/freebsd-11-stable/contrib/llvm-project/llvm/include/llvm/IR/ |
H A D | InstrTypes.h | 1001 ArrayRef<Use> Inputs; 1004 explicit OperandBundleUse(StringMapEntry<uint32_t> *Tag, ArrayRef<Use> Inputs) 1188 bool isDataOperand(const Use *U) const { 1205 unsigned getDataOperandNo(const Use *U) const { 1256 /// Wrappers for getting the \c Use of a call argument. 1257 const Use &getArgOperandUse(unsigned i) const { 1261 Use &getArgOperandUse(unsigned i) { 1266 bool isArgOperand(const Use *U) const { 1277 unsigned getArgOperandNo(const Use *U) const { 1300 const Use [all...] |