/macosx-10.9.5/llvmCore-3425.0.33/lib/VMCore/ |
H A D | Instruction.cpp | 220 if (const AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(this)) 221 return CXI->isVolatile() == cast<AtomicCmpXchgInst>(I)->isVolatile() && 222 CXI->getOrdering() == cast<AtomicCmpXchgInst>(I)->getOrdering() && 223 CXI->getSynchScope() == cast<AtomicCmpXchgInst>(I)->getSynchScope(); 294 if (const AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(this)) 295 return CXI->isVolatile() == cast<AtomicCmpXchgInst>(I)->isVolatile() && 296 CXI->getOrdering() == cast<AtomicCmpXchgInst>(I)->getOrdering() && 297 CXI->getSynchScope() == cast<AtomicCmpXchgInst>( [all...] |
H A D | Instructions.cpp | 1192 // AtomicCmpXchgInst Implementation 1195 void AtomicCmpXchgInst::Init(Value *Ptr, Value *Cmp, Value *NewVal, 1218 AtomicCmpXchgInst::AtomicCmpXchgInst(Value *Ptr, Value *Cmp, Value *NewVal, function in class:AtomicCmpXchgInst 1223 OperandTraits<AtomicCmpXchgInst>::op_begin(this), 1224 OperandTraits<AtomicCmpXchgInst>::operands(this), 1229 AtomicCmpXchgInst::AtomicCmpXchgInst(Value *Ptr, Value *Cmp, Value *NewVal, function in class:AtomicCmpXchgInst 1234 OperandTraits<AtomicCmpXchgInst>::op_begin(this), 1235 OperandTraits<AtomicCmpXchgInst> [all...] |
H A D | AsmWriter.cpp | 1742 (isa<AtomicCmpXchgInst>(I) && cast<AtomicCmpXchgInst>(I).isVolatile()) || 1996 } else if (const AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(&I)) {
|
H A D | Verifier.cpp | 286 void visitAtomicCmpXchgInst(AtomicCmpXchgInst &CXI); 1485 void Verifier::visitAtomicCmpXchgInst(AtomicCmpXchgInst &CXI) {
|
/macosx-10.9.5/llvmCore-3425.0.33/lib/Transforms/Scalar/ |
H A D | LowerAtomic.cpp | 23 static bool LowerAtomicCmpXchgInst(AtomicCmpXchgInst *CXI) { 120 else if (AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(Inst))
|
H A D | SCCP.cpp | 504 void visitAtomicCmpXchgInst (AtomicCmpXchgInst &I) { markOverdefined(&I); }
|
/macosx-10.9.5/llvmCore-3425.0.33/lib/Transforms/Instrumentation/ |
H A D | BoundsChecking.cpp | 185 if (isa<LoadInst>(I) || isa<StoreInst>(I) || isa<AtomicCmpXchgInst>(I) || 200 } else if (AtomicCmpXchgInst *AI = dyn_cast<AtomicCmpXchgInst>(Inst)) {
|
H A D | ThreadSanitizer.cpp | 245 if (isa<AtomicCmpXchgInst>(I)) 390 } else if (isa<AtomicCmpXchgInst>(I)) {
|
H A D | AddressSanitizer.cpp | 362 if (AtomicCmpXchgInst *XCHG = dyn_cast<AtomicCmpXchgInst>(I)) {
|
/macosx-10.9.5/llvmCore-3425.0.33/include/llvm/Analysis/ |
H A D | AliasAnalysis.h | 147 Location getLocation(const AtomicCmpXchgInst *CXI); 361 return getModRefInfo((const AtomicCmpXchgInst*)I, Loc); 444 ModRefResult getModRefInfo(const AtomicCmpXchgInst *CX, const Location &Loc); 447 ModRefResult getModRefInfo(const AtomicCmpXchgInst *CX,
|
/macosx-10.9.5/llvmCore-3425.0.33/lib/Transforms/IPO/ |
H A D | MergeFunctions.cpp | 331 if (const AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(I1)) 332 return CXI->isVolatile() == cast<AtomicCmpXchgInst>(I2)->isVolatile() && 333 CXI->getOrdering() == cast<AtomicCmpXchgInst>(I2)->getOrdering() && 334 CXI->getSynchScope() == cast<AtomicCmpXchgInst>(I2)->getSynchScope();
|
/macosx-10.9.5/llvmCore-3425.0.33/lib/Analysis/ |
H A D | AliasAnalysis.cpp | 245 AliasAnalysis::getLocation(const AtomicCmpXchgInst *CXI) { 338 AliasAnalysis::getModRefInfo(const AtomicCmpXchgInst *CX, const Location &Loc) {
|
/macosx-10.9.5/llvmCore-3425.0.33/include/llvm/Support/ |
H A D | InstVisitor.h | 177 RetTy visitAtomicCmpXchgInst(AtomicCmpXchgInst &I) { DELEGATE(Instruction);}
|
/macosx-10.9.5/llvmCore-3425.0.33/include/llvm/ |
H A D | Instructions.h | 445 // AtomicCmpXchgInst Class 448 /// AtomicCmpXchgInst - an instruction that atomically checks whether a 452 class AtomicCmpXchgInst : public Instruction { class in namespace:llvm 457 virtual AtomicCmpXchgInst *clone_impl() const; 463 AtomicCmpXchgInst(Value *Ptr, Value *Cmp, Value *NewVal, 466 AtomicCmpXchgInst(Value *Ptr, Value *Cmp, Value *NewVal, 529 static inline bool classof(const AtomicCmpXchgInst *) { return true; } 545 struct OperandTraits<AtomicCmpXchgInst> : 546 public FixedNumOperandTraits<AtomicCmpXchgInst, 3> { 549 DEFINE_TRANSPARENT_OPERAND_ACCESSORS(AtomicCmpXchgInst, Valu [all...] |
H A D | IRBuilder.h | 845 AtomicCmpXchgInst *CreateAtomicCmpXchg(Value *Ptr, Value *Cmp, Value *New, 848 return Insert(new AtomicCmpXchgInst(Ptr, Cmp, New, Ordering, SynchScope));
|
/macosx-10.9.5/llvmCore-3425.0.33/lib/CodeGen/SelectionDAG/ |
H A D | SelectionDAGBuilder.h | 522 void visitAtomicCmpXchg(const AtomicCmpXchgInst &I);
|
H A D | SelectionDAGBuilder.cpp | 3417 void SelectionDAGBuilder::visitAtomicCmpXchg(const AtomicCmpXchgInst &I) {
|
/macosx-10.9.5/llvmCore-3425.0.33/lib/Bitcode/Writer/ |
H A D | BitcodeWriter.cpp | 1327 Vals.push_back(cast<AtomicCmpXchgInst>(I).isVolatile()); 1329 cast<AtomicCmpXchgInst>(I).getOrdering())); 1331 cast<AtomicCmpXchgInst>(I).getSynchScope()));
|
/macosx-10.9.5/llvmCore-3425.0.33/lib/Target/CppBackend/ |
H A D | CPPBackend.cpp | 1538 const AtomicCmpXchgInst *cxi = cast<AtomicCmpXchgInst>(I); 1541 Out << "AtomicCmpXchgInst* " << iName 1542 << " = new AtomicCmpXchgInst("
|
/macosx-10.9.5/llvmCore-3425.0.33/lib/AsmParser/ |
H A D | LLParser.cpp | 3890 AtomicCmpXchgInst *CXI = 3891 new AtomicCmpXchgInst(Ptr, Cmp, New, Ordering, Scope);
|
/macosx-10.9.5/llvmCore-3425.0.33/lib/Bitcode/Reader/ |
H A D | BitcodeReader.cpp | 2592 I = new AtomicCmpXchgInst(Ptr, Cmp, New, Ordering, SynchScope); 2593 cast<AtomicCmpXchgInst>(I)->setVolatile(Record[OpNum]);
|
/macosx-10.9.5/llvmCore-3425.0.33/lib/Transforms/Utils/ |
H A D | SimplifyCFG.cpp | 2864 } else if (AtomicCmpXchgInst *CXI = dyn_cast<AtomicCmpXchgInst>(BBI)) {
|