1//===- VPlan.h - Represent A Vectorizer Plan --------------------*- C++ -*-===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9/// \file
10/// This file contains the declarations of the Vectorization Plan base classes:
11/// 1. VPBasicBlock and VPRegionBlock that inherit from a common pure virtual
12///    VPBlockBase, together implementing a Hierarchical CFG;
13/// 2. Specializations of GraphTraits that allow VPBlockBase graphs to be
14///    treated as proper graphs for generic algorithms;
15/// 3. Pure virtual VPRecipeBase serving as the base class for recipes contained
16///    within VPBasicBlocks;
17/// 4. VPInstruction, a concrete Recipe and VPUser modeling a single planned
18///    instruction;
19/// 5. The VPlan class holding a candidate for vectorization;
20/// 6. The VPlanPrinter class providing a way to print a plan in dot format;
21/// These are documented in docs/VectorizationPlan.rst.
22//
23//===----------------------------------------------------------------------===//
24
25#ifndef LLVM_TRANSFORMS_VECTORIZE_VPLAN_H
26#define LLVM_TRANSFORMS_VECTORIZE_VPLAN_H
27
28#include "VPlanLoopInfo.h"
29#include "VPlanValue.h"
30#include "llvm/ADT/DenseMap.h"
31#include "llvm/ADT/DepthFirstIterator.h"
32#include "llvm/ADT/GraphTraits.h"
33#include "llvm/ADT/Optional.h"
34#include "llvm/ADT/SmallBitVector.h"
35#include "llvm/ADT/SmallPtrSet.h"
36#include "llvm/ADT/SmallSet.h"
37#include "llvm/ADT/SmallVector.h"
38#include "llvm/ADT/Twine.h"
39#include "llvm/ADT/ilist.h"
40#include "llvm/ADT/ilist_node.h"
41#include "llvm/Analysis/VectorUtils.h"
42#include "llvm/IR/IRBuilder.h"
43#include "llvm/Support/InstructionCost.h"
44#include <algorithm>
45#include <cassert>
46#include <cstddef>
47#include <map>
48#include <string>
49
50namespace llvm {
51
52class BasicBlock;
53class DominatorTree;
54class InnerLoopVectorizer;
55class LoopInfo;
56class raw_ostream;
57class RecurrenceDescriptor;
58class Value;
59class VPBasicBlock;
60class VPRegionBlock;
61class VPlan;
62class VPlanSlp;
63
64/// Returns a calculation for the total number of elements for a given \p VF.
65/// For fixed width vectors this value is a constant, whereas for scalable
66/// vectors it is an expression determined at runtime.
67Value *getRuntimeVF(IRBuilder<> &B, Type *Ty, ElementCount VF);
68
69/// A range of powers-of-2 vectorization factors with fixed start and
70/// adjustable end. The range includes start and excludes end, e.g.,:
71/// [1, 9) = {1, 2, 4, 8}
72struct VFRange {
73  // A power of 2.
74  const ElementCount Start;
75
76  // Need not be a power of 2. If End <= Start range is empty.
77  ElementCount End;
78
79  bool isEmpty() const {
80    return End.getKnownMinValue() <= Start.getKnownMinValue();
81  }
82
83  VFRange(const ElementCount &Start, const ElementCount &End)
84      : Start(Start), End(End) {
85    assert(Start.isScalable() == End.isScalable() &&
86           "Both Start and End should have the same scalable flag");
87    assert(isPowerOf2_32(Start.getKnownMinValue()) &&
88           "Expected Start to be a power of 2");
89  }
90};
91
92using VPlanPtr = std::unique_ptr<VPlan>;
93
94/// In what follows, the term "input IR" refers to code that is fed into the
95/// vectorizer whereas the term "output IR" refers to code that is generated by
96/// the vectorizer.
97
98/// VPLane provides a way to access lanes in both fixed width and scalable
99/// vectors, where for the latter the lane index sometimes needs calculating
100/// as a runtime expression.
101class VPLane {
102public:
103  /// Kind describes how to interpret Lane.
104  enum class Kind : uint8_t {
105    /// For First, Lane is the index into the first N elements of a
106    /// fixed-vector <N x <ElTy>> or a scalable vector <vscale x N x <ElTy>>.
107    First,
108    /// For ScalableLast, Lane is the offset from the start of the last
109    /// N-element subvector in a scalable vector <vscale x N x <ElTy>>. For
110    /// example, a Lane of 0 corresponds to lane `(vscale - 1) * N`, a Lane of
111    /// 1 corresponds to `((vscale - 1) * N) + 1`, etc.
112    ScalableLast
113  };
114
115private:
116  /// in [0..VF)
117  unsigned Lane;
118
119  /// Indicates how the Lane should be interpreted, as described above.
120  Kind LaneKind;
121
122public:
123  VPLane(unsigned Lane, Kind LaneKind) : Lane(Lane), LaneKind(LaneKind) {}
124
125  static VPLane getFirstLane() { return VPLane(0, VPLane::Kind::First); }
126
127  static VPLane getLastLaneForVF(const ElementCount &VF) {
128    unsigned LaneOffset = VF.getKnownMinValue() - 1;
129    Kind LaneKind;
130    if (VF.isScalable())
131      // In this case 'LaneOffset' refers to the offset from the start of the
132      // last subvector with VF.getKnownMinValue() elements.
133      LaneKind = VPLane::Kind::ScalableLast;
134    else
135      LaneKind = VPLane::Kind::First;
136    return VPLane(LaneOffset, LaneKind);
137  }
138
139  /// Returns a compile-time known value for the lane index and asserts if the
140  /// lane can only be calculated at runtime.
141  unsigned getKnownLane() const {
142    assert(LaneKind == Kind::First);
143    return Lane;
144  }
145
146  /// Returns an expression describing the lane index that can be used at
147  /// runtime.
148  Value *getAsRuntimeExpr(IRBuilder<> &Builder, const ElementCount &VF) const;
149
150  /// Returns the Kind of lane offset.
151  Kind getKind() const { return LaneKind; }
152
153  /// Returns true if this is the first lane of the whole vector.
154  bool isFirstLane() const { return Lane == 0 && LaneKind == Kind::First; }
155
156  /// Maps the lane to a cache index based on \p VF.
157  unsigned mapToCacheIndex(const ElementCount &VF) const {
158    switch (LaneKind) {
159    case VPLane::Kind::ScalableLast:
160      assert(VF.isScalable() && Lane < VF.getKnownMinValue());
161      return VF.getKnownMinValue() + Lane;
162    default:
163      assert(Lane < VF.getKnownMinValue());
164      return Lane;
165    }
166  }
167
168  /// Returns the maxmimum number of lanes that we are able to consider
169  /// caching for \p VF.
170  static unsigned getNumCachedLanes(const ElementCount &VF) {
171    return VF.getKnownMinValue() * (VF.isScalable() ? 2 : 1);
172  }
173};
174
175/// VPIteration represents a single point in the iteration space of the output
176/// (vectorized and/or unrolled) IR loop.
177struct VPIteration {
178  /// in [0..UF)
179  unsigned Part;
180
181  VPLane Lane;
182
183  VPIteration(unsigned Part, unsigned Lane,
184              VPLane::Kind Kind = VPLane::Kind::First)
185      : Part(Part), Lane(Lane, Kind) {}
186
187  VPIteration(unsigned Part, const VPLane &Lane) : Part(Part), Lane(Lane) {}
188
189  bool isFirstIteration() const { return Part == 0 && Lane.isFirstLane(); }
190};
191
192/// VPTransformState holds information passed down when "executing" a VPlan,
193/// needed for generating the output IR.
194struct VPTransformState {
195  VPTransformState(ElementCount VF, unsigned UF, LoopInfo *LI,
196                   DominatorTree *DT, IRBuilder<> &Builder,
197                   InnerLoopVectorizer *ILV, VPlan *Plan)
198      : VF(VF), UF(UF), Instance(), LI(LI), DT(DT), Builder(Builder), ILV(ILV),
199        Plan(Plan) {}
200
201  /// The chosen Vectorization and Unroll Factors of the loop being vectorized.
202  ElementCount VF;
203  unsigned UF;
204
205  /// Hold the indices to generate specific scalar instructions. Null indicates
206  /// that all instances are to be generated, using either scalar or vector
207  /// instructions.
208  Optional<VPIteration> Instance;
209
210  struct DataState {
211    /// A type for vectorized values in the new loop. Each value from the
212    /// original loop, when vectorized, is represented by UF vector values in
213    /// the new unrolled loop, where UF is the unroll factor.
214    typedef SmallVector<Value *, 2> PerPartValuesTy;
215
216    DenseMap<VPValue *, PerPartValuesTy> PerPartOutput;
217
218    using ScalarsPerPartValuesTy = SmallVector<SmallVector<Value *, 4>, 2>;
219    DenseMap<VPValue *, ScalarsPerPartValuesTy> PerPartScalars;
220  } Data;
221
222  /// Get the generated Value for a given VPValue and a given Part. Note that
223  /// as some Defs are still created by ILV and managed in its ValueMap, this
224  /// method will delegate the call to ILV in such cases in order to provide
225  /// callers a consistent API.
226  /// \see set.
227  Value *get(VPValue *Def, unsigned Part);
228
229  /// Get the generated Value for a given VPValue and given Part and Lane.
230  Value *get(VPValue *Def, const VPIteration &Instance);
231
232  bool hasVectorValue(VPValue *Def, unsigned Part) {
233    auto I = Data.PerPartOutput.find(Def);
234    return I != Data.PerPartOutput.end() && Part < I->second.size() &&
235           I->second[Part];
236  }
237
238  bool hasAnyVectorValue(VPValue *Def) const {
239    return Data.PerPartOutput.find(Def) != Data.PerPartOutput.end();
240  }
241
242  bool hasScalarValue(VPValue *Def, VPIteration Instance) {
243    auto I = Data.PerPartScalars.find(Def);
244    if (I == Data.PerPartScalars.end())
245      return false;
246    unsigned CacheIdx = Instance.Lane.mapToCacheIndex(VF);
247    return Instance.Part < I->second.size() &&
248           CacheIdx < I->second[Instance.Part].size() &&
249           I->second[Instance.Part][CacheIdx];
250  }
251
252  /// Set the generated Value for a given VPValue and a given Part.
253  void set(VPValue *Def, Value *V, unsigned Part) {
254    if (!Data.PerPartOutput.count(Def)) {
255      DataState::PerPartValuesTy Entry(UF);
256      Data.PerPartOutput[Def] = Entry;
257    }
258    Data.PerPartOutput[Def][Part] = V;
259  }
260  /// Reset an existing vector value for \p Def and a given \p Part.
261  void reset(VPValue *Def, Value *V, unsigned Part) {
262    auto Iter = Data.PerPartOutput.find(Def);
263    assert(Iter != Data.PerPartOutput.end() &&
264           "need to overwrite existing value");
265    Iter->second[Part] = V;
266  }
267
268  /// Set the generated scalar \p V for \p Def and the given \p Instance.
269  void set(VPValue *Def, Value *V, const VPIteration &Instance) {
270    auto Iter = Data.PerPartScalars.insert({Def, {}});
271    auto &PerPartVec = Iter.first->second;
272    while (PerPartVec.size() <= Instance.Part)
273      PerPartVec.emplace_back();
274    auto &Scalars = PerPartVec[Instance.Part];
275    unsigned CacheIdx = Instance.Lane.mapToCacheIndex(VF);
276    while (Scalars.size() <= CacheIdx)
277      Scalars.push_back(nullptr);
278    assert(!Scalars[CacheIdx] && "should overwrite existing value");
279    Scalars[CacheIdx] = V;
280  }
281
282  /// Reset an existing scalar value for \p Def and a given \p Instance.
283  void reset(VPValue *Def, Value *V, const VPIteration &Instance) {
284    auto Iter = Data.PerPartScalars.find(Def);
285    assert(Iter != Data.PerPartScalars.end() &&
286           "need to overwrite existing value");
287    assert(Instance.Part < Iter->second.size() &&
288           "need to overwrite existing value");
289    unsigned CacheIdx = Instance.Lane.mapToCacheIndex(VF);
290    assert(CacheIdx < Iter->second[Instance.Part].size() &&
291           "need to overwrite existing value");
292    Iter->second[Instance.Part][CacheIdx] = V;
293  }
294
295  /// Hold state information used when constructing the CFG of the output IR,
296  /// traversing the VPBasicBlocks and generating corresponding IR BasicBlocks.
297  struct CFGState {
298    /// The previous VPBasicBlock visited. Initially set to null.
299    VPBasicBlock *PrevVPBB = nullptr;
300
301    /// The previous IR BasicBlock created or used. Initially set to the new
302    /// header BasicBlock.
303    BasicBlock *PrevBB = nullptr;
304
305    /// The last IR BasicBlock in the output IR. Set to the new latch
306    /// BasicBlock, used for placing the newly created BasicBlocks.
307    BasicBlock *LastBB = nullptr;
308
309    /// A mapping of each VPBasicBlock to the corresponding BasicBlock. In case
310    /// of replication, maps the BasicBlock of the last replica created.
311    SmallDenseMap<VPBasicBlock *, BasicBlock *> VPBB2IRBB;
312
313    /// Vector of VPBasicBlocks whose terminator instruction needs to be fixed
314    /// up at the end of vector code generation.
315    SmallVector<VPBasicBlock *, 8> VPBBsToFix;
316
317    CFGState() = default;
318  } CFG;
319
320  /// Hold a pointer to LoopInfo to register new basic blocks in the loop.
321  LoopInfo *LI;
322
323  /// Hold a pointer to Dominator Tree to register new basic blocks in the loop.
324  DominatorTree *DT;
325
326  /// Hold a reference to the IRBuilder used to generate output IR code.
327  IRBuilder<> &Builder;
328
329  VPValue2ValueTy VPValue2Value;
330
331  /// Hold the canonical scalar IV of the vector loop (start=0, step=VF*UF).
332  Value *CanonicalIV = nullptr;
333
334  /// Hold the trip count of the scalar loop.
335  Value *TripCount = nullptr;
336
337  /// Hold a pointer to InnerLoopVectorizer to reuse its IR generation methods.
338  InnerLoopVectorizer *ILV;
339
340  /// Pointer to the VPlan code is generated for.
341  VPlan *Plan;
342};
343
344/// VPUsers instance used by VPBlockBase to manage CondBit and the block
345/// predicate. Currently VPBlockUsers are used in VPBlockBase for historical
346/// reasons, but in the future the only VPUsers should either be recipes or
347/// live-outs.VPBlockBase uses.
348struct VPBlockUser : public VPUser {
349  VPBlockUser() : VPUser({}, VPUserID::Block) {}
350
351  VPValue *getSingleOperandOrNull() {
352    if (getNumOperands() == 1)
353      return getOperand(0);
354
355    return nullptr;
356  }
357  const VPValue *getSingleOperandOrNull() const {
358    if (getNumOperands() == 1)
359      return getOperand(0);
360
361    return nullptr;
362  }
363
364  void resetSingleOpUser(VPValue *NewVal) {
365    assert(getNumOperands() <= 1 && "Didn't expect more than one operand!");
366    if (!NewVal) {
367      if (getNumOperands() == 1)
368        removeLastOperand();
369      return;
370    }
371
372    if (getNumOperands() == 1)
373      setOperand(0, NewVal);
374    else
375      addOperand(NewVal);
376  }
377};
378
379/// VPBlockBase is the building block of the Hierarchical Control-Flow Graph.
380/// A VPBlockBase can be either a VPBasicBlock or a VPRegionBlock.
381class VPBlockBase {
382  friend class VPBlockUtils;
383
384  const unsigned char SubclassID; ///< Subclass identifier (for isa/dyn_cast).
385
386  /// An optional name for the block.
387  std::string Name;
388
389  /// The immediate VPRegionBlock which this VPBlockBase belongs to, or null if
390  /// it is a topmost VPBlockBase.
391  VPRegionBlock *Parent = nullptr;
392
393  /// List of predecessor blocks.
394  SmallVector<VPBlockBase *, 1> Predecessors;
395
396  /// List of successor blocks.
397  SmallVector<VPBlockBase *, 1> Successors;
398
399  /// Successor selector managed by a VPUser. For blocks with zero or one
400  /// successors, there is no operand. Otherwise there is exactly one operand
401  /// which is the branch condition.
402  VPBlockUser CondBitUser;
403
404  /// If the block is predicated, its predicate is stored as an operand of this
405  /// VPUser to maintain the def-use relations. Otherwise there is no operand
406  /// here.
407  VPBlockUser PredicateUser;
408
409  /// VPlan containing the block. Can only be set on the entry block of the
410  /// plan.
411  VPlan *Plan = nullptr;
412
413  /// Add \p Successor as the last successor to this block.
414  void appendSuccessor(VPBlockBase *Successor) {
415    assert(Successor && "Cannot add nullptr successor!");
416    Successors.push_back(Successor);
417  }
418
419  /// Add \p Predecessor as the last predecessor to this block.
420  void appendPredecessor(VPBlockBase *Predecessor) {
421    assert(Predecessor && "Cannot add nullptr predecessor!");
422    Predecessors.push_back(Predecessor);
423  }
424
425  /// Remove \p Predecessor from the predecessors of this block.
426  void removePredecessor(VPBlockBase *Predecessor) {
427    auto Pos = find(Predecessors, Predecessor);
428    assert(Pos && "Predecessor does not exist");
429    Predecessors.erase(Pos);
430  }
431
432  /// Remove \p Successor from the successors of this block.
433  void removeSuccessor(VPBlockBase *Successor) {
434    auto Pos = find(Successors, Successor);
435    assert(Pos && "Successor does not exist");
436    Successors.erase(Pos);
437  }
438
439protected:
440  VPBlockBase(const unsigned char SC, const std::string &N)
441      : SubclassID(SC), Name(N) {}
442
443public:
444  /// An enumeration for keeping track of the concrete subclass of VPBlockBase
445  /// that are actually instantiated. Values of this enumeration are kept in the
446  /// SubclassID field of the VPBlockBase objects. They are used for concrete
447  /// type identification.
448  using VPBlockTy = enum { VPBasicBlockSC, VPRegionBlockSC };
449
450  using VPBlocksTy = SmallVectorImpl<VPBlockBase *>;
451
452  virtual ~VPBlockBase() = default;
453
454  const std::string &getName() const { return Name; }
455
456  void setName(const Twine &newName) { Name = newName.str(); }
457
458  /// \return an ID for the concrete type of this object.
459  /// This is used to implement the classof checks. This should not be used
460  /// for any other purpose, as the values may change as LLVM evolves.
461  unsigned getVPBlockID() const { return SubclassID; }
462
463  VPRegionBlock *getParent() { return Parent; }
464  const VPRegionBlock *getParent() const { return Parent; }
465
466  /// \return A pointer to the plan containing the current block.
467  VPlan *getPlan();
468  const VPlan *getPlan() const;
469
470  /// Sets the pointer of the plan containing the block. The block must be the
471  /// entry block into the VPlan.
472  void setPlan(VPlan *ParentPlan);
473
474  void setParent(VPRegionBlock *P) { Parent = P; }
475
476  /// \return the VPBasicBlock that is the entry of this VPBlockBase,
477  /// recursively, if the latter is a VPRegionBlock. Otherwise, if this
478  /// VPBlockBase is a VPBasicBlock, it is returned.
479  const VPBasicBlock *getEntryBasicBlock() const;
480  VPBasicBlock *getEntryBasicBlock();
481
482  /// \return the VPBasicBlock that is the exit of this VPBlockBase,
483  /// recursively, if the latter is a VPRegionBlock. Otherwise, if this
484  /// VPBlockBase is a VPBasicBlock, it is returned.
485  const VPBasicBlock *getExitBasicBlock() const;
486  VPBasicBlock *getExitBasicBlock();
487
488  const VPBlocksTy &getSuccessors() const { return Successors; }
489  VPBlocksTy &getSuccessors() { return Successors; }
490
491  const VPBlocksTy &getPredecessors() const { return Predecessors; }
492  VPBlocksTy &getPredecessors() { return Predecessors; }
493
494  /// \return the successor of this VPBlockBase if it has a single successor.
495  /// Otherwise return a null pointer.
496  VPBlockBase *getSingleSuccessor() const {
497    return (Successors.size() == 1 ? *Successors.begin() : nullptr);
498  }
499
500  /// \return the predecessor of this VPBlockBase if it has a single
501  /// predecessor. Otherwise return a null pointer.
502  VPBlockBase *getSinglePredecessor() const {
503    return (Predecessors.size() == 1 ? *Predecessors.begin() : nullptr);
504  }
505
506  size_t getNumSuccessors() const { return Successors.size(); }
507  size_t getNumPredecessors() const { return Predecessors.size(); }
508
509  /// An Enclosing Block of a block B is any block containing B, including B
510  /// itself. \return the closest enclosing block starting from "this", which
511  /// has successors. \return the root enclosing block if all enclosing blocks
512  /// have no successors.
513  VPBlockBase *getEnclosingBlockWithSuccessors();
514
515  /// \return the closest enclosing block starting from "this", which has
516  /// predecessors. \return the root enclosing block if all enclosing blocks
517  /// have no predecessors.
518  VPBlockBase *getEnclosingBlockWithPredecessors();
519
520  /// \return the successors either attached directly to this VPBlockBase or, if
521  /// this VPBlockBase is the exit block of a VPRegionBlock and has no
522  /// successors of its own, search recursively for the first enclosing
523  /// VPRegionBlock that has successors and return them. If no such
524  /// VPRegionBlock exists, return the (empty) successors of the topmost
525  /// VPBlockBase reached.
526  const VPBlocksTy &getHierarchicalSuccessors() {
527    return getEnclosingBlockWithSuccessors()->getSuccessors();
528  }
529
530  /// \return the hierarchical successor of this VPBlockBase if it has a single
531  /// hierarchical successor. Otherwise return a null pointer.
532  VPBlockBase *getSingleHierarchicalSuccessor() {
533    return getEnclosingBlockWithSuccessors()->getSingleSuccessor();
534  }
535
536  /// \return the predecessors either attached directly to this VPBlockBase or,
537  /// if this VPBlockBase is the entry block of a VPRegionBlock and has no
538  /// predecessors of its own, search recursively for the first enclosing
539  /// VPRegionBlock that has predecessors and return them. If no such
540  /// VPRegionBlock exists, return the (empty) predecessors of the topmost
541  /// VPBlockBase reached.
542  const VPBlocksTy &getHierarchicalPredecessors() {
543    return getEnclosingBlockWithPredecessors()->getPredecessors();
544  }
545
546  /// \return the hierarchical predecessor of this VPBlockBase if it has a
547  /// single hierarchical predecessor. Otherwise return a null pointer.
548  VPBlockBase *getSingleHierarchicalPredecessor() {
549    return getEnclosingBlockWithPredecessors()->getSinglePredecessor();
550  }
551
552  /// \return the condition bit selecting the successor.
553  VPValue *getCondBit();
554  /// \return the condition bit selecting the successor.
555  const VPValue *getCondBit() const;
556  /// Set the condition bit selecting the successor.
557  void setCondBit(VPValue *CV);
558
559  /// \return the block's predicate.
560  VPValue *getPredicate();
561  /// \return the block's predicate.
562  const VPValue *getPredicate() const;
563  /// Set the block's predicate.
564  void setPredicate(VPValue *Pred);
565
566  /// Set a given VPBlockBase \p Successor as the single successor of this
567  /// VPBlockBase. This VPBlockBase is not added as predecessor of \p Successor.
568  /// This VPBlockBase must have no successors.
569  void setOneSuccessor(VPBlockBase *Successor) {
570    assert(Successors.empty() && "Setting one successor when others exist.");
571    appendSuccessor(Successor);
572  }
573
574  /// Set two given VPBlockBases \p IfTrue and \p IfFalse to be the two
575  /// successors of this VPBlockBase. \p Condition is set as the successor
576  /// selector. This VPBlockBase is not added as predecessor of \p IfTrue or \p
577  /// IfFalse. This VPBlockBase must have no successors.
578  void setTwoSuccessors(VPBlockBase *IfTrue, VPBlockBase *IfFalse,
579                        VPValue *Condition) {
580    assert(Successors.empty() && "Setting two successors when others exist.");
581    assert(Condition && "Setting two successors without condition!");
582    setCondBit(Condition);
583    appendSuccessor(IfTrue);
584    appendSuccessor(IfFalse);
585  }
586
587  /// Set each VPBasicBlock in \p NewPreds as predecessor of this VPBlockBase.
588  /// This VPBlockBase must have no predecessors. This VPBlockBase is not added
589  /// as successor of any VPBasicBlock in \p NewPreds.
590  void setPredecessors(ArrayRef<VPBlockBase *> NewPreds) {
591    assert(Predecessors.empty() && "Block predecessors already set.");
592    for (auto *Pred : NewPreds)
593      appendPredecessor(Pred);
594  }
595
596  /// Remove all the predecessor of this block.
597  void clearPredecessors() { Predecessors.clear(); }
598
599  /// Remove all the successors of this block and set to null its condition bit
600  void clearSuccessors() {
601    Successors.clear();
602    setCondBit(nullptr);
603  }
604
605  /// The method which generates the output IR that correspond to this
606  /// VPBlockBase, thereby "executing" the VPlan.
607  virtual void execute(struct VPTransformState *State) = 0;
608
609  /// Delete all blocks reachable from a given VPBlockBase, inclusive.
610  static void deleteCFG(VPBlockBase *Entry);
611
612  /// Return true if it is legal to hoist instructions into this block.
613  bool isLegalToHoistInto() {
614    // There are currently no constraints that prevent an instruction to be
615    // hoisted into a VPBlockBase.
616    return true;
617  }
618
619  /// Replace all operands of VPUsers in the block with \p NewValue and also
620  /// replaces all uses of VPValues defined in the block with NewValue.
621  virtual void dropAllReferences(VPValue *NewValue) = 0;
622
623#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
624  void printAsOperand(raw_ostream &OS, bool PrintType) const {
625    OS << getName();
626  }
627
628  /// Print plain-text dump of this VPBlockBase to \p O, prefixing all lines
629  /// with \p Indent. \p SlotTracker is used to print unnamed VPValue's using
630  /// consequtive numbers.
631  ///
632  /// Note that the numbering is applied to the whole VPlan, so printing
633  /// individual blocks is consistent with the whole VPlan printing.
634  virtual void print(raw_ostream &O, const Twine &Indent,
635                     VPSlotTracker &SlotTracker) const = 0;
636
637  /// Print plain-text dump of this VPlan to \p O.
638  void print(raw_ostream &O) const {
639    VPSlotTracker SlotTracker(getPlan());
640    print(O, "", SlotTracker);
641  }
642
643  /// Dump this VPBlockBase to dbgs().
644  LLVM_DUMP_METHOD void dump() const { print(dbgs()); }
645#endif
646};
647
648/// VPRecipeBase is a base class modeling a sequence of one or more output IR
649/// instructions. VPRecipeBase owns the the VPValues it defines through VPDef
650/// and is responsible for deleting its defined values. Single-value
651/// VPRecipeBases that also inherit from VPValue must make sure to inherit from
652/// VPRecipeBase before VPValue.
653class VPRecipeBase : public ilist_node_with_parent<VPRecipeBase, VPBasicBlock>,
654                     public VPDef,
655                     public VPUser {
656  friend VPBasicBlock;
657  friend class VPBlockUtils;
658
659  /// Each VPRecipe belongs to a single VPBasicBlock.
660  VPBasicBlock *Parent = nullptr;
661
662public:
663  VPRecipeBase(const unsigned char SC, ArrayRef<VPValue *> Operands)
664      : VPDef(SC), VPUser(Operands, VPUser::VPUserID::Recipe) {}
665
666  template <typename IterT>
667  VPRecipeBase(const unsigned char SC, iterator_range<IterT> Operands)
668      : VPDef(SC), VPUser(Operands, VPUser::VPUserID::Recipe) {}
669  virtual ~VPRecipeBase() = default;
670
671  /// \return the VPBasicBlock which this VPRecipe belongs to.
672  VPBasicBlock *getParent() { return Parent; }
673  const VPBasicBlock *getParent() const { return Parent; }
674
675  /// The method which generates the output IR instructions that correspond to
676  /// this VPRecipe, thereby "executing" the VPlan.
677  virtual void execute(struct VPTransformState &State) = 0;
678
679  /// Insert an unlinked recipe into a basic block immediately before
680  /// the specified recipe.
681  void insertBefore(VPRecipeBase *InsertPos);
682
683  /// Insert an unlinked Recipe into a basic block immediately after
684  /// the specified Recipe.
685  void insertAfter(VPRecipeBase *InsertPos);
686
687  /// Unlink this recipe from its current VPBasicBlock and insert it into
688  /// the VPBasicBlock that MovePos lives in, right after MovePos.
689  void moveAfter(VPRecipeBase *MovePos);
690
691  /// Unlink this recipe and insert into BB before I.
692  ///
693  /// \pre I is a valid iterator into BB.
694  void moveBefore(VPBasicBlock &BB, iplist<VPRecipeBase>::iterator I);
695
696  /// This method unlinks 'this' from the containing basic block, but does not
697  /// delete it.
698  void removeFromParent();
699
700  /// This method unlinks 'this' from the containing basic block and deletes it.
701  ///
702  /// \returns an iterator pointing to the element after the erased one
703  iplist<VPRecipeBase>::iterator eraseFromParent();
704
705  /// Returns the underlying instruction, if the recipe is a VPValue or nullptr
706  /// otherwise.
707  Instruction *getUnderlyingInstr() {
708    return cast<Instruction>(getVPSingleValue()->getUnderlyingValue());
709  }
710  const Instruction *getUnderlyingInstr() const {
711    return cast<Instruction>(getVPSingleValue()->getUnderlyingValue());
712  }
713
714  /// Method to support type inquiry through isa, cast, and dyn_cast.
715  static inline bool classof(const VPDef *D) {
716    // All VPDefs are also VPRecipeBases.
717    return true;
718  }
719
720  static inline bool classof(const VPUser *U) {
721    return U->getVPUserID() == VPUser::VPUserID::Recipe;
722  }
723
724  /// Returns true if the recipe may have side-effects.
725  bool mayHaveSideEffects() const;
726
727  /// Returns true for PHI-like recipes.
728  bool isPhi() const {
729    return getVPDefID() == VPWidenIntOrFpInductionSC || getVPDefID() == VPWidenPHISC ||
730      getVPDefID() == VPPredInstPHISC || getVPDefID() == VPWidenCanonicalIVSC;
731  }
732};
733
734inline bool VPUser::classof(const VPDef *Def) {
735  return Def->getVPDefID() == VPRecipeBase::VPInstructionSC ||
736         Def->getVPDefID() == VPRecipeBase::VPWidenSC ||
737         Def->getVPDefID() == VPRecipeBase::VPWidenCallSC ||
738         Def->getVPDefID() == VPRecipeBase::VPWidenSelectSC ||
739         Def->getVPDefID() == VPRecipeBase::VPWidenGEPSC ||
740         Def->getVPDefID() == VPRecipeBase::VPBlendSC ||
741         Def->getVPDefID() == VPRecipeBase::VPInterleaveSC ||
742         Def->getVPDefID() == VPRecipeBase::VPReplicateSC ||
743         Def->getVPDefID() == VPRecipeBase::VPReductionSC ||
744         Def->getVPDefID() == VPRecipeBase::VPBranchOnMaskSC ||
745         Def->getVPDefID() == VPRecipeBase::VPWidenMemoryInstructionSC;
746}
747
748/// This is a concrete Recipe that models a single VPlan-level instruction.
749/// While as any Recipe it may generate a sequence of IR instructions when
750/// executed, these instructions would always form a single-def expression as
751/// the VPInstruction is also a single def-use vertex.
752class VPInstruction : public VPRecipeBase, public VPValue {
753  friend class VPlanSlp;
754
755public:
756  /// VPlan opcodes, extending LLVM IR with idiomatics instructions.
757  enum {
758    Not = Instruction::OtherOpsEnd + 1,
759    ICmpULE,
760    SLPLoad,
761    SLPStore,
762    ActiveLaneMask,
763  };
764
765private:
766  typedef unsigned char OpcodeTy;
767  OpcodeTy Opcode;
768
769  /// Utility method serving execute(): generates a single instance of the
770  /// modeled instruction.
771  void generateInstruction(VPTransformState &State, unsigned Part);
772
773protected:
774  void setUnderlyingInstr(Instruction *I) { setUnderlyingValue(I); }
775
776public:
777  VPInstruction(unsigned Opcode, ArrayRef<VPValue *> Operands)
778      : VPRecipeBase(VPRecipeBase::VPInstructionSC, Operands),
779        VPValue(VPValue::VPVInstructionSC, nullptr, this), Opcode(Opcode) {}
780
781  VPInstruction(unsigned Opcode, ArrayRef<VPInstruction *> Operands)
782      : VPRecipeBase(VPRecipeBase::VPInstructionSC, {}),
783        VPValue(VPValue::VPVInstructionSC, nullptr, this), Opcode(Opcode) {
784    for (auto *I : Operands)
785      addOperand(I->getVPSingleValue());
786  }
787
788  VPInstruction(unsigned Opcode, std::initializer_list<VPValue *> Operands)
789      : VPInstruction(Opcode, ArrayRef<VPValue *>(Operands)) {}
790
791  /// Method to support type inquiry through isa, cast, and dyn_cast.
792  static inline bool classof(const VPValue *V) {
793    return V->getVPValueID() == VPValue::VPVInstructionSC;
794  }
795
796  VPInstruction *clone() const {
797    SmallVector<VPValue *, 2> Operands(operands());
798    return new VPInstruction(Opcode, Operands);
799  }
800
801  /// Method to support type inquiry through isa, cast, and dyn_cast.
802  static inline bool classof(const VPDef *R) {
803    return R->getVPDefID() == VPRecipeBase::VPInstructionSC;
804  }
805
806  unsigned getOpcode() const { return Opcode; }
807
808  /// Generate the instruction.
809  /// TODO: We currently execute only per-part unless a specific instance is
810  /// provided.
811  void execute(VPTransformState &State) override;
812
813#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
814  /// Print the VPInstruction to \p O.
815  void print(raw_ostream &O, const Twine &Indent,
816             VPSlotTracker &SlotTracker) const override;
817
818  /// Print the VPInstruction to dbgs() (for debugging).
819  LLVM_DUMP_METHOD void dump() const;
820#endif
821
822  /// Return true if this instruction may modify memory.
823  bool mayWriteToMemory() const {
824    // TODO: we can use attributes of the called function to rule out memory
825    //       modifications.
826    return Opcode == Instruction::Store || Opcode == Instruction::Call ||
827           Opcode == Instruction::Invoke || Opcode == SLPStore;
828  }
829
830  bool hasResult() const {
831    // CallInst may or may not have a result, depending on the called function.
832    // Conservatively return calls have results for now.
833    switch (getOpcode()) {
834    case Instruction::Ret:
835    case Instruction::Br:
836    case Instruction::Store:
837    case Instruction::Switch:
838    case Instruction::IndirectBr:
839    case Instruction::Resume:
840    case Instruction::CatchRet:
841    case Instruction::Unreachable:
842    case Instruction::Fence:
843    case Instruction::AtomicRMW:
844      return false;
845    default:
846      return true;
847    }
848  }
849};
850
851/// VPWidenRecipe is a recipe for producing a copy of vector type its
852/// ingredient. This recipe covers most of the traditional vectorization cases
853/// where each ingredient transforms into a vectorized version of itself.
854class VPWidenRecipe : public VPRecipeBase, public VPValue {
855public:
856  template <typename IterT>
857  VPWidenRecipe(Instruction &I, iterator_range<IterT> Operands)
858      : VPRecipeBase(VPRecipeBase::VPWidenSC, Operands),
859        VPValue(VPValue::VPVWidenSC, &I, this) {}
860
861  ~VPWidenRecipe() override = default;
862
863  /// Method to support type inquiry through isa, cast, and dyn_cast.
864  static inline bool classof(const VPDef *D) {
865    return D->getVPDefID() == VPRecipeBase::VPWidenSC;
866  }
867  static inline bool classof(const VPValue *V) {
868    return V->getVPValueID() == VPValue::VPVWidenSC;
869  }
870
871  /// Produce widened copies of all Ingredients.
872  void execute(VPTransformState &State) override;
873
874#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
875  /// Print the recipe.
876  void print(raw_ostream &O, const Twine &Indent,
877             VPSlotTracker &SlotTracker) const override;
878#endif
879};
880
881/// A recipe for widening Call instructions.
882class VPWidenCallRecipe : public VPRecipeBase, public VPValue {
883
884public:
885  template <typename IterT>
886  VPWidenCallRecipe(CallInst &I, iterator_range<IterT> CallArguments)
887      : VPRecipeBase(VPRecipeBase::VPWidenCallSC, CallArguments),
888        VPValue(VPValue::VPVWidenCallSC, &I, this) {}
889
890  ~VPWidenCallRecipe() override = default;
891
892  /// Method to support type inquiry through isa, cast, and dyn_cast.
893  static inline bool classof(const VPDef *D) {
894    return D->getVPDefID() == VPRecipeBase::VPWidenCallSC;
895  }
896
897  /// Produce a widened version of the call instruction.
898  void execute(VPTransformState &State) override;
899
900#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
901  /// Print the recipe.
902  void print(raw_ostream &O, const Twine &Indent,
903             VPSlotTracker &SlotTracker) const override;
904#endif
905};
906
907/// A recipe for widening select instructions.
908class VPWidenSelectRecipe : public VPRecipeBase, public VPValue {
909
910  /// Is the condition of the select loop invariant?
911  bool InvariantCond;
912
913public:
914  template <typename IterT>
915  VPWidenSelectRecipe(SelectInst &I, iterator_range<IterT> Operands,
916                      bool InvariantCond)
917      : VPRecipeBase(VPRecipeBase::VPWidenSelectSC, Operands),
918        VPValue(VPValue::VPVWidenSelectSC, &I, this),
919        InvariantCond(InvariantCond) {}
920
921  ~VPWidenSelectRecipe() override = default;
922
923  /// Method to support type inquiry through isa, cast, and dyn_cast.
924  static inline bool classof(const VPDef *D) {
925    return D->getVPDefID() == VPRecipeBase::VPWidenSelectSC;
926  }
927
928  /// Produce a widened version of the select instruction.
929  void execute(VPTransformState &State) override;
930
931#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
932  /// Print the recipe.
933  void print(raw_ostream &O, const Twine &Indent,
934             VPSlotTracker &SlotTracker) const override;
935#endif
936};
937
938/// A recipe for handling GEP instructions.
939class VPWidenGEPRecipe : public VPRecipeBase, public VPValue {
940  bool IsPtrLoopInvariant;
941  SmallBitVector IsIndexLoopInvariant;
942
943public:
944  template <typename IterT>
945  VPWidenGEPRecipe(GetElementPtrInst *GEP, iterator_range<IterT> Operands)
946      : VPRecipeBase(VPRecipeBase::VPWidenGEPSC, Operands),
947        VPValue(VPWidenGEPSC, GEP, this),
948        IsIndexLoopInvariant(GEP->getNumIndices(), false) {}
949
950  template <typename IterT>
951  VPWidenGEPRecipe(GetElementPtrInst *GEP, iterator_range<IterT> Operands,
952                   Loop *OrigLoop)
953      : VPRecipeBase(VPRecipeBase::VPWidenGEPSC, Operands),
954        VPValue(VPValue::VPVWidenGEPSC, GEP, this),
955        IsIndexLoopInvariant(GEP->getNumIndices(), false) {
956    IsPtrLoopInvariant = OrigLoop->isLoopInvariant(GEP->getPointerOperand());
957    for (auto Index : enumerate(GEP->indices()))
958      IsIndexLoopInvariant[Index.index()] =
959          OrigLoop->isLoopInvariant(Index.value().get());
960  }
961  ~VPWidenGEPRecipe() override = default;
962
963  /// Method to support type inquiry through isa, cast, and dyn_cast.
964  static inline bool classof(const VPDef *D) {
965    return D->getVPDefID() == VPRecipeBase::VPWidenGEPSC;
966  }
967
968  /// Generate the gep nodes.
969  void execute(VPTransformState &State) override;
970
971#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
972  /// Print the recipe.
973  void print(raw_ostream &O, const Twine &Indent,
974             VPSlotTracker &SlotTracker) const override;
975#endif
976};
977
978/// A recipe for handling phi nodes of integer and floating-point inductions,
979/// producing their vector and scalar values.
980class VPWidenIntOrFpInductionRecipe : public VPRecipeBase {
981  PHINode *IV;
982
983public:
984  VPWidenIntOrFpInductionRecipe(PHINode *IV, VPValue *Start, Instruction *Cast,
985                                TruncInst *Trunc = nullptr)
986      : VPRecipeBase(VPWidenIntOrFpInductionSC, {Start}), IV(IV) {
987    if (Trunc)
988      new VPValue(Trunc, this);
989    else
990      new VPValue(IV, this);
991
992    if (Cast)
993      new VPValue(Cast, this);
994  }
995  ~VPWidenIntOrFpInductionRecipe() override = default;
996
997  /// Method to support type inquiry through isa, cast, and dyn_cast.
998  static inline bool classof(const VPDef *D) {
999    return D->getVPDefID() == VPRecipeBase::VPWidenIntOrFpInductionSC;
1000  }
1001
1002  /// Generate the vectorized and scalarized versions of the phi node as
1003  /// needed by their users.
1004  void execute(VPTransformState &State) override;
1005
1006#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1007  /// Print the recipe.
1008  void print(raw_ostream &O, const Twine &Indent,
1009             VPSlotTracker &SlotTracker) const override;
1010#endif
1011
1012  /// Returns the start value of the induction.
1013  VPValue *getStartValue() { return getOperand(0); }
1014
1015  /// Returns the cast VPValue, if one is attached, or nullptr otherwise.
1016  VPValue *getCastValue() {
1017    if (getNumDefinedValues() != 2)
1018      return nullptr;
1019    return getVPValue(1);
1020  }
1021
1022  /// Returns the first defined value as TruncInst, if it is one or nullptr
1023  /// otherwise.
1024  TruncInst *getTruncInst() {
1025    return dyn_cast_or_null<TruncInst>(getVPValue(0)->getUnderlyingValue());
1026  }
1027  const TruncInst *getTruncInst() const {
1028    return dyn_cast_or_null<TruncInst>(getVPValue(0)->getUnderlyingValue());
1029  }
1030};
1031
1032/// A recipe for handling all phi nodes except for integer and FP inductions.
1033/// For reduction PHIs, RdxDesc must point to the corresponding recurrence
1034/// descriptor, the start value is the first operand of the recipe and the
1035/// incoming value from the backedge is the second operand. In the VPlan native
1036/// path, all incoming VPValues & VPBasicBlock pairs are managed in the recipe
1037/// directly.
1038class VPWidenPHIRecipe : public VPRecipeBase, public VPValue {
1039  /// Descriptor for a reduction PHI.
1040  RecurrenceDescriptor *RdxDesc = nullptr;
1041
1042  /// List of incoming blocks. Only used in the VPlan native path.
1043  SmallVector<VPBasicBlock *, 2> IncomingBlocks;
1044
1045public:
1046  /// Create a new VPWidenPHIRecipe for the reduction \p Phi described by \p
1047  /// RdxDesc.
1048  VPWidenPHIRecipe(PHINode *Phi, RecurrenceDescriptor &RdxDesc, VPValue &Start)
1049      : VPWidenPHIRecipe(Phi) {
1050    this->RdxDesc = &RdxDesc;
1051    addOperand(&Start);
1052  }
1053
1054  /// Create a VPWidenPHIRecipe for \p Phi
1055  VPWidenPHIRecipe(PHINode *Phi)
1056      : VPRecipeBase(VPWidenPHISC, {}),
1057        VPValue(VPValue::VPVWidenPHISC, Phi, this) {}
1058  ~VPWidenPHIRecipe() override = default;
1059
1060  /// Method to support type inquiry through isa, cast, and dyn_cast.
1061  static inline bool classof(const VPDef *D) {
1062    return D->getVPDefID() == VPRecipeBase::VPWidenPHISC;
1063  }
1064  static inline bool classof(const VPValue *V) {
1065    return V->getVPValueID() == VPValue::VPVWidenPHISC;
1066  }
1067
1068  /// Generate the phi/select nodes.
1069  void execute(VPTransformState &State) override;
1070
1071#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1072  /// Print the recipe.
1073  void print(raw_ostream &O, const Twine &Indent,
1074             VPSlotTracker &SlotTracker) const override;
1075#endif
1076
1077  /// Returns the start value of the phi, if it is a reduction.
1078  VPValue *getStartValue() {
1079    return getNumOperands() == 0 ? nullptr : getOperand(0);
1080  }
1081
1082  /// Returns the incoming value from the loop backedge, if it is a reduction.
1083  VPValue *getBackedgeValue() {
1084    assert(RdxDesc && "second incoming value is only guaranteed to be backedge "
1085                      "value for reductions");
1086    return getOperand(1);
1087  }
1088
1089  /// Adds a pair (\p IncomingV, \p IncomingBlock) to the phi.
1090  void addIncoming(VPValue *IncomingV, VPBasicBlock *IncomingBlock) {
1091    addOperand(IncomingV);
1092    IncomingBlocks.push_back(IncomingBlock);
1093  }
1094
1095  /// Returns the \p I th incoming VPValue.
1096  VPValue *getIncomingValue(unsigned I) { return getOperand(I); }
1097
1098  /// Returns the \p I th incoming VPBasicBlock.
1099  VPBasicBlock *getIncomingBlock(unsigned I) { return IncomingBlocks[I]; }
1100
1101  RecurrenceDescriptor *getRecurrenceDescriptor() { return RdxDesc; }
1102};
1103
1104/// A recipe for vectorizing a phi-node as a sequence of mask-based select
1105/// instructions.
1106class VPBlendRecipe : public VPRecipeBase, public VPValue {
1107  PHINode *Phi;
1108
1109public:
1110  /// The blend operation is a User of the incoming values and of their
1111  /// respective masks, ordered [I0, M0, I1, M1, ...]. Note that a single value
1112  /// might be incoming with a full mask for which there is no VPValue.
1113  VPBlendRecipe(PHINode *Phi, ArrayRef<VPValue *> Operands)
1114      : VPRecipeBase(VPBlendSC, Operands),
1115        VPValue(VPValue::VPVBlendSC, Phi, this), Phi(Phi) {
1116    assert(Operands.size() > 0 &&
1117           ((Operands.size() == 1) || (Operands.size() % 2 == 0)) &&
1118           "Expected either a single incoming value or a positive even number "
1119           "of operands");
1120  }
1121
1122  /// Method to support type inquiry through isa, cast, and dyn_cast.
1123  static inline bool classof(const VPDef *D) {
1124    return D->getVPDefID() == VPRecipeBase::VPBlendSC;
1125  }
1126
1127  /// Return the number of incoming values, taking into account that a single
1128  /// incoming value has no mask.
1129  unsigned getNumIncomingValues() const { return (getNumOperands() + 1) / 2; }
1130
1131  /// Return incoming value number \p Idx.
1132  VPValue *getIncomingValue(unsigned Idx) const { return getOperand(Idx * 2); }
1133
1134  /// Return mask number \p Idx.
1135  VPValue *getMask(unsigned Idx) const { return getOperand(Idx * 2 + 1); }
1136
1137  /// Generate the phi/select nodes.
1138  void execute(VPTransformState &State) override;
1139
1140#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1141  /// Print the recipe.
1142  void print(raw_ostream &O, const Twine &Indent,
1143             VPSlotTracker &SlotTracker) const override;
1144#endif
1145};
1146
1147/// VPInterleaveRecipe is a recipe for transforming an interleave group of load
1148/// or stores into one wide load/store and shuffles. The first operand of a
1149/// VPInterleave recipe is the address, followed by the stored values, followed
1150/// by an optional mask.
1151class VPInterleaveRecipe : public VPRecipeBase {
1152  const InterleaveGroup<Instruction> *IG;
1153
1154  bool HasMask = false;
1155
1156public:
1157  VPInterleaveRecipe(const InterleaveGroup<Instruction> *IG, VPValue *Addr,
1158                     ArrayRef<VPValue *> StoredValues, VPValue *Mask)
1159      : VPRecipeBase(VPInterleaveSC, {Addr}), IG(IG) {
1160    for (unsigned i = 0; i < IG->getFactor(); ++i)
1161      if (Instruction *I = IG->getMember(i)) {
1162        if (I->getType()->isVoidTy())
1163          continue;
1164        new VPValue(I, this);
1165      }
1166
1167    for (auto *SV : StoredValues)
1168      addOperand(SV);
1169    if (Mask) {
1170      HasMask = true;
1171      addOperand(Mask);
1172    }
1173  }
1174  ~VPInterleaveRecipe() override = default;
1175
1176  /// Method to support type inquiry through isa, cast, and dyn_cast.
1177  static inline bool classof(const VPDef *D) {
1178    return D->getVPDefID() == VPRecipeBase::VPInterleaveSC;
1179  }
1180
1181  /// Return the address accessed by this recipe.
1182  VPValue *getAddr() const {
1183    return getOperand(0); // Address is the 1st, mandatory operand.
1184  }
1185
1186  /// Return the mask used by this recipe. Note that a full mask is represented
1187  /// by a nullptr.
1188  VPValue *getMask() const {
1189    // Mask is optional and therefore the last, currently 2nd operand.
1190    return HasMask ? getOperand(getNumOperands() - 1) : nullptr;
1191  }
1192
1193  /// Return the VPValues stored by this interleave group. If it is a load
1194  /// interleave group, return an empty ArrayRef.
1195  ArrayRef<VPValue *> getStoredValues() const {
1196    // The first operand is the address, followed by the stored values, followed
1197    // by an optional mask.
1198    return ArrayRef<VPValue *>(op_begin(), getNumOperands())
1199        .slice(1, getNumOperands() - (HasMask ? 2 : 1));
1200  }
1201
1202  /// Generate the wide load or store, and shuffles.
1203  void execute(VPTransformState &State) override;
1204
1205#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1206  /// Print the recipe.
1207  void print(raw_ostream &O, const Twine &Indent,
1208             VPSlotTracker &SlotTracker) const override;
1209#endif
1210
1211  const InterleaveGroup<Instruction> *getInterleaveGroup() { return IG; }
1212};
1213
1214/// A recipe to represent inloop reduction operations, performing a reduction on
1215/// a vector operand into a scalar value, and adding the result to a chain.
1216/// The Operands are {ChainOp, VecOp, [Condition]}.
1217class VPReductionRecipe : public VPRecipeBase, public VPValue {
1218  /// The recurrence decriptor for the reduction in question.
1219  RecurrenceDescriptor *RdxDesc;
1220  /// Pointer to the TTI, needed to create the target reduction
1221  const TargetTransformInfo *TTI;
1222
1223public:
1224  VPReductionRecipe(RecurrenceDescriptor *R, Instruction *I, VPValue *ChainOp,
1225                    VPValue *VecOp, VPValue *CondOp,
1226                    const TargetTransformInfo *TTI)
1227      : VPRecipeBase(VPRecipeBase::VPReductionSC, {ChainOp, VecOp}),
1228        VPValue(VPValue::VPVReductionSC, I, this), RdxDesc(R), TTI(TTI) {
1229    if (CondOp)
1230      addOperand(CondOp);
1231  }
1232
1233  ~VPReductionRecipe() override = default;
1234
1235  /// Method to support type inquiry through isa, cast, and dyn_cast.
1236  static inline bool classof(const VPValue *V) {
1237    return V->getVPValueID() == VPValue::VPVReductionSC;
1238  }
1239
1240  static inline bool classof(const VPDef *D) {
1241    return D->getVPDefID() == VPRecipeBase::VPReductionSC;
1242  }
1243
1244  /// Generate the reduction in the loop
1245  void execute(VPTransformState &State) override;
1246
1247#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1248  /// Print the recipe.
1249  void print(raw_ostream &O, const Twine &Indent,
1250             VPSlotTracker &SlotTracker) const override;
1251#endif
1252
1253  /// The VPValue of the scalar Chain being accumulated.
1254  VPValue *getChainOp() const { return getOperand(0); }
1255  /// The VPValue of the vector value to be reduced.
1256  VPValue *getVecOp() const { return getOperand(1); }
1257  /// The VPValue of the condition for the block.
1258  VPValue *getCondOp() const {
1259    return getNumOperands() > 2 ? getOperand(2) : nullptr;
1260  }
1261};
1262
1263/// VPReplicateRecipe replicates a given instruction producing multiple scalar
1264/// copies of the original scalar type, one per lane, instead of producing a
1265/// single copy of widened type for all lanes. If the instruction is known to be
1266/// uniform only one copy, per lane zero, will be generated.
1267class VPReplicateRecipe : public VPRecipeBase, public VPValue {
1268  /// Indicator if only a single replica per lane is needed.
1269  bool IsUniform;
1270
1271  /// Indicator if the replicas are also predicated.
1272  bool IsPredicated;
1273
1274  /// Indicator if the scalar values should also be packed into a vector.
1275  bool AlsoPack;
1276
1277public:
1278  template <typename IterT>
1279  VPReplicateRecipe(Instruction *I, iterator_range<IterT> Operands,
1280                    bool IsUniform, bool IsPredicated = false)
1281      : VPRecipeBase(VPReplicateSC, Operands), VPValue(VPVReplicateSC, I, this),
1282        IsUniform(IsUniform), IsPredicated(IsPredicated) {
1283    // Retain the previous behavior of predicateInstructions(), where an
1284    // insert-element of a predicated instruction got hoisted into the
1285    // predicated basic block iff it was its only user. This is achieved by
1286    // having predicated instructions also pack their values into a vector by
1287    // default unless they have a replicated user which uses their scalar value.
1288    AlsoPack = IsPredicated && !I->use_empty();
1289  }
1290
1291  ~VPReplicateRecipe() override = default;
1292
1293  /// Method to support type inquiry through isa, cast, and dyn_cast.
1294  static inline bool classof(const VPDef *D) {
1295    return D->getVPDefID() == VPRecipeBase::VPReplicateSC;
1296  }
1297
1298  static inline bool classof(const VPValue *V) {
1299    return V->getVPValueID() == VPValue::VPVReplicateSC;
1300  }
1301
1302  /// Generate replicas of the desired Ingredient. Replicas will be generated
1303  /// for all parts and lanes unless a specific part and lane are specified in
1304  /// the \p State.
1305  void execute(VPTransformState &State) override;
1306
1307  void setAlsoPack(bool Pack) { AlsoPack = Pack; }
1308
1309#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1310  /// Print the recipe.
1311  void print(raw_ostream &O, const Twine &Indent,
1312             VPSlotTracker &SlotTracker) const override;
1313#endif
1314
1315  bool isUniform() const { return IsUniform; }
1316
1317  bool isPacked() const { return AlsoPack; }
1318
1319  bool isPredicated() const { return IsPredicated; }
1320};
1321
1322/// A recipe for generating conditional branches on the bits of a mask.
1323class VPBranchOnMaskRecipe : public VPRecipeBase {
1324public:
1325  VPBranchOnMaskRecipe(VPValue *BlockInMask)
1326      : VPRecipeBase(VPBranchOnMaskSC, {}) {
1327    if (BlockInMask) // nullptr means all-one mask.
1328      addOperand(BlockInMask);
1329  }
1330
1331  /// Method to support type inquiry through isa, cast, and dyn_cast.
1332  static inline bool classof(const VPDef *D) {
1333    return D->getVPDefID() == VPRecipeBase::VPBranchOnMaskSC;
1334  }
1335
1336  /// Generate the extraction of the appropriate bit from the block mask and the
1337  /// conditional branch.
1338  void execute(VPTransformState &State) override;
1339
1340#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1341  /// Print the recipe.
1342  void print(raw_ostream &O, const Twine &Indent,
1343             VPSlotTracker &SlotTracker) const override {
1344    O << Indent << "BRANCH-ON-MASK ";
1345    if (VPValue *Mask = getMask())
1346      Mask->printAsOperand(O, SlotTracker);
1347    else
1348      O << " All-One";
1349  }
1350#endif
1351
1352  /// Return the mask used by this recipe. Note that a full mask is represented
1353  /// by a nullptr.
1354  VPValue *getMask() const {
1355    assert(getNumOperands() <= 1 && "should have either 0 or 1 operands");
1356    // Mask is optional.
1357    return getNumOperands() == 1 ? getOperand(0) : nullptr;
1358  }
1359};
1360
1361/// VPPredInstPHIRecipe is a recipe for generating the phi nodes needed when
1362/// control converges back from a Branch-on-Mask. The phi nodes are needed in
1363/// order to merge values that are set under such a branch and feed their uses.
1364/// The phi nodes can be scalar or vector depending on the users of the value.
1365/// This recipe works in concert with VPBranchOnMaskRecipe.
1366class VPPredInstPHIRecipe : public VPRecipeBase, public VPValue {
1367public:
1368  /// Construct a VPPredInstPHIRecipe given \p PredInst whose value needs a phi
1369  /// nodes after merging back from a Branch-on-Mask.
1370  VPPredInstPHIRecipe(VPValue *PredV)
1371      : VPRecipeBase(VPPredInstPHISC, PredV),
1372        VPValue(VPValue::VPVPredInstPHI, nullptr, this) {}
1373  ~VPPredInstPHIRecipe() override = default;
1374
1375  /// Method to support type inquiry through isa, cast, and dyn_cast.
1376  static inline bool classof(const VPDef *D) {
1377    return D->getVPDefID() == VPRecipeBase::VPPredInstPHISC;
1378  }
1379
1380  /// Generates phi nodes for live-outs as needed to retain SSA form.
1381  void execute(VPTransformState &State) override;
1382
1383#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1384  /// Print the recipe.
1385  void print(raw_ostream &O, const Twine &Indent,
1386             VPSlotTracker &SlotTracker) const override;
1387#endif
1388};
1389
1390/// A Recipe for widening load/store operations.
1391/// The recipe uses the following VPValues:
1392/// - For load: Address, optional mask
1393/// - For store: Address, stored value, optional mask
1394/// TODO: We currently execute only per-part unless a specific instance is
1395/// provided.
1396class VPWidenMemoryInstructionRecipe : public VPRecipeBase {
1397  Instruction &Ingredient;
1398
1399  void setMask(VPValue *Mask) {
1400    if (!Mask)
1401      return;
1402    addOperand(Mask);
1403  }
1404
1405  bool isMasked() const {
1406    return isStore() ? getNumOperands() == 3 : getNumOperands() == 2;
1407  }
1408
1409public:
1410  VPWidenMemoryInstructionRecipe(LoadInst &Load, VPValue *Addr, VPValue *Mask)
1411      : VPRecipeBase(VPWidenMemoryInstructionSC, {Addr}), Ingredient(Load) {
1412    new VPValue(VPValue::VPVMemoryInstructionSC, &Load, this);
1413    setMask(Mask);
1414  }
1415
1416  VPWidenMemoryInstructionRecipe(StoreInst &Store, VPValue *Addr,
1417                                 VPValue *StoredValue, VPValue *Mask)
1418      : VPRecipeBase(VPWidenMemoryInstructionSC, {Addr, StoredValue}),
1419        Ingredient(Store) {
1420    setMask(Mask);
1421  }
1422
1423  /// Method to support type inquiry through isa, cast, and dyn_cast.
1424  static inline bool classof(const VPDef *D) {
1425    return D->getVPDefID() == VPRecipeBase::VPWidenMemoryInstructionSC;
1426  }
1427
1428  /// Return the address accessed by this recipe.
1429  VPValue *getAddr() const {
1430    return getOperand(0); // Address is the 1st, mandatory operand.
1431  }
1432
1433  /// Return the mask used by this recipe. Note that a full mask is represented
1434  /// by a nullptr.
1435  VPValue *getMask() const {
1436    // Mask is optional and therefore the last operand.
1437    return isMasked() ? getOperand(getNumOperands() - 1) : nullptr;
1438  }
1439
1440  /// Returns true if this recipe is a store.
1441  bool isStore() const { return isa<StoreInst>(Ingredient); }
1442
1443  /// Return the address accessed by this recipe.
1444  VPValue *getStoredValue() const {
1445    assert(isStore() && "Stored value only available for store instructions");
1446    return getOperand(1); // Stored value is the 2nd, mandatory operand.
1447  }
1448
1449  /// Generate the wide load/store.
1450  void execute(VPTransformState &State) override;
1451
1452#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1453  /// Print the recipe.
1454  void print(raw_ostream &O, const Twine &Indent,
1455             VPSlotTracker &SlotTracker) const override;
1456#endif
1457};
1458
1459/// A Recipe for widening the canonical induction variable of the vector loop.
1460class VPWidenCanonicalIVRecipe : public VPRecipeBase {
1461public:
1462  VPWidenCanonicalIVRecipe() : VPRecipeBase(VPWidenCanonicalIVSC, {}) {
1463    new VPValue(nullptr, this);
1464  }
1465
1466  ~VPWidenCanonicalIVRecipe() override = default;
1467
1468  /// Method to support type inquiry through isa, cast, and dyn_cast.
1469  static inline bool classof(const VPDef *D) {
1470    return D->getVPDefID() == VPRecipeBase::VPWidenCanonicalIVSC;
1471  }
1472
1473  /// Generate a canonical vector induction variable of the vector loop, with
1474  /// start = {<Part*VF, Part*VF+1, ..., Part*VF+VF-1> for 0 <= Part < UF}, and
1475  /// step = <VF*UF, VF*UF, ..., VF*UF>.
1476  void execute(VPTransformState &State) override;
1477
1478#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1479  /// Print the recipe.
1480  void print(raw_ostream &O, const Twine &Indent,
1481             VPSlotTracker &SlotTracker) const override;
1482#endif
1483};
1484
1485/// VPBasicBlock serves as the leaf of the Hierarchical Control-Flow Graph. It
1486/// holds a sequence of zero or more VPRecipe's each representing a sequence of
1487/// output IR instructions. All PHI-like recipes must come before any non-PHI recipes.
1488class VPBasicBlock : public VPBlockBase {
1489public:
1490  using RecipeListTy = iplist<VPRecipeBase>;
1491
1492private:
1493  /// The VPRecipes held in the order of output instructions to generate.
1494  RecipeListTy Recipes;
1495
1496public:
1497  VPBasicBlock(const Twine &Name = "", VPRecipeBase *Recipe = nullptr)
1498      : VPBlockBase(VPBasicBlockSC, Name.str()) {
1499    if (Recipe)
1500      appendRecipe(Recipe);
1501  }
1502
1503  ~VPBasicBlock() override {
1504    while (!Recipes.empty())
1505      Recipes.pop_back();
1506  }
1507
1508  /// Instruction iterators...
1509  using iterator = RecipeListTy::iterator;
1510  using const_iterator = RecipeListTy::const_iterator;
1511  using reverse_iterator = RecipeListTy::reverse_iterator;
1512  using const_reverse_iterator = RecipeListTy::const_reverse_iterator;
1513
1514  //===--------------------------------------------------------------------===//
1515  /// Recipe iterator methods
1516  ///
1517  inline iterator begin() { return Recipes.begin(); }
1518  inline const_iterator begin() const { return Recipes.begin(); }
1519  inline iterator end() { return Recipes.end(); }
1520  inline const_iterator end() const { return Recipes.end(); }
1521
1522  inline reverse_iterator rbegin() { return Recipes.rbegin(); }
1523  inline const_reverse_iterator rbegin() const { return Recipes.rbegin(); }
1524  inline reverse_iterator rend() { return Recipes.rend(); }
1525  inline const_reverse_iterator rend() const { return Recipes.rend(); }
1526
1527  inline size_t size() const { return Recipes.size(); }
1528  inline bool empty() const { return Recipes.empty(); }
1529  inline const VPRecipeBase &front() const { return Recipes.front(); }
1530  inline VPRecipeBase &front() { return Recipes.front(); }
1531  inline const VPRecipeBase &back() const { return Recipes.back(); }
1532  inline VPRecipeBase &back() { return Recipes.back(); }
1533
1534  /// Returns a reference to the list of recipes.
1535  RecipeListTy &getRecipeList() { return Recipes; }
1536
1537  /// Returns a pointer to a member of the recipe list.
1538  static RecipeListTy VPBasicBlock::*getSublistAccess(VPRecipeBase *) {
1539    return &VPBasicBlock::Recipes;
1540  }
1541
1542  /// Method to support type inquiry through isa, cast, and dyn_cast.
1543  static inline bool classof(const VPBlockBase *V) {
1544    return V->getVPBlockID() == VPBlockBase::VPBasicBlockSC;
1545  }
1546
1547  void insert(VPRecipeBase *Recipe, iterator InsertPt) {
1548    assert(Recipe && "No recipe to append.");
1549    assert(!Recipe->Parent && "Recipe already in VPlan");
1550    Recipe->Parent = this;
1551    Recipes.insert(InsertPt, Recipe);
1552  }
1553
1554  /// Augment the existing recipes of a VPBasicBlock with an additional
1555  /// \p Recipe as the last recipe.
1556  void appendRecipe(VPRecipeBase *Recipe) { insert(Recipe, end()); }
1557
1558  /// The method which generates the output IR instructions that correspond to
1559  /// this VPBasicBlock, thereby "executing" the VPlan.
1560  void execute(struct VPTransformState *State) override;
1561
1562  /// Return the position of the first non-phi node recipe in the block.
1563  iterator getFirstNonPhi();
1564
1565  /// Returns an iterator range over the PHI-like recipes in the block.
1566  iterator_range<iterator> phis() {
1567    return make_range(begin(), getFirstNonPhi());
1568  }
1569
1570  void dropAllReferences(VPValue *NewValue) override;
1571
1572  /// Split current block at \p SplitAt by inserting a new block between the
1573  /// current block and its successors and moving all recipes starting at
1574  /// SplitAt to the new block. Returns the new block.
1575  VPBasicBlock *splitAt(iterator SplitAt);
1576
1577#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1578  /// Print this VPBsicBlock to \p O, prefixing all lines with \p Indent. \p
1579  /// SlotTracker is used to print unnamed VPValue's using consequtive numbers.
1580  ///
1581  /// Note that the numbering is applied to the whole VPlan, so printing
1582  /// individual blocks is consistent with the whole VPlan printing.
1583  void print(raw_ostream &O, const Twine &Indent,
1584             VPSlotTracker &SlotTracker) const override;
1585  using VPBlockBase::print; // Get the print(raw_stream &O) version.
1586#endif
1587
1588private:
1589  /// Create an IR BasicBlock to hold the output instructions generated by this
1590  /// VPBasicBlock, and return it. Update the CFGState accordingly.
1591  BasicBlock *createEmptyBasicBlock(VPTransformState::CFGState &CFG);
1592};
1593
1594/// VPRegionBlock represents a collection of VPBasicBlocks and VPRegionBlocks
1595/// which form a Single-Entry-Single-Exit subgraph of the output IR CFG.
1596/// A VPRegionBlock may indicate that its contents are to be replicated several
1597/// times. This is designed to support predicated scalarization, in which a
1598/// scalar if-then code structure needs to be generated VF * UF times. Having
1599/// this replication indicator helps to keep a single model for multiple
1600/// candidate VF's. The actual replication takes place only once the desired VF
1601/// and UF have been determined.
1602class VPRegionBlock : public VPBlockBase {
1603  /// Hold the Single Entry of the SESE region modelled by the VPRegionBlock.
1604  VPBlockBase *Entry;
1605
1606  /// Hold the Single Exit of the SESE region modelled by the VPRegionBlock.
1607  VPBlockBase *Exit;
1608
1609  /// An indicator whether this region is to generate multiple replicated
1610  /// instances of output IR corresponding to its VPBlockBases.
1611  bool IsReplicator;
1612
1613public:
1614  VPRegionBlock(VPBlockBase *Entry, VPBlockBase *Exit,
1615                const std::string &Name = "", bool IsReplicator = false)
1616      : VPBlockBase(VPRegionBlockSC, Name), Entry(Entry), Exit(Exit),
1617        IsReplicator(IsReplicator) {
1618    assert(Entry->getPredecessors().empty() && "Entry block has predecessors.");
1619    assert(Exit->getSuccessors().empty() && "Exit block has successors.");
1620    Entry->setParent(this);
1621    Exit->setParent(this);
1622  }
1623  VPRegionBlock(const std::string &Name = "", bool IsReplicator = false)
1624      : VPBlockBase(VPRegionBlockSC, Name), Entry(nullptr), Exit(nullptr),
1625        IsReplicator(IsReplicator) {}
1626
1627  ~VPRegionBlock() override {
1628    if (Entry) {
1629      VPValue DummyValue;
1630      Entry->dropAllReferences(&DummyValue);
1631      deleteCFG(Entry);
1632    }
1633  }
1634
1635  /// Method to support type inquiry through isa, cast, and dyn_cast.
1636  static inline bool classof(const VPBlockBase *V) {
1637    return V->getVPBlockID() == VPBlockBase::VPRegionBlockSC;
1638  }
1639
1640  const VPBlockBase *getEntry() const { return Entry; }
1641  VPBlockBase *getEntry() { return Entry; }
1642
1643  /// Set \p EntryBlock as the entry VPBlockBase of this VPRegionBlock. \p
1644  /// EntryBlock must have no predecessors.
1645  void setEntry(VPBlockBase *EntryBlock) {
1646    assert(EntryBlock->getPredecessors().empty() &&
1647           "Entry block cannot have predecessors.");
1648    Entry = EntryBlock;
1649    EntryBlock->setParent(this);
1650  }
1651
1652  // FIXME: DominatorTreeBase is doing 'A->getParent()->front()'. 'front' is a
1653  // specific interface of llvm::Function, instead of using
1654  // GraphTraints::getEntryNode. We should add a new template parameter to
1655  // DominatorTreeBase representing the Graph type.
1656  VPBlockBase &front() const { return *Entry; }
1657
1658  const VPBlockBase *getExit() const { return Exit; }
1659  VPBlockBase *getExit() { return Exit; }
1660
1661  /// Set \p ExitBlock as the exit VPBlockBase of this VPRegionBlock. \p
1662  /// ExitBlock must have no successors.
1663  void setExit(VPBlockBase *ExitBlock) {
1664    assert(ExitBlock->getSuccessors().empty() &&
1665           "Exit block cannot have successors.");
1666    Exit = ExitBlock;
1667    ExitBlock->setParent(this);
1668  }
1669
1670  /// An indicator whether this region is to generate multiple replicated
1671  /// instances of output IR corresponding to its VPBlockBases.
1672  bool isReplicator() const { return IsReplicator; }
1673
1674  /// The method which generates the output IR instructions that correspond to
1675  /// this VPRegionBlock, thereby "executing" the VPlan.
1676  void execute(struct VPTransformState *State) override;
1677
1678  void dropAllReferences(VPValue *NewValue) override;
1679
1680#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1681  /// Print this VPRegionBlock to \p O (recursively), prefixing all lines with
1682  /// \p Indent. \p SlotTracker is used to print unnamed VPValue's using
1683  /// consequtive numbers.
1684  ///
1685  /// Note that the numbering is applied to the whole VPlan, so printing
1686  /// individual regions is consistent with the whole VPlan printing.
1687  void print(raw_ostream &O, const Twine &Indent,
1688             VPSlotTracker &SlotTracker) const override;
1689  using VPBlockBase::print; // Get the print(raw_stream &O) version.
1690#endif
1691};
1692
1693//===----------------------------------------------------------------------===//
1694// GraphTraits specializations for VPlan Hierarchical Control-Flow Graphs     //
1695//===----------------------------------------------------------------------===//
1696
1697// The following set of template specializations implement GraphTraits to treat
1698// any VPBlockBase as a node in a graph of VPBlockBases. It's important to note
1699// that VPBlockBase traits don't recurse into VPRegioBlocks, i.e., if the
1700// VPBlockBase is a VPRegionBlock, this specialization provides access to its
1701// successors/predecessors but not to the blocks inside the region.
1702
1703template <> struct GraphTraits<VPBlockBase *> {
1704  using NodeRef = VPBlockBase *;
1705  using ChildIteratorType = SmallVectorImpl<VPBlockBase *>::iterator;
1706
1707  static NodeRef getEntryNode(NodeRef N) { return N; }
1708
1709  static inline ChildIteratorType child_begin(NodeRef N) {
1710    return N->getSuccessors().begin();
1711  }
1712
1713  static inline ChildIteratorType child_end(NodeRef N) {
1714    return N->getSuccessors().end();
1715  }
1716};
1717
1718template <> struct GraphTraits<const VPBlockBase *> {
1719  using NodeRef = const VPBlockBase *;
1720  using ChildIteratorType = SmallVectorImpl<VPBlockBase *>::const_iterator;
1721
1722  static NodeRef getEntryNode(NodeRef N) { return N; }
1723
1724  static inline ChildIteratorType child_begin(NodeRef N) {
1725    return N->getSuccessors().begin();
1726  }
1727
1728  static inline ChildIteratorType child_end(NodeRef N) {
1729    return N->getSuccessors().end();
1730  }
1731};
1732
1733// Inverse order specialization for VPBasicBlocks. Predecessors are used instead
1734// of successors for the inverse traversal.
1735template <> struct GraphTraits<Inverse<VPBlockBase *>> {
1736  using NodeRef = VPBlockBase *;
1737  using ChildIteratorType = SmallVectorImpl<VPBlockBase *>::iterator;
1738
1739  static NodeRef getEntryNode(Inverse<NodeRef> B) { return B.Graph; }
1740
1741  static inline ChildIteratorType child_begin(NodeRef N) {
1742    return N->getPredecessors().begin();
1743  }
1744
1745  static inline ChildIteratorType child_end(NodeRef N) {
1746    return N->getPredecessors().end();
1747  }
1748};
1749
1750// The following set of template specializations implement GraphTraits to
1751// treat VPRegionBlock as a graph and recurse inside its nodes. It's important
1752// to note that the blocks inside the VPRegionBlock are treated as VPBlockBases
1753// (i.e., no dyn_cast is performed, VPBlockBases specialization is used), so
1754// there won't be automatic recursion into other VPBlockBases that turn to be
1755// VPRegionBlocks.
1756
1757template <>
1758struct GraphTraits<VPRegionBlock *> : public GraphTraits<VPBlockBase *> {
1759  using GraphRef = VPRegionBlock *;
1760  using nodes_iterator = df_iterator<NodeRef>;
1761
1762  static NodeRef getEntryNode(GraphRef N) { return N->getEntry(); }
1763
1764  static nodes_iterator nodes_begin(GraphRef N) {
1765    return nodes_iterator::begin(N->getEntry());
1766  }
1767
1768  static nodes_iterator nodes_end(GraphRef N) {
1769    // df_iterator::end() returns an empty iterator so the node used doesn't
1770    // matter.
1771    return nodes_iterator::end(N);
1772  }
1773};
1774
1775template <>
1776struct GraphTraits<const VPRegionBlock *>
1777    : public GraphTraits<const VPBlockBase *> {
1778  using GraphRef = const VPRegionBlock *;
1779  using nodes_iterator = df_iterator<NodeRef>;
1780
1781  static NodeRef getEntryNode(GraphRef N) { return N->getEntry(); }
1782
1783  static nodes_iterator nodes_begin(GraphRef N) {
1784    return nodes_iterator::begin(N->getEntry());
1785  }
1786
1787  static nodes_iterator nodes_end(GraphRef N) {
1788    // df_iterator::end() returns an empty iterator so the node used doesn't
1789    // matter.
1790    return nodes_iterator::end(N);
1791  }
1792};
1793
1794template <>
1795struct GraphTraits<Inverse<VPRegionBlock *>>
1796    : public GraphTraits<Inverse<VPBlockBase *>> {
1797  using GraphRef = VPRegionBlock *;
1798  using nodes_iterator = df_iterator<NodeRef>;
1799
1800  static NodeRef getEntryNode(Inverse<GraphRef> N) {
1801    return N.Graph->getExit();
1802  }
1803
1804  static nodes_iterator nodes_begin(GraphRef N) {
1805    return nodes_iterator::begin(N->getExit());
1806  }
1807
1808  static nodes_iterator nodes_end(GraphRef N) {
1809    // df_iterator::end() returns an empty iterator so the node used doesn't
1810    // matter.
1811    return nodes_iterator::end(N);
1812  }
1813};
1814
1815/// Iterator to traverse all successors of a VPBlockBase node. This includes the
1816/// entry node of VPRegionBlocks. Exit blocks of a region implicitly have their
1817/// parent region's successors. This ensures all blocks in a region are visited
1818/// before any blocks in a successor region when doing a reverse post-order
1819// traversal of the graph.
1820template <typename BlockPtrTy>
1821class VPAllSuccessorsIterator
1822    : public iterator_facade_base<VPAllSuccessorsIterator<BlockPtrTy>,
1823                                  std::forward_iterator_tag, VPBlockBase> {
1824  BlockPtrTy Block;
1825  /// Index of the current successor. For VPBasicBlock nodes, this simply is the
1826  /// index for the successor array. For VPRegionBlock, SuccessorIdx == 0 is
1827  /// used for the region's entry block, and SuccessorIdx - 1 are the indices
1828  /// for the successor array.
1829  size_t SuccessorIdx;
1830
1831  static BlockPtrTy getBlockWithSuccs(BlockPtrTy Current) {
1832    while (Current && Current->getNumSuccessors() == 0)
1833      Current = Current->getParent();
1834    return Current;
1835  }
1836
1837  /// Templated helper to dereference successor \p SuccIdx of \p Block. Used by
1838  /// both the const and non-const operator* implementations.
1839  template <typename T1> static T1 deref(T1 Block, unsigned SuccIdx) {
1840    if (auto *R = dyn_cast<VPRegionBlock>(Block)) {
1841      if (SuccIdx == 0)
1842        return R->getEntry();
1843      SuccIdx--;
1844    }
1845
1846    // For exit blocks, use the next parent region with successors.
1847    return getBlockWithSuccs(Block)->getSuccessors()[SuccIdx];
1848  }
1849
1850public:
1851  VPAllSuccessorsIterator(BlockPtrTy Block, size_t Idx = 0)
1852      : Block(Block), SuccessorIdx(Idx) {}
1853  VPAllSuccessorsIterator(const VPAllSuccessorsIterator &Other)
1854      : Block(Other.Block), SuccessorIdx(Other.SuccessorIdx) {}
1855
1856  VPAllSuccessorsIterator &operator=(const VPAllSuccessorsIterator &R) {
1857    Block = R.Block;
1858    SuccessorIdx = R.SuccessorIdx;
1859    return *this;
1860  }
1861
1862  static VPAllSuccessorsIterator end(BlockPtrTy Block) {
1863    BlockPtrTy ParentWithSuccs = getBlockWithSuccs(Block);
1864    unsigned NumSuccessors = ParentWithSuccs
1865                                 ? ParentWithSuccs->getNumSuccessors()
1866                                 : Block->getNumSuccessors();
1867
1868    if (auto *R = dyn_cast<VPRegionBlock>(Block))
1869      return {R, NumSuccessors + 1};
1870    return {Block, NumSuccessors};
1871  }
1872
1873  bool operator==(const VPAllSuccessorsIterator &R) const {
1874    return Block == R.Block && SuccessorIdx == R.SuccessorIdx;
1875  }
1876
1877  const VPBlockBase *operator*() const { return deref(Block, SuccessorIdx); }
1878
1879  BlockPtrTy operator*() { return deref(Block, SuccessorIdx); }
1880
1881  VPAllSuccessorsIterator &operator++() {
1882    SuccessorIdx++;
1883    return *this;
1884  }
1885
1886  VPAllSuccessorsIterator operator++(int X) {
1887    VPAllSuccessorsIterator Orig = *this;
1888    SuccessorIdx++;
1889    return Orig;
1890  }
1891};
1892
1893/// Helper for GraphTraits specialization that traverses through VPRegionBlocks.
1894template <typename BlockTy> class VPBlockRecursiveTraversalWrapper {
1895  BlockTy Entry;
1896
1897public:
1898  VPBlockRecursiveTraversalWrapper(BlockTy Entry) : Entry(Entry) {}
1899  BlockTy getEntry() { return Entry; }
1900};
1901
1902/// GraphTraits specialization to recursively traverse VPBlockBase nodes,
1903/// including traversing through VPRegionBlocks.  Exit blocks of a region
1904/// implicitly have their parent region's successors. This ensures all blocks in
1905/// a region are visited before any blocks in a successor region when doing a
1906/// reverse post-order traversal of the graph.
1907template <>
1908struct GraphTraits<VPBlockRecursiveTraversalWrapper<VPBlockBase *>> {
1909  using NodeRef = VPBlockBase *;
1910  using ChildIteratorType = VPAllSuccessorsIterator<VPBlockBase *>;
1911
1912  static NodeRef
1913  getEntryNode(VPBlockRecursiveTraversalWrapper<VPBlockBase *> N) {
1914    return N.getEntry();
1915  }
1916
1917  static inline ChildIteratorType child_begin(NodeRef N) {
1918    return ChildIteratorType(N);
1919  }
1920
1921  static inline ChildIteratorType child_end(NodeRef N) {
1922    return ChildIteratorType::end(N);
1923  }
1924};
1925
1926template <>
1927struct GraphTraits<VPBlockRecursiveTraversalWrapper<const VPBlockBase *>> {
1928  using NodeRef = const VPBlockBase *;
1929  using ChildIteratorType = VPAllSuccessorsIterator<const VPBlockBase *>;
1930
1931  static NodeRef
1932  getEntryNode(VPBlockRecursiveTraversalWrapper<const VPBlockBase *> N) {
1933    return N.getEntry();
1934  }
1935
1936  static inline ChildIteratorType child_begin(NodeRef N) {
1937    return ChildIteratorType(N);
1938  }
1939
1940  static inline ChildIteratorType child_end(NodeRef N) {
1941    return ChildIteratorType::end(N);
1942  }
1943};
1944
1945/// VPlan models a candidate for vectorization, encoding various decisions take
1946/// to produce efficient output IR, including which branches, basic-blocks and
1947/// output IR instructions to generate, and their cost. VPlan holds a
1948/// Hierarchical-CFG of VPBasicBlocks and VPRegionBlocks rooted at an Entry
1949/// VPBlock.
1950class VPlan {
1951  friend class VPlanPrinter;
1952  friend class VPSlotTracker;
1953
1954  /// Hold the single entry to the Hierarchical CFG of the VPlan.
1955  VPBlockBase *Entry;
1956
1957  /// Holds the VFs applicable to this VPlan.
1958  SmallSetVector<ElementCount, 2> VFs;
1959
1960  /// Holds the name of the VPlan, for printing.
1961  std::string Name;
1962
1963  /// Holds all the external definitions created for this VPlan.
1964  // TODO: Introduce a specific representation for external definitions in
1965  // VPlan. External definitions must be immutable and hold a pointer to its
1966  // underlying IR that will be used to implement its structural comparison
1967  // (operators '==' and '<').
1968  SetVector<VPValue *> VPExternalDefs;
1969
1970  /// Represents the backedge taken count of the original loop, for folding
1971  /// the tail.
1972  VPValue *BackedgeTakenCount = nullptr;
1973
1974  /// Holds a mapping between Values and their corresponding VPValue inside
1975  /// VPlan.
1976  Value2VPValueTy Value2VPValue;
1977
1978  /// Contains all VPValues that been allocated by addVPValue directly and need
1979  /// to be free when the plan's destructor is called.
1980  SmallVector<VPValue *, 16> VPValuesToFree;
1981
1982  /// Holds the VPLoopInfo analysis for this VPlan.
1983  VPLoopInfo VPLInfo;
1984
1985public:
1986  VPlan(VPBlockBase *Entry = nullptr) : Entry(Entry) {
1987    if (Entry)
1988      Entry->setPlan(this);
1989  }
1990
1991  ~VPlan() {
1992    if (Entry) {
1993      VPValue DummyValue;
1994      for (VPBlockBase *Block : depth_first(Entry))
1995        Block->dropAllReferences(&DummyValue);
1996
1997      VPBlockBase::deleteCFG(Entry);
1998    }
1999    for (VPValue *VPV : VPValuesToFree)
2000      delete VPV;
2001    if (BackedgeTakenCount)
2002      delete BackedgeTakenCount;
2003    for (VPValue *Def : VPExternalDefs)
2004      delete Def;
2005  }
2006
2007  /// Generate the IR code for this VPlan.
2008  void execute(struct VPTransformState *State);
2009
2010  VPBlockBase *getEntry() { return Entry; }
2011  const VPBlockBase *getEntry() const { return Entry; }
2012
2013  VPBlockBase *setEntry(VPBlockBase *Block) {
2014    Entry = Block;
2015    Block->setPlan(this);
2016    return Entry;
2017  }
2018
2019  /// The backedge taken count of the original loop.
2020  VPValue *getOrCreateBackedgeTakenCount() {
2021    if (!BackedgeTakenCount)
2022      BackedgeTakenCount = new VPValue();
2023    return BackedgeTakenCount;
2024  }
2025
2026  void addVF(ElementCount VF) { VFs.insert(VF); }
2027
2028  bool hasVF(ElementCount VF) { return VFs.count(VF); }
2029
2030  const std::string &getName() const { return Name; }
2031
2032  void setName(const Twine &newName) { Name = newName.str(); }
2033
2034  /// Add \p VPVal to the pool of external definitions if it's not already
2035  /// in the pool.
2036  void addExternalDef(VPValue *VPVal) { VPExternalDefs.insert(VPVal); }
2037
2038  void addVPValue(Value *V) {
2039    assert(V && "Trying to add a null Value to VPlan");
2040    assert(!Value2VPValue.count(V) && "Value already exists in VPlan");
2041    VPValue *VPV = new VPValue(V);
2042    Value2VPValue[V] = VPV;
2043    VPValuesToFree.push_back(VPV);
2044  }
2045
2046  void addVPValue(Value *V, VPValue *VPV) {
2047    assert(V && "Trying to add a null Value to VPlan");
2048    assert(!Value2VPValue.count(V) && "Value already exists in VPlan");
2049    Value2VPValue[V] = VPV;
2050  }
2051
2052  VPValue *getVPValue(Value *V) {
2053    assert(V && "Trying to get the VPValue of a null Value");
2054    assert(Value2VPValue.count(V) && "Value does not exist in VPlan");
2055    return Value2VPValue[V];
2056  }
2057
2058  VPValue *getOrAddVPValue(Value *V) {
2059    assert(V && "Trying to get or add the VPValue of a null Value");
2060    if (!Value2VPValue.count(V))
2061      addVPValue(V);
2062    return getVPValue(V);
2063  }
2064
2065  void removeVPValueFor(Value *V) { Value2VPValue.erase(V); }
2066
2067  /// Return the VPLoopInfo analysis for this VPlan.
2068  VPLoopInfo &getVPLoopInfo() { return VPLInfo; }
2069  const VPLoopInfo &getVPLoopInfo() const { return VPLInfo; }
2070
2071#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
2072  /// Print this VPlan to \p O.
2073  void print(raw_ostream &O) const;
2074
2075  /// Print this VPlan in DOT format to \p O.
2076  void printDOT(raw_ostream &O) const;
2077
2078  /// Dump the plan to stderr (for debugging).
2079  LLVM_DUMP_METHOD void dump() const;
2080#endif
2081
2082  /// Returns a range mapping the values the range \p Operands to their
2083  /// corresponding VPValues.
2084  iterator_range<mapped_iterator<Use *, std::function<VPValue *(Value *)>>>
2085  mapToVPValues(User::op_range Operands) {
2086    std::function<VPValue *(Value *)> Fn = [this](Value *Op) {
2087      return getOrAddVPValue(Op);
2088    };
2089    return map_range(Operands, Fn);
2090  }
2091
2092private:
2093  /// Add to the given dominator tree the header block and every new basic block
2094  /// that was created between it and the latch block, inclusive.
2095  static void updateDominatorTree(DominatorTree *DT, BasicBlock *LoopLatchBB,
2096                                  BasicBlock *LoopPreHeaderBB,
2097                                  BasicBlock *LoopExitBB);
2098};
2099
2100#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
2101/// VPlanPrinter prints a given VPlan to a given output stream. The printing is
2102/// indented and follows the dot format.
2103class VPlanPrinter {
2104  raw_ostream &OS;
2105  const VPlan &Plan;
2106  unsigned Depth = 0;
2107  unsigned TabWidth = 2;
2108  std::string Indent;
2109  unsigned BID = 0;
2110  SmallDenseMap<const VPBlockBase *, unsigned> BlockID;
2111
2112  VPSlotTracker SlotTracker;
2113
2114  /// Handle indentation.
2115  void bumpIndent(int b) { Indent = std::string((Depth += b) * TabWidth, ' '); }
2116
2117  /// Print a given \p Block of the Plan.
2118  void dumpBlock(const VPBlockBase *Block);
2119
2120  /// Print the information related to the CFG edges going out of a given
2121  /// \p Block, followed by printing the successor blocks themselves.
2122  void dumpEdges(const VPBlockBase *Block);
2123
2124  /// Print a given \p BasicBlock, including its VPRecipes, followed by printing
2125  /// its successor blocks.
2126  void dumpBasicBlock(const VPBasicBlock *BasicBlock);
2127
2128  /// Print a given \p Region of the Plan.
2129  void dumpRegion(const VPRegionBlock *Region);
2130
2131  unsigned getOrCreateBID(const VPBlockBase *Block) {
2132    return BlockID.count(Block) ? BlockID[Block] : BlockID[Block] = BID++;
2133  }
2134
2135  const Twine getOrCreateName(const VPBlockBase *Block);
2136
2137  const Twine getUID(const VPBlockBase *Block);
2138
2139  /// Print the information related to a CFG edge between two VPBlockBases.
2140  void drawEdge(const VPBlockBase *From, const VPBlockBase *To, bool Hidden,
2141                const Twine &Label);
2142
2143public:
2144  VPlanPrinter(raw_ostream &O, const VPlan &P)
2145      : OS(O), Plan(P), SlotTracker(&P) {}
2146
2147  LLVM_DUMP_METHOD void dump();
2148};
2149
2150struct VPlanIngredient {
2151  const Value *V;
2152
2153  VPlanIngredient(const Value *V) : V(V) {}
2154
2155  void print(raw_ostream &O) const;
2156};
2157
2158inline raw_ostream &operator<<(raw_ostream &OS, const VPlanIngredient &I) {
2159  I.print(OS);
2160  return OS;
2161}
2162
2163inline raw_ostream &operator<<(raw_ostream &OS, const VPlan &Plan) {
2164  Plan.print(OS);
2165  return OS;
2166}
2167#endif
2168
2169//===----------------------------------------------------------------------===//
2170// VPlan Utilities
2171//===----------------------------------------------------------------------===//
2172
2173/// Class that provides utilities for VPBlockBases in VPlan.
2174class VPBlockUtils {
2175public:
2176  VPBlockUtils() = delete;
2177
2178  /// Insert disconnected VPBlockBase \p NewBlock after \p BlockPtr. Add \p
2179  /// NewBlock as successor of \p BlockPtr and \p BlockPtr as predecessor of \p
2180  /// NewBlock, and propagate \p BlockPtr parent to \p NewBlock. If \p BlockPtr
2181  /// has more than one successor, its conditional bit is propagated to \p
2182  /// NewBlock. \p NewBlock must have neither successors nor predecessors.
2183  static void insertBlockAfter(VPBlockBase *NewBlock, VPBlockBase *BlockPtr) {
2184    assert(NewBlock->getSuccessors().empty() &&
2185           "Can't insert new block with successors.");
2186    // TODO: move successors from BlockPtr to NewBlock when this functionality
2187    // is necessary. For now, setBlockSingleSuccessor will assert if BlockPtr
2188    // already has successors.
2189    BlockPtr->setOneSuccessor(NewBlock);
2190    NewBlock->setPredecessors({BlockPtr});
2191    NewBlock->setParent(BlockPtr->getParent());
2192  }
2193
2194  /// Insert disconnected VPBlockBases \p IfTrue and \p IfFalse after \p
2195  /// BlockPtr. Add \p IfTrue and \p IfFalse as succesors of \p BlockPtr and \p
2196  /// BlockPtr as predecessor of \p IfTrue and \p IfFalse. Propagate \p BlockPtr
2197  /// parent to \p IfTrue and \p IfFalse. \p Condition is set as the successor
2198  /// selector. \p BlockPtr must have no successors and \p IfTrue and \p IfFalse
2199  /// must have neither successors nor predecessors.
2200  static void insertTwoBlocksAfter(VPBlockBase *IfTrue, VPBlockBase *IfFalse,
2201                                   VPValue *Condition, VPBlockBase *BlockPtr) {
2202    assert(IfTrue->getSuccessors().empty() &&
2203           "Can't insert IfTrue with successors.");
2204    assert(IfFalse->getSuccessors().empty() &&
2205           "Can't insert IfFalse with successors.");
2206    BlockPtr->setTwoSuccessors(IfTrue, IfFalse, Condition);
2207    IfTrue->setPredecessors({BlockPtr});
2208    IfFalse->setPredecessors({BlockPtr});
2209    IfTrue->setParent(BlockPtr->getParent());
2210    IfFalse->setParent(BlockPtr->getParent());
2211  }
2212
2213  /// Connect VPBlockBases \p From and \p To bi-directionally. Append \p To to
2214  /// the successors of \p From and \p From to the predecessors of \p To. Both
2215  /// VPBlockBases must have the same parent, which can be null. Both
2216  /// VPBlockBases can be already connected to other VPBlockBases.
2217  static void connectBlocks(VPBlockBase *From, VPBlockBase *To) {
2218    assert((From->getParent() == To->getParent()) &&
2219           "Can't connect two block with different parents");
2220    assert(From->getNumSuccessors() < 2 &&
2221           "Blocks can't have more than two successors.");
2222    From->appendSuccessor(To);
2223    To->appendPredecessor(From);
2224  }
2225
2226  /// Disconnect VPBlockBases \p From and \p To bi-directionally. Remove \p To
2227  /// from the successors of \p From and \p From from the predecessors of \p To.
2228  static void disconnectBlocks(VPBlockBase *From, VPBlockBase *To) {
2229    assert(To && "Successor to disconnect is null.");
2230    From->removeSuccessor(To);
2231    To->removePredecessor(From);
2232  }
2233
2234  /// Returns true if the edge \p FromBlock -> \p ToBlock is a back-edge.
2235  static bool isBackEdge(const VPBlockBase *FromBlock,
2236                         const VPBlockBase *ToBlock, const VPLoopInfo *VPLI) {
2237    assert(FromBlock->getParent() == ToBlock->getParent() &&
2238           FromBlock->getParent() && "Must be in same region");
2239    const VPLoop *FromLoop = VPLI->getLoopFor(FromBlock);
2240    const VPLoop *ToLoop = VPLI->getLoopFor(ToBlock);
2241    if (!FromLoop || !ToLoop || FromLoop != ToLoop)
2242      return false;
2243
2244    // A back-edge is a branch from the loop latch to its header.
2245    return ToLoop->isLoopLatch(FromBlock) && ToBlock == ToLoop->getHeader();
2246  }
2247
2248  /// Returns true if \p Block is a loop latch
2249  static bool blockIsLoopLatch(const VPBlockBase *Block,
2250                               const VPLoopInfo *VPLInfo) {
2251    if (const VPLoop *ParentVPL = VPLInfo->getLoopFor(Block))
2252      return ParentVPL->isLoopLatch(Block);
2253
2254    return false;
2255  }
2256
2257  /// Count and return the number of succesors of \p PredBlock excluding any
2258  /// backedges.
2259  static unsigned countSuccessorsNoBE(VPBlockBase *PredBlock,
2260                                      VPLoopInfo *VPLI) {
2261    unsigned Count = 0;
2262    for (VPBlockBase *SuccBlock : PredBlock->getSuccessors()) {
2263      if (!VPBlockUtils::isBackEdge(PredBlock, SuccBlock, VPLI))
2264        Count++;
2265    }
2266    return Count;
2267  }
2268
2269  /// Return an iterator range over \p Range which only includes \p BlockTy
2270  /// blocks. The accesses are casted to \p BlockTy.
2271  template <typename BlockTy, typename T>
2272  static auto blocksOnly(const T &Range) {
2273    // Create BaseTy with correct const-ness based on BlockTy.
2274    using BaseTy =
2275        typename std::conditional<std::is_const<BlockTy>::value,
2276                                  const VPBlockBase, VPBlockBase>::type;
2277
2278    // We need to first create an iterator range over (const) BlocktTy & instead
2279    // of (const) BlockTy * for filter_range to work properly.
2280    auto Mapped =
2281        map_range(Range, [](BaseTy *Block) -> BaseTy & { return *Block; });
2282    auto Filter = make_filter_range(
2283        Mapped, [](BaseTy &Block) { return isa<BlockTy>(&Block); });
2284    return map_range(Filter, [](BaseTy &Block) -> BlockTy * {
2285      return cast<BlockTy>(&Block);
2286    });
2287  }
2288};
2289
2290class VPInterleavedAccessInfo {
2291  DenseMap<VPInstruction *, InterleaveGroup<VPInstruction> *>
2292      InterleaveGroupMap;
2293
2294  /// Type for mapping of instruction based interleave groups to VPInstruction
2295  /// interleave groups
2296  using Old2NewTy = DenseMap<InterleaveGroup<Instruction> *,
2297                             InterleaveGroup<VPInstruction> *>;
2298
2299  /// Recursively \p Region and populate VPlan based interleave groups based on
2300  /// \p IAI.
2301  void visitRegion(VPRegionBlock *Region, Old2NewTy &Old2New,
2302                   InterleavedAccessInfo &IAI);
2303  /// Recursively traverse \p Block and populate VPlan based interleave groups
2304  /// based on \p IAI.
2305  void visitBlock(VPBlockBase *Block, Old2NewTy &Old2New,
2306                  InterleavedAccessInfo &IAI);
2307
2308public:
2309  VPInterleavedAccessInfo(VPlan &Plan, InterleavedAccessInfo &IAI);
2310
2311  ~VPInterleavedAccessInfo() {
2312    SmallPtrSet<InterleaveGroup<VPInstruction> *, 4> DelSet;
2313    // Avoid releasing a pointer twice.
2314    for (auto &I : InterleaveGroupMap)
2315      DelSet.insert(I.second);
2316    for (auto *Ptr : DelSet)
2317      delete Ptr;
2318  }
2319
2320  /// Get the interleave group that \p Instr belongs to.
2321  ///
2322  /// \returns nullptr if doesn't have such group.
2323  InterleaveGroup<VPInstruction> *
2324  getInterleaveGroup(VPInstruction *Instr) const {
2325    return InterleaveGroupMap.lookup(Instr);
2326  }
2327};
2328
2329/// Class that maps (parts of) an existing VPlan to trees of combined
2330/// VPInstructions.
2331class VPlanSlp {
2332  enum class OpMode { Failed, Load, Opcode };
2333
2334  /// A DenseMapInfo implementation for using SmallVector<VPValue *, 4> as
2335  /// DenseMap keys.
2336  struct BundleDenseMapInfo {
2337    static SmallVector<VPValue *, 4> getEmptyKey() {
2338      return {reinterpret_cast<VPValue *>(-1)};
2339    }
2340
2341    static SmallVector<VPValue *, 4> getTombstoneKey() {
2342      return {reinterpret_cast<VPValue *>(-2)};
2343    }
2344
2345    static unsigned getHashValue(const SmallVector<VPValue *, 4> &V) {
2346      return static_cast<unsigned>(hash_combine_range(V.begin(), V.end()));
2347    }
2348
2349    static bool isEqual(const SmallVector<VPValue *, 4> &LHS,
2350                        const SmallVector<VPValue *, 4> &RHS) {
2351      return LHS == RHS;
2352    }
2353  };
2354
2355  /// Mapping of values in the original VPlan to a combined VPInstruction.
2356  DenseMap<SmallVector<VPValue *, 4>, VPInstruction *, BundleDenseMapInfo>
2357      BundleToCombined;
2358
2359  VPInterleavedAccessInfo &IAI;
2360
2361  /// Basic block to operate on. For now, only instructions in a single BB are
2362  /// considered.
2363  const VPBasicBlock &BB;
2364
2365  /// Indicates whether we managed to combine all visited instructions or not.
2366  bool CompletelySLP = true;
2367
2368  /// Width of the widest combined bundle in bits.
2369  unsigned WidestBundleBits = 0;
2370
2371  using MultiNodeOpTy =
2372      typename std::pair<VPInstruction *, SmallVector<VPValue *, 4>>;
2373
2374  // Input operand bundles for the current multi node. Each multi node operand
2375  // bundle contains values not matching the multi node's opcode. They will
2376  // be reordered in reorderMultiNodeOps, once we completed building a
2377  // multi node.
2378  SmallVector<MultiNodeOpTy, 4> MultiNodeOps;
2379
2380  /// Indicates whether we are building a multi node currently.
2381  bool MultiNodeActive = false;
2382
2383  /// Check if we can vectorize Operands together.
2384  bool areVectorizable(ArrayRef<VPValue *> Operands) const;
2385
2386  /// Add combined instruction \p New for the bundle \p Operands.
2387  void addCombined(ArrayRef<VPValue *> Operands, VPInstruction *New);
2388
2389  /// Indicate we hit a bundle we failed to combine. Returns nullptr for now.
2390  VPInstruction *markFailed();
2391
2392  /// Reorder operands in the multi node to maximize sequential memory access
2393  /// and commutative operations.
2394  SmallVector<MultiNodeOpTy, 4> reorderMultiNodeOps();
2395
2396  /// Choose the best candidate to use for the lane after \p Last. The set of
2397  /// candidates to choose from are values with an opcode matching \p Last's
2398  /// or loads consecutive to \p Last.
2399  std::pair<OpMode, VPValue *> getBest(OpMode Mode, VPValue *Last,
2400                                       SmallPtrSetImpl<VPValue *> &Candidates,
2401                                       VPInterleavedAccessInfo &IAI);
2402
2403#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
2404  /// Print bundle \p Values to dbgs().
2405  void dumpBundle(ArrayRef<VPValue *> Values);
2406#endif
2407
2408public:
2409  VPlanSlp(VPInterleavedAccessInfo &IAI, VPBasicBlock &BB) : IAI(IAI), BB(BB) {}
2410
2411  ~VPlanSlp() = default;
2412
2413  /// Tries to build an SLP tree rooted at \p Operands and returns a
2414  /// VPInstruction combining \p Operands, if they can be combined.
2415  VPInstruction *buildGraph(ArrayRef<VPValue *> Operands);
2416
2417  /// Return the width of the widest combined bundle in bits.
2418  unsigned getWidestBundleBits() const { return WidestBundleBits; }
2419
2420  /// Return true if all visited instruction can be combined.
2421  bool isCompletelySLP() const { return CompletelySLP; }
2422};
2423} // end namespace llvm
2424
2425#endif // LLVM_TRANSFORMS_VECTORIZE_VPLAN_H
2426