1/* 2 * Copyright (c) 2000, 2014, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25#ifndef SHARE_VM_OPTO_OUTPUT_HPP 26#define SHARE_VM_OPTO_OUTPUT_HPP 27 28#include "opto/ad.hpp" 29#include "opto/block.hpp" 30#include "opto/node.hpp" 31 32class Arena; 33class Bundle; 34class Block; 35class Block_Array; 36class Node; 37class Node_Array; 38class Node_List; 39class PhaseCFG; 40class PhaseChaitin; 41class Pipeline_Use_Element; 42class Pipeline_Use; 43 44#ifndef PRODUCT 45#define DEBUG_ARG(x) , x 46#else 47#define DEBUG_ARG(x) 48#endif 49 50// Define the initial sizes for allocation of the resizable code buffer 51enum { 52 initial_code_capacity = 16 * 1024, 53 initial_stub_capacity = 4 * 1024, 54 initial_const_capacity = 4 * 1024, 55 initial_locs_capacity = 3 * 1024 56}; 57 58//------------------------------Scheduling---------------------------------- 59// This class contains all the information necessary to implement instruction 60// scheduling and bundling. 61class Scheduling { 62 63private: 64 // Arena to use 65 Arena *_arena; 66 67 // Control-Flow Graph info 68 PhaseCFG *_cfg; 69 70 // Register Allocation info 71 PhaseRegAlloc *_regalloc; 72 73 // Number of nodes in the method 74 uint _node_bundling_limit; 75 76 // List of scheduled nodes. Generated in reverse order 77 Node_List _scheduled; 78 79 // List of nodes currently available for choosing for scheduling 80 Node_List _available; 81 82 // For each instruction beginning a bundle, the number of following 83 // nodes to be bundled with it. 84 Bundle *_node_bundling_base; 85 86 // Mapping from register to Node 87 Node_List _reg_node; 88 89 // Free list for pinch nodes. 90 Node_List _pinch_free_list; 91 92 // Latency from the beginning of the containing basic block (base 1) 93 // for each node. 94 unsigned short *_node_latency; 95 96 // Number of uses of this node within the containing basic block. 97 short *_uses; 98 99 // Schedulable portion of current block. Skips Region/Phi/CreateEx up 100 // front, branch+proj at end. Also skips Catch/CProj (same as 101 // branch-at-end), plus just-prior exception-throwing call. 102 uint _bb_start, _bb_end; 103 104 // Latency from the end of the basic block as scheduled 105 unsigned short *_current_latency; 106 107 // Remember the next node 108 Node *_next_node; 109 110 // Use this for an unconditional branch delay slot 111 Node *_unconditional_delay_slot; 112 113 // Pointer to a Nop 114 MachNopNode *_nop; 115 116 // Length of the current bundle, in instructions 117 uint _bundle_instr_count; 118 119 // Current Cycle number, for computing latencies and bundling 120 uint _bundle_cycle_number; 121 122 // Bundle information 123 Pipeline_Use_Element _bundle_use_elements[resource_count]; 124 Pipeline_Use _bundle_use; 125 126 // Dump the available list 127 void dump_available() const; 128 129public: 130 Scheduling(Arena *arena, Compile &compile); 131 132 // Destructor 133 NOT_PRODUCT( ~Scheduling(); ) 134 135 // Step ahead "i" cycles 136 void step(uint i); 137 138 // Step ahead 1 cycle, and clear the bundle state (for example, 139 // at a branch target) 140 void step_and_clear(); 141 142 Bundle* node_bundling(const Node *n) { 143 assert(valid_bundle_info(n), "oob"); 144 return (&_node_bundling_base[n->_idx]); 145 } 146 147 bool valid_bundle_info(const Node *n) const { 148 return (_node_bundling_limit > n->_idx); 149 } 150 151 bool starts_bundle(const Node *n) const { 152 return (_node_bundling_limit > n->_idx && _node_bundling_base[n->_idx].starts_bundle()); 153 } 154 155 // Do the scheduling 156 void DoScheduling(); 157 158 // Compute the local latencies walking forward over the list of 159 // nodes for a basic block 160 void ComputeLocalLatenciesForward(const Block *bb); 161 162 // Compute the register antidependencies within a basic block 163 void ComputeRegisterAntidependencies(Block *bb); 164 void verify_do_def( Node *n, OptoReg::Name def, const char *msg ); 165 void verify_good_schedule( Block *b, const char *msg ); 166 void anti_do_def( Block *b, Node *def, OptoReg::Name def_reg, int is_def ); 167 void anti_do_use( Block *b, Node *use, OptoReg::Name use_reg ); 168 169 // Add a node to the current bundle 170 void AddNodeToBundle(Node *n, const Block *bb); 171 172 // Add a node to the list of available nodes 173 void AddNodeToAvailableList(Node *n); 174 175 // Compute the local use count for the nodes in a block, and compute 176 // the list of instructions with no uses in the block as available 177 void ComputeUseCount(const Block *bb); 178 179 // Choose an instruction from the available list to add to the bundle 180 Node * ChooseNodeToBundle(); 181 182 // See if this Node fits into the currently accumulating bundle 183 bool NodeFitsInBundle(Node *n); 184 185 // Decrement the use count for a node 186 void DecrementUseCounts(Node *n, const Block *bb); 187 188 // Garbage collect pinch nodes for reuse by other blocks. 189 void garbage_collect_pinch_nodes(); 190 // Clean up a pinch node for reuse (helper for above). 191 void cleanup_pinch( Node *pinch ); 192 193 // Information for statistics gathering 194#ifndef PRODUCT 195private: 196 // Gather information on size of nops relative to total 197 uint _branches, _unconditional_delays; 198 199 static uint _total_nop_size, _total_method_size; 200 static uint _total_branches, _total_unconditional_delays; 201 static uint _total_instructions_per_bundle[Pipeline::_max_instrs_per_cycle+1]; 202 203public: 204 static void print_statistics(); 205 206 static void increment_instructions_per_bundle(uint i) { 207 _total_instructions_per_bundle[i]++; 208 } 209 210 static void increment_nop_size(uint s) { 211 _total_nop_size += s; 212 } 213 214 static void increment_method_size(uint s) { 215 _total_method_size += s; 216 } 217#endif 218 219}; 220 221#endif // SHARE_VM_OPTO_OUTPUT_HPP 222