/netbsd-current/external/gpl3/gcc.old/dist/gcc/ |
H A D | tree-cfgcleanup.c | 94 basic_block default_bb = gimple_switch_default_bb (cfun, swtch); 95 basic_block case_bb = label_to_block (cfun, CASE_LABEL (label)); 271 if (DECL_CONTEXT (label) != cfun->decl) 273 target_block = label_to_block (cfun, label); 333 || single_succ (bb) == EXIT_BLOCK_PTR_FOR_FN (cfun) 340 gcc_checking_assert (bb != ENTRY_BLOCK_PTR_FOR_FN (cfun)); 350 if (e->src == ENTRY_BLOCK_PTR_FOR_FN (cfun) || (e->flags & EDGE_EH)) 695 set_ssa_default_def (cfun, new_var, lhs); 838 basic_block recv_bb = label_to_block (cfun, TREE_OPERAND (arg, 0)); 863 auto_vec<edge_iterator, 20> stack (n_basic_blocks_for_fn (cfun) [all...] |
H A D | cfgloop.c | 48 FOR_EACH_BB_FN (bb, cfun) 171 fprintf (file, ";; %d loops found\n", number_of_loops (cfun)); 407 if (latch != ENTRY_BLOCK_PTR_FOR_FN (cfun) 437 init_loops_structure (cfun, loops, 1); 445 if (n_basic_blocks_for_fn (cfun) == NUM_FIXED_BLOCKS) 449 loops->tree_root->num_nodes = n_basic_blocks_for_fn (cfun); 453 rc_order = XNEWVEC (int, n_basic_blocks_for_fn (cfun)); 459 for (b = 0; b < n_basic_blocks_for_fn (cfun) - NUM_FIXED_BLOCKS; b++) 461 basic_block header = BASIC_BLOCK_FOR_FN (cfun, rc_order[b]); 555 sort_sibling_loops_cmp_rpo = XNEWVEC (int, last_basic_block_for_fn (cfun)); [all...] |
H A D | except.c | 355 cfun->eh = ggc_cleared_alloc<eh_status> (); 358 vec_safe_push (cfun->eh->region_array, (eh_region)0); 359 vec_safe_push (cfun->eh->lp_array, (eh_landing_pad)0); 383 new_eh->next_peer = cfun->eh->region_tree; 384 cfun->eh->region_tree = new_eh; 387 new_eh->index = vec_safe_length (cfun->eh->region_array); 388 vec_safe_push (cfun->eh->region_array, new_eh); 468 lp->index = vec_safe_length (cfun->eh->lp_array); 471 vec_safe_push (cfun->eh->lp_array, lp); 485 return get_eh_region_from_number_fn (cfun, [all...] |
H A D | value-prof.c | 367 new_val = gimple_alloc_histogram_value (cfun, type, stmt); 401 gimple_add_histogram_value (cfun, stmt, new_val); 498 FOR_EACH_BB_FN (bb, cfun) 503 for (hist = gimple_histogram_value (cfun, stmt); hist; 517 if (VALUE_HISTOGRAMS (cfun)) 518 htab_traverse (VALUE_HISTOGRAMS (cfun), visit_hist, &visited_hists); 600 FOR_EACH_BB_FN (bb, cfun) 605 histogram_value th = gimple_histogram_value (cfun, stmt); 613 dump_histograms_for_stmt (cfun, dump_file, stmt); 797 histogram = gimple_histogram_value_of_type (cfun, stm [all...] |
H A D | store-motion.c | 541 if (cfun->can_throw_non_call_exceptions && may_trap_p (dest)) 639 FOR_EACH_BB_FN (bb, cfun) 790 if (!tmp && bb != EXIT_BLOCK_PTR_FOR_FN (cfun)) 829 auto_sbitmap visited (last_basic_block_for_fn (cfun)); 834 stack = XNEWVEC (edge_iterator, n_basic_blocks_for_fn (cfun)); 856 if (bb == EXIT_BLOCK_PTR_FOR_FN (cfun) 1011 st_avloc = sbitmap_vector_alloc (last_basic_block_for_fn (cfun), 1013 bitmap_vector_clear (st_avloc, last_basic_block_for_fn (cfun)); 1015 st_antloc = sbitmap_vector_alloc (last_basic_block_for_fn (cfun), 1017 bitmap_vector_clear (st_antloc, last_basic_block_for_fn (cfun)); [all...] |
H A D | tree-ssa-propagate.c | 186 if (bb == EXIT_BLOCK_PTR_FOR_FN (cfun)) 316 if (block == EXIT_BLOCK_PTR_FOR_FN (cfun)) 388 bb_to_cfg_order = XNEWVEC (int, last_basic_block_for_fn (cfun) + 1); 389 cfg_order_to_bb = XNEWVEC (int, n_basic_blocks_for_fn (cfun)); 390 int n = pre_and_rev_post_order_compute_fn (cfun, NULL, 401 set_gimple_stmt_max_uid (cfun, 0); 405 bb = BASIC_BLOCK_FOR_FN (cfun, cfg_order_to_bb[i]); 410 gimple_set_uid (stmt, inc_gimple_stmt_max_uid (cfun)); 416 gimple_set_uid (stmt, inc_gimple_stmt_max_uid (cfun)); 423 uid_to_stmt.safe_grow (gimple_stmt_max_uid (cfun)); 758 FOR_EACH_EDGE(e, ei, ENTRY_BLOCK_PTR_FOR_FN (cfun)->succs) argument [all...] |
H A D | gcse.c | 629 modify_mem_list = GCNEWVEC (vec_rtx_heap, last_basic_block_for_fn (cfun)); 631 last_basic_block_for_fn (cfun)); 683 bitmap_vector_ones (transp, last_basic_block_for_fn (cfun)); 687 bitmap_vector_clear (comp, last_basic_block_for_fn (cfun)); 689 bitmap_vector_clear (antloc, last_basic_block_for_fn (cfun)); 806 gcc_assert (!optimize_function_for_speed_p (cfun) 807 && optimize_function_for_size_p (cfun)); 1525 FOR_EACH_BB_FN (current_bb, cfun) 1774 FOR_EACH_BB_FN (bb, cfun) 1837 for (i = 0; i < (unsigned) n_edges_for_fn (cfun); [all...] |
H A D | gimplify-me.c | 59 push_gimplify_context (gimple_in_ssa_p (cfun), true); 65 if (gimple_in_ssa_p (cfun) && is_gimple_reg (var)) 160 push_gimplify_context (gimple_in_ssa_p (cfun)); 276 if (stmt_can_throw_internal (cfun, stmt)) 302 if (gimple_in_ssa_p (cfun)
|
H A D | tree-stdarg.c | 66 auto_sbitmap visited (last_basic_block_for_fn (cfun)); 96 gcc_assert (src != ENTRY_BLOCK_PTR_FOR_FN (cfun)); 132 counter_val = gpr_p ? cfun->va_list_gpr_size : cfun->va_list_fpr_size; 286 cfun->va_list_{g,f}pr_size. AP is a va_list GPR/FPR counter, 315 if (gpr_p && cfun->va_list_gpr_size + increment < VA_LIST_MAX_GPR_SIZE) 317 cfun->va_list_gpr_size += increment; 321 if (!gpr_p && cfun->va_list_fpr_size + increment < VA_LIST_MAX_FPR_SIZE) 323 cfun->va_list_fpr_size += increment; 331 cfun [all...] |
H A D | ipa-split.c | 225 if (e->src != ENTRY_BLOCK_PTR_FOR_FN (cfun) 236 if (e->src != ENTRY_BLOCK_PTR_FOR_FN (cfun) 300 FOR_EACH_BB_FN (bb, cfun) 398 BASIC_BLOCK_FOR_FN (cfun, dom_bb))) 454 < (ENTRY_BLOCK_PTR_FOR_FN (cfun)->count.apply_scale 463 && profile_status_for_fn (cfun) != PROFILE_READ 465 < ENTRY_BLOCK_PTR_FOR_FN (cfun)->count) 473 ENTRY_BLOCK_PTR_FOR_FN (cfun)->count.dump (dump_file); 538 tree ddef = ssa_default_def (cfun, parm); 651 if (return_bb != EXIT_BLOCK_PTR_FOR_FN (cfun) [all...] |
/netbsd-current/external/gpl3/gcc/dist/gcc/ |
H A D | tree-cfgcleanup.cc | 94 basic_block default_bb = gimple_switch_default_bb (cfun, swtch); 95 basic_block case_bb = label_to_block (cfun, CASE_LABEL (label)); 293 if (DECL_CONTEXT (label) != cfun->decl) 295 target_block = label_to_block (cfun, label); 355 || single_succ (bb) == EXIT_BLOCK_PTR_FOR_FN (cfun) 362 gcc_checking_assert (bb != ENTRY_BLOCK_PTR_FOR_FN (cfun)); 372 if (e->src == ENTRY_BLOCK_PTR_FOR_FN (cfun) || (e->flags & EDGE_EH)) 717 set_ssa_default_def (cfun, new_var, lhs); 860 basic_block recv_bb = label_to_block (cfun, TREE_OPERAND (arg, 0)); 885 auto_vec<edge_iterator, 20> stack (n_basic_blocks_for_fn (cfun) [all...] |
H A D | except.cc | 356 cfun->eh = ggc_cleared_alloc<eh_status> (); 359 vec_safe_push (cfun->eh->region_array, (eh_region)0); 360 vec_safe_push (cfun->eh->lp_array, (eh_landing_pad)0); 384 new_eh->next_peer = cfun->eh->region_tree; 385 cfun->eh->region_tree = new_eh; 388 new_eh->index = vec_safe_length (cfun->eh->region_array); 389 vec_safe_push (cfun->eh->region_array, new_eh); 469 lp->index = vec_safe_length (cfun->eh->lp_array); 472 vec_safe_push (cfun->eh->lp_array, lp); 486 return get_eh_region_from_number_fn (cfun, [all...] |
H A D | value-prof.cc | 370 new_val = gimple_alloc_histogram_value (cfun, type, stmt); 422 gimple_add_histogram_value (cfun, stmt, new_val); 519 FOR_EACH_BB_FN (bb, cfun) 524 for (hist = gimple_histogram_value (cfun, stmt); hist; 538 if (VALUE_HISTOGRAMS (cfun)) 539 htab_traverse (VALUE_HISTOGRAMS (cfun), visit_hist, &visited_hists); 621 FOR_EACH_BB_FN (bb, cfun) 626 histogram_value th = gimple_histogram_value (cfun, stmt); 634 dump_histograms_for_stmt (cfun, dump_file, stmt); 834 histogram = gimple_histogram_value_of_type (cfun, stm [all...] |
H A D | store-motion.cc | 537 if (cfun->can_throw_non_call_exceptions && may_trap_p (dest)) 635 FOR_EACH_BB_FN (bb, cfun) 786 if (!tmp && bb != EXIT_BLOCK_PTR_FOR_FN (cfun)) 825 auto_sbitmap visited (last_basic_block_for_fn (cfun)); 830 stack = XNEWVEC (edge_iterator, n_basic_blocks_for_fn (cfun)); 852 if (bb == EXIT_BLOCK_PTR_FOR_FN (cfun) 1007 st_avloc = sbitmap_vector_alloc (last_basic_block_for_fn (cfun), 1009 bitmap_vector_clear (st_avloc, last_basic_block_for_fn (cfun)); 1011 st_antloc = sbitmap_vector_alloc (last_basic_block_for_fn (cfun), 1013 bitmap_vector_clear (st_antloc, last_basic_block_for_fn (cfun)); [all...] |
H A D | gcse.cc | 629 modify_mem_list = GCNEWVEC (vec_rtx_heap, last_basic_block_for_fn (cfun)); 631 last_basic_block_for_fn (cfun)); 683 bitmap_vector_ones (transp, last_basic_block_for_fn (cfun)); 687 bitmap_vector_clear (comp, last_basic_block_for_fn (cfun)); 689 bitmap_vector_clear (antloc, last_basic_block_for_fn (cfun)); 806 gcc_assert (!optimize_function_for_speed_p (cfun) 807 && optimize_function_for_size_p (cfun)); 1524 FOR_EACH_BB_FN (current_bb, cfun) 1774 FOR_EACH_BB_FN (bb, cfun) 1837 for (i = 0; i < (unsigned) n_edges_for_fn (cfun); [all...] |
H A D | graphite.cc | 78 FOR_ALL_BB_FN (bb, cfun) 142 FOR_ALL_BB_FN (bb, cfun) 380 for (auto loop : loops_list (cfun, LI_FROM_INNERMOST)) 421 if (parallelized_function_p (cfun->decl)) 434 sort_sibling_loops (cfun); 486 mark_virtual_operands_for_renaming (cfun); 498 for (auto loop : loops_list (cfun, 0)) 513 profile_status_for_fn (cfun) = PROFILE_ABSENT; 514 release_recorded_exits (cfun);
|
H A D | tree-stdarg.cc | 66 auto_sbitmap visited (last_basic_block_for_fn (cfun)); 96 gcc_assert (src != ENTRY_BLOCK_PTR_FOR_FN (cfun)); 132 counter_val = gpr_p ? cfun->va_list_gpr_size : cfun->va_list_fpr_size; 286 cfun->va_list_{g,f}pr_size. AP is a va_list GPR/FPR counter, 315 if (gpr_p && cfun->va_list_gpr_size + increment < VA_LIST_MAX_GPR_SIZE) 317 cfun->va_list_gpr_size += increment; 321 if (!gpr_p && cfun->va_list_fpr_size + increment < VA_LIST_MAX_FPR_SIZE) 323 cfun->va_list_fpr_size += increment; 331 cfun [all...] |
H A D | gimplify-me.cc | 59 push_gimplify_context (gimple_in_ssa_p (cfun), true); 65 if (gimple_in_ssa_p (cfun) && is_gimple_reg (var)) 160 push_gimplify_context (gimple_in_ssa_p (cfun)); 272 if (stmt_can_throw_internal (cfun, stmt)) 298 if (gimple_in_ssa_p (cfun)
|
H A D | ipa-split.cc | 225 if (e->src != ENTRY_BLOCK_PTR_FOR_FN (cfun) 236 if (e->src != ENTRY_BLOCK_PTR_FOR_FN (cfun) 300 FOR_EACH_BB_FN (bb, cfun) 398 BASIC_BLOCK_FOR_FN (cfun, dom_bb))) 454 < (ENTRY_BLOCK_PTR_FOR_FN (cfun)->count.apply_scale 463 && profile_status_for_fn (cfun) != PROFILE_READ 465 < ENTRY_BLOCK_PTR_FOR_FN (cfun)->count) 473 ENTRY_BLOCK_PTR_FOR_FN (cfun)->count.dump (dump_file); 538 tree ddef = ssa_default_def (cfun, parm); 652 if (return_bb != EXIT_BLOCK_PTR_FOR_FN (cfun) [all...] |
/netbsd-current/external/gpl3/gcc.old/dist/gcc/config/pru/ |
H A D | pru.c | 127 gcc_assert (!cfun->machine->initialized); 129 save_mask = &cfun->machine->save_mask; 150 cfun->machine->total_size = total_size; 151 cfun->machine->var_size = var_size; 152 cfun->machine->out_args_size = out_args_size; 153 cfun->machine->save_reg_size = save_reg_size; 154 cfun->machine->initialized = reload_completed; 155 cfun->machine->save_regs_offset = out_args_size + var_size; 224 if (TEST_HARD_REG_BIT (cfun->machine->save_mask, regno_start)) 231 if (TEST_HARD_REG_BIT (cfun [all...] |
/netbsd-current/external/gpl3/gcc/dist/gcc/config/pru/ |
H A D | pru.cc | 127 gcc_assert (!cfun->machine->initialized); 129 save_mask = &cfun->machine->save_mask; 150 cfun->machine->total_size = total_size; 151 cfun->machine->var_size = var_size; 152 cfun->machine->out_args_size = out_args_size; 153 cfun->machine->save_reg_size = save_reg_size; 154 cfun->machine->initialized = reload_completed; 155 cfun->machine->save_regs_offset = out_args_size + var_size; 224 if (TEST_HARD_REG_BIT (cfun->machine->save_mask, regno_start)) 231 if (TEST_HARD_REG_BIT (cfun [all...] |
/netbsd-current/external/gpl3/gcc/dist/gcc/config/nvptx/ |
H A D | nvptx.cc | 752 if (!cfun || !cfun->machine->doing_call) 771 gcc_assert (cfun); 772 cfun->machine->return_mode = mode; 934 mode = (machine_mode) cfun->machine->return_mode; 941 cfun->machine->return_mode = VOIDmode; 1301 cfun->machine->has_softstack = true; 1313 if (strcmp (name, "x") == 0 && cfun->machine->red_partition) 1320 if (strcmp (name, "x") == 0 && cfun->machine->red_partition) 1326 REGNO (cfun [all...] |
/netbsd-current/external/gpl3/gcc.old/dist/gcc/config/gcn/ |
H A D | gcn.c | 1474 cfun->machine->args.reg[QUEUE_PTR_ARG]); 2073 if (!cfun || !cfun->machine) 2076 if (cfun->machine->normal_function) 2092 if (cfun->machine->args.requested == default_requested_args) 2106 if (cfun->machine->args.reg[PRIVATE_SEGMENT_WAVE_OFFSET_ARG] >= 0) 2107 fixed_regs[cfun->machine->args.reg[PRIVATE_SEGMENT_WAVE_OFFSET_ARG]] = 1; 2108 if (cfun->machine->args.reg[PRIVATE_SEGMENT_BUFFER_ARG] >= 0) 2112 fixed_regs[cfun->machine->args.reg[PRIVATE_SEGMENT_BUFFER_ARG]] = 1; 2113 fixed_regs[cfun [all...] |
/netbsd-current/external/gpl3/gcc.old/dist/gcc/config/nds32/ |
H A D | nds32-md-auxiliary.c | 1232 unsigned int regno = cfun->machine->callee_saved_first_fpr_regno; 1233 unsigned int last_fpr = cfun->machine->callee_saved_last_fpr_regno; 1255 unsigned int regno = cfun->machine->callee_saved_first_fpr_regno; 1256 unsigned int last_fpr = cfun->machine->callee_saved_last_fpr_regno; 1296 regno = cfun->machine->callee_saved_first_fpr_regno; 1297 while (regno <= cfun->machine->callee_saved_last_fpr_regno) 1987 int rb_va_args = cfun->machine->va_args_first_regno; 1988 int re_va_args = cfun->machine->va_args_last_regno; 1993 int rb_eh_data = cfun->machine->eh_return_data_first_regno; 1994 int re_eh_data = cfun [all...] |
/netbsd-current/external/gpl3/gcc/dist/gcc/c/ |
H A D | gimple-parser.cc | 223 cfun->pass_startwith = gimple_pass; 224 init_tree_ssa (cfun); 232 cfun->curr_properties = PROP_gimple_any; 238 parser.current_bb = ENTRY_BLOCK_PTR_FOR_FN (cfun); 241 set_loops_for_fn (cfun, ggc_cleared_alloc<struct loops> ()); 242 init_loops_structure (cfun, loops_for_fn (cfun), 1); 243 loops_state_set (cfun, LOOPS_NEED_FIXUP|LOOPS_MAY_HAVE_MULTIPLE_LATCHES); 244 cfun->curr_properties 248 init_ssa_operands (cfun); [all...] |