Lines Matching refs:n2

186   if (fd.collapse > 1 && TREE_CODE (fd.loop.n2) != INTEGER_CST)
197 || !is_gimple_min_invariant (fd.loop.n2)
239 tree n1, n2;
243 n2 = fd.loop.n2;
255 n2 = OMP_CLAUSE_DECL (innerc);
263 t = fold_convert_loc (loc, long_integer_type_node, n2);
1603 tree e = loop->n2;
1809 && TREE_CODE (fd->loop.n2) != INTEGER_CST)
1854 fold_convert (itype, fd->loops[i].n2));
1879 if ((SSA_VAR_P (fd->loop.n2) || i >= fd->collapse)
1882 fold_convert (itype, fd->loops[i].n2)))
1886 tree n1, n2;
1890 n2 = fold_convert (itype, unshare_expr (fd->loops[i].n2));
1891 n2 = force_gimple_operand_gsi (gsi, n2, true, NULL_TREE,
1894 n1, n2);
1908 assign_stmt = gimple_build_assign (fd->loop.n2,
1937 fold_convert (itype, fd->loops[i].n2));
1963 if (SSA_VAR_P (fd->loop.n2) && i < fd->collapse)
1973 t = fold_build2 (MULT_EXPR, type, fd->loop.n2, counts[i]);
1974 expand_omp_build_assign (gsi, fd->loop.n2, t);
1977 if (fd->non_rect && SSA_VAR_P (fd->loop.n2))
2000 t = fold_convert (itype, unshare_expr (fd->loops[o].n2));
2045 tree n1, n2, n1e, n2e;
2057 t = fold_convert (itype, unshare_expr (fd->loops[i].n2));
2060 n2 = fold_convert (itype, unshare_expr (fd->loops[i].m2));
2061 n2 = fold_build2 (MULT_EXPR, itype, n1o, n2);
2062 n2 = fold_build2 (PLUS_EXPR, itype, n2, t);
2065 n2 = t;
2066 n2 = force_gimple_operand_gsi (&gsi2, n2, true, NULL_TREE,
2079 t = fold_convert (itype, unshare_expr (fd->loops[i].n2));
2092 n1, n2);
2110 t = fold_build2 (PLUS_EXPR, itype, t, n2);
2172 unshare_expr (fd->loops[i].n2));
2193 t = fold_convert (itype, unshare_expr (fd->loops[i].n2));
2196 n2 = fold_convert (itype, unshare_expr (fd->loops[i].m2));
2197 n2 = fold_build2 (MULT_EXPR, itype, tem, n2);
2198 n2 = fold_build2 (PLUS_EXPR, itype, n2, t);
2201 n2 = t;
2202 n2 = force_gimple_operand_gsi (&gsi2, n2, true, NULL_TREE,
2207 n1, n2);
2261 tree n1, n2;
2286 t = fold_convert (itype, unshare_expr (fd->loops[i].n2));
2288 n2 = t;
2293 unshare_expr (fd->loops[i].n2));
2294 n2 = fold_build_pointer_plus (vs[i - fd->loops[i].outer], t);
2298 n2 = fold_convert (itype, unshare_expr (fd->loops[i].m2));
2299 n2 = fold_build2 (MULT_EXPR, itype,
2300 vs[i - fd->loops[i].outer], n2);
2301 n2 = fold_build2 (PLUS_EXPR, itype, n2, t);
2303 n2 = force_gimple_operand_gsi (&gsi2, n2, true, NULL_TREE,
2311 n1, n2);
2324 fold_convert (itype, n2));
2365 expand_omp_build_cond (&gsi2, fd->loops[i].cond_code, vs[i], n2);
2407 t = fold_build2 (MULT_EXPR, type, fd->loop.n2,
2409 expand_omp_build_assign (gsi, fd->loop.n2, t);
2413 tree t = fd->loop.n2;
2482 /* If fd->loop.n2 is constant, then no propagation of the counts
2484 if (TREE_CODE (fd->loop.n2) == INTEGER_CST)
2555 && (TREE_CODE (fd->loop.n2) == INTEGER_CST
2559 && !integer_zerop (fd->loop.n2))
2780 tree n1 = NULL_TREE, n2 = NULL_TREE;
2814 t = fold_convert (itype, unshare_expr (fd->loops[j].n2));
2816 n2 = rect_p ? counts[j] : t;
2821 unshare_expr (fd->loops[j].n2));
2822 n2 = fold_build_pointer_plus (vs[j - fd->loops[j].outer], t);
2826 n2 = fold_convert (itype, unshare_expr (fd->loops[j].m2));
2827 n2 = fold_build2 (MULT_EXPR, itype,
2828 vs[j - fd->loops[j].outer], n2);
2829 n2 = fold_build2 (PLUS_EXPR, itype, n2, t);
2831 n2 = force_gimple_operand_gsi (&gsi2, n2, true, NULL_TREE,
2839 n1, n2);
2852 fold_convert (itype, n2));
2920 vs[j], n2, NULL_TREE, NULL_TREE);
3037 t = fold_convert (sizetype, unshare_expr (fd->loops[i].n2));
3048 unshare_expr (fd->loops[i].n2)));
3173 tree n1, n2;
3215 l->n2));
3222 t, unshare_expr (l->n2));
3224 n2 = force_gimple_operand_gsi (&gsi2, t, true, NULL_TREE,
3227 stmt = gimple_build_assign (nonrect_bounds[j], n2);
3229 n2 = nonrect_bounds[j];
3232 n2 = force_gimple_operand_gsi (&gsi2, unshare_expr (l->n2),
3236 = gimple_build_cond (l->cond_code, n1, n2,
3259 t = unshare_expr (fd->loops[i].n2);
3416 fd->loops[i].n2);
3419 fd->loops[i].n2);
3444 fd->loops[i].n2);
3448 fd->loops[i].n2);
3679 tree n2
3680 = force_gimple_operand_gsi (&gsi, fold_convert (type, fd->loops[i].n2),
3682 t = build2 (fd->loops[i].cond_code, boolean_type_node, v, n2);
3724 fold_convert (type, fd->loops[i].n2));
3974 counts[0] = fd->loop.n2;
3991 tree n1, n2;
3996 n2 = fold_build2 (PLUS_EXPR, type, fd->loop.n2, fd->loop.step);
4000 n1 = fold_build2 (MINUS_EXPR, type, fd->loop.n2, fd->loop.step);
4001 n2 = fd->loop.n1;
4004 || TREE_CODE (n2) != INTEGER_CST
4005 || ((tree_int_cst_sgn (n1) < 0) ^ (tree_int_cst_sgn (n2) < 0)))
4048 c = fd->loop.n2;
4068 t1 = fd->loop.n2;
4382 fold_convert (type, fd->loops[i].n2));
4864 tree n1, n2, step;
4910 fold_convert (type, fd->loop.n2));
4918 n2 = fold_convert (type, unshare_expr (fd->loop.n2));
4919 n2 = force_gimple_operand_gsi (&gsi, n2, true, NULL_TREE,
4922 n1, n2);
5066 n2 = fd->loop.n2;
5077 n2 = OMP_CLAUSE_DECL (innerc);
5081 n2 = force_gimple_operand_gsi (&gsi, fold_convert (itype, n2),
5088 t = fold_build2 (PLUS_EXPR, itype, t, n2);
5224 if (fd->collapse > 1 && TREE_CODE (fd->loop.n2) != INTEGER_CST
5233 fd->loop.n2);
5655 tree n1, n2, step;
5705 fold_convert (type, fd->loop.n2));
5713 n2 = fold_convert (type, unshare_expr (fd->loop.n2));
5714 n2 = force_gimple_operand_gsi (&gsi, n2, true, NULL_TREE,
5717 n1, n2);
5824 n2 = fd->loop.n2;
5835 n2 = OMP_CLAUSE_DECL (innerc);
5839 n2 = force_gimple_operand_gsi (&gsi, fold_convert (itype, n2),
5851 t = fold_build2 (PLUS_EXPR, itype, t, n2);
5930 if (fd->collapse > 1 && TREE_CODE (fd->loop.n2) != INTEGER_CST
5939 fd->loop.n2);
6363 tree n1, n2;
6427 n2 = fd->loop.n2;
6437 n2 = OMP_CLAUSE_DECL (innerc);
6506 i.e. fd->loop.v + MIN (n2 - fd->loop.v, cnt)
6519 fold_convert (itype, fd->loops[i].n2));
6540 tree t2 = fold_build2 (MINUS_EXPR, type, n2, n1);
6548 if (TREE_CODE (n2) == INTEGER_CST)
6552 if (integer_nonzerop (n2))
6554 fold_convert (type, n2));
6564 expand_omp_build_assign (&gsi, n2, build_one_cst (type));
6598 altn2 = n1 < n2 ? (n2 - n1 + step - 1) / step : 0;
6608 t = fold_build2 (PLUS_EXPR, itype, t, fold_convert (itype, n2));
6622 tree t2 = fold_convert (TREE_TYPE (fd->loop.v), n2);
6670 else if (TREE_CODE (n2) != INTEGER_CST)
6728 t = fold_convert (itype, unshare_expr (fd->loops[i].n2));
6742 t = fold_convert (type, unshare_expr (n2));
6854 t = fold_convert (itype, unshare_expr (fd->loops[i].n2));
6915 t = fold_build_pointer_plus (t2, fd->loops[i + 1].n2);
6919 fd->loops[i + 1].n2);
6928 /* For composite simd, n2 is the first iteration the current
6936 tree t2 = fold_build2 (MINUS_EXPR, type, n2, fd->loop.v);
6950 fd->loops[i + 1].n2));
7004 ? n2v : fd->loops[i + 1].n2));
7019 ? n2v : fd->loops[i + 1].n2);
7108 tree n1, n2;
7113 n2 = fold_build2 (PLUS_EXPR, type, fd->loop.n2, fd->loop.step);
7117 n1 = fold_build2 (MINUS_EXPR, type, fd->loop.n2, fd->loop.step);
7118 n2 = fd->loop.n1;
7121 || TREE_CODE (n2) != INTEGER_CST
7122 || ((tree_int_cst_sgn (n1) < 0) ^ (tree_int_cst_sgn (n2) < 0)))
7165 t1 = fd->loop.n2;
7194 if (fd->collapse > 1 && TREE_CODE (fd->loop.n2) != INTEGER_CST)
7201 tree t = force_gimple_operand_gsi (&gsi, fd->loop.n2, false,
7258 tree n1, n2, step;
7269 tree n1, n2;
7274 n2 = fold_build2 (PLUS_EXPR, type, fd->loop.n2, fd->loop.step);
7278 n1 = fold_build2 (MINUS_EXPR, type, fd->loop.n2, fd->loop.step);
7279 n2 = fd->loop.n1;
7282 || TREE_CODE (n2) != INTEGER_CST
7283 || ((tree_int_cst_sgn (n1) < 0) ^ (tree_int_cst_sgn (n2) < 0)))
7326 n2 = OMP_CLAUSE_DECL (innerc);
7330 n2 = fold_build2 (PLUS_EXPR, fd->iter_type, n2, bias);
7334 n2 = force_gimple_operand_gsi (&gsi, fold_convert (itype, n2),
7361 t = fold_convert (TREE_TYPE (startvar), n2);
7652 TREE_TYPE (fd->loop.n2), loc);
7654 if (SSA_VAR_P (fd->loop.n2))
7658 ass = gimple_build_assign (fd->loop.n2, total);
7664 tree e = fd->loop.n2;