Lines Matching refs:mode1
3889 enum machine_mode mode1, rtx str_rtx,
3898 if (mode1 != VOIDmode
4062 enum machine_mode mode1;
4070 tem = get_inner_reference (to, &bitsize, &bitpos, &offset, &mode1,
4107 && (bitsize % GET_MODE_ALIGNMENT (mode1)) == 0
4108 && MEM_ALIGN (to_rtx) == GET_MODE_ALIGNMENT (mode1))
4110 to_rtx = adjust_address (to_rtx, mode1, bitpos / BITS_PER_UNIT);
4129 gcc_assert (bitpos == 0 || bitpos == GET_MODE_BITSIZE (mode1));
4151 if (optimize_bitfield_assignment_op (bitsize, bitpos, mode1,
4155 result = store_field (to_rtx, bitsize, bitpos, mode1, from,
6521 enum machine_mode mode1;
6599 &mode1, &unsignedp, &volatilep, false);
7458 enum machine_mode mode1;
7463 &mode1, &unsignedp, &volatilep, true);
7544 && (bitsize % GET_MODE_ALIGNMENT (mode1)) == 0
7545 && MEM_ALIGN (op0) == GET_MODE_ALIGNMENT (mode1))
7547 op0 = adjust_address (op0, mode1, bitpos / BITS_PER_UNIT);
7586 if (mode1 == VOIDmode
7588 || (mode1 != BLKmode && ! direct_load[(int) mode1]
7595 || (mode1 != BLKmode
7599 && (MEM_ALIGN (op0) < GET_MODE_ALIGNMENT (mode1)
7600 || (bitpos % GET_MODE_ALIGNMENT (mode1) != 0))))
7604 : SLOW_UNALIGNED_ACCESS (mode1, MEM_ALIGN (op0))))
7695 mode1 = BLKmode;
7700 op0 = adjust_address_nv (op0, mode1, bitpos / BITS_PER_UNIT);
7702 op0 = adjust_address (op0, mode1, bitpos / BITS_PER_UNIT);
7712 if (mode == mode1 || mode1 == BLKmode || mode1 == tmode
8522 enum machine_mode mode1 = GET_MODE (temp);
8523 if (mode1 == VOIDmode)
8524 mode1 = tmode != VOIDmode ? tmode : mode;
8526 temp = copy_to_mode_reg (mode1, temp);