Lines Matching refs:G1

397     __ mov(O0, G1);  // g1 = o0
404 __ ldsw(G1, 0, O0); // o0 = [g1]
408 __ ldx(G1, 0, O0); // o0 = [g1]
660 // Overwrites: G1,G2,G3
839 // With G1, don't generate the call if we statically know that the target in uninitialized
1016 const Register left_shift = G1; // left shift bit counter
1066 __ andcc(from, 7, G1); // misaligned bytes
1070 const Register left_shift = G1; // left shift bit counter
1073 __ sll(G1, LogBitsPerByte, left_shift);
1125 __ andcc(end_from, 7, G1); // misaligned bytes
1129 const Register left_shift = G1; // left shift bit counter
1132 __ sll(G1, LogBitsPerByte, left_shift);
1225 __ andcc(to, 7, G1); // misaligned bytes
1227 __ delayed()->neg(G1);
1228 __ inc(G1, 8); // bytes need to copy to next 8-bytes alignment
1229 __ sub(count, G1, count);
1232 __ deccc(G1);
1322 __ andcc(end_to, 7, G1); // misaligned bytes
1325 __ sub(count, G1, count);
1330 __ deccc(G1);
2373 __ mov(to, G1);
2375 gen_write_ref_array_pre_barrier(G1, G5, dest_uninitialized);
2383 gen_write_ref_array_post_barrier(G1, G5, O0);
2423 __ mov(to, G1);
2425 gen_write_ref_array_pre_barrier(G1, G5, dest_uninitialized);
2434 gen_write_ref_array_post_barrier(G1, G5, O0);
2499 const Register G1_remain = G1; // loop var, with stride -1
2515 assert_clean_int(O2_count, G1); // Make sure 'count' is clean int.
2521 __ mov(O3, G1); // spill: overlap test smashes O3
2526 __ mov(G1, O3);
2576 // Register G1 has number of *remaining* oops, O2 number of *total* oops.
2577 // Emit GC store barriers for the oops we have copied (O2 minus G1),
2618 const Register G1_bits = G1; // test copy of low bits
2625 inc_counter_np(SharedRuntime::_unsafe_array_copy_ctr, G1, G3);
2735 inc_counter_np(SharedRuntime::_generic_array_copy_ctr, G1, G3);
2738 //assert_clean_int(src_pos, G1);
2739 //assert_clean_int(dst_pos, G1);
2740 //assert_clean_int(length, G1);
3283 __ mov(from, G1);
3373 __ mov(G1, from);
3484 __ ldx(key,0,G1);
3534 __ xor3(G1,G3,G3);
3627 __ xor3(G1,G3,G3);
3720 __ xor3(G1,G3,G3);
3903 __ delayed()->mov(from, G1); // save original 'from' address before alignaddr
3920 __ mov(G1, from);
3923 __ xor3(L2,L4,G1);
3924 __ movxtod(G1,F60);
3925 __ xor3(L3,L5,G1);
3926 __ movxtod(G1,F62);
3966 __ andcc(to, 7, G1);
3978 __ sub(G4, G1, G4);
3982 __ mov(to, G1);
3992 __ mov(G1, to);
4014 __ delayed()->mov(from, G1); // save original 'from' address before alignaddr
4039 __ mov(G1, from);
4043 __ xor3(L2,G4,G1);
4044 __ movxtod(G1,F40);
4045 __ xor3(L3,G5,G1);
4046 __ movxtod(G1,F42);
4049 __ xor3(L2,L4,G1);
4050 __ movxtod(G1,F60);
4051 __ xor3(L3,L5,G1);
4052 __ movxtod(G1,F62);
4091 __ andcc(to, 7, G1);
4104 __ sub(G4, G1, G4);
4110 __ mov(to, G1);
4119 __ mov(G1, to);
4136 __ delayed()->mov(from, G1); // save original 'from' address before alignaddr
4161 __ mov(G1, from);
4165 __ xor3(L2,G4,G1);
4166 __ movxtod(G1,F48);
4167 __ xor3(L3,G5,G1);
4168 __ movxtod(G1,F50);
4171 __ xor3(L2,L4,G1);
4172 __ movxtod(G1,F60);
4173 __ xor3(L3,L5,G1);
4174 __ movxtod(G1,F62);
4207 __ andcc(to, 7, G1);
4220 __ sub(G4, G1, G4);
4226 __ mov(to, G1);
4235 __ mov(G1, to);
4252 __ delayed()->mov(from, G1); // save original 'from' address before alignaddr
4277 __ mov(G1, from);
4281 __ xor3(L2,G4,G1);
4282 __ movxtod(G1,F0);
4283 __ xor3(L3,G5,G1);
4284 __ movxtod(G1,F2);
4287 __ xor3(L2,L4,G1);
4288 __ movxtod(G1,F60);
4289 __ xor3(L3,L5,G1);
4290 __ movxtod(G1,F62);
4304 __ movdtox(F48,G1);
4332 __ movxtod(G1,F48);
4347 __ andcc(to, 7, G1);
4360 __ sub(G4, G1, G4);
4366 __ mov(to, G1);
4375 __ mov(G1, to);
4629 __ andcc(data, 7, G1);
4638 __ sll(G1, LogBitsPerByte, left_shift);
4643 __ sub(data, G1, data_ptr);
4675 __ xmulx(O0, O3, G1);
4685 __ xor3(G5, G1, G1);
4686 __ xor3(G1, O3, G1);
4688 __ srlx(G1, 63, G3);
4693 __ sllx(G1, 1, G1);
4694 __ or3(G1, O1, G1);
4696 __ xor3(G1, O2, G1);
4700 __ xmulxhi(G1, O0, O1);
4701 __ xmulx(G1, O0, O2);
4703 __ xmulx(G2, O0, G1);
4715 __ srlx(G1, 63, O2);