/linux-master/arch/sparc/include/asm/ |
H A D | sfp-machine_32.h | 78 #define __FP_FRAC_ADD_3(r2,r1,r0,x2,x1,x0,y2,y1,y0) \ 86 "rI" ((USItype)(y2)), \ 93 #define __FP_FRAC_SUB_3(r2,r1,r0,x2,x1,x0,y2,y1,y0) \ 101 "rI" ((USItype)(y2)), \ 108 #define __FP_FRAC_ADD_4(r3,r2,r1,r0,x3,x2,x1,x0,y3,y2,y1,y0) \ 123 "rI" ((USItype)(y2)), \ 133 #define __FP_FRAC_SUB_4(r3,r2,r1,r0,x3,x2,x1,x0,y3,y2,y1,y0) \ 148 "rI" ((USItype)(y2)), \ 158 #define __FP_FRAC_DEC_3(x2,x1,x0,y2,y1,y0) __FP_FRAC_SUB_3(x2,x1,x0,x2,x1,x0,y2,y [all...] |
/linux-master/drivers/gpu/drm/nouveau/dispnv04/ |
H A D | tvnv17.h | 117 static inline int interpolate(int y0, int y1, int y2, int x) argument 119 return y1 + (x < 50 ? y1 - y0 : y2 - y1) * (x - 50) / 50;
|
/linux-master/drivers/net/wireless/ath/carl9170/ |
H A D | phy.c | 1068 static s32 carl9170_interpolate_s32(s32 x, s32 x1, s32 y1, s32 x2, s32 y2) argument 1071 if (y2 == y1) 1078 return y2; 1084 return y1 + (((y2 - y1) * (x - x1)) / (x2 - x1)); 1087 static u8 carl9170_interpolate_u8(u8 x, u8 x1, u8 y1, u8 x2, u8 y2) argument 1093 y1 << SHIFT, x2 << SHIFT, y2 << SHIFT);
|
/linux-master/drivers/video/fbdev/nvidia/ |
H A D | nv_accel.c | 190 int x2, int y2) 193 int h = y2 - y1 + 1; 189 NVSetClippingRectangle(struct fb_info *info, int x1, int y1, int x2, int y2) argument
|
/linux-master/drivers/video/fbdev/ |
H A D | pxa168fb.h | 113 #define CFG_CKEY_Y2(y2) ((y2) << 24)
|
/linux-master/include/math-emu/ |
H A D | op-4.h | 516 #define __FP_FRAC_ADD_3(r2,r1,r0,x2,x1,x0,y2,y1,y0) \ 525 r2 = x2 + y2 + _c2; \ 530 #define __FP_FRAC_ADD_4(r3,r2,r1,r0,x3,x2,x1,x0,y3,y2,y1,y0) \ 539 r2 = x2 + y2; \ 548 #define __FP_FRAC_SUB_3(r2,r1,r0,x2,x1,x0,y2,y1,y0) \ 557 r2 = x2 - y2 - _c2; \ 562 #define __FP_FRAC_SUB_4(r3,r2,r1,r0,x3,x2,x1,x0,y3,y2,y1,y0) \ 571 r2 = x2 - y2; \ 580 #define __FP_FRAC_DEC_3(x2,x1,x0,y2,y1,y0) \ 584 __FP_FRAC_SUB_3 (x2, x1, x0, _t2, _t1, _t0, y2, y [all...] |
/linux-master/include/video/ |
H A D | udlfb.h | 17 int x2, y2; member in struct:dloarea
|
/linux-master/scripts/ |
H A D | bootgraph.pl | 177 my ($s, $s2, $s3, $e, $w, $y, $y2, $style); 191 $y2 = $y + 4; 205 print "<text transform=\"translate($s2,$y2) rotate(90)\">$key</text>\n"; 207 print "<text transform=\"translate($s3,$y2) rotate(90)\" font-size=\"3pt\">$key</text>\n";
|
/linux-master/arch/arm/crypto/ |
H A D | aes-neonbs-core.S | 151 .macro mul_gf4_n_gf4, x0, x1, y0, y1, t0, x2, x3, y2, y3, t1 153 veor \t1, \y2, \y3 159 vand \x3, \x3, \y2 169 y0, y1, y2, y3, t0, t1, t2, t3 173 veor \y0, \y0, \y2 175 mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x2, \x3, \y2, \y3, \t2 182 mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x6, \x7, \y2, \y3, \t2 183 veor \y0, \y0, \y2
|
/linux-master/arch/arm64/crypto/ |
H A D | aes-neonbs-core.S | 97 .macro mul_gf4_n_gf4, x0, x1, y0, y1, t0, x2, x3, y2, y3, t1 99 eor \t1, \y2, \y3 105 and \x3, \x3, \y2 115 y0, y1, y2, y3, t0, t1, t2, t3 119 eor \y0, \y0, \y2 121 mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x2, \x3, \y2, \y3, \t2 128 mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x6, \x7, \y2, \y3, \t2 129 eor \y0, \y0, \y2
|
/linux-master/arch/x86/crypto/ |
H A D | aria-aesni-avx-asm_64.S | 173 y0, y1, y2, y3, \ 186 vmovdqu (10 * 16)(rio), y2; \ 196 y0, y1, y2, y3, \ 201 y0, y1, y2, y3, \ 215 vmovdqu y2, 2 * 16(mem_cd); \ 224 y0, y1, y2, y3, \ 237 vmovdqu y2, 10 * 16(mem); \ 381 y0, y1, y2, y3, \ 386 vpxor y2, x6, x6; \ 392 vpxor y6, y2, y [all...] |
H A D | aria-aesni-avx2-asm_64.S | 189 y0, y1, y2, y3, \ 202 vmovdqu (10 * 32)(rio), y2; \ 212 y0, y1, y2, y3, \ 217 y0, y1, y2, y3, \ 231 vmovdqu y2, 2 * 32(mem_cd); \ 240 y0, y1, y2, y3, \ 253 vmovdqu y2, 10 * 32(mem); \ 422 y0, y1, y2, y3, \ 427 vpxor y2, x6, x6; \ 433 vpxor y6, y2, y [all...] |
H A D | aria-gfni-avx512-asm_64.S | 173 y0, y1, y2, y3, \ 186 vmovdqu64 (10 * 64)(rio), y2; \ 196 y0, y1, y2, y3, \ 201 y0, y1, y2, y3, \ 215 vmovdqu64 y2, 2 * 64(mem_cd); \ 224 y0, y1, y2, y3, \ 237 vmovdqu64 y2, 10 * 64(mem); \ 270 y0, y1, y2, y3, \ 295 vpxorq t0, y2, y2; \ [all...] |
H A D | camellia-aesni-avx-asm_64.S | 212 #define two_roundsm16(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 244 #define enc_rounds16(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 246 two_roundsm16(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 248 two_roundsm16(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 250 two_roundsm16(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 253 #define dec_rounds16(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 255 two_roundsm16(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 257 two_roundsm16(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 259 two_roundsm16(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 482 #define inpack16_pre(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y [all...] |
H A D | camellia-aesni-avx2-asm_64.S | 244 #define two_roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 276 #define enc_rounds32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 278 two_roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 280 two_roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 282 two_roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 285 #define dec_rounds32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 287 two_roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 289 two_roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 291 two_roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y3, y4, y5, \ 514 #define inpack32_pre(x0, x1, x2, x3, x4, x5, x6, x7, y0, y1, y2, y [all...] |
H A D | sha256-avx-asm.S | 116 y2 = %r15d define 164 mov f, y2 # y2 = f 167 xor g, y2 # y2 = f^g 170 and e, y2 # y2 = (f^g)&e 176 xor g, y2 # y2 = CH = ((f^g)&e)^g 178 add y0, y2 # y [all...] |
H A D | sha256-avx2-asm.S | 113 y2 = %r15d define 164 mov f, y2 # y2 = f # CH 168 xor g, y2 # y2 = f^g # CH 172 and e, y2 # y2 = (f^g)&e # CH 182 xor g, y2 # y2 = CH = ((f^g)&e)^g # CH 188 add y0, y2 # y [all...] |
H A D | sha256-ssse3-asm.S | 109 y2 = %r15d define 158 mov f, y2 # y2 = f 162 xor g, y2 # y2 = f^g 165 and e, y2 # y2 = (f^g)&e 171 xor g, y2 # y2 = CH = ((f^g)&e)^g 174 add y0, y2 # y [all...] |
H A D | sha512-avx2-asm.S | 98 y2 = %r15 define 187 mov f, y2 # y2 = f # CH 191 xor g, y2 # y2 = f^g # CH 194 and e, y2 # y2 = (f^g)&e # CH 203 xor g, y2 # y2 = CH = ((f^g)&e)^g # CH 208 add y0, y2 # y [all...] |
H A D | twofish-avx-x86_64-asm_64.S | 115 #define round_head_2(a, b, x1, y1, x2, y2) \ 135 G(RGI3, RGI4, y2, s1, s2, s3, s0); \ 136 vmovq RGS2, y2; \ 137 vpinsrq $1, RGS3, y2, y2;
|
/linux-master/crypto/ |
H A D | ecc.c | 1155 /* P = (x1, y1) => 2P, (x2, y2) => P' */ 1156 static void xycz_initial_double(u64 *x1, u64 *y1, u64 *x2, u64 *y2, argument 1163 vli_set(y2, y1, ndigits); 1175 apply_z(x2, y2, z, curve); 1178 /* Input P = (x1, y1, Z), Q = (x2, y2, Z) 1182 static void xycz_add(u64 *x1, u64 *y1, u64 *x2, u64 *y2, argument 1198 /* t4 = y2 - y1 */ 1199 vli_mod_sub(y2, y2, y1, curve_prime, ndigits); 1200 /* t5 = (y2 1225 xycz_add_c(u64 *x1, u64 *y1, u64 *x2, u64 *y2, const struct ecc_curve *curve) argument [all...] |
/linux-master/drivers/gpu/drm/amd/display/amdgpu_dm/ |
H A D | amdgpu_dm.c | 5322 clips->x2 - clips->x1, clips->y2 - clips->y1, 5372 clips->y2 - clips->y1, &i, false);
|
/linux-master/drivers/gpu/drm/armada/ |
H A D | armada_plane.c | 111 if ((new_plane_state->dst.y1 | new_plane_state->dst.y2) & 1)
|
/linux-master/drivers/gpu/drm/ |
H A D | drm_atomic.c | 691 clips->y1 >= clips->y2 || 695 clips->y2 > fb_height) { 699 clips->y1, clips->x2, clips->y2);
|
H A D | drm_atomic_helper.c | 909 drm_mode_get_hv_timing(&crtc_state->mode, &clip.x2, &clip.y2);
|