• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /netgear-R7000-V1.0.7.12_1.2.5/ap/gpl/amule/wxWidgets-2.8.12/src/png/

Lines Matching defs:movq

335                   movq       mm0,mask0
347 movq mm4,[esi]
349 movq mm6,mm0
352 movq [ebx],mm4
439 movq mm0,mask0
440 movq mm1,mask1
455 movq mm4,[esi]
457 movq mm6,mm0
458 movq mm7,[ebx]
461 movq [ebx],mm4
463 movq mm5,[esi+8]
465 movq mm7,mm1
466 movq mm6,[ebx+8]
469 movq [ebx+8],mm5
558 movq mm0,mask0
559 movq mm1,mask1
560 movq mm2,mask2
577 movq mm4,[esi]
579 movq mm6,mm0
580 movq mm7,[ebx]
583 movq [ebx],mm4
586 movq mm5,[esi+8]
588 movq mm7,mm1
589 movq mm6,[ebx+8]
592 movq [ebx+8],mm5
594 movq mm6,[esi+16]
596 movq mm4,mm2
597 movq mm7,[ebx+16]
600 movq [ebx+16],mm6
694 movq mm0,mask0
695 movq mm1,mask1
696 movq mm2,mask2
697 movq mm3,mask3
717 movq mm4,[esi]
719 movq mm6,mm0
720 movq mm7,[ebx]
723 movq [ebx],mm4
725 movq mm5,[esi+8]
727 movq mm7,mm1
728 movq mm6,[ebx+8]
731 movq [ebx+8],mm5
733 movq mm6,[esi+16]
735 movq mm4,mm2
736 movq mm7,[ebx+16]
739 movq [ebx+16],mm6
741 movq mm7,[esi+24]
743 movq mm5,mm3
744 movq mm4,[ebx+24]
747 movq [ebx+24],mm7
839 movq mm0,mask0
840 movq mm1,mask1
841 movq mm2,mask2
842 movq mm3,mask3
843 movq mm4,mask4
844 movq mm5,mask5
868 movq mm7,[esi]
870 movq mm6,mm0
873 movq [ebx],mm7
875 movq mm6,[esi+8]
877 movq mm7,mm1
880 movq [ebx+8],mm6
882 movq mm6,[esi+16]
884 movq mm7,mm2
887 movq [ebx+16],mm6
889 movq mm7,[esi+24]
891 movq mm6,mm3
894 movq [ebx+24],mm7
896 movq mm6,[esi+32]
898 movq mm7,mm4
901 movq [ebx+32],mm6
903 movq mm7,[esi+40]
905 movq mm6,mm5
908 movq [ebx+40],mm7
1237 movq mm1, mm0 ; 0 0 0 0 0 v2 v1 v0
1239 movq mm2, mm0 ; 0 0 0 v2 v1 v0 0 0
1244 movq mm3, mm0 ; v2 v1 v0 v2 v1 v0 v2 v1
1246 movq mm4, mm3 ; v2 v1 v0 v2 v1 v0 v2 v1
1248 movq [edi+16] , mm4
1250 movq [edi+8] , mm3
1253 movq [edi], mm0
1272 movq mm1, mm0 ; 0 0 0 0 0 v2 v1 v0
1274 movq mm2, mm0 ; 0 0 0 v2 v1 v0 0 0
1279 movq [edi+4], mm0 ; move to memory
1305 movq mm0, [esi] ; X X v2 v1 v0 v5 v4 v3
1306 movq mm7, mm0 ; X X v2 v1 v0 v5 v4 v3
1307 movq mm6, mm0 ; X X v2 v1 v0 v5 v4 v3
1312 movq mm5, mm6 ; 0 0 0 X X v2 v1 v0
1314 movq [edi], mm0 ; move quad to memory
1362 movq mm1, mm0 ; X X X X v0 v1 v2 v3
1364 movq mm2, mm0 ; v0 v0 v1 v1 v2 v2 v3 v3
1366 movq mm3, mm0 ; v2 v2 v2 v2 v3 v3 v3 v3
1369 movq [edi], mm0 ; move to memory v3
1371 movq [edi+8], mm3 ; move to memory v2
1372 movq mm4, mm2 ; v0 v0 v0 v0 v1 v1 v1 v1
1375 movq [edi+16], mm2 ; move to memory v1
1376 movq [edi+24], mm4 ; move to memory v0
1430 movq mm1, mm0 ; v0 v0 v1 v1 v2 v2 v3 v3
1433 movq [edi], mm0 ; move to memory v2 and v3
1435 movq [edi+8], mm1 ; move to memory v1 and v0
1470 movq mm0, [esi] ; v0 v1 v2 v3 v4 v5 v6 v7
1471 movq mm1, mm0 ; v0 v1 v2 v3 v4 v5 v6 v7
1473 /*movq mm1, mm0 ; v0 v0 v1 v1 v2 v2 v3 v3 */
1475 movq [edi+8], mm1 ; move to memory v0 v1 v2 and v3
1477 movq [edi], mm0 ; move to memory v4 v5 v6 and v7
1519 movq mm1, mm0 ; v1 v0 v1 v0 v3 v2 v3 v2
1522 movq [edi], mm0
1523 movq [edi + 8], mm0
1524 movq [edi + 16], mm1
1525 movq [edi + 24], mm1
1565 movq mm1, mm0 ; v1 v0 v1 v0 v3 v2 v3 v2
1568 movq [edi], mm0
1570 movq [edi + 8], mm1
1611 movq [edi], mm0
1652 movq mm0, [esi] ; v3 v2 v1 v0 v7 v6 v5 v4
1653 movq mm1, mm0 ; v3 v2 v1 v0 v7 v6 v5 v4
1656 movq [edi], mm0
1657 movq [edi + 8], mm0
1658 movq [edi + 16], mm0
1659 movq [edi + 24], mm0
1660 movq [edi+32], mm1
1661 movq [edi + 40], mm1
1662 movq [edi+ 48], mm1
1664 movq [edi + 56], mm1
1701 movq mm0, [esi] ; v3 v2 v1 v0 v7 v6 v5 v4
1702 movq mm1, mm0 ; v3 v2 v1 v0 v7 v6 v5 v4
1705 movq [edi], mm0
1706 movq [edi + 8], mm0
1707 movq [edi+16], mm1
1708 movq [edi + 24], mm1
1746 movq mm0, [esi] ; v3 v2 v1 v0 v7 v6 v5 v4
1747 movq mm1, mm0 ; v3 v2 v1 v0 v7 v6 v5 v4
1750 movq [edi], mm0
1752 movq [edi + 8], mm1
2000 movq mm7, ActiveMask
2002 movq mm5, LBCarryMask
2004 movq mm4, HBClearMask
2007 movq mm2, [edi + ebx - 8] /* Load previous aligned 8 bytes */
2010 movq mm0, [edi + ebx] /* Load mm0 with Avg(x) */
2012 movq mm3, mm5
2014 movq mm1, [esi + ebx] /* Load mm1 with Prior(x) */
2015 movq mm6, mm7
2021 movq mm1, mm3 /* now use mm1 for getting LBCarrys */
2032 movq mm2, mm0 /* mov updated Raws to mm2 */
2034 movq mm1, mm3 /* now use mm1 for getting LBCarrys */
2047 movq mm2, mm0 /* mov updated Raws to mm2 */
2051 movq mm1, mm3 /* now use mm1 for getting LBCarrys */
2063 movq [edi + ebx - 8], mm0
2066 movq mm2, mm0 /* mov updated Raw(x) to mm2 */
2082 movq mm4, HBClearMask
2086 movq mm7, ActiveMask
2090 movq mm6, mm7
2091 movq mm5, LBCarryMask
2094 movq mm2, [edi + ebx - 8] /* Load previous aligned 8 bytes */
2097 movq mm0, [edi + ebx]
2099 movq mm1, [esi + ebx]
2101 movq mm3, mm5
2107 movq mm1, mm3 /* now use mm1 for getting LBCarrys */
2117 movq mm2, mm0 /* mov updated Raws to mm2 */
2120 movq mm1, mm3 /* now use mm1 for getting LBCarrys */
2131 movq [edi + ebx - 8], mm0
2133 movq mm2, mm0 /* mov updated Raws to mm2 */
2145 movq mm7, ActiveMask
2148 movq mm5, LBCarryMask
2150 movq mm4, HBClearMask
2153 movq mm2, [edi + ebx - 8] /* Load previous aligned 8 bytes */
2156 movq mm0, [edi + ebx]
2158 movq mm1, [esi + ebx]
2160 movq mm3, mm5
2164 movq mm6, mm7
2167 movq mm1, mm3 /* now use mm1 for getting LBCarrys */
2177 movq mm2, mm0 /* mov updated Raws to mm2 */
2179 movq mm1, mm3 /* now use mm1 for getting LBCarrys */
2190 movq mm2, mm0 /* mov updated Raws to mm2 */
2194 movq mm1, mm3 /* now use mm1 for getting LBCarrys */
2205 movq mm2, mm0 /* mov updated Raws to mm2 */
2210 movq mm1, mm3 /* now use mm1 for getting LBCarrys */
2221 movq [edi + ebx - 8], mm0
2223 movq mm2, mm0 /* mov updated Raws to mm2 */
2265 movq mm5, LBCarryMask
2267 movq mm4, HBClearMask
2270 movq mm2, [edi + ebx - 8] /* Load previous aligned 8 bytes */
2273 movq mm0, [edi + ebx]
2274 movq mm3, mm5
2275 movq mm1, [esi + ebx]
2288 movq [edi + ebx - 8], mm0
2289 movq mm2, mm0 /* reuse as Raw(x-bpp) */
2297 movq mm5, LBCarryMask
2301 movq mm4, HBClearMask
2306 movq mm0, [edi + ebx]
2307 movq mm3, mm5
2308 movq mm1, [esi + ebx]
2310 movq mm2, [edx + ebx]
2322 movq [edi + ebx - 8], mm0
2488 movq mm1, [edi+ebx-8]
2491 movq mm2, [esi + ebx] /* load b=Prior(x) */
2493 movq mm3, [esi+ebx-8] /* Prep c=Prior(x-bpp) bytes */
2497 movq mm4, mm2
2500 movq mm5, mm1
2504 movq mm6, mm4
2524 movq mm7, mm4
2527 movq mm0, mm7
2544 movq mm3, [esi + ebx] /* load c=Prior(x-bpp) */
2546 movq mm2, mm3 /* load b=Prior(x) step 1 */
2549 movq [edi + ebx], mm7 /* write back updated value */
2550 movq mm1, mm7 /* Now mm1 will be used as Raw(x-bpp) */
2557 movq mm5, mm1
2559 movq mm4, mm2
2564 movq mm6, mm5
2583 movq mm7, mm4
2586 movq mm0, mm7
2597 movq mm2, [esi + ebx] /* load b=Prior(x) */
2604 movq mm3, mm2 /* load c=Prior(x-bpp) step 1 */
2609 movq mm4, mm2
2612 movq [edi + ebx], mm7 /* write back updated value */
2613 movq mm1, mm7
2622 movq mm5, mm1
2624 movq mm6, mm4
2645 movq mm7, mm4
2648 movq mm0, mm7
2671 movq [edi + ebx - 8], mm1 /* write back updated value */
2693 movq mm1, [edi+ebx-8]
2699 movq mm3, [esi+ebx-8] /* read c=Prior(x-bpp) bytes */
2701 movq mm2, [esi + ebx] /* load b=Prior(x) */
2706 movq mm4, mm2
2709 movq mm5, mm1
2713 movq mm6, mm4
2732 movq mm7, mm4
2735 movq mm0, mm7
2752 movq mm3, [esi + ebx - 8] /* load c=Prior(x-bpp) */
2755 movq mm2, [esi + ebx] /* load b=Prior(x) step 1 */
2757 movq mm6, mm2
2758 movq [edi + ebx], mm7 /* write back updated value */
2759 movq mm1, [edi+ebx-8]
2761 movq mm5, mm7
2771 movq mm4, mm2
2773 movq mm5, mm1
2777 movq mm6, mm4
2796 movq mm7, mm4
2799 movq mm0, mm7
2821 movq [edi + ebx - 8], mm1 /* write back updated value */
2837 movq mm1, [edi+ebx-8] /* Only time should need to read */
2841 movq mm3, [esi+ebx-8] /* read c=Prior(x-bpp) bytes */
2843 movq mm2, [esi + ebx] /* load b=Prior(x) */
2846 movq mm4, mm2
2849 movq mm5, mm1
2853 movq mm6, mm4
2872 movq mm7, mm4
2875 movq mm0, mm7
2892 movq mm3, [esi + ebx] /* load c=Prior(x-bpp) */
2894 movq mm2, mm3 /* load b=Prior(x) step 1 */
2897 movq [edi + ebx], mm7 /* write back updated value */
2898 movq mm1, mm7 /* Now mm1 will be used as Raw(x-bpp) */
2903 movq mm4, mm2
2905 movq mm5, mm1
2909 movq mm6, mm4
2928 movq mm7, mm4
2931 movq mm0, mm7
2953 movq [edi + ebx - 8], mm1 /* write back updated value */
2968 movq mm1, [edi+ebx-8] /* Only time should need to read */
2972 movq mm3, [esi+ebx-8] /* read c=Prior(x-bpp) bytes */
2974 movq mm2, [esi + ebx] /* load b=Prior(x) */
2977 movq mm4, mm2
2980 movq mm5, mm1
2984 movq mm6, mm4
3003 movq mm7, mm4
3006 movq mm0, mm7
3023 movq mm3, [esi+ebx-8] /* read c=Prior(x-bpp) bytes */
3025 movq mm2, [esi + ebx] /* load b=Prior(x) */
3028 movq [edi + ebx], mm7 /* write back updated value */
3029 movq mm1, [edi+ebx-8] /* read a=Raw(x-bpp) bytes */
3035 movq mm4, mm2
3037 movq mm5, mm1
3041 movq mm6, mm4
3060 movq mm7, mm4
3063 movq mm0, mm7
3085 movq [edi + ebx - 8], mm1 /* write back updated value */
3307 movq mm7, ActiveMask /* Load ActiveMask for 2nd active byte group */
3310 movq mm6, mm7
3315 movq mm1, [edi+ebx-8]
3320 movq mm0, [edi+ebx]
3323 movq mm1, mm0 /* mov updated Raws to mm1 */
3328 movq mm1, mm0 /* mov updated Raws to mm1 */
3334 movq [edi+ebx-8], mm0 /* Write updated Raws back to array */
3336 movq mm1, mm0
3388 movq mm1, [edi+ebx-8]
3392 movq mm0, [edi+ebx]
3395 movq mm1, mm0 /* mov updated Raws to mm1 */
3402 movq [edi+ebx-8], mm0
3403 movq mm1, mm0 /* Prep for doing 1st add at top of loop */
3415 movq mm7, ActiveMask /* Load ActiveMask for 2nd active byte group */
3417 movq mm6, mm7
3422 movq mm5, mm6
3427 movq mm1, [edi+ebx-8]
3433 movq mm0, [edi+ebx]
3436 movq mm1, mm0 /* mov updated Raws to mm1 */
3441 movq mm1, mm0 /* mov updated Raws to mm1 */
3446 movq mm1, mm0 /* mov updated Raws to mm1 */
3452 movq [edi+ebx-8], mm0 /* Write updated Raws back to array */
3453 movq mm1, mm0 /* Prep for doing 1st add at top of loop */
3466 movq mm7, [edi+ebx-8] /* PRIME the pump (load the first */
3470 movq mm0, [edi+ebx] /* Load Sub(x) for 1st 8 bytes */
3472 movq mm1, [edi+ebx+8] /* Load Sub(x) for 2nd 8 bytes */
3473 movq [edi+ebx], mm0 /* Write Raw(x) for 1st 8 bytes */
3480 movq mm2, [edi+ebx+16] /* Load Sub(x) for 3rd 8 bytes */
3481 movq [edi+ebx+8], mm1 /* Write Raw(x) for 2nd 8 bytes */
3483 movq mm3, [edi+ebx+24] /* Load Sub(x) for 4th 8 bytes */
3484 movq [edi+ebx+16], mm2 /* Write Raw(x) for 3rd 8 bytes */
3486 movq mm4, [edi+ebx+32] /* Load Sub(x) for 5th 8 bytes */
3487 movq [edi+ebx+24], mm3 /* Write Raw(x) for 4th 8 bytes */
3489 movq mm5, [edi+ebx+40] /* Load Sub(x) for 6th 8 bytes */
3490 movq [edi+ebx+32], mm4 /* Write Raw(x) for 5th 8 bytes */
3492 movq mm6, [edi+ebx+48] /* Load Sub(x) for 7th 8 bytes */
3493 movq [edi+ebx+40], mm5 /* Write Raw(x) for 6th 8 bytes */
3495 movq mm7, [edi+ebx+56] /* Load Sub(x) for 8th 8 bytes */
3496 movq [edi+ebx+48], mm6 /* Write Raw(x) for 7th 8 bytes */
3500 movq [edi+ebx-8], mm7 /* Write Raw(x) for 8th 8 bytes */
3505 movq mm0, [edi+ebx]
3509 movq [edi+ebx-8], mm0 /* use -8 to offset early add to ebx */
3510 movq mm7, mm0 /* Move calculated Raw(x) data to mm1 to */
3526 movq mm0, [edi+ebx]
3527 movq mm1, [esi+ebx]
3531 movq [edi+ebx-8], mm0 /* mov does not affect flags; -8 to offset */
3594 movq mm1, [esi+ebx]
3595 movq mm0, [edi+ebx]
3596 movq mm3, [esi+ebx+8]
3598 movq mm2, [edi+ebx+8]
3599 movq [edi+ebx], mm0
3601 movq mm5, [esi+ebx+16]
3602 movq [edi+ebx+8], mm2
3603 movq mm4, [edi+ebx+16]
3604 movq mm7, [esi+ebx+24]
3606 movq mm6, [edi+ebx+24]
3607 movq [edi+ebx+16], mm4
3609 movq mm1, [esi+ebx+32]
3610 movq [edi+ebx+24], mm6
3611 movq mm0, [edi+ebx+32]
3612 movq mm3, [esi+ebx+40]
3614 movq mm2, [edi+ebx+40]
3615 movq [edi+ebx+32], mm0
3617 movq mm5, [esi+ebx+48]
3618 movq [edi+ebx+40], mm2
3619 movq mm4, [edi+ebx+48]
3620 movq mm7, [esi+ebx+56]
3622 movq mm6, [edi+ebx+56]
3623 movq [edi+ebx+48], mm4
3627 movq [edi+ebx-8], mm6 /* (+56)movq does not affect flags; */
3647 movq mm1, [esi+ebx]
3648 movq mm0, [edi+ebx]
3652 movq [edi+ebx-8], mm0 /* movq does not affect flags; -8 to offset add ebx */