1/* 2 * This file is part of FFmpeg. 3 * 4 * FFmpeg is free software; you can redistribute it and/or 5 * modify it under the terms of the GNU Lesser General Public 6 * License as published by the Free Software Foundation; either 7 * version 2.1 of the License, or (at your option) any later version. 8 * 9 * FFmpeg is distributed in the hope that it will be useful, 10 * but WITHOUT ANY WARRANTY; without even the implied warranty of 11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 12 * Lesser General Public License for more details. 13 * 14 * You should have received a copy of the GNU Lesser General Public 15 * License along with FFmpeg; if not, write to the Free Software 16 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA 17 */ 18 19#include "config.h" 20#include "libavutil/attributes.h" 21#include "libavutil/cpu.h" 22#include "libavutil/x86/asm.h" 23#include "libavutil/x86/cpu.h" 24#include "libavcodec/mpegvideodsp.h" 25#include "libavcodec/videodsp.h" 26 27#if HAVE_INLINE_ASM 28 29static void gmc_mmx(uint8_t *dst, uint8_t *src, 30 int stride, int h, int ox, int oy, 31 int dxx, int dxy, int dyx, int dyy, 32 int shift, int r, int width, int height) 33{ 34 const int w = 8; 35 const int ix = ox >> (16 + shift); 36 const int iy = oy >> (16 + shift); 37 const int oxs = ox >> 4; 38 const int oys = oy >> 4; 39 const int dxxs = dxx >> 4; 40 const int dxys = dxy >> 4; 41 const int dyxs = dyx >> 4; 42 const int dyys = dyy >> 4; 43 const uint16_t r4[4] = { r, r, r, r }; 44 const uint16_t dxy4[4] = { dxys, dxys, dxys, dxys }; 45 const uint16_t dyy4[4] = { dyys, dyys, dyys, dyys }; 46 const uint64_t shift2 = 2 * shift; 47#define MAX_STRIDE 4096U 48#define MAX_H 8U 49 uint8_t edge_buf[(MAX_H + 1) * MAX_STRIDE]; 50 int x, y; 51 52 const int dxw = (dxx - (1 << (16 + shift))) * (w - 1); 53 const int dyh = (dyy - (1 << (16 + shift))) * (h - 1); 54 const int dxh = dxy * (h - 1); 55 const int dyw = dyx * (w - 1); 56 int need_emu = (unsigned) ix >= width - w || 57 (unsigned) iy >= height - h; 58 59 if ( // non-constant fullpel offset (3% of blocks) 60 ((ox ^ (ox + dxw)) | (ox ^ (ox + dxh)) | (ox ^ (ox + dxw + dxh)) | 61 (oy ^ (oy + dyw)) | (oy ^ (oy + dyh)) | (oy ^ (oy + dyw + dyh))) >> (16 + shift) || 62 // uses more than 16 bits of subpel mv (only at huge resolution) 63 (dxx | dxy | dyx | dyy) & 15 || 64 (need_emu && (h > MAX_H || stride > MAX_STRIDE))) { 65 // FIXME could still use mmx for some of the rows 66 ff_gmc_c(dst, src, stride, h, ox, oy, dxx, dxy, dyx, dyy, 67 shift, r, width, height); 68 return; 69 } 70 71 src += ix + iy * stride; 72 if (need_emu) { 73 ff_emulated_edge_mc_8(edge_buf, src, stride, stride, w + 1, h + 1, ix, iy, width, height); 74 src = edge_buf; 75 } 76 77 __asm__ volatile ( 78 "movd %0, %%mm6 \n\t" 79 "pxor %%mm7, %%mm7 \n\t" 80 "punpcklwd %%mm6, %%mm6 \n\t" 81 "punpcklwd %%mm6, %%mm6 \n\t" 82 :: "r" (1 << shift)); 83 84 for (x = 0; x < w; x += 4) { 85 uint16_t dx4[4] = { oxs - dxys + dxxs * (x + 0), 86 oxs - dxys + dxxs * (x + 1), 87 oxs - dxys + dxxs * (x + 2), 88 oxs - dxys + dxxs * (x + 3) }; 89 uint16_t dy4[4] = { oys - dyys + dyxs * (x + 0), 90 oys - dyys + dyxs * (x + 1), 91 oys - dyys + dyxs * (x + 2), 92 oys - dyys + dyxs * (x + 3) }; 93 94 for (y = 0; y < h; y++) { 95 __asm__ volatile ( 96 "movq %0, %%mm4 \n\t" 97 "movq %1, %%mm5 \n\t" 98 "paddw %2, %%mm4 \n\t" 99 "paddw %3, %%mm5 \n\t" 100 "movq %%mm4, %0 \n\t" 101 "movq %%mm5, %1 \n\t" 102 "psrlw $12, %%mm4 \n\t" 103 "psrlw $12, %%mm5 \n\t" 104 : "+m" (*dx4), "+m" (*dy4) 105 : "m" (*dxy4), "m" (*dyy4)); 106 107 __asm__ volatile ( 108 "movq %%mm6, %%mm2 \n\t" 109 "movq %%mm6, %%mm1 \n\t" 110 "psubw %%mm4, %%mm2 \n\t" 111 "psubw %%mm5, %%mm1 \n\t" 112 "movq %%mm2, %%mm0 \n\t" 113 "movq %%mm4, %%mm3 \n\t" 114 "pmullw %%mm1, %%mm0 \n\t" // (s - dx) * (s - dy) 115 "pmullw %%mm5, %%mm3 \n\t" // dx * dy 116 "pmullw %%mm5, %%mm2 \n\t" // (s - dx) * dy 117 "pmullw %%mm4, %%mm1 \n\t" // dx * (s - dy) 118 119 "movd %4, %%mm5 \n\t" 120 "movd %3, %%mm4 \n\t" 121 "punpcklbw %%mm7, %%mm5 \n\t" 122 "punpcklbw %%mm7, %%mm4 \n\t" 123 "pmullw %%mm5, %%mm3 \n\t" // src[1, 1] * dx * dy 124 "pmullw %%mm4, %%mm2 \n\t" // src[0, 1] * (s - dx) * dy 125 126 "movd %2, %%mm5 \n\t" 127 "movd %1, %%mm4 \n\t" 128 "punpcklbw %%mm7, %%mm5 \n\t" 129 "punpcklbw %%mm7, %%mm4 \n\t" 130 "pmullw %%mm5, %%mm1 \n\t" // src[1, 0] * dx * (s - dy) 131 "pmullw %%mm4, %%mm0 \n\t" // src[0, 0] * (s - dx) * (s - dy) 132 "paddw %5, %%mm1 \n\t" 133 "paddw %%mm3, %%mm2 \n\t" 134 "paddw %%mm1, %%mm0 \n\t" 135 "paddw %%mm2, %%mm0 \n\t" 136 137 "psrlw %6, %%mm0 \n\t" 138 "packuswb %%mm0, %%mm0 \n\t" 139 "movd %%mm0, %0 \n\t" 140 141 : "=m" (dst[x + y * stride]) 142 : "m" (src[0]), "m" (src[1]), 143 "m" (src[stride]), "m" (src[stride + 1]), 144 "m" (*r4), "m" (shift2)); 145 src += stride; 146 } 147 src += 4 - h * stride; 148 } 149} 150 151#endif /* HAVE_INLINE_ASM */ 152 153av_cold void ff_mpegvideodsp_init_x86(MpegVideoDSPContext *c) 154{ 155#if HAVE_INLINE_ASM 156 int cpu_flags = av_get_cpu_flags(); 157 158 if (INLINE_MMX(cpu_flags)) 159 c->gmc = gmc_mmx; 160#endif /* HAVE_INLINE_ASM */ 161} 162