1/* Copyright (C) 2018-2020 Free Software Foundation, Inc. 2 3 This file is part of GCC. 4 5 GCC is free software; you can redistribute it and/or modify 6 it under the terms of the GNU General Public License as published by 7 the Free Software Foundation; either version 3, or (at your option) 8 any later version. 9 10 GCC is distributed in the hope that it will be useful, 11 but WITHOUT ANY WARRANTY; without even the implied warranty of 12 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 13 GNU General Public License for more details. 14 15 Under Section 7 of GPL version 3, you are granted additional 16 permissions described in the GCC Runtime Library Exception, version 17 3.1, as published by the Free Software Foundation. 18 19 You should have received a copy of the GNU General Public License and 20 a copy of the GCC Runtime Library Exception along with this program; 21 see the files COPYING3 and COPYING.RUNTIME respectively. If not, see 22 <http://www.gnu.org/licenses/>. */ 23 24/* Implemented from the specification included in the Intel C++ Compiler 25 User Guide and Reference, version 9.0. 26 27 NOTE: This is NOT a complete implementation of the SSE4 intrinsics! */ 28 29#ifndef NO_WARN_X86_INTRINSICS 30/* This header is distributed to simplify porting x86_64 code that 31 makes explicit use of Intel intrinsics to powerpc64le. 32 It is the user's responsibility to determine if the results are 33 acceptable and make additional changes as necessary. 34 Note that much code that uses Intel intrinsics can be rewritten in 35 standard C or GNU C extensions, which are more portable and better 36 optimized across multiple targets. */ 37#endif 38 39#ifndef SMMINTRIN_H_ 40#define SMMINTRIN_H_ 41 42#include <altivec.h> 43#include <tmmintrin.h> 44 45extern __inline int __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 46_mm_extract_epi8 (__m128i __X, const int __N) 47{ 48 return (unsigned char) ((__v16qi)__X)[__N & 15]; 49} 50 51extern __inline int __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 52_mm_extract_epi32 (__m128i __X, const int __N) 53{ 54 return ((__v4si)__X)[__N & 3]; 55} 56 57extern __inline int __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 58_mm_extract_epi64 (__m128i __X, const int __N) 59{ 60 return ((__v2di)__X)[__N & 1]; 61} 62 63extern __inline int __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 64_mm_extract_ps (__m128 __X, const int __N) 65{ 66 return ((__v4si)__X)[__N & 3]; 67} 68 69extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 70_mm_blend_epi16 (__m128i __A, __m128i __B, const int __imm8) 71{ 72 __v16qi __charmask = vec_splats ((signed char) __imm8); 73 __charmask = vec_gb (__charmask); 74 __v8hu __shortmask = (__v8hu) vec_unpackh (__charmask); 75 #ifdef __BIG_ENDIAN__ 76 __shortmask = vec_reve (__shortmask); 77 #endif 78 return (__m128i) vec_sel ((__v8hu) __A, (__v8hu) __B, __shortmask); 79} 80 81extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 82_mm_blendv_epi8 (__m128i __A, __m128i __B, __m128i __mask) 83{ 84 const __v16qu __seven = vec_splats ((unsigned char) 0x07); 85 __v16qu __lmask = vec_sra ((__v16qu) __mask, __seven); 86 return (__m128i) vec_sel ((__v16qu) __A, (__v16qu) __B, __lmask); 87} 88 89#endif 90