forked from OSchip/llvm-project
2782 lines
109 KiB
C
2782 lines
109 KiB
C
/*===---- avx512vlbwintrin.h - AVX512VL and AVX512BW intrinsics ------------===
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
* of this software and associated documentation files (the "Software"), to deal
|
|
* in the Software without restriction, including without limitation the rights
|
|
* to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
* copies of the Software, and to permit persons to whom the Software is
|
|
* furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
|
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
|
* THE SOFTWARE.
|
|
*
|
|
*===-----------------------------------------------------------------------===
|
|
*/
|
|
|
|
#ifndef __IMMINTRIN_H
|
|
#error "Never use <avx512vlbwintrin.h> directly; include <immintrin.h> instead."
|
|
#endif
|
|
|
|
#ifndef __AVX512VLBWINTRIN_H
|
|
#define __AVX512VLBWINTRIN_H
|
|
|
|
/* Define the default attributes for the functions in this file. */
|
|
#define __DEFAULT_FN_ATTRS __attribute__((__always_inline__, __nodebug__, __target__("avx512vl,avx512bw")))
|
|
|
|
static __inline __m128i __DEFAULT_FN_ATTRS
|
|
_mm_setzero_hi(void){
|
|
return (__m128i)(__v8hi){ 0, 0, 0, 0, 0, 0, 0, 0 };
|
|
}
|
|
|
|
/* Integer compare */
|
|
|
|
#define _mm_cmp_epi8_mask(a, b, p) __extension__ ({ \
|
|
(__mmask16)__builtin_ia32_cmpb128_mask((__v16qi)(__m128i)(a), \
|
|
(__v16qi)(__m128i)(b), (int)(p), \
|
|
(__mmask16)-1); })
|
|
|
|
#define _mm_mask_cmp_epi8_mask(m, a, b, p) __extension__ ({ \
|
|
(__mmask16)__builtin_ia32_cmpb128_mask((__v16qi)(__m128i)(a), \
|
|
(__v16qi)(__m128i)(b), (int)(p), \
|
|
(__mmask16)(m)); })
|
|
|
|
#define _mm_cmp_epu8_mask(a, b, p) __extension__ ({ \
|
|
(__mmask16)__builtin_ia32_ucmpb128_mask((__v16qi)(__m128i)(a), \
|
|
(__v16qi)(__m128i)(b), (int)(p), \
|
|
(__mmask16)-1); })
|
|
|
|
#define _mm_mask_cmp_epu8_mask(m, a, b, p) __extension__ ({ \
|
|
(__mmask16)__builtin_ia32_ucmpb128_mask((__v16qi)(__m128i)(a), \
|
|
(__v16qi)(__m128i)(b), (int)(p), \
|
|
(__mmask16)(m)); })
|
|
|
|
#define _mm256_cmp_epi8_mask(a, b, p) __extension__ ({ \
|
|
(__mmask32)__builtin_ia32_cmpb256_mask((__v32qi)(__m256i)(a), \
|
|
(__v32qi)(__m256i)(b), (int)(p), \
|
|
(__mmask32)-1); })
|
|
|
|
#define _mm256_mask_cmp_epi8_mask(m, a, b, p) __extension__ ({ \
|
|
(__mmask32)__builtin_ia32_cmpb256_mask((__v32qi)(__m256i)(a), \
|
|
(__v32qi)(__m256i)(b), (int)(p), \
|
|
(__mmask32)(m)); })
|
|
|
|
#define _mm256_cmp_epu8_mask(a, b, p) __extension__ ({ \
|
|
(__mmask32)__builtin_ia32_ucmpb256_mask((__v32qi)(__m256i)(a), \
|
|
(__v32qi)(__m256i)(b), (int)(p), \
|
|
(__mmask32)-1); })
|
|
|
|
#define _mm256_mask_cmp_epu8_mask(m, a, b, p) __extension__ ({ \
|
|
(__mmask32)__builtin_ia32_ucmpb256_mask((__v32qi)(__m256i)(a), \
|
|
(__v32qi)(__m256i)(b), (int)(p), \
|
|
(__mmask32)(m)); })
|
|
|
|
#define _mm_cmp_epi16_mask(a, b, p) __extension__ ({ \
|
|
(__mmask8)__builtin_ia32_cmpw128_mask((__v8hi)(__m128i)(a), \
|
|
(__v8hi)(__m128i)(b), (int)(p), \
|
|
(__mmask8)-1); })
|
|
|
|
#define _mm_mask_cmp_epi16_mask(m, a, b, p) __extension__ ({ \
|
|
(__mmask8)__builtin_ia32_cmpw128_mask((__v8hi)(__m128i)(a), \
|
|
(__v8hi)(__m128i)(b), (int)(p), \
|
|
(__mmask8)(m)); })
|
|
|
|
#define _mm_cmp_epu16_mask(a, b, p) __extension__ ({ \
|
|
(__mmask8)__builtin_ia32_ucmpw128_mask((__v8hi)(__m128i)(a), \
|
|
(__v8hi)(__m128i)(b), (int)(p), \
|
|
(__mmask8)-1); })
|
|
|
|
#define _mm_mask_cmp_epu16_mask(m, a, b, p) __extension__ ({ \
|
|
(__mmask8)__builtin_ia32_ucmpw128_mask((__v8hi)(__m128i)(a), \
|
|
(__v8hi)(__m128i)(b), (int)(p), \
|
|
(__mmask8)(m)); })
|
|
|
|
#define _mm256_cmp_epi16_mask(a, b, p) __extension__ ({ \
|
|
(__mmask16)__builtin_ia32_cmpw256_mask((__v16hi)(__m256i)(a), \
|
|
(__v16hi)(__m256i)(b), (int)(p), \
|
|
(__mmask16)-1); })
|
|
|
|
#define _mm256_mask_cmp_epi16_mask(m, a, b, p) __extension__ ({ \
|
|
(__mmask16)__builtin_ia32_cmpw256_mask((__v16hi)(__m256i)(a), \
|
|
(__v16hi)(__m256i)(b), (int)(p), \
|
|
(__mmask16)(m)); })
|
|
|
|
#define _mm256_cmp_epu16_mask(a, b, p) __extension__ ({ \
|
|
(__mmask16)__builtin_ia32_ucmpw256_mask((__v16hi)(__m256i)(a), \
|
|
(__v16hi)(__m256i)(b), (int)(p), \
|
|
(__mmask16)-1); })
|
|
|
|
#define _mm256_mask_cmp_epu16_mask(m, a, b, p) __extension__ ({ \
|
|
(__mmask16)__builtin_ia32_ucmpw256_mask((__v16hi)(__m256i)(a), \
|
|
(__v16hi)(__m256i)(b), (int)(p), \
|
|
(__mmask16)(m)); })
|
|
|
|
#define _mm_cmpeq_epi8_mask(A, B) \
|
|
_mm_cmp_epi8_mask((A), (B), _MM_CMPINT_EQ)
|
|
#define _mm_mask_cmpeq_epi8_mask(k, A, B) \
|
|
_mm_mask_cmp_epi8_mask((k), (A), (B), _MM_CMPINT_EQ)
|
|
#define _mm_cmpge_epi8_mask(A, B) \
|
|
_mm_cmp_epi8_mask((A), (B), _MM_CMPINT_GE)
|
|
#define _mm_mask_cmpge_epi8_mask(k, A, B) \
|
|
_mm_mask_cmp_epi8_mask((k), (A), (B), _MM_CMPINT_GE)
|
|
#define _mm_cmpgt_epi8_mask(A, B) \
|
|
_mm_cmp_epi8_mask((A), (B), _MM_CMPINT_GT)
|
|
#define _mm_mask_cmpgt_epi8_mask(k, A, B) \
|
|
_mm_mask_cmp_epi8_mask((k), (A), (B), _MM_CMPINT_GT)
|
|
#define _mm_cmple_epi8_mask(A, B) \
|
|
_mm_cmp_epi8_mask((A), (B), _MM_CMPINT_LE)
|
|
#define _mm_mask_cmple_epi8_mask(k, A, B) \
|
|
_mm_mask_cmp_epi8_mask((k), (A), (B), _MM_CMPINT_LE)
|
|
#define _mm_cmplt_epi8_mask(A, B) \
|
|
_mm_cmp_epi8_mask((A), (B), _MM_CMPINT_LT)
|
|
#define _mm_mask_cmplt_epi8_mask(k, A, B) \
|
|
_mm_mask_cmp_epi8_mask((k), (A), (B), _MM_CMPINT_LT)
|
|
#define _mm_cmpneq_epi8_mask(A, B) \
|
|
_mm_cmp_epi8_mask((A), (B), _MM_CMPINT_NE)
|
|
#define _mm_mask_cmpneq_epi8_mask(k, A, B) \
|
|
_mm_mask_cmp_epi8_mask((k), (A), (B), _MM_CMPINT_NE)
|
|
|
|
#define _mm256_cmpeq_epi8_mask(A, B) \
|
|
_mm256_cmp_epi8_mask((A), (B), _MM_CMPINT_EQ)
|
|
#define _mm256_mask_cmpeq_epi8_mask(k, A, B) \
|
|
_mm256_mask_cmp_epi8_mask((k), (A), (B), _MM_CMPINT_EQ)
|
|
#define _mm256_cmpge_epi8_mask(A, B) \
|
|
_mm256_cmp_epi8_mask((A), (B), _MM_CMPINT_GE)
|
|
#define _mm256_mask_cmpge_epi8_mask(k, A, B) \
|
|
_mm256_mask_cmp_epi8_mask((k), (A), (B), _MM_CMPINT_GE)
|
|
#define _mm256_cmpgt_epi8_mask(A, B) \
|
|
_mm256_cmp_epi8_mask((A), (B), _MM_CMPINT_GT)
|
|
#define _mm256_mask_cmpgt_epi8_mask(k, A, B) \
|
|
_mm256_mask_cmp_epi8_mask((k), (A), (B), _MM_CMPINT_GT)
|
|
#define _mm256_cmple_epi8_mask(A, B) \
|
|
_mm256_cmp_epi8_mask((A), (B), _MM_CMPINT_LE)
|
|
#define _mm256_mask_cmple_epi8_mask(k, A, B) \
|
|
_mm256_mask_cmp_epi8_mask((k), (A), (B), _MM_CMPINT_LE)
|
|
#define _mm256_cmplt_epi8_mask(A, B) \
|
|
_mm256_cmp_epi8_mask((A), (B), _MM_CMPINT_LT)
|
|
#define _mm256_mask_cmplt_epi8_mask(k, A, B) \
|
|
_mm256_mask_cmp_epi8_mask((k), (A), (B), _MM_CMPINT_LT)
|
|
#define _mm256_cmpneq_epi8_mask(A, B) \
|
|
_mm256_cmp_epi8_mask((A), (B), _MM_CMPINT_NE)
|
|
#define _mm256_mask_cmpneq_epi8_mask(k, A, B) \
|
|
_mm256_mask_cmp_epi8_mask((k), (A), (B), _MM_CMPINT_NE)
|
|
|
|
#define _mm_cmpeq_epu8_mask(A, B) \
|
|
_mm_cmp_epu8_mask((A), (B), _MM_CMPINT_EQ)
|
|
#define _mm_mask_cmpeq_epu8_mask(k, A, B) \
|
|
_mm_mask_cmp_epu8_mask((k), (A), (B), _MM_CMPINT_EQ)
|
|
#define _mm_cmpge_epu8_mask(A, B) \
|
|
_mm_cmp_epu8_mask((A), (B), _MM_CMPINT_GE)
|
|
#define _mm_mask_cmpge_epu8_mask(k, A, B) \
|
|
_mm_mask_cmp_epu8_mask((k), (A), (B), _MM_CMPINT_GE)
|
|
#define _mm_cmpgt_epu8_mask(A, B) \
|
|
_mm_cmp_epu8_mask((A), (B), _MM_CMPINT_GT)
|
|
#define _mm_mask_cmpgt_epu8_mask(k, A, B) \
|
|
_mm_mask_cmp_epu8_mask((k), (A), (B), _MM_CMPINT_GT)
|
|
#define _mm_cmple_epu8_mask(A, B) \
|
|
_mm_cmp_epu8_mask((A), (B), _MM_CMPINT_LE)
|
|
#define _mm_mask_cmple_epu8_mask(k, A, B) \
|
|
_mm_mask_cmp_epu8_mask((k), (A), (B), _MM_CMPINT_LE)
|
|
#define _mm_cmplt_epu8_mask(A, B) \
|
|
_mm_cmp_epu8_mask((A), (B), _MM_CMPINT_LT)
|
|
#define _mm_mask_cmplt_epu8_mask(k, A, B) \
|
|
_mm_mask_cmp_epu8_mask((k), (A), (B), _MM_CMPINT_LT)
|
|
#define _mm_cmpneq_epu8_mask(A, B) \
|
|
_mm_cmp_epu8_mask((A), (B), _MM_CMPINT_NE)
|
|
#define _mm_mask_cmpneq_epu8_mask(k, A, B) \
|
|
_mm_mask_cmp_epu8_mask((k), (A), (B), _MM_CMPINT_NE)
|
|
|
|
#define _mm256_cmpeq_epu8_mask(A, B) \
|
|
_mm256_cmp_epu8_mask((A), (B), _MM_CMPINT_EQ)
|
|
#define _mm256_mask_cmpeq_epu8_mask(k, A, B) \
|
|
_mm256_mask_cmp_epu8_mask((k), (A), (B), _MM_CMPINT_EQ)
|
|
#define _mm256_cmpge_epu8_mask(A, B) \
|
|
_mm256_cmp_epu8_mask((A), (B), _MM_CMPINT_GE)
|
|
#define _mm256_mask_cmpge_epu8_mask(k, A, B) \
|
|
_mm256_mask_cmp_epu8_mask((k), (A), (B), _MM_CMPINT_GE)
|
|
#define _mm256_cmpgt_epu8_mask(A, B) \
|
|
_mm256_cmp_epu8_mask((A), (B), _MM_CMPINT_GT)
|
|
#define _mm256_mask_cmpgt_epu8_mask(k, A, B) \
|
|
_mm256_mask_cmp_epu8_mask((k), (A), (B), _MM_CMPINT_GT)
|
|
#define _mm256_cmple_epu8_mask(A, B) \
|
|
_mm256_cmp_epu8_mask((A), (B), _MM_CMPINT_LE)
|
|
#define _mm256_mask_cmple_epu8_mask(k, A, B) \
|
|
_mm256_mask_cmp_epu8_mask((k), (A), (B), _MM_CMPINT_LE)
|
|
#define _mm256_cmplt_epu8_mask(A, B) \
|
|
_mm256_cmp_epu8_mask((A), (B), _MM_CMPINT_LT)
|
|
#define _mm256_mask_cmplt_epu8_mask(k, A, B) \
|
|
_mm256_mask_cmp_epu8_mask((k), (A), (B), _MM_CMPINT_LT)
|
|
#define _mm256_cmpneq_epu8_mask(A, B) \
|
|
_mm256_cmp_epu8_mask((A), (B), _MM_CMPINT_NE)
|
|
#define _mm256_mask_cmpneq_epu8_mask(k, A, B) \
|
|
_mm256_mask_cmp_epu8_mask((k), (A), (B), _MM_CMPINT_NE)
|
|
|
|
#define _mm_cmpeq_epi16_mask(A, B) \
|
|
_mm_cmp_epi16_mask((A), (B), _MM_CMPINT_EQ)
|
|
#define _mm_mask_cmpeq_epi16_mask(k, A, B) \
|
|
_mm_mask_cmp_epi16_mask((k), (A), (B), _MM_CMPINT_EQ)
|
|
#define _mm_cmpge_epi16_mask(A, B) \
|
|
_mm_cmp_epi16_mask((A), (B), _MM_CMPINT_GE)
|
|
#define _mm_mask_cmpge_epi16_mask(k, A, B) \
|
|
_mm_mask_cmp_epi16_mask((k), (A), (B), _MM_CMPINT_GE)
|
|
#define _mm_cmpgt_epi16_mask(A, B) \
|
|
_mm_cmp_epi16_mask((A), (B), _MM_CMPINT_GT)
|
|
#define _mm_mask_cmpgt_epi16_mask(k, A, B) \
|
|
_mm_mask_cmp_epi16_mask((k), (A), (B), _MM_CMPINT_GT)
|
|
#define _mm_cmple_epi16_mask(A, B) \
|
|
_mm_cmp_epi16_mask((A), (B), _MM_CMPINT_LE)
|
|
#define _mm_mask_cmple_epi16_mask(k, A, B) \
|
|
_mm_mask_cmp_epi16_mask((k), (A), (B), _MM_CMPINT_LE)
|
|
#define _mm_cmplt_epi16_mask(A, B) \
|
|
_mm_cmp_epi16_mask((A), (B), _MM_CMPINT_LT)
|
|
#define _mm_mask_cmplt_epi16_mask(k, A, B) \
|
|
_mm_mask_cmp_epi16_mask((k), (A), (B), _MM_CMPINT_LT)
|
|
#define _mm_cmpneq_epi16_mask(A, B) \
|
|
_mm_cmp_epi16_mask((A), (B), _MM_CMPINT_NE)
|
|
#define _mm_mask_cmpneq_epi16_mask(k, A, B) \
|
|
_mm_mask_cmp_epi16_mask((k), (A), (B), _MM_CMPINT_NE)
|
|
|
|
#define _mm256_cmpeq_epi16_mask(A, B) \
|
|
_mm256_cmp_epi16_mask((A), (B), _MM_CMPINT_EQ)
|
|
#define _mm256_mask_cmpeq_epi16_mask(k, A, B) \
|
|
_mm256_mask_cmp_epi16_mask((k), (A), (B), _MM_CMPINT_EQ)
|
|
#define _mm256_cmpge_epi16_mask(A, B) \
|
|
_mm256_cmp_epi16_mask((A), (B), _MM_CMPINT_GE)
|
|
#define _mm256_mask_cmpge_epi16_mask(k, A, B) \
|
|
_mm256_mask_cmp_epi16_mask((k), (A), (B), _MM_CMPINT_GE)
|
|
#define _mm256_cmpgt_epi16_mask(A, B) \
|
|
_mm256_cmp_epi16_mask((A), (B), _MM_CMPINT_GT)
|
|
#define _mm256_mask_cmpgt_epi16_mask(k, A, B) \
|
|
_mm256_mask_cmp_epi16_mask((k), (A), (B), _MM_CMPINT_GT)
|
|
#define _mm256_cmple_epi16_mask(A, B) \
|
|
_mm256_cmp_epi16_mask((A), (B), _MM_CMPINT_LE)
|
|
#define _mm256_mask_cmple_epi16_mask(k, A, B) \
|
|
_mm256_mask_cmp_epi16_mask((k), (A), (B), _MM_CMPINT_LE)
|
|
#define _mm256_cmplt_epi16_mask(A, B) \
|
|
_mm256_cmp_epi16_mask((A), (B), _MM_CMPINT_LT)
|
|
#define _mm256_mask_cmplt_epi16_mask(k, A, B) \
|
|
_mm256_mask_cmp_epi16_mask((k), (A), (B), _MM_CMPINT_LT)
|
|
#define _mm256_cmpneq_epi16_mask(A, B) \
|
|
_mm256_cmp_epi16_mask((A), (B), _MM_CMPINT_NE)
|
|
#define _mm256_mask_cmpneq_epi16_mask(k, A, B) \
|
|
_mm256_mask_cmp_epi16_mask((k), (A), (B), _MM_CMPINT_NE)
|
|
|
|
#define _mm_cmpeq_epu16_mask(A, B) \
|
|
_mm_cmp_epu16_mask((A), (B), _MM_CMPINT_EQ)
|
|
#define _mm_mask_cmpeq_epu16_mask(k, A, B) \
|
|
_mm_mask_cmp_epu16_mask((k), (A), (B), _MM_CMPINT_EQ)
|
|
#define _mm_cmpge_epu16_mask(A, B) \
|
|
_mm_cmp_epu16_mask((A), (B), _MM_CMPINT_GE)
|
|
#define _mm_mask_cmpge_epu16_mask(k, A, B) \
|
|
_mm_mask_cmp_epu16_mask((k), (A), (B), _MM_CMPINT_GE)
|
|
#define _mm_cmpgt_epu16_mask(A, B) \
|
|
_mm_cmp_epu16_mask((A), (B), _MM_CMPINT_GT)
|
|
#define _mm_mask_cmpgt_epu16_mask(k, A, B) \
|
|
_mm_mask_cmp_epu16_mask((k), (A), (B), _MM_CMPINT_GT)
|
|
#define _mm_cmple_epu16_mask(A, B) \
|
|
_mm_cmp_epu16_mask((A), (B), _MM_CMPINT_LE)
|
|
#define _mm_mask_cmple_epu16_mask(k, A, B) \
|
|
_mm_mask_cmp_epu16_mask((k), (A), (B), _MM_CMPINT_LE)
|
|
#define _mm_cmplt_epu16_mask(A, B) \
|
|
_mm_cmp_epu16_mask((A), (B), _MM_CMPINT_LT)
|
|
#define _mm_mask_cmplt_epu16_mask(k, A, B) \
|
|
_mm_mask_cmp_epu16_mask((k), (A), (B), _MM_CMPINT_LT)
|
|
#define _mm_cmpneq_epu16_mask(A, B) \
|
|
_mm_cmp_epu16_mask((A), (B), _MM_CMPINT_NE)
|
|
#define _mm_mask_cmpneq_epu16_mask(k, A, B) \
|
|
_mm_mask_cmp_epu16_mask((k), (A), (B), _MM_CMPINT_NE)
|
|
|
|
#define _mm256_cmpeq_epu16_mask(A, B) \
|
|
_mm256_cmp_epu16_mask((A), (B), _MM_CMPINT_EQ)
|
|
#define _mm256_mask_cmpeq_epu16_mask(k, A, B) \
|
|
_mm256_mask_cmp_epu16_mask((k), (A), (B), _MM_CMPINT_EQ)
|
|
#define _mm256_cmpge_epu16_mask(A, B) \
|
|
_mm256_cmp_epu16_mask((A), (B), _MM_CMPINT_GE)
|
|
#define _mm256_mask_cmpge_epu16_mask(k, A, B) \
|
|
_mm256_mask_cmp_epu16_mask((k), (A), (B), _MM_CMPINT_GE)
|
|
#define _mm256_cmpgt_epu16_mask(A, B) \
|
|
_mm256_cmp_epu16_mask((A), (B), _MM_CMPINT_GT)
|
|
#define _mm256_mask_cmpgt_epu16_mask(k, A, B) \
|
|
_mm256_mask_cmp_epu16_mask((k), (A), (B), _MM_CMPINT_GT)
|
|
#define _mm256_cmple_epu16_mask(A, B) \
|
|
_mm256_cmp_epu16_mask((A), (B), _MM_CMPINT_LE)
|
|
#define _mm256_mask_cmple_epu16_mask(k, A, B) \
|
|
_mm256_mask_cmp_epu16_mask((k), (A), (B), _MM_CMPINT_LE)
|
|
#define _mm256_cmplt_epu16_mask(A, B) \
|
|
_mm256_cmp_epu16_mask((A), (B), _MM_CMPINT_LT)
|
|
#define _mm256_mask_cmplt_epu16_mask(k, A, B) \
|
|
_mm256_mask_cmp_epu16_mask((k), (A), (B), _MM_CMPINT_LT)
|
|
#define _mm256_cmpneq_epu16_mask(A, B) \
|
|
_mm256_cmp_epu16_mask((A), (B), _MM_CMPINT_NE)
|
|
#define _mm256_mask_cmpneq_epu16_mask(k, A, B) \
|
|
_mm256_mask_cmp_epu16_mask((k), (A), (B), _MM_CMPINT_NE)
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_add_epi8(__m256i __W, __mmask32 __U, __m256i __A, __m256i __B){
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_add_epi8(__A, __B),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_add_epi8(__mmask32 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_add_epi8(__A, __B),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_add_epi16(__m256i __W, __mmask16 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_add_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_add_epi16(__mmask16 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_add_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_sub_epi8(__m256i __W, __mmask32 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_sub_epi8(__A, __B),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_sub_epi8(__mmask32 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_sub_epi8(__A, __B),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_sub_epi16(__m256i __W, __mmask16 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_sub_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_sub_epi16(__mmask16 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_sub_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_add_epi8(__m128i __W, __mmask16 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_add_epi8(__A, __B),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_add_epi8(__mmask16 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_add_epi8(__A, __B),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_add_epi16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_add_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_add_epi16(__mmask8 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_add_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_sub_epi8(__m128i __W, __mmask16 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_sub_epi8(__A, __B),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_sub_epi8(__mmask16 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_sub_epi8(__A, __B),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_sub_epi16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_sub_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_sub_epi16(__mmask8 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_sub_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_mullo_epi16(__m256i __W, __mmask16 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_mullo_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_mullo_epi16(__mmask16 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_mullo_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_mullo_epi16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_mullo_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_mullo_epi16(__mmask8 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_mullo_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_blend_epi8 (__mmask16 __U, __m128i __A, __m128i __W)
|
|
{
|
|
return (__m128i) __builtin_ia32_selectb_128 ((__mmask16) __U,
|
|
(__v16qi) __W,
|
|
(__v16qi) __A);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_blend_epi8 (__mmask32 __U, __m256i __A, __m256i __W)
|
|
{
|
|
return (__m256i) __builtin_ia32_selectb_256 ((__mmask32) __U,
|
|
(__v32qi) __W,
|
|
(__v32qi) __A);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_blend_epi16 (__mmask8 __U, __m128i __A, __m128i __W)
|
|
{
|
|
return (__m128i) __builtin_ia32_selectw_128 ((__mmask8) __U,
|
|
(__v8hi) __W,
|
|
(__v8hi) __A);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_blend_epi16 (__mmask16 __U, __m256i __A, __m256i __W)
|
|
{
|
|
return (__m256i) __builtin_ia32_selectw_256 ((__mmask16) __U,
|
|
(__v16hi) __W,
|
|
(__v16hi) __A);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_abs_epi8(__m128i __W, __mmask16 __U, __m128i __A)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_abs_epi8(__A),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_abs_epi8(__mmask16 __U, __m128i __A)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_abs_epi8(__A),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_abs_epi8(__m256i __W, __mmask32 __U, __m256i __A)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_abs_epi8(__A),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_abs_epi8 (__mmask32 __U, __m256i __A)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_abs_epi8(__A),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_abs_epi16(__m128i __W, __mmask8 __U, __m128i __A)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_abs_epi16(__A),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_abs_epi16(__mmask8 __U, __m128i __A)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_abs_epi16(__A),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_abs_epi16(__m256i __W, __mmask16 __U, __m256i __A)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_abs_epi16(__A),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_abs_epi16(__mmask16 __U, __m256i __A)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_abs_epi16(__A),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_packs_epi32(__mmask8 __M, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__M,
|
|
(__v8hi)_mm_packs_epi32(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_packs_epi32(__m128i __W, __mmask16 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__M,
|
|
(__v8hi)_mm_packs_epi32(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_packs_epi32(__mmask16 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__M,
|
|
(__v16hi)_mm256_packs_epi32(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_packs_epi32(__m256i __W, __mmask16 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__M,
|
|
(__v16hi)_mm256_packs_epi32(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_packs_epi16(__mmask16 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__M,
|
|
(__v16qi)_mm_packs_epi16(__A, __B),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_packs_epi16(__m128i __W, __mmask16 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__M,
|
|
(__v16qi)_mm_packs_epi16(__A, __B),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_packs_epi16(__mmask32 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__M,
|
|
(__v32qi)_mm256_packs_epi16(__A, __B),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_packs_epi16(__m256i __W, __mmask32 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__M,
|
|
(__v32qi)_mm256_packs_epi16(__A, __B),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_packus_epi32(__mmask8 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__M,
|
|
(__v8hi)_mm_packus_epi32(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_packus_epi32(__m128i __W, __mmask16 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__M,
|
|
(__v8hi)_mm_packus_epi32(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_packus_epi32(__mmask16 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__M,
|
|
(__v16hi)_mm256_packus_epi32(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_packus_epi32(__m256i __W, __mmask16 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__M,
|
|
(__v16hi)_mm256_packus_epi32(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_packus_epi16(__mmask16 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__M,
|
|
(__v16qi)_mm_packus_epi16(__A, __B),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_packus_epi16(__m128i __W, __mmask16 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__M,
|
|
(__v16qi)_mm_packus_epi16(__A, __B),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_packus_epi16(__mmask32 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__M,
|
|
(__v32qi)_mm256_packus_epi16(__A, __B),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_packus_epi16(__m256i __W, __mmask32 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__M,
|
|
(__v32qi)_mm256_packus_epi16(__A, __B),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_adds_epi8(__m128i __W, __mmask16 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_adds_epi8(__A, __B),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_adds_epi8(__mmask16 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_adds_epi8(__A, __B),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_adds_epi8(__m256i __W, __mmask32 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_adds_epi8(__A, __B),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_adds_epi8(__mmask32 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_adds_epi8(__A, __B),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_adds_epi16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_adds_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_adds_epi16(__mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_adds_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_adds_epi16(__m256i __W, __mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_adds_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_adds_epi16(__mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_adds_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_adds_epu8(__m128i __W, __mmask16 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_adds_epu8(__A, __B),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_adds_epu8(__mmask16 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_adds_epu8(__A, __B),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_adds_epu8(__m256i __W, __mmask32 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_adds_epu8(__A, __B),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_adds_epu8(__mmask32 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_adds_epu8(__A, __B),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_adds_epu16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_adds_epu16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_adds_epu16(__mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_adds_epu16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_adds_epu16(__m256i __W, __mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_adds_epu16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_adds_epu16(__mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_adds_epu16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_avg_epu8(__m128i __W, __mmask16 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_avg_epu8(__A, __B),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_avg_epu8(__mmask16 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_avg_epu8(__A, __B),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_avg_epu8(__m256i __W, __mmask32 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_avg_epu8(__A, __B),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_avg_epu8(__mmask32 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_avg_epu8(__A, __B),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_avg_epu16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_avg_epu16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_avg_epu16(__mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_avg_epu16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_avg_epu16(__m256i __W, __mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_avg_epu16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_avg_epu16(__mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_avg_epu16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_max_epi8(__mmask16 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__M,
|
|
(__v16qi)_mm_max_epi8(__A, __B),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_max_epi8(__m128i __W, __mmask16 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__M,
|
|
(__v16qi)_mm_max_epi8(__A, __B),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_max_epi8(__mmask32 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__M,
|
|
(__v32qi)_mm256_max_epi8(__A, __B),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_max_epi8(__m256i __W, __mmask32 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__M,
|
|
(__v32qi)_mm256_max_epi8(__A, __B),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_max_epi16(__mmask8 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__M,
|
|
(__v8hi)_mm_max_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_max_epi16(__m128i __W, __mmask8 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__M,
|
|
(__v8hi)_mm_max_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_max_epi16(__mmask16 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__M,
|
|
(__v16hi)_mm256_max_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_max_epi16(__m256i __W, __mmask16 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__M,
|
|
(__v16hi)_mm256_max_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_max_epu8(__mmask16 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__M,
|
|
(__v16qi)_mm_max_epu8(__A, __B),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_max_epu8(__m128i __W, __mmask16 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__M,
|
|
(__v16qi)_mm_max_epu8(__A, __B),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_max_epu8 (__mmask32 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__M,
|
|
(__v32qi)_mm256_max_epu8(__A, __B),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_max_epu8(__m256i __W, __mmask32 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__M,
|
|
(__v32qi)_mm256_max_epu8(__A, __B),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_max_epu16(__mmask8 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__M,
|
|
(__v8hi)_mm_max_epu16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_max_epu16(__m128i __W, __mmask8 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__M,
|
|
(__v8hi)_mm_max_epu16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_max_epu16(__mmask16 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__M,
|
|
(__v16hi)_mm256_max_epu16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_max_epu16(__m256i __W, __mmask16 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__M,
|
|
(__v16hi)_mm256_max_epu16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_min_epi8(__mmask16 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__M,
|
|
(__v16qi)_mm_min_epi8(__A, __B),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_min_epi8(__m128i __W, __mmask16 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__M,
|
|
(__v16qi)_mm_min_epi8(__A, __B),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_min_epi8(__mmask32 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__M,
|
|
(__v32qi)_mm256_min_epi8(__A, __B),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_min_epi8(__m256i __W, __mmask32 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__M,
|
|
(__v32qi)_mm256_min_epi8(__A, __B),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_min_epi16(__mmask8 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__M,
|
|
(__v8hi)_mm_min_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_min_epi16(__m128i __W, __mmask8 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__M,
|
|
(__v8hi)_mm_min_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_min_epi16(__mmask16 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__M,
|
|
(__v16hi)_mm256_min_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_min_epi16(__m256i __W, __mmask16 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__M,
|
|
(__v16hi)_mm256_min_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_min_epu8(__mmask16 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__M,
|
|
(__v16qi)_mm_min_epu8(__A, __B),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_min_epu8(__m128i __W, __mmask16 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__M,
|
|
(__v16qi)_mm_min_epu8(__A, __B),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_min_epu8 (__mmask32 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__M,
|
|
(__v32qi)_mm256_min_epu8(__A, __B),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_min_epu8(__m256i __W, __mmask32 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__M,
|
|
(__v32qi)_mm256_min_epu8(__A, __B),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_min_epu16(__mmask8 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__M,
|
|
(__v8hi)_mm_min_epu16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_min_epu16(__m128i __W, __mmask8 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__M,
|
|
(__v8hi)_mm_min_epu16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_min_epu16(__mmask16 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__M,
|
|
(__v16hi)_mm256_min_epu16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_min_epu16(__m256i __W, __mmask16 __M, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__M,
|
|
(__v16hi)_mm256_min_epu16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_shuffle_epi8(__m128i __W, __mmask16 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_shuffle_epi8(__A, __B),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_shuffle_epi8(__mmask16 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_shuffle_epi8(__A, __B),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_shuffle_epi8(__m256i __W, __mmask32 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_shuffle_epi8(__A, __B),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_shuffle_epi8(__mmask32 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_shuffle_epi8(__A, __B),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_subs_epi8(__m128i __W, __mmask16 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_subs_epi8(__A, __B),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_subs_epi8(__mmask16 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_subs_epi8(__A, __B),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_subs_epi8(__m256i __W, __mmask32 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_subs_epi8(__A, __B),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_subs_epi8(__mmask32 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_subs_epi8(__A, __B),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_subs_epi16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_subs_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_subs_epi16(__mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_subs_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_subs_epi16(__m256i __W, __mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_subs_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_subs_epi16(__mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_subs_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_subs_epu8(__m128i __W, __mmask16 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_subs_epu8(__A, __B),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_subs_epu8(__mmask16 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_subs_epu8(__A, __B),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_subs_epu8(__m256i __W, __mmask32 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_subs_epu8(__A, __B),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_subs_epu8(__mmask32 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_subs_epu8(__A, __B),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_subs_epu16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_subs_epu16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_subs_epu16(__mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_subs_epu16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_subs_epu16(__m256i __W, __mmask16 __U, __m256i __A,
|
|
__m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_subs_epu16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_subs_epu16(__mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_subs_epu16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask2_permutex2var_epi16 (__m128i __A, __m128i __I, __mmask8 __U,
|
|
__m128i __B)
|
|
{
|
|
return (__m128i) __builtin_ia32_vpermi2varhi128_mask ((__v8hi) __A,
|
|
(__v8hi) __I /* idx */ ,
|
|
(__v8hi) __B,
|
|
(__mmask8) __U);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask2_permutex2var_epi16 (__m256i __A, __m256i __I,
|
|
__mmask16 __U, __m256i __B)
|
|
{
|
|
return (__m256i) __builtin_ia32_vpermi2varhi256_mask ((__v16hi) __A,
|
|
(__v16hi) __I /* idx */ ,
|
|
(__v16hi) __B,
|
|
(__mmask16) __U);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_permutex2var_epi16 (__m128i __A, __m128i __I, __m128i __B)
|
|
{
|
|
return (__m128i) __builtin_ia32_vpermt2varhi128_mask ((__v8hi) __I/* idx */,
|
|
(__v8hi) __A,
|
|
(__v8hi) __B,
|
|
(__mmask8) -1);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_permutex2var_epi16 (__m128i __A, __mmask8 __U, __m128i __I,
|
|
__m128i __B)
|
|
{
|
|
return (__m128i) __builtin_ia32_vpermt2varhi128_mask ((__v8hi) __I/* idx */,
|
|
(__v8hi) __A,
|
|
(__v8hi) __B,
|
|
(__mmask8) __U);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_permutex2var_epi16 (__mmask8 __U, __m128i __A, __m128i __I,
|
|
__m128i __B)
|
|
{
|
|
return (__m128i) __builtin_ia32_vpermt2varhi128_maskz ((__v8hi) __I/* idx */,
|
|
(__v8hi) __A,
|
|
(__v8hi) __B,
|
|
(__mmask8) __U);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_permutex2var_epi16 (__m256i __A, __m256i __I, __m256i __B)
|
|
{
|
|
return (__m256i) __builtin_ia32_vpermt2varhi256_mask ((__v16hi) __I/* idx */,
|
|
(__v16hi) __A,
|
|
(__v16hi) __B,
|
|
(__mmask16) -1);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_permutex2var_epi16 (__m256i __A, __mmask16 __U,
|
|
__m256i __I, __m256i __B)
|
|
{
|
|
return (__m256i) __builtin_ia32_vpermt2varhi256_mask ((__v16hi) __I/* idx */,
|
|
(__v16hi) __A,
|
|
(__v16hi) __B,
|
|
(__mmask16) __U);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_permutex2var_epi16 (__mmask16 __U, __m256i __A,
|
|
__m256i __I, __m256i __B)
|
|
{
|
|
return (__m256i) __builtin_ia32_vpermt2varhi256_maskz ((__v16hi) __I/* idx */,
|
|
(__v16hi) __A,
|
|
(__v16hi) __B,
|
|
(__mmask16) __U);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_maddubs_epi16(__m128i __W, __mmask8 __U, __m128i __X, __m128i __Y) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_maddubs_epi16(__X, __Y),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_maddubs_epi16(__mmask8 __U, __m128i __X, __m128i __Y) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_maddubs_epi16(__X, __Y),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_maddubs_epi16(__m256i __W, __mmask16 __U, __m256i __X,
|
|
__m256i __Y) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_maddubs_epi16(__X, __Y),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_maddubs_epi16(__mmask16 __U, __m256i __X, __m256i __Y) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_maddubs_epi16(__X, __Y),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_madd_epi16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectd_128((__mmask8)__U,
|
|
(__v4si)_mm_madd_epi16(__A, __B),
|
|
(__v4si)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_madd_epi16(__mmask8 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectd_128((__mmask8)__U,
|
|
(__v4si)_mm_madd_epi16(__A, __B),
|
|
(__v4si)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_madd_epi16(__m256i __W, __mmask8 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectd_256((__mmask8)__U,
|
|
(__v8si)_mm256_madd_epi16(__A, __B),
|
|
(__v8si)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_madd_epi16(__mmask8 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectd_256((__mmask8)__U,
|
|
(__v8si)_mm256_madd_epi16(__A, __B),
|
|
(__v8si)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_cvtsepi16_epi8 (__m128i __A) {
|
|
return (__m128i) __builtin_ia32_pmovswb128_mask ((__v8hi) __A,
|
|
(__v16qi) _mm_setzero_si128(),
|
|
(__mmask8) -1);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_cvtsepi16_epi8 (__m128i __O, __mmask8 __M, __m128i __A) {
|
|
return (__m128i) __builtin_ia32_pmovswb128_mask ((__v8hi) __A,
|
|
(__v16qi) __O,
|
|
__M);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_cvtsepi16_epi8 (__mmask8 __M, __m128i __A) {
|
|
return (__m128i) __builtin_ia32_pmovswb128_mask ((__v8hi) __A,
|
|
(__v16qi) _mm_setzero_si128(),
|
|
__M);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm256_cvtsepi16_epi8 (__m256i __A) {
|
|
return (__m128i) __builtin_ia32_pmovswb256_mask ((__v16hi) __A,
|
|
(__v16qi) _mm_setzero_si128(),
|
|
(__mmask16) -1);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_cvtsepi16_epi8 (__m128i __O, __mmask16 __M, __m256i __A) {
|
|
return (__m128i) __builtin_ia32_pmovswb256_mask ((__v16hi) __A,
|
|
(__v16qi) __O,
|
|
__M);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_cvtsepi16_epi8 (__mmask16 __M, __m256i __A) {
|
|
return (__m128i) __builtin_ia32_pmovswb256_mask ((__v16hi) __A,
|
|
(__v16qi) _mm_setzero_si128(),
|
|
__M);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_cvtusepi16_epi8 (__m128i __A) {
|
|
return (__m128i) __builtin_ia32_pmovuswb128_mask ((__v8hi) __A,
|
|
(__v16qi) _mm_setzero_si128(),
|
|
(__mmask8) -1);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_cvtusepi16_epi8 (__m128i __O, __mmask8 __M, __m128i __A) {
|
|
return (__m128i) __builtin_ia32_pmovuswb128_mask ((__v8hi) __A,
|
|
(__v16qi) __O,
|
|
__M);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_cvtusepi16_epi8 (__mmask8 __M, __m128i __A) {
|
|
return (__m128i) __builtin_ia32_pmovuswb128_mask ((__v8hi) __A,
|
|
(__v16qi) _mm_setzero_si128(),
|
|
__M);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm256_cvtusepi16_epi8 (__m256i __A) {
|
|
return (__m128i) __builtin_ia32_pmovuswb256_mask ((__v16hi) __A,
|
|
(__v16qi) _mm_setzero_si128(),
|
|
(__mmask16) -1);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_cvtusepi16_epi8 (__m128i __O, __mmask16 __M, __m256i __A) {
|
|
return (__m128i) __builtin_ia32_pmovuswb256_mask ((__v16hi) __A,
|
|
(__v16qi) __O,
|
|
__M);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_cvtusepi16_epi8 (__mmask16 __M, __m256i __A) {
|
|
return (__m128i) __builtin_ia32_pmovuswb256_mask ((__v16hi) __A,
|
|
(__v16qi) _mm_setzero_si128(),
|
|
__M);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_cvtepi16_epi8 (__m128i __A) {
|
|
|
|
return (__m128i) __builtin_ia32_pmovwb128_mask ((__v8hi) __A,
|
|
(__v16qi) _mm_setzero_si128(),
|
|
(__mmask8) -1);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_cvtepi16_epi8 (__m128i __O, __mmask8 __M, __m128i __A) {
|
|
return (__m128i) __builtin_ia32_pmovwb128_mask ((__v8hi) __A,
|
|
(__v16qi) __O,
|
|
__M);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_cvtepi16_epi8 (__mmask8 __M, __m128i __A) {
|
|
return (__m128i) __builtin_ia32_pmovwb128_mask ((__v8hi) __A,
|
|
(__v16qi) _mm_setzero_si128(),
|
|
__M);
|
|
}
|
|
|
|
static __inline__ void __DEFAULT_FN_ATTRS
|
|
_mm_mask_cvtepi16_storeu_epi8 (void * __P, __mmask8 __M, __m128i __A)
|
|
{
|
|
__builtin_ia32_pmovwb128mem_mask ((__v16qi *) __P, (__v8hi) __A, __M);
|
|
}
|
|
|
|
|
|
static __inline__ void __DEFAULT_FN_ATTRS
|
|
_mm_mask_cvtsepi16_storeu_epi8 (void * __P, __mmask8 __M, __m128i __A)
|
|
{
|
|
__builtin_ia32_pmovswb128mem_mask ((__v16qi *) __P, (__v8hi) __A, __M);
|
|
}
|
|
|
|
static __inline__ void __DEFAULT_FN_ATTRS
|
|
_mm_mask_cvtusepi16_storeu_epi8 (void * __P, __mmask8 __M, __m128i __A)
|
|
{
|
|
__builtin_ia32_pmovuswb128mem_mask ((__v16qi *) __P, (__v8hi) __A, __M);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm256_cvtepi16_epi8 (__m256i __A) {
|
|
return (__m128i) __builtin_ia32_pmovwb256_mask ((__v16hi) __A,
|
|
(__v16qi) _mm_setzero_si128(),
|
|
(__mmask16) -1);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_cvtepi16_epi8 (__m128i __O, __mmask16 __M, __m256i __A) {
|
|
return (__m128i) __builtin_ia32_pmovwb256_mask ((__v16hi) __A,
|
|
(__v16qi) __O,
|
|
__M);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_cvtepi16_epi8 (__mmask16 __M, __m256i __A) {
|
|
return (__m128i) __builtin_ia32_pmovwb256_mask ((__v16hi) __A,
|
|
(__v16qi) _mm_setzero_si128(),
|
|
__M);
|
|
}
|
|
|
|
static __inline__ void __DEFAULT_FN_ATTRS
|
|
_mm256_mask_cvtepi16_storeu_epi8 (void * __P, __mmask16 __M, __m256i __A)
|
|
{
|
|
__builtin_ia32_pmovwb256mem_mask ((__v16qi *) __P, (__v16hi) __A, __M);
|
|
}
|
|
|
|
static __inline__ void __DEFAULT_FN_ATTRS
|
|
_mm256_mask_cvtsepi16_storeu_epi8 (void * __P, __mmask16 __M, __m256i __A)
|
|
{
|
|
__builtin_ia32_pmovswb256mem_mask ((__v16qi *) __P, (__v16hi) __A, __M);
|
|
}
|
|
|
|
static __inline__ void __DEFAULT_FN_ATTRS
|
|
_mm256_mask_cvtusepi16_storeu_epi8 (void * __P, __mmask8 __M, __m256i __A)
|
|
{
|
|
__builtin_ia32_pmovuswb256mem_mask ((__v16qi*) __P, (__v16hi) __A, __M);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_mulhrs_epi16(__m128i __W, __mmask8 __U, __m128i __X, __m128i __Y) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_mulhrs_epi16(__X, __Y),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_mulhrs_epi16(__mmask8 __U, __m128i __X, __m128i __Y) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_mulhrs_epi16(__X, __Y),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_mulhrs_epi16(__m256i __W, __mmask16 __U, __m256i __X, __m256i __Y) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_mulhrs_epi16(__X, __Y),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_mulhrs_epi16(__mmask16 __U, __m256i __X, __m256i __Y) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_mulhrs_epi16(__X, __Y),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_mulhi_epu16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_mulhi_epu16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_mulhi_epu16(__mmask8 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_mulhi_epu16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_mulhi_epu16(__m256i __W, __mmask16 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_mulhi_epu16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_mulhi_epu16(__mmask16 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_mulhi_epu16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_mulhi_epi16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_mulhi_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_mulhi_epi16(__mmask8 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_mulhi_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_mulhi_epi16(__m256i __W, __mmask16 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_mulhi_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_mulhi_epi16(__mmask16 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_mulhi_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_unpackhi_epi8(__m128i __W, __mmask16 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_unpackhi_epi8(__A, __B),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_unpackhi_epi8(__mmask16 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_unpackhi_epi8(__A, __B),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_unpackhi_epi8(__m256i __W, __mmask32 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_unpackhi_epi8(__A, __B),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_unpackhi_epi8(__mmask32 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_unpackhi_epi8(__A, __B),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_unpackhi_epi16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_unpackhi_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_unpackhi_epi16(__mmask8 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_unpackhi_epi16(__A, __B),
|
|
(__v8hi) _mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_unpackhi_epi16(__m256i __W, __mmask16 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_unpackhi_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_unpackhi_epi16(__mmask16 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_unpackhi_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_unpacklo_epi8(__m128i __W, __mmask16 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_unpacklo_epi8(__A, __B),
|
|
(__v16qi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_unpacklo_epi8(__mmask16 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectb_128((__mmask16)__U,
|
|
(__v16qi)_mm_unpacklo_epi8(__A, __B),
|
|
(__v16qi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_unpacklo_epi8(__m256i __W, __mmask32 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_unpacklo_epi8(__A, __B),
|
|
(__v32qi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_unpacklo_epi8(__mmask32 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectb_256((__mmask32)__U,
|
|
(__v32qi)_mm256_unpacklo_epi8(__A, __B),
|
|
(__v32qi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_unpacklo_epi16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_unpacklo_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_unpacklo_epi16(__mmask8 __U, __m128i __A, __m128i __B) {
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_unpacklo_epi16(__A, __B),
|
|
(__v8hi) _mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_unpacklo_epi16(__m256i __W, __mmask16 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_unpacklo_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_unpacklo_epi16(__mmask16 __U, __m256i __A, __m256i __B) {
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_unpacklo_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_cvtepi8_epi16(__m128i __W, __mmask8 __U, __m128i __A)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_cvtepi8_epi16(__A),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_cvtepi8_epi16(__mmask8 __U, __m128i __A)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_cvtepi8_epi16(__A),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_cvtepi8_epi16(__m256i __W, __mmask16 __U, __m128i __A)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_cvtepi8_epi16(__A),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_cvtepi8_epi16(__mmask16 __U, __m128i __A)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_cvtepi8_epi16(__A),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_cvtepu8_epi16(__m128i __W, __mmask8 __U, __m128i __A)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_cvtepu8_epi16(__A),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_cvtepu8_epi16(__mmask8 __U, __m128i __A)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_cvtepu8_epi16(__A),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_cvtepu8_epi16(__m256i __W, __mmask16 __U, __m128i __A)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_cvtepu8_epi16(__A),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_cvtepu8_epi16 (__mmask16 __U, __m128i __A)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_cvtepu8_epi16(__A),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
|
|
#define _mm_mask_shufflehi_epi16(W, U, A, imm) __extension__ ({ \
|
|
(__m128i)__builtin_ia32_selectw_128((__mmask8)(U), \
|
|
(__v8hi)_mm_shufflehi_epi16((A), (imm)), \
|
|
(__v8hi)(__m128i)(W)); })
|
|
|
|
#define _mm_maskz_shufflehi_epi16(U, A, imm) __extension__ ({ \
|
|
(__m128i)__builtin_ia32_selectw_128((__mmask8)(U), \
|
|
(__v8hi)_mm_shufflehi_epi16((A), (imm)), \
|
|
(__v8hi)_mm_setzero_hi()); })
|
|
|
|
#define _mm256_mask_shufflehi_epi16(W, U, A, imm) __extension__ ({ \
|
|
(__m256i)__builtin_ia32_selectw_256((__mmask16)(U), \
|
|
(__v16hi)_mm256_shufflehi_epi16((A), (imm)), \
|
|
(__v16hi)(__m256i)(W)); })
|
|
|
|
#define _mm256_maskz_shufflehi_epi16(U, A, imm) __extension__ ({ \
|
|
(__m256i)__builtin_ia32_selectw_256((__mmask16)(U), \
|
|
(__v16hi)_mm256_shufflehi_epi16((A), (imm)), \
|
|
(__v16hi)_mm256_setzero_si256()); })
|
|
|
|
#define _mm_mask_shufflelo_epi16(W, U, A, imm) __extension__ ({ \
|
|
(__m128i)__builtin_ia32_selectw_128((__mmask8)(U), \
|
|
(__v8hi)_mm_shufflelo_epi16((A), (imm)), \
|
|
(__v8hi)(__m128i)(W)); })
|
|
|
|
#define _mm_maskz_shufflelo_epi16(U, A, imm) __extension__ ({ \
|
|
(__m128i)__builtin_ia32_selectw_128((__mmask8)(U), \
|
|
(__v8hi)_mm_shufflelo_epi16((A), (imm)), \
|
|
(__v8hi)_mm_setzero_hi()); })
|
|
|
|
#define _mm256_mask_shufflelo_epi16(W, U, A, imm) __extension__ ({ \
|
|
(__m256i)__builtin_ia32_selectw_256((__mmask16)(U), \
|
|
(__v16hi)_mm256_shufflelo_epi16((A), \
|
|
(imm)), \
|
|
(__v16hi)(__m256i)(W)); })
|
|
|
|
#define _mm256_maskz_shufflelo_epi16(U, A, imm) __extension__ ({ \
|
|
(__m256i)__builtin_ia32_selectw_256((__mmask16)(U), \
|
|
(__v16hi)_mm256_shufflelo_epi16((A), \
|
|
(imm)), \
|
|
(__v16hi)_mm256_setzero_si256()); })
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_sllv_epi16(__m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_psllv16hi((__v16hi)__A, (__v16hi)__B);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_sllv_epi16(__m256i __W, __mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_sllv_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_sllv_epi16(__mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_sllv_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_sllv_epi16(__m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_psllv8hi((__v8hi)__A, (__v8hi)__B);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_sllv_epi16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_sllv_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_sllv_epi16(__mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_sllv_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_sll_epi16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_sll_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_sll_epi16 (__mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_sll_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_sll_epi16(__m256i __W, __mmask16 __U, __m256i __A, __m128i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_sll_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_sll_epi16(__mmask16 __U, __m256i __A, __m128i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_sll_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_slli_epi16(__m128i __W, __mmask8 __U, __m128i __A, int __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_slli_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_slli_epi16 (__mmask8 __U, __m128i __A, int __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_slli_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_slli_epi16(__m256i __W, __mmask16 __U, __m256i __A, int __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_slli_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_slli_epi16(__mmask16 __U, __m256i __A, int __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_slli_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_srlv_epi16(__m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_psrlv16hi((__v16hi)__A, (__v16hi)__B);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_srlv_epi16(__m256i __W, __mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_srlv_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_srlv_epi16(__mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_srlv_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_srlv_epi16(__m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_psrlv8hi((__v8hi)__A, (__v8hi)__B);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_srlv_epi16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_srlv_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_srlv_epi16(__mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_srlv_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_srav_epi16(__m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_psrav16hi((__v16hi)__A, (__v16hi)__B);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_srav_epi16(__m256i __W, __mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_srav_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_srav_epi16(__mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_srav_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_srav_epi16(__m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_psrav8hi((__v8hi)__A, (__v8hi)__B);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_srav_epi16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_srav_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_srav_epi16(__mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_srav_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_sra_epi16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_sra_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_sra_epi16(__mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_sra_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_sra_epi16(__m256i __W, __mmask16 __U, __m256i __A, __m128i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_sra_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_sra_epi16(__mmask16 __U, __m256i __A, __m128i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_sra_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_srai_epi16(__m128i __W, __mmask8 __U, __m128i __A, int __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_srai_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_srai_epi16(__mmask8 __U, __m128i __A, int __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_srai_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_srai_epi16(__m256i __W, __mmask16 __U, __m256i __A, int __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_srai_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_srai_epi16(__mmask16 __U, __m256i __A, int __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_srai_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_srl_epi16(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_srl_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_srl_epi16 (__mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_srl_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_srl_epi16(__m256i __W, __mmask16 __U, __m256i __A, __m128i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_srl_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_srl_epi16(__mmask16 __U, __m256i __A, __m128i __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_srl_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_srli_epi16(__m128i __W, __mmask8 __U, __m128i __A, int __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_srli_epi16(__A, __B),
|
|
(__v8hi)__W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_srli_epi16 (__mmask8 __U, __m128i __A, int __B)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128((__mmask8)__U,
|
|
(__v8hi)_mm_srli_epi16(__A, __B),
|
|
(__v8hi)_mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_srli_epi16(__m256i __W, __mmask16 __U, __m256i __A, int __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_srli_epi16(__A, __B),
|
|
(__v16hi)__W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_srli_epi16(__mmask16 __U, __m256i __A, int __B)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256((__mmask16)__U,
|
|
(__v16hi)_mm256_srli_epi16(__A, __B),
|
|
(__v16hi)_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_mov_epi16 (__m128i __W, __mmask8 __U, __m128i __A)
|
|
{
|
|
return (__m128i) __builtin_ia32_selectw_128 ((__mmask8) __U,
|
|
(__v8hi) __A,
|
|
(__v8hi) __W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_mov_epi16 (__mmask8 __U, __m128i __A)
|
|
{
|
|
return (__m128i) __builtin_ia32_selectw_128 ((__mmask8) __U,
|
|
(__v8hi) __A,
|
|
(__v8hi) _mm_setzero_hi ());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_mov_epi16 (__m256i __W, __mmask16 __U, __m256i __A)
|
|
{
|
|
return (__m256i) __builtin_ia32_selectw_256 ((__mmask16) __U,
|
|
(__v16hi) __A,
|
|
(__v16hi) __W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_mov_epi16 (__mmask16 __U, __m256i __A)
|
|
{
|
|
return (__m256i) __builtin_ia32_selectw_256 ((__mmask16) __U,
|
|
(__v16hi) __A,
|
|
(__v16hi) _mm256_setzero_si256 ());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_mov_epi8 (__m128i __W, __mmask16 __U, __m128i __A)
|
|
{
|
|
return (__m128i) __builtin_ia32_selectb_128 ((__mmask16) __U,
|
|
(__v16qi) __A,
|
|
(__v16qi) __W);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_mov_epi8 (__mmask16 __U, __m128i __A)
|
|
{
|
|
return (__m128i) __builtin_ia32_selectb_128 ((__mmask16) __U,
|
|
(__v16qi) __A,
|
|
(__v16qi) _mm_setzero_hi ());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_mov_epi8 (__m256i __W, __mmask32 __U, __m256i __A)
|
|
{
|
|
return (__m256i) __builtin_ia32_selectb_256 ((__mmask32) __U,
|
|
(__v32qi) __A,
|
|
(__v32qi) __W);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_mov_epi8 (__mmask32 __U, __m256i __A)
|
|
{
|
|
return (__m256i) __builtin_ia32_selectb_256 ((__mmask32) __U,
|
|
(__v32qi) __A,
|
|
(__v32qi) _mm256_setzero_si256 ());
|
|
}
|
|
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_set1_epi8 (__m128i __O, __mmask16 __M, char __A)
|
|
{
|
|
return (__m128i) __builtin_ia32_selectb_128(__M,
|
|
(__v16qi) _mm_set1_epi8(__A),
|
|
(__v16qi) __O);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_set1_epi8 (__mmask16 __M, char __A)
|
|
{
|
|
return (__m128i) __builtin_ia32_selectb_128(__M,
|
|
(__v16qi) _mm_set1_epi8(__A),
|
|
(__v16qi) _mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_set1_epi8 (__m256i __O, __mmask32 __M, char __A)
|
|
{
|
|
return (__m256i) __builtin_ia32_selectb_256(__M,
|
|
(__v32qi) _mm256_set1_epi8(__A),
|
|
(__v32qi) __O);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_set1_epi8 (__mmask32 __M, char __A)
|
|
{
|
|
return (__m256i) __builtin_ia32_selectb_256(__M,
|
|
(__v32qi) _mm256_set1_epi8(__A),
|
|
(__v32qi) _mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_loadu_epi16 (__m128i __W, __mmask8 __U, void const *__P)
|
|
{
|
|
return (__m128i) __builtin_ia32_loaddquhi128_mask ((__v8hi *) __P,
|
|
(__v8hi) __W,
|
|
(__mmask8) __U);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_loadu_epi16 (__mmask8 __U, void const *__P)
|
|
{
|
|
return (__m128i) __builtin_ia32_loaddquhi128_mask ((__v8hi *) __P,
|
|
(__v8hi)
|
|
_mm_setzero_hi (),
|
|
(__mmask8) __U);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_loadu_epi16 (__m256i __W, __mmask16 __U, void const *__P)
|
|
{
|
|
return (__m256i) __builtin_ia32_loaddquhi256_mask ((__v16hi *) __P,
|
|
(__v16hi) __W,
|
|
(__mmask16) __U);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_loadu_epi16 (__mmask16 __U, void const *__P)
|
|
{
|
|
return (__m256i) __builtin_ia32_loaddquhi256_mask ((__v16hi *) __P,
|
|
(__v16hi)
|
|
_mm256_setzero_si256 (),
|
|
(__mmask16) __U);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_loadu_epi8 (__m128i __W, __mmask16 __U, void const *__P)
|
|
{
|
|
return (__m128i) __builtin_ia32_loaddquqi128_mask ((__v16qi *) __P,
|
|
(__v16qi) __W,
|
|
(__mmask16) __U);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_loadu_epi8 (__mmask16 __U, void const *__P)
|
|
{
|
|
return (__m128i) __builtin_ia32_loaddquqi128_mask ((__v16qi *) __P,
|
|
(__v16qi)
|
|
_mm_setzero_si128 (),
|
|
(__mmask16) __U);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_loadu_epi8 (__m256i __W, __mmask32 __U, void const *__P)
|
|
{
|
|
return (__m256i) __builtin_ia32_loaddquqi256_mask ((__v32qi *) __P,
|
|
(__v32qi) __W,
|
|
(__mmask32) __U);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_loadu_epi8 (__mmask32 __U, void const *__P)
|
|
{
|
|
return (__m256i) __builtin_ia32_loaddquqi256_mask ((__v32qi *) __P,
|
|
(__v32qi)
|
|
_mm256_setzero_si256 (),
|
|
(__mmask32) __U);
|
|
}
|
|
static __inline__ void __DEFAULT_FN_ATTRS
|
|
_mm_mask_storeu_epi16 (void *__P, __mmask8 __U, __m128i __A)
|
|
{
|
|
__builtin_ia32_storedquhi128_mask ((__v8hi *) __P,
|
|
(__v8hi) __A,
|
|
(__mmask8) __U);
|
|
}
|
|
|
|
static __inline__ void __DEFAULT_FN_ATTRS
|
|
_mm256_mask_storeu_epi16 (void *__P, __mmask16 __U, __m256i __A)
|
|
{
|
|
__builtin_ia32_storedquhi256_mask ((__v16hi *) __P,
|
|
(__v16hi) __A,
|
|
(__mmask16) __U);
|
|
}
|
|
|
|
static __inline__ void __DEFAULT_FN_ATTRS
|
|
_mm_mask_storeu_epi8 (void *__P, __mmask16 __U, __m128i __A)
|
|
{
|
|
__builtin_ia32_storedquqi128_mask ((__v16qi *) __P,
|
|
(__v16qi) __A,
|
|
(__mmask16) __U);
|
|
}
|
|
|
|
static __inline__ void __DEFAULT_FN_ATTRS
|
|
_mm256_mask_storeu_epi8 (void *__P, __mmask32 __U, __m256i __A)
|
|
{
|
|
__builtin_ia32_storedquqi256_mask ((__v32qi *) __P,
|
|
(__v32qi) __A,
|
|
(__mmask32) __U);
|
|
}
|
|
|
|
static __inline__ __mmask16 __DEFAULT_FN_ATTRS
|
|
_mm_test_epi8_mask (__m128i __A, __m128i __B)
|
|
{
|
|
return _mm_cmpneq_epi8_mask (_mm_and_si128(__A, __B), _mm_setzero_hi());
|
|
}
|
|
|
|
static __inline__ __mmask16 __DEFAULT_FN_ATTRS
|
|
_mm_mask_test_epi8_mask (__mmask16 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return _mm_mask_cmpneq_epi8_mask (__U, _mm_and_si128 (__A, __B),
|
|
_mm_setzero_hi());
|
|
}
|
|
|
|
static __inline__ __mmask32 __DEFAULT_FN_ATTRS
|
|
_mm256_test_epi8_mask (__m256i __A, __m256i __B)
|
|
{
|
|
return _mm256_cmpneq_epi8_mask (_mm256_and_si256(__A, __B),
|
|
_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __mmask32 __DEFAULT_FN_ATTRS
|
|
_mm256_mask_test_epi8_mask (__mmask32 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return _mm256_mask_cmpneq_epi8_mask (__U, _mm256_and_si256(__A, __B),
|
|
_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __mmask8 __DEFAULT_FN_ATTRS
|
|
_mm_test_epi16_mask (__m128i __A, __m128i __B)
|
|
{
|
|
return _mm_cmpneq_epi16_mask (_mm_and_si128 (__A, __B), _mm_setzero_hi());
|
|
}
|
|
|
|
static __inline__ __mmask8 __DEFAULT_FN_ATTRS
|
|
_mm_mask_test_epi16_mask (__mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return _mm_mask_cmpneq_epi16_mask (__U, _mm_and_si128 (__A, __B),
|
|
_mm_setzero_hi());
|
|
}
|
|
|
|
static __inline__ __mmask16 __DEFAULT_FN_ATTRS
|
|
_mm256_test_epi16_mask (__m256i __A, __m256i __B)
|
|
{
|
|
return _mm256_cmpneq_epi16_mask (_mm256_and_si256 (__A, __B),
|
|
_mm256_setzero_si256 ());
|
|
}
|
|
|
|
static __inline__ __mmask16 __DEFAULT_FN_ATTRS
|
|
_mm256_mask_test_epi16_mask (__mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return _mm256_mask_cmpneq_epi16_mask (__U, _mm256_and_si256(__A, __B),
|
|
_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __mmask16 __DEFAULT_FN_ATTRS
|
|
_mm_testn_epi8_mask (__m128i __A, __m128i __B)
|
|
{
|
|
return _mm_cmpeq_epi8_mask (_mm_and_si128 (__A, __B), _mm_setzero_hi());
|
|
}
|
|
|
|
static __inline__ __mmask16 __DEFAULT_FN_ATTRS
|
|
_mm_mask_testn_epi8_mask (__mmask16 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return _mm_mask_cmpeq_epi8_mask (__U, _mm_and_si128 (__A, __B),
|
|
_mm_setzero_hi());
|
|
}
|
|
|
|
static __inline__ __mmask32 __DEFAULT_FN_ATTRS
|
|
_mm256_testn_epi8_mask (__m256i __A, __m256i __B)
|
|
{
|
|
return _mm256_cmpeq_epi8_mask (_mm256_and_si256 (__A, __B),
|
|
_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __mmask32 __DEFAULT_FN_ATTRS
|
|
_mm256_mask_testn_epi8_mask (__mmask32 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return _mm256_mask_cmpeq_epi8_mask (__U, _mm256_and_si256 (__A, __B),
|
|
_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __mmask8 __DEFAULT_FN_ATTRS
|
|
_mm_testn_epi16_mask (__m128i __A, __m128i __B)
|
|
{
|
|
return _mm_cmpeq_epi16_mask (_mm_and_si128 (__A, __B), _mm_setzero_hi());
|
|
}
|
|
|
|
static __inline__ __mmask8 __DEFAULT_FN_ATTRS
|
|
_mm_mask_testn_epi16_mask (__mmask8 __U, __m128i __A, __m128i __B)
|
|
{
|
|
return _mm_mask_cmpeq_epi16_mask (__U, _mm_and_si128(__A, __B), _mm_setzero_hi());
|
|
}
|
|
|
|
static __inline__ __mmask16 __DEFAULT_FN_ATTRS
|
|
_mm256_testn_epi16_mask (__m256i __A, __m256i __B)
|
|
{
|
|
return _mm256_cmpeq_epi16_mask (_mm256_and_si256(__A, __B),
|
|
_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __mmask16 __DEFAULT_FN_ATTRS
|
|
_mm256_mask_testn_epi16_mask (__mmask16 __U, __m256i __A, __m256i __B)
|
|
{
|
|
return _mm256_mask_cmpeq_epi16_mask (__U, _mm256_and_si256 (__A, __B),
|
|
_mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __mmask16 __DEFAULT_FN_ATTRS
|
|
_mm_movepi8_mask (__m128i __A)
|
|
{
|
|
return (__mmask16) __builtin_ia32_cvtb2mask128 ((__v16qi) __A);
|
|
}
|
|
|
|
static __inline__ __mmask32 __DEFAULT_FN_ATTRS
|
|
_mm256_movepi8_mask (__m256i __A)
|
|
{
|
|
return (__mmask32) __builtin_ia32_cvtb2mask256 ((__v32qi) __A);
|
|
}
|
|
|
|
static __inline__ __mmask8 __DEFAULT_FN_ATTRS
|
|
_mm_movepi16_mask (__m128i __A)
|
|
{
|
|
return (__mmask8) __builtin_ia32_cvtw2mask128 ((__v8hi) __A);
|
|
}
|
|
|
|
static __inline__ __mmask16 __DEFAULT_FN_ATTRS
|
|
_mm256_movepi16_mask (__m256i __A)
|
|
{
|
|
return (__mmask16) __builtin_ia32_cvtw2mask256 ((__v16hi) __A);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_movm_epi8 (__mmask16 __A)
|
|
{
|
|
return (__m128i) __builtin_ia32_cvtmask2b128 (__A);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_movm_epi8 (__mmask32 __A)
|
|
{
|
|
return (__m256i) __builtin_ia32_cvtmask2b256 (__A);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_movm_epi16 (__mmask8 __A)
|
|
{
|
|
return (__m128i) __builtin_ia32_cvtmask2w128 (__A);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_movm_epi16 (__mmask16 __A)
|
|
{
|
|
return (__m256i) __builtin_ia32_cvtmask2w256 (__A);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_broadcastb_epi8 (__m128i __O, __mmask16 __M, __m128i __A)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128(__M,
|
|
(__v16qi) _mm_broadcastb_epi8(__A),
|
|
(__v16qi) __O);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_broadcastb_epi8 (__mmask16 __M, __m128i __A)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectb_128(__M,
|
|
(__v16qi) _mm_broadcastb_epi8(__A),
|
|
(__v16qi) _mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_broadcastb_epi8 (__m256i __O, __mmask32 __M, __m128i __A)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256(__M,
|
|
(__v32qi) _mm256_broadcastb_epi8(__A),
|
|
(__v32qi) __O);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_broadcastb_epi8 (__mmask32 __M, __m128i __A)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectb_256(__M,
|
|
(__v32qi) _mm256_broadcastb_epi8(__A),
|
|
(__v32qi) _mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_broadcastw_epi16 (__m128i __O, __mmask8 __M, __m128i __A)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128(__M,
|
|
(__v8hi) _mm_broadcastw_epi16(__A),
|
|
(__v8hi) __O);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_broadcastw_epi16 (__mmask8 __M, __m128i __A)
|
|
{
|
|
return (__m128i)__builtin_ia32_selectw_128(__M,
|
|
(__v8hi) _mm_broadcastw_epi16(__A),
|
|
(__v8hi) _mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_broadcastw_epi16 (__m256i __O, __mmask16 __M, __m128i __A)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256(__M,
|
|
(__v16hi) _mm256_broadcastw_epi16(__A),
|
|
(__v16hi) __O);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_broadcastw_epi16 (__mmask16 __M, __m128i __A)
|
|
{
|
|
return (__m256i)__builtin_ia32_selectw_256(__M,
|
|
(__v16hi) _mm256_broadcastw_epi16(__A),
|
|
(__v16hi) _mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_set1_epi16 (__m256i __O, __mmask16 __M, short __A)
|
|
{
|
|
return (__m256i) __builtin_ia32_selectw_256 (__M,
|
|
(__v16hi) _mm256_set1_epi16(__A),
|
|
(__v16hi) __O);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_set1_epi16 (__mmask16 __M, short __A)
|
|
{
|
|
return (__m256i) __builtin_ia32_selectw_256(__M,
|
|
(__v16hi)_mm256_set1_epi16(__A),
|
|
(__v16hi) _mm256_setzero_si256());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_set1_epi16 (__m128i __O, __mmask8 __M, short __A)
|
|
{
|
|
return (__m128i) __builtin_ia32_selectw_128(__M,
|
|
(__v8hi) _mm_set1_epi16(__A),
|
|
(__v8hi) __O);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_set1_epi16 (__mmask8 __M, short __A)
|
|
{
|
|
return (__m128i) __builtin_ia32_selectw_128(__M,
|
|
(__v8hi) _mm_set1_epi16(__A),
|
|
(__v8hi) _mm_setzero_si128());
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_permutexvar_epi16 (__m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i) __builtin_ia32_permvarhi128_mask ((__v8hi) __B,
|
|
(__v8hi) __A,
|
|
(__v8hi) _mm_undefined_si128 (),
|
|
(__mmask8) -1);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_maskz_permutexvar_epi16 (__mmask8 __M, __m128i __A, __m128i __B)
|
|
{
|
|
return (__m128i) __builtin_ia32_permvarhi128_mask ((__v8hi) __B,
|
|
(__v8hi) __A,
|
|
(__v8hi) _mm_setzero_si128 (),
|
|
(__mmask8) __M);
|
|
}
|
|
|
|
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
|
_mm_mask_permutexvar_epi16 (__m128i __W, __mmask8 __M, __m128i __A,
|
|
__m128i __B)
|
|
{
|
|
return (__m128i) __builtin_ia32_permvarhi128_mask ((__v8hi) __B,
|
|
(__v8hi) __A,
|
|
(__v8hi) __W,
|
|
(__mmask8) __M);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_permutexvar_epi16 (__m256i __A, __m256i __B)
|
|
{
|
|
return (__m256i) __builtin_ia32_permvarhi256_mask ((__v16hi) __B,
|
|
(__v16hi) __A,
|
|
(__v16hi) _mm256_undefined_si256 (),
|
|
(__mmask16) -1);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_maskz_permutexvar_epi16 (__mmask16 __M, __m256i __A,
|
|
__m256i __B)
|
|
{
|
|
return (__m256i) __builtin_ia32_permvarhi256_mask ((__v16hi) __B,
|
|
(__v16hi) __A,
|
|
(__v16hi) _mm256_setzero_si256 (),
|
|
(__mmask16) __M);
|
|
}
|
|
|
|
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
|
_mm256_mask_permutexvar_epi16 (__m256i __W, __mmask16 __M, __m256i __A,
|
|
__m256i __B)
|
|
{
|
|
return (__m256i) __builtin_ia32_permvarhi256_mask ((__v16hi) __B,
|
|
(__v16hi) __A,
|
|
(__v16hi) __W,
|
|
(__mmask16) __M);
|
|
}
|
|
|
|
#define _mm_mask_alignr_epi8(W, U, A, B, N) __extension__ ({ \
|
|
(__m128i)__builtin_ia32_selectb_128((__mmask16)(U), \
|
|
(__v16qi)_mm_alignr_epi8((A), (B), (int)(N)), \
|
|
(__v16qi)(__m128i)(W)); })
|
|
|
|
#define _mm_maskz_alignr_epi8(U, A, B, N) __extension__ ({ \
|
|
(__m128i)__builtin_ia32_selectb_128((__mmask16)(U), \
|
|
(__v16qi)_mm_alignr_epi8((A), (B), (int)(N)), \
|
|
(__v16qi)_mm_setzero_si128()); })
|
|
|
|
#define _mm256_mask_alignr_epi8(W, U, A, B, N) __extension__ ({ \
|
|
(__m256i)__builtin_ia32_selectb_256((__mmask32)(U), \
|
|
(__v32qi)_mm256_alignr_epi8((A), (B), (int)(N)), \
|
|
(__v32qi)(__m256i)(W)); })
|
|
|
|
#define _mm256_maskz_alignr_epi8(U, A, B, N) __extension__ ({ \
|
|
(__m256i)__builtin_ia32_selectb_256((__mmask32)(U), \
|
|
(__v32qi)_mm256_alignr_epi8((A), (B), (int)(N)), \
|
|
(__v32qi)_mm256_setzero_si256()); })
|
|
|
|
#define _mm_dbsad_epu8(A, B, imm) __extension__ ({ \
|
|
(__m128i)__builtin_ia32_dbpsadbw128_mask((__v16qi)(__m128i)(A), \
|
|
(__v16qi)(__m128i)(B), (int)(imm), \
|
|
(__v8hi)_mm_setzero_hi(), \
|
|
(__mmask8)-1); })
|
|
|
|
#define _mm_mask_dbsad_epu8(W, U, A, B, imm) __extension__ ({ \
|
|
(__m128i)__builtin_ia32_dbpsadbw128_mask((__v16qi)(__m128i)(A), \
|
|
(__v16qi)(__m128i)(B), (int)(imm), \
|
|
(__v8hi)(__m128i)(W), \
|
|
(__mmask8)(U)); })
|
|
|
|
#define _mm_maskz_dbsad_epu8(U, A, B, imm) __extension__ ({ \
|
|
(__m128i)__builtin_ia32_dbpsadbw128_mask((__v16qi)(__m128i)(A), \
|
|
(__v16qi)(__m128i)(B), (int)(imm), \
|
|
(__v8hi)_mm_setzero_si128(), \
|
|
(__mmask8)(U)); })
|
|
|
|
#define _mm256_dbsad_epu8(A, B, imm) __extension__ ({ \
|
|
(__m256i)__builtin_ia32_dbpsadbw256_mask((__v32qi)(__m256i)(A), \
|
|
(__v32qi)(__m256i)(B), (int)(imm), \
|
|
(__v16hi)_mm256_setzero_si256(), \
|
|
(__mmask16)-1); })
|
|
|
|
#define _mm256_mask_dbsad_epu8(W, U, A, B, imm) __extension__ ({ \
|
|
(__m256i)__builtin_ia32_dbpsadbw256_mask((__v32qi)(__m256i)(A), \
|
|
(__v32qi)(__m256i)(B), (int)(imm), \
|
|
(__v16hi)(__m256i)(W), \
|
|
(__mmask16)(U)); })
|
|
|
|
#define _mm256_maskz_dbsad_epu8(U, A, B, imm) __extension__ ({ \
|
|
(__m256i)__builtin_ia32_dbpsadbw256_mask((__v32qi)(__m256i)(A), \
|
|
(__v32qi)(__m256i)(B), (int)(imm), \
|
|
(__v16hi)_mm256_setzero_si256(), \
|
|
(__mmask16)(U)); })
|
|
|
|
#undef __DEFAULT_FN_ATTRS
|
|
|
|
#endif /* __AVX512VLBWINTRIN_H */
|