|
#define | SSE2NEON_PRECISE_MINMAX (0) |
|
#define | SSE2NEON_PRECISE_DIV (0) |
|
#define | SSE2NEON_PRECISE_SQRT (0) |
|
#define | SSE2NEON_PRECISE_DP (0) |
|
#define | FORCE_INLINE static inline |
|
#define | ALIGN_STRUCT(x) __declspec(align(x)) |
|
#define | _sse2neon_likely(x) (x) |
|
#define | _sse2neon_unlikely(x) (x) |
|
#define | _sse2neon_const const |
|
#define | __has_builtin(x) 0 |
|
#define | _MM_SHUFFLE(fp3, fp2, fp1, fp0) (((fp3) << 6) | ((fp2) << 4) | ((fp1) << 2) | ((fp0))) |
|
#define | _MM_FROUND_TO_NEAREST_INT 0x00 |
|
#define | _MM_FROUND_TO_NEG_INF 0x01 |
|
#define | _MM_FROUND_TO_POS_INF 0x02 |
|
#define | _MM_FROUND_TO_ZERO 0x03 |
|
#define | _MM_FROUND_CUR_DIRECTION 0x04 |
|
#define | _MM_FROUND_NO_EXC 0x08 |
|
#define | _MM_FROUND_RAISE_EXC 0x00 |
|
#define | _MM_FROUND_NINT (_MM_FROUND_TO_NEAREST_INT | _MM_FROUND_RAISE_EXC) |
|
#define | _MM_FROUND_FLOOR (_MM_FROUND_TO_NEG_INF | _MM_FROUND_RAISE_EXC) |
|
#define | _MM_FROUND_CEIL (_MM_FROUND_TO_POS_INF | _MM_FROUND_RAISE_EXC) |
|
#define | _MM_FROUND_TRUNC (_MM_FROUND_TO_ZERO | _MM_FROUND_RAISE_EXC) |
|
#define | _MM_FROUND_RINT (_MM_FROUND_CUR_DIRECTION | _MM_FROUND_RAISE_EXC) |
|
#define | _MM_FROUND_NEARBYINT (_MM_FROUND_CUR_DIRECTION | _MM_FROUND_NO_EXC) |
|
#define | _MM_ROUND_NEAREST 0x0000 |
|
#define | _MM_ROUND_DOWN 0x2000 |
|
#define | _MM_ROUND_UP 0x4000 |
|
#define | _MM_ROUND_TOWARD_ZERO 0x6000 |
|
#define | _MM_FLUSH_ZERO_MASK 0x8000 |
|
#define | _MM_FLUSH_ZERO_ON 0x8000 |
|
#define | _MM_FLUSH_ZERO_OFF 0x0000 |
|
#define | _MM_DENORMALS_ZERO_MASK 0x0040 |
|
#define | _MM_DENORMALS_ZERO_ON 0x0040 |
|
#define | _MM_DENORMALS_ZERO_OFF 0x0000 |
|
#define | __constrange(a, b) const |
|
#define | __int64 int64_t |
|
#define | vreinterpretq_m128_f16(x) vreinterpretq_f32_f16(x) |
|
#define | vreinterpretq_m128_f32(x) (x) |
|
#define | vreinterpretq_m128_f64(x) vreinterpretq_f32_f64(x) |
|
#define | vreinterpretq_m128_u8(x) vreinterpretq_f32_u8(x) |
|
#define | vreinterpretq_m128_u16(x) vreinterpretq_f32_u16(x) |
|
#define | vreinterpretq_m128_u32(x) vreinterpretq_f32_u32(x) |
|
#define | vreinterpretq_m128_u64(x) vreinterpretq_f32_u64(x) |
|
#define | vreinterpretq_m128_s8(x) vreinterpretq_f32_s8(x) |
|
#define | vreinterpretq_m128_s16(x) vreinterpretq_f32_s16(x) |
|
#define | vreinterpretq_m128_s32(x) vreinterpretq_f32_s32(x) |
|
#define | vreinterpretq_m128_s64(x) vreinterpretq_f32_s64(x) |
|
#define | vreinterpretq_f16_m128(x) vreinterpretq_f16_f32(x) |
|
#define | vreinterpretq_f32_m128(x) (x) |
|
#define | vreinterpretq_f64_m128(x) vreinterpretq_f64_f32(x) |
|
#define | vreinterpretq_u8_m128(x) vreinterpretq_u8_f32(x) |
|
#define | vreinterpretq_u16_m128(x) vreinterpretq_u16_f32(x) |
|
#define | vreinterpretq_u32_m128(x) vreinterpretq_u32_f32(x) |
|
#define | vreinterpretq_u64_m128(x) vreinterpretq_u64_f32(x) |
|
#define | vreinterpretq_s8_m128(x) vreinterpretq_s8_f32(x) |
|
#define | vreinterpretq_s16_m128(x) vreinterpretq_s16_f32(x) |
|
#define | vreinterpretq_s32_m128(x) vreinterpretq_s32_f32(x) |
|
#define | vreinterpretq_s64_m128(x) vreinterpretq_s64_f32(x) |
|
#define | vreinterpretq_m128i_s8(x) vreinterpretq_s64_s8(x) |
|
#define | vreinterpretq_m128i_s16(x) vreinterpretq_s64_s16(x) |
|
#define | vreinterpretq_m128i_s32(x) vreinterpretq_s64_s32(x) |
|
#define | vreinterpretq_m128i_s64(x) (x) |
|
#define | vreinterpretq_m128i_u8(x) vreinterpretq_s64_u8(x) |
|
#define | vreinterpretq_m128i_u16(x) vreinterpretq_s64_u16(x) |
|
#define | vreinterpretq_m128i_u32(x) vreinterpretq_s64_u32(x) |
|
#define | vreinterpretq_m128i_u64(x) vreinterpretq_s64_u64(x) |
|
#define | vreinterpretq_f32_m128i(x) vreinterpretq_f32_s64(x) |
|
#define | vreinterpretq_f64_m128i(x) vreinterpretq_f64_s64(x) |
|
#define | vreinterpretq_s8_m128i(x) vreinterpretq_s8_s64(x) |
|
#define | vreinterpretq_s16_m128i(x) vreinterpretq_s16_s64(x) |
|
#define | vreinterpretq_s32_m128i(x) vreinterpretq_s32_s64(x) |
|
#define | vreinterpretq_s64_m128i(x) (x) |
|
#define | vreinterpretq_u8_m128i(x) vreinterpretq_u8_s64(x) |
|
#define | vreinterpretq_u16_m128i(x) vreinterpretq_u16_s64(x) |
|
#define | vreinterpretq_u32_m128i(x) vreinterpretq_u32_s64(x) |
|
#define | vreinterpretq_u64_m128i(x) vreinterpretq_u64_s64(x) |
|
#define | vreinterpret_m64_s8(x) vreinterpret_s64_s8(x) |
|
#define | vreinterpret_m64_s16(x) vreinterpret_s64_s16(x) |
|
#define | vreinterpret_m64_s32(x) vreinterpret_s64_s32(x) |
|
#define | vreinterpret_m64_s64(x) (x) |
|
#define | vreinterpret_m64_u8(x) vreinterpret_s64_u8(x) |
|
#define | vreinterpret_m64_u16(x) vreinterpret_s64_u16(x) |
|
#define | vreinterpret_m64_u32(x) vreinterpret_s64_u32(x) |
|
#define | vreinterpret_m64_u64(x) vreinterpret_s64_u64(x) |
|
#define | vreinterpret_m64_f16(x) vreinterpret_s64_f16(x) |
|
#define | vreinterpret_m64_f32(x) vreinterpret_s64_f32(x) |
|
#define | vreinterpret_m64_f64(x) vreinterpret_s64_f64(x) |
|
#define | vreinterpret_u8_m64(x) vreinterpret_u8_s64(x) |
|
#define | vreinterpret_u16_m64(x) vreinterpret_u16_s64(x) |
|
#define | vreinterpret_u32_m64(x) vreinterpret_u32_s64(x) |
|
#define | vreinterpret_u64_m64(x) vreinterpret_u64_s64(x) |
|
#define | vreinterpret_s8_m64(x) vreinterpret_s8_s64(x) |
|
#define | vreinterpret_s16_m64(x) vreinterpret_s16_s64(x) |
|
#define | vreinterpret_s32_m64(x) vreinterpret_s32_s64(x) |
|
#define | vreinterpret_s64_m64(x) (x) |
|
#define | vreinterpret_f32_m64(x) vreinterpret_f32_s64(x) |
|
#define | vreinterpretq_m128d_s32(x) vreinterpretq_f32_s32(x) |
|
#define | vreinterpretq_m128d_s64(x) vreinterpretq_f32_s64(x) |
|
#define | vreinterpretq_m128d_u32(x) vreinterpretq_f32_u32(x) |
|
#define | vreinterpretq_m128d_u64(x) vreinterpretq_f32_u64(x) |
|
#define | vreinterpretq_m128d_f32(x) (x) |
|
#define | vreinterpretq_s64_m128d(x) vreinterpretq_s64_f32(x) |
|
#define | vreinterpretq_u32_m128d(x) vreinterpretq_u32_f32(x) |
|
#define | vreinterpretq_u64_m128d(x) vreinterpretq_u64_f32(x) |
|
#define | vreinterpretq_f32_m128d(x) (x) |
|
#define | vreinterpretq_nth_u64_m128i(x, n) (((SIMDVec *) &x)->m128_u64[n]) |
|
#define | vreinterpretq_nth_u32_m128i(x, n) (((SIMDVec *) &x)->m128_u32[n]) |
|
#define | vreinterpretq_nth_u8_m128i(x, n) (((SIMDVec *) &x)->m128_u8[n]) |
|
#define | _MM_GET_FLUSH_ZERO_MODE _sse2neon_mm_get_flush_zero_mode |
|
#define | _MM_SET_FLUSH_ZERO_MODE _sse2neon_mm_set_flush_zero_mode |
|
#define | _MM_GET_DENORMALS_ZERO_MODE _sse2neon_mm_get_denormals_zero_mode |
|
#define | _MM_SET_DENORMALS_ZERO_MODE _sse2neon_mm_set_denormals_zero_mode |
|
#define | _mm_shuffle_epi32_default(a, imm) |
|
#define | _mm_shuffle_epi32_splat(a, imm) |
|
#define | _mm_shuffle_ps_default(a, b, imm) |
|
#define | _mm_shufflelo_epi16_function(a, imm) |
|
#define | _mm_shufflehi_epi16_function(a, imm) |
|
#define | _mm_cvtps_pi32(a) _mm_cvt_ps2pi(a) |
|
#define | _mm_cvtsi32_ss(a, b) _mm_cvt_si2ss(a, b) |
|
#define | _mm_cvtss_si32(a) _mm_cvt_ss2si(a) |
|
#define | _mm_cvttps_pi32(a) _mm_cvtt_ps2pi(a) |
|
#define | _mm_cvttss_si32(a) _mm_cvtt_ss2si(a) |
|
#define | _mm_extract_pi16(a, imm) (int32_t) vget_lane_u16(vreinterpret_u16_m64(a), (imm)) |
|
#define | _mm_insert_pi16(a, b, imm) |
|
#define | _mm_load_ps1 _mm_load1_ps |
|
#define | _m_maskmovq(a, mask, mem_addr) _mm_maskmove_si64(a, mask, mem_addr) |
|
#define | _m_pavgb(a, b) _mm_avg_pu8(a, b) |
|
#define | _m_pavgw(a, b) _mm_avg_pu16(a, b) |
|
#define | _m_pextrw(a, imm) _mm_extract_pi16(a, imm) |
|
#define | _m_pinsrw(a, i, imm) _mm_insert_pi16(a, i, imm) |
|
#define | _m_pmaxsw(a, b) _mm_max_pi16(a, b) |
|
#define | _m_pmaxub(a, b) _mm_max_pu8(a, b) |
|
#define | _m_pminsw(a, b) _mm_min_pi16(a, b) |
|
#define | _m_pminub(a, b) _mm_min_pu8(a, b) |
|
#define | _m_pmovmskb(a) _mm_movemask_pi8(a) |
|
#define | _m_pmulhuw(a, b) _mm_mulhi_pu16(a, b) |
|
#define | _m_psadbw(a, b) _mm_sad_pu8(a, b) |
|
#define | _m_pshufw(a, imm) _mm_shuffle_pi16(a, imm) |
|
#define | _mm_shuffle_pi16(a, imm) |
|
#define | _mm_shuffle_ps(a, b, imm) |
|
#define | _mm_store1_ps _mm_store_ps1 |
|
#define | _MM_TRANSPOSE4_PS(row0, row1, row2, row3) |
|
#define | _mm_ucomieq_ss _mm_comieq_ss |
|
#define | _mm_ucomige_ss _mm_comige_ss |
|
#define | _mm_ucomigt_ss _mm_comigt_ss |
|
#define | _mm_ucomile_ss _mm_comile_ss |
|
#define | _mm_ucomilt_ss _mm_comilt_ss |
|
#define | _mm_ucomineq_ss _mm_comineq_ss |
|
#define | _mm_bslli_si128(a, imm) _mm_slli_si128(a, imm) |
|
#define | _mm_bsrli_si128(a, imm) _mm_srli_si128(a, imm) |
|
#define | _mm_cvtsd_si64x _mm_cvtsd_si64 |
|
#define | _mm_cvtsi128_si64x(a) _mm_cvtsi128_si64(a) |
|
#define | _mm_cvtsi128_si64x(a) _mm_cvtsi128_si64(a) |
|
#define | _mm_cvtsi64x_si128(a) _mm_cvtsi64_si128(a) |
|
#define | _mm_cvtsi64x_sd(a, b) _mm_cvtsi64_sd(a, b) |
|
#define | _mm_cvttsd_si64x(a) _mm_cvttsd_si64(a) |
|
#define | _mm_extract_epi16(a, imm) vgetq_lane_u16(vreinterpretq_u16_m128i(a), (imm)) |
|
#define | _mm_insert_epi16(a, b, imm) |
|
#define | _mm_load_pd1 _mm_load1_pd |
|
#define | _mm_set_pd1 _mm_set1_pd |
|
#define | _mm_shuffle_epi32(a, imm) |
|
#define | _mm_shuffle_pd(a, b, imm8) |
|
#define | _mm_shufflehi_epi16(a, imm) _mm_shufflehi_epi16_function((a), (imm)) |
|
#define | _mm_shufflelo_epi16(a, imm) _mm_shufflelo_epi16_function((a), (imm)) |
|
#define | _mm_srai_epi32(a, imm) |
|
#define | _mm_srli_epi16(a, imm) |
|
#define | _mm_srli_epi32(a, imm) |
|
#define | _mm_srli_epi64(a, imm) |
|
#define | _mm_store1_pd _mm_store_pd1 |
|
#define | _mm_ucomieq_sd _mm_comieq_sd |
|
#define | _mm_ucomige_sd _mm_comige_sd |
|
#define | _mm_ucomigt_sd _mm_comigt_sd |
|
#define | _mm_ucomile_sd _mm_comile_sd |
|
#define | _mm_ucomilt_sd _mm_comilt_sd |
|
#define | _mm_ucomineq_sd _mm_comineq_sd |
|
#define | _mm_lddqu_si128 _mm_loadu_si128 |
|
#define | _mm_loaddup_pd _mm_load1_pd |
|
#define | _mm_alignr_pi8(a, b, imm) |
|
#define | _mm_blend_epi16(a, b, imm) |
|
#define | _mm_blend_pd(a, b, imm) |
|
#define | _mm_extract_epi32(a, imm) vgetq_lane_s32(vreinterpretq_s32_m128i(a), (imm)) |
|
#define | _mm_extract_epi64(a, imm) vgetq_lane_s64(vreinterpretq_s64_m128i(a), (imm)) |
|
#define | _mm_extract_epi8(a, imm) vgetq_lane_u8(vreinterpretq_u8_m128i(a), (imm)) |
|
#define | _mm_extract_ps(a, imm) vgetq_lane_s32(vreinterpretq_s32_m128(a), (imm)) |
|
#define | _mm_insert_epi32(a, b, imm) |
|
#define | _mm_insert_epi64(a, b, imm) |
|
#define | _mm_insert_epi8(a, b, imm) |
|
#define | _mm_insert_ps(a, b, imm8) |
|
#define | _mm_testnzc_si128(a, b) _mm_test_mix_ones_zeros(a, b) |
|
#define | SSE2NEON_AES_DATA(w) |
|
#define | SSE2NEON_AES_H0(x) (x) |
|
#define | SSE2NEON_AES_B2W(b0, b1, b2, b3) |
|
#define | SSE2NEON_AES_F2(x) ((x << 1) ^ (((x >> 7) & 1) * 0x011b /* WPOLY */)) |
|
#define | SSE2NEON_AES_F3(x) (SSE2NEON_AES_F2(x) ^ x) |
|
#define | SSE2NEON_AES_U0(p) SSE2NEON_AES_B2W(SSE2NEON_AES_F2(p), p, p, SSE2NEON_AES_F3(p)) |
|
#define | SSE2NEON_AES_U1(p) SSE2NEON_AES_B2W(SSE2NEON_AES_F3(p), SSE2NEON_AES_F2(p), p, p) |
|
#define | SSE2NEON_AES_U2(p) SSE2NEON_AES_B2W(p, SSE2NEON_AES_F3(p), SSE2NEON_AES_F2(p), p) |
|
#define | SSE2NEON_AES_U3(p) SSE2NEON_AES_B2W(p, p, SSE2NEON_AES_F3(p), SSE2NEON_AES_F2(p)) |
|
|
union | ALIGN_STRUCT (16) SIMDVec |
|
FORCE_INLINE unsigned int | _MM_GET_ROUNDING_MODE () |
|
FORCE_INLINE __m128 | _mm_move_ss (__m128, __m128) |
|
FORCE_INLINE __m128 | _mm_or_ps (__m128, __m128) |
|
FORCE_INLINE __m128 | _mm_set_ps1 (float) |
|
FORCE_INLINE __m128 | _mm_setzero_ps (void) |
|
FORCE_INLINE __m128i | _mm_and_si128 (__m128i, __m128i) |
|
FORCE_INLINE __m128i | _mm_castps_si128 (__m128) |
|
FORCE_INLINE __m128i | _mm_cmpeq_epi32 (__m128i, __m128i) |
|
FORCE_INLINE __m128i | _mm_cvtps_epi32 (__m128) |
|
FORCE_INLINE __m128d | _mm_move_sd (__m128d, __m128d) |
|
FORCE_INLINE __m128i | _mm_or_si128 (__m128i, __m128i) |
|
FORCE_INLINE __m128i | _mm_set_epi32 (int, int, int, int) |
|
FORCE_INLINE __m128i | _mm_set_epi64x (int64_t, int64_t) |
|
FORCE_INLINE __m128d | _mm_set_pd (double, double) |
|
FORCE_INLINE __m128i | _mm_set1_epi32 (int) |
|
FORCE_INLINE __m128i | _mm_setzero_si128 () |
|
FORCE_INLINE __m128d | _mm_ceil_pd (__m128d) |
|
FORCE_INLINE __m128 | _mm_ceil_ps (__m128) |
|
FORCE_INLINE __m128d | _mm_floor_pd (__m128d) |
|
FORCE_INLINE __m128 | _mm_floor_ps (__m128) |
|
FORCE_INLINE __m128d | _mm_round_pd (__m128d, int) |
|
FORCE_INLINE __m128 | _mm_round_ps (__m128, int) |
|
FORCE_INLINE uint32_t | _mm_crc32_u8 (uint32_t, uint8_t) |
|
FORCE_INLINE uint8x16x4_t | _sse2neon_vld1q_u8_x4 (const uint8_t *p) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_1032 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_2301 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_0321 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_2103 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_1010 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_1001 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_0101 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_3210 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_0011 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_0022 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_2200 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_3202 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_1133 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_2010 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_2001 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_2032 (__m128 a, __m128 b) |
|
FORCE_INLINE void | _sse2neon_kadd_f32 (float *sum, float *c, float y) |
|
static uint64x2_t | _sse2neon_vmull_p64 (uint64x1_t _a, uint64x1_t _b) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_1032 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_2301 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_0321 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_2103 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_1010 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_1001 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_0101 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_2211 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_0122 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_3332 (__m128i a) |
|
FORCE_INLINE void | _mm_empty (void) |
|
FORCE_INLINE __m128 | _mm_add_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_add_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_and_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_andnot_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m64 | _mm_avg_pu16 (__m64 a, __m64 b) |
|
FORCE_INLINE __m64 | _mm_avg_pu8 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128 | _mm_cmpeq_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpeq_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpge_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpge_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpgt_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpgt_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmple_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmple_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmplt_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmplt_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpneq_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpneq_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpnge_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpnge_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpngt_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpngt_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpnle_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpnle_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpnlt_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpnlt_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpord_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpord_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpunord_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpunord_ss (__m128 a, __m128 b) |
|
FORCE_INLINE int | _mm_comieq_ss (__m128 a, __m128 b) |
|
FORCE_INLINE int | _mm_comige_ss (__m128 a, __m128 b) |
|
FORCE_INLINE int | _mm_comigt_ss (__m128 a, __m128 b) |
|
FORCE_INLINE int | _mm_comile_ss (__m128 a, __m128 b) |
|
FORCE_INLINE int | _mm_comilt_ss (__m128 a, __m128 b) |
|
FORCE_INLINE int | _mm_comineq_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cvt_pi2ps (__m128 a, __m64 b) |
|
FORCE_INLINE __m64 | _mm_cvt_ps2pi (__m128 a) |
|
FORCE_INLINE __m128 | _mm_cvt_si2ss (__m128 a, int b) |
|
FORCE_INLINE int | _mm_cvt_ss2si (__m128 a) |
|
FORCE_INLINE __m128 | _mm_cvtpi16_ps (__m64 a) |
|
FORCE_INLINE __m128 | _mm_cvtpi32_ps (__m128 a, __m64 b) |
|
FORCE_INLINE __m128 | _mm_cvtpi32x2_ps (__m64 a, __m64 b) |
|
FORCE_INLINE __m128 | _mm_cvtpi8_ps (__m64 a) |
|
FORCE_INLINE __m64 | _mm_cvtps_pi16 (__m128 a) |
|
FORCE_INLINE __m64 | _mm_cvtps_pi8 (__m128 a) |
|
FORCE_INLINE __m128 | _mm_cvtpu16_ps (__m64 a) |
|
FORCE_INLINE __m128 | _mm_cvtpu8_ps (__m64 a) |
|
FORCE_INLINE __m128 | _mm_cvtsi64_ss (__m128 a, int64_t b) |
|
FORCE_INLINE float | _mm_cvtss_f32 (__m128 a) |
|
FORCE_INLINE int64_t | _mm_cvtss_si64 (__m128 a) |
|
FORCE_INLINE __m64 | _mm_cvtt_ps2pi (__m128 a) |
|
FORCE_INLINE int | _mm_cvtt_ss2si (__m128 a) |
|
FORCE_INLINE int64_t | _mm_cvttss_si64 (__m128 a) |
|
FORCE_INLINE __m128 | _mm_div_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_div_ss (__m128 a, __m128 b) |
|
FORCE_INLINE void | _mm_free (void *addr) |
|
FORCE_INLINE unsigned int | _sse2neon_mm_get_flush_zero_mode () |
|
FORCE_INLINE __m128 | _mm_load_ps (const float *p) |
|
FORCE_INLINE __m128 | _mm_load_ss (const float *p) |
|
FORCE_INLINE __m128 | _mm_load1_ps (const float *p) |
|
FORCE_INLINE __m128 | _mm_loadh_pi (__m128 a, __m64 const *p) |
|
FORCE_INLINE __m128 | _mm_loadl_pi (__m128 a, __m64 const *p) |
|
FORCE_INLINE __m128 | _mm_loadr_ps (const float *p) |
|
FORCE_INLINE __m128 | _mm_loadu_ps (const float *p) |
|
FORCE_INLINE __m128i | _mm_loadu_si16 (const void *p) |
|
FORCE_INLINE __m128i | _mm_loadu_si64 (const void *p) |
|
FORCE_INLINE void * | _mm_malloc (size_t size, size_t align) |
|
FORCE_INLINE void | _mm_maskmove_si64 (__m64 a, __m64 mask, char *mem_addr) |
|
FORCE_INLINE __m64 | _mm_max_pi16 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128 | _mm_max_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m64 | _mm_max_pu8 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128 | _mm_max_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m64 | _mm_min_pi16 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128 | _mm_min_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m64 | _mm_min_pu8 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128 | _mm_min_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_movehl_ps (__m128 __A, __m128 __B) |
|
FORCE_INLINE __m128 | _mm_movelh_ps (__m128 __A, __m128 __B) |
|
FORCE_INLINE int | _mm_movemask_pi8 (__m64 a) |
|
FORCE_INLINE int | _mm_movemask_ps (__m128 a) |
|
FORCE_INLINE __m128 | _mm_mul_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_mul_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m64 | _mm_mulhi_pu16 (__m64 a, __m64 b) |
|
FORCE_INLINE void | _mm_prefetch (const void *p, int i) |
|
FORCE_INLINE __m128 | _mm_rcp_ps (__m128 in) |
|
FORCE_INLINE __m128 | _mm_rcp_ss (__m128 a) |
|
FORCE_INLINE __m128 | _mm_rsqrt_ps (__m128 in) |
|
FORCE_INLINE __m128 | _mm_rsqrt_ss (__m128 in) |
|
FORCE_INLINE __m64 | _mm_sad_pu8 (__m64 a, __m64 b) |
|
FORCE_INLINE void | _sse2neon_mm_set_flush_zero_mode (unsigned int flag) |
|
FORCE_INLINE __m128 | _mm_set_ps (float w, float z, float y, float x) |
|
FORCE_INLINE void | _MM_SET_ROUNDING_MODE (int rounding) |
|
FORCE_INLINE __m128 | _mm_set_ss (float a) |
|
FORCE_INLINE __m128 | _mm_set1_ps (float _w) |
|
FORCE_INLINE void | _mm_setcsr (unsigned int a) |
|
FORCE_INLINE unsigned int | _mm_getcsr () |
|
FORCE_INLINE __m128 | _mm_setr_ps (float w, float z, float y, float x) |
|
FORCE_INLINE void | _mm_sfence (void) |
|
FORCE_INLINE __m128 | _mm_sqrt_ps (__m128 in) |
|
FORCE_INLINE __m128 | _mm_sqrt_ss (__m128 in) |
|
FORCE_INLINE void | _mm_store_ps (float *p, __m128 a) |
|
FORCE_INLINE void | _mm_store_ps1 (float *p, __m128 a) |
|
FORCE_INLINE void | _mm_store_ss (float *p, __m128 a) |
|
FORCE_INLINE void | _mm_storeh_pi (__m64 *p, __m128 a) |
|
FORCE_INLINE void | _mm_storel_pi (__m64 *p, __m128 a) |
|
FORCE_INLINE void | _mm_storer_ps (float *p, __m128 a) |
|
FORCE_INLINE void | _mm_storeu_ps (float *p, __m128 a) |
|
FORCE_INLINE void | _mm_storeu_si16 (void *p, __m128i a) |
|
FORCE_INLINE void | _mm_storeu_si64 (void *p, __m128i a) |
|
FORCE_INLINE void | _mm_stream_pi (__m64 *p, __m64 a) |
|
FORCE_INLINE void | _mm_stream_ps (float *p, __m128 a) |
|
FORCE_INLINE __m128 | _mm_sub_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_sub_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128i | _mm_undefined_si128 (void) |
|
FORCE_INLINE __m128 | _mm_undefined_ps (void) |
|
FORCE_INLINE __m128 | _mm_unpackhi_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_unpacklo_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_xor_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128i | _mm_add_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_add_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_add_epi64 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_add_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_add_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_add_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m64 | _mm_add_si64 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128i | _mm_adds_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_adds_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_adds_epu16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_adds_epu8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_and_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_andnot_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_andnot_si128 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_avg_epu16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_avg_epu8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128 | _mm_castpd_ps (__m128d a) |
|
FORCE_INLINE __m128i | _mm_castpd_si128 (__m128d a) |
|
FORCE_INLINE __m128d | _mm_castps_pd (__m128 a) |
|
FORCE_INLINE __m128d | _mm_castsi128_pd (__m128i a) |
|
FORCE_INLINE __m128 | _mm_castsi128_ps (__m128i a) |
|
FORCE_INLINE void | _mm_clflush (void const *p) |
|
FORCE_INLINE __m128i | _mm_cmpeq_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_cmpeq_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_cmpeq_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpeq_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpge_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpge_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_cmpgt_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_cmpgt_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_cmpgt_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_cmpgt_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpgt_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmple_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmple_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_cmplt_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_cmplt_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_cmplt_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_cmplt_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmplt_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpneq_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpneq_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpnge_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpnge_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpngt_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpngt_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpnle_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpnle_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpnlt_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpnlt_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpord_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpord_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpunord_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpunord_sd (__m128d a, __m128d b) |
|
FORCE_INLINE int | _mm_comige_sd (__m128d a, __m128d b) |
|
FORCE_INLINE int | _mm_comigt_sd (__m128d a, __m128d b) |
|
FORCE_INLINE int | _mm_comile_sd (__m128d a, __m128d b) |
|
FORCE_INLINE int | _mm_comilt_sd (__m128d a, __m128d b) |
|
FORCE_INLINE int | _mm_comieq_sd (__m128d a, __m128d b) |
|
FORCE_INLINE int | _mm_comineq_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cvtepi32_pd (__m128i a) |
|
FORCE_INLINE __m128 | _mm_cvtepi32_ps (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtpd_epi32 (__m128d a) |
|
FORCE_INLINE __m64 | _mm_cvtpd_pi32 (__m128d a) |
|
FORCE_INLINE __m128 | _mm_cvtpd_ps (__m128d a) |
|
FORCE_INLINE __m128d | _mm_cvtpi32_pd (__m64 a) |
|
FORCE_INLINE __m128d | _mm_cvtps_pd (__m128 a) |
|
FORCE_INLINE double | _mm_cvtsd_f64 (__m128d a) |
|
FORCE_INLINE int32_t | _mm_cvtsd_si32 (__m128d a) |
|
FORCE_INLINE int64_t | _mm_cvtsd_si64 (__m128d a) |
|
FORCE_INLINE __m128 | _mm_cvtsd_ss (__m128 a, __m128d b) |
|
FORCE_INLINE int | _mm_cvtsi128_si32 (__m128i a) |
|
FORCE_INLINE int64_t | _mm_cvtsi128_si64 (__m128i a) |
|
FORCE_INLINE __m128d | _mm_cvtsi32_sd (__m128d a, int32_t b) |
|
FORCE_INLINE __m128i | _mm_cvtsi32_si128 (int a) |
|
FORCE_INLINE __m128d | _mm_cvtsi64_sd (__m128d a, int64_t b) |
|
FORCE_INLINE __m128i | _mm_cvtsi64_si128 (int64_t a) |
|
FORCE_INLINE __m128d | _mm_cvtss_sd (__m128d a, __m128 b) |
|
FORCE_INLINE __m128i | _mm_cvttpd_epi32 (__m128d a) |
|
FORCE_INLINE __m64 | _mm_cvttpd_pi32 (__m128d a) |
|
FORCE_INLINE __m128i | _mm_cvttps_epi32 (__m128 a) |
|
FORCE_INLINE int32_t | _mm_cvttsd_si32 (__m128d a) |
|
FORCE_INLINE int64_t | _mm_cvttsd_si64 (__m128d a) |
|
FORCE_INLINE __m128d | _mm_div_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_div_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_load_pd (const double *p) |
|
FORCE_INLINE __m128d | _mm_load_sd (const double *p) |
|
FORCE_INLINE __m128i | _mm_load_si128 (const __m128i *p) |
|
FORCE_INLINE __m128d | _mm_load1_pd (const double *p) |
|
FORCE_INLINE __m128d | _mm_loadh_pd (__m128d a, const double *p) |
|
FORCE_INLINE __m128i | _mm_loadl_epi64 (__m128i const *p) |
|
FORCE_INLINE __m128d | _mm_loadl_pd (__m128d a, const double *p) |
|
FORCE_INLINE __m128d | _mm_loadr_pd (const double *p) |
|
FORCE_INLINE __m128d | _mm_loadu_pd (const double *p) |
|
FORCE_INLINE __m128i | _mm_loadu_si128 (const __m128i *p) |
|
FORCE_INLINE __m128i | _mm_loadu_si32 (const void *p) |
|
FORCE_INLINE __m128i | _mm_madd_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE void | _mm_maskmoveu_si128 (__m128i a, __m128i mask, char *mem_addr) |
|
FORCE_INLINE __m128i | _mm_max_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_max_epu8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_max_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_max_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_min_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_min_epu8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_min_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_min_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_move_epi64 (__m128i a) |
|
FORCE_INLINE int | _mm_movemask_epi8 (__m128i a) |
|
FORCE_INLINE int | _mm_movemask_pd (__m128d a) |
|
FORCE_INLINE __m64 | _mm_movepi64_pi64 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_movpi64_epi64 (__m64 a) |
|
FORCE_INLINE __m128i | _mm_mul_epu32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_mul_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_mul_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m64 | _mm_mul_su32 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128i | _mm_mulhi_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_mulhi_epu16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_mullo_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_or_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_packs_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_packs_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_packus_epi16 (const __m128i a, const __m128i b) |
|
FORCE_INLINE void | _mm_pause () |
|
FORCE_INLINE __m128i | _mm_sad_epu8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_set_epi16 (short i7, short i6, short i5, short i4, short i3, short i2, short i1, short i0) |
|
FORCE_INLINE __m128i | _mm_set_epi64 (__m64 i1, __m64 i2) |
|
FORCE_INLINE __m128i | _mm_set_epi8 (signed char b15, signed char b14, signed char b13, signed char b12, signed char b11, signed char b10, signed char b9, signed char b8, signed char b7, signed char b6, signed char b5, signed char b4, signed char b3, signed char b2, signed char b1, signed char b0) |
|
FORCE_INLINE __m128d | _mm_set_sd (double a) |
|
FORCE_INLINE __m128i | _mm_set1_epi16 (short w) |
|
FORCE_INLINE __m128i | _mm_set1_epi64 (__m64 _i) |
|
FORCE_INLINE __m128i | _mm_set1_epi64x (int64_t _i) |
|
FORCE_INLINE __m128i | _mm_set1_epi8 (signed char w) |
|
FORCE_INLINE __m128d | _mm_set1_pd (double d) |
|
FORCE_INLINE __m128i | _mm_setr_epi16 (short w0, short w1, short w2, short w3, short w4, short w5, short w6, short w7) |
|
FORCE_INLINE __m128i | _mm_setr_epi32 (int i3, int i2, int i1, int i0) |
|
FORCE_INLINE __m128i | _mm_setr_epi64 (__m64 e1, __m64 e0) |
|
FORCE_INLINE __m128i | _mm_setr_epi8 (signed char b0, signed char b1, signed char b2, signed char b3, signed char b4, signed char b5, signed char b6, signed char b7, signed char b8, signed char b9, signed char b10, signed char b11, signed char b12, signed char b13, signed char b14, signed char b15) |
|
FORCE_INLINE __m128d | _mm_setr_pd (double e1, double e0) |
|
FORCE_INLINE __m128d | _mm_setzero_pd (void) |
|
FORCE_INLINE __m128i | _mm_sll_epi16 (__m128i a, __m128i count) |
|
FORCE_INLINE __m128i | _mm_sll_epi32 (__m128i a, __m128i count) |
|
FORCE_INLINE __m128i | _mm_sll_epi64 (__m128i a, __m128i count) |
|
FORCE_INLINE __m128i | _mm_slli_epi16 (__m128i a, int imm) |
|
FORCE_INLINE __m128i | _mm_slli_epi32 (__m128i a, int imm) |
|
FORCE_INLINE __m128i | _mm_slli_epi64 (__m128i a, int imm) |
|
FORCE_INLINE __m128i | _mm_slli_si128 (__m128i a, int imm) |
|
FORCE_INLINE __m128d | _mm_sqrt_pd (__m128d a) |
|
FORCE_INLINE __m128d | _mm_sqrt_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_sra_epi16 (__m128i a, __m128i count) |
|
FORCE_INLINE __m128i | _mm_sra_epi32 (__m128i a, __m128i count) |
|
FORCE_INLINE __m128i | _mm_srai_epi16 (__m128i a, int imm) |
|
FORCE_INLINE __m128i | _mm_srl_epi16 (__m128i a, __m128i count) |
|
FORCE_INLINE __m128i | _mm_srl_epi32 (__m128i a, __m128i count) |
|
FORCE_INLINE __m128i | _mm_srl_epi64 (__m128i a, __m128i count) |
|
FORCE_INLINE __m128i | _mm_srli_si128 (__m128i a, int imm) |
|
FORCE_INLINE void | _mm_store_pd (double *mem_addr, __m128d a) |
|
FORCE_INLINE void | _mm_store_pd1 (double *mem_addr, __m128d a) |
|
FORCE_INLINE void | _mm_store_sd (double *mem_addr, __m128d a) |
|
FORCE_INLINE void | _mm_store_si128 (__m128i *p, __m128i a) |
|
FORCE_INLINE void | _mm_storeh_pd (double *mem_addr, __m128d a) |
|
FORCE_INLINE void | _mm_storel_epi64 (__m128i *a, __m128i b) |
|
FORCE_INLINE void | _mm_storel_pd (double *mem_addr, __m128d a) |
|
FORCE_INLINE void | _mm_storer_pd (double *mem_addr, __m128d a) |
|
FORCE_INLINE void | _mm_storeu_pd (double *mem_addr, __m128d a) |
|
FORCE_INLINE void | _mm_storeu_si128 (__m128i *p, __m128i a) |
|
FORCE_INLINE void | _mm_storeu_si32 (void *p, __m128i a) |
|
FORCE_INLINE void | _mm_stream_pd (double *p, __m128d a) |
|
FORCE_INLINE void | _mm_stream_si128 (__m128i *p, __m128i a) |
|
FORCE_INLINE void | _mm_stream_si32 (int *p, int a) |
|
FORCE_INLINE void | _mm_stream_si64 (__int64 *p, __int64 a) |
|
FORCE_INLINE __m128i | _mm_sub_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_sub_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_sub_epi64 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_sub_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_sub_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_sub_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m64 | _mm_sub_si64 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128i | _mm_subs_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_subs_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_subs_epu16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_subs_epu8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_undefined_pd (void) |
|
FORCE_INLINE __m128i | _mm_unpackhi_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_unpackhi_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_unpackhi_epi64 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_unpackhi_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_unpackhi_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_unpacklo_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_unpacklo_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_unpacklo_epi64 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_unpacklo_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_unpacklo_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_xor_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_xor_si128 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_addsub_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128 | _mm_addsub_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128d | _mm_hadd_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128 | _mm_hadd_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128d | _mm_hsub_pd (__m128d _a, __m128d _b) |
|
FORCE_INLINE __m128 | _mm_hsub_ps (__m128 _a, __m128 _b) |
|
FORCE_INLINE __m128d | _mm_movedup_pd (__m128d a) |
|
FORCE_INLINE __m128 | _mm_movehdup_ps (__m128 a) |
|
FORCE_INLINE __m128 | _mm_moveldup_ps (__m128 a) |
|
FORCE_INLINE __m128i | _mm_abs_epi16 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_abs_epi32 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_abs_epi8 (__m128i a) |
|
FORCE_INLINE __m64 | _mm_abs_pi16 (__m64 a) |
|
FORCE_INLINE __m64 | _mm_abs_pi32 (__m64 a) |
|
FORCE_INLINE __m64 | _mm_abs_pi8 (__m64 a) |
|
FORCE_INLINE __m128i | _mm_alignr_epi8 (__m128i a, __m128i b, int imm) |
|
FORCE_INLINE __m128i | _mm_hadd_epi16 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m128i | _mm_hadd_epi32 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m64 | _mm_hadd_pi16 (__m64 a, __m64 b) |
|
FORCE_INLINE __m64 | _mm_hadd_pi32 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128i | _mm_hadds_epi16 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m64 | _mm_hadds_pi16 (__m64 _a, __m64 _b) |
|
FORCE_INLINE __m128i | _mm_hsub_epi16 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m128i | _mm_hsub_epi32 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m64 | _mm_hsub_pi16 (__m64 _a, __m64 _b) |
|
FORCE_INLINE __m64 | _mm_hsub_pi32 (__m64 _a, __m64 _b) |
|
FORCE_INLINE __m128i | _mm_hsubs_epi16 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m64 | _mm_hsubs_pi16 (__m64 _a, __m64 _b) |
|
FORCE_INLINE __m128i | _mm_maddubs_epi16 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m64 | _mm_maddubs_pi16 (__m64 _a, __m64 _b) |
|
FORCE_INLINE __m128i | _mm_mulhrs_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m64 | _mm_mulhrs_pi16 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m64 | _mm_shuffle_pi8 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128i | _mm_sign_epi16 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m128i | _mm_sign_epi32 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m128i | _mm_sign_epi8 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m64 | _mm_sign_pi16 (__m64 _a, __m64 _b) |
|
FORCE_INLINE __m64 | _mm_sign_pi32 (__m64 _a, __m64 _b) |
|
FORCE_INLINE __m64 | _mm_sign_pi8 (__m64 _a, __m64 _b) |
|
FORCE_INLINE __m128 | _mm_blend_ps (__m128 _a, __m128 _b, const char imm8) |
|
FORCE_INLINE __m128i | _mm_blendv_epi8 (__m128i _a, __m128i _b, __m128i _mask) |
|
FORCE_INLINE __m128d | _mm_blendv_pd (__m128d _a, __m128d _b, __m128d _mask) |
|
FORCE_INLINE __m128 | _mm_blendv_ps (__m128 _a, __m128 _b, __m128 _mask) |
|
FORCE_INLINE __m128d | _mm_ceil_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128 | _mm_ceil_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128i | _mm_cmpeq_epi64 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_cvtepi16_epi32 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepi16_epi64 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepi32_epi64 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepi8_epi16 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepi8_epi32 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepi8_epi64 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepu16_epi32 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepu16_epi64 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepu32_epi64 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepu8_epi16 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepu8_epi32 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepu8_epi64 (__m128i a) |
|
FORCE_INLINE __m128d | _mm_dp_pd (__m128d a, __m128d b, const int imm) |
|
FORCE_INLINE __m128 | _mm_dp_ps (__m128 a, __m128 b, const int imm) |
|
FORCE_INLINE __m128d | _mm_floor_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128 | _mm_floor_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128i | _mm_max_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_max_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_max_epu16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_max_epu32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_min_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_min_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_min_epu16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_min_epu32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_minpos_epu16 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_mpsadbw_epu8 (__m128i a, __m128i b, const int imm) |
|
FORCE_INLINE __m128i | _mm_mul_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_mullo_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_packus_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_round_sd (__m128d a, __m128d b, int rounding) |
|
FORCE_INLINE __m128 | _mm_round_ss (__m128 a, __m128 b, int rounding) |
|
FORCE_INLINE __m128i | _mm_stream_load_si128 (__m128i *p) |
|
FORCE_INLINE int | _mm_test_all_ones (__m128i a) |
|
FORCE_INLINE int | _mm_test_all_zeros (__m128i a, __m128i mask) |
|
FORCE_INLINE int | _mm_test_mix_ones_zeros (__m128i a, __m128i mask) |
|
FORCE_INLINE int | _mm_testc_si128 (__m128i a, __m128i b) |
|
FORCE_INLINE int | _mm_testz_si128 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_cmpgt_epi64 (__m128i a, __m128i b) |
|
FORCE_INLINE uint32_t | _mm_crc32_u16 (uint32_t crc, uint16_t v) |
|
FORCE_INLINE uint32_t | _mm_crc32_u32 (uint32_t crc, uint32_t v) |
|
FORCE_INLINE uint64_t | _mm_crc32_u64 (uint64_t crc, uint64_t v) |
|
FORCE_INLINE __m128i | _mm_aesenc_si128 (__m128i EncBlock, __m128i RoundKey) |
|
FORCE_INLINE __m128i | _mm_aesenclast_si128 (__m128i a, __m128i RoundKey) |
|
FORCE_INLINE __m128i | _mm_aeskeygenassist_si128 (__m128i key, const int rcon) |
|
FORCE_INLINE __m128i | _mm_clmulepi64_si128 (__m128i _a, __m128i _b, const int imm) |
|
FORCE_INLINE unsigned int | _sse2neon_mm_get_denormals_zero_mode () |
|
FORCE_INLINE int | _mm_popcnt_u32 (unsigned int a) |
|
FORCE_INLINE int64_t | _mm_popcnt_u64 (uint64_t a) |
|
FORCE_INLINE void | _sse2neon_mm_set_denormals_zero_mode (unsigned int flag) |
|
FORCE_INLINE uint64_t | _rdtsc (void) |
|