diff options
Diffstat (limited to 'aom_dsp/x86/loopfilter_avx2.c')
-rw-r--r-- | aom_dsp/x86/loopfilter_avx2.c | 516 |
1 files changed, 307 insertions, 209 deletions
diff --git a/aom_dsp/x86/loopfilter_avx2.c b/aom_dsp/x86/loopfilter_avx2.c index b59381928..6e77742e3 100644 --- a/aom_dsp/x86/loopfilter_avx2.c +++ b/aom_dsp/x86/loopfilter_avx2.c @@ -32,7 +32,7 @@ void aom_lpf_horizontal_6_quad_avx2(unsigned char *s, int p, _mm_broadcastb_epi8(_mm_cvtsi32_si128((int)_limit0[0])); const __m128i blimit_v = _mm_broadcastb_epi8(_mm_cvtsi32_si128((int)_blimit0[0])); - const __m128i zero = _mm_set1_epi16(0); + const __m128i zero = _mm_setzero_si128(); const __m128i ff = _mm_cmpeq_epi8(zero, zero); p256_2 = @@ -239,7 +239,7 @@ void aom_lpf_horizontal_8_quad_avx2(unsigned char *s, int p, _mm_broadcastb_epi8(_mm_cvtsi32_si128((int)_limit0[0])); const __m128i blimit_v = _mm_broadcastb_epi8(_mm_cvtsi32_si128((int)_blimit0[0])); - const __m128i zero = _mm_set1_epi16(0); + const __m128i zero = _mm_setzero_si128(); const __m128i ff = _mm_cmpeq_epi8(zero, zero); p256_3 = @@ -472,51 +472,162 @@ void aom_lpf_horizontal_8_quad_avx2(unsigned char *s, int p, } } +static INLINE void trans_store_16x16_lpf_vert14(unsigned char *in0, int in_p, + unsigned char *out, int out_p, + int is_store_avx2) { + const __m128i x0 = _mm_loadu_si128((__m128i *)in0); + const __m128i x1 = _mm_loadu_si128((__m128i *)(in0 + in_p * 1)); + const __m128i x2 = _mm_loadu_si128((__m128i *)(in0 + in_p * 2)); + const __m128i x3 = _mm_loadu_si128((__m128i *)(in0 + in_p * 3)); + const __m128i x4 = _mm_loadu_si128((__m128i *)(in0 + in_p * 4)); + const __m128i x5 = _mm_loadu_si128((__m128i *)(in0 + in_p * 5)); + const __m128i x6 = _mm_loadu_si128((__m128i *)(in0 + in_p * 6)); + const __m128i x7 = _mm_loadu_si128((__m128i *)(in0 + in_p * 7)); + + const __m256i y0 = _mm256_insertf128_si256( + _mm256_castsi128_si256(x0), _mm_loadu_si128((__m128i *)(in0 + in_p * 8)), + 0x1); + const __m256i y1 = _mm256_insertf128_si256( + _mm256_castsi128_si256(x1), _mm_loadu_si128((__m128i *)(in0 + in_p * 9)), + 0x1); + const __m256i y2 = _mm256_insertf128_si256( + _mm256_castsi128_si256(x2), _mm_loadu_si128((__m128i *)(in0 + in_p * 10)), + 0x1); + const __m256i y3 = _mm256_insertf128_si256( + _mm256_castsi128_si256(x3), _mm_loadu_si128((__m128i *)(in0 + in_p * 11)), + 0x1); + const __m256i y4 = _mm256_insertf128_si256( + _mm256_castsi128_si256(x4), _mm_loadu_si128((__m128i *)(in0 + in_p * 12)), + 0x1); + const __m256i y5 = _mm256_insertf128_si256( + _mm256_castsi128_si256(x5), _mm_loadu_si128((__m128i *)(in0 + in_p * 13)), + 0x1); + const __m256i y6 = _mm256_insertf128_si256( + _mm256_castsi128_si256(x6), _mm_loadu_si128((__m128i *)(in0 + in_p * 14)), + 0x1); + const __m256i y7 = _mm256_insertf128_si256( + _mm256_castsi128_si256(x7), _mm_loadu_si128((__m128i *)(in0 + in_p * 15)), + 0x1); + + const __m256i y_s00 = _mm256_unpacklo_epi8(y0, y1); + const __m256i y_s01 = _mm256_unpackhi_epi8(y0, y1); + const __m256i y_s02 = _mm256_unpacklo_epi8(y2, y3); + const __m256i y_s03 = _mm256_unpackhi_epi8(y2, y3); + const __m256i y_s04 = _mm256_unpacklo_epi8(y4, y5); + const __m256i y_s05 = _mm256_unpackhi_epi8(y4, y5); + const __m256i y_s06 = _mm256_unpacklo_epi8(y6, y7); + const __m256i y_s07 = _mm256_unpackhi_epi8(y6, y7); + + const __m256i y_s10 = _mm256_unpacklo_epi16(y_s00, y_s02); + const __m256i y_s11 = _mm256_unpackhi_epi16(y_s00, y_s02); + const __m256i y_s12 = _mm256_unpacklo_epi16(y_s01, y_s03); + const __m256i y_s13 = _mm256_unpackhi_epi16(y_s01, y_s03); + const __m256i y_s14 = _mm256_unpacklo_epi16(y_s04, y_s06); + const __m256i y_s15 = _mm256_unpackhi_epi16(y_s04, y_s06); + const __m256i y_s16 = _mm256_unpacklo_epi16(y_s05, y_s07); + const __m256i y_s17 = _mm256_unpackhi_epi16(y_s05, y_s07); + + const __m256i y_s20 = _mm256_unpacklo_epi32(y_s10, y_s14); + const __m256i y_s21 = _mm256_unpackhi_epi32(y_s10, y_s14); + const __m256i y_s22 = _mm256_unpacklo_epi32(y_s11, y_s15); + const __m256i y_s23 = _mm256_unpackhi_epi32(y_s11, y_s15); + const __m256i y_s24 = _mm256_unpacklo_epi32(y_s12, y_s16); + const __m256i y_s25 = _mm256_unpackhi_epi32(y_s12, y_s16); + const __m256i y_s26 = _mm256_unpacklo_epi32(y_s13, y_s17); + const __m256i y_s27 = _mm256_unpackhi_epi32(y_s13, y_s17); + + const __m256i row_s01 = _mm256_permute4x64_epi64(y_s20, 0xd8); + const __m256i row_s23 = _mm256_permute4x64_epi64(y_s21, 0xd8); + const __m256i row_s45 = _mm256_permute4x64_epi64(y_s22, 0xd8); + const __m256i row_s67 = _mm256_permute4x64_epi64(y_s23, 0xd8); + const __m256i row_s89 = _mm256_permute4x64_epi64(y_s24, 0xd8); + const __m256i row_s1011 = _mm256_permute4x64_epi64(y_s25, 0xd8); + const __m256i row_s1213 = _mm256_permute4x64_epi64(y_s26, 0xd8); + const __m256i row_s1415 = _mm256_permute4x64_epi64(y_s27, 0xd8); + + if (is_store_avx2) { + _mm256_storeu_si256((__m256i *)(out), row_s01); + _mm256_storeu_si256((__m256i *)(out + (2 * out_p)), row_s23); + _mm256_storeu_si256((__m256i *)(out + (4 * out_p)), row_s45); + _mm256_storeu_si256((__m256i *)(out + (6 * out_p)), row_s67); + _mm256_storeu_si256((__m256i *)(out + (8 * out_p)), row_s89); + _mm256_storeu_si256((__m256i *)(out + (10 * out_p)), row_s1011); + _mm256_storeu_si256((__m256i *)(out + (12 * out_p)), row_s1213); + _mm256_storeu_si256((__m256i *)(out + (14 * out_p)), row_s1415); + } else { + _mm_storeu_si128((__m128i *)(out), _mm256_castsi256_si128(row_s01)); + _mm_storeu_si128((__m128i *)(out + (2 * out_p)), + _mm256_castsi256_si128(row_s23)); + _mm_storeu_si128((__m128i *)(out + (4 * out_p)), + _mm256_castsi256_si128(row_s45)); + _mm_storeu_si128((__m128i *)(out + (6 * out_p)), + _mm256_castsi256_si128(row_s67)); + _mm_storeu_si128((__m128i *)(out + (8 * out_p)), + _mm256_castsi256_si128(row_s89)); + _mm_storeu_si128((__m128i *)(out + (10 * out_p)), + _mm256_castsi256_si128(row_s1011)); + _mm_storeu_si128((__m128i *)(out + (12 * out_p)), + _mm256_castsi256_si128(row_s1213)); + _mm_storeu_si128((__m128i *)(out + (14 * out_p)), + _mm256_castsi256_si128(row_s1415)); + _mm_storeu_si128((__m128i *)(out + (1 * out_p)), + _mm256_extracti128_si256(row_s01, 1)); + _mm_storeu_si128((__m128i *)(out + (3 * out_p)), + _mm256_extracti128_si256(row_s23, 1)); + _mm_storeu_si128((__m128i *)(out + (5 * out_p)), + _mm256_extracti128_si256(row_s45, 1)); + _mm_storeu_si128((__m128i *)(out + (7 * out_p)), + _mm256_extracti128_si256(row_s67, 1)); + _mm_storeu_si128((__m128i *)(out + (9 * out_p)), + _mm256_extracti128_si256(row_s89, 1)); + _mm_storeu_si128((__m128i *)(out + (11 * out_p)), + _mm256_extracti128_si256(row_s1011, 1)); + _mm_storeu_si128((__m128i *)(out + (13 * out_p)), + _mm256_extracti128_si256(row_s1213, 1)); + _mm_storeu_si128((__m128i *)(out + (15 * out_p)), + _mm256_extracti128_si256(row_s1415, 1)); + } +} + void aom_lpf_horizontal_14_quad_avx2(unsigned char *s, int p, const unsigned char *_blimit0, const unsigned char *_limit0, const unsigned char *_thresh0) { - __m256i p256_3, q256_3, p256_2, q256_2, p256_1, q256_1, p256_0, q256_0; - __m128i p3, p2, p1, p0, q0, q1, q2, q3; - __m128i mask, flat, flat2; - - const __m128i thresh_v = - _mm_broadcastb_epi8(_mm_cvtsi32_si128((int)_thresh0[0])); - const __m128i limit_v = - _mm_broadcastb_epi8(_mm_cvtsi32_si128((int)_limit0[0])); - const __m128i blimit_v = - _mm_broadcastb_epi8(_mm_cvtsi32_si128((int)_blimit0[0])); - const __m128i zero = _mm_set1_epi16(0); + __m128i mask, flat; + const __m128i zero = _mm_setzero_si128(); const __m128i ff = _mm_cmpeq_epi8(zero, zero); - p256_3 = + __m256i p256_3 = _mm256_castpd_si256(_mm256_broadcast_pd((__m128d const *)(s - 4 * p))); - p256_2 = + __m256i p256_2 = _mm256_castpd_si256(_mm256_broadcast_pd((__m128d const *)(s - 3 * p))); - p256_1 = + __m256i p256_1 = _mm256_castpd_si256(_mm256_broadcast_pd((__m128d const *)(s - 2 * p))); - p256_0 = + __m256i p256_0 = _mm256_castpd_si256(_mm256_broadcast_pd((__m128d const *)(s - 1 * p))); - q256_0 = + __m256i q256_0 = _mm256_castpd_si256(_mm256_broadcast_pd((__m128d const *)(s - 0 * p))); - q256_1 = + __m256i q256_1 = _mm256_castpd_si256(_mm256_broadcast_pd((__m128d const *)(s + 1 * p))); - q256_2 = + __m256i q256_2 = _mm256_castpd_si256(_mm256_broadcast_pd((__m128d const *)(s + 2 * p))); - q256_3 = + __m256i q256_3 = _mm256_castpd_si256(_mm256_broadcast_pd((__m128d const *)(s + 3 * p))); - p3 = _mm256_castsi256_si128(p256_3); - p2 = _mm256_castsi256_si128(p256_2); - p1 = _mm256_castsi256_si128(p256_1); - p0 = _mm256_castsi256_si128(p256_0); - q0 = _mm256_castsi256_si128(q256_0); - q1 = _mm256_castsi256_si128(q256_1); - q2 = _mm256_castsi256_si128(q256_2); - q3 = _mm256_castsi256_si128(q256_3); + __m128i p3 = _mm256_castsi256_si128(p256_3); + __m128i p2 = _mm256_castsi256_si128(p256_2); + __m128i p1 = _mm256_castsi256_si128(p256_1); + __m128i p0 = _mm256_castsi256_si128(p256_0); + __m128i q0 = _mm256_castsi256_si128(q256_0); + __m128i q1 = _mm256_castsi256_si128(q256_1); + __m128i q2 = _mm256_castsi256_si128(q256_2); + __m128i q3 = _mm256_castsi256_si128(q256_3); { - __m128i work; + const __m128i limit_v = + _mm_broadcastb_epi8(_mm_cvtsi32_si128((int)_limit0[0])); + const __m128i blimit_v = + _mm_broadcastb_epi8(_mm_cvtsi32_si128((int)_blimit0[0])); const __m128i fe = _mm_set1_epi8((int8_t)0xfe); const __m128i abs_p1p0 = _mm_or_si128(_mm_subs_epu8(p1, p0), _mm_subs_epu8(p0, p1)); @@ -537,7 +648,7 @@ void aom_lpf_horizontal_14_quad_avx2(unsigned char *s, int p, mask = _mm_max_epu8(flat, mask); // mask |= (abs(p1 - p0) > limit) * -1; // mask |= (abs(q1 - q0) > limit) * -1; - work = _mm_max_epu8( + __m128i work = _mm_max_epu8( _mm_or_si128(_mm_subs_epu8(p2, p1), _mm_subs_epu8(p1, p2)), _mm_or_si128(_mm_subs_epu8(p3, p2), _mm_subs_epu8(p2, p3))); mask = _mm_max_epu8(work, mask); @@ -553,39 +664,33 @@ void aom_lpf_horizontal_14_quad_avx2(unsigned char *s, int p, // loop filter { - const __m128i t4 = _mm_set1_epi8(4); + const __m128i thresh_v = + _mm_broadcastb_epi8(_mm_cvtsi32_si128((int)_thresh0[0])); + const __m128i one = _mm_set1_epi8(1); const __m128i t3 = _mm_set1_epi8(3); + const __m128i t4 = _mm_add_epi8(one, t3); const __m128i t80 = _mm_set1_epi8((int8_t)0x80); const __m128i te0 = _mm_set1_epi8((int8_t)0xe0); const __m128i t1f = _mm_set1_epi8(0x1f); - const __m128i t1 = _mm_set1_epi8(0x1); - const __m128i t7f = _mm_set1_epi8(0x7f); - const __m128i one = _mm_set1_epi8(1); - __m128i hev; + const __m128i t7f = _mm_sub_epi8(t80, one); - hev = _mm_subs_epu8(flat, thresh_v); + __m128i hev = _mm_subs_epu8(flat, thresh_v); hev = _mm_xor_si128(_mm_cmpeq_epi8(hev, zero), ff); __m128i ps1 = _mm_xor_si128(p1, t80); __m128i ps0 = _mm_xor_si128(p0, t80); __m128i qs0 = _mm_xor_si128(q0, t80); __m128i qs1 = _mm_xor_si128(q1, t80); - __m128i filt; - __m128i work_a; - __m128i filter1, filter2; - __m128i flat2_p5, flat2_p4, flat2_p3, flat2_p2, flat2_p1, flat2_p0, - flat2_q0, flat2_q1, flat2_q2, flat2_q3, flat2_q4, flat2_q5, flat_p2, - flat_p1, flat_p0, flat_q0, flat_q1, flat_q2; - filt = _mm_and_si128(_mm_subs_epi8(ps1, qs1), hev); - work_a = _mm_subs_epi8(qs0, ps0); + __m128i filt = _mm_and_si128(_mm_subs_epi8(ps1, qs1), hev); + __m128i work_a = _mm_subs_epi8(qs0, ps0); filt = _mm_adds_epi8(filt, work_a); filt = _mm_adds_epi8(filt, work_a); filt = _mm_adds_epi8(filt, work_a); filt = _mm_and_si128(filt, mask); - filter1 = _mm_adds_epi8(filt, t4); - filter2 = _mm_adds_epi8(filt, t3); + __m128i filter1 = _mm_adds_epi8(filt, t4); + __m128i filter2 = _mm_adds_epi8(filt, t3); work_a = _mm_cmpgt_epi8(zero, filter1); filter1 = _mm_srli_epi16(filter1, 3); @@ -601,7 +706,7 @@ void aom_lpf_horizontal_14_quad_avx2(unsigned char *s, int p, filter2 = _mm_or_si128(filter2, work_a); ps0 = _mm_xor_si128(_mm_adds_epi8(ps0, filter2), t80); - filt = _mm_adds_epi8(filter1, t1); + filt = _mm_adds_epi8(filter1, one); work_a = _mm_cmpgt_epi8(zero, filt); filt = _mm_srli_epi16(filt, 1); work_a = _mm_and_si128(work_a, t80); @@ -611,15 +716,23 @@ void aom_lpf_horizontal_14_quad_avx2(unsigned char *s, int p, ps1 = _mm_xor_si128(_mm_adds_epi8(ps1, filt), t80); qs1 = _mm_xor_si128(_mm_subs_epi8(qs1, filt), t80); - __m128i work; - work = _mm_max_epu8( - _mm_or_si128(_mm_subs_epu8(p2, p0), _mm_subs_epu8(p0, p2)), - _mm_or_si128(_mm_subs_epu8(q2, q0), _mm_subs_epu8(q0, q2))); - flat = _mm_max_epu8(work, flat); - work = _mm_max_epu8( - _mm_or_si128(_mm_subs_epu8(p3, p0), _mm_subs_epu8(p0, p3)), - _mm_or_si128(_mm_subs_epu8(q3, q0), _mm_subs_epu8(q0, q3))); - flat = _mm_max_epu8(work, flat); + // Derive flat + __m256i p0q0256 = _mm256_blend_epi32(p256_0, q256_0, 0xf0); + __m256i p2q2256 = _mm256_blend_epi32(p256_2, q256_2, 0xf0); + __m256i p3q3256 = _mm256_blend_epi32(p256_3, q256_3, 0xf0); + const __m256i ps0qs0256 = + _mm256_insertf128_si256(_mm256_castsi128_si256(ps0), qs0, 0x1); + const __m256i ps1qs1256 = + _mm256_insertf128_si256(_mm256_castsi128_si256(ps1), qs1, 0x1); + const __m256i work01 = _mm256_or_si256(_mm256_subs_epu8(p2q2256, p0q0256), + _mm256_subs_epu8(p0q0256, p2q2256)); + const __m256i work02 = _mm256_or_si256(_mm256_subs_epu8(p3q3256, p0q0256), + _mm256_subs_epu8(p0q0256, p3q3256)); + const __m256i max0_256 = _mm256_max_epu8(work01, work02); + const __m128i max1_256 = + _mm_max_epu8(_mm256_castsi256_si128(max0_256), + _mm256_extractf128_si256(max0_256, 1)); + flat = _mm_max_epu8(max1_256, flat); flat = _mm_subs_epu8(flat, one); flat = _mm_cmpeq_epi8(flat, zero); flat = _mm_and_si128(flat, mask); @@ -627,47 +740,38 @@ void aom_lpf_horizontal_14_quad_avx2(unsigned char *s, int p, // ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ // flat and wide flat calculations if (0xffff != _mm_movemask_epi8(_mm_cmpeq_epi8(flat, zero))) { + const __m256i flat256 = + _mm256_insertf128_si256(_mm256_castsi128_si256(flat), flat, 0x1); const __m256i eight = _mm256_set1_epi16(8); const __m256i four = _mm256_set1_epi16(4); - __m256i pixelFilter_p, pixelFilter_q, pixetFilter_p2p1p0, p2p1p0, q2q1q0, - pixetFilter_q2q1q0, sum_p, sum_q, res_p, res_q; - __m256i p256_4, q256_4, p256_5, q256_5, p256_6, q256_6; - __m128i p4, q4, p5, q5, p6, q6; - p256_4 = _mm256_castpd_si256( + __m256i p256_4 = _mm256_castpd_si256( _mm256_broadcast_pd((__m128d const *)(s - 5 * p))); - q256_4 = _mm256_castpd_si256( + __m256i q256_4 = _mm256_castpd_si256( _mm256_broadcast_pd((__m128d const *)(s + 4 * p))); - p4 = _mm256_castsi256_si128(p256_4); - q4 = _mm256_castsi256_si128(q256_4); - work = _mm_max_epu8( - _mm_or_si128(_mm_subs_epu8(p4, p0), _mm_subs_epu8(p0, p4)), - _mm_or_si128(_mm_subs_epu8(q4, q0), _mm_subs_epu8(q0, q4))); - - p256_5 = _mm256_castpd_si256( + __m256i p256_5 = _mm256_castpd_si256( _mm256_broadcast_pd((__m128d const *)(s - 6 * p))); - q256_5 = _mm256_castpd_si256( + __m256i q256_5 = _mm256_castpd_si256( _mm256_broadcast_pd((__m128d const *)(s + 5 * p))); - p5 = _mm256_castsi256_si128(p256_5); - q5 = _mm256_castsi256_si128(q256_5); - flat2 = _mm_max_epu8( - _mm_or_si128(_mm_subs_epu8(p5, p0), _mm_subs_epu8(p0, p5)), - _mm_or_si128(_mm_subs_epu8(q5, q0), _mm_subs_epu8(q0, q5))); - - flat2 = _mm_max_epu8(work, flat2); - - p256_6 = _mm256_castpd_si256( + __m256i p256_6 = _mm256_castpd_si256( _mm256_broadcast_pd((__m128d const *)(s - 7 * p))); - q256_6 = _mm256_castpd_si256( + __m256i q256_6 = _mm256_castpd_si256( _mm256_broadcast_pd((__m128d const *)(s + 6 * p))); - p6 = _mm256_castsi256_si128(p256_6); - q6 = _mm256_castsi256_si128(q256_6); - work = _mm_max_epu8( - _mm_or_si128(_mm_subs_epu8(p6, p0), _mm_subs_epu8(p0, p6)), - _mm_or_si128(_mm_subs_epu8(q6, q0), _mm_subs_epu8(q0, q6))); - - flat2 = _mm_max_epu8(work, flat2); + // Derive flat2 + __m256i p4q4256 = _mm256_blend_epi32(p256_4, q256_4, 0xf0); + __m256i p5q5256 = _mm256_blend_epi32(p256_5, q256_5, 0xf0); + const __m256i p6q6256 = _mm256_blend_epi32(p256_6, q256_6, 0xf0); + const __m256i work1 = _mm256_or_si256(_mm256_subs_epu8(p4q4256, p0q0256), + _mm256_subs_epu8(p0q0256, p4q4256)); + const __m256i work2 = _mm256_or_si256(_mm256_subs_epu8(p5q5256, p0q0256), + _mm256_subs_epu8(p0q0256, p5q5256)); + const __m256i work3 = _mm256_or_si256(_mm256_subs_epu8(p6q6256, p0q0256), + _mm256_subs_epu8(p0q0256, p6q6256)); + __m256i flat2_256 = _mm256_max_epu8(work1, work2); + flat2_256 = _mm256_max_epu8(flat2_256, work3); + __m128i flat2 = _mm_max_epu8(_mm256_castsi256_si128(flat2_256), + _mm256_extractf128_si256(flat2_256, 1)); flat2 = _mm_subs_epu8(flat2, one); flat2 = _mm_cmpeq_epi8(flat2, zero); flat2 = _mm_and_si128(flat2, flat); // flat2 & flat & mask @@ -684,72 +788,66 @@ void aom_lpf_horizontal_14_quad_avx2(unsigned char *s, int p, q256_2 = _mm256_shuffle_epi8(q256_2, filter); q256_3 = _mm256_shuffle_epi8(q256_3, filter); - p2p1p0 = _mm256_add_epi16(p256_0, _mm256_add_epi16(p256_2, p256_1)); - q2q1q0 = _mm256_add_epi16(q256_0, _mm256_add_epi16(q256_2, q256_1)); + const __m256i p2p1p0 = + _mm256_add_epi16(p256_0, _mm256_add_epi16(p256_2, p256_1)); + const __m256i q2q1q0 = + _mm256_add_epi16(q256_0, _mm256_add_epi16(q256_2, q256_1)); - pixetFilter_p2p1p0 = + __m256i pixetFilter_p2p1p0 = _mm256_add_epi16(four, _mm256_add_epi16(p2p1p0, q2q1q0)); - pixetFilter_q2q1q0 = pixetFilter_p2p1p0; + __m256i pixetFilter_q2q1q0 = pixetFilter_p2p1p0; + // Derive p0 and q0 pixetFilter_p2p1p0 = _mm256_add_epi16(pixetFilter_p2p1p0, p256_3); - res_p = + __m256i res_p = _mm256_srli_epi16(_mm256_add_epi16(pixetFilter_p2p1p0, p256_0), 3); - flat_p0 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_p), 168)); - p0 = _mm_andnot_si128(flat, ps0); - flat_p0 = _mm_and_si128(flat, flat_p0); - p0 = _mm_or_si128(flat_p0, p0); - pixetFilter_q2q1q0 = _mm256_add_epi16(pixetFilter_q2q1q0, q256_3); - res_q = + __m256i res_q = _mm256_srli_epi16(_mm256_add_epi16(pixetFilter_q2q1q0, q256_0), 3); - flat_q0 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_q, res_q), 168)); - q0 = _mm_andnot_si128(flat, qs0); - flat_q0 = _mm_and_si128(flat, flat_q0); - q0 = _mm_or_si128(flat_q0, q0); - - sum_p = _mm256_sub_epi16(p256_3, q256_2); + __m256i flat_p0q0 = + _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_q), 0xd8); + p0q0256 = _mm256_andnot_si256(flat256, ps0qs0256); + flat_p0q0 = _mm256_and_si256(flat256, flat_p0q0); + p0q0256 = _mm256_or_si256(flat_p0q0, p0q0256); + p0 = _mm256_castsi256_si128(p0q0256); + q0 = _mm256_extractf128_si256(p0q0256, 1); + + // Derive p1 and q1 + __m256i sum_p = _mm256_sub_epi16(p256_3, q256_2); pixetFilter_p2p1p0 = _mm256_add_epi16(pixetFilter_p2p1p0, sum_p); res_p = _mm256_srli_epi16(_mm256_add_epi16(pixetFilter_p2p1p0, p256_1), 3); - flat_p1 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_p), 168)); - p1 = _mm_andnot_si128(flat, ps1); - flat_p1 = _mm_and_si128(flat, flat_p1); - p1 = _mm_or_si128(flat_p1, p1); - - sum_q = _mm256_sub_epi16(q256_3, p256_2); + __m256i sum_q = _mm256_sub_epi16(q256_3, p256_2); pixetFilter_q2q1q0 = _mm256_add_epi16(pixetFilter_q2q1q0, sum_q); res_q = _mm256_srli_epi16(_mm256_add_epi16(pixetFilter_q2q1q0, q256_1), 3); - flat_q1 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_q, res_q), 168)); - q1 = _mm_andnot_si128(flat, qs1); - flat_q1 = _mm_and_si128(flat, flat_q1); - q1 = _mm_or_si128(flat_q1, q1); - + __m256i flat_p1q1 = + _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_q), 0xd8); + __m256i p1q1256 = _mm256_andnot_si256(flat256, ps1qs1256); + flat_p1q1 = _mm256_and_si256(flat256, flat_p1q1); + p1q1256 = _mm256_or_si256(flat_p1q1, p1q1256); + p1 = _mm256_castsi256_si128(p1q1256); + q1 = _mm256_extractf128_si256(p1q1256, 1); + + // Derive p2 and q2 sum_p = _mm256_sub_epi16(p256_3, q256_1); pixetFilter_p2p1p0 = _mm256_add_epi16(pixetFilter_p2p1p0, sum_p); res_p = _mm256_srli_epi16(_mm256_add_epi16(pixetFilter_p2p1p0, p256_2), 3); - flat_p2 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_p), 168)); - p2 = _mm_andnot_si128(flat, p2); - flat_p2 = _mm_and_si128(flat, flat_p2); - p2 = _mm_or_si128(flat_p2, p2); - sum_q = _mm256_sub_epi16(q256_3, p256_1); pixetFilter_q2q1q0 = _mm256_add_epi16(pixetFilter_q2q1q0, sum_q); res_q = _mm256_srli_epi16(_mm256_add_epi16(pixetFilter_q2q1q0, q256_2), 3); - flat_q2 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_q, res_q), 168)); - q2 = _mm_andnot_si128(flat, q2); - flat_q2 = _mm_and_si128(flat, flat_q2); - q2 = _mm_or_si128(flat_q2, q2); - + __m256i flat_p2q2 = + _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_q), 0xd8); + p2q2256 = _mm256_andnot_si256(flat256, p2q2256); + flat_p2q2 = _mm256_and_si256(flat256, flat_p2q2); + p2q2256 = _mm256_or_si256(flat_p2q2, p2q2256); + p2 = _mm256_castsi256_si128(p2q2256); + q2 = _mm256_extractf128_si256(p2q2256, 1); if (0xffff != _mm_movemask_epi8(_mm_cmpeq_epi8(flat2, zero))) { + flat2_256 = + _mm256_insertf128_si256(_mm256_castsi128_si256(flat2), flat2, 0x1); p256_6 = _mm256_shuffle_epi8(p256_6, filter); p256_5 = _mm256_shuffle_epi8(p256_5, filter); p256_4 = _mm256_shuffle_epi8(p256_4, filter); @@ -757,9 +855,9 @@ void aom_lpf_horizontal_14_quad_avx2(unsigned char *s, int p, q256_5 = _mm256_shuffle_epi8(q256_5, filter); q256_6 = _mm256_shuffle_epi8(q256_6, filter); - pixelFilter_p = + __m256i pixelFilter_p = _mm256_add_epi16(p256_5, _mm256_add_epi16(p256_4, p256_3)); - pixelFilter_q = + __m256i pixelFilter_q = _mm256_add_epi16(q256_5, _mm256_add_epi16(q256_4, q256_3)); pixelFilter_p = _mm256_add_epi16(pixelFilter_p, p2p1p0); @@ -771,135 +869,118 @@ void aom_lpf_horizontal_14_quad_avx2(unsigned char *s, int p, eight, _mm256_add_epi16(pixelFilter_p, pixelFilter_q)); pixelFilter_q = pixelFilter_p; + // Derive p0 and q0 pixelFilter_p = _mm256_add_epi16(_mm256_add_epi16(p256_6, p256_1), pixelFilter_p); res_p = _mm256_srli_epi16(pixelFilter_p, 4); - flat2_p0 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_p), 168)); - p0 = _mm_andnot_si128(flat2, p0); - flat2_p0 = _mm_and_si128(flat2, flat2_p0); - p0 = _mm_or_si128(flat2_p0, p0); - _mm_storeu_si128((__m128i *)(s - 1 * p), p0); - pixelFilter_q = _mm256_add_epi16(_mm256_add_epi16(q256_6, q256_1), pixelFilter_q); res_q = _mm256_srli_epi16(pixelFilter_q, 4); - flat2_q0 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_q, res_q), 168)); - q0 = _mm_andnot_si128(flat2, q0); - flat2_q0 = _mm_and_si128(flat2, flat2_q0); - q0 = _mm_or_si128(flat2_q0, q0); + __m256i flat2_p0q0 = + _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_q), 0xd8); + p0q0256 = _mm256_andnot_si256(flat2_256, p0q0256); + flat2_p0q0 = _mm256_and_si256(flat2_256, flat2_p0q0); + p0q0256 = _mm256_or_si256(flat2_p0q0, p0q0256); + + p0 = _mm256_castsi256_si128(p0q0256); + q0 = _mm256_extractf128_si256(p0q0256, 1); + _mm_storeu_si128((__m128i *)(s - 1 * p), p0); _mm_storeu_si128((__m128i *)(s - 0 * p), q0); + // Derive p1 and q1 sum_p = _mm256_add_epi16(_mm256_sub_epi16(p256_6, q256_5), _mm256_sub_epi16(p256_2, q256_0)); pixelFilter_p = _mm256_add_epi16(pixelFilter_p, sum_p); res_p = _mm256_srli_epi16(pixelFilter_p, 4); - flat2_p1 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_p), 168)); - p1 = _mm_andnot_si128(flat2, p1); - flat2_p1 = _mm_and_si128(flat2, flat2_p1); - p1 = _mm_or_si128(flat2_p1, p1); - _mm_storeu_si128((__m128i *)(s - 2 * p), p1); - sum_q = _mm256_add_epi16(_mm256_sub_epi16(q256_6, p256_5), _mm256_sub_epi16(q256_2, p256_0)); pixelFilter_q = _mm256_add_epi16(pixelFilter_q, sum_q); res_q = _mm256_srli_epi16(pixelFilter_q, 4); - flat2_q1 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_q, res_q), 168)); - q1 = _mm_andnot_si128(flat2, q1); - flat2_q1 = _mm_and_si128(flat2, flat2_q1); - q1 = _mm_or_si128(flat2_q1, q1); + __m256i flat2_p1q1 = + _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_q), 0xd8); + p1q1256 = _mm256_andnot_si256(flat2_256, p1q1256); + flat2_p1q1 = _mm256_and_si256(flat2_256, flat2_p1q1); + p1q1256 = _mm256_or_si256(flat2_p1q1, p1q1256); + p1 = _mm256_castsi256_si128(p1q1256); + q1 = _mm256_extractf128_si256(p1q1256, 1); + _mm_storeu_si128((__m128i *)(s - 2 * p), p1); _mm_storeu_si128((__m128i *)(s + 1 * p), q1); + // Derive p2 and q2 sum_p = _mm256_add_epi16(_mm256_sub_epi16(p256_6, q256_4), _mm256_sub_epi16(p256_3, p256_0)); pixelFilter_p = _mm256_add_epi16(pixelFilter_p, sum_p); res_p = _mm256_srli_epi16(pixelFilter_p, 4); - flat2_p2 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_p), 168)); - p2 = _mm_andnot_si128(flat2, p2); - flat2_p2 = _mm_and_si128(flat2, flat2_p2); - p2 = _mm_or_si128(flat2_p2, p2); - _mm_storeu_si128((__m128i *)(s - 3 * p), p2); - sum_q = _mm256_add_epi16(_mm256_sub_epi16(q256_6, p256_4), _mm256_sub_epi16(q256_3, q256_0)); pixelFilter_q = _mm256_add_epi16(pixelFilter_q, sum_q); res_q = _mm256_srli_epi16(pixelFilter_q, 4); - flat2_q2 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_q, res_q), 168)); - q2 = _mm_andnot_si128(flat2, q2); - flat2_q2 = _mm_and_si128(flat2, flat2_q2); - q2 = _mm_or_si128(flat2_q2, q2); + __m256i flat2_p2q2 = + _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_q), 0xd8); + p2q2256 = _mm256_andnot_si256(flat2_256, p2q2256); + flat2_p2q2 = _mm256_and_si256(flat2_256, flat2_p2q2); + p2q2256 = _mm256_or_si256(flat2_p2q2, p2q2256); + p2 = _mm256_castsi256_si128(p2q2256); + q2 = _mm256_extractf128_si256(p2q2256, 1); + _mm_storeu_si128((__m128i *)(s - 3 * p), p2); _mm_storeu_si128((__m128i *)(s + 2 * p), q2); + // Derive p3 and q3 sum_p = _mm256_add_epi16(_mm256_sub_epi16(p256_6, q256_3), _mm256_sub_epi16(p256_4, p256_1)); pixelFilter_p = _mm256_add_epi16(pixelFilter_p, sum_p); res_p = _mm256_srli_epi16(pixelFilter_p, 4); - flat2_p3 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_p), 168)); - p3 = _mm_andnot_si128(flat2, p3); - flat2_p3 = _mm_and_si128(flat2, flat2_p3); - p3 = _mm_or_si128(flat2_p3, p3); - _mm_storeu_si128((__m128i *)(s - 4 * p), p3); - sum_q = _mm256_add_epi16(_mm256_sub_epi16(q256_6, p256_3), _mm256_sub_epi16(q256_4, q256_1)); pixelFilter_q = _mm256_add_epi16(pixelFilter_q, sum_q); res_q = _mm256_srli_epi16(pixelFilter_q, 4); - flat2_q3 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_q, res_q), 168)); - q3 = _mm_andnot_si128(flat2, q3); - flat2_q3 = _mm_and_si128(flat2, flat2_q3); - q3 = _mm_or_si128(flat2_q3, q3); + __m256i flat2_p3q3 = + _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_q), 0xd8); + p3q3256 = _mm256_andnot_si256(flat2_256, p3q3256); + flat2_p3q3 = _mm256_and_si256(flat2_256, flat2_p3q3); + p3q3256 = _mm256_or_si256(flat2_p3q3, p3q3256); + p3 = _mm256_castsi256_si128(p3q3256); + q3 = _mm256_extractf128_si256(p3q3256, 1); + _mm_storeu_si128((__m128i *)(s - 4 * p), p3); _mm_storeu_si128((__m128i *)(s + 3 * p), q3); + // Derive p4 and q4 sum_p = _mm256_add_epi16(_mm256_sub_epi16(p256_6, q256_2), _mm256_sub_epi16(p256_5, p256_2)); pixelFilter_p = _mm256_add_epi16(pixelFilter_p, sum_p); res_p = _mm256_srli_epi16(pixelFilter_p, 4); - flat2_p4 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_p), 168)); - p4 = _mm_andnot_si128(flat2, p4); - flat2_p4 = _mm_and_si128(flat2, flat2_p4); - p4 = _mm_or_si128(flat2_p4, p4); - _mm_storeu_si128((__m128i *)(s - 5 * p), p4); - sum_q = _mm256_add_epi16(_mm256_sub_epi16(q256_6, p256_2), _mm256_sub_epi16(q256_5, q256_2)); pixelFilter_q = _mm256_add_epi16(pixelFilter_q, sum_q); res_q = _mm256_srli_epi16(pixelFilter_q, 4); - flat2_q4 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_q, res_q), 168)); - q4 = _mm_andnot_si128(flat2, q4); - flat2_q4 = _mm_and_si128(flat2, flat2_q4); - q4 = _mm_or_si128(flat2_q4, q4); - _mm_storeu_si128((__m128i *)(s + 4 * p), q4); - + __m256i flat2_p4q4 = + _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_q), 0xd8); + p4q4256 = _mm256_andnot_si256(flat2_256, p4q4256); + flat2_p4q4 = _mm256_and_si256(flat2_256, flat2_p4q4); + p4q4256 = _mm256_or_si256(flat2_p4q4, p4q4256); + _mm_storeu_si128((__m128i *)(s - 5 * p), + _mm256_castsi256_si128(p4q4256)); + _mm_storeu_si128((__m128i *)(s + 4 * p), + _mm256_extractf128_si256(p4q4256, 1)); + + // Derive p5 and q5 sum_p = _mm256_add_epi16(_mm256_sub_epi16(p256_6, q256_1), _mm256_sub_epi16(p256_6, p256_3)); pixelFilter_p = _mm256_add_epi16(pixelFilter_p, sum_p); res_p = _mm256_srli_epi16(pixelFilter_p, 4); - flat2_p5 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_p), 168)); - p5 = _mm_andnot_si128(flat2, p5); - flat2_p5 = _mm_and_si128(flat2, flat2_p5); - p5 = _mm_or_si128(flat2_p5, p5); - _mm_storeu_si128((__m128i *)(s - 6 * p), p5); - sum_q = _mm256_add_epi16(_mm256_sub_epi16(q256_6, p256_1), _mm256_sub_epi16(q256_6, q256_3)); pixelFilter_q = _mm256_add_epi16(pixelFilter_q, sum_q); res_q = _mm256_srli_epi16(pixelFilter_q, 4); - flat2_q5 = _mm256_castsi256_si128( - _mm256_permute4x64_epi64(_mm256_packus_epi16(res_q, res_q), 168)); - q5 = _mm_andnot_si128(flat2, q5); - flat2_q5 = _mm_and_si128(flat2, flat2_q5); - q5 = _mm_or_si128(flat2_q5, q5); - _mm_storeu_si128((__m128i *)(s + 5 * p), q5); + __m256i flat2_p5q5 = + _mm256_permute4x64_epi64(_mm256_packus_epi16(res_p, res_q), 0xd8); + p5q5256 = _mm256_andnot_si256(flat2_256, p5q5256); + flat2_p5q5 = _mm256_and_si256(flat2_256, flat2_p5q5); + p5q5256 = _mm256_or_si256(flat2_p5q5, p5q5256); + _mm_storeu_si128((__m128i *)(s - 6 * p), + _mm256_castsi256_si128(p5q5256)); + _mm_storeu_si128((__m128i *)(s + 5 * p), + _mm256_extractf128_si256(p5q5256, 1)); } else { _mm_storeu_si128((__m128i *)(s - 3 * p), p2); _mm_storeu_si128((__m128i *)(s - 2 * p), p1); @@ -916,3 +997,20 @@ void aom_lpf_horizontal_14_quad_avx2(unsigned char *s, int p, } } } + +void aom_lpf_vertical_14_quad_avx2(unsigned char *s, int pitch, + const uint8_t *_blimit0, + const uint8_t *_limit0, + const uint8_t *_thresh0) { + DECLARE_ALIGNED(16, unsigned char, t_dst[256]); + + // Transpose 16x16 + trans_store_16x16_lpf_vert14(s - 8, pitch, t_dst, 16, 1); + + // Loop filtering + aom_lpf_horizontal_14_quad_avx2(t_dst + 8 * 16, 16, _blimit0, _limit0, + _thresh0); + + // Transpose back + trans_store_16x16_lpf_vert14(t_dst, 16, s - 8, pitch, 0); +} |