diff options
author | johannkoenig@chromium.org <johannkoenig@chromium.org@4ff67af0-8c30-449e-8e8b-ad334ec8d88c> | 2014-04-10 17:14:25 +0000 |
---|---|---|
committer | johannkoenig@chromium.org <johannkoenig@chromium.org@4ff67af0-8c30-449e-8e8b-ad334ec8d88c> | 2014-04-10 17:14:25 +0000 |
commit | 93a74791c8e808ea76001ee07693aa2a5fdd3500 (patch) | |
tree | 88c3a21369388876dccedda352d254b16007ba22 /source/config/win/ia32/vp8_rtcd.h | |
parent | ef98d99073c8ddc400dac9bd4a1b31fb7240d861 (diff) | |
download | libvpx-93a74791c8e808ea76001ee07693aa2a5fdd3500.tar.gz |
libvpx: Pull from upstream
Update grep pattern in update_libvpx.sh to account for
alternative locales.
Update generate_gypi.sh to use the new perl rtcd scripts.
Current HEAD: 23ccf7192434399e5db3a981fbfde40e1712ed5f
git log from upstream:
23ccf71 Merge "Fix encoder uninitialized read errors reported by drmemory"
4b8ad4a Merge "Fix coding format in vp9_rc_regulate_q"
675d95f Merge "Prevent the usage of invalid best_mode_index"
9034094 Merge "Remove duplicate code"
81056e2 Merge "Minor code cleanup"
65e650e Merge "Revert "Converting set_prev_mi() to get_prev_mi().""
3dff8aa Merge "Moving q_trans[] table to vp9_quantize.{c, h}."
f10c173 Merge "Removing unused code from vp9_onyx_if.c."
1dcc1aa Prevent the usage of invalid best_mode_index
41ea9ef Merge "Removing redundant assignments."
71ffc7d Merge "Remove unused tile arguments from vp_rdopt."
1eee13c Merge "Cleanup vp9_rd_pick_inter_mode_sub8x8()."
2255085 Fix coding format in vp9_rc_regulate_q
60def47 Revert "Converting set_prev_mi() to get_prev_mi()."
2dc9248 Merge "Fix the setting of mode_skip_mask"
b60d23f Removing unused code from vp9_onyx_if.c.
d1a396d Moving q_trans[] table to vp9_quantize.{c, h}.
4fffefe Merge "Fix avx builds on macosx with clang 5.0."
585e01b Remove duplicate code
7cc78c0 Merge "Adding vp9_inc_frame_in_layer() function."
7c891ed Minor code cleanup
3ab4d57 Remove unused tile arguments from vp_rdopt.
<...>
TBR=tomfinegan@chromium.org
Review URL: https://codereview.chromium.org/232133009
git-svn-id: http://src.chromium.org/svn/trunk/deps/third_party/libvpx@263021 4ff67af0-8c30-449e-8e8b-ad334ec8d88c
Diffstat (limited to 'source/config/win/ia32/vp8_rtcd.h')
-rw-r--r-- | source/config/win/ia32/vp8_rtcd.h | 98 |
1 files changed, 0 insertions, 98 deletions
diff --git a/source/config/win/ia32/vp8_rtcd.h b/source/config/win/ia32/vp8_rtcd.h index 7a3e0f4..7e90462 100644 --- a/source/config/win/ia32/vp8_rtcd.h +++ b/source/config/win/ia32/vp8_rtcd.h @@ -492,337 +492,239 @@ static void setup_rtcd_internal(void) if (flags & HAS_MMX) vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16_mmx; if (flags & HAS_SSE2) vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16_sse2; if (flags & HAS_SSSE3) vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16_ssse3; - vp8_bilinear_predict4x4 = vp8_bilinear_predict4x4_c; if (flags & HAS_MMX) vp8_bilinear_predict4x4 = vp8_bilinear_predict4x4_mmx; - vp8_bilinear_predict8x4 = vp8_bilinear_predict8x4_c; if (flags & HAS_MMX) vp8_bilinear_predict8x4 = vp8_bilinear_predict8x4_mmx; - vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_c; if (flags & HAS_MMX) vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_mmx; if (flags & HAS_SSE2) vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_sse2; if (flags & HAS_SSSE3) vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_ssse3; - - - - vp8_block_error = vp8_block_error_c; if (flags & HAS_MMX) vp8_block_error = vp8_block_error_mmx; if (flags & HAS_SSE2) vp8_block_error = vp8_block_error_xmm; - vp8_build_intra_predictors_mbuv_s = vp8_build_intra_predictors_mbuv_s_c; if (flags & HAS_SSE2) vp8_build_intra_predictors_mbuv_s = vp8_build_intra_predictors_mbuv_s_sse2; if (flags & HAS_SSSE3) vp8_build_intra_predictors_mbuv_s = vp8_build_intra_predictors_mbuv_s_ssse3; - vp8_build_intra_predictors_mby_s = vp8_build_intra_predictors_mby_s_c; if (flags & HAS_SSE2) vp8_build_intra_predictors_mby_s = vp8_build_intra_predictors_mby_s_sse2; if (flags & HAS_SSSE3) vp8_build_intra_predictors_mby_s = vp8_build_intra_predictors_mby_s_ssse3; - vp8_clear_system_state = vp8_clear_system_state_c; if (flags & HAS_MMX) vp8_clear_system_state = vpx_reset_mmx_state; - vp8_copy32xn = vp8_copy32xn_c; if (flags & HAS_SSE2) vp8_copy32xn = vp8_copy32xn_sse2; if (flags & HAS_SSE3) vp8_copy32xn = vp8_copy32xn_sse3; - vp8_copy_mem16x16 = vp8_copy_mem16x16_c; if (flags & HAS_MMX) vp8_copy_mem16x16 = vp8_copy_mem16x16_mmx; if (flags & HAS_SSE2) vp8_copy_mem16x16 = vp8_copy_mem16x16_sse2; - vp8_copy_mem8x4 = vp8_copy_mem8x4_c; if (flags & HAS_MMX) vp8_copy_mem8x4 = vp8_copy_mem8x4_mmx; - vp8_copy_mem8x8 = vp8_copy_mem8x8_c; if (flags & HAS_MMX) vp8_copy_mem8x8 = vp8_copy_mem8x8_mmx; - vp8_dc_only_idct_add = vp8_dc_only_idct_add_c; if (flags & HAS_MMX) vp8_dc_only_idct_add = vp8_dc_only_idct_add_mmx; - vp8_denoiser_filter = vp8_denoiser_filter_c; if (flags & HAS_SSE2) vp8_denoiser_filter = vp8_denoiser_filter_sse2; - vp8_dequant_idct_add = vp8_dequant_idct_add_c; if (flags & HAS_MMX) vp8_dequant_idct_add = vp8_dequant_idct_add_mmx; - vp8_dequant_idct_add_uv_block = vp8_dequant_idct_add_uv_block_c; if (flags & HAS_MMX) vp8_dequant_idct_add_uv_block = vp8_dequant_idct_add_uv_block_mmx; if (flags & HAS_SSE2) vp8_dequant_idct_add_uv_block = vp8_dequant_idct_add_uv_block_sse2; - vp8_dequant_idct_add_y_block = vp8_dequant_idct_add_y_block_c; if (flags & HAS_MMX) vp8_dequant_idct_add_y_block = vp8_dequant_idct_add_y_block_mmx; if (flags & HAS_SSE2) vp8_dequant_idct_add_y_block = vp8_dequant_idct_add_y_block_sse2; - vp8_dequantize_b = vp8_dequantize_b_c; if (flags & HAS_MMX) vp8_dequantize_b = vp8_dequantize_b_mmx; - vp8_diamond_search_sad = vp8_diamond_search_sad_c; if (flags & HAS_SSE3) vp8_diamond_search_sad = vp8_diamond_search_sadx4; - vp8_fast_quantize_b = vp8_fast_quantize_b_c; if (flags & HAS_SSE2) vp8_fast_quantize_b = vp8_fast_quantize_b_sse2; if (flags & HAS_SSSE3) vp8_fast_quantize_b = vp8_fast_quantize_b_ssse3; - - vp8_filter_by_weight16x16 = vp8_filter_by_weight16x16_c; if (flags & HAS_SSE2) vp8_filter_by_weight16x16 = vp8_filter_by_weight16x16_sse2; - - vp8_filter_by_weight8x8 = vp8_filter_by_weight8x8_c; if (flags & HAS_SSE2) vp8_filter_by_weight8x8 = vp8_filter_by_weight8x8_sse2; - vp8_full_search_sad = vp8_full_search_sad_c; if (flags & HAS_SSE3) vp8_full_search_sad = vp8_full_search_sadx3; if (flags & HAS_SSE4_1) vp8_full_search_sad = vp8_full_search_sadx8; - vp8_get4x4sse_cs = vp8_get4x4sse_cs_c; if (flags & HAS_MMX) vp8_get4x4sse_cs = vp8_get4x4sse_cs_mmx; - vp8_get_mb_ss = vp8_get_mb_ss_c; if (flags & HAS_MMX) vp8_get_mb_ss = vp8_get_mb_ss_mmx; if (flags & HAS_SSE2) vp8_get_mb_ss = vp8_get_mb_ss_sse2; - - vp8_loop_filter_bh = vp8_loop_filter_bh_c; if (flags & HAS_MMX) vp8_loop_filter_bh = vp8_loop_filter_bh_mmx; if (flags & HAS_SSE2) vp8_loop_filter_bh = vp8_loop_filter_bh_sse2; - vp8_loop_filter_bv = vp8_loop_filter_bv_c; if (flags & HAS_MMX) vp8_loop_filter_bv = vp8_loop_filter_bv_mmx; if (flags & HAS_SSE2) vp8_loop_filter_bv = vp8_loop_filter_bv_sse2; - vp8_loop_filter_mbh = vp8_loop_filter_mbh_c; if (flags & HAS_MMX) vp8_loop_filter_mbh = vp8_loop_filter_mbh_mmx; if (flags & HAS_SSE2) vp8_loop_filter_mbh = vp8_loop_filter_mbh_sse2; - vp8_loop_filter_mbv = vp8_loop_filter_mbv_c; if (flags & HAS_MMX) vp8_loop_filter_mbv = vp8_loop_filter_mbv_mmx; if (flags & HAS_SSE2) vp8_loop_filter_mbv = vp8_loop_filter_mbv_sse2; - vp8_loop_filter_simple_bh = vp8_loop_filter_bhs_c; if (flags & HAS_MMX) vp8_loop_filter_simple_bh = vp8_loop_filter_bhs_mmx; if (flags & HAS_SSE2) vp8_loop_filter_simple_bh = vp8_loop_filter_bhs_sse2; - vp8_loop_filter_simple_bv = vp8_loop_filter_bvs_c; if (flags & HAS_MMX) vp8_loop_filter_simple_bv = vp8_loop_filter_bvs_mmx; if (flags & HAS_SSE2) vp8_loop_filter_simple_bv = vp8_loop_filter_bvs_sse2; - vp8_loop_filter_simple_mbh = vp8_loop_filter_simple_horizontal_edge_c; if (flags & HAS_MMX) vp8_loop_filter_simple_mbh = vp8_loop_filter_simple_horizontal_edge_mmx; if (flags & HAS_SSE2) vp8_loop_filter_simple_mbh = vp8_loop_filter_simple_horizontal_edge_sse2; - vp8_loop_filter_simple_mbv = vp8_loop_filter_simple_vertical_edge_c; if (flags & HAS_MMX) vp8_loop_filter_simple_mbv = vp8_loop_filter_simple_vertical_edge_mmx; if (flags & HAS_SSE2) vp8_loop_filter_simple_mbv = vp8_loop_filter_simple_vertical_edge_sse2; - vp8_mbblock_error = vp8_mbblock_error_c; if (flags & HAS_MMX) vp8_mbblock_error = vp8_mbblock_error_mmx; if (flags & HAS_SSE2) vp8_mbblock_error = vp8_mbblock_error_xmm; - vp8_mbpost_proc_across_ip = vp8_mbpost_proc_across_ip_c; if (flags & HAS_SSE2) vp8_mbpost_proc_across_ip = vp8_mbpost_proc_across_ip_xmm; - vp8_mbpost_proc_down = vp8_mbpost_proc_down_c; if (flags & HAS_MMX) vp8_mbpost_proc_down = vp8_mbpost_proc_down_mmx; if (flags & HAS_SSE2) vp8_mbpost_proc_down = vp8_mbpost_proc_down_xmm; - vp8_mbuverror = vp8_mbuverror_c; if (flags & HAS_MMX) vp8_mbuverror = vp8_mbuverror_mmx; if (flags & HAS_SSE2) vp8_mbuverror = vp8_mbuverror_xmm; - vp8_mse16x16 = vp8_mse16x16_c; if (flags & HAS_MMX) vp8_mse16x16 = vp8_mse16x16_mmx; if (flags & HAS_SSE2) vp8_mse16x16 = vp8_mse16x16_wmt; - vp8_plane_add_noise = vp8_plane_add_noise_c; if (flags & HAS_MMX) vp8_plane_add_noise = vp8_plane_add_noise_mmx; if (flags & HAS_SSE2) vp8_plane_add_noise = vp8_plane_add_noise_wmt; - vp8_post_proc_down_and_across_mb_row = vp8_post_proc_down_and_across_mb_row_c; if (flags & HAS_SSE2) vp8_post_proc_down_and_across_mb_row = vp8_post_proc_down_and_across_mb_row_sse2; - - - - vp8_refining_search_sad = vp8_refining_search_sad_c; if (flags & HAS_SSE3) vp8_refining_search_sad = vp8_refining_search_sadx4; - vp8_regular_quantize_b = vp8_regular_quantize_b_c; if (flags & HAS_SSE2) vp8_regular_quantize_b = vp8_regular_quantize_b_sse2; - - vp8_sad16x16 = vp8_sad16x16_c; if (flags & HAS_MMX) vp8_sad16x16 = vp8_sad16x16_mmx; if (flags & HAS_SSE2) vp8_sad16x16 = vp8_sad16x16_wmt; if (flags & HAS_SSE3) vp8_sad16x16 = vp8_sad16x16_sse3; - vp8_sad16x16x3 = vp8_sad16x16x3_c; if (flags & HAS_SSE3) vp8_sad16x16x3 = vp8_sad16x16x3_sse3; if (flags & HAS_SSSE3) vp8_sad16x16x3 = vp8_sad16x16x3_ssse3; - vp8_sad16x16x4d = vp8_sad16x16x4d_c; if (flags & HAS_SSE3) vp8_sad16x16x4d = vp8_sad16x16x4d_sse3; - vp8_sad16x16x8 = vp8_sad16x16x8_c; if (flags & HAS_SSE4_1) vp8_sad16x16x8 = vp8_sad16x16x8_sse4; - vp8_sad16x8 = vp8_sad16x8_c; if (flags & HAS_MMX) vp8_sad16x8 = vp8_sad16x8_mmx; if (flags & HAS_SSE2) vp8_sad16x8 = vp8_sad16x8_wmt; - vp8_sad16x8x3 = vp8_sad16x8x3_c; if (flags & HAS_SSE3) vp8_sad16x8x3 = vp8_sad16x8x3_sse3; if (flags & HAS_SSSE3) vp8_sad16x8x3 = vp8_sad16x8x3_ssse3; - vp8_sad16x8x4d = vp8_sad16x8x4d_c; if (flags & HAS_SSE3) vp8_sad16x8x4d = vp8_sad16x8x4d_sse3; - vp8_sad16x8x8 = vp8_sad16x8x8_c; if (flags & HAS_SSE4_1) vp8_sad16x8x8 = vp8_sad16x8x8_sse4; - vp8_sad4x4 = vp8_sad4x4_c; if (flags & HAS_MMX) vp8_sad4x4 = vp8_sad4x4_mmx; if (flags & HAS_SSE2) vp8_sad4x4 = vp8_sad4x4_wmt; - vp8_sad4x4x3 = vp8_sad4x4x3_c; if (flags & HAS_SSE3) vp8_sad4x4x3 = vp8_sad4x4x3_sse3; - vp8_sad4x4x4d = vp8_sad4x4x4d_c; if (flags & HAS_SSE3) vp8_sad4x4x4d = vp8_sad4x4x4d_sse3; - vp8_sad4x4x8 = vp8_sad4x4x8_c; if (flags & HAS_SSE4_1) vp8_sad4x4x8 = vp8_sad4x4x8_sse4; - vp8_sad8x16 = vp8_sad8x16_c; if (flags & HAS_MMX) vp8_sad8x16 = vp8_sad8x16_mmx; if (flags & HAS_SSE2) vp8_sad8x16 = vp8_sad8x16_wmt; - vp8_sad8x16x3 = vp8_sad8x16x3_c; if (flags & HAS_SSE3) vp8_sad8x16x3 = vp8_sad8x16x3_sse3; - vp8_sad8x16x4d = vp8_sad8x16x4d_c; if (flags & HAS_SSE3) vp8_sad8x16x4d = vp8_sad8x16x4d_sse3; - vp8_sad8x16x8 = vp8_sad8x16x8_c; if (flags & HAS_SSE4_1) vp8_sad8x16x8 = vp8_sad8x16x8_sse4; - vp8_sad8x8 = vp8_sad8x8_c; if (flags & HAS_MMX) vp8_sad8x8 = vp8_sad8x8_mmx; if (flags & HAS_SSE2) vp8_sad8x8 = vp8_sad8x8_wmt; - vp8_sad8x8x3 = vp8_sad8x8x3_c; if (flags & HAS_SSE3) vp8_sad8x8x3 = vp8_sad8x8x3_sse3; - vp8_sad8x8x4d = vp8_sad8x8x4d_c; if (flags & HAS_SSE3) vp8_sad8x8x4d = vp8_sad8x8x4d_sse3; - vp8_sad8x8x8 = vp8_sad8x8x8_c; if (flags & HAS_SSE4_1) vp8_sad8x8x8 = vp8_sad8x8x8_sse4; - vp8_short_fdct4x4 = vp8_short_fdct4x4_c; if (flags & HAS_MMX) vp8_short_fdct4x4 = vp8_short_fdct4x4_mmx; if (flags & HAS_SSE2) vp8_short_fdct4x4 = vp8_short_fdct4x4_sse2; - vp8_short_fdct8x4 = vp8_short_fdct8x4_c; if (flags & HAS_MMX) vp8_short_fdct8x4 = vp8_short_fdct8x4_mmx; if (flags & HAS_SSE2) vp8_short_fdct8x4 = vp8_short_fdct8x4_sse2; - vp8_short_idct4x4llm = vp8_short_idct4x4llm_c; if (flags & HAS_MMX) vp8_short_idct4x4llm = vp8_short_idct4x4llm_mmx; - vp8_short_inv_walsh4x4 = vp8_short_inv_walsh4x4_c; if (flags & HAS_MMX) vp8_short_inv_walsh4x4 = vp8_short_inv_walsh4x4_mmx; if (flags & HAS_SSE2) vp8_short_inv_walsh4x4 = vp8_short_inv_walsh4x4_sse2; - - vp8_short_walsh4x4 = vp8_short_walsh4x4_c; if (flags & HAS_SSE2) vp8_short_walsh4x4 = vp8_short_walsh4x4_sse2; - vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_c; if (flags & HAS_MMX) vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_mmx; if (flags & HAS_SSE2) vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_sse2; if (flags & HAS_SSSE3) vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_ssse3; - vp8_sixtap_predict4x4 = vp8_sixtap_predict4x4_c; if (flags & HAS_MMX) vp8_sixtap_predict4x4 = vp8_sixtap_predict4x4_mmx; if (flags & HAS_SSSE3) vp8_sixtap_predict4x4 = vp8_sixtap_predict4x4_ssse3; - vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_c; if (flags & HAS_MMX) vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_mmx; if (flags & HAS_SSE2) vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_sse2; if (flags & HAS_SSSE3) vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_ssse3; - vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_c; if (flags & HAS_MMX) vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_mmx; if (flags & HAS_SSE2) vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_sse2; if (flags & HAS_SSSE3) vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_ssse3; - vp8_sub_pixel_mse16x16 = vp8_sub_pixel_mse16x16_c; if (flags & HAS_MMX) vp8_sub_pixel_mse16x16 = vp8_sub_pixel_mse16x16_mmx; if (flags & HAS_SSE2) vp8_sub_pixel_mse16x16 = vp8_sub_pixel_mse16x16_wmt; - vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_c; if (flags & HAS_MMX) vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_mmx; if (flags & HAS_SSE2) vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_wmt; if (flags & HAS_SSSE3) vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_ssse3; - vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_c; if (flags & HAS_MMX) vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_mmx; if (flags & HAS_SSE2) vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_wmt; if (flags & HAS_SSSE3) vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_ssse3; - vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_c; if (flags & HAS_MMX) vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_mmx; if (flags & HAS_SSE2) vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_wmt; - vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x16_c; if (flags & HAS_MMX) vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x16_mmx; if (flags & HAS_SSE2) vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x16_wmt; - vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_c; if (flags & HAS_MMX) vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_mmx; if (flags & HAS_SSE2) vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_wmt; - vp8_subtract_b = vp8_subtract_b_c; if (flags & HAS_MMX) vp8_subtract_b = vp8_subtract_b_mmx; if (flags & HAS_SSE2) vp8_subtract_b = vp8_subtract_b_sse2; - vp8_subtract_mbuv = vp8_subtract_mbuv_c; if (flags & HAS_MMX) vp8_subtract_mbuv = vp8_subtract_mbuv_mmx; if (flags & HAS_SSE2) vp8_subtract_mbuv = vp8_subtract_mbuv_sse2; - vp8_subtract_mby = vp8_subtract_mby_c; if (flags & HAS_MMX) vp8_subtract_mby = vp8_subtract_mby_mmx; if (flags & HAS_SSE2) vp8_subtract_mby = vp8_subtract_mby_sse2; - vp8_variance16x16 = vp8_variance16x16_c; if (flags & HAS_MMX) vp8_variance16x16 = vp8_variance16x16_mmx; if (flags & HAS_SSE2) vp8_variance16x16 = vp8_variance16x16_wmt; - vp8_variance16x8 = vp8_variance16x8_c; if (flags & HAS_MMX) vp8_variance16x8 = vp8_variance16x8_mmx; if (flags & HAS_SSE2) vp8_variance16x8 = vp8_variance16x8_wmt; - vp8_variance4x4 = vp8_variance4x4_c; if (flags & HAS_MMX) vp8_variance4x4 = vp8_variance4x4_mmx; if (flags & HAS_SSE2) vp8_variance4x4 = vp8_variance4x4_wmt; - vp8_variance8x16 = vp8_variance8x16_c; if (flags & HAS_MMX) vp8_variance8x16 = vp8_variance8x16_mmx; if (flags & HAS_SSE2) vp8_variance8x16 = vp8_variance8x16_wmt; - vp8_variance8x8 = vp8_variance8x8_c; if (flags & HAS_MMX) vp8_variance8x8 = vp8_variance8x8_mmx; if (flags & HAS_SSE2) vp8_variance8x8 = vp8_variance8x8_wmt; - vp8_variance_halfpixvar16x16_h = vp8_variance_halfpixvar16x16_h_c; if (flags & HAS_MMX) vp8_variance_halfpixvar16x16_h = vp8_variance_halfpixvar16x16_h_mmx; if (flags & HAS_SSE2) vp8_variance_halfpixvar16x16_h = vp8_variance_halfpixvar16x16_h_wmt; - vp8_variance_halfpixvar16x16_hv = vp8_variance_halfpixvar16x16_hv_c; if (flags & HAS_MMX) vp8_variance_halfpixvar16x16_hv = vp8_variance_halfpixvar16x16_hv_mmx; if (flags & HAS_SSE2) vp8_variance_halfpixvar16x16_hv = vp8_variance_halfpixvar16x16_hv_wmt; - vp8_variance_halfpixvar16x16_v = vp8_variance_halfpixvar16x16_v_c; if (flags & HAS_MMX) vp8_variance_halfpixvar16x16_v = vp8_variance_halfpixvar16x16_v_mmx; if (flags & HAS_SSE2) vp8_variance_halfpixvar16x16_v = vp8_variance_halfpixvar16x16_v_wmt; |