summaryrefslogtreecommitdiff
path: root/source/config/win/ia32/vp8_rtcd.h
diff options
context:
space:
mode:
authorjohannkoenig@chromium.org <johannkoenig@chromium.org@4ff67af0-8c30-449e-8e8b-ad334ec8d88c>2014-04-10 17:14:25 +0000
committerjohannkoenig@chromium.org <johannkoenig@chromium.org@4ff67af0-8c30-449e-8e8b-ad334ec8d88c>2014-04-10 17:14:25 +0000
commit93a74791c8e808ea76001ee07693aa2a5fdd3500 (patch)
tree88c3a21369388876dccedda352d254b16007ba22 /source/config/win/ia32/vp8_rtcd.h
parentef98d99073c8ddc400dac9bd4a1b31fb7240d861 (diff)
downloadlibvpx-93a74791c8e808ea76001ee07693aa2a5fdd3500.tar.gz
libvpx: Pull from upstream
Update grep pattern in update_libvpx.sh to account for alternative locales. Update generate_gypi.sh to use the new perl rtcd scripts. Current HEAD: 23ccf7192434399e5db3a981fbfde40e1712ed5f git log from upstream: 23ccf71 Merge "Fix encoder uninitialized read errors reported by drmemory" 4b8ad4a Merge "Fix coding format in vp9_rc_regulate_q" 675d95f Merge "Prevent the usage of invalid best_mode_index" 9034094 Merge "Remove duplicate code" 81056e2 Merge "Minor code cleanup" 65e650e Merge "Revert "Converting set_prev_mi() to get_prev_mi()."" 3dff8aa Merge "Moving q_trans[] table to vp9_quantize.{c, h}." f10c173 Merge "Removing unused code from vp9_onyx_if.c." 1dcc1aa Prevent the usage of invalid best_mode_index 41ea9ef Merge "Removing redundant assignments." 71ffc7d Merge "Remove unused tile arguments from vp_rdopt." 1eee13c Merge "Cleanup vp9_rd_pick_inter_mode_sub8x8()." 2255085 Fix coding format in vp9_rc_regulate_q 60def47 Revert "Converting set_prev_mi() to get_prev_mi()." 2dc9248 Merge "Fix the setting of mode_skip_mask" b60d23f Removing unused code from vp9_onyx_if.c. d1a396d Moving q_trans[] table to vp9_quantize.{c, h}. 4fffefe Merge "Fix avx builds on macosx with clang 5.0." 585e01b Remove duplicate code 7cc78c0 Merge "Adding vp9_inc_frame_in_layer() function." 7c891ed Minor code cleanup 3ab4d57 Remove unused tile arguments from vp_rdopt. <...> TBR=tomfinegan@chromium.org Review URL: https://codereview.chromium.org/232133009 git-svn-id: http://src.chromium.org/svn/trunk/deps/third_party/libvpx@263021 4ff67af0-8c30-449e-8e8b-ad334ec8d88c
Diffstat (limited to 'source/config/win/ia32/vp8_rtcd.h')
-rw-r--r--source/config/win/ia32/vp8_rtcd.h98
1 files changed, 0 insertions, 98 deletions
diff --git a/source/config/win/ia32/vp8_rtcd.h b/source/config/win/ia32/vp8_rtcd.h
index 7a3e0f4..7e90462 100644
--- a/source/config/win/ia32/vp8_rtcd.h
+++ b/source/config/win/ia32/vp8_rtcd.h
@@ -492,337 +492,239 @@ static void setup_rtcd_internal(void)
if (flags & HAS_MMX) vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16_mmx;
if (flags & HAS_SSE2) vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16_sse2;
if (flags & HAS_SSSE3) vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16_ssse3;
-
vp8_bilinear_predict4x4 = vp8_bilinear_predict4x4_c;
if (flags & HAS_MMX) vp8_bilinear_predict4x4 = vp8_bilinear_predict4x4_mmx;
-
vp8_bilinear_predict8x4 = vp8_bilinear_predict8x4_c;
if (flags & HAS_MMX) vp8_bilinear_predict8x4 = vp8_bilinear_predict8x4_mmx;
-
vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_c;
if (flags & HAS_MMX) vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_mmx;
if (flags & HAS_SSE2) vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_sse2;
if (flags & HAS_SSSE3) vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_ssse3;
-
-
-
-
vp8_block_error = vp8_block_error_c;
if (flags & HAS_MMX) vp8_block_error = vp8_block_error_mmx;
if (flags & HAS_SSE2) vp8_block_error = vp8_block_error_xmm;
-
vp8_build_intra_predictors_mbuv_s = vp8_build_intra_predictors_mbuv_s_c;
if (flags & HAS_SSE2) vp8_build_intra_predictors_mbuv_s = vp8_build_intra_predictors_mbuv_s_sse2;
if (flags & HAS_SSSE3) vp8_build_intra_predictors_mbuv_s = vp8_build_intra_predictors_mbuv_s_ssse3;
-
vp8_build_intra_predictors_mby_s = vp8_build_intra_predictors_mby_s_c;
if (flags & HAS_SSE2) vp8_build_intra_predictors_mby_s = vp8_build_intra_predictors_mby_s_sse2;
if (flags & HAS_SSSE3) vp8_build_intra_predictors_mby_s = vp8_build_intra_predictors_mby_s_ssse3;
-
vp8_clear_system_state = vp8_clear_system_state_c;
if (flags & HAS_MMX) vp8_clear_system_state = vpx_reset_mmx_state;
-
vp8_copy32xn = vp8_copy32xn_c;
if (flags & HAS_SSE2) vp8_copy32xn = vp8_copy32xn_sse2;
if (flags & HAS_SSE3) vp8_copy32xn = vp8_copy32xn_sse3;
-
vp8_copy_mem16x16 = vp8_copy_mem16x16_c;
if (flags & HAS_MMX) vp8_copy_mem16x16 = vp8_copy_mem16x16_mmx;
if (flags & HAS_SSE2) vp8_copy_mem16x16 = vp8_copy_mem16x16_sse2;
-
vp8_copy_mem8x4 = vp8_copy_mem8x4_c;
if (flags & HAS_MMX) vp8_copy_mem8x4 = vp8_copy_mem8x4_mmx;
-
vp8_copy_mem8x8 = vp8_copy_mem8x8_c;
if (flags & HAS_MMX) vp8_copy_mem8x8 = vp8_copy_mem8x8_mmx;
-
vp8_dc_only_idct_add = vp8_dc_only_idct_add_c;
if (flags & HAS_MMX) vp8_dc_only_idct_add = vp8_dc_only_idct_add_mmx;
-
vp8_denoiser_filter = vp8_denoiser_filter_c;
if (flags & HAS_SSE2) vp8_denoiser_filter = vp8_denoiser_filter_sse2;
-
vp8_dequant_idct_add = vp8_dequant_idct_add_c;
if (flags & HAS_MMX) vp8_dequant_idct_add = vp8_dequant_idct_add_mmx;
-
vp8_dequant_idct_add_uv_block = vp8_dequant_idct_add_uv_block_c;
if (flags & HAS_MMX) vp8_dequant_idct_add_uv_block = vp8_dequant_idct_add_uv_block_mmx;
if (flags & HAS_SSE2) vp8_dequant_idct_add_uv_block = vp8_dequant_idct_add_uv_block_sse2;
-
vp8_dequant_idct_add_y_block = vp8_dequant_idct_add_y_block_c;
if (flags & HAS_MMX) vp8_dequant_idct_add_y_block = vp8_dequant_idct_add_y_block_mmx;
if (flags & HAS_SSE2) vp8_dequant_idct_add_y_block = vp8_dequant_idct_add_y_block_sse2;
-
vp8_dequantize_b = vp8_dequantize_b_c;
if (flags & HAS_MMX) vp8_dequantize_b = vp8_dequantize_b_mmx;
-
vp8_diamond_search_sad = vp8_diamond_search_sad_c;
if (flags & HAS_SSE3) vp8_diamond_search_sad = vp8_diamond_search_sadx4;
-
vp8_fast_quantize_b = vp8_fast_quantize_b_c;
if (flags & HAS_SSE2) vp8_fast_quantize_b = vp8_fast_quantize_b_sse2;
if (flags & HAS_SSSE3) vp8_fast_quantize_b = vp8_fast_quantize_b_ssse3;
-
-
vp8_filter_by_weight16x16 = vp8_filter_by_weight16x16_c;
if (flags & HAS_SSE2) vp8_filter_by_weight16x16 = vp8_filter_by_weight16x16_sse2;
-
-
vp8_filter_by_weight8x8 = vp8_filter_by_weight8x8_c;
if (flags & HAS_SSE2) vp8_filter_by_weight8x8 = vp8_filter_by_weight8x8_sse2;
-
vp8_full_search_sad = vp8_full_search_sad_c;
if (flags & HAS_SSE3) vp8_full_search_sad = vp8_full_search_sadx3;
if (flags & HAS_SSE4_1) vp8_full_search_sad = vp8_full_search_sadx8;
-
vp8_get4x4sse_cs = vp8_get4x4sse_cs_c;
if (flags & HAS_MMX) vp8_get4x4sse_cs = vp8_get4x4sse_cs_mmx;
-
vp8_get_mb_ss = vp8_get_mb_ss_c;
if (flags & HAS_MMX) vp8_get_mb_ss = vp8_get_mb_ss_mmx;
if (flags & HAS_SSE2) vp8_get_mb_ss = vp8_get_mb_ss_sse2;
-
-
vp8_loop_filter_bh = vp8_loop_filter_bh_c;
if (flags & HAS_MMX) vp8_loop_filter_bh = vp8_loop_filter_bh_mmx;
if (flags & HAS_SSE2) vp8_loop_filter_bh = vp8_loop_filter_bh_sse2;
-
vp8_loop_filter_bv = vp8_loop_filter_bv_c;
if (flags & HAS_MMX) vp8_loop_filter_bv = vp8_loop_filter_bv_mmx;
if (flags & HAS_SSE2) vp8_loop_filter_bv = vp8_loop_filter_bv_sse2;
-
vp8_loop_filter_mbh = vp8_loop_filter_mbh_c;
if (flags & HAS_MMX) vp8_loop_filter_mbh = vp8_loop_filter_mbh_mmx;
if (flags & HAS_SSE2) vp8_loop_filter_mbh = vp8_loop_filter_mbh_sse2;
-
vp8_loop_filter_mbv = vp8_loop_filter_mbv_c;
if (flags & HAS_MMX) vp8_loop_filter_mbv = vp8_loop_filter_mbv_mmx;
if (flags & HAS_SSE2) vp8_loop_filter_mbv = vp8_loop_filter_mbv_sse2;
-
vp8_loop_filter_simple_bh = vp8_loop_filter_bhs_c;
if (flags & HAS_MMX) vp8_loop_filter_simple_bh = vp8_loop_filter_bhs_mmx;
if (flags & HAS_SSE2) vp8_loop_filter_simple_bh = vp8_loop_filter_bhs_sse2;
-
vp8_loop_filter_simple_bv = vp8_loop_filter_bvs_c;
if (flags & HAS_MMX) vp8_loop_filter_simple_bv = vp8_loop_filter_bvs_mmx;
if (flags & HAS_SSE2) vp8_loop_filter_simple_bv = vp8_loop_filter_bvs_sse2;
-
vp8_loop_filter_simple_mbh = vp8_loop_filter_simple_horizontal_edge_c;
if (flags & HAS_MMX) vp8_loop_filter_simple_mbh = vp8_loop_filter_simple_horizontal_edge_mmx;
if (flags & HAS_SSE2) vp8_loop_filter_simple_mbh = vp8_loop_filter_simple_horizontal_edge_sse2;
-
vp8_loop_filter_simple_mbv = vp8_loop_filter_simple_vertical_edge_c;
if (flags & HAS_MMX) vp8_loop_filter_simple_mbv = vp8_loop_filter_simple_vertical_edge_mmx;
if (flags & HAS_SSE2) vp8_loop_filter_simple_mbv = vp8_loop_filter_simple_vertical_edge_sse2;
-
vp8_mbblock_error = vp8_mbblock_error_c;
if (flags & HAS_MMX) vp8_mbblock_error = vp8_mbblock_error_mmx;
if (flags & HAS_SSE2) vp8_mbblock_error = vp8_mbblock_error_xmm;
-
vp8_mbpost_proc_across_ip = vp8_mbpost_proc_across_ip_c;
if (flags & HAS_SSE2) vp8_mbpost_proc_across_ip = vp8_mbpost_proc_across_ip_xmm;
-
vp8_mbpost_proc_down = vp8_mbpost_proc_down_c;
if (flags & HAS_MMX) vp8_mbpost_proc_down = vp8_mbpost_proc_down_mmx;
if (flags & HAS_SSE2) vp8_mbpost_proc_down = vp8_mbpost_proc_down_xmm;
-
vp8_mbuverror = vp8_mbuverror_c;
if (flags & HAS_MMX) vp8_mbuverror = vp8_mbuverror_mmx;
if (flags & HAS_SSE2) vp8_mbuverror = vp8_mbuverror_xmm;
-
vp8_mse16x16 = vp8_mse16x16_c;
if (flags & HAS_MMX) vp8_mse16x16 = vp8_mse16x16_mmx;
if (flags & HAS_SSE2) vp8_mse16x16 = vp8_mse16x16_wmt;
-
vp8_plane_add_noise = vp8_plane_add_noise_c;
if (flags & HAS_MMX) vp8_plane_add_noise = vp8_plane_add_noise_mmx;
if (flags & HAS_SSE2) vp8_plane_add_noise = vp8_plane_add_noise_wmt;
-
vp8_post_proc_down_and_across_mb_row = vp8_post_proc_down_and_across_mb_row_c;
if (flags & HAS_SSE2) vp8_post_proc_down_and_across_mb_row = vp8_post_proc_down_and_across_mb_row_sse2;
-
-
-
-
vp8_refining_search_sad = vp8_refining_search_sad_c;
if (flags & HAS_SSE3) vp8_refining_search_sad = vp8_refining_search_sadx4;
-
vp8_regular_quantize_b = vp8_regular_quantize_b_c;
if (flags & HAS_SSE2) vp8_regular_quantize_b = vp8_regular_quantize_b_sse2;
-
-
vp8_sad16x16 = vp8_sad16x16_c;
if (flags & HAS_MMX) vp8_sad16x16 = vp8_sad16x16_mmx;
if (flags & HAS_SSE2) vp8_sad16x16 = vp8_sad16x16_wmt;
if (flags & HAS_SSE3) vp8_sad16x16 = vp8_sad16x16_sse3;
-
vp8_sad16x16x3 = vp8_sad16x16x3_c;
if (flags & HAS_SSE3) vp8_sad16x16x3 = vp8_sad16x16x3_sse3;
if (flags & HAS_SSSE3) vp8_sad16x16x3 = vp8_sad16x16x3_ssse3;
-
vp8_sad16x16x4d = vp8_sad16x16x4d_c;
if (flags & HAS_SSE3) vp8_sad16x16x4d = vp8_sad16x16x4d_sse3;
-
vp8_sad16x16x8 = vp8_sad16x16x8_c;
if (flags & HAS_SSE4_1) vp8_sad16x16x8 = vp8_sad16x16x8_sse4;
-
vp8_sad16x8 = vp8_sad16x8_c;
if (flags & HAS_MMX) vp8_sad16x8 = vp8_sad16x8_mmx;
if (flags & HAS_SSE2) vp8_sad16x8 = vp8_sad16x8_wmt;
-
vp8_sad16x8x3 = vp8_sad16x8x3_c;
if (flags & HAS_SSE3) vp8_sad16x8x3 = vp8_sad16x8x3_sse3;
if (flags & HAS_SSSE3) vp8_sad16x8x3 = vp8_sad16x8x3_ssse3;
-
vp8_sad16x8x4d = vp8_sad16x8x4d_c;
if (flags & HAS_SSE3) vp8_sad16x8x4d = vp8_sad16x8x4d_sse3;
-
vp8_sad16x8x8 = vp8_sad16x8x8_c;
if (flags & HAS_SSE4_1) vp8_sad16x8x8 = vp8_sad16x8x8_sse4;
-
vp8_sad4x4 = vp8_sad4x4_c;
if (flags & HAS_MMX) vp8_sad4x4 = vp8_sad4x4_mmx;
if (flags & HAS_SSE2) vp8_sad4x4 = vp8_sad4x4_wmt;
-
vp8_sad4x4x3 = vp8_sad4x4x3_c;
if (flags & HAS_SSE3) vp8_sad4x4x3 = vp8_sad4x4x3_sse3;
-
vp8_sad4x4x4d = vp8_sad4x4x4d_c;
if (flags & HAS_SSE3) vp8_sad4x4x4d = vp8_sad4x4x4d_sse3;
-
vp8_sad4x4x8 = vp8_sad4x4x8_c;
if (flags & HAS_SSE4_1) vp8_sad4x4x8 = vp8_sad4x4x8_sse4;
-
vp8_sad8x16 = vp8_sad8x16_c;
if (flags & HAS_MMX) vp8_sad8x16 = vp8_sad8x16_mmx;
if (flags & HAS_SSE2) vp8_sad8x16 = vp8_sad8x16_wmt;
-
vp8_sad8x16x3 = vp8_sad8x16x3_c;
if (flags & HAS_SSE3) vp8_sad8x16x3 = vp8_sad8x16x3_sse3;
-
vp8_sad8x16x4d = vp8_sad8x16x4d_c;
if (flags & HAS_SSE3) vp8_sad8x16x4d = vp8_sad8x16x4d_sse3;
-
vp8_sad8x16x8 = vp8_sad8x16x8_c;
if (flags & HAS_SSE4_1) vp8_sad8x16x8 = vp8_sad8x16x8_sse4;
-
vp8_sad8x8 = vp8_sad8x8_c;
if (flags & HAS_MMX) vp8_sad8x8 = vp8_sad8x8_mmx;
if (flags & HAS_SSE2) vp8_sad8x8 = vp8_sad8x8_wmt;
-
vp8_sad8x8x3 = vp8_sad8x8x3_c;
if (flags & HAS_SSE3) vp8_sad8x8x3 = vp8_sad8x8x3_sse3;
-
vp8_sad8x8x4d = vp8_sad8x8x4d_c;
if (flags & HAS_SSE3) vp8_sad8x8x4d = vp8_sad8x8x4d_sse3;
-
vp8_sad8x8x8 = vp8_sad8x8x8_c;
if (flags & HAS_SSE4_1) vp8_sad8x8x8 = vp8_sad8x8x8_sse4;
-
vp8_short_fdct4x4 = vp8_short_fdct4x4_c;
if (flags & HAS_MMX) vp8_short_fdct4x4 = vp8_short_fdct4x4_mmx;
if (flags & HAS_SSE2) vp8_short_fdct4x4 = vp8_short_fdct4x4_sse2;
-
vp8_short_fdct8x4 = vp8_short_fdct8x4_c;
if (flags & HAS_MMX) vp8_short_fdct8x4 = vp8_short_fdct8x4_mmx;
if (flags & HAS_SSE2) vp8_short_fdct8x4 = vp8_short_fdct8x4_sse2;
-
vp8_short_idct4x4llm = vp8_short_idct4x4llm_c;
if (flags & HAS_MMX) vp8_short_idct4x4llm = vp8_short_idct4x4llm_mmx;
-
vp8_short_inv_walsh4x4 = vp8_short_inv_walsh4x4_c;
if (flags & HAS_MMX) vp8_short_inv_walsh4x4 = vp8_short_inv_walsh4x4_mmx;
if (flags & HAS_SSE2) vp8_short_inv_walsh4x4 = vp8_short_inv_walsh4x4_sse2;
-
-
vp8_short_walsh4x4 = vp8_short_walsh4x4_c;
if (flags & HAS_SSE2) vp8_short_walsh4x4 = vp8_short_walsh4x4_sse2;
-
vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_c;
if (flags & HAS_MMX) vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_mmx;
if (flags & HAS_SSE2) vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_sse2;
if (flags & HAS_SSSE3) vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_ssse3;
-
vp8_sixtap_predict4x4 = vp8_sixtap_predict4x4_c;
if (flags & HAS_MMX) vp8_sixtap_predict4x4 = vp8_sixtap_predict4x4_mmx;
if (flags & HAS_SSSE3) vp8_sixtap_predict4x4 = vp8_sixtap_predict4x4_ssse3;
-
vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_c;
if (flags & HAS_MMX) vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_mmx;
if (flags & HAS_SSE2) vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_sse2;
if (flags & HAS_SSSE3) vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_ssse3;
-
vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_c;
if (flags & HAS_MMX) vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_mmx;
if (flags & HAS_SSE2) vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_sse2;
if (flags & HAS_SSSE3) vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_ssse3;
-
vp8_sub_pixel_mse16x16 = vp8_sub_pixel_mse16x16_c;
if (flags & HAS_MMX) vp8_sub_pixel_mse16x16 = vp8_sub_pixel_mse16x16_mmx;
if (flags & HAS_SSE2) vp8_sub_pixel_mse16x16 = vp8_sub_pixel_mse16x16_wmt;
-
vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_c;
if (flags & HAS_MMX) vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_mmx;
if (flags & HAS_SSE2) vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_wmt;
if (flags & HAS_SSSE3) vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_ssse3;
-
vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_c;
if (flags & HAS_MMX) vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_mmx;
if (flags & HAS_SSE2) vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_wmt;
if (flags & HAS_SSSE3) vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_ssse3;
-
vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_c;
if (flags & HAS_MMX) vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_mmx;
if (flags & HAS_SSE2) vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_wmt;
-
vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x16_c;
if (flags & HAS_MMX) vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x16_mmx;
if (flags & HAS_SSE2) vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x16_wmt;
-
vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_c;
if (flags & HAS_MMX) vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_mmx;
if (flags & HAS_SSE2) vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_wmt;
-
vp8_subtract_b = vp8_subtract_b_c;
if (flags & HAS_MMX) vp8_subtract_b = vp8_subtract_b_mmx;
if (flags & HAS_SSE2) vp8_subtract_b = vp8_subtract_b_sse2;
-
vp8_subtract_mbuv = vp8_subtract_mbuv_c;
if (flags & HAS_MMX) vp8_subtract_mbuv = vp8_subtract_mbuv_mmx;
if (flags & HAS_SSE2) vp8_subtract_mbuv = vp8_subtract_mbuv_sse2;
-
vp8_subtract_mby = vp8_subtract_mby_c;
if (flags & HAS_MMX) vp8_subtract_mby = vp8_subtract_mby_mmx;
if (flags & HAS_SSE2) vp8_subtract_mby = vp8_subtract_mby_sse2;
-
vp8_variance16x16 = vp8_variance16x16_c;
if (flags & HAS_MMX) vp8_variance16x16 = vp8_variance16x16_mmx;
if (flags & HAS_SSE2) vp8_variance16x16 = vp8_variance16x16_wmt;
-
vp8_variance16x8 = vp8_variance16x8_c;
if (flags & HAS_MMX) vp8_variance16x8 = vp8_variance16x8_mmx;
if (flags & HAS_SSE2) vp8_variance16x8 = vp8_variance16x8_wmt;
-
vp8_variance4x4 = vp8_variance4x4_c;
if (flags & HAS_MMX) vp8_variance4x4 = vp8_variance4x4_mmx;
if (flags & HAS_SSE2) vp8_variance4x4 = vp8_variance4x4_wmt;
-
vp8_variance8x16 = vp8_variance8x16_c;
if (flags & HAS_MMX) vp8_variance8x16 = vp8_variance8x16_mmx;
if (flags & HAS_SSE2) vp8_variance8x16 = vp8_variance8x16_wmt;
-
vp8_variance8x8 = vp8_variance8x8_c;
if (flags & HAS_MMX) vp8_variance8x8 = vp8_variance8x8_mmx;
if (flags & HAS_SSE2) vp8_variance8x8 = vp8_variance8x8_wmt;
-
vp8_variance_halfpixvar16x16_h = vp8_variance_halfpixvar16x16_h_c;
if (flags & HAS_MMX) vp8_variance_halfpixvar16x16_h = vp8_variance_halfpixvar16x16_h_mmx;
if (flags & HAS_SSE2) vp8_variance_halfpixvar16x16_h = vp8_variance_halfpixvar16x16_h_wmt;
-
vp8_variance_halfpixvar16x16_hv = vp8_variance_halfpixvar16x16_hv_c;
if (flags & HAS_MMX) vp8_variance_halfpixvar16x16_hv = vp8_variance_halfpixvar16x16_hv_mmx;
if (flags & HAS_SSE2) vp8_variance_halfpixvar16x16_hv = vp8_variance_halfpixvar16x16_hv_wmt;
-
vp8_variance_halfpixvar16x16_v = vp8_variance_halfpixvar16x16_v_c;
if (flags & HAS_MMX) vp8_variance_halfpixvar16x16_v = vp8_variance_halfpixvar16x16_v_mmx;
if (flags & HAS_SSE2) vp8_variance_halfpixvar16x16_v = vp8_variance_halfpixvar16x16_v_wmt;