diff options
Diffstat (limited to 'webrtc/modules/video_processing/main/test/unit_test/video_processing_unittest.cc')
-rw-r--r-- | webrtc/modules/video_processing/main/test/unit_test/video_processing_unittest.cc | 390 |
1 files changed, 0 insertions, 390 deletions
diff --git a/webrtc/modules/video_processing/main/test/unit_test/video_processing_unittest.cc b/webrtc/modules/video_processing/main/test/unit_test/video_processing_unittest.cc deleted file mode 100644 index 11ccc4891b..0000000000 --- a/webrtc/modules/video_processing/main/test/unit_test/video_processing_unittest.cc +++ /dev/null @@ -1,390 +0,0 @@ -/* - * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. - * - * Use of this source code is governed by a BSD-style license - * that can be found in the LICENSE file in the root of the source - * tree. An additional intellectual property rights grant can be found - * in the file PATENTS. All contributing project authors may - * be found in the AUTHORS file in the root of the source tree. - */ - -#include "webrtc/modules/video_processing/main/test/unit_test/video_processing_unittest.h" - -#include <string> - -#include <gflags/gflags.h> -#include "webrtc/common_video/libyuv/include/webrtc_libyuv.h" -#include "webrtc/system_wrappers/include/tick_util.h" -#include "webrtc/test/testsupport/fileutils.h" -#include "webrtc/test/testsupport/gtest_disable.h" - -namespace webrtc { - -namespace { - -// Define command line flag 'gen_files' (default value: false). -DEFINE_bool(gen_files, false, "Output files for visual inspection."); - -} // namespace - -static void PreprocessFrameAndVerify(const VideoFrame& source, - int target_width, - int target_height, - VideoProcessingModule* vpm, - VideoFrame** out_frame); -static void CropFrame(const uint8_t* source_data, - int source_width, - int source_height, - int offset_x, - int offset_y, - int cropped_width, - int cropped_height, - VideoFrame* cropped_frame); -// The |source_data| is cropped and scaled to |target_width| x |target_height|, -// and then scaled back to the expected cropped size. |expected_psnr| is used to -// verify basic quality, and is set to be ~0.1/0.05dB lower than actual PSNR -// verified under the same conditions. -static void TestSize(const VideoFrame& source_frame, - const VideoFrame& cropped_source_frame, - int target_width, - int target_height, - double expected_psnr, - VideoProcessingModule* vpm); -static bool CompareFrames(const webrtc::VideoFrame& frame1, - const webrtc::VideoFrame& frame2); -static void WriteProcessedFrameForVisualInspection(const VideoFrame& source, - const VideoFrame& processed); - -VideoProcessingModuleTest::VideoProcessingModuleTest() - : vpm_(NULL), - source_file_(NULL), - width_(352), - half_width_((width_ + 1) / 2), - height_(288), - size_y_(width_ * height_), - size_uv_(half_width_ * ((height_ + 1) / 2)), - frame_length_(CalcBufferSize(kI420, width_, height_)) {} - -void VideoProcessingModuleTest::SetUp() { - vpm_ = VideoProcessingModule::Create(); - ASSERT_TRUE(vpm_ != NULL); - - ASSERT_EQ(0, video_frame_.CreateEmptyFrame(width_, height_, width_, - half_width_, half_width_)); - // Clear video frame so DrMemory/Valgrind will allow reads of the buffer. - memset(video_frame_.buffer(kYPlane), 0, video_frame_.allocated_size(kYPlane)); - memset(video_frame_.buffer(kUPlane), 0, video_frame_.allocated_size(kUPlane)); - memset(video_frame_.buffer(kVPlane), 0, video_frame_.allocated_size(kVPlane)); - const std::string video_file = - webrtc::test::ResourcePath("foreman_cif", "yuv"); - source_file_ = fopen(video_file.c_str(),"rb"); - ASSERT_TRUE(source_file_ != NULL) << - "Cannot read source file: " + video_file + "\n"; -} - -void VideoProcessingModuleTest::TearDown() { - if (source_file_ != NULL) { - ASSERT_EQ(0, fclose(source_file_)); - } - source_file_ = NULL; - - if (vpm_ != NULL) { - VideoProcessingModule::Destroy(vpm_); - } - vpm_ = NULL; -} - -TEST_F(VideoProcessingModuleTest, DISABLED_ON_IOS(HandleNullBuffer)) { - // TODO(mikhal/stefan): Do we need this one? - VideoProcessingModule::FrameStats stats; - // Video frame with unallocated buffer. - VideoFrame videoFrame; - - EXPECT_EQ(-3, vpm_->GetFrameStats(&stats, videoFrame)); - - EXPECT_EQ(-1, vpm_->Deflickering(&videoFrame, &stats)); - - EXPECT_EQ(-3, vpm_->BrightnessDetection(videoFrame, stats)); -} - -TEST_F(VideoProcessingModuleTest, DISABLED_ON_IOS(HandleBadStats)) { - VideoProcessingModule::FrameStats stats; - rtc::scoped_ptr<uint8_t[]> video_buffer(new uint8_t[frame_length_]); - ASSERT_EQ(frame_length_, fread(video_buffer.get(), 1, frame_length_, - source_file_)); - EXPECT_EQ(0, ConvertToI420(kI420, video_buffer.get(), 0, 0, width_, height_, - 0, kVideoRotation_0, &video_frame_)); - - EXPECT_EQ(-1, vpm_->Deflickering(&video_frame_, &stats)); - - EXPECT_EQ(-3, vpm_->BrightnessDetection(video_frame_, stats)); -} - -TEST_F(VideoProcessingModuleTest, DISABLED_ON_IOS(IdenticalResultsAfterReset)) { - VideoFrame video_frame2; - VideoProcessingModule::FrameStats stats; - // Only testing non-static functions here. - rtc::scoped_ptr<uint8_t[]> video_buffer(new uint8_t[frame_length_]); - ASSERT_EQ(frame_length_, fread(video_buffer.get(), 1, frame_length_, - source_file_)); - EXPECT_EQ(0, ConvertToI420(kI420, video_buffer.get(), 0, 0, width_, height_, - 0, kVideoRotation_0, &video_frame_)); - ASSERT_EQ(0, vpm_->GetFrameStats(&stats, video_frame_)); - ASSERT_EQ(0, video_frame2.CopyFrame(video_frame_)); - ASSERT_EQ(0, vpm_->Deflickering(&video_frame_, &stats)); - vpm_->Reset(); - // Retrieve frame stats again in case Deflickering() has zeroed them. - ASSERT_EQ(0, vpm_->GetFrameStats(&stats, video_frame2)); - ASSERT_EQ(0, vpm_->Deflickering(&video_frame2, &stats)); - EXPECT_TRUE(CompareFrames(video_frame_, video_frame2)); - - ASSERT_EQ(frame_length_, fread(video_buffer.get(), 1, frame_length_, - source_file_)); - EXPECT_EQ(0, ConvertToI420(kI420, video_buffer.get(), 0, 0, width_, height_, - 0, kVideoRotation_0, &video_frame_)); - ASSERT_EQ(0, vpm_->GetFrameStats(&stats, video_frame_)); - video_frame2.CopyFrame(video_frame_); - ASSERT_EQ(0, vpm_->BrightnessDetection(video_frame_, stats)); - vpm_->Reset(); - ASSERT_EQ(0, vpm_->BrightnessDetection(video_frame2, stats)); - EXPECT_TRUE(CompareFrames(video_frame_, video_frame2)); -} - -TEST_F(VideoProcessingModuleTest, DISABLED_ON_IOS(FrameStats)) { - VideoProcessingModule::FrameStats stats; - rtc::scoped_ptr<uint8_t[]> video_buffer(new uint8_t[frame_length_]); - ASSERT_EQ(frame_length_, fread(video_buffer.get(), 1, frame_length_, - source_file_)); - EXPECT_EQ(0, ConvertToI420(kI420, video_buffer.get(), 0, 0, width_, height_, - 0, kVideoRotation_0, &video_frame_)); - - EXPECT_FALSE(vpm_->ValidFrameStats(stats)); - EXPECT_EQ(0, vpm_->GetFrameStats(&stats, video_frame_)); - EXPECT_TRUE(vpm_->ValidFrameStats(stats)); - - printf("\nFrameStats\n"); - printf("mean: %u\nnum_pixels: %u\nsubSamplWidth: " - "%u\nsumSamplHeight: %u\nsum: %u\n\n", - static_cast<unsigned int>(stats.mean), - static_cast<unsigned int>(stats.num_pixels), - static_cast<unsigned int>(stats.subSamplHeight), - static_cast<unsigned int>(stats.subSamplWidth), - static_cast<unsigned int>(stats.sum)); - - vpm_->ClearFrameStats(&stats); - EXPECT_FALSE(vpm_->ValidFrameStats(stats)); -} - -TEST_F(VideoProcessingModuleTest, DISABLED_ON_IOS(PreprocessorLogic)) { - // Disable temporal sampling (frame dropping). - vpm_->EnableTemporalDecimation(false); - int resolution = 100; - EXPECT_EQ(VPM_OK, vpm_->SetTargetResolution(resolution, resolution, 15)); - EXPECT_EQ(VPM_OK, vpm_->SetTargetResolution(resolution, resolution, 30)); - // Disable spatial sampling. - vpm_->SetInputFrameResampleMode(kNoRescaling); - EXPECT_EQ(VPM_OK, vpm_->SetTargetResolution(resolution, resolution, 30)); - VideoFrame* out_frame = NULL; - // Set rescaling => output frame != NULL. - vpm_->SetInputFrameResampleMode(kFastRescaling); - PreprocessFrameAndVerify(video_frame_, resolution, resolution, vpm_, - &out_frame); - // No rescaling=> output frame = NULL. - vpm_->SetInputFrameResampleMode(kNoRescaling); - EXPECT_EQ(VPM_OK, vpm_->PreprocessFrame(video_frame_, &out_frame)); - EXPECT_TRUE(out_frame == NULL); -} - -TEST_F(VideoProcessingModuleTest, DISABLED_ON_IOS(Resampler)) { - enum { NumRuns = 1 }; - - int64_t min_runtime = 0; - int64_t total_runtime = 0; - - rewind(source_file_); - ASSERT_TRUE(source_file_ != NULL) << - "Cannot read input file \n"; - - // CA not needed here - vpm_->EnableContentAnalysis(false); - // no temporal decimation - vpm_->EnableTemporalDecimation(false); - - // Reading test frame - rtc::scoped_ptr<uint8_t[]> video_buffer(new uint8_t[frame_length_]); - ASSERT_EQ(frame_length_, fread(video_buffer.get(), 1, frame_length_, - source_file_)); - // Using ConvertToI420 to add stride to the image. - EXPECT_EQ(0, ConvertToI420(kI420, video_buffer.get(), 0, 0, width_, height_, - 0, kVideoRotation_0, &video_frame_)); - // Cropped source frame that will contain the expected visible region. - VideoFrame cropped_source_frame; - cropped_source_frame.CopyFrame(video_frame_); - - for (uint32_t run_idx = 0; run_idx < NumRuns; run_idx++) { - // Initiate test timer. - const TickTime time_start = TickTime::Now(); - - // Init the sourceFrame with a timestamp. - video_frame_.set_render_time_ms(time_start.MillisecondTimestamp()); - video_frame_.set_timestamp(time_start.MillisecondTimestamp() * 90); - - // Test scaling to different sizes: source is of |width|/|height| = 352/288. - // Pure scaling: - TestSize(video_frame_, video_frame_, width_ / 4, height_ / 4, 25.2, vpm_); - TestSize(video_frame_, video_frame_, width_ / 2, height_ / 2, 28.1, vpm_); - // No resampling: - TestSize(video_frame_, video_frame_, width_, height_, -1, vpm_); - TestSize(video_frame_, video_frame_, 2 * width_, 2 * height_, 32.2, vpm_); - - // Scaling and cropping. The cropped source frame is the largest center - // aligned region that can be used from the source while preserving aspect - // ratio. - CropFrame(video_buffer.get(), width_, height_, 0, 56, 352, 176, - &cropped_source_frame); - TestSize(video_frame_, cropped_source_frame, 100, 50, 24.0, vpm_); - - CropFrame(video_buffer.get(), width_, height_, 0, 30, 352, 225, - &cropped_source_frame); - TestSize(video_frame_, cropped_source_frame, 400, 256, 31.3, vpm_); - - CropFrame(video_buffer.get(), width_, height_, 68, 0, 216, 288, - &cropped_source_frame); - TestSize(video_frame_, cropped_source_frame, 480, 640, 32.15, vpm_); - - CropFrame(video_buffer.get(), width_, height_, 0, 12, 352, 264, - &cropped_source_frame); - TestSize(video_frame_, cropped_source_frame, 960, 720, 32.2, vpm_); - - CropFrame(video_buffer.get(), width_, height_, 0, 44, 352, 198, - &cropped_source_frame); - TestSize(video_frame_, cropped_source_frame, 1280, 720, 32.15, vpm_); - - // Upsampling to odd size. - CropFrame(video_buffer.get(), width_, height_, 0, 26, 352, 233, - &cropped_source_frame); - TestSize(video_frame_, cropped_source_frame, 501, 333, 32.05, vpm_); - // Downsample to odd size. - CropFrame(video_buffer.get(), width_, height_, 0, 34, 352, 219, - &cropped_source_frame); - TestSize(video_frame_, cropped_source_frame, 281, 175, 29.3, vpm_); - - // Stop timer. - const int64_t runtime = (TickTime::Now() - time_start).Microseconds(); - if (runtime < min_runtime || run_idx == 0) { - min_runtime = runtime; - } - total_runtime += runtime; - } - - printf("\nAverage run time = %d us / frame\n", - static_cast<int>(total_runtime)); - printf("Min run time = %d us / frame\n\n", - static_cast<int>(min_runtime)); -} - -void PreprocessFrameAndVerify(const VideoFrame& source, - int target_width, - int target_height, - VideoProcessingModule* vpm, - VideoFrame** out_frame) { - ASSERT_EQ(VPM_OK, vpm->SetTargetResolution(target_width, target_height, 30)); - ASSERT_EQ(VPM_OK, vpm->PreprocessFrame(source, out_frame)); - - // If no resizing is needed, expect NULL. - if (target_width == source.width() && target_height == source.height()) { - EXPECT_EQ(NULL, *out_frame); - return; - } - - // Verify the resampled frame. - EXPECT_TRUE(*out_frame != NULL); - EXPECT_EQ(source.render_time_ms(), (*out_frame)->render_time_ms()); - EXPECT_EQ(source.timestamp(), (*out_frame)->timestamp()); - EXPECT_EQ(target_width, (*out_frame)->width()); - EXPECT_EQ(target_height, (*out_frame)->height()); -} - -void CropFrame(const uint8_t* source_data, - int source_width, - int source_height, - int offset_x, - int offset_y, - int cropped_width, - int cropped_height, - VideoFrame* cropped_frame) { - cropped_frame->CreateEmptyFrame(cropped_width, cropped_height, cropped_width, - (cropped_width + 1) / 2, - (cropped_width + 1) / 2); - EXPECT_EQ(0, - ConvertToI420(kI420, source_data, offset_x, offset_y, source_width, - source_height, 0, kVideoRotation_0, cropped_frame)); -} - -void TestSize(const VideoFrame& source_frame, - const VideoFrame& cropped_source_frame, - int target_width, - int target_height, - double expected_psnr, - VideoProcessingModule* vpm) { - // Resample source_frame to out_frame. - VideoFrame* out_frame = NULL; - vpm->SetInputFrameResampleMode(kBox); - PreprocessFrameAndVerify(source_frame, target_width, target_height, vpm, - &out_frame); - if (out_frame == NULL) - return; - WriteProcessedFrameForVisualInspection(source_frame, *out_frame); - - // Scale |resampled_source_frame| back to the source scale. - VideoFrame resampled_source_frame; - resampled_source_frame.CopyFrame(*out_frame); - PreprocessFrameAndVerify(resampled_source_frame, cropped_source_frame.width(), - cropped_source_frame.height(), vpm, &out_frame); - WriteProcessedFrameForVisualInspection(resampled_source_frame, *out_frame); - - // Compute PSNR against the cropped source frame and check expectation. - double psnr = I420PSNR(&cropped_source_frame, out_frame); - EXPECT_GT(psnr, expected_psnr); - printf("PSNR: %f. PSNR is between source of size %d %d, and a modified " - "source which is scaled down/up to: %d %d, and back to source size \n", - psnr, source_frame.width(), source_frame.height(), - target_width, target_height); -} - -bool CompareFrames(const webrtc::VideoFrame& frame1, - const webrtc::VideoFrame& frame2) { - for (int plane = 0; plane < webrtc::kNumOfPlanes; plane ++) { - webrtc::PlaneType plane_type = static_cast<webrtc::PlaneType>(plane); - int allocated_size1 = frame1.allocated_size(plane_type); - int allocated_size2 = frame2.allocated_size(plane_type); - if (allocated_size1 != allocated_size2) - return false; - const uint8_t* plane_buffer1 = frame1.buffer(plane_type); - const uint8_t* plane_buffer2 = frame2.buffer(plane_type); - if (memcmp(plane_buffer1, plane_buffer2, allocated_size1)) - return false; - } - return true; -} - -void WriteProcessedFrameForVisualInspection(const VideoFrame& source, - const VideoFrame& processed) { - // Skip if writing to files is not enabled. - if (!FLAGS_gen_files) - return; - // Write the processed frame to file for visual inspection. - std::ostringstream filename; - filename << webrtc::test::OutputPath() << "Resampler_from_" << source.width() - << "x" << source.height() << "_to_" << processed.width() << "x" - << processed.height() << "_30Hz_P420.yuv"; - std::cout << "Watch " << filename.str() << " and verify that it is okay." - << std::endl; - FILE* stand_alone_file = fopen(filename.str().c_str(), "wb"); - if (PrintVideoFrame(processed, stand_alone_file) < 0) - std::cerr << "Failed to write: " << filename.str() << std::endl; - if (stand_alone_file) - fclose(stand_alone_file); -} - -} // namespace webrtc |