| // Copyright (c) 2012 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include <stddef.h> |
| #include <stdint.h> |
| |
| #include <memory> |
| #include <utility> |
| |
| #include "base/bind.h" |
| #include "base/command_line.h" |
| #include "base/macros.h" |
| #include "base/memory/ptr_util.h" |
| #include "base/memory/ref_counted.h" |
| #include "base/run_loop.h" |
| #include "base/strings/string_split.h" |
| #include "base/strings/string_util.h" |
| #include "base/threading/thread_task_runner_handle.h" |
| #include "build/build_config.h" |
| #include "media/base/cdm_callback_promise.h" |
| #include "media/base/cdm_key_information.h" |
| #include "media/base/decoder_buffer.h" |
| #include "media/base/media.h" |
| #include "media/base/media_switches.h" |
| #include "media/base/media_tracks.h" |
| #include "media/base/mock_media_log.h" |
| #include "media/base/test_data_util.h" |
| #include "media/base/timestamp_constants.h" |
| #include "media/cdm/aes_decryptor.h" |
| #include "media/cdm/json_web_key.h" |
| #include "media/media_features.h" |
| #include "media/renderers/renderer_impl.h" |
| #include "media/test/fake_encrypted_media.h" |
| #include "media/test/mock_media_source.h" |
| #include "media/test/pipeline_integration_test_base.h" |
| #include "testing/gmock/include/gmock/gmock.h" |
| #include "third_party/libaom/av1_features.h" |
| #include "url/gurl.h" |
| |
| #if defined(MOJO_RENDERER) |
| #include "media/mojo/clients/mojo_renderer.h" |
| #include "media/mojo/interfaces/interface_factory.mojom.h" |
| #include "media/mojo/interfaces/renderer.mojom.h" |
| #include "services/service_manager/public/cpp/connect.h" |
| #include "services/service_manager/public/cpp/service_test.h" |
| |
| // TODO(dalecurtis): The mojo renderer is in another process, so we have no way |
| // currently to get hashes for video and audio samples. This also means that |
| // real audio plays out for each test. |
| #define EXPECT_HASH_EQ(a, b) |
| #define EXPECT_VIDEO_FORMAT_EQ(a, b) |
| #define EXPECT_COLOR_SPACE_EQ(a, b) |
| |
| // TODO(xhwang): EME support is not complete for the mojo renderer, so all |
| // encrypted tests are currently disabled. |
| #define DISABLE_EME_TESTS 1 |
| |
| // TODO(xhwang,dalecurtis): Text tracks are not currently supported by the mojo |
| // renderer. |
| #define DISABLE_TEXT_TRACK_TESTS 1 |
| |
| #else |
| #define EXPECT_HASH_EQ(a, b) EXPECT_EQ(a, b) |
| #define EXPECT_VIDEO_FORMAT_EQ(a, b) EXPECT_EQ(a, b) |
| #define EXPECT_COLOR_SPACE_EQ(a, b) EXPECT_EQ(a, b) |
| #endif // defined(MOJO_RENDERER) |
| |
| #if defined(DISABLE_EME_TESTS) |
| #define MAYBE_EME(test) DISABLED_##test |
| #else |
| #define MAYBE_EME(test) test |
| #endif |
| |
| // To use MAYBE_EME in a parameterized test fixture, don't directly use TEST_P |
| // (because "MAYBE_EME" will literally be used as part of the test name). |
| // Instead, use this wrapper macro to ensure that this first level of parameter |
| // expansion is done before the INSTANTIATE_TEST_CASE_P macro is processed. |
| // For precedent, see similar IN_PROC_BROWSER_TEST_P definition. |
| #define MAYBE_EME_TEST_P(fixture, test) TEST_P(fixture, test) |
| |
| #if defined(DISABLE_TEXT_TRACK_TESTS) |
| #define MAYBE_TEXT(test) DISABLED_##test |
| #else |
| #define MAYBE_TEXT(test) test |
| #endif |
| |
| using ::testing::_; |
| using ::testing::AnyNumber; |
| using ::testing::AtLeast; |
| using ::testing::AtMost; |
| using ::testing::HasSubstr; |
| using ::testing::SaveArg; |
| |
| namespace { |
| |
| // See https://crbug.com/718641 and kMseBufferByPts. This controls which kind of |
| // buffering implementation is constructed and tested in MSE portion of these |
| // tests. |
| enum class BufferingApi { kLegacyByDts, kNewByPts }; |
| |
| } // namespace |
| |
| namespace media { |
| |
| #if BUILDFLAG(ENABLE_AV1_DECODER) |
| // TODO(dalecurtis): This is not the correct final string. Fix before enabling |
| // by default. http://crbug.com/784607 |
| const char kWebMAV1[] = "video/webm; codecs=\"av1\""; |
| #endif |
| |
| const char kWebM[] = "video/webm; codecs=\"vp8,vorbis\""; |
| const char kWebMVP9[] = "video/webm; codecs=\"vp9\""; |
| const char kAudioOnlyWebM[] = "video/webm; codecs=\"vorbis\""; |
| const char kOpusAudioOnlyWebM[] = "video/webm; codecs=\"opus\""; |
| const char kVideoOnlyWebM[] = "video/webm; codecs=\"vp8\""; |
| const char kMP4VideoVP9[] = |
| "video/mp4; codecs=\"vp09.00.10.08.01.02.02.02.00\""; |
| const char kMP4AudioFlac[] = "audio/mp4; codecs=\"flac\""; |
| const char kMP3[] = "audio/mpeg"; |
| #if BUILDFLAG(ENABLE_AV1_DECODER) |
| // TODO(dalecurtis): This is not the correct final string. Fix before enabling |
| // by default. http://crbug.com/784607 |
| const char kMP4AV1[] = "video/mp4; codecs=\"av1\""; |
| #endif // BUILDFLAG(ENABLE_AV1_DECODER) |
| #if BUILDFLAG(USE_PROPRIETARY_CODECS) |
| const char kADTS[] = "audio/aac"; |
| const char kMP4[] = "video/mp4; codecs=\"avc1.4D4041,mp4a.40.2\""; |
| const char kMP4VideoAVC3[] = "video/mp4; codecs=\"avc3.64001f\""; |
| const char kMP4VideoHEVC1[] = "video/mp4; codecs=\"hvc1.1.6.L93.B0\""; |
| const char kMP4VideoHEVC2[] = "video/mp4; codecs=\"hev1.1.6.L93.B0\""; |
| const char kMP4Video[] = "video/mp4; codecs=\"avc1.4D4041\""; |
| const char kMP4Audio[] = "audio/mp4; codecs=\"mp4a.40.2\""; |
| const char kMP2AudioSBR[] = "video/mp2t; codecs=\"avc1.4D4041,mp4a.40.5\""; |
| #endif // BUILDFLAG(USE_PROPRIETARY_CODECS) |
| |
| // Constants for the Media Source config change tests. |
| const int kAppendTimeSec = 1; |
| const int kAppendTimeMs = kAppendTimeSec * 1000; |
| const int k320WebMFileDurationMs = 2736; |
| const int k640WebMFileDurationMs = 2749; |
| const int kOpusEndTrimmingWebMFileDurationMs = 2741; |
| const int kVP9WebMFileDurationMs = 2736; |
| const int kVP8AWebMFileDurationMs = 2734; |
| |
| #if !defined(MOJO_RENDERER) |
| static const char kSfxLosslessHash[] = "3.03,2.86,2.99,3.31,3.57,4.06,"; |
| |
| #if defined(OPUS_FIXED_POINT) |
| // NOTE: Hashes are specific to ARM devices. x86 will not match. |
| static const char kOpusEndTrimmingHash_1[] = |
| "-4.57,-5.66,-6.52,-6.30,-4.37,-3.61,"; |
| static const char kOpusEndTrimmingHash_2[] = |
| "-11.91,-11.11,-8.27,-7.13,-7.86,-10.00,"; |
| static const char kOpusEndTrimmingHash_3[] = |
| "-13.31,-14.38,-13.70,-11.71,-10.21,-10.49,"; |
| static const char kOpusSmallCodecDelayHash_1[] = |
| "-0.48,-0.09,1.27,1.06,1.54,-0.22,"; |
| static const char kOpusSmallCodecDelayHash_2[] = |
| "0.29,0.14,-0.20,0.24,0.68,0.83,"; |
| #else |
| // Hash for a full playthrough of "opus-trimming-test.(webm|ogg)". |
| static const char kOpusEndTrimmingHash_1[] = |
| "-4.56,-5.65,-6.51,-6.29,-4.36,-3.59,"; |
| // The above hash, plus an additional playthrough starting from T=1s. |
| static const char kOpusEndTrimmingHash_2[] = |
| "-11.89,-11.09,-8.25,-7.11,-7.84,-9.97,"; |
| // The above hash, plus an additional playthrough starting from T=6.36s. |
| static const char kOpusEndTrimmingHash_3[] = |
| "-13.28,-14.35,-13.67,-11.68,-10.18,-10.46,"; |
| // Hash for a full playthrough of "bear-opus.webm". |
| static const char kOpusSmallCodecDelayHash_1[] = |
| "-0.47,-0.09,1.28,1.07,1.55,-0.22,"; |
| // The above hash, plus an additional playthrough starting from T=1.414s. |
| static const char kOpusSmallCodecDelayHash_2[] = |
| "0.31,0.15,-0.18,0.25,0.70,0.84,"; |
| #endif // defined(OPUS_FIXED_POINT) |
| #endif // !defined(MOJO_RENDERER) |
| |
| #if BUILDFLAG(USE_PROPRIETARY_CODECS) |
| const int k640IsoFileDurationMs = 2737; |
| const int k640IsoCencFileDurationMsByDts = 2736; |
| const int k640IsoCencFileDurationMsByPts = 2769; |
| const int k1280IsoFileDurationMs = 2736; |
| const int k1280IsoFileDurationMsAVByPts = 2763; |
| const int k1280IsoAVC3FileDurationMs = 2736; |
| #endif // BUILDFLAG(USE_PROPRIETARY_CODECS) |
| |
| // Return a timeline offset for bear-320x240-live.webm. |
| static base::Time kLiveTimelineOffset() { |
| // The file contians the following UTC timeline offset: |
| // 2012-11-10 12:34:56.789123456 |
| // Since base::Time only has a resolution of microseconds, |
| // construct a base::Time for 2012-11-10 12:34:56.789123. |
| base::Time::Exploded exploded_time; |
| exploded_time.year = 2012; |
| exploded_time.month = 11; |
| exploded_time.day_of_month = 10; |
| exploded_time.day_of_week = 6; |
| exploded_time.hour = 12; |
| exploded_time.minute = 34; |
| exploded_time.second = 56; |
| exploded_time.millisecond = 789; |
| base::Time timeline_offset; |
| EXPECT_TRUE(base::Time::FromUTCExploded(exploded_time, &timeline_offset)); |
| |
| timeline_offset += base::TimeDelta::FromMicroseconds(123); |
| |
| return timeline_offset; |
| } |
| |
| enum PromiseResult { RESOLVED, REJECTED }; |
| |
| // Provides the test key in response to the encrypted event. |
| class KeyProvidingApp : public FakeEncryptedMedia::AppBase { |
| public: |
| KeyProvidingApp() = default; |
| |
| void OnResolveWithSession(PromiseResult expected, |
| const std::string& session_id) { |
| EXPECT_EQ(expected, RESOLVED); |
| EXPECT_GT(session_id.length(), 0ul); |
| current_session_id_ = session_id; |
| } |
| |
| void OnResolve(PromiseResult expected) { EXPECT_EQ(expected, RESOLVED); } |
| |
| void OnReject(PromiseResult expected, |
| media::CdmPromise::Exception exception_code, |
| uint32_t system_code, |
| const std::string& error_message) { |
| EXPECT_EQ(expected, REJECTED) << error_message; |
| } |
| |
| std::unique_ptr<SimpleCdmPromise> CreatePromise(PromiseResult expected) { |
| std::unique_ptr<media::SimpleCdmPromise> promise( |
| new media::CdmCallbackPromise<>( |
| base::Bind(&KeyProvidingApp::OnResolve, base::Unretained(this), |
| expected), |
| base::Bind(&KeyProvidingApp::OnReject, base::Unretained(this), |
| expected))); |
| return promise; |
| } |
| |
| std::unique_ptr<NewSessionCdmPromise> CreateSessionPromise( |
| PromiseResult expected) { |
| std::unique_ptr<media::NewSessionCdmPromise> promise( |
| new media::CdmCallbackPromise<std::string>( |
| base::Bind(&KeyProvidingApp::OnResolveWithSession, |
| base::Unretained(this), expected), |
| base::Bind(&KeyProvidingApp::OnReject, base::Unretained(this), |
| expected))); |
| return promise; |
| } |
| |
| void OnSessionMessage(const std::string& session_id, |
| CdmMessageType message_type, |
| const std::vector<uint8_t>& message, |
| AesDecryptor* decryptor) override { |
| EXPECT_FALSE(session_id.empty()); |
| EXPECT_FALSE(message.empty()); |
| EXPECT_EQ(current_session_id_, session_id); |
| EXPECT_EQ(CdmMessageType::LICENSE_REQUEST, message_type); |
| |
| // Extract the key ID from |message|. For Clear Key this is a JSON object |
| // containing a set of "kids". There should only be 1 key ID in |message|. |
| std::string message_string(message.begin(), message.end()); |
| KeyIdList key_ids; |
| std::string error_message; |
| EXPECT_TRUE(ExtractKeyIdsFromKeyIdsInitData(message_string, &key_ids, |
| &error_message)) |
| << error_message; |
| EXPECT_EQ(1u, key_ids.size()); |
| |
| // Determine the key that matches the key ID |key_ids[0]|. |
| std::vector<uint8_t> key; |
| EXPECT_TRUE(LookupKey(key_ids[0], &key)); |
| |
| // Update the session with the key ID and key. |
| std::string jwk = GenerateJWKSet(key.data(), key.size(), key_ids[0].data(), |
| key_ids[0].size()); |
| decryptor->UpdateSession(session_id, |
| std::vector<uint8_t>(jwk.begin(), jwk.end()), |
| CreatePromise(RESOLVED)); |
| } |
| |
| void OnSessionClosed(const std::string& session_id) override { |
| EXPECT_EQ(current_session_id_, session_id); |
| } |
| |
| void OnSessionKeysChange(const std::string& session_id, |
| bool has_additional_usable_key, |
| CdmKeysInfo keys_info) override { |
| EXPECT_EQ(current_session_id_, session_id); |
| EXPECT_EQ(has_additional_usable_key, true); |
| } |
| |
| void OnSessionExpirationUpdate(const std::string& session_id, |
| base::Time new_expiry_time) override { |
| EXPECT_EQ(current_session_id_, session_id); |
| } |
| |
| void OnEncryptedMediaInitData(EmeInitDataType init_data_type, |
| const std::vector<uint8_t>& init_data, |
| AesDecryptor* decryptor) override { |
| // Since only 1 session is created, skip the request if the |init_data| |
| // has been seen before (no need to add the same key again). |
| if (init_data == prev_init_data_) |
| return; |
| prev_init_data_ = init_data; |
| |
| if (current_session_id_.empty()) { |
| decryptor->CreateSessionAndGenerateRequest( |
| CdmSessionType::TEMPORARY_SESSION, init_data_type, init_data, |
| CreateSessionPromise(RESOLVED)); |
| EXPECT_FALSE(current_session_id_.empty()); |
| } |
| } |
| |
| virtual bool LookupKey(const std::vector<uint8_t>& key_id, |
| std::vector<uint8_t>* key) { |
| // No key rotation. |
| return LookupTestKeyVector(key_id, false, key); |
| } |
| |
| std::string current_session_id_; |
| std::vector<uint8_t> prev_init_data_; |
| }; |
| |
| class RotatingKeyProvidingApp : public KeyProvidingApp { |
| public: |
| RotatingKeyProvidingApp() : num_distinct_need_key_calls_(0) {} |
| ~RotatingKeyProvidingApp() override { |
| // Expect that OnEncryptedMediaInitData is fired multiple times with |
| // different |init_data|. |
| EXPECT_GT(num_distinct_need_key_calls_, 1u); |
| } |
| |
| void OnEncryptedMediaInitData(EmeInitDataType init_data_type, |
| const std::vector<uint8_t>& init_data, |
| AesDecryptor* decryptor) override { |
| // Skip the request if the |init_data| has been seen. |
| if (init_data == prev_init_data_) |
| return; |
| prev_init_data_ = init_data; |
| ++num_distinct_need_key_calls_; |
| |
| decryptor->CreateSessionAndGenerateRequest( |
| CdmSessionType::TEMPORARY_SESSION, init_data_type, init_data, |
| CreateSessionPromise(RESOLVED)); |
| } |
| |
| bool LookupKey(const std::vector<uint8_t>& key_id, |
| std::vector<uint8_t>* key) override { |
| // With key rotation. |
| return LookupTestKeyVector(key_id, true, key); |
| } |
| |
| uint32_t num_distinct_need_key_calls_; |
| }; |
| |
| // Ignores the encrypted event and does not perform a license request. |
| class NoResponseApp : public FakeEncryptedMedia::AppBase { |
| public: |
| void OnSessionMessage(const std::string& session_id, |
| CdmMessageType message_type, |
| const std::vector<uint8_t>& message, |
| AesDecryptor* decryptor) override { |
| EXPECT_FALSE(session_id.empty()); |
| EXPECT_FALSE(message.empty()); |
| FAIL() << "Unexpected Message"; |
| } |
| |
| void OnSessionClosed(const std::string& session_id) override { |
| EXPECT_FALSE(session_id.empty()); |
| FAIL() << "Unexpected Closed"; |
| } |
| |
| void OnSessionKeysChange(const std::string& session_id, |
| bool has_additional_usable_key, |
| CdmKeysInfo keys_info) override { |
| EXPECT_FALSE(session_id.empty()); |
| EXPECT_EQ(has_additional_usable_key, true); |
| } |
| |
| void OnSessionExpirationUpdate(const std::string& session_id, |
| base::Time new_expiry_time) override {} |
| |
| void OnEncryptedMediaInitData(EmeInitDataType init_data_type, |
| const std::vector<uint8_t>& init_data, |
| AesDecryptor* decryptor) override {} |
| }; |
| |
| // A rough simulation of GpuVideoDecoder that fails every Decode() request. This |
| // is used to test post-Initialize() fallback paths. |
| class FailingVideoDecoder : public VideoDecoder { |
| public: |
| std::string GetDisplayName() const override { return "FailingVideoDecoder"; } |
| void Initialize(const VideoDecoderConfig& config, |
| bool low_delay, |
| CdmContext* cdm_context, |
| const InitCB& init_cb, |
| const OutputCB& output_cb) override { |
| init_cb.Run(true); |
| } |
| void Decode(const scoped_refptr<DecoderBuffer>& buffer, |
| const DecodeCB& decode_cb) override { |
| base::ThreadTaskRunnerHandle::Get()->PostTask( |
| FROM_HERE, base::Bind(decode_cb, DecodeStatus::DECODE_ERROR)); |
| } |
| void Reset(const base::Closure& closure) override { closure.Run(); } |
| bool NeedsBitstreamConversion() const override { return true; } |
| }; |
| |
| // TODO(xhwang): These tests have been disabled for some time as apptests and no |
| // longer pass. They need to be reconstituted as shell tests. |
| // Currently there are compile issues which must be resolved, |
| // preferably by eliminating multiple inheritance here which is |
| // banned by Google C++ style. |
| #if defined(MOJO_RENDERER) && defined(ENABLE_MOJO_PIPELINE_INTEGRATION_TEST) |
| class PipelineIntegrationTest : public service_manager::test::ServiceTest, |
| public PipelineIntegrationTestBase { |
| public: |
| PipelineIntegrationTest() |
| : service_manager::test::ServiceTest( |
| "media_pipeline_integration_shelltests") {} |
| |
| void SetUp() override { |
| ServiceTest::SetUp(); |
| InitializeMediaLibrary(); |
| } |
| |
| protected: |
| std::unique_ptr<Renderer> CreateRenderer( |
| CreateVideoDecodersCB prepend_video_decoders_cb, |
| CreateAudioDecodersCB prepend_audio_decoders_cb) override { |
| connector()->BindInterface("media", &media_interface_factory_); |
| |
| mojom::RendererPtr mojo_renderer; |
| media_interface_factory_->CreateRenderer(std::string(), |
| mojo::MakeRequest(&mojo_renderer)); |
| |
| return base::MakeUnique<MojoRenderer>(message_loop_.task_runner(), |
| std::move(mojo_renderer)); |
| } |
| |
| private: |
| mojom::InterfaceFactoryPtr media_interface_factory_; |
| }; |
| #else |
| class PipelineIntegrationTest : public testing::Test, |
| public PipelineIntegrationTestBase { |
| public: |
| // Verifies that seeking works properly for ChunkDemuxer when the |
| // seek happens while there is a pending read on the ChunkDemuxer |
| // and no data is available. |
| bool TestSeekDuringRead(const std::string& filename, |
| const std::string& mimetype, |
| int initial_append_size, |
| base::TimeDelta start_seek_time, |
| base::TimeDelta seek_time, |
| int seek_file_position, |
| int seek_append_size) { |
| MockMediaSource source(filename, mimetype, initial_append_size); |
| |
| if (StartPipelineWithMediaSource(&source, kNoClockless, nullptr) != |
| PIPELINE_OK) { |
| return false; |
| } |
| |
| Play(); |
| if (!WaitUntilCurrentTimeIsAfter(start_seek_time)) |
| return false; |
| |
| source.Seek(seek_time, seek_file_position, seek_append_size); |
| if (!Seek(seek_time)) |
| return false; |
| |
| source.EndOfStream(); |
| |
| source.Shutdown(); |
| Stop(); |
| return true; |
| } |
| }; |
| #endif // defined(MOJO_RENDERER) |
| |
| struct PlaybackTestData { |
| const std::string filename; |
| const uint32_t start_time_ms; |
| const uint32_t duration_ms; |
| }; |
| |
| struct MSEPlaybackTestData { |
| const std::string filename; |
| const std::string mimetype; |
| const size_t append_bytes; |
| const uint32_t duration_ms; |
| }; |
| |
| // Tells gtest how to print our PlaybackTestData structure. |
| std::ostream& operator<<(std::ostream& os, const PlaybackTestData& data) { |
| return os << data.filename; |
| } |
| |
| std::ostream& operator<<(std::ostream& os, const MSEPlaybackTestData& data) { |
| return os << data.filename; |
| } |
| |
| class BasicPlaybackTest : public PipelineIntegrationTest, |
| public testing::WithParamInterface<PlaybackTestData> { |
| }; |
| |
| class BasicMSEPlaybackTest |
| : public ::testing::WithParamInterface<MSEPlaybackTestData>, |
| public PipelineIntegrationTest { |
| protected: |
| void PlayToEnd() { |
| MSEPlaybackTestData data = GetParam(); |
| |
| MockMediaSource source(data.filename, data.mimetype, data.append_bytes); |
| ASSERT_EQ(PIPELINE_OK, |
| StartPipelineWithMediaSource(&source, kNormal, nullptr)); |
| source.EndOfStream(); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(data.duration_ms, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_TRUE(demuxer_->GetTimelineOffset().is_null()); |
| source.Shutdown(); |
| Stop(); |
| } |
| }; |
| |
| TEST_P(BasicPlaybackTest, PlayToEnd) { |
| PlaybackTestData data = GetParam(); |
| |
| ASSERT_EQ(PIPELINE_OK, Start(data.filename, kUnreliableDuration)); |
| EXPECT_EQ(data.start_time_ms, demuxer_->GetStartTime().InMilliseconds()); |
| EXPECT_EQ(data.duration_ms, pipeline_->GetMediaDuration().InMilliseconds()); |
| |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| TEST_P(BasicMSEPlaybackTest, LegacyByDts_PlayToEnd) { |
| base::test::ScopedFeatureList scoped_feature_list; |
| scoped_feature_list.InitAndDisableFeature(media::kMseBufferByPts); |
| PlayToEnd(); |
| } |
| |
| TEST_P(BasicMSEPlaybackTest, NewByPts_PlayToEnd) { |
| base::test::ScopedFeatureList scoped_feature_list; |
| scoped_feature_list.InitAndEnableFeature(media::kMseBufferByPts); |
| PlayToEnd(); |
| } |
| |
| #if BUILDFLAG(USE_PROPRIETARY_CODECS) |
| |
| const PlaybackTestData kADTSTests[] = { |
| {"bear-audio-main-aac.aac", 0, 2724}, |
| {"bear-audio-lc-aac.aac", 0, 2717}, |
| {"bear-audio-implicit-he-aac-v1.aac", 0, 2812}, |
| {"bear-audio-implicit-he-aac-v2.aac", 0, 3047}, |
| }; |
| |
| // TODO(chcunningham): Migrate other basic playback tests to TEST_P. |
| INSTANTIATE_TEST_CASE_P(ProprietaryCodecs, |
| BasicPlaybackTest, |
| testing::ValuesIn(kADTSTests)); |
| |
| const MSEPlaybackTestData kMediaSourceADTSTests[] = { |
| {"bear-audio-main-aac.aac", kADTS, kAppendWholeFile, 2773}, |
| {"bear-audio-lc-aac.aac", kADTS, kAppendWholeFile, 2794}, |
| {"bear-audio-implicit-he-aac-v1.aac", kADTS, kAppendWholeFile, 2858}, |
| {"bear-audio-implicit-he-aac-v2.aac", kADTS, kAppendWholeFile, 2901}, |
| }; |
| |
| // TODO(chcunningham): Migrate other basic MSE playback tests to TEST_P. |
| INSTANTIATE_TEST_CASE_P(ProprietaryCodecs, |
| BasicMSEPlaybackTest, |
| testing::ValuesIn(kMediaSourceADTSTests)); |
| |
| #endif // BUILDFLAG(USE_PROPRIETARY_CODECS) |
| |
| // Test parameter determines if media::kMseBufferByPts feature should be forced |
| // on or off for the test. |
| // Note, the BasicMSEPlaybackTest test fixture defines its own parameter type, |
| // so we need this sibling fixture type to test other MSE/EME cases in this file |
| // that use a different parameter type. |
| class MSEPipelineIntegrationTest |
| : public ::testing::WithParamInterface<BufferingApi>, |
| public PipelineIntegrationTest { |
| public: |
| MSEPipelineIntegrationTest() { |
| buffering_api_ = GetParam(); |
| switch (buffering_api_) { |
| case BufferingApi::kLegacyByDts: |
| scoped_feature_list_.InitAndDisableFeature(media::kMseBufferByPts); |
| break; |
| case BufferingApi::kNewByPts: |
| scoped_feature_list_.InitAndEnableFeature(media::kMseBufferByPts); |
| break; |
| } |
| } |
| |
| protected: |
| BufferingApi buffering_api_; |
| base::test::ScopedFeatureList scoped_feature_list_; |
| }; |
| |
| TEST_F(PipelineIntegrationTest, BasicPlayback) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240.webm")); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, BasicPlaybackOpusOgg) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-opus.ogg")); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, BasicPlaybackOpusOgg_4ch_ChannelMapping2) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-opus-4ch-channelmapping2.ogg", kWebAudio)); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, BasicPlaybackOpusOgg_11ch_ChannelMapping2) { |
| ASSERT_EQ(PIPELINE_OK, |
| Start("bear-opus-11ch-channelmapping2.ogg", kWebAudio)); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, BasicPlaybackHashed) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240.webm", kHashed)); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_HASH_EQ("f0be120a90a811506777c99a2cdf7cc1", GetVideoHash()); |
| EXPECT_HASH_EQ("-3.59,-2.06,-0.43,2.15,0.77,-0.95,", GetAudioHash()); |
| EXPECT_TRUE(demuxer_->GetTimelineOffset().is_null()); |
| } |
| |
| base::TimeDelta TimestampMs(int milliseconds) { |
| return base::TimeDelta::FromMilliseconds(milliseconds); |
| } |
| |
| TEST_F(PipelineIntegrationTest, WaveLayoutChange) { |
| ASSERT_EQ(PIPELINE_OK, Start("layout_change.wav")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, PlaybackTooManyChannels) { |
| EXPECT_EQ(PIPELINE_ERROR_INITIALIZATION_FAILED, Start("9ch.wav")); |
| } |
| |
| TEST_F(PipelineIntegrationTest, PlaybackWithAudioTrackDisabledThenEnabled) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240.webm", kHashed | kNoClockless)); |
| |
| // Disable audio. |
| std::vector<MediaTrack::Id> empty; |
| pipeline_->OnEnabledAudioTracksChanged(empty); |
| scoped_task_environment_.RunUntilIdle(); |
| |
| // Seek to flush the pipeline and ensure there's no prerolled audio data. |
| ASSERT_TRUE(Seek(base::TimeDelta())); |
| |
| Play(); |
| const base::TimeDelta k500ms = TimestampMs(500); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(k500ms)); |
| Pause(); |
| |
| // Verify that no audio has been played, since we disabled audio tracks. |
| EXPECT_HASH_EQ(kNullAudioHash, GetAudioHash()); |
| |
| // Re-enable audio. |
| std::vector<MediaTrack::Id> audio_track_id; |
| audio_track_id.push_back("2"); |
| pipeline_->OnEnabledAudioTracksChanged(audio_track_id); |
| scoped_task_environment_.RunUntilIdle(); |
| |
| // Restart playback from 500ms position. |
| ASSERT_TRUE(Seek(k500ms)); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| |
| // Verify that audio has been playing after being enabled. |
| EXPECT_HASH_EQ("-1.53,0.21,1.23,1.56,-0.34,-0.94,", GetAudioHash()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, PlaybackWithVideoTrackDisabledThenEnabled) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240.webm", kHashed | kNoClockless)); |
| |
| // Disable video. |
| pipeline_->OnSelectedVideoTrackChanged(base::nullopt); |
| scoped_task_environment_.RunUntilIdle(); |
| |
| // Seek to flush the pipeline and ensure there's no prerolled video data. |
| ASSERT_TRUE(Seek(base::TimeDelta())); |
| |
| // Reset the video hash in case some of the prerolled video frames have been |
| // hashed already. |
| ResetVideoHash(); |
| |
| Play(); |
| const base::TimeDelta k500ms = TimestampMs(500); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(k500ms)); |
| Pause(); |
| |
| // Verify that no video has been rendered, since we disabled video tracks. |
| EXPECT_HASH_EQ(kNullVideoHash, GetVideoHash()); |
| |
| // Re-enable video. |
| pipeline_->OnSelectedVideoTrackChanged(MediaTrack::Id("1")); |
| scoped_task_environment_.RunUntilIdle(); |
| |
| // Seek to flush video pipeline and reset the video hash again to clear state |
| // if some prerolled frames got hashed after enabling video. |
| ASSERT_TRUE(Seek(base::TimeDelta())); |
| ResetVideoHash(); |
| |
| // Restart playback from 500ms position. |
| ASSERT_TRUE(Seek(k500ms)); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| |
| // Verify that video has been rendered after being enabled. |
| EXPECT_HASH_EQ("fd59357dfd9c144ab4fb8181b2de32c3", GetVideoHash()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, TrackStatusChangesBeforePipelineStarted) { |
| std::vector<MediaTrack::Id> empty_track_ids; |
| pipeline_->OnEnabledAudioTracksChanged(empty_track_ids); |
| pipeline_->OnSelectedVideoTrackChanged(base::nullopt); |
| } |
| |
| TEST_F(PipelineIntegrationTest, TrackStatusChangesAfterPipelineEnded) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240.webm", kHashed)); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| std::vector<MediaTrack::Id> track_ids; |
| // Disable audio track. |
| pipeline_->OnEnabledAudioTracksChanged(track_ids); |
| // Re-enable audio track. |
| track_ids.push_back("2"); |
| pipeline_->OnEnabledAudioTracksChanged(track_ids); |
| // Disable video track. |
| pipeline_->OnSelectedVideoTrackChanged(base::nullopt); |
| // Re-enable video track. |
| pipeline_->OnSelectedVideoTrackChanged(MediaTrack::Id("1")); |
| } |
| |
| TEST_F(PipelineIntegrationTest, TrackStatusChangesWhileSuspended) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240.webm", kHashed)); |
| Play(); |
| |
| ASSERT_TRUE(Suspend()); |
| |
| // These get triggered every time playback is resumed. |
| EXPECT_CALL(*this, OnVideoNaturalSizeChange(gfx::Size(320, 240))) |
| .Times(AnyNumber()); |
| EXPECT_CALL(*this, OnVideoOpacityChange(true)).Times(AnyNumber()); |
| |
| std::vector<MediaTrack::Id> track_ids; |
| |
| // Disable audio track. |
| pipeline_->OnEnabledAudioTracksChanged(track_ids); |
| ASSERT_TRUE(Resume(TimestampMs(100))); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(TimestampMs(200))); |
| ASSERT_TRUE(Suspend()); |
| |
| // Re-enable audio track. |
| track_ids.push_back("2"); |
| pipeline_->OnEnabledAudioTracksChanged(track_ids); |
| ASSERT_TRUE(Resume(TimestampMs(200))); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(TimestampMs(300))); |
| ASSERT_TRUE(Suspend()); |
| |
| // Disable video track. |
| pipeline_->OnSelectedVideoTrackChanged(base::nullopt); |
| ASSERT_TRUE(Resume(TimestampMs(300))); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(TimestampMs(400))); |
| ASSERT_TRUE(Suspend()); |
| |
| // Re-enable video track. |
| pipeline_->OnSelectedVideoTrackChanged(MediaTrack::Id("1")); |
| ASSERT_TRUE(Resume(TimestampMs(400))); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, ReinitRenderersWhileAudioTrackIsDisabled) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240.webm", kNoClockless)); |
| Play(); |
| |
| // These get triggered every time playback is resumed. |
| EXPECT_CALL(*this, OnVideoNaturalSizeChange(gfx::Size(320, 240))) |
| .Times(AnyNumber()); |
| EXPECT_CALL(*this, OnVideoOpacityChange(true)).Times(AnyNumber()); |
| |
| // Disable the audio track. |
| std::vector<MediaTrack::Id> track_ids; |
| pipeline_->OnEnabledAudioTracksChanged(track_ids); |
| // pipeline.Suspend() releases renderers and pipeline.Resume() recreates and |
| // reinitializes renderers while the audio track is disabled. |
| ASSERT_TRUE(Suspend()); |
| ASSERT_TRUE(Resume(TimestampMs(100))); |
| // Now re-enable the audio track, playback should continue successfully. |
| EXPECT_CALL(*this, OnBufferingStateChange(BUFFERING_HAVE_ENOUGH)).Times(1); |
| track_ids.push_back("2"); |
| pipeline_->OnEnabledAudioTracksChanged(track_ids); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(TimestampMs(200))); |
| |
| Stop(); |
| } |
| |
| TEST_F(PipelineIntegrationTest, ReinitRenderersWhileVideoTrackIsDisabled) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240.webm", kHashed | kNoClockless)); |
| Play(); |
| |
| // These get triggered every time playback is resumed. |
| EXPECT_CALL(*this, OnVideoNaturalSizeChange(gfx::Size(320, 240))) |
| .Times(AnyNumber()); |
| EXPECT_CALL(*this, OnVideoOpacityChange(true)).Times(AnyNumber()); |
| |
| // Disable the video track. |
| pipeline_->OnSelectedVideoTrackChanged(base::nullopt); |
| // pipeline.Suspend() releases renderers and pipeline.Resume() recreates and |
| // reinitializes renderers while the video track is disabled. |
| ASSERT_TRUE(Suspend()); |
| ASSERT_TRUE(Resume(TimestampMs(100))); |
| // Now re-enable the video track, playback should continue successfully. |
| pipeline_->OnSelectedVideoTrackChanged(MediaTrack::Id("1")); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(TimestampMs(200))); |
| |
| Stop(); |
| } |
| |
| TEST_F(PipelineIntegrationTest, PipelineStoppedWhileAudioRestartPending) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240.webm")); |
| Play(); |
| |
| // Disable audio track first, to re-enable it later and stop the pipeline |
| // (which destroys the media renderer) while audio restart is pending. |
| std::vector<MediaTrack::Id> track_ids; |
| pipeline_->OnEnabledAudioTracksChanged(track_ids); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(TimestampMs(200))); |
| |
| track_ids.push_back("2"); |
| pipeline_->OnEnabledAudioTracksChanged(track_ids); |
| Stop(); |
| } |
| |
| TEST_F(PipelineIntegrationTest, PipelineStoppedWhileVideoRestartPending) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240.webm")); |
| Play(); |
| |
| // Disable video track first, to re-enable it later and stop the pipeline |
| // (which destroys the media renderer) while video restart is pending. |
| pipeline_->OnSelectedVideoTrackChanged(base::nullopt); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(TimestampMs(200))); |
| |
| pipeline_->OnSelectedVideoTrackChanged(MediaTrack::Id("1")); |
| Stop(); |
| } |
| |
| TEST_F(PipelineIntegrationTest, SwitchAudioTrackDuringPlayback) { |
| ASSERT_EQ(PIPELINE_OK, |
| Start("multitrack-3video-2audio.webm", kHashed | kNoClockless)); |
| Play(); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(TimestampMs(100))); |
| // The first audio track (TrackId=4) is enabled by default. This should |
| // disable TrackId=4 and enable TrackId=5. |
| std::vector<MediaTrack::Id> track_ids; |
| track_ids.push_back("5"); |
| pipeline_->OnEnabledAudioTracksChanged(track_ids); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(TimestampMs(200))); |
| Stop(); |
| } |
| |
| TEST_F(PipelineIntegrationTest, SwitchVideoTrackDuringPlayback) { |
| ASSERT_EQ(PIPELINE_OK, |
| Start("multitrack-3video-2audio.webm", kHashed | kNoClockless)); |
| Play(); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(TimestampMs(100))); |
| // The first video track (TrackId=1) is enabled by default. This should |
| // disable TrackId=1 and enable TrackId=2. |
| pipeline_->OnSelectedVideoTrackChanged(MediaTrack::Id("2")); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(TimestampMs(200))); |
| Stop(); |
| } |
| |
| TEST_F(PipelineIntegrationTest, BasicPlaybackOpusOggTrimmingHashed) { |
| ASSERT_EQ(PIPELINE_OK, Start("opus-trimming-test.webm", kHashed)); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(kOpusEndTrimmingHash_1, GetAudioHash()); |
| |
| // Seek within the pre-skip section, this should not cause a beep. |
| ASSERT_TRUE(Seek(base::TimeDelta::FromSeconds(1))); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(kOpusEndTrimmingHash_2, GetAudioHash()); |
| |
| // Seek somewhere outside of the pre-skip / end-trim section, demxuer should |
| // correctly preroll enough to accurately decode this segment. |
| ASSERT_TRUE(Seek(base::TimeDelta::FromMilliseconds(6360))); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(kOpusEndTrimmingHash_3, GetAudioHash()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, BasicPlaybackOpusWebmTrimmingHashed) { |
| ASSERT_EQ(PIPELINE_OK, Start("opus-trimming-test.webm", kHashed)); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(kOpusEndTrimmingHash_1, GetAudioHash()); |
| |
| // Seek within the pre-skip section, this should not cause a beep. |
| ASSERT_TRUE(Seek(base::TimeDelta::FromSeconds(1))); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(kOpusEndTrimmingHash_2, GetAudioHash()); |
| |
| // Seek somewhere outside of the pre-skip / end-trim section, demxuer should |
| // correctly preroll enough to accurately decode this segment. |
| ASSERT_TRUE(Seek(base::TimeDelta::FromMilliseconds(6360))); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(kOpusEndTrimmingHash_3, GetAudioHash()); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, BasicPlaybackOpusWebmTrimmingHashed) { |
| MockMediaSource source("opus-trimming-test.webm", kOpusAudioOnlyWebM, |
| kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithMediaSource(&source, kHashed, nullptr)); |
| source.EndOfStream(); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(kOpusEndTrimmingHash_1, GetAudioHash()); |
| |
| // Seek within the pre-skip section, this should not cause a beep. |
| base::TimeDelta seek_time = base::TimeDelta::FromSeconds(1); |
| source.Seek(seek_time); |
| ASSERT_TRUE(Seek(seek_time)); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(kOpusEndTrimmingHash_2, GetAudioHash()); |
| |
| // Seek somewhere outside of the pre-skip / end-trim section, demuxer should |
| // correctly preroll enough to accurately decode this segment. |
| seek_time = base::TimeDelta::FromMilliseconds(6360); |
| source.Seek(seek_time); |
| ASSERT_TRUE(Seek(seek_time)); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(kOpusEndTrimmingHash_3, GetAudioHash()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, BasicPlaybackOpusPrerollExceedsCodecDelay) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-opus.webm", kHashed)); |
| |
| AudioDecoderConfig config = |
| demuxer_->GetFirstStream(DemuxerStream::AUDIO)->audio_decoder_config(); |
| |
| // Verify that this file's preroll is not eclipsed by the codec delay so we |
| // can detect when preroll is not properly performed. |
| base::TimeDelta codec_delay = base::TimeDelta::FromSecondsD( |
| static_cast<double>(config.codec_delay()) / config.samples_per_second()); |
| ASSERT_GT(config.seek_preroll(), codec_delay); |
| |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(kOpusSmallCodecDelayHash_1, GetAudioHash()); |
| |
| // Seek halfway through the file to invoke seek preroll. |
| ASSERT_TRUE(Seek(base::TimeDelta::FromSecondsD(1.414))); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(kOpusSmallCodecDelayHash_2, GetAudioHash()); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, BasicPlaybackOpusPrerollExceedsCodecDelay) { |
| MockMediaSource source("bear-opus.webm", kOpusAudioOnlyWebM, |
| kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithMediaSource(&source, kHashed, nullptr)); |
| source.EndOfStream(); |
| |
| AudioDecoderConfig config = |
| demuxer_->GetFirstStream(DemuxerStream::AUDIO)->audio_decoder_config(); |
| |
| // Verify that this file's preroll is not eclipsed by the codec delay so we |
| // can detect when preroll is not properly performed. |
| base::TimeDelta codec_delay = base::TimeDelta::FromSecondsD( |
| static_cast<double>(config.codec_delay()) / config.samples_per_second()); |
| ASSERT_GT(config.seek_preroll(), codec_delay); |
| |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(kOpusSmallCodecDelayHash_1, GetAudioHash()); |
| |
| // Seek halfway through the file to invoke seek preroll. |
| base::TimeDelta seek_time = base::TimeDelta::FromSecondsD(1.414); |
| source.Seek(seek_time); |
| ASSERT_TRUE(Seek(seek_time)); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(kOpusSmallCodecDelayHash_2, GetAudioHash()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, BasicPlaybackLive) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240-live.webm", kHashed)); |
| |
| // Live stream does not have duration in the initialization segment. |
| // It will be set after the entire file is available. |
| EXPECT_CALL(*this, OnDurationChange()).Times(1); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_HASH_EQ("f0be120a90a811506777c99a2cdf7cc1", GetVideoHash()); |
| EXPECT_HASH_EQ("-3.59,-2.06,-0.43,2.15,0.77,-0.95,", GetAudioHash()); |
| EXPECT_EQ(kLiveTimelineOffset(), demuxer_->GetTimelineOffset()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, S32PlaybackHashed) { |
| ASSERT_EQ(PIPELINE_OK, Start("sfx_s32le.wav", kHashed)); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(std::string(kNullVideoHash), GetVideoHash()); |
| EXPECT_HASH_EQ(kSfxLosslessHash, GetAudioHash()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, F32PlaybackHashed) { |
| ASSERT_EQ(PIPELINE_OK, Start("sfx_f32le.wav", kHashed)); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(std::string(kNullVideoHash), GetVideoHash()); |
| EXPECT_HASH_EQ(kSfxLosslessHash, GetAudioHash()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, MAYBE_EME(BasicPlaybackEncrypted)) { |
| FakeEncryptedMedia encrypted_media(new KeyProvidingApp()); |
| set_encrypted_media_init_data_cb( |
| base::Bind(&FakeEncryptedMedia::OnEncryptedMediaInitData, |
| base::Unretained(&encrypted_media))); |
| |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240-av_enc-av.webm", |
| encrypted_media.GetCdmContext())); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| Stop(); |
| } |
| |
| TEST_F(PipelineIntegrationTest, FlacPlaybackHashed) { |
| ASSERT_EQ(PIPELINE_OK, Start("sfx.flac", kHashed)); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(std::string(kNullVideoHash), GetVideoHash()); |
| EXPECT_HASH_EQ(kSfxLosslessHash, GetAudioHash()); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, BasicPlayback) { |
| MockMediaSource source("bear-320x240.webm", kWebM, 219229); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| source.EndOfStream(); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(k320WebMFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_TRUE(demuxer_->GetTimelineOffset().is_null()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, Eos_Before_Demuxer_Opened) { |
| // After appending only a partial initialization segment, marking end of |
| // stream should let the test complete with error indicating failure to open |
| // demuxer. Here we append only the first 10 bytes of a test WebM, definitely |
| // less than the ~4400 bytes needed to parse its full initialization segment. |
| MockMediaSource source("bear-320x240.webm", kWebM, 10); |
| source.set_do_eos_after_next_append(true); |
| EXPECT_EQ( |
| DEMUXER_ERROR_COULD_NOT_OPEN, |
| StartPipelineWithMediaSource(&source, kExpectDemuxerFailure, nullptr)); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, Corrupted_First_Media_Segment) { |
| // After successful initialization segment append completing demuxer opening, |
| // immediately append a corrupted media segment to trigger parse error while |
| // pipeline is still completing renderer setup. |
| MockMediaSource source("bear-320x240_corrupted_after_init_segment.webm", |
| kWebM, 4380); |
| source.set_expect_append_success(false); |
| EXPECT_EQ(CHUNK_DEMUXER_ERROR_APPEND_FAILED, |
| StartPipelineWithMediaSource(&source)); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, BasicPlayback_Live) { |
| MockMediaSource source("bear-320x240-live.webm", kWebM, 219221); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| source.EndOfStream(); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(k320WebMFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_EQ(kLiveTimelineOffset(), demuxer_->GetTimelineOffset()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| #if BUILDFLAG(ENABLE_AV1_DECODER) |
| TEST_P(MSEPipelineIntegrationTest, BasicPlayback_AV1_WebM) { |
| base::test::ScopedFeatureList scoped_feature_list_; |
| scoped_feature_list_.InitAndEnableFeature(kAv1Decoder); |
| |
| MockMediaSource source("bear-av1.webm", kWebMAV1, 79943); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| source.EndOfStream(); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(kVP9WebMFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| #endif |
| |
| TEST_P(MSEPipelineIntegrationTest, BasicPlayback_VP9_WebM) { |
| MockMediaSource source("bear-vp9.webm", kWebMVP9, 67504); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| source.EndOfStream(); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(kVP9WebMFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, BasicPlayback_VP9_BlockGroup_WebM) { |
| MockMediaSource source("bear-vp9-blockgroup.webm", kWebMVP9, 67871); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| source.EndOfStream(); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(kVP9WebMFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, BasicPlayback_VP8A_WebM) { |
| MockMediaSource source("bear-vp8a.webm", kVideoOnlyWebM, kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| source.EndOfStream(); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(kVP8AWebMFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, BasicPlayback_Opus_WebM) { |
| MockMediaSource source("bear-opus-end-trimming.webm", kOpusAudioOnlyWebM, |
| kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| source.EndOfStream(); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(kOpusEndTrimmingWebMFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| // Flaky. http://crbug.com/304776 |
| TEST_P(MSEPipelineIntegrationTest, DISABLED_Opus_Seeking_WebM) { |
| MockMediaSource source("bear-opus-end-trimming.webm", kOpusAudioOnlyWebM, |
| kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithMediaSource(&source, kHashed, nullptr)); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(kOpusEndTrimmingWebMFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| base::TimeDelta start_seek_time = base::TimeDelta::FromMilliseconds(1000); |
| base::TimeDelta seek_time = base::TimeDelta::FromMilliseconds(2000); |
| |
| Play(); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(start_seek_time)); |
| source.Seek(seek_time, 0x1D5, 34017); |
| source.EndOfStream(); |
| ASSERT_TRUE(Seek(seek_time)); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_HASH_EQ("0.76,0.20,-0.82,-0.58,-1.29,-0.29,", GetAudioHash()); |
| |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, ConfigChange_WebM) { |
| MockMediaSource source("bear-320x240-16x9-aspect.webm", kWebM, |
| kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| |
| const gfx::Size kNewSize(640, 360); |
| EXPECT_CALL(*this, OnVideoConfigChange(::testing::Property( |
| &VideoDecoderConfig::natural_size, kNewSize))) |
| .Times(1); |
| EXPECT_CALL(*this, OnVideoNaturalSizeChange(kNewSize)).Times(1); |
| scoped_refptr<DecoderBuffer> second_file = |
| ReadTestDataFile("bear-640x360.webm"); |
| source.AppendAtTime(base::TimeDelta::FromSeconds(kAppendTimeSec), |
| second_file->data(), second_file->data_size()); |
| source.EndOfStream(); |
| |
| Play(); |
| EXPECT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(kAppendTimeMs + k640WebMFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, AudioConfigChange_WebM) { |
| MockMediaSource source("bear-320x240-audio-only.webm", kAudioOnlyWebM, |
| kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| |
| const int kNewSampleRate = 48000; |
| EXPECT_CALL(*this, |
| OnAudioConfigChange(::testing::Property( |
| &AudioDecoderConfig::samples_per_second, kNewSampleRate))) |
| .Times(1); |
| |
| // A higher sample rate will cause the audio buffer durations to change. This |
| // should not manifest as a timestamp gap in AudioTimestampValidator. |
| // Timestamp expectations should be reset across config changes. |
| EXPECT_MEDIA_LOG(Not(HasSubstr("Large timestamp gap detected"))) |
| .Times(AnyNumber()); |
| |
| scoped_refptr<DecoderBuffer> second_file = |
| ReadTestDataFile("bear-320x240-audio-only-48khz.webm"); |
| ASSERT_TRUE(source.AppendAtTime(base::TimeDelta::FromSeconds(kAppendTimeSec), |
| second_file->data(), |
| second_file->data_size())); |
| source.EndOfStream(); |
| |
| Play(); |
| EXPECT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(3773, pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, Remove_Updates_BufferedRanges) { |
| const char* input_filename = "bear-320x240.webm"; |
| MockMediaSource source(input_filename, kWebM, kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| |
| auto buffered_ranges = pipeline_->GetBufferedTimeRanges(); |
| EXPECT_EQ(1u, buffered_ranges.size()); |
| EXPECT_EQ(0, buffered_ranges.start(0).InMilliseconds()); |
| EXPECT_EQ(k320WebMFileDurationMs, buffered_ranges.end(0).InMilliseconds()); |
| |
| source.RemoveRange(base::TimeDelta::FromMilliseconds(1000), |
| base::TimeDelta::FromMilliseconds(k320WebMFileDurationMs)); |
| scoped_task_environment_.RunUntilIdle(); |
| |
| buffered_ranges = pipeline_->GetBufferedTimeRanges(); |
| EXPECT_EQ(1u, buffered_ranges.size()); |
| EXPECT_EQ(0, buffered_ranges.start(0).InMilliseconds()); |
| EXPECT_EQ(1001, buffered_ranges.end(0).InMilliseconds()); |
| |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| // This test case imitates media playback with advancing media_time and |
| // continuously adding new data. At some point we should reach the buffering |
| // limit, after that MediaSource should evict some buffered data and that |
| // evicted data shold be reflected in the change of media::Pipeline buffered |
| // ranges (returned by GetBufferedTimeRanges). At that point the buffered ranges |
| // will no longer start at 0. |
| TEST_P(MSEPipelineIntegrationTest, FillUp_Buffer) { |
| const char* input_filename = "bear-320x240.webm"; |
| MockMediaSource source(input_filename, kWebM, kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| source.SetMemoryLimits(1048576); |
| |
| scoped_refptr<DecoderBuffer> file = ReadTestDataFile(input_filename); |
| |
| auto buffered_ranges = pipeline_->GetBufferedTimeRanges(); |
| EXPECT_EQ(1u, buffered_ranges.size()); |
| do { |
| // Advance media_time to the end of the currently buffered data |
| base::TimeDelta media_time = buffered_ranges.end(0); |
| source.Seek(media_time); |
| // Ask MediaSource to evict buffered data if buffering limit has been |
| // reached (the data will be evicted from the front of the buffered range). |
| source.EvictCodedFrames(media_time, file->data_size()); |
| source.AppendAtTime(media_time, file->data(), file->data_size()); |
| scoped_task_environment_.RunUntilIdle(); |
| |
| buffered_ranges = pipeline_->GetBufferedTimeRanges(); |
| } while (buffered_ranges.size() == 1 && |
| buffered_ranges.start(0) == base::TimeDelta::FromSeconds(0)); |
| |
| EXPECT_EQ(1u, buffered_ranges.size()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, GCWithDisabledVideoStream) { |
| const char* input_filename = "bear-320x240.webm"; |
| MockMediaSource source(input_filename, kWebM, kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| scoped_refptr<DecoderBuffer> file = ReadTestDataFile(input_filename); |
| // The input file contains audio + video data. Assuming video data size is |
| // larger than audio, so setting memory limits to half of file data_size will |
| // ensure that video SourceBuffer is above memory limit and the audio |
| // SourceBuffer is below the memory limit. |
| source.SetMemoryLimits(file->data_size() / 2); |
| |
| // Disable the video track and start playback. Renderer won't read from the |
| // disabled video stream, so the video stream read position should be 0. |
| pipeline_->OnSelectedVideoTrackChanged(base::nullopt); |
| Play(); |
| |
| // Wait until audio playback advances past 2 seconds and call MSE GC algorithm |
| // to prepare for more data to be appended. |
| base::TimeDelta media_time = base::TimeDelta::FromSeconds(2); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(media_time)); |
| // At this point the video SourceBuffer is over the memory limit (see the |
| // SetMemoryLimits comment above), but MSE GC should be able to remove some |
| // of video data and return true indicating success, even though no data has |
| // been read from the disabled video stream and its read position is 0. |
| ASSERT_TRUE(source.EvictCodedFrames(media_time, 10)); |
| |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| MAYBE_EME_TEST_P(MSEPipelineIntegrationTest, |
| MAYBE_EME(ConfigChange_Encrypted_WebM)) { |
| MockMediaSource source("bear-320x240-16x9-aspect-av_enc-av.webm", kWebM, |
| kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new KeyProvidingApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| const gfx::Size kNewSize(640, 360); |
| EXPECT_CALL(*this, OnVideoConfigChange(::testing::Property( |
| &VideoDecoderConfig::natural_size, kNewSize))) |
| .Times(1); |
| EXPECT_CALL(*this, OnVideoNaturalSizeChange(kNewSize)).Times(1); |
| scoped_refptr<DecoderBuffer> second_file = |
| ReadTestDataFile("bear-640x360-av_enc-av.webm"); |
| |
| source.AppendAtTime(base::TimeDelta::FromSeconds(kAppendTimeSec), |
| second_file->data(), second_file->data_size()); |
| source.EndOfStream(); |
| |
| Play(); |
| EXPECT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(kAppendTimeMs + k640WebMFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| MAYBE_EME_TEST_P(MSEPipelineIntegrationTest, |
| MAYBE_EME(ConfigChange_ClearThenEncrypted_WebM)) { |
| MockMediaSource source("bear-320x240-16x9-aspect.webm", kWebM, |
| kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new KeyProvidingApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| const gfx::Size kNewSize(640, 360); |
| EXPECT_CALL(*this, OnVideoConfigChange(::testing::Property( |
| &VideoDecoderConfig::natural_size, kNewSize))) |
| .Times(1); |
| EXPECT_CALL(*this, OnVideoNaturalSizeChange(kNewSize)).Times(1); |
| scoped_refptr<DecoderBuffer> second_file = |
| ReadTestDataFile("bear-640x360-av_enc-av.webm"); |
| |
| source.AppendAtTime(base::TimeDelta::FromSeconds(kAppendTimeSec), |
| second_file->data(), second_file->data_size()); |
| source.EndOfStream(); |
| |
| Play(); |
| EXPECT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(kAppendTimeMs + k640WebMFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| // Config change from encrypted to clear is allowed by the demuxer, and is |
| // supported by the Renderer. |
| MAYBE_EME_TEST_P(MSEPipelineIntegrationTest, |
| MAYBE_EME(ConfigChange_EncryptedThenClear_WebM)) { |
| MockMediaSource source("bear-320x240-16x9-aspect-av_enc-av.webm", kWebM, |
| kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new KeyProvidingApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| const gfx::Size kNewSize(640, 360); |
| EXPECT_CALL(*this, OnVideoConfigChange(::testing::Property( |
| &VideoDecoderConfig::natural_size, kNewSize))) |
| .Times(1); |
| EXPECT_CALL(*this, OnVideoNaturalSizeChange(kNewSize)).Times(1); |
| scoped_refptr<DecoderBuffer> second_file = |
| ReadTestDataFile("bear-640x360.webm"); |
| |
| source.AppendAtTime(base::TimeDelta::FromSeconds(kAppendTimeSec), |
| second_file->data(), second_file->data_size()); |
| source.EndOfStream(); |
| |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(kAppendTimeMs + k640WebMFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| #if defined(ARCH_CPU_X86_FAMILY) && !defined(OS_ANDROID) |
| TEST_F(PipelineIntegrationTest, BasicPlaybackHi10PVP9) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x180-hi10p-vp9.webm")); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, BasicPlaybackHi12PVP9) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x180-hi12p-vp9.webm")); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| #endif |
| |
| #if BUILDFLAG(ENABLE_AV1_DECODER) |
| TEST_P(MSEPipelineIntegrationTest, BasicPlayback_AV1_MP4) { |
| base::test::ScopedFeatureList scoped_feature_list_; |
| scoped_feature_list_.InitAndEnableFeature(kAv1Decoder); |
| MockMediaSource source("bear-av1.mp4", kMP4AV1, 80496); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| source.EndOfStream(); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(kVP9WebMFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| #endif |
| |
| TEST_P(MSEPipelineIntegrationTest, FlacInMp4_Hashed) { |
| // The feature is disabled by default. Enable it. |
| base::test::ScopedFeatureList features; |
| features.InitAndEnableFeature(kMseFlacInIsobmff); |
| |
| MockMediaSource source("sfx-flac_frag.mp4", kMP4AudioFlac, kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithMediaSource(&source, kHashed, nullptr)); |
| source.EndOfStream(); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(288, pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(std::string(kNullVideoHash), GetVideoHash()); |
| EXPECT_HASH_EQ(kSfxLosslessHash, GetAudioHash()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, BasicPlaybackHashed_MP3) { |
| ASSERT_EQ(PIPELINE_OK, Start("sfx.mp3", kHashed)); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| |
| // Verify codec delay and preroll are stripped. |
| EXPECT_HASH_EQ("1.30,2.72,4.56,5.08,3.74,2.03,", GetAudioHash()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, BasicPlaybackHashed_FlacInMp4) { |
| ASSERT_EQ(PIPELINE_OK, Start("sfx-flac.mp4", kHashed)); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ(std::string(kNullVideoHash), GetVideoHash()); |
| EXPECT_HASH_EQ(kSfxLosslessHash, GetAudioHash()); |
| } |
| |
| #if BUILDFLAG(ENABLE_AV1_DECODER) |
| TEST_F(PipelineIntegrationTest, BasicPlayback_VideoOnly_AV1_Mp4) { |
| base::test::ScopedFeatureList scoped_feature_list_; |
| scoped_feature_list_.InitAndEnableFeature(kAv1Decoder); |
| ASSERT_EQ(PIPELINE_OK, Start("bear-av1.mp4")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| #endif |
| |
| class Mp3FastSeekParams { |
| public: |
| Mp3FastSeekParams(const char* filename, const char* hash) |
| : filename(filename), hash(hash) {} |
| const char* filename; |
| const char* hash; |
| }; |
| |
| class Mp3FastSeekIntegrationTest |
| : public PipelineIntegrationTest, |
| public testing::WithParamInterface<Mp3FastSeekParams> {}; |
| |
| TEST_P(Mp3FastSeekIntegrationTest, FastSeekAccuracy_MP3) { |
| Mp3FastSeekParams config = GetParam(); |
| ASSERT_EQ(PIPELINE_OK, Start(config.filename, kHashed)); |
| |
| // The XING TOC is inaccurate. We don't use it for CBR, we tolerate it for VBR |
| // (best option for fast seeking; see Mp3SeekFFmpegDemuxerTest). The chosen |
| // seek time exposes inaccuracy in TOC such that the hash will change if seek |
| // logic is regressed. See https://crbug.com/545914. |
| // |
| // Quick TOC design (not pretty!): |
| // - All MP3 TOCs are 100 bytes |
| // - Each byte is read as a uint8_t; value between 0 - 255. |
| // - The index into this array is the numerator in the ratio: index / 100. |
| // This fraction represents a playback time as a percentage of duration. |
| // - The value at the given index is the numerator in the ratio: value / 256. |
| // This fraction represents a byte offset as a percentage of the file size. |
| // |
| // For CBR files, each frame is the same size, so the offset for time of |
| // (0.98 * duration) should be around (0.98 * file size). This is 250.88 / 256 |
| // but the numerator will be truncated in the TOC as 250, losing precision. |
| base::TimeDelta seek_time(0.98 * pipeline_->GetMediaDuration()); |
| |
| ASSERT_TRUE(Seek(seek_time)); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_HASH_EQ(config.hash, GetAudioHash()); |
| } |
| |
| // CBR seeks should always be fast and accurate. |
| INSTANTIATE_TEST_CASE_P( |
| CBRSeek_HasTOC, |
| Mp3FastSeekIntegrationTest, |
| ::testing::Values(Mp3FastSeekParams("bear-audio-10s-CBR-has-TOC.mp3", |
| "-0.71,0.36,2.96,2.68,2.11,-1.08,"))); |
| |
| INSTANTIATE_TEST_CASE_P( |
| CBRSeeks_NoTOC, |
| Mp3FastSeekIntegrationTest, |
| ::testing::Values(Mp3FastSeekParams("bear-audio-10s-CBR-no-TOC.mp3", |
| "0.95,0.56,1.34,0.47,1.77,0.84,"))); |
| |
| // VBR seeks can be fast *OR* accurate, but not both. We chose fast. |
| INSTANTIATE_TEST_CASE_P( |
| VBRSeeks_HasTOC, |
| Mp3FastSeekIntegrationTest, |
| ::testing::Values(Mp3FastSeekParams("bear-audio-10s-VBR-has-TOC.mp3", |
| "-0.15,-0.83,0.54,1.00,1.94,0.93,"))); |
| |
| INSTANTIATE_TEST_CASE_P( |
| VBRSeeks_NoTOC, |
| Mp3FastSeekIntegrationTest, |
| ::testing::Values(Mp3FastSeekParams("bear-audio-10s-VBR-no-TOC.mp3", |
| "-0.22,0.80,1.19,0.73,-0.31,-1.12,"))); |
| |
| TEST_P(MSEPipelineIntegrationTest, MP3) { |
| MockMediaSource source("sfx.mp3", kMP3, kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithMediaSource(&source, kHashed, nullptr)); |
| source.EndOfStream(); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(313, pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| Play(); |
| |
| EXPECT_TRUE(WaitUntilOnEnded()); |
| |
| // Verify that codec delay was stripped. |
| EXPECT_HASH_EQ("1.01,2.71,4.18,4.32,3.04,1.12,", GetAudioHash()); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, MP3_TimestampOffset) { |
| MockMediaSource source("sfx.mp3", kMP3, kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| EXPECT_EQ(313, source.last_timestamp_offset().InMilliseconds()); |
| |
| // There are 576 silent frames at the start of this mp3. The second append |
| // should trim them off. |
| const base::TimeDelta mp3_preroll_duration = |
| base::TimeDelta::FromSecondsD(576.0 / 44100); |
| const base::TimeDelta append_time = |
| source.last_timestamp_offset() - mp3_preroll_duration; |
| |
| scoped_refptr<DecoderBuffer> second_file = ReadTestDataFile("sfx.mp3"); |
| source.AppendAtTimeWithWindow(append_time, append_time + mp3_preroll_duration, |
| kInfiniteDuration, second_file->data(), |
| second_file->data_size()); |
| source.EndOfStream(); |
| |
| Play(); |
| EXPECT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_EQ(613, source.last_timestamp_offset().InMilliseconds()); |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(613, pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, MP3_Icecast) { |
| MockMediaSource source("icy_sfx.mp3", kMP3, kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| source.EndOfStream(); |
| |
| Play(); |
| |
| EXPECT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| #if BUILDFLAG(USE_PROPRIETARY_CODECS) |
| |
| TEST_P(MSEPipelineIntegrationTest, ADTS) { |
| MockMediaSource source("sfx.adts", kADTS, kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithMediaSource(&source, kHashed, nullptr)); |
| source.EndOfStream(); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(325, pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| Play(); |
| |
| EXPECT_TRUE(WaitUntilOnEnded()); |
| |
| // Verify that nothing was stripped. |
| EXPECT_HASH_EQ("0.46,1.72,4.26,4.57,3.39,1.53,", GetAudioHash()); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, ADTS_TimestampOffset) { |
| MockMediaSource source("sfx.adts", kADTS, kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithMediaSource(&source, kHashed, nullptr)); |
| EXPECT_EQ(325, source.last_timestamp_offset().InMilliseconds()); |
| |
| // Trim off multiple frames off the beginning of the segment which will cause |
| // the first decoded frame to be incorrect if preroll isn't implemented. |
| const base::TimeDelta adts_preroll_duration = |
| base::TimeDelta::FromSecondsD(2.5 * 1024 / 44100); |
| const base::TimeDelta append_time = |
| source.last_timestamp_offset() - adts_preroll_duration; |
| |
| scoped_refptr<DecoderBuffer> second_file = ReadTestDataFile("sfx.adts"); |
| source.AppendAtTimeWithWindow( |
| append_time, append_time + adts_preroll_duration, kInfiniteDuration, |
| second_file->data(), second_file->data_size()); |
| source.EndOfStream(); |
| |
| Play(); |
| EXPECT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_EQ(592, source.last_timestamp_offset().InMilliseconds()); |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(592, pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| // Verify preroll is stripped. |
| EXPECT_HASH_EQ("-1.76,-1.35,-0.72,0.70,1.24,0.52,", GetAudioHash()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, BasicPlaybackHashed_ADTS) { |
| ASSERT_EQ(PIPELINE_OK, Start("sfx.adts", kHashed)); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| |
| // Verify codec delay and preroll are stripped. |
| EXPECT_HASH_EQ("1.80,1.66,2.31,3.26,4.46,3.36,", GetAudioHash()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, BasicPlaybackHi10P) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x180-hi10p.mp4")); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| std::vector<std::unique_ptr<VideoDecoder>> CreateFailingVideoDecoder() { |
| std::vector<std::unique_ptr<VideoDecoder>> failing_video_decoder; |
| failing_video_decoder.push_back(base::MakeUnique<FailingVideoDecoder>()); |
| return failing_video_decoder; |
| } |
| |
| TEST_F(PipelineIntegrationTest, BasicFallback) { |
| ASSERT_EQ(PIPELINE_OK, |
| Start("bear.mp4", kNormal, base::Bind(&CreateFailingVideoDecoder))); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| }; |
| |
| TEST_P(MSEPipelineIntegrationTest, ConfigChange_MP4) { |
| MockMediaSource source("bear-640x360-av_frag.mp4", kMP4, kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| |
| const gfx::Size kNewSize(1280, 720); |
| EXPECT_CALL(*this, OnVideoConfigChange(::testing::Property( |
| &VideoDecoderConfig::natural_size, kNewSize))) |
| .Times(1); |
| EXPECT_CALL(*this, OnVideoNaturalSizeChange(kNewSize)).Times(1); |
| scoped_refptr<DecoderBuffer> second_file = |
| ReadTestDataFile("bear-1280x720-av_frag.mp4"); |
| source.AppendAtTime(base::TimeDelta::FromSeconds(kAppendTimeSec), |
| second_file->data(), second_file->data_size()); |
| source.EndOfStream(); |
| |
| Play(); |
| EXPECT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| if (buffering_api_ == BufferingApi::kLegacyByDts) { |
| EXPECT_EQ(kAppendTimeMs + k1280IsoFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| } else { |
| EXPECT_EQ(kAppendTimeMs + k1280IsoFileDurationMsAVByPts, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| } |
| |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| MAYBE_EME_TEST_P(MSEPipelineIntegrationTest, |
| MAYBE_EME(ConfigChange_Encrypted_MP4_CENC_VideoOnly)) { |
| MockMediaSource source("bear-640x360-v_frag-cenc.mp4", kMP4Video, |
| kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new KeyProvidingApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| const gfx::Size kNewSize(1280, 720); |
| EXPECT_CALL(*this, OnVideoConfigChange(::testing::Property( |
| &VideoDecoderConfig::natural_size, kNewSize))) |
| .Times(1); |
| EXPECT_CALL(*this, OnVideoNaturalSizeChange(kNewSize)).Times(1); |
| scoped_refptr<DecoderBuffer> second_file = |
| ReadTestDataFile("bear-1280x720-v_frag-cenc.mp4"); |
| source.AppendAtTime(base::TimeDelta::FromSeconds(kAppendTimeSec), |
| second_file->data(), second_file->data_size()); |
| source.EndOfStream(); |
| |
| Play(); |
| EXPECT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| if (buffering_api_ == BufferingApi::kLegacyByDts) |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| else |
| EXPECT_EQ(33, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(kAppendTimeMs + k1280IsoFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| MAYBE_EME_TEST_P( |
| MSEPipelineIntegrationTest, |
| MAYBE_EME(ConfigChange_Encrypted_MP4_CENC_KeyRotation_VideoOnly)) { |
| MockMediaSource source("bear-640x360-v_frag-cenc-key_rotation.mp4", kMP4Video, |
| kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new RotatingKeyProvidingApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| EXPECT_CALL(*this, OnVideoNaturalSizeChange(gfx::Size(1280, 720))).Times(1); |
| scoped_refptr<DecoderBuffer> second_file = |
| ReadTestDataFile("bear-1280x720-v_frag-cenc-key_rotation.mp4"); |
| source.AppendAtTime(base::TimeDelta::FromSeconds(kAppendTimeSec), |
| second_file->data(), second_file->data_size()); |
| source.EndOfStream(); |
| |
| Play(); |
| EXPECT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| if (buffering_api_ == BufferingApi::kLegacyByDts) |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| else |
| EXPECT_EQ(33, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(kAppendTimeMs + k1280IsoFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| // Config changes from clear to encrypted are not currently supported. |
| // TODO(ddorwin): Figure out why this CHECKs in AppendAtTime(). |
| TEST_P(MSEPipelineIntegrationTest, |
| DISABLED_ConfigChange_ClearThenEncrypted_MP4_CENC) { |
| MockMediaSource source("bear-640x360-av_frag.mp4", kMP4Video, |
| kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new KeyProvidingApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| EXPECT_CALL(*this, OnVideoNaturalSizeChange(gfx::Size(1280, 720))).Times(1); |
| scoped_refptr<DecoderBuffer> second_file = |
| ReadTestDataFile("bear-1280x720-v_frag-cenc.mp4"); |
| source.set_expect_append_success(false); |
| source.AppendAtTime(base::TimeDelta::FromSeconds(kAppendTimeSec), |
| second_file->data(), second_file->data_size()); |
| |
| source.EndOfStream(); |
| |
| base::RunLoop().Run(); |
| EXPECT_EQ(CHUNK_DEMUXER_ERROR_APPEND_FAILED, pipeline_status_); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| // The second video was not added, so its time has not been added. |
| EXPECT_EQ(k640IsoFileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| Play(); |
| |
| EXPECT_EQ(CHUNK_DEMUXER_ERROR_APPEND_FAILED, WaitUntilEndedOrError()); |
| source.Shutdown(); |
| } |
| |
| // Config changes from encrypted to clear are not currently supported. |
| MAYBE_EME_TEST_P(MSEPipelineIntegrationTest, |
| MAYBE_EME(ConfigChange_EncryptedThenClear_MP4_CENC)) { |
| MockMediaSource source("bear-640x360-v_frag-cenc.mp4", kMP4Video, |
| kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new KeyProvidingApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| scoped_refptr<DecoderBuffer> second_file = |
| ReadTestDataFile("bear-1280x720-av_frag.mp4"); |
| |
| source.set_expect_append_success(false); |
| source.AppendAtTime(base::TimeDelta::FromSeconds(kAppendTimeSec), |
| second_file->data(), second_file->data_size()); |
| |
| source.EndOfStream(); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| if (buffering_api_ == BufferingApi::kLegacyByDts) |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| else |
| EXPECT_EQ(33, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| |
| // The second video was not added, so its time has not been added. |
| if (buffering_api_ == BufferingApi::kLegacyByDts) { |
| EXPECT_EQ(k640IsoCencFileDurationMsByDts, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| } else { |
| EXPECT_EQ(k640IsoCencFileDurationMsByPts, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| } |
| |
| Play(); |
| |
| EXPECT_EQ(CHUNK_DEMUXER_ERROR_APPEND_FAILED, WaitUntilEndedOrError()); |
| source.Shutdown(); |
| } |
| |
| // Verify files which change configuration midstream fail gracefully. |
| TEST_F(PipelineIntegrationTest, MidStreamConfigChangesFail) { |
| ASSERT_EQ(PIPELINE_OK, Start("midstream_config_change.mp3")); |
| Play(); |
| ASSERT_EQ(WaitUntilEndedOrError(), PIPELINE_ERROR_DECODE); |
| } |
| #endif // BUILDFLAG(USE_PROPRIETARY_CODECS) |
| |
| TEST_F(PipelineIntegrationTest, BasicPlayback_16x9AspectRatio) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240-16x9-aspect.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| MAYBE_EME_TEST_P(MSEPipelineIntegrationTest, |
| MAYBE_EME(EncryptedPlayback_WebM)) { |
| MockMediaSource source("bear-320x240-av_enc-av.webm", kWebM, 219816); |
| FakeEncryptedMedia encrypted_media(new KeyProvidingApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| source.EndOfStream(); |
| ASSERT_EQ(PIPELINE_OK, pipeline_status_); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| MAYBE_EME_TEST_P(MSEPipelineIntegrationTest, |
| MAYBE_EME(EncryptedPlayback_ClearStart_WebM)) { |
| MockMediaSource source("bear-320x240-av_enc-av_clear-1s.webm", kWebM, |
| kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new KeyProvidingApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| source.EndOfStream(); |
| ASSERT_EQ(PIPELINE_OK, pipeline_status_); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| MAYBE_EME_TEST_P(MSEPipelineIntegrationTest, |
| MAYBE_EME(EncryptedPlayback_NoEncryptedFrames_WebM)) { |
| MockMediaSource source("bear-320x240-av_enc-av_clear-all.webm", kWebM, |
| kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new NoResponseApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| source.EndOfStream(); |
| ASSERT_EQ(PIPELINE_OK, pipeline_status_); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| MAYBE_EME_TEST_P(MSEPipelineIntegrationTest, |
| MAYBE_EME(EncryptedPlayback_MP4_VP9_CENC_VideoOnly)) { |
| MockMediaSource source("bear-320x240-v_frag-vp9-cenc.mp4", kMP4VideoVP9, |
| kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new KeyProvidingApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| source.EndOfStream(); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, BasicPlayback_VideoOnly_MP4_VP9) { |
| MockMediaSource source("bear-320x240-v_frag-vp9.mp4", kMP4VideoVP9, |
| kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| source.EndOfStream(); |
| ASSERT_EQ(PIPELINE_OK, pipeline_status_); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| #if BUILDFLAG(USE_PROPRIETARY_CODECS) |
| MAYBE_EME_TEST_P(MSEPipelineIntegrationTest, |
| MAYBE_EME(EncryptedPlayback_MP4_CENC_VideoOnly)) { |
| MockMediaSource source("bear-1280x720-v_frag-cenc.mp4", kMP4Video, |
| kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new KeyProvidingApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| source.EndOfStream(); |
| ASSERT_EQ(PIPELINE_OK, pipeline_status_); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| MAYBE_EME_TEST_P(MSEPipelineIntegrationTest, |
| MAYBE_EME(EncryptedPlayback_MP4_CENC_AudioOnly)) { |
| MockMediaSource source("bear-1280x720-a_frag-cenc.mp4", kMP4Audio, |
| kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new KeyProvidingApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| source.EndOfStream(); |
| ASSERT_EQ(PIPELINE_OK, pipeline_status_); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| MAYBE_EME_TEST_P( |
| MSEPipelineIntegrationTest, |
| MAYBE_EME(EncryptedPlayback_NoEncryptedFrames_MP4_CENC_VideoOnly)) { |
| MockMediaSource source("bear-1280x720-v_frag-cenc_clear-all.mp4", kMP4Video, |
| kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new NoResponseApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| source.EndOfStream(); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, Mp2ts_AAC_HE_SBR_Audio) { |
| MockMediaSource source("bear-1280x720-aac_he.ts", kMP2AudioSBR, |
| kAppendWholeFile); |
| #if BUILDFLAG(ENABLE_MSE_MPEG2TS_STREAM_PARSER) |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| source.EndOfStream(); |
| ASSERT_EQ(PIPELINE_OK, pipeline_status_); |
| |
| // Check that SBR is taken into account correctly by mpeg2ts parser. When an |
| // SBR stream is parsed as non-SBR stream, then audio frame durations are |
| // calculated incorrectly and that leads to gaps in buffered ranges (so this |
| // check will fail) and eventually leads to stalled playback. |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| #else |
| EXPECT_EQ( |
| DEMUXER_ERROR_COULD_NOT_OPEN, |
| StartPipelineWithMediaSource(&source, kExpectDemuxerFailure, nullptr)); |
| #endif |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, Mpeg2ts_MP3Audio_Mp4a_6B) { |
| MockMediaSource source("bear-audio-mp4a.6B.ts", |
| "video/mp2t; codecs=\"mp4a.6B\"", kAppendWholeFile); |
| #if BUILDFLAG(ENABLE_MSE_MPEG2TS_STREAM_PARSER) |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| source.EndOfStream(); |
| ASSERT_EQ(PIPELINE_OK, pipeline_status_); |
| #else |
| EXPECT_EQ( |
| DEMUXER_ERROR_COULD_NOT_OPEN, |
| StartPipelineWithMediaSource(&source, kExpectDemuxerFailure, nullptr)); |
| #endif |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, Mpeg2ts_MP3Audio_Mp4a_69) { |
| MockMediaSource source("bear-audio-mp4a.69.ts", |
| "video/mp2t; codecs=\"mp4a.69\"", kAppendWholeFile); |
| #if BUILDFLAG(ENABLE_MSE_MPEG2TS_STREAM_PARSER) |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| source.EndOfStream(); |
| ASSERT_EQ(PIPELINE_OK, pipeline_status_); |
| #else |
| EXPECT_EQ( |
| DEMUXER_ERROR_COULD_NOT_OPEN, |
| StartPipelineWithMediaSource(&source, kExpectDemuxerFailure, nullptr)); |
| #endif |
| } |
| |
| MAYBE_EME_TEST_P( |
| MSEPipelineIntegrationTest, |
| MAYBE_EME(EncryptedPlayback_NoEncryptedFrames_MP4_CENC_AudioOnly)) { |
| MockMediaSource source("bear-1280x720-a_frag-cenc_clear-all.mp4", kMP4Audio, |
| kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new NoResponseApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| source.EndOfStream(); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| MAYBE_EME_TEST_P(MSEPipelineIntegrationTest, |
| MAYBE_EME(EncryptedPlayback_MP4_CENC_SENC_Video)) { |
| MockMediaSource source("bear-640x360-v_frag-cenc-senc.mp4", kMP4Video, |
| kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new KeyProvidingApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| source.EndOfStream(); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| // 'SAIZ' and 'SAIO' boxes contain redundant information which is already |
| // available in 'SENC' box. Although 'SAIZ' and 'SAIO' boxes are required per |
| // CENC spec for backward compatibility reasons, but we do not use the two |
| // boxes if 'SENC' box is present, so the code should work even if the two |
| // boxes are not present. |
| MAYBE_EME_TEST_P( |
| MSEPipelineIntegrationTest, |
| MAYBE_EME(EncryptedPlayback_MP4_CENC_SENC_NO_SAIZ_SAIO_Video)) { |
| MockMediaSource source("bear-640x360-v_frag-cenc-senc-no-saiz-saio.mp4", |
| kMP4Video, kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new KeyProvidingApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| source.EndOfStream(); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| MAYBE_EME_TEST_P(MSEPipelineIntegrationTest, |
| MAYBE_EME(EncryptedPlayback_MP4_CENC_KeyRotation_Video)) { |
| MockMediaSource source("bear-1280x720-v_frag-cenc-key_rotation.mp4", |
| kMP4Video, kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new RotatingKeyProvidingApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| source.EndOfStream(); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| MAYBE_EME_TEST_P(MSEPipelineIntegrationTest, |
| MAYBE_EME(EncryptedPlayback_MP4_CENC_KeyRotation_Audio)) { |
| MockMediaSource source("bear-1280x720-a_frag-cenc-key_rotation.mp4", |
| kMP4Audio, kAppendWholeFile); |
| FakeEncryptedMedia encrypted_media(new RotatingKeyProvidingApp()); |
| EXPECT_EQ(PIPELINE_OK, |
| StartPipelineWithEncryptedMedia(&source, &encrypted_media)); |
| |
| source.EndOfStream(); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, BasicPlayback_VideoOnly_MP4_AVC3) { |
| MockMediaSource source("bear-1280x720-v_frag-avc3.mp4", kMP4VideoAVC3, |
| kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| source.EndOfStream(); |
| |
| EXPECT_EQ(1u, pipeline_->GetBufferedTimeRanges().size()); |
| EXPECT_EQ(0, pipeline_->GetBufferedTimeRanges().start(0).InMilliseconds()); |
| EXPECT_EQ(k1280IsoAVC3FileDurationMs, |
| pipeline_->GetBufferedTimeRanges().end(0).InMilliseconds()); |
| |
| Play(); |
| |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, BasicPlayback_VideoOnly_MP4_HEVC1) { |
| // HEVC demuxing might be enabled even on platforms that don't support HEVC |
| // decoding. For those cases we'll get DECODER_ERROR_NOT_SUPPORTED, which |
| // indicates indicates that we did pass media mime type checks and attempted |
| // to actually demux and decode the stream. On platforms that support both |
| // demuxing and decoding we'll get PIPELINE_OK. |
| MockMediaSource source("bear-320x240-v_frag-hevc.mp4", kMP4VideoHEVC1, |
| kAppendWholeFile); |
| #if BUILDFLAG(ENABLE_HEVC_DEMUXING) |
| PipelineStatus status = StartPipelineWithMediaSource(&source); |
| EXPECT_TRUE(status == PIPELINE_OK || status == DECODER_ERROR_NOT_SUPPORTED); |
| #else |
| EXPECT_EQ( |
| DEMUXER_ERROR_COULD_NOT_OPEN, |
| StartPipelineWithMediaSource(&source, kExpectDemuxerFailure, nullptr)); |
| #endif |
| } |
| |
| TEST_P(MSEPipelineIntegrationTest, BasicPlayback_VideoOnly_MP4_HEVC2) { |
| // HEVC demuxing might be enabled even on platforms that don't support HEVC |
| // decoding. For those cases we'll get DECODER_ERROR_NOT_SUPPORTED, which |
| // indicates indicates that we did pass media mime type checks and attempted |
| // to actually demux and decode the stream. On platforms that support both |
| // demuxing and decoding we'll get PIPELINE_OK. |
| MockMediaSource source("bear-320x240-v_frag-hevc.mp4", kMP4VideoHEVC2, |
| kAppendWholeFile); |
| #if BUILDFLAG(ENABLE_HEVC_DEMUXING) |
| PipelineStatus status = StartPipelineWithMediaSource(&source); |
| EXPECT_TRUE(status == PIPELINE_OK || status == DECODER_ERROR_NOT_SUPPORTED); |
| #else |
| EXPECT_EQ( |
| DEMUXER_ERROR_COULD_NOT_OPEN, |
| StartPipelineWithMediaSource(&source, kExpectDemuxerFailure, nullptr)); |
| #endif |
| } |
| |
| #endif // BUILDFLAG(USE_PROPRIETARY_CODECS) |
| |
| TEST_F(PipelineIntegrationTest, SeekWhilePaused) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240.webm")); |
| |
| base::TimeDelta duration(pipeline_->GetMediaDuration()); |
| base::TimeDelta start_seek_time(duration / 4); |
| base::TimeDelta seek_time(duration * 3 / 4); |
| |
| Play(); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(start_seek_time)); |
| Pause(); |
| ASSERT_TRUE(Seek(seek_time)); |
| EXPECT_EQ(seek_time, pipeline_->GetMediaTime()); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| |
| // Make sure seeking after reaching the end works as expected. |
| Pause(); |
| ASSERT_TRUE(Seek(seek_time)); |
| EXPECT_EQ(seek_time, pipeline_->GetMediaTime()); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, SeekWhilePlaying) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240.webm")); |
| |
| base::TimeDelta duration(pipeline_->GetMediaDuration()); |
| base::TimeDelta start_seek_time(duration / 4); |
| base::TimeDelta seek_time(duration * 3 / 4); |
| |
| Play(); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(start_seek_time)); |
| ASSERT_TRUE(Seek(seek_time)); |
| EXPECT_GE(pipeline_->GetMediaTime(), seek_time); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| |
| // Make sure seeking after reaching the end works as expected. |
| ASSERT_TRUE(Seek(seek_time)); |
| EXPECT_GE(pipeline_->GetMediaTime(), seek_time); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, SuspendWhilePaused) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240.webm")); |
| |
| base::TimeDelta duration(pipeline_->GetMediaDuration()); |
| base::TimeDelta start_seek_time(duration / 4); |
| base::TimeDelta seek_time(duration * 3 / 4); |
| |
| Play(); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(start_seek_time)); |
| Pause(); |
| |
| // Suspend while paused. |
| ASSERT_TRUE(Suspend()); |
| |
| // Resuming the pipeline will create a new Renderer, |
| // which in turn will trigger video size and opacity notifications. |
| EXPECT_CALL(*this, OnVideoNaturalSizeChange(gfx::Size(320, 240))).Times(1); |
| EXPECT_CALL(*this, OnVideoOpacityChange(true)).Times(1); |
| |
| ASSERT_TRUE(Resume(seek_time)); |
| EXPECT_GE(pipeline_->GetMediaTime(), seek_time); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, SuspendWhilePlaying) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240.webm")); |
| |
| base::TimeDelta duration(pipeline_->GetMediaDuration()); |
| base::TimeDelta start_seek_time(duration / 4); |
| base::TimeDelta seek_time(duration * 3 / 4); |
| |
| Play(); |
| ASSERT_TRUE(WaitUntilCurrentTimeIsAfter(start_seek_time)); |
| ASSERT_TRUE(Suspend()); |
| |
| // Resuming the pipeline will create a new Renderer, |
| // which in turn will trigger video size and opacity notifications. |
| EXPECT_CALL(*this, OnVideoNaturalSizeChange(gfx::Size(320, 240))).Times(1); |
| EXPECT_CALL(*this, OnVideoOpacityChange(true)).Times(1); |
| |
| ASSERT_TRUE(Resume(seek_time)); |
| EXPECT_GE(pipeline_->GetMediaTime(), seek_time); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| #if BUILDFLAG(USE_PROPRIETARY_CODECS) |
| TEST_F(PipelineIntegrationTest, Rotated_Metadata_0) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear_rotate_0.mp4")); |
| ASSERT_EQ(VIDEO_ROTATION_0, metadata_.video_decoder_config.video_rotation()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, Rotated_Metadata_90) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear_rotate_90.mp4")); |
| ASSERT_EQ(VIDEO_ROTATION_90, metadata_.video_decoder_config.video_rotation()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, Rotated_Metadata_180) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear_rotate_180.mp4")); |
| ASSERT_EQ(VIDEO_ROTATION_180, |
| metadata_.video_decoder_config.video_rotation()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, Rotated_Metadata_270) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear_rotate_270.mp4")); |
| ASSERT_EQ(VIDEO_ROTATION_270, |
| metadata_.video_decoder_config.video_rotation()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, Spherical) { |
| ASSERT_EQ(PIPELINE_OK, Start("spherical.mp4", kHashed)); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_HASH_EQ("1cb7f980020d99ea852e22dd6bd8d9de", GetVideoHash()); |
| } |
| #endif // BUILDFLAG(USE_PROPRIETARY_CODECS) |
| |
| // Verify audio decoder & renderer can handle aborted demuxer reads. |
| TEST_P(MSEPipelineIntegrationTest, ChunkDemuxerAbortRead_AudioOnly) { |
| ASSERT_TRUE(TestSeekDuringRead("bear-320x240-audio-only.webm", kAudioOnlyWebM, |
| 16384, base::TimeDelta::FromMilliseconds(464), |
| base::TimeDelta::FromMilliseconds(617), 0x10CA, |
| 19730)); |
| } |
| |
| // Verify video decoder & renderer can handle aborted demuxer reads. |
| TEST_P(MSEPipelineIntegrationTest, ChunkDemuxerAbortRead_VideoOnly) { |
| ASSERT_TRUE(TestSeekDuringRead("bear-320x240-video-only.webm", kVideoOnlyWebM, |
| 32768, base::TimeDelta::FromMilliseconds(167), |
| base::TimeDelta::FromMilliseconds(1668), |
| 0x1C896, 65536)); |
| } |
| |
| // Verify that Opus audio in WebM containers can be played back. |
| TEST_F(PipelineIntegrationTest, BasicPlayback_AudioOnly_Opus_WebM) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-opus-end-trimming.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, |
| BasicPlayback_AudioOnly_Opus_4ch_ChannelMapping2_WebM) { |
| ASSERT_EQ( |
| PIPELINE_OK, |
| Start("bear-opus-end-trimming-4ch-channelmapping2.webm", kWebAudio)); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, |
| BasicPlayback_AudioOnly_Opus_11ch_ChannelMapping2_WebM) { |
| ASSERT_EQ( |
| PIPELINE_OK, |
| Start("bear-opus-end-trimming-11ch-channelmapping2.webm", kWebAudio)); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| // Verify that VP9 video in WebM containers can be played back. |
| TEST_F(PipelineIntegrationTest, BasicPlayback_VideoOnly_VP9_WebM) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-vp9.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| #if BUILDFLAG(ENABLE_AV1_DECODER) |
| TEST_F(PipelineIntegrationTest, BasicPlayback_VideoOnly_AV1_WebM) { |
| base::test::ScopedFeatureList scoped_feature_list_; |
| scoped_feature_list_.InitAndEnableFeature(kAv1Decoder); |
| ASSERT_EQ(PIPELINE_OK, Start("bear-av1.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| #endif |
| |
| // Verify that VP9 video and Opus audio in the same WebM container can be played |
| // back. |
| TEST_F(PipelineIntegrationTest, BasicPlayback_VP9_Opus_WebM) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-vp9-opus.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| // Verify that VP8 video with alpha channel can be played back. |
| TEST_F(PipelineIntegrationTest, BasicPlayback_VP8A_WebM) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-vp8a.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_VIDEO_FORMAT_EQ(last_video_frame_format_, PIXEL_FORMAT_YV12A); |
| } |
| |
| // Verify that VP8A video with odd width/height can be played back. |
| TEST_F(PipelineIntegrationTest, BasicPlayback_VP8A_Odd_WebM) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-vp8a-odd-dimensions.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_VIDEO_FORMAT_EQ(last_video_frame_format_, PIXEL_FORMAT_YV12A); |
| } |
| |
| // Verify that VP9 video with odd width/height can be played back. |
| TEST_F(PipelineIntegrationTest, BasicPlayback_VP9_Odd_WebM) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-vp9-odd-dimensions.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| // Verify that VP9 video with alpha channel can be played back. |
| TEST_F(PipelineIntegrationTest, BasicPlayback_VP9A_WebM) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-vp9a.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_VIDEO_FORMAT_EQ(last_video_frame_format_, PIXEL_FORMAT_YV12A); |
| } |
| |
| // Verify that VP9A video with odd width/height can be played back. |
| TEST_F(PipelineIntegrationTest, BasicPlayback_VP9A_Odd_WebM) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-vp9a-odd-dimensions.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_VIDEO_FORMAT_EQ(last_video_frame_format_, PIXEL_FORMAT_YV12A); |
| } |
| |
| // Verify that VP8 video with inband text track can be played back. |
| TEST_F(PipelineIntegrationTest, MAYBE_TEXT(BasicPlayback_VP8_WebVTT_WebM)) { |
| EXPECT_CALL(*this, OnAddTextTrack(_, _)); |
| ASSERT_EQ(PIPELINE_OK, Start("bear-vp8-webvtt.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| // Verify that VP9 video with 4:4:4 subsampling can be played back. |
| TEST_F(PipelineIntegrationTest, P444_VP9_WebM) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-320x240-P444.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_VIDEO_FORMAT_EQ(last_video_frame_format_, PIXEL_FORMAT_YV24); |
| } |
| |
| // Verify that frames of VP9 video in the BT.709 color space have the YV12HD |
| // format. |
| TEST_F(PipelineIntegrationTest, BT709_VP9_WebM) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-vp9-bt709.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| EXPECT_VIDEO_FORMAT_EQ(last_video_frame_format_, PIXEL_FORMAT_YV12); |
| EXPECT_COLOR_SPACE_EQ(last_video_frame_color_space_, COLOR_SPACE_HD_REC709); |
| } |
| |
| TEST_F(PipelineIntegrationTest, HD_VP9_WebM) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear-1280x720.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| // Verify that videos with an odd frame size playback successfully. |
| TEST_F(PipelineIntegrationTest, BasicPlayback_OddVideoSize) { |
| ASSERT_EQ(PIPELINE_OK, Start("butterfly-853x480.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| // Verify that OPUS audio in a webm which reports a 44.1kHz sample rate plays |
| // correctly at 48kHz |
| TEST_F(PipelineIntegrationTest, BasicPlayback_Opus441kHz) { |
| ASSERT_EQ(PIPELINE_OK, Start("sfx-opus-441.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| |
| EXPECT_EQ(48000, demuxer_->GetFirstStream(DemuxerStream::AUDIO) |
| ->audio_decoder_config() |
| .samples_per_second()); |
| } |
| |
| // Same as above but using MediaSource. |
| TEST_P(MSEPipelineIntegrationTest, BasicPlayback_Opus441kHz) { |
| MockMediaSource source("sfx-opus-441.webm", kOpusAudioOnlyWebM, |
| kAppendWholeFile); |
| EXPECT_EQ(PIPELINE_OK, StartPipelineWithMediaSource(&source)); |
| source.EndOfStream(); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| source.Shutdown(); |
| Stop(); |
| EXPECT_EQ(48000, demuxer_->GetFirstStream(DemuxerStream::AUDIO) |
| ->audio_decoder_config() |
| .samples_per_second()); |
| } |
| |
| // Ensures audio-only playback with missing or negative timestamps works. Tests |
| // the common live-streaming case for chained ogg. See http://crbug.com/396864. |
| TEST_F(PipelineIntegrationTest, BasicPlaybackChainedOgg) { |
| ASSERT_EQ(PIPELINE_OK, Start("double-sfx.ogg", kUnreliableDuration)); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| ASSERT_EQ(base::TimeDelta(), demuxer_->GetStartTime()); |
| } |
| |
| // Ensures audio-video playback with missing or negative timestamps fails softly |
| // instead of crashing. See http://crbug.com/396864. |
| TEST_F(PipelineIntegrationTest, BasicPlaybackChainedOggVideo) { |
| ASSERT_EQ(PIPELINE_OK, Start("double-bear.ogv", kUnreliableDuration)); |
| Play(); |
| EXPECT_EQ(PIPELINE_ERROR_DECODE, WaitUntilEndedOrError()); |
| ASSERT_EQ(base::TimeDelta(), demuxer_->GetStartTime()); |
| } |
| |
| // Tests that we signal ended even when audio runs longer than video track. |
| TEST_F(PipelineIntegrationTest, BasicPlaybackAudioLongerThanVideo) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear_audio_longer_than_video.ogv")); |
| // Audio track is 2000ms. Video track is 1001ms. Duration should be higher |
| // of the two. |
| EXPECT_EQ(2000, pipeline_->GetMediaDuration().InMilliseconds()); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| // Tests that we signal ended even when audio runs shorter than video track. |
| TEST_F(PipelineIntegrationTest, BasicPlaybackAudioShorterThanVideo) { |
| ASSERT_EQ(PIPELINE_OK, Start("bear_audio_shorter_than_video.ogv")); |
| // Audio track is 500ms. Video track is 1001ms. Duration should be higher of |
| // the two. |
| EXPECT_EQ(1001, pipeline_->GetMediaDuration().InMilliseconds()); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| } |
| |
| TEST_F(PipelineIntegrationTest, BasicPlaybackPositiveStartTime) { |
| ASSERT_EQ(PIPELINE_OK, Start("nonzero-start-time.webm")); |
| Play(); |
| ASSERT_TRUE(WaitUntilOnEnded()); |
| ASSERT_EQ(base::TimeDelta::FromMicroseconds(396000), |
| demuxer_->GetStartTime()); |
| } |
| |
| INSTANTIATE_TEST_CASE_P(LegacyByDts, |
| MSEPipelineIntegrationTest, |
| ::testing::Values(BufferingApi::kLegacyByDts)); |
| INSTANTIATE_TEST_CASE_P(NewByPts, |
| MSEPipelineIntegrationTest, |
| ::testing::Values(BufferingApi::kNewByPts)); |
| |
| } // namespace media |