/* * libjingle * Copyright 2012, Google Inc. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * 1. Redistributions of source code must retain the above copyright notice, * this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright notice, * this list of conditions and the following disclaimer in the documentation * and/or other materials provided with the distribution. * 3. The name of the author may not be used to endorse or promote products * derived from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR IMPLIED * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO * EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR * OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include #include #include "talk/app/webrtc/audiotrack.h" #include "talk/app/webrtc/mediastream.h" #include "talk/app/webrtc/mediastreamsignaling.h" #include "talk/app/webrtc/sctputils.h" #include "talk/app/webrtc/streamcollection.h" #include "talk/app/webrtc/test/fakeconstraints.h" #include "talk/app/webrtc/test/fakedatachannelprovider.h" #include "talk/app/webrtc/videotrack.h" #include "talk/media/base/fakemediaengine.h" #include "talk/media/devices/fakedevicemanager.h" #include "webrtc/p2p/base/constants.h" #include "webrtc/p2p/base/sessiondescription.h" #include "talk/session/media/channelmanager.h" #include "webrtc/base/gunit.h" #include "webrtc/base/scoped_ptr.h" #include "webrtc/base/stringutils.h" #include "webrtc/base/thread.h" static const char kStreams[][8] = {"stream1", "stream2"}; static const char kAudioTracks[][32] = {"audiotrack0", "audiotrack1"}; static const char kVideoTracks[][32] = {"videotrack0", "videotrack1"}; using webrtc::AudioTrack; using webrtc::AudioTrackInterface; using webrtc::AudioTrackVector; using webrtc::VideoTrack; using webrtc::VideoTrackInterface; using webrtc::VideoTrackVector; using webrtc::DataChannelInterface; using webrtc::FakeConstraints; using webrtc::IceCandidateInterface; using webrtc::MediaConstraintsInterface; using webrtc::MediaStreamInterface; using webrtc::MediaStreamTrackInterface; using webrtc::PeerConnectionInterface; using webrtc::SdpParseError; using webrtc::SessionDescriptionInterface; using webrtc::StreamCollection; using webrtc::StreamCollectionInterface; typedef PeerConnectionInterface::RTCOfferAnswerOptions RTCOfferAnswerOptions; // Reference SDP with a MediaStream with label "stream1" and audio track with // id "audio_1" and a video track with id "video_1; static const char kSdpStringWithStream1[] = "v=0\r\n" "o=- 0 0 IN IP4 127.0.0.1\r\n" "s=-\r\n" "t=0 0\r\n" "m=audio 1 RTP/AVPF 103\r\n" "a=mid:audio\r\n" "a=rtpmap:103 ISAC/16000\r\n" "a=ssrc:1 cname:stream1\r\n" "a=ssrc:1 mslabel:stream1\r\n" "a=ssrc:1 label:audiotrack0\r\n" "m=video 1 RTP/AVPF 120\r\n" "a=mid:video\r\n" "a=rtpmap:120 VP8/90000\r\n" "a=ssrc:2 cname:stream1\r\n" "a=ssrc:2 mslabel:stream1\r\n" "a=ssrc:2 label:videotrack0\r\n"; // Reference SDP with two MediaStreams with label "stream1" and "stream2. Each // MediaStreams have one audio track and one video track. // This uses MSID. static const char kSdpStringWith2Stream[] = "v=0\r\n" "o=- 0 0 IN IP4 127.0.0.1\r\n" "s=-\r\n" "t=0 0\r\n" "a=msid-semantic: WMS stream1 stream2\r\n" "m=audio 1 RTP/AVPF 103\r\n" "a=mid:audio\r\n" "a=rtpmap:103 ISAC/16000\r\n" "a=ssrc:1 cname:stream1\r\n" "a=ssrc:1 msid:stream1 audiotrack0\r\n" "a=ssrc:3 cname:stream2\r\n" "a=ssrc:3 msid:stream2 audiotrack1\r\n" "m=video 1 RTP/AVPF 120\r\n" "a=mid:video\r\n" "a=rtpmap:120 VP8/0\r\n" "a=ssrc:2 cname:stream1\r\n" "a=ssrc:2 msid:stream1 videotrack0\r\n" "a=ssrc:4 cname:stream2\r\n" "a=ssrc:4 msid:stream2 videotrack1\r\n"; // Reference SDP without MediaStreams. Msid is not supported. static const char kSdpStringWithoutStreams[] = "v=0\r\n" "o=- 0 0 IN IP4 127.0.0.1\r\n" "s=-\r\n" "t=0 0\r\n" "m=audio 1 RTP/AVPF 103\r\n" "a=mid:audio\r\n" "a=rtpmap:103 ISAC/16000\r\n" "m=video 1 RTP/AVPF 120\r\n" "a=mid:video\r\n" "a=rtpmap:120 VP8/90000\r\n"; // Reference SDP without MediaStreams. Msid is supported. static const char kSdpStringWithMsidWithoutStreams[] = "v=0\r\n" "o=- 0 0 IN IP4 127.0.0.1\r\n" "s=-\r\n" "t=0 0\r\n" "a=msid-semantic: WMS\r\n" "m=audio 1 RTP/AVPF 103\r\n" "a=mid:audio\r\n" "a=rtpmap:103 ISAC/16000\r\n" "m=video 1 RTP/AVPF 120\r\n" "a=mid:video\r\n" "a=rtpmap:120 VP8/90000\r\n"; // Reference SDP without MediaStreams and audio only. static const char kSdpStringWithoutStreamsAudioOnly[] = "v=0\r\n" "o=- 0 0 IN IP4 127.0.0.1\r\n" "s=-\r\n" "t=0 0\r\n" "m=audio 1 RTP/AVPF 103\r\n" "a=mid:audio\r\n" "a=rtpmap:103 ISAC/16000\r\n"; // Reference SENDONLY SDP without MediaStreams. Msid is not supported. static const char kSdpStringSendOnlyWithWithoutStreams[] = "v=0\r\n" "o=- 0 0 IN IP4 127.0.0.1\r\n" "s=-\r\n" "t=0 0\r\n" "m=audio 1 RTP/AVPF 103\r\n" "a=mid:audio\r\n" "a=sendonly" "a=rtpmap:103 ISAC/16000\r\n" "m=video 1 RTP/AVPF 120\r\n" "a=mid:video\r\n" "a=sendonly" "a=rtpmap:120 VP8/90000\r\n"; static const char kSdpStringInit[] = "v=0\r\n" "o=- 0 0 IN IP4 127.0.0.1\r\n" "s=-\r\n" "t=0 0\r\n" "a=msid-semantic: WMS\r\n"; static const char kSdpStringAudio[] = "m=audio 1 RTP/AVPF 103\r\n" "a=mid:audio\r\n" "a=rtpmap:103 ISAC/16000\r\n"; static const char kSdpStringVideo[] = "m=video 1 RTP/AVPF 120\r\n" "a=mid:video\r\n" "a=rtpmap:120 VP8/90000\r\n"; static const char kSdpStringMs1Audio0[] = "a=ssrc:1 cname:stream1\r\n" "a=ssrc:1 msid:stream1 audiotrack0\r\n"; static const char kSdpStringMs1Video0[] = "a=ssrc:2 cname:stream1\r\n" "a=ssrc:2 msid:stream1 videotrack0\r\n"; static const char kSdpStringMs1Audio1[] = "a=ssrc:3 cname:stream1\r\n" "a=ssrc:3 msid:stream1 audiotrack1\r\n"; static const char kSdpStringMs1Video1[] = "a=ssrc:4 cname:stream1\r\n" "a=ssrc:4 msid:stream1 videotrack1\r\n"; // Verifies that |options| contain all tracks in |collection| and that // the |options| has set the the has_audio and has_video flags correct. static void VerifyMediaOptions(StreamCollectionInterface* collection, const cricket::MediaSessionOptions& options) { if (!collection) { return; } size_t stream_index = 0; for (size_t i = 0; i < collection->count(); ++i) { MediaStreamInterface* stream = collection->at(i); AudioTrackVector audio_tracks = stream->GetAudioTracks(); ASSERT_GE(options.streams.size(), stream_index + audio_tracks.size()); for (size_t j = 0; j < audio_tracks.size(); ++j) { webrtc::AudioTrackInterface* audio = audio_tracks[j]; EXPECT_EQ(options.streams[stream_index].sync_label, stream->label()); EXPECT_EQ(options.streams[stream_index++].id, audio->id()); EXPECT_TRUE(options.has_audio()); } VideoTrackVector video_tracks = stream->GetVideoTracks(); ASSERT_GE(options.streams.size(), stream_index + video_tracks.size()); for (size_t j = 0; j < video_tracks.size(); ++j) { webrtc::VideoTrackInterface* video = video_tracks[j]; EXPECT_EQ(options.streams[stream_index].sync_label, stream->label()); EXPECT_EQ(options.streams[stream_index++].id, video->id()); EXPECT_TRUE(options.has_video()); } } } static bool CompareStreamCollections(StreamCollectionInterface* s1, StreamCollectionInterface* s2) { if (s1 == NULL || s2 == NULL || s1->count() != s2->count()) return false; for (size_t i = 0; i != s1->count(); ++i) { if (s1->at(i)->label() != s2->at(i)->label()) return false; webrtc::AudioTrackVector audio_tracks1 = s1->at(i)->GetAudioTracks(); webrtc::AudioTrackVector audio_tracks2 = s2->at(i)->GetAudioTracks(); webrtc::VideoTrackVector video_tracks1 = s1->at(i)->GetVideoTracks(); webrtc::VideoTrackVector video_tracks2 = s2->at(i)->GetVideoTracks(); if (audio_tracks1.size() != audio_tracks2.size()) return false; for (size_t j = 0; j != audio_tracks1.size(); ++j) { if (audio_tracks1[j]->id() != audio_tracks2[j]->id()) return false; } if (video_tracks1.size() != video_tracks2.size()) return false; for (size_t j = 0; j != video_tracks1.size(); ++j) { if (video_tracks1[j]->id() != video_tracks2[j]->id()) return false; } } return true; } class FakeDataChannelFactory : public webrtc::DataChannelFactory { public: FakeDataChannelFactory(FakeDataChannelProvider* provider, cricket::DataChannelType dct, webrtc::MediaStreamSignaling* media_stream_signaling) : provider_(provider), type_(dct), media_stream_signaling_(media_stream_signaling) {} virtual rtc::scoped_refptr CreateDataChannel( const std::string& label, const webrtc::InternalDataChannelInit* config) { last_init_ = *config; rtc::scoped_refptr data_channel = webrtc::DataChannel::Create(provider_, type_, label, *config); media_stream_signaling_->AddDataChannel(data_channel); return data_channel; } const webrtc::InternalDataChannelInit& last_init() const { return last_init_; } private: FakeDataChannelProvider* provider_; cricket::DataChannelType type_; webrtc::MediaStreamSignaling* media_stream_signaling_; webrtc::InternalDataChannelInit last_init_; }; class MockSignalingObserver : public webrtc::MediaStreamSignalingObserver { public: MockSignalingObserver() : remote_media_streams_(StreamCollection::Create()) { } virtual ~MockSignalingObserver() { } // New remote stream have been discovered. virtual void OnAddRemoteStream(MediaStreamInterface* remote_stream) { remote_media_streams_->AddStream(remote_stream); } // Remote stream is no longer available. virtual void OnRemoveRemoteStream(MediaStreamInterface* remote_stream) { remote_media_streams_->RemoveStream(remote_stream); } virtual void OnAddDataChannel(DataChannelInterface* data_channel) { } virtual void OnAddLocalAudioTrack(MediaStreamInterface* stream, AudioTrackInterface* audio_track, uint32 ssrc) { AddTrack(&local_audio_tracks_, stream, audio_track, ssrc); } virtual void OnAddLocalVideoTrack(MediaStreamInterface* stream, VideoTrackInterface* video_track, uint32 ssrc) { AddTrack(&local_video_tracks_, stream, video_track, ssrc); } virtual void OnRemoveLocalAudioTrack(MediaStreamInterface* stream, AudioTrackInterface* audio_track, uint32 ssrc) { RemoveTrack(&local_audio_tracks_, stream, audio_track); } virtual void OnRemoveLocalVideoTrack(MediaStreamInterface* stream, VideoTrackInterface* video_track) { RemoveTrack(&local_video_tracks_, stream, video_track); } virtual void OnAddRemoteAudioTrack(MediaStreamInterface* stream, AudioTrackInterface* audio_track, uint32 ssrc) { AddTrack(&remote_audio_tracks_, stream, audio_track, ssrc); } virtual void OnAddRemoteVideoTrack(MediaStreamInterface* stream, VideoTrackInterface* video_track, uint32 ssrc) { AddTrack(&remote_video_tracks_, stream, video_track, ssrc); } virtual void OnRemoveRemoteAudioTrack(MediaStreamInterface* stream, AudioTrackInterface* audio_track) { RemoveTrack(&remote_audio_tracks_, stream, audio_track); } virtual void OnRemoveRemoteVideoTrack(MediaStreamInterface* stream, VideoTrackInterface* video_track) { RemoveTrack(&remote_video_tracks_, stream, video_track); } virtual void OnRemoveLocalStream(MediaStreamInterface* stream) { } MediaStreamInterface* RemoteStream(const std::string& label) { return remote_media_streams_->find(label); } StreamCollectionInterface* remote_streams() const { return remote_media_streams_; } size_t NumberOfRemoteAudioTracks() { return remote_audio_tracks_.size(); } void VerifyRemoteAudioTrack(const std::string& stream_label, const std::string& track_id, uint32 ssrc) { VerifyTrack(remote_audio_tracks_, stream_label, track_id, ssrc); } size_t NumberOfRemoteVideoTracks() { return remote_video_tracks_.size(); } void VerifyRemoteVideoTrack(const std::string& stream_label, const std::string& track_id, uint32 ssrc) { VerifyTrack(remote_video_tracks_, stream_label, track_id, ssrc); } size_t NumberOfLocalAudioTracks() { return local_audio_tracks_.size(); } void VerifyLocalAudioTrack(const std::string& stream_label, const std::string& track_id, uint32 ssrc) { VerifyTrack(local_audio_tracks_, stream_label, track_id, ssrc); } size_t NumberOfLocalVideoTracks() { return local_video_tracks_.size(); } void VerifyLocalVideoTrack(const std::string& stream_label, const std::string& track_id, uint32 ssrc) { VerifyTrack(local_video_tracks_, stream_label, track_id, ssrc); } private: struct TrackInfo { TrackInfo() {} TrackInfo(const std::string& stream_label, const std::string track_id, uint32 ssrc) : stream_label(stream_label), track_id(track_id), ssrc(ssrc) { } std::string stream_label; std::string track_id; uint32 ssrc; }; typedef std::vector TrackInfos; void AddTrack(TrackInfos* track_infos, MediaStreamInterface* stream, MediaStreamTrackInterface* track, uint32 ssrc) { (*track_infos).push_back(TrackInfo(stream->label(), track->id(), ssrc)); } void RemoveTrack(TrackInfos* track_infos, MediaStreamInterface* stream, MediaStreamTrackInterface* track) { for (TrackInfos::iterator it = track_infos->begin(); it != track_infos->end(); ++it) { if (it->stream_label == stream->label() && it->track_id == track->id()) { track_infos->erase(it); return; } } ADD_FAILURE(); } const TrackInfo* FindTrackInfo(const TrackInfos& infos, const std::string& stream_label, const std::string track_id) const { for (TrackInfos::const_iterator it = infos.begin(); it != infos.end(); ++it) { if (it->stream_label == stream_label && it->track_id == track_id) return &*it; } return NULL; } void VerifyTrack(const TrackInfos& track_infos, const std::string& stream_label, const std::string& track_id, uint32 ssrc) { const TrackInfo* track_info = FindTrackInfo(track_infos, stream_label, track_id); ASSERT_TRUE(track_info != NULL); EXPECT_EQ(ssrc, track_info->ssrc); } TrackInfos remote_audio_tracks_; TrackInfos remote_video_tracks_; TrackInfos local_audio_tracks_; TrackInfos local_video_tracks_; rtc::scoped_refptr remote_media_streams_; }; class MediaStreamSignalingForTest : public webrtc::MediaStreamSignaling { public: MediaStreamSignalingForTest(MockSignalingObserver* observer, cricket::ChannelManager* channel_manager) : webrtc::MediaStreamSignaling(rtc::Thread::Current(), observer, channel_manager) { }; using webrtc::MediaStreamSignaling::GetOptionsForOffer; using webrtc::MediaStreamSignaling::GetOptionsForAnswer; using webrtc::MediaStreamSignaling::OnRemoteDescriptionChanged; using webrtc::MediaStreamSignaling::remote_streams; }; class MediaStreamSignalingTest: public testing::Test { protected: virtual void SetUp() { observer_.reset(new MockSignalingObserver()); channel_manager_.reset( new cricket::ChannelManager(new cricket::FakeMediaEngine(), new cricket::FakeDeviceManager(), rtc::Thread::Current())); signaling_.reset(new MediaStreamSignalingForTest(observer_.get(), channel_manager_.get())); data_channel_provider_.reset(new FakeDataChannelProvider()); } // Create a collection of streams. // CreateStreamCollection(1) creates a collection that // correspond to kSdpString1. // CreateStreamCollection(2) correspond to kSdpString2. rtc::scoped_refptr CreateStreamCollection(int number_of_streams) { rtc::scoped_refptr local_collection( StreamCollection::Create()); for (int i = 0; i < number_of_streams; ++i) { rtc::scoped_refptr stream( webrtc::MediaStream::Create(kStreams[i])); // Add a local audio track. rtc::scoped_refptr audio_track( webrtc::AudioTrack::Create(kAudioTracks[i], NULL)); stream->AddTrack(audio_track); // Add a local video track. rtc::scoped_refptr video_track( webrtc::VideoTrack::Create(kVideoTracks[i], NULL)); stream->AddTrack(video_track); local_collection->AddStream(stream); } return local_collection; } // This functions Creates a MediaStream with label kStreams[0] and // |number_of_audio_tracks| and |number_of_video_tracks| tracks and the // corresponding SessionDescriptionInterface. The SessionDescriptionInterface // is returned in |desc| and the MediaStream is stored in // |reference_collection_| void CreateSessionDescriptionAndReference( size_t number_of_audio_tracks, size_t number_of_video_tracks, SessionDescriptionInterface** desc) { ASSERT_TRUE(desc != NULL); ASSERT_LE(number_of_audio_tracks, 2u); ASSERT_LE(number_of_video_tracks, 2u); reference_collection_ = StreamCollection::Create(); std::string sdp_ms1 = std::string(kSdpStringInit); std::string mediastream_label = kStreams[0]; rtc::scoped_refptr stream( webrtc::MediaStream::Create(mediastream_label)); reference_collection_->AddStream(stream); if (number_of_audio_tracks > 0) { sdp_ms1 += std::string(kSdpStringAudio); sdp_ms1 += std::string(kSdpStringMs1Audio0); AddAudioTrack(kAudioTracks[0], stream); } if (number_of_audio_tracks > 1) { sdp_ms1 += kSdpStringMs1Audio1; AddAudioTrack(kAudioTracks[1], stream); } if (number_of_video_tracks > 0) { sdp_ms1 += std::string(kSdpStringVideo); sdp_ms1 += std::string(kSdpStringMs1Video0); AddVideoTrack(kVideoTracks[0], stream); } if (number_of_video_tracks > 1) { sdp_ms1 += kSdpStringMs1Video1; AddVideoTrack(kVideoTracks[1], stream); } *desc = webrtc::CreateSessionDescription( SessionDescriptionInterface::kOffer, sdp_ms1, NULL); } void AddAudioTrack(const std::string& track_id, MediaStreamInterface* stream) { rtc::scoped_refptr audio_track( webrtc::AudioTrack::Create(track_id, NULL)); ASSERT_TRUE(stream->AddTrack(audio_track)); } void AddVideoTrack(const std::string& track_id, MediaStreamInterface* stream) { rtc::scoped_refptr video_track( webrtc::VideoTrack::Create(track_id, NULL)); ASSERT_TRUE(stream->AddTrack(video_track)); } rtc::scoped_refptr AddDataChannel( cricket::DataChannelType type, const std::string& label, int id) { webrtc::InternalDataChannelInit config; config.id = id; rtc::scoped_refptr data_channel( webrtc::DataChannel::Create( data_channel_provider_.get(), type, label, config)); EXPECT_TRUE(data_channel.get() != NULL); EXPECT_TRUE(signaling_->AddDataChannel(data_channel.get())); return data_channel; } // ChannelManager is used by VideoSource, so it should be released after all // the video tracks. Put it as the first private variable should ensure that. rtc::scoped_ptr channel_manager_; rtc::scoped_refptr reference_collection_; rtc::scoped_ptr observer_; rtc::scoped_ptr signaling_; rtc::scoped_ptr data_channel_provider_; }; TEST_F(MediaStreamSignalingTest, GetOptionsForOfferWithInvalidAudioOption) { RTCOfferAnswerOptions rtc_options; rtc_options.offer_to_receive_audio = RTCOfferAnswerOptions::kUndefined - 1; cricket::MediaSessionOptions options; EXPECT_FALSE(signaling_->GetOptionsForOffer(rtc_options, &options)); rtc_options.offer_to_receive_audio = RTCOfferAnswerOptions::kMaxOfferToReceiveMedia + 1; EXPECT_FALSE(signaling_->GetOptionsForOffer(rtc_options, &options)); } TEST_F(MediaStreamSignalingTest, GetOptionsForOfferWithInvalidVideoOption) { RTCOfferAnswerOptions rtc_options; rtc_options.offer_to_receive_video = RTCOfferAnswerOptions::kUndefined - 1; cricket::MediaSessionOptions options; EXPECT_FALSE(signaling_->GetOptionsForOffer(rtc_options, &options)); rtc_options.offer_to_receive_video = RTCOfferAnswerOptions::kMaxOfferToReceiveMedia + 1; EXPECT_FALSE(signaling_->GetOptionsForOffer(rtc_options, &options)); } // Test that a MediaSessionOptions is created for an offer if // OfferToReceiveAudio and OfferToReceiveVideo options are set but no // MediaStreams are sent. TEST_F(MediaStreamSignalingTest, GetMediaSessionOptionsForOfferWithAudioVideo) { RTCOfferAnswerOptions rtc_options; rtc_options.offer_to_receive_audio = 1; rtc_options.offer_to_receive_video = 1; cricket::MediaSessionOptions options; EXPECT_TRUE(signaling_->GetOptionsForOffer(rtc_options, &options)); EXPECT_TRUE(options.has_audio()); EXPECT_TRUE(options.has_video()); EXPECT_TRUE(options.bundle_enabled); } // Test that a correct MediaSessionOptions is created for an offer if // OfferToReceiveAudio is set but no MediaStreams are sent. TEST_F(MediaStreamSignalingTest, GetMediaSessionOptionsForOfferWithAudio) { RTCOfferAnswerOptions rtc_options; rtc_options.offer_to_receive_audio = 1; cricket::MediaSessionOptions options; EXPECT_TRUE(signaling_->GetOptionsForOffer(rtc_options, &options)); EXPECT_TRUE(options.has_audio()); EXPECT_FALSE(options.has_video()); EXPECT_TRUE(options.bundle_enabled); } // Test that a correct MediaSessionOptions is created for an offer if // the default OfferOptons is used or MediaStreams are sent. TEST_F(MediaStreamSignalingTest, GetDefaultMediaSessionOptionsForOffer) { RTCOfferAnswerOptions rtc_options; cricket::MediaSessionOptions options; EXPECT_TRUE(signaling_->GetOptionsForOffer(rtc_options, &options)); EXPECT_FALSE(options.has_audio()); EXPECT_FALSE(options.has_video()); EXPECT_FALSE(options.bundle_enabled); EXPECT_TRUE(options.vad_enabled); EXPECT_FALSE(options.transport_options.ice_restart); } // Test that a correct MediaSessionOptions is created for an offer if // OfferToReceiveVideo is set but no MediaStreams are sent. TEST_F(MediaStreamSignalingTest, GetMediaSessionOptionsForOfferWithVideo) { RTCOfferAnswerOptions rtc_options; rtc_options.offer_to_receive_audio = 0; rtc_options.offer_to_receive_video = 1; cricket::MediaSessionOptions options; EXPECT_TRUE(signaling_->GetOptionsForOffer(rtc_options, &options)); EXPECT_FALSE(options.has_audio()); EXPECT_TRUE(options.has_video()); EXPECT_TRUE(options.bundle_enabled); } // Test that a correct MediaSessionOptions is created for an offer if // UseRtpMux is set to false. TEST_F(MediaStreamSignalingTest, GetMediaSessionOptionsForOfferWithBundleDisabled) { RTCOfferAnswerOptions rtc_options; rtc_options.offer_to_receive_audio = 1; rtc_options.offer_to_receive_video = 1; rtc_options.use_rtp_mux = false; cricket::MediaSessionOptions options; EXPECT_TRUE(signaling_->GetOptionsForOffer(rtc_options, &options)); EXPECT_TRUE(options.has_audio()); EXPECT_TRUE(options.has_video()); EXPECT_FALSE(options.bundle_enabled); } // Test that a correct MediaSessionOptions is created to restart ice if // IceRestart is set. It also tests that subsequent MediaSessionOptions don't // have |transport_options.ice_restart| set. TEST_F(MediaStreamSignalingTest, GetMediaSessionOptionsForOfferWithIceRestart) { RTCOfferAnswerOptions rtc_options; rtc_options.ice_restart = true; cricket::MediaSessionOptions options; EXPECT_TRUE(signaling_->GetOptionsForOffer(rtc_options, &options)); EXPECT_TRUE(options.transport_options.ice_restart); rtc_options = RTCOfferAnswerOptions(); EXPECT_TRUE(signaling_->GetOptionsForOffer(rtc_options, &options)); EXPECT_FALSE(options.transport_options.ice_restart); } // Test that a correct MediaSessionOptions are created for an offer if // a MediaStream is sent and later updated with a new track. // MediaConstraints are not used. TEST_F(MediaStreamSignalingTest, AddTrackToLocalMediaStream) { RTCOfferAnswerOptions rtc_options; rtc::scoped_refptr local_streams( CreateStreamCollection(1)); MediaStreamInterface* local_stream = local_streams->at(0); EXPECT_TRUE(signaling_->AddLocalStream(local_stream)); cricket::MediaSessionOptions options; EXPECT_TRUE(signaling_->GetOptionsForOffer(rtc_options, &options)); VerifyMediaOptions(local_streams, options); cricket::MediaSessionOptions updated_options; local_stream->AddTrack(AudioTrack::Create(kAudioTracks[1], NULL)); EXPECT_TRUE(signaling_->GetOptionsForOffer(rtc_options, &options)); VerifyMediaOptions(local_streams, options); } // Test that the MediaConstraints in an answer don't affect if audio and video // is offered in an offer but that if kOfferToReceiveAudio or // kOfferToReceiveVideo constraints are true in an offer, the media type will be // included in subsequent answers. TEST_F(MediaStreamSignalingTest, MediaConstraintsInAnswer) { FakeConstraints answer_c; answer_c.SetMandatoryReceiveAudio(true); answer_c.SetMandatoryReceiveVideo(true); cricket::MediaSessionOptions answer_options; EXPECT_TRUE(signaling_->GetOptionsForAnswer(&answer_c, &answer_options)); EXPECT_TRUE(answer_options.has_audio()); EXPECT_TRUE(answer_options.has_video()); RTCOfferAnswerOptions rtc_offer_optoins; cricket::MediaSessionOptions offer_options; EXPECT_TRUE( signaling_->GetOptionsForOffer(rtc_offer_optoins, &offer_options)); EXPECT_FALSE(offer_options.has_audio()); EXPECT_FALSE(offer_options.has_video()); RTCOfferAnswerOptions updated_rtc_offer_optoins; updated_rtc_offer_optoins.offer_to_receive_audio = 1; updated_rtc_offer_optoins.offer_to_receive_video = 1; cricket::MediaSessionOptions updated_offer_options; EXPECT_TRUE(signaling_->GetOptionsForOffer(updated_rtc_offer_optoins, &updated_offer_options)); EXPECT_TRUE(updated_offer_options.has_audio()); EXPECT_TRUE(updated_offer_options.has_video()); // Since an offer has been created with both audio and video, subsequent // offers and answers should contain both audio and video. // Answers will only contain the media types that exist in the offer // regardless of the value of |updated_answer_options.has_audio| and // |updated_answer_options.has_video|. FakeConstraints updated_answer_c; answer_c.SetMandatoryReceiveAudio(false); answer_c.SetMandatoryReceiveVideo(false); cricket::MediaSessionOptions updated_answer_options; EXPECT_TRUE(signaling_->GetOptionsForAnswer(&updated_answer_c, &updated_answer_options)); EXPECT_TRUE(updated_answer_options.has_audio()); EXPECT_TRUE(updated_answer_options.has_video()); RTCOfferAnswerOptions default_rtc_options; EXPECT_TRUE(signaling_->GetOptionsForOffer(default_rtc_options, &updated_offer_options)); // By default, |has_audio| or |has_video| are false if there is no media // track. EXPECT_FALSE(updated_offer_options.has_audio()); EXPECT_FALSE(updated_offer_options.has_video()); } // This test verifies that the remote MediaStreams corresponding to a received // SDP string is created. In this test the two separate MediaStreams are // signaled. TEST_F(MediaStreamSignalingTest, UpdateRemoteStreams) { rtc::scoped_ptr desc( webrtc::CreateSessionDescription(SessionDescriptionInterface::kOffer, kSdpStringWithStream1, NULL)); EXPECT_TRUE(desc != NULL); signaling_->OnRemoteDescriptionChanged(desc.get()); rtc::scoped_refptr reference( CreateStreamCollection(1)); EXPECT_TRUE(CompareStreamCollections(signaling_->remote_streams(), reference.get())); EXPECT_TRUE(CompareStreamCollections(observer_->remote_streams(), reference.get())); EXPECT_EQ(1u, observer_->NumberOfRemoteAudioTracks()); observer_->VerifyRemoteAudioTrack(kStreams[0], kAudioTracks[0], 1); EXPECT_EQ(1u, observer_->NumberOfRemoteVideoTracks()); observer_->VerifyRemoteVideoTrack(kStreams[0], kVideoTracks[0], 2); ASSERT_EQ(1u, observer_->remote_streams()->count()); MediaStreamInterface* remote_stream = observer_->remote_streams()->at(0); EXPECT_TRUE(remote_stream->GetVideoTracks()[0]->GetSource() != NULL); // Create a session description based on another SDP with another // MediaStream. rtc::scoped_ptr update_desc( webrtc::CreateSessionDescription(SessionDescriptionInterface::kOffer, kSdpStringWith2Stream, NULL)); EXPECT_TRUE(update_desc != NULL); signaling_->OnRemoteDescriptionChanged(update_desc.get()); rtc::scoped_refptr reference2( CreateStreamCollection(2)); EXPECT_TRUE(CompareStreamCollections(signaling_->remote_streams(), reference2.get())); EXPECT_TRUE(CompareStreamCollections(observer_->remote_streams(), reference2.get())); EXPECT_EQ(2u, observer_->NumberOfRemoteAudioTracks()); observer_->VerifyRemoteAudioTrack(kStreams[0], kAudioTracks[0], 1); observer_->VerifyRemoteAudioTrack(kStreams[1], kAudioTracks[1], 3); EXPECT_EQ(2u, observer_->NumberOfRemoteVideoTracks()); observer_->VerifyRemoteVideoTrack(kStreams[0], kVideoTracks[0], 2); observer_->VerifyRemoteVideoTrack(kStreams[1], kVideoTracks[1], 4); } // This test verifies that the remote MediaStreams corresponding to a received // SDP string is created. In this test the same remote MediaStream is signaled // but MediaStream tracks are added and removed. TEST_F(MediaStreamSignalingTest, AddRemoveTrackFromExistingRemoteMediaStream) { rtc::scoped_ptr desc_ms1; CreateSessionDescriptionAndReference(1, 1, desc_ms1.use()); signaling_->OnRemoteDescriptionChanged(desc_ms1.get()); EXPECT_TRUE(CompareStreamCollections(signaling_->remote_streams(), reference_collection_)); // Add extra audio and video tracks to the same MediaStream. rtc::scoped_ptr desc_ms1_two_tracks; CreateSessionDescriptionAndReference(2, 2, desc_ms1_two_tracks.use()); signaling_->OnRemoteDescriptionChanged(desc_ms1_two_tracks.get()); EXPECT_TRUE(CompareStreamCollections(signaling_->remote_streams(), reference_collection_)); EXPECT_TRUE(CompareStreamCollections(observer_->remote_streams(), reference_collection_)); // Remove the extra audio and video tracks again. rtc::scoped_ptr desc_ms2; CreateSessionDescriptionAndReference(1, 1, desc_ms2.use()); signaling_->OnRemoteDescriptionChanged(desc_ms2.get()); EXPECT_TRUE(CompareStreamCollections(signaling_->remote_streams(), reference_collection_)); EXPECT_TRUE(CompareStreamCollections(observer_->remote_streams(), reference_collection_)); } // This test that remote tracks are ended if a // local session description is set that rejects the media content type. TEST_F(MediaStreamSignalingTest, RejectMediaContent) { rtc::scoped_ptr desc( webrtc::CreateSessionDescription(SessionDescriptionInterface::kOffer, kSdpStringWithStream1, NULL)); EXPECT_TRUE(desc != NULL); signaling_->OnRemoteDescriptionChanged(desc.get()); ASSERT_EQ(1u, observer_->remote_streams()->count()); MediaStreamInterface* remote_stream = observer_->remote_streams()->at(0); ASSERT_EQ(1u, remote_stream->GetVideoTracks().size()); ASSERT_EQ(1u, remote_stream->GetAudioTracks().size()); rtc::scoped_refptr remote_video = remote_stream->GetVideoTracks()[0]; EXPECT_EQ(webrtc::MediaStreamTrackInterface::kLive, remote_video->state()); rtc::scoped_refptr remote_audio = remote_stream->GetAudioTracks()[0]; EXPECT_EQ(webrtc::MediaStreamTrackInterface::kLive, remote_audio->state()); cricket::ContentInfo* video_info = desc->description()->GetContentByName("video"); ASSERT_TRUE(video_info != NULL); video_info->rejected = true; signaling_->OnLocalDescriptionChanged(desc.get()); EXPECT_EQ(webrtc::MediaStreamTrackInterface::kEnded, remote_video->state()); EXPECT_EQ(webrtc::MediaStreamTrackInterface::kLive, remote_audio->state()); cricket::ContentInfo* audio_info = desc->description()->GetContentByName("audio"); ASSERT_TRUE(audio_info != NULL); audio_info->rejected = true; signaling_->OnLocalDescriptionChanged(desc.get()); EXPECT_EQ(webrtc::MediaStreamTrackInterface::kEnded, remote_audio->state()); } // This test that it won't crash if the remote track as been removed outside // of MediaStreamSignaling and then MediaStreamSignaling tries to reject // this track. TEST_F(MediaStreamSignalingTest, RemoveTrackThenRejectMediaContent) { rtc::scoped_ptr desc( webrtc::CreateSessionDescription(SessionDescriptionInterface::kOffer, kSdpStringWithStream1, NULL)); EXPECT_TRUE(desc != NULL); signaling_->OnRemoteDescriptionChanged(desc.get()); MediaStreamInterface* remote_stream = observer_->remote_streams()->at(0); remote_stream->RemoveTrack(remote_stream->GetVideoTracks()[0]); remote_stream->RemoveTrack(remote_stream->GetAudioTracks()[0]); cricket::ContentInfo* video_info = desc->description()->GetContentByName("video"); video_info->rejected = true; signaling_->OnLocalDescriptionChanged(desc.get()); cricket::ContentInfo* audio_info = desc->description()->GetContentByName("audio"); audio_info->rejected = true; signaling_->OnLocalDescriptionChanged(desc.get()); // No crash is a pass. } // This tests that a default MediaStream is created if a remote session // description doesn't contain any streams and no MSID support. // It also tests that the default stream is updated if a video m-line is added // in a subsequent session description. TEST_F(MediaStreamSignalingTest, SdpWithoutMsidCreatesDefaultStream) { rtc::scoped_ptr desc_audio_only( webrtc::CreateSessionDescription(SessionDescriptionInterface::kOffer, kSdpStringWithoutStreamsAudioOnly, NULL)); ASSERT_TRUE(desc_audio_only != NULL); signaling_->OnRemoteDescriptionChanged(desc_audio_only.get()); EXPECT_EQ(1u, signaling_->remote_streams()->count()); ASSERT_EQ(1u, observer_->remote_streams()->count()); MediaStreamInterface* remote_stream = observer_->remote_streams()->at(0); EXPECT_EQ(1u, remote_stream->GetAudioTracks().size()); EXPECT_EQ(0u, remote_stream->GetVideoTracks().size()); EXPECT_EQ("default", remote_stream->label()); rtc::scoped_ptr desc( webrtc::CreateSessionDescription(SessionDescriptionInterface::kOffer, kSdpStringWithoutStreams, NULL)); ASSERT_TRUE(desc != NULL); signaling_->OnRemoteDescriptionChanged(desc.get()); EXPECT_EQ(1u, signaling_->remote_streams()->count()); ASSERT_EQ(1u, remote_stream->GetAudioTracks().size()); EXPECT_EQ("defaulta0", remote_stream->GetAudioTracks()[0]->id()); ASSERT_EQ(1u, remote_stream->GetVideoTracks().size()); EXPECT_EQ("defaultv0", remote_stream->GetVideoTracks()[0]->id()); observer_->VerifyRemoteAudioTrack("default", "defaulta0", 0); observer_->VerifyRemoteVideoTrack("default", "defaultv0", 0); } // This tests that a default MediaStream is created if a remote session // description doesn't contain any streams and media direction is send only. TEST_F(MediaStreamSignalingTest, RecvOnlySdpWithoutMsidCreatesDefaultStream) { rtc::scoped_ptr desc( webrtc::CreateSessionDescription(SessionDescriptionInterface::kOffer, kSdpStringSendOnlyWithWithoutStreams, NULL)); ASSERT_TRUE(desc != NULL); signaling_->OnRemoteDescriptionChanged(desc.get()); EXPECT_EQ(1u, signaling_->remote_streams()->count()); ASSERT_EQ(1u, observer_->remote_streams()->count()); MediaStreamInterface* remote_stream = observer_->remote_streams()->at(0); EXPECT_EQ(1u, remote_stream->GetAudioTracks().size()); EXPECT_EQ(1u, remote_stream->GetVideoTracks().size()); EXPECT_EQ("default", remote_stream->label()); } // This tests that it won't crash when MediaStreamSignaling tries to remove // a remote track that as already been removed from the mediastream. TEST_F(MediaStreamSignalingTest, RemoveAlreadyGoneRemoteStream) { rtc::scoped_ptr desc_audio_only( webrtc::CreateSessionDescription(SessionDescriptionInterface::kOffer, kSdpStringWithoutStreams, NULL)); ASSERT_TRUE(desc_audio_only != NULL); signaling_->OnRemoteDescriptionChanged(desc_audio_only.get()); MediaStreamInterface* remote_stream = observer_->remote_streams()->at(0); remote_stream->RemoveTrack(remote_stream->GetAudioTracks()[0]); remote_stream->RemoveTrack(remote_stream->GetVideoTracks()[0]); rtc::scoped_ptr desc( webrtc::CreateSessionDescription(SessionDescriptionInterface::kOffer, kSdpStringWithoutStreams, NULL)); ASSERT_TRUE(desc != NULL); signaling_->OnRemoteDescriptionChanged(desc.get()); // No crash is a pass. } // This tests that a default MediaStream is created if the remote session // description doesn't contain any streams and don't contain an indication if // MSID is supported. TEST_F(MediaStreamSignalingTest, SdpWithoutMsidAndStreamsCreatesDefaultStream) { rtc::scoped_ptr desc( webrtc::CreateSessionDescription(SessionDescriptionInterface::kOffer, kSdpStringWithoutStreams, NULL)); ASSERT_TRUE(desc != NULL); signaling_->OnRemoteDescriptionChanged(desc.get()); ASSERT_EQ(1u, observer_->remote_streams()->count()); MediaStreamInterface* remote_stream = observer_->remote_streams()->at(0); EXPECT_EQ(1u, remote_stream->GetAudioTracks().size()); EXPECT_EQ(1u, remote_stream->GetVideoTracks().size()); } // This tests that a default MediaStream is not created if the remote session // description doesn't contain any streams but does support MSID. TEST_F(MediaStreamSignalingTest, SdpWitMsidDontCreatesDefaultStream) { rtc::scoped_ptr desc_msid_without_streams( webrtc::CreateSessionDescription(SessionDescriptionInterface::kOffer, kSdpStringWithMsidWithoutStreams, NULL)); signaling_->OnRemoteDescriptionChanged(desc_msid_without_streams.get()); EXPECT_EQ(0u, observer_->remote_streams()->count()); } // This test that a default MediaStream is not created if a remote session // description is updated to not have any MediaStreams. TEST_F(MediaStreamSignalingTest, VerifyDefaultStreamIsNotCreated) { rtc::scoped_ptr desc( webrtc::CreateSessionDescription(SessionDescriptionInterface::kOffer, kSdpStringWithStream1, NULL)); ASSERT_TRUE(desc != NULL); signaling_->OnRemoteDescriptionChanged(desc.get()); rtc::scoped_refptr reference( CreateStreamCollection(1)); EXPECT_TRUE(CompareStreamCollections(observer_->remote_streams(), reference.get())); rtc::scoped_ptr desc_without_streams( webrtc::CreateSessionDescription(SessionDescriptionInterface::kOffer, kSdpStringWithoutStreams, NULL)); signaling_->OnRemoteDescriptionChanged(desc_without_streams.get()); EXPECT_EQ(0u, observer_->remote_streams()->count()); } // This test that the correct MediaStreamSignalingObserver methods are called // when MediaStreamSignaling::OnLocalDescriptionChanged is called with an // updated local session description. TEST_F(MediaStreamSignalingTest, LocalDescriptionChanged) { rtc::scoped_ptr desc_1; CreateSessionDescriptionAndReference(2, 2, desc_1.use()); signaling_->AddLocalStream(reference_collection_->at(0)); signaling_->OnLocalDescriptionChanged(desc_1.get()); EXPECT_EQ(2u, observer_->NumberOfLocalAudioTracks()); EXPECT_EQ(2u, observer_->NumberOfLocalVideoTracks()); observer_->VerifyLocalAudioTrack(kStreams[0], kAudioTracks[0], 1); observer_->VerifyLocalVideoTrack(kStreams[0], kVideoTracks[0], 2); observer_->VerifyLocalAudioTrack(kStreams[0], kAudioTracks[1], 3); observer_->VerifyLocalVideoTrack(kStreams[0], kVideoTracks[1], 4); // Remove an audio and video track. rtc::scoped_ptr desc_2; CreateSessionDescriptionAndReference(1, 1, desc_2.use()); signaling_->OnLocalDescriptionChanged(desc_2.get()); EXPECT_EQ(1u, observer_->NumberOfLocalAudioTracks()); EXPECT_EQ(1u, observer_->NumberOfLocalVideoTracks()); observer_->VerifyLocalAudioTrack(kStreams[0], kAudioTracks[0], 1); observer_->VerifyLocalVideoTrack(kStreams[0], kVideoTracks[0], 2); } // This test that the correct MediaStreamSignalingObserver methods are called // when MediaStreamSignaling::AddLocalStream is called after // MediaStreamSignaling::OnLocalDescriptionChanged is called. TEST_F(MediaStreamSignalingTest, AddLocalStreamAfterLocalDescriptionChanged) { rtc::scoped_ptr desc_1; CreateSessionDescriptionAndReference(2, 2, desc_1.use()); signaling_->OnLocalDescriptionChanged(desc_1.get()); EXPECT_EQ(0u, observer_->NumberOfLocalAudioTracks()); EXPECT_EQ(0u, observer_->NumberOfLocalVideoTracks()); signaling_->AddLocalStream(reference_collection_->at(0)); EXPECT_EQ(2u, observer_->NumberOfLocalAudioTracks()); EXPECT_EQ(2u, observer_->NumberOfLocalVideoTracks()); observer_->VerifyLocalAudioTrack(kStreams[0], kAudioTracks[0], 1); observer_->VerifyLocalVideoTrack(kStreams[0], kVideoTracks[0], 2); observer_->VerifyLocalAudioTrack(kStreams[0], kAudioTracks[1], 3); observer_->VerifyLocalVideoTrack(kStreams[0], kVideoTracks[1], 4); } // This test that the correct MediaStreamSignalingObserver methods are called // if the ssrc on a local track is changed when // MediaStreamSignaling::OnLocalDescriptionChanged is called. TEST_F(MediaStreamSignalingTest, ChangeSsrcOnTrackInLocalSessionDescription) { rtc::scoped_ptr desc; CreateSessionDescriptionAndReference(1, 1, desc.use()); signaling_->AddLocalStream(reference_collection_->at(0)); signaling_->OnLocalDescriptionChanged(desc.get()); EXPECT_EQ(1u, observer_->NumberOfLocalAudioTracks()); EXPECT_EQ(1u, observer_->NumberOfLocalVideoTracks()); observer_->VerifyLocalAudioTrack(kStreams[0], kAudioTracks[0], 1); observer_->VerifyLocalVideoTrack(kStreams[0], kVideoTracks[0], 2); // Change the ssrc of the audio and video track. std::string sdp; desc->ToString(&sdp); std::string ssrc_org = "a=ssrc:1"; std::string ssrc_to = "a=ssrc:97"; rtc::replace_substrs(ssrc_org.c_str(), ssrc_org.length(), ssrc_to.c_str(), ssrc_to.length(), &sdp); ssrc_org = "a=ssrc:2"; ssrc_to = "a=ssrc:98"; rtc::replace_substrs(ssrc_org.c_str(), ssrc_org.length(), ssrc_to.c_str(), ssrc_to.length(), &sdp); rtc::scoped_ptr updated_desc( webrtc::CreateSessionDescription(SessionDescriptionInterface::kOffer, sdp, NULL)); signaling_->OnLocalDescriptionChanged(updated_desc.get()); EXPECT_EQ(1u, observer_->NumberOfLocalAudioTracks()); EXPECT_EQ(1u, observer_->NumberOfLocalVideoTracks()); observer_->VerifyLocalAudioTrack(kStreams[0], kAudioTracks[0], 97); observer_->VerifyLocalVideoTrack(kStreams[0], kVideoTracks[0], 98); } // This test that the correct MediaStreamSignalingObserver methods are called // if a new session description is set with the same tracks but they are now // sent on a another MediaStream. TEST_F(MediaStreamSignalingTest, SignalSameTracksInSeparateMediaStream) { rtc::scoped_ptr desc; CreateSessionDescriptionAndReference(1, 1, desc.use()); signaling_->AddLocalStream(reference_collection_->at(0)); signaling_->OnLocalDescriptionChanged(desc.get()); EXPECT_EQ(1u, observer_->NumberOfLocalAudioTracks()); EXPECT_EQ(1u, observer_->NumberOfLocalVideoTracks()); std::string stream_label_0 = kStreams[0]; observer_->VerifyLocalAudioTrack(stream_label_0, kAudioTracks[0], 1); observer_->VerifyLocalVideoTrack(stream_label_0, kVideoTracks[0], 2); // Add a new MediaStream but with the same tracks as in the first stream. std::string stream_label_1 = kStreams[1]; rtc::scoped_refptr stream_1( webrtc::MediaStream::Create(kStreams[1])); stream_1->AddTrack(reference_collection_->at(0)->GetVideoTracks()[0]); stream_1->AddTrack(reference_collection_->at(0)->GetAudioTracks()[0]); signaling_->AddLocalStream(stream_1); // Replace msid in the original SDP. std::string sdp; desc->ToString(&sdp); rtc::replace_substrs( kStreams[0], strlen(kStreams[0]), kStreams[1], strlen(kStreams[1]), &sdp); rtc::scoped_ptr updated_desc( webrtc::CreateSessionDescription(SessionDescriptionInterface::kOffer, sdp, NULL)); signaling_->OnLocalDescriptionChanged(updated_desc.get()); observer_->VerifyLocalAudioTrack(kStreams[1], kAudioTracks[0], 1); observer_->VerifyLocalVideoTrack(kStreams[1], kVideoTracks[0], 2); EXPECT_EQ(1u, observer_->NumberOfLocalAudioTracks()); EXPECT_EQ(1u, observer_->NumberOfLocalVideoTracks()); } // Verifies that an even SCTP id is allocated for SSL_CLIENT and an odd id for // SSL_SERVER. TEST_F(MediaStreamSignalingTest, SctpIdAllocationBasedOnRole) { int id; ASSERT_TRUE(signaling_->AllocateSctpSid(rtc::SSL_SERVER, &id)); EXPECT_EQ(1, id); ASSERT_TRUE(signaling_->AllocateSctpSid(rtc::SSL_CLIENT, &id)); EXPECT_EQ(0, id); ASSERT_TRUE(signaling_->AllocateSctpSid(rtc::SSL_SERVER, &id)); EXPECT_EQ(3, id); ASSERT_TRUE(signaling_->AllocateSctpSid(rtc::SSL_CLIENT, &id)); EXPECT_EQ(2, id); } // Verifies that SCTP ids of existing DataChannels are not reused. TEST_F(MediaStreamSignalingTest, SctpIdAllocationNoReuse) { int old_id = 1; AddDataChannel(cricket::DCT_SCTP, "a", old_id); int new_id; ASSERT_TRUE(signaling_->AllocateSctpSid(rtc::SSL_SERVER, &new_id)); EXPECT_NE(old_id, new_id); // Creates a DataChannel with id 0. old_id = 0; AddDataChannel(cricket::DCT_SCTP, "a", old_id); ASSERT_TRUE(signaling_->AllocateSctpSid(rtc::SSL_CLIENT, &new_id)); EXPECT_NE(old_id, new_id); } // Verifies that SCTP ids of removed DataChannels can be reused. TEST_F(MediaStreamSignalingTest, SctpIdReusedForRemovedDataChannel) { int odd_id = 1; int even_id = 0; AddDataChannel(cricket::DCT_SCTP, "a", odd_id); AddDataChannel(cricket::DCT_SCTP, "a", even_id); int allocated_id = -1; ASSERT_TRUE(signaling_->AllocateSctpSid(rtc::SSL_SERVER, &allocated_id)); EXPECT_EQ(odd_id + 2, allocated_id); AddDataChannel(cricket::DCT_SCTP, "a", allocated_id); ASSERT_TRUE(signaling_->AllocateSctpSid(rtc::SSL_CLIENT, &allocated_id)); EXPECT_EQ(even_id + 2, allocated_id); AddDataChannel(cricket::DCT_SCTP, "a", allocated_id); signaling_->RemoveSctpDataChannel(odd_id); signaling_->RemoveSctpDataChannel(even_id); // Verifies that removed DataChannel ids are reused. ASSERT_TRUE(signaling_->AllocateSctpSid(rtc::SSL_SERVER, &allocated_id)); EXPECT_EQ(odd_id, allocated_id); ASSERT_TRUE(signaling_->AllocateSctpSid(rtc::SSL_CLIENT, &allocated_id)); EXPECT_EQ(even_id, allocated_id); // Verifies that used higher DataChannel ids are not reused. ASSERT_TRUE(signaling_->AllocateSctpSid(rtc::SSL_SERVER, &allocated_id)); EXPECT_NE(odd_id + 2, allocated_id); ASSERT_TRUE(signaling_->AllocateSctpSid(rtc::SSL_CLIENT, &allocated_id)); EXPECT_NE(even_id + 2, allocated_id); } // Verifies that duplicated label is not allowed for RTP data channel. TEST_F(MediaStreamSignalingTest, RtpDuplicatedLabelNotAllowed) { AddDataChannel(cricket::DCT_RTP, "a", -1); webrtc::InternalDataChannelInit config; rtc::scoped_refptr data_channel = webrtc::DataChannel::Create( data_channel_provider_.get(), cricket::DCT_RTP, "a", config); ASSERT_TRUE(data_channel.get() != NULL); EXPECT_FALSE(signaling_->AddDataChannel(data_channel.get())); } // Verifies that duplicated label is allowed for SCTP data channel. TEST_F(MediaStreamSignalingTest, SctpDuplicatedLabelAllowed) { AddDataChannel(cricket::DCT_SCTP, "a", -1); AddDataChannel(cricket::DCT_SCTP, "a", -1); } // Verifies the correct configuration is used to create DataChannel from an OPEN // message. TEST_F(MediaStreamSignalingTest, CreateDataChannelFromOpenMessage) { FakeDataChannelFactory fake_factory(data_channel_provider_.get(), cricket::DCT_SCTP, signaling_.get()); signaling_->SetDataChannelFactory(&fake_factory); webrtc::DataChannelInit config; config.id = 1; rtc::Buffer payload; webrtc::WriteDataChannelOpenMessage("a", config, &payload); cricket::ReceiveDataParams params; params.ssrc = config.id; EXPECT_TRUE(signaling_->AddDataChannelFromOpenMessage(params, payload)); EXPECT_EQ(config.id, fake_factory.last_init().id); EXPECT_FALSE(fake_factory.last_init().negotiated); EXPECT_EQ(webrtc::InternalDataChannelInit::kAcker, fake_factory.last_init().open_handshake_role); } // Verifies that duplicated label from OPEN message is allowed. TEST_F(MediaStreamSignalingTest, DuplicatedLabelFromOpenMessageAllowed) { AddDataChannel(cricket::DCT_SCTP, "a", -1); FakeDataChannelFactory fake_factory(data_channel_provider_.get(), cricket::DCT_SCTP, signaling_.get()); signaling_->SetDataChannelFactory(&fake_factory); webrtc::DataChannelInit config; config.id = 0; rtc::Buffer payload; webrtc::WriteDataChannelOpenMessage("a", config, &payload); cricket::ReceiveDataParams params; params.ssrc = config.id; EXPECT_TRUE(signaling_->AddDataChannelFromOpenMessage(params, payload)); } // Verifies that a DataChannel closed remotely is closed locally. TEST_F(MediaStreamSignalingTest, SctpDataChannelClosedLocallyWhenClosedRemotely) { webrtc::InternalDataChannelInit config; config.id = 0; rtc::scoped_refptr data_channel = webrtc::DataChannel::Create( data_channel_provider_.get(), cricket::DCT_SCTP, "a", config); ASSERT_TRUE(data_channel.get() != NULL); EXPECT_EQ(webrtc::DataChannelInterface::kConnecting, data_channel->state()); EXPECT_TRUE(signaling_->AddDataChannel(data_channel.get())); signaling_->OnRemoteSctpDataChannelClosed(config.id); EXPECT_EQ(webrtc::DataChannelInterface::kClosed, data_channel->state()); } // Verifies that DataChannel added from OPEN message is added to // MediaStreamSignaling only once (webrtc issue 3778). TEST_F(MediaStreamSignalingTest, DataChannelFromOpenMessageAddedOnce) { FakeDataChannelFactory fake_factory(data_channel_provider_.get(), cricket::DCT_SCTP, signaling_.get()); signaling_->SetDataChannelFactory(&fake_factory); webrtc::DataChannelInit config; config.id = 1; rtc::Buffer payload; webrtc::WriteDataChannelOpenMessage("a", config, &payload); cricket::ReceiveDataParams params; params.ssrc = config.id; EXPECT_TRUE(signaling_->AddDataChannelFromOpenMessage(params, payload)); EXPECT_TRUE(signaling_->HasDataChannels()); // Removes the DataChannel and verifies that no DataChannel is left. signaling_->RemoveSctpDataChannel(config.id); EXPECT_FALSE(signaling_->HasDataChannels()); }