/* * libjingle * Copyright 2012, Google Inc. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * 1. Redistributions of source code must retain the above copyright notice, * this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright notice, * this list of conditions and the following disclaimer in the documentation * and/or other materials provided with the distribution. * 3. The name of the author may not be used to endorse or promote products * derived from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR IMPLIED * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO * EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR * OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include "talk/app/webrtc/mediastreamsignaling.h" #include #include "talk/app/webrtc/audiotrack.h" #include "talk/app/webrtc/mediaconstraintsinterface.h" #include "talk/app/webrtc/mediastreamproxy.h" #include "talk/app/webrtc/mediastreamtrackproxy.h" #include "talk/app/webrtc/remoteaudiosource.h" #include "talk/app/webrtc/remotevideocapturer.h" #include "talk/app/webrtc/sctputils.h" #include "talk/app/webrtc/videosource.h" #include "talk/app/webrtc/videotrack.h" #include "talk/media/sctp/sctpdataengine.h" #include "webrtc/base/bytebuffer.h" #include "webrtc/base/stringutils.h" static const char kDefaultStreamLabel[] = "default"; static const char kDefaultAudioTrackLabel[] = "defaulta0"; static const char kDefaultVideoTrackLabel[] = "defaultv0"; namespace webrtc { using rtc::scoped_ptr; using rtc::scoped_refptr; static bool ParseConstraintsForAnswer( const MediaConstraintsInterface* constraints, cricket::MediaSessionOptions* options) { bool value; size_t mandatory_constraints_satisfied = 0; // kOfferToReceiveAudio defaults to true according to spec. if (!FindConstraint(constraints, MediaConstraintsInterface::kOfferToReceiveAudio, &value, &mandatory_constraints_satisfied) || value) { options->recv_audio = true; } // kOfferToReceiveVideo defaults to false according to spec. But // if it is an answer and video is offered, we should still accept video // per default. if (!FindConstraint(constraints, MediaConstraintsInterface::kOfferToReceiveVideo, &value, &mandatory_constraints_satisfied) || value) { options->recv_video = true; } if (FindConstraint(constraints, MediaConstraintsInterface::kVoiceActivityDetection, &value, &mandatory_constraints_satisfied)) { options->vad_enabled = value; } if (FindConstraint(constraints, MediaConstraintsInterface::kUseRtpMux, &value, &mandatory_constraints_satisfied)) { options->bundle_enabled = value; } else { // kUseRtpMux defaults to true according to spec. options->bundle_enabled = true; } if (FindConstraint(constraints, MediaConstraintsInterface::kIceRestart, &value, &mandatory_constraints_satisfied)) { options->transport_options.ice_restart = value; } else { // kIceRestart defaults to false according to spec. options->transport_options.ice_restart = false; } if (!constraints) { return true; } return mandatory_constraints_satisfied == constraints->GetMandatory().size(); } // Returns true if if at least one media content is present and // |options.bundle_enabled| is true. // Bundle will be enabled by default if at least one media content is present // and the constraint kUseRtpMux has not disabled bundle. static bool EvaluateNeedForBundle(const cricket::MediaSessionOptions& options) { return options.bundle_enabled && (options.has_audio() || options.has_video() || options.has_data()); } static bool MediaContentDirectionHasSend(cricket::MediaContentDirection dir) { return dir == cricket::MD_SENDONLY || dir == cricket::MD_SENDRECV; } static bool IsValidOfferToReceiveMedia(int value) { typedef PeerConnectionInterface::RTCOfferAnswerOptions Options; return (value >= Options::kUndefined) && (value <= Options::kMaxOfferToReceiveMedia); } // Add the stream and RTP data channel info to |session_options|. static void SetStreams( cricket::MediaSessionOptions* session_options, rtc::scoped_refptr streams, const MediaStreamSignaling::RtpDataChannels& rtp_data_channels) { session_options->streams.clear(); if (streams != NULL) { for (size_t i = 0; i < streams->count(); ++i) { MediaStreamInterface* stream = streams->at(i); AudioTrackVector audio_tracks(stream->GetAudioTracks()); // For each audio track in the stream, add it to the MediaSessionOptions. for (size_t j = 0; j < audio_tracks.size(); ++j) { scoped_refptr track(audio_tracks[j]); session_options->AddSendStream( cricket::MEDIA_TYPE_AUDIO, track->id(), stream->label()); } VideoTrackVector video_tracks(stream->GetVideoTracks()); // For each video track in the stream, add it to the MediaSessionOptions. for (size_t j = 0; j < video_tracks.size(); ++j) { scoped_refptr track(video_tracks[j]); session_options->AddSendStream( cricket::MEDIA_TYPE_VIDEO, track->id(), stream->label()); } } } // Check for data channels. MediaStreamSignaling::RtpDataChannels::const_iterator data_channel_it = rtp_data_channels.begin(); for (; data_channel_it != rtp_data_channels.end(); ++data_channel_it) { const DataChannel* channel = data_channel_it->second; if (channel->state() == DataChannel::kConnecting || channel->state() == DataChannel::kOpen) { // |streamid| and |sync_label| are both set to the DataChannel label // here so they can be signaled the same way as MediaStreams and Tracks. // For MediaStreams, the sync_label is the MediaStream label and the // track label is the same as |streamid|. const std::string& streamid = channel->label(); const std::string& sync_label = channel->label(); session_options->AddSendStream( cricket::MEDIA_TYPE_DATA, streamid, sync_label); } } } // Factory class for creating remote MediaStreams and MediaStreamTracks. class RemoteMediaStreamFactory { public: explicit RemoteMediaStreamFactory(rtc::Thread* signaling_thread, cricket::ChannelManager* channel_manager) : signaling_thread_(signaling_thread), channel_manager_(channel_manager) { } rtc::scoped_refptr CreateMediaStream( const std::string& stream_label) { return MediaStreamProxy::Create( signaling_thread_, MediaStream::Create(stream_label)); } AudioTrackInterface* AddAudioTrack(webrtc::MediaStreamInterface* stream, const std::string& track_id) { return AddTrack( stream, track_id, RemoteAudioSource::Create().get()); } VideoTrackInterface* AddVideoTrack(webrtc::MediaStreamInterface* stream, const std::string& track_id) { return AddTrack( stream, track_id, VideoSource::Create(channel_manager_, new RemoteVideoCapturer(), NULL).get()); } private: template TI* AddTrack(MediaStreamInterface* stream, const std::string& track_id, S* source) { rtc::scoped_refptr track( TP::Create(signaling_thread_, T::Create(track_id, source))); track->set_state(webrtc::MediaStreamTrackInterface::kLive); if (stream->AddTrack(track)) { return track; } return NULL; } rtc::Thread* signaling_thread_; cricket::ChannelManager* channel_manager_; }; MediaStreamSignaling::MediaStreamSignaling( rtc::Thread* signaling_thread, MediaStreamSignalingObserver* stream_observer, cricket::ChannelManager* channel_manager) : signaling_thread_(signaling_thread), data_channel_factory_(NULL), stream_observer_(stream_observer), local_streams_(StreamCollection::Create()), remote_streams_(StreamCollection::Create()), remote_stream_factory_(new RemoteMediaStreamFactory(signaling_thread, channel_manager)), last_allocated_sctp_even_sid_(-2), last_allocated_sctp_odd_sid_(-1) { } MediaStreamSignaling::~MediaStreamSignaling() { } void MediaStreamSignaling::TearDown() { OnAudioChannelClose(); OnVideoChannelClose(); OnDataChannelClose(); } bool MediaStreamSignaling::IsSctpSidAvailable(int sid) const { if (sid < 0 || sid > static_cast(cricket::kMaxSctpSid)) return false; return FindDataChannelBySid(sid) < 0; } // Gets the first unused odd/even id based on the DTLS role. If |role| is // SSL_CLIENT, the allocated id starts from 0 and takes even numbers; otherwise, // the id starts from 1 and takes odd numbers. Returns false if no id can be // allocated. bool MediaStreamSignaling::AllocateSctpSid(rtc::SSLRole role, int* sid) { int& last_id = (role == rtc::SSL_CLIENT) ? last_allocated_sctp_even_sid_ : last_allocated_sctp_odd_sid_; do { last_id += 2; } while (last_id <= static_cast(cricket::kMaxSctpSid) && !IsSctpSidAvailable(last_id)); if (last_id > static_cast(cricket::kMaxSctpSid)) { return false; } *sid = last_id; return true; } bool MediaStreamSignaling::HasDataChannels() const { return !rtp_data_channels_.empty() || !sctp_data_channels_.empty(); } bool MediaStreamSignaling::AddDataChannel(DataChannel* data_channel) { ASSERT(data_channel != NULL); if (data_channel->data_channel_type() == cricket::DCT_RTP) { if (rtp_data_channels_.find(data_channel->label()) != rtp_data_channels_.end()) { LOG(LS_ERROR) << "DataChannel with label " << data_channel->label() << " already exists."; return false; } rtp_data_channels_[data_channel->label()] = data_channel; } else { ASSERT(data_channel->data_channel_type() == cricket::DCT_SCTP); sctp_data_channels_.push_back(data_channel); } return true; } bool MediaStreamSignaling::AddDataChannelFromOpenMessage( const cricket::ReceiveDataParams& params, const rtc::Buffer& payload) { if (!data_channel_factory_) { LOG(LS_WARNING) << "Remote peer requested a DataChannel but DataChannels " << "are not supported."; return false; } std::string label; InternalDataChannelInit config; config.id = params.ssrc; if (!ParseDataChannelOpenMessage(payload, &label, &config)) { LOG(LS_WARNING) << "Failed to parse the OPEN message for sid " << params.ssrc; return false; } config.open_handshake_role = InternalDataChannelInit::kAcker; scoped_refptr channel( data_channel_factory_->CreateDataChannel(label, &config)); if (!channel.get()) { LOG(LS_ERROR) << "Failed to create DataChannel from the OPEN message."; return false; } stream_observer_->OnAddDataChannel(channel); return true; } void MediaStreamSignaling::RemoveSctpDataChannel(int sid) { ASSERT(sid >= 0); for (SctpDataChannels::iterator iter = sctp_data_channels_.begin(); iter != sctp_data_channels_.end(); ++iter) { if ((*iter)->id() == sid) { sctp_data_channels_.erase(iter); if (rtc::IsEven(sid) && sid <= last_allocated_sctp_even_sid_) { last_allocated_sctp_even_sid_ = sid - 2; } else if (rtc::IsOdd(sid) && sid <= last_allocated_sctp_odd_sid_) { last_allocated_sctp_odd_sid_ = sid - 2; } return; } } } bool MediaStreamSignaling::AddLocalStream(MediaStreamInterface* local_stream) { if (local_streams_->find(local_stream->label()) != NULL) { LOG(LS_WARNING) << "MediaStream with label " << local_stream->label() << "already exist."; return false; } local_streams_->AddStream(local_stream); // Find tracks that has already been configured in SDP. This can occur if a // local session description that contains the MSID of these tracks is set // before AddLocalStream is called. It can also occur if the local session // description is not changed and RemoveLocalStream // is called and later AddLocalStream is called again with the same stream. AudioTrackVector audio_tracks = local_stream->GetAudioTracks(); for (AudioTrackVector::const_iterator it = audio_tracks.begin(); it != audio_tracks.end(); ++it) { const TrackInfo* track_info = FindTrackInfo(local_audio_tracks_, local_stream->label(), (*it)->id()); if (track_info) { OnLocalTrackSeen(track_info->stream_label, track_info->track_id, track_info->ssrc, cricket::MEDIA_TYPE_AUDIO); } } VideoTrackVector video_tracks = local_stream->GetVideoTracks(); for (VideoTrackVector::const_iterator it = video_tracks.begin(); it != video_tracks.end(); ++it) { const TrackInfo* track_info = FindTrackInfo(local_video_tracks_, local_stream->label(), (*it)->id()); if (track_info) { OnLocalTrackSeen(track_info->stream_label, track_info->track_id, track_info->ssrc, cricket::MEDIA_TYPE_VIDEO); } } return true; } void MediaStreamSignaling::RemoveLocalStream( MediaStreamInterface* local_stream) { AudioTrackVector audio_tracks = local_stream->GetAudioTracks(); for (AudioTrackVector::const_iterator it = audio_tracks.begin(); it != audio_tracks.end(); ++it) { const TrackInfo* track_info = FindTrackInfo(local_audio_tracks_, local_stream->label(), (*it)->id()); if (track_info) { stream_observer_->OnRemoveLocalAudioTrack(local_stream, *it, track_info->ssrc); } } VideoTrackVector video_tracks = local_stream->GetVideoTracks(); for (VideoTrackVector::const_iterator it = video_tracks.begin(); it != video_tracks.end(); ++it) { const TrackInfo* track_info = FindTrackInfo(local_video_tracks_, local_stream->label(), (*it)->id()); if (track_info) { stream_observer_->OnRemoveLocalVideoTrack(local_stream, *it); } } local_streams_->RemoveStream(local_stream); stream_observer_->OnRemoveLocalStream(local_stream); } bool MediaStreamSignaling::GetOptionsForOffer( const PeerConnectionInterface::RTCOfferAnswerOptions& rtc_options, cricket::MediaSessionOptions* session_options) { typedef PeerConnectionInterface::RTCOfferAnswerOptions RTCOfferAnswerOptions; if (!IsValidOfferToReceiveMedia(rtc_options.offer_to_receive_audio) || !IsValidOfferToReceiveMedia(rtc_options.offer_to_receive_video)) { return false; } SetStreams(session_options, local_streams_, rtp_data_channels_); // According to the spec, offer to receive audio/video if the constraint is // not set and there are send streams. if (rtc_options.offer_to_receive_audio == RTCOfferAnswerOptions::kUndefined) { session_options->recv_audio = session_options->HasSendMediaStream(cricket::MEDIA_TYPE_AUDIO); } else { session_options->recv_audio = (rtc_options.offer_to_receive_audio > 0); } if (rtc_options.offer_to_receive_video == RTCOfferAnswerOptions::kUndefined) { session_options->recv_video = session_options->HasSendMediaStream(cricket::MEDIA_TYPE_VIDEO); } else { session_options->recv_video = (rtc_options.offer_to_receive_video > 0); } session_options->vad_enabled = rtc_options.voice_activity_detection; session_options->transport_options.ice_restart = rtc_options.ice_restart; session_options->bundle_enabled = rtc_options.use_rtp_mux; session_options->bundle_enabled = EvaluateNeedForBundle(*session_options); return true; } bool MediaStreamSignaling::GetOptionsForAnswer( const MediaConstraintsInterface* constraints, cricket::MediaSessionOptions* options) { SetStreams(options, local_streams_, rtp_data_channels_); options->recv_audio = false; options->recv_video = false; if (!ParseConstraintsForAnswer(constraints, options)) { return false; } options->bundle_enabled = EvaluateNeedForBundle(*options); return true; } // Updates or creates remote MediaStream objects given a // remote SessionDesription. // If the remote SessionDesription contains new remote MediaStreams // the observer OnAddStream method is called. If a remote MediaStream is missing // from the remote SessionDescription OnRemoveStream is called. void MediaStreamSignaling::OnRemoteDescriptionChanged( const SessionDescriptionInterface* desc) { const cricket::SessionDescription* remote_desc = desc->description(); rtc::scoped_refptr new_streams( StreamCollection::Create()); // Find all audio rtp streams and create corresponding remote AudioTracks // and MediaStreams. const cricket::ContentInfo* audio_content = GetFirstAudioContent(remote_desc); if (audio_content) { const cricket::AudioContentDescription* desc = static_cast( audio_content->description); UpdateRemoteStreamsList(desc->streams(), desc->type(), new_streams); remote_info_.default_audio_track_needed = MediaContentDirectionHasSend(desc->direction()) && desc->streams().empty(); } // Find all video rtp streams and create corresponding remote VideoTracks // and MediaStreams. const cricket::ContentInfo* video_content = GetFirstVideoContent(remote_desc); if (video_content) { const cricket::VideoContentDescription* desc = static_cast( video_content->description); UpdateRemoteStreamsList(desc->streams(), desc->type(), new_streams); remote_info_.default_video_track_needed = MediaContentDirectionHasSend(desc->direction()) && desc->streams().empty(); } // Update the DataChannels with the information from the remote peer. const cricket::ContentInfo* data_content = GetFirstDataContent(remote_desc); if (data_content) { const cricket::DataContentDescription* data_desc = static_cast( data_content->description); if (rtc::starts_with( data_desc->protocol().data(), cricket::kMediaProtocolRtpPrefix)) { UpdateRemoteRtpDataChannels(data_desc->streams()); } } // Iterate new_streams and notify the observer about new MediaStreams. for (size_t i = 0; i < new_streams->count(); ++i) { MediaStreamInterface* new_stream = new_streams->at(i); stream_observer_->OnAddRemoteStream(new_stream); } // Find removed MediaStreams. if (remote_info_.IsDefaultMediaStreamNeeded() && remote_streams_->find(kDefaultStreamLabel) != NULL) { // The default media stream already exists. No need to do anything. } else { UpdateEndedRemoteMediaStreams(); remote_info_.msid_supported |= remote_streams_->count() > 0; } MaybeCreateDefaultStream(); } void MediaStreamSignaling::OnLocalDescriptionChanged( const SessionDescriptionInterface* desc) { const cricket::ContentInfo* audio_content = GetFirstAudioContent(desc->description()); if (audio_content) { if (audio_content->rejected) { RejectRemoteTracks(cricket::MEDIA_TYPE_AUDIO); } const cricket::AudioContentDescription* audio_desc = static_cast( audio_content->description); UpdateLocalTracks(audio_desc->streams(), audio_desc->type()); } const cricket::ContentInfo* video_content = GetFirstVideoContent(desc->description()); if (video_content) { if (video_content->rejected) { RejectRemoteTracks(cricket::MEDIA_TYPE_VIDEO); } const cricket::VideoContentDescription* video_desc = static_cast( video_content->description); UpdateLocalTracks(video_desc->streams(), video_desc->type()); } const cricket::ContentInfo* data_content = GetFirstDataContent(desc->description()); if (data_content) { const cricket::DataContentDescription* data_desc = static_cast( data_content->description); if (rtc::starts_with( data_desc->protocol().data(), cricket::kMediaProtocolRtpPrefix)) { UpdateLocalRtpDataChannels(data_desc->streams()); } } } void MediaStreamSignaling::OnAudioChannelClose() { RejectRemoteTracks(cricket::MEDIA_TYPE_AUDIO); } void MediaStreamSignaling::OnVideoChannelClose() { RejectRemoteTracks(cricket::MEDIA_TYPE_VIDEO); } void MediaStreamSignaling::OnDataChannelClose() { // Use a temporary copy of the RTP/SCTP DataChannel list because the // DataChannel may callback to us and try to modify the list. RtpDataChannels temp_rtp_dcs; temp_rtp_dcs.swap(rtp_data_channels_); RtpDataChannels::iterator it1 = temp_rtp_dcs.begin(); for (; it1 != temp_rtp_dcs.end(); ++it1) { it1->second->OnDataEngineClose(); } SctpDataChannels temp_sctp_dcs; temp_sctp_dcs.swap(sctp_data_channels_); SctpDataChannels::iterator it2 = temp_sctp_dcs.begin(); for (; it2 != temp_sctp_dcs.end(); ++it2) { (*it2)->OnDataEngineClose(); } } void MediaStreamSignaling::UpdateRemoteStreamsList( const cricket::StreamParamsVec& streams, cricket::MediaType media_type, StreamCollection* new_streams) { TrackInfos* current_tracks = GetRemoteTracks(media_type); // Find removed tracks. Ie tracks where the track id or ssrc don't match the // new StreamParam. TrackInfos::iterator track_it = current_tracks->begin(); while (track_it != current_tracks->end()) { const TrackInfo& info = *track_it; cricket::StreamParams params; if (!cricket::GetStreamBySsrc(streams, info.ssrc, ¶ms) || params.id != info.track_id) { OnRemoteTrackRemoved(info.stream_label, info.track_id, media_type); track_it = current_tracks->erase(track_it); } else { ++track_it; } } // Find new and active tracks. for (cricket::StreamParamsVec::const_iterator it = streams.begin(); it != streams.end(); ++it) { // The sync_label is the MediaStream label and the |stream.id| is the // track id. const std::string& stream_label = it->sync_label; const std::string& track_id = it->id; uint32 ssrc = it->first_ssrc(); rtc::scoped_refptr stream = remote_streams_->find(stream_label); if (!stream) { // This is a new MediaStream. Create a new remote MediaStream. stream = remote_stream_factory_->CreateMediaStream(stream_label); remote_streams_->AddStream(stream); new_streams->AddStream(stream); } const TrackInfo* track_info = FindTrackInfo(*current_tracks, stream_label, track_id); if (!track_info) { current_tracks->push_back(TrackInfo(stream_label, track_id, ssrc)); OnRemoteTrackSeen(stream_label, track_id, it->first_ssrc(), media_type); } } } void MediaStreamSignaling::OnRemoteTrackSeen(const std::string& stream_label, const std::string& track_id, uint32 ssrc, cricket::MediaType media_type) { MediaStreamInterface* stream = remote_streams_->find(stream_label); if (media_type == cricket::MEDIA_TYPE_AUDIO) { AudioTrackInterface* audio_track = remote_stream_factory_->AddAudioTrack(stream, track_id); stream_observer_->OnAddRemoteAudioTrack(stream, audio_track, ssrc); } else if (media_type == cricket::MEDIA_TYPE_VIDEO) { VideoTrackInterface* video_track = remote_stream_factory_->AddVideoTrack(stream, track_id); stream_observer_->OnAddRemoteVideoTrack(stream, video_track, ssrc); } else { ASSERT(false && "Invalid media type"); } } void MediaStreamSignaling::OnRemoteTrackRemoved( const std::string& stream_label, const std::string& track_id, cricket::MediaType media_type) { MediaStreamInterface* stream = remote_streams_->find(stream_label); if (media_type == cricket::MEDIA_TYPE_AUDIO) { rtc::scoped_refptr audio_track = stream->FindAudioTrack(track_id); if (audio_track) { audio_track->set_state(webrtc::MediaStreamTrackInterface::kEnded); stream->RemoveTrack(audio_track); stream_observer_->OnRemoveRemoteAudioTrack(stream, audio_track); } } else if (media_type == cricket::MEDIA_TYPE_VIDEO) { rtc::scoped_refptr video_track = stream->FindVideoTrack(track_id); if (video_track) { video_track->set_state(webrtc::MediaStreamTrackInterface::kEnded); stream->RemoveTrack(video_track); stream_observer_->OnRemoveRemoteVideoTrack(stream, video_track); } } else { ASSERT(false && "Invalid media type"); } } void MediaStreamSignaling::RejectRemoteTracks(cricket::MediaType media_type) { TrackInfos* current_tracks = GetRemoteTracks(media_type); for (TrackInfos::iterator track_it = current_tracks->begin(); track_it != current_tracks->end(); ++track_it) { const TrackInfo& info = *track_it; MediaStreamInterface* stream = remote_streams_->find(info.stream_label); if (media_type == cricket::MEDIA_TYPE_AUDIO) { AudioTrackInterface* track = stream->FindAudioTrack(info.track_id); // There's no guarantee the track is still available, e.g. the track may // have been removed from the stream by javascript. if (track) { track->set_state(webrtc::MediaStreamTrackInterface::kEnded); } } if (media_type == cricket::MEDIA_TYPE_VIDEO) { VideoTrackInterface* track = stream->FindVideoTrack(info.track_id); // There's no guarantee the track is still available, e.g. the track may // have been removed from the stream by javascript. if (track) { track->set_state(webrtc::MediaStreamTrackInterface::kEnded); } } } } void MediaStreamSignaling::UpdateEndedRemoteMediaStreams() { std::vector > streams_to_remove; for (size_t i = 0; i < remote_streams_->count(); ++i) { MediaStreamInterface*stream = remote_streams_->at(i); if (stream->GetAudioTracks().empty() && stream->GetVideoTracks().empty()) { streams_to_remove.push_back(stream); } } std::vector >::const_iterator it; for (it = streams_to_remove.begin(); it != streams_to_remove.end(); ++it) { remote_streams_->RemoveStream(*it); stream_observer_->OnRemoveRemoteStream(*it); } } void MediaStreamSignaling::MaybeCreateDefaultStream() { if (!remote_info_.IsDefaultMediaStreamNeeded()) return; bool default_created = false; scoped_refptr default_remote_stream = remote_streams_->find(kDefaultStreamLabel); if (default_remote_stream == NULL) { default_created = true; default_remote_stream = remote_stream_factory_->CreateMediaStream(kDefaultStreamLabel); remote_streams_->AddStream(default_remote_stream); } if (remote_info_.default_audio_track_needed && default_remote_stream->GetAudioTracks().size() == 0) { remote_audio_tracks_.push_back(TrackInfo(kDefaultStreamLabel, kDefaultAudioTrackLabel, 0)); OnRemoteTrackSeen(kDefaultStreamLabel, kDefaultAudioTrackLabel, 0, cricket::MEDIA_TYPE_AUDIO); } if (remote_info_.default_video_track_needed && default_remote_stream->GetVideoTracks().size() == 0) { remote_video_tracks_.push_back(TrackInfo(kDefaultStreamLabel, kDefaultVideoTrackLabel, 0)); OnRemoteTrackSeen(kDefaultStreamLabel, kDefaultVideoTrackLabel, 0, cricket::MEDIA_TYPE_VIDEO); } if (default_created) { stream_observer_->OnAddRemoteStream(default_remote_stream); } } MediaStreamSignaling::TrackInfos* MediaStreamSignaling::GetRemoteTracks( cricket::MediaType type) { if (type == cricket::MEDIA_TYPE_AUDIO) return &remote_audio_tracks_; else if (type == cricket::MEDIA_TYPE_VIDEO) return &remote_video_tracks_; ASSERT(false && "Unknown MediaType"); return NULL; } MediaStreamSignaling::TrackInfos* MediaStreamSignaling::GetLocalTracks( cricket::MediaType media_type) { ASSERT(media_type == cricket::MEDIA_TYPE_AUDIO || media_type == cricket::MEDIA_TYPE_VIDEO); return (media_type == cricket::MEDIA_TYPE_AUDIO) ? &local_audio_tracks_ : &local_video_tracks_; } void MediaStreamSignaling::UpdateLocalTracks( const std::vector& streams, cricket::MediaType media_type) { TrackInfos* current_tracks = GetLocalTracks(media_type); // Find removed tracks. Ie tracks where the track id, stream label or ssrc // don't match the new StreamParam. TrackInfos::iterator track_it = current_tracks->begin(); while (track_it != current_tracks->end()) { const TrackInfo& info = *track_it; cricket::StreamParams params; if (!cricket::GetStreamBySsrc(streams, info.ssrc, ¶ms) || params.id != info.track_id || params.sync_label != info.stream_label) { OnLocalTrackRemoved(info.stream_label, info.track_id, info.ssrc, media_type); track_it = current_tracks->erase(track_it); } else { ++track_it; } } // Find new and active tracks. for (cricket::StreamParamsVec::const_iterator it = streams.begin(); it != streams.end(); ++it) { // The sync_label is the MediaStream label and the |stream.id| is the // track id. const std::string& stream_label = it->sync_label; const std::string& track_id = it->id; uint32 ssrc = it->first_ssrc(); const TrackInfo* track_info = FindTrackInfo(*current_tracks, stream_label, track_id); if (!track_info) { current_tracks->push_back(TrackInfo(stream_label, track_id, ssrc)); OnLocalTrackSeen(stream_label, track_id, it->first_ssrc(), media_type); } } } void MediaStreamSignaling::OnLocalTrackSeen( const std::string& stream_label, const std::string& track_id, uint32 ssrc, cricket::MediaType media_type) { MediaStreamInterface* stream = local_streams_->find(stream_label); if (!stream) { LOG(LS_WARNING) << "An unknown local MediaStream with label " << stream_label << " has been configured."; return; } if (media_type == cricket::MEDIA_TYPE_AUDIO) { AudioTrackInterface* audio_track = stream->FindAudioTrack(track_id); if (!audio_track) { LOG(LS_WARNING) << "An unknown local AudioTrack with id , " << track_id << " has been configured."; return; } stream_observer_->OnAddLocalAudioTrack(stream, audio_track, ssrc); } else if (media_type == cricket::MEDIA_TYPE_VIDEO) { VideoTrackInterface* video_track = stream->FindVideoTrack(track_id); if (!video_track) { LOG(LS_WARNING) << "An unknown local VideoTrack with id , " << track_id << " has been configured."; return; } stream_observer_->OnAddLocalVideoTrack(stream, video_track, ssrc); } else { ASSERT(false && "Invalid media type"); } } void MediaStreamSignaling::OnLocalTrackRemoved( const std::string& stream_label, const std::string& track_id, uint32 ssrc, cricket::MediaType media_type) { MediaStreamInterface* stream = local_streams_->find(stream_label); if (!stream) { // This is the normal case. Ie RemoveLocalStream has been called and the // SessionDescriptions has been renegotiated. return; } // A track has been removed from the SessionDescription but the MediaStream // is still associated with MediaStreamSignaling. This only occurs if the SDP // doesn't match with the calls to AddLocalStream and RemoveLocalStream. if (media_type == cricket::MEDIA_TYPE_AUDIO) { AudioTrackInterface* audio_track = stream->FindAudioTrack(track_id); if (!audio_track) { return; } stream_observer_->OnRemoveLocalAudioTrack(stream, audio_track, ssrc); } else if (media_type == cricket::MEDIA_TYPE_VIDEO) { VideoTrackInterface* video_track = stream->FindVideoTrack(track_id); if (!video_track) { return; } stream_observer_->OnRemoveLocalVideoTrack(stream, video_track); } else { ASSERT(false && "Invalid media type."); } } void MediaStreamSignaling::UpdateLocalRtpDataChannels( const cricket::StreamParamsVec& streams) { std::vector existing_channels; // Find new and active data channels. for (cricket::StreamParamsVec::const_iterator it =streams.begin(); it != streams.end(); ++it) { // |it->sync_label| is actually the data channel label. The reason is that // we use the same naming of data channels as we do for // MediaStreams and Tracks. // For MediaStreams, the sync_label is the MediaStream label and the // track label is the same as |streamid|. const std::string& channel_label = it->sync_label; RtpDataChannels::iterator data_channel_it = rtp_data_channels_.find(channel_label); if (!VERIFY(data_channel_it != rtp_data_channels_.end())) { continue; } // Set the SSRC the data channel should use for sending. data_channel_it->second->SetSendSsrc(it->first_ssrc()); existing_channels.push_back(data_channel_it->first); } UpdateClosingDataChannels(existing_channels, true); } void MediaStreamSignaling::UpdateRemoteRtpDataChannels( const cricket::StreamParamsVec& streams) { std::vector existing_channels; // Find new and active data channels. for (cricket::StreamParamsVec::const_iterator it = streams.begin(); it != streams.end(); ++it) { // The data channel label is either the mslabel or the SSRC if the mslabel // does not exist. Ex a=ssrc:444330170 mslabel:test1. std::string label = it->sync_label.empty() ? rtc::ToString(it->first_ssrc()) : it->sync_label; RtpDataChannels::iterator data_channel_it = rtp_data_channels_.find(label); if (data_channel_it == rtp_data_channels_.end()) { // This is a new data channel. CreateRemoteDataChannel(label, it->first_ssrc()); } else { data_channel_it->second->SetReceiveSsrc(it->first_ssrc()); } existing_channels.push_back(label); } UpdateClosingDataChannels(existing_channels, false); } void MediaStreamSignaling::UpdateClosingDataChannels( const std::vector& active_channels, bool is_local_update) { RtpDataChannels::iterator it = rtp_data_channels_.begin(); while (it != rtp_data_channels_.end()) { DataChannel* data_channel = it->second; if (std::find(active_channels.begin(), active_channels.end(), data_channel->label()) != active_channels.end()) { ++it; continue; } if (is_local_update) data_channel->SetSendSsrc(0); else data_channel->RemotePeerRequestClose(); if (data_channel->state() == DataChannel::kClosed) { rtp_data_channels_.erase(it); it = rtp_data_channels_.begin(); } else { ++it; } } } void MediaStreamSignaling::CreateRemoteDataChannel(const std::string& label, uint32 remote_ssrc) { if (!data_channel_factory_) { LOG(LS_WARNING) << "Remote peer requested a DataChannel but DataChannels " << "are not supported."; return; } scoped_refptr channel( data_channel_factory_->CreateDataChannel(label, NULL)); if (!channel.get()) { LOG(LS_WARNING) << "Remote peer requested a DataChannel but" << "CreateDataChannel failed."; return; } channel->SetReceiveSsrc(remote_ssrc); stream_observer_->OnAddDataChannel(channel); } void MediaStreamSignaling::OnDataTransportCreatedForSctp() { SctpDataChannels::iterator it = sctp_data_channels_.begin(); for (; it != sctp_data_channels_.end(); ++it) { (*it)->OnTransportChannelCreated(); } } void MediaStreamSignaling::OnDtlsRoleReadyForSctp(rtc::SSLRole role) { SctpDataChannels::iterator it = sctp_data_channels_.begin(); for (; it != sctp_data_channels_.end(); ++it) { if ((*it)->id() < 0) { int sid; if (!AllocateSctpSid(role, &sid)) { LOG(LS_ERROR) << "Failed to allocate SCTP sid."; continue; } (*it)->SetSctpSid(sid); } } } void MediaStreamSignaling::OnRemoteSctpDataChannelClosed(uint32 sid) { int index = FindDataChannelBySid(sid); if (index < 0) { LOG(LS_WARNING) << "Unexpected sid " << sid << " of the remotely closed DataChannel."; return; } sctp_data_channels_[index]->Close(); } const MediaStreamSignaling::TrackInfo* MediaStreamSignaling::FindTrackInfo( const MediaStreamSignaling::TrackInfos& infos, const std::string& stream_label, const std::string track_id) const { for (TrackInfos::const_iterator it = infos.begin(); it != infos.end(); ++it) { if (it->stream_label == stream_label && it->track_id == track_id) return &*it; } return NULL; } int MediaStreamSignaling::FindDataChannelBySid(int sid) const { for (size_t i = 0; i < sctp_data_channels_.size(); ++i) { if (sctp_data_channels_[i]->id() == sid) { return static_cast(i); } } return -1; } } // namespace webrtc