3352 lines
92 KiB
C++
3352 lines
92 KiB
C++
/*
|
|
This file is part of Telegram Desktop,
|
|
the official desktop application for the Telegram messaging service.
|
|
|
|
For license and copyright information please follow this link:
|
|
https://github.com/telegramdesktop/tdesktop/blob/master/LEGAL
|
|
*/
|
|
#include "calls/group/calls_group_call.h"
|
|
|
|
#include "calls/group/calls_group_common.h"
|
|
#include "main/main_session.h"
|
|
#include "api/api_send_progress.h"
|
|
#include "api/api_updates.h"
|
|
#include "apiwrap.h"
|
|
#include "lang/lang_keys.h"
|
|
#include "lang/lang_hardcoded.h"
|
|
#include "boxes/peers/edit_participants_box.h" // SubscribeToMigration.
|
|
#include "ui/toasts/common_toasts.h"
|
|
#include "base/unixtime.h"
|
|
#include "core/application.h"
|
|
#include "core/core_settings.h"
|
|
#include "data/data_changes.h"
|
|
#include "data/data_user.h"
|
|
#include "data/data_chat.h"
|
|
#include "data/data_channel.h"
|
|
#include "data/data_group_call.h"
|
|
#include "data/data_peer_values.h"
|
|
#include "data/data_session.h"
|
|
#include "base/global_shortcuts.h"
|
|
#include "base/random.h"
|
|
#include "webrtc/webrtc_video_track.h"
|
|
#include "webrtc/webrtc_media_devices.h"
|
|
#include "webrtc/webrtc_create_adm.h"
|
|
|
|
#include <tgcalls/group/GroupInstanceCustomImpl.h>
|
|
#include <tgcalls/VideoCaptureInterface.h>
|
|
#include <tgcalls/StaticThreads.h>
|
|
#include <QtCore/QJsonDocument>
|
|
#include <QtCore/QJsonObject>
|
|
#include <QtCore/QJsonArray>
|
|
|
|
namespace Calls {
|
|
namespace {
|
|
|
|
constexpr auto kMaxInvitePerSlice = 10;
|
|
constexpr auto kCheckLastSpokeInterval = crl::time(1000);
|
|
constexpr auto kCheckJoinedTimeout = 4 * crl::time(1000);
|
|
constexpr auto kUpdateSendActionEach = crl::time(500);
|
|
constexpr auto kPlayConnectingEach = crl::time(1056) + 2 * crl::time(1000);
|
|
constexpr auto kFixManualLargeVideoDuration = 5 * crl::time(1000);
|
|
constexpr auto kFixSpeakingLargeVideoDuration = 3 * crl::time(1000);
|
|
constexpr auto kFullAsMediumsCount = 4; // 1 Full is like 4 Mediums.
|
|
constexpr auto kMaxMediumQualities = 16; // 4 Fulls or 16 Mediums.
|
|
|
|
[[nodiscard]] std::unique_ptr<Webrtc::MediaDevices> CreateMediaDevices() {
|
|
const auto &settings = Core::App().settings();
|
|
return Webrtc::CreateMediaDevices(
|
|
settings.callInputDeviceId(),
|
|
settings.callOutputDeviceId(),
|
|
settings.callVideoInputDeviceId());
|
|
}
|
|
|
|
[[nodiscard]] const Data::GroupCallParticipant *LookupParticipant(
|
|
not_null<PeerData*> peer,
|
|
uint64 id,
|
|
not_null<PeerData*> participantPeer) {
|
|
const auto call = peer->groupCall();
|
|
return (id && call && call->id() == id)
|
|
? call->participantByPeer(participantPeer)
|
|
: nullptr;
|
|
}
|
|
|
|
[[nodiscard]] double TimestampFromMsgId(mtpMsgId msgId) {
|
|
return msgId / double(1ULL << 32);
|
|
}
|
|
|
|
[[nodiscard]] int64 TimestampInMsFromMsgId(mtpMsgId msgId) {
|
|
// return (msgId * 1000) / (1ULL << 32); // Almost... But this overflows.
|
|
return ((msgId / (1ULL << 10)) * 1000) / (1ULL << 22);
|
|
}
|
|
|
|
[[nodiscard]] uint64 FindLocalRaisedHandRating(
|
|
const std::vector<Data::GroupCallParticipant> &list) {
|
|
const auto i = ranges::max_element(
|
|
list,
|
|
ranges::less(),
|
|
&Data::GroupCallParticipant::raisedHandRating);
|
|
return (i == end(list)) ? 1 : (i->raisedHandRating + 1);
|
|
}
|
|
|
|
struct JoinVideoEndpoint {
|
|
std::string id;
|
|
};
|
|
|
|
struct JoinBroadcastStream {
|
|
};
|
|
|
|
using JoinClientFields = std::variant<
|
|
v::null_t,
|
|
JoinVideoEndpoint,
|
|
JoinBroadcastStream>;
|
|
|
|
class RequestCurrentTimeTask final : public tgcalls::BroadcastPartTask {
|
|
public:
|
|
RequestCurrentTimeTask(
|
|
base::weak_ptr<GroupCall> call,
|
|
Fn<void(int64)> done);
|
|
|
|
void done(int64 value);
|
|
void cancel() override;
|
|
|
|
private:
|
|
const base::weak_ptr<GroupCall> _call;
|
|
Fn<void(int64)> _done;
|
|
QMutex _mutex;
|
|
|
|
};
|
|
|
|
RequestCurrentTimeTask::RequestCurrentTimeTask(
|
|
base::weak_ptr<GroupCall> call,
|
|
Fn<void(int64)> done)
|
|
: _call(call)
|
|
, _done(std::move(done)) {
|
|
}
|
|
|
|
void RequestCurrentTimeTask::done(int64 value) {
|
|
QMutexLocker lock(&_mutex);
|
|
if (_done) {
|
|
base::take(_done)(value);
|
|
}
|
|
}
|
|
|
|
void RequestCurrentTimeTask::cancel() {
|
|
QMutexLocker lock(&_mutex);
|
|
_done = nullptr;
|
|
}
|
|
|
|
[[nodiscard]] JoinClientFields ParseJoinResponse(const QByteArray &json) {
|
|
auto error = QJsonParseError{ 0, QJsonParseError::NoError };
|
|
const auto document = QJsonDocument::fromJson(json, &error);
|
|
if (error.error != QJsonParseError::NoError) {
|
|
LOG(("API Error: "
|
|
"Failed to parse join response params, error: %1."
|
|
).arg(error.errorString()));
|
|
return {};
|
|
} else if (!document.isObject()) {
|
|
LOG(("API Error: "
|
|
"Not an object received in join response params."));
|
|
return {};
|
|
}
|
|
if (document.object().value("stream").toBool()) {
|
|
return JoinBroadcastStream{};
|
|
}
|
|
const auto video = document.object().value("video").toObject();
|
|
return JoinVideoEndpoint{
|
|
video.value("endpoint").toString().toStdString(),
|
|
};
|
|
}
|
|
|
|
[[nodiscard]] const std::string &EmptyString() {
|
|
static const auto result = std::string();
|
|
return result;
|
|
}
|
|
|
|
} // namespace
|
|
|
|
class GroupCall::LoadPartTask final : public tgcalls::BroadcastPartTask {
|
|
public:
|
|
using Quality = tgcalls::VideoChannelDescription::Quality;
|
|
LoadPartTask(
|
|
base::weak_ptr<GroupCall> call,
|
|
int64 time,
|
|
int64 period,
|
|
Fn<void(tgcalls::BroadcastPart&&)> done);
|
|
LoadPartTask(
|
|
base::weak_ptr<GroupCall> call,
|
|
int64 time,
|
|
int64 period,
|
|
int32 videoChannel,
|
|
Quality videoQuality,
|
|
Fn<void(tgcalls::BroadcastPart&&)> done);
|
|
|
|
[[nodiscard]] int64 time() const {
|
|
return _time;
|
|
}
|
|
[[nodiscard]] int32 scale() const {
|
|
return _scale;
|
|
}
|
|
[[nodiscard]] int32 videoChannel() const {
|
|
return _videoChannel;
|
|
}
|
|
[[nodiscard]] Quality videoQuality() const {
|
|
return _videoQuality;
|
|
}
|
|
|
|
void done(tgcalls::BroadcastPart &&part);
|
|
void cancel() override;
|
|
|
|
private:
|
|
const base::weak_ptr<GroupCall> _call;
|
|
const int64 _time = 0;
|
|
const int32 _scale = 0;
|
|
const int32 _videoChannel = 0;
|
|
const Quality _videoQuality = {};
|
|
Fn<void(tgcalls::BroadcastPart &&)> _done;
|
|
QMutex _mutex;
|
|
|
|
};
|
|
|
|
class GroupCall::MediaChannelDescriptionsTask final
|
|
: public tgcalls::RequestMediaChannelDescriptionTask {
|
|
public:
|
|
MediaChannelDescriptionsTask(
|
|
base::weak_ptr<GroupCall> call,
|
|
const std::vector<std::uint32_t> &ssrcs,
|
|
Fn<void(std::vector<tgcalls::MediaChannelDescription>&&)> done);
|
|
|
|
[[nodiscard]] base::flat_set<uint32> ssrcs() const;
|
|
|
|
[[nodiscard]] bool finishWithAdding(
|
|
uint32 ssrc,
|
|
std::optional<tgcalls::MediaChannelDescription> description,
|
|
bool screen = false);
|
|
|
|
void cancel() override;
|
|
|
|
private:
|
|
const base::weak_ptr<GroupCall> _call;
|
|
base::flat_set<uint32> _ssrcs;
|
|
base::flat_set<uint32> _cameraAdded;
|
|
base::flat_set<uint32> _screenAdded;
|
|
std::vector<tgcalls::MediaChannelDescription> _result;
|
|
Fn<void(std::vector<tgcalls::MediaChannelDescription>&&)> _done;
|
|
QMutex _mutex;
|
|
|
|
};
|
|
|
|
struct GroupCall::SinkPointer {
|
|
std::weak_ptr<Webrtc::SinkInterface> data;
|
|
};
|
|
|
|
struct GroupCall::VideoTrack {
|
|
VideoTrack(bool paused, bool requireARGB32, not_null<PeerData*> peer);
|
|
|
|
Webrtc::VideoTrack track;
|
|
rpl::variable<QSize> trackSize;
|
|
not_null<PeerData*> peer;
|
|
rpl::lifetime lifetime;
|
|
Group::VideoQuality quality = Group::VideoQuality();
|
|
bool shown = false;
|
|
};
|
|
|
|
GroupCall::VideoTrack::VideoTrack(
|
|
bool paused,
|
|
bool requireARGB32,
|
|
not_null<PeerData*> peer)
|
|
: track((paused
|
|
? Webrtc::VideoState::Paused
|
|
: Webrtc::VideoState::Active),
|
|
requireARGB32)
|
|
, peer(peer) {
|
|
}
|
|
|
|
[[nodiscard]] bool IsGroupCallAdmin(
|
|
not_null<PeerData*> peer,
|
|
not_null<PeerData*> participantPeer) {
|
|
const auto user = participantPeer->asUser();
|
|
if (!user) {
|
|
return (peer == participantPeer);
|
|
}
|
|
if (const auto chat = peer->asChat()) {
|
|
return chat->admins.contains(user)
|
|
|| (chat->creator == peerToUser(user->id));
|
|
} else if (const auto group = peer->asChannel()) {
|
|
if (const auto mgInfo = group->mgInfo.get()) {
|
|
if (mgInfo->creator == user) {
|
|
return true;
|
|
}
|
|
const auto i = mgInfo->lastAdmins.find(user);
|
|
if (i == mgInfo->lastAdmins.end()) {
|
|
return false;
|
|
}
|
|
return (i->second.rights.flags & ChatAdminRight::ManageCall);
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
struct VideoParams {
|
|
std::string endpointId;
|
|
std::vector<tgcalls::MediaSsrcGroup> ssrcGroups;
|
|
uint32 additionalSsrc = 0;
|
|
bool paused = false;
|
|
|
|
[[nodiscard]] bool empty() const {
|
|
return !additionalSsrc && (endpointId.empty() || ssrcGroups.empty());
|
|
}
|
|
[[nodiscard]] explicit operator bool() const {
|
|
return !empty();
|
|
}
|
|
};
|
|
|
|
struct ParticipantVideoParams {
|
|
VideoParams camera;
|
|
VideoParams screen;
|
|
};
|
|
|
|
[[nodiscard]] bool VideoParamsAreEqual(
|
|
const VideoParams &was,
|
|
const tl::conditional<MTPGroupCallParticipantVideo> &now) {
|
|
if (!now) {
|
|
return !was;
|
|
}
|
|
return now->match([&](const MTPDgroupCallParticipantVideo &data) {
|
|
if (data.is_paused() != was.paused
|
|
|| data.vaudio_source().value_or_empty() != was.additionalSsrc) {
|
|
return false;
|
|
}
|
|
if (gsl::make_span(data.vendpoint().v)
|
|
!= gsl::make_span(was.endpointId)) {
|
|
return false;
|
|
}
|
|
const auto &list = data.vsource_groups().v;
|
|
if (list.size() != was.ssrcGroups.size()) {
|
|
return false;
|
|
}
|
|
auto index = 0;
|
|
for (const auto &group : list) {
|
|
const auto equal = group.match([&](
|
|
const MTPDgroupCallParticipantVideoSourceGroup &data) {
|
|
const auto &group = was.ssrcGroups[index++];
|
|
if (gsl::make_span(data.vsemantics().v)
|
|
!= gsl::make_span(group.semantics)) {
|
|
return false;
|
|
}
|
|
const auto list = data.vsources().v;
|
|
if (list.size() != group.ssrcs.size()) {
|
|
return false;
|
|
}
|
|
auto i = 0;
|
|
for (const auto &ssrc : list) {
|
|
if (ssrc.v != group.ssrcs[i++]) {
|
|
return false;
|
|
}
|
|
}
|
|
return true;
|
|
});
|
|
if (!equal) {
|
|
return false;
|
|
}
|
|
}
|
|
return true;
|
|
});
|
|
}
|
|
|
|
[[nodiscard]] VideoParams ParseVideoParams(
|
|
const tl::conditional<MTPGroupCallParticipantVideo> ¶ms) {
|
|
if (!params) {
|
|
return VideoParams();
|
|
}
|
|
auto result = VideoParams();
|
|
params->match([&](const MTPDgroupCallParticipantVideo &data) {
|
|
result.paused = data.is_paused();
|
|
result.endpointId = data.vendpoint().v.toStdString();
|
|
result.additionalSsrc = data.vaudio_source().value_or_empty();
|
|
const auto &list = data.vsource_groups().v;
|
|
result.ssrcGroups.reserve(list.size());
|
|
for (const auto &group : list) {
|
|
group.match([&](
|
|
const MTPDgroupCallParticipantVideoSourceGroup &data) {
|
|
const auto &list = data.vsources().v;
|
|
auto ssrcs = std::vector<uint32_t>();
|
|
ssrcs.reserve(list.size());
|
|
for (const auto &ssrc : list) {
|
|
ssrcs.push_back(ssrc.v);
|
|
}
|
|
result.ssrcGroups.push_back({
|
|
.semantics = data.vsemantics().v.toStdString(),
|
|
.ssrcs = std::move(ssrcs),
|
|
});
|
|
});
|
|
}
|
|
});
|
|
return result;
|
|
}
|
|
|
|
const std::string &GetCameraEndpoint(
|
|
const std::shared_ptr<ParticipantVideoParams> ¶ms) {
|
|
return params ? params->camera.endpointId : EmptyString();
|
|
}
|
|
|
|
const std::string &GetScreenEndpoint(
|
|
const std::shared_ptr<ParticipantVideoParams> ¶ms) {
|
|
return params ? params->screen.endpointId : EmptyString();
|
|
}
|
|
|
|
bool IsCameraPaused(const std::shared_ptr<ParticipantVideoParams> ¶ms) {
|
|
return params && params->camera.paused;
|
|
}
|
|
|
|
bool IsScreenPaused(const std::shared_ptr<ParticipantVideoParams> ¶ms) {
|
|
return params && params->screen.paused;
|
|
}
|
|
|
|
uint32 GetAdditionalAudioSsrc(
|
|
const std::shared_ptr<ParticipantVideoParams> ¶ms) {
|
|
return params ? params->screen.additionalSsrc : 0;
|
|
}
|
|
|
|
std::shared_ptr<ParticipantVideoParams> ParseVideoParams(
|
|
const tl::conditional<MTPGroupCallParticipantVideo> &camera,
|
|
const tl::conditional<MTPGroupCallParticipantVideo> &screen,
|
|
const std::shared_ptr<ParticipantVideoParams> &existing) {
|
|
using namespace tgcalls;
|
|
|
|
if (!camera && !screen) {
|
|
return nullptr;
|
|
}
|
|
if (existing
|
|
&& VideoParamsAreEqual(existing->camera, camera)
|
|
&& VideoParamsAreEqual(existing->screen, screen)) {
|
|
return existing;
|
|
}
|
|
// We don't reuse existing pointer, that way we can compare pointers
|
|
// to see if anything was changed in video params.
|
|
const auto data = /*existing
|
|
? existing
|
|
: */std::make_shared<ParticipantVideoParams>();
|
|
data->camera = ParseVideoParams(camera);
|
|
data->screen = ParseVideoParams(screen);
|
|
return data;
|
|
}
|
|
|
|
GroupCall::LoadPartTask::LoadPartTask(
|
|
base::weak_ptr<GroupCall> call,
|
|
int64 time,
|
|
int64 period,
|
|
Fn<void(tgcalls::BroadcastPart&&)> done)
|
|
: LoadPartTask(std::move(call), time, period, 0, {}, std::move(done)) {
|
|
}
|
|
|
|
GroupCall::LoadPartTask::LoadPartTask(
|
|
base::weak_ptr<GroupCall> call,
|
|
int64 time,
|
|
int64 period,
|
|
int32 videoChannel,
|
|
tgcalls::VideoChannelDescription::Quality videoQuality,
|
|
Fn<void(tgcalls::BroadcastPart&&)> done)
|
|
: _call(std::move(call))
|
|
, _time(time ? time : (base::unixtime::now() * int64(1000)))
|
|
, _scale([&] {
|
|
switch (period) {
|
|
case 1000: return 0;
|
|
case 500: return 1;
|
|
case 250: return 2;
|
|
case 125: return 3;
|
|
}
|
|
Unexpected("Period in LoadPartTask.");
|
|
}())
|
|
, _videoChannel(videoChannel)
|
|
, _videoQuality(videoQuality)
|
|
, _done(std::move(done)) {
|
|
}
|
|
|
|
void GroupCall::LoadPartTask::done(tgcalls::BroadcastPart &&part) {
|
|
QMutexLocker lock(&_mutex);
|
|
if (_done) {
|
|
base::take(_done)(std::move(part));
|
|
}
|
|
}
|
|
|
|
void GroupCall::LoadPartTask::cancel() {
|
|
QMutexLocker lock(&_mutex);
|
|
if (!_done) {
|
|
return;
|
|
}
|
|
_done = nullptr;
|
|
lock.unlock();
|
|
|
|
if (_call) {
|
|
const auto that = this;
|
|
crl::on_main(_call, [weak = _call, that] {
|
|
if (const auto strong = weak.get()) {
|
|
strong->broadcastPartCancel(that);
|
|
}
|
|
});
|
|
}
|
|
}
|
|
|
|
GroupCall::MediaChannelDescriptionsTask::MediaChannelDescriptionsTask(
|
|
base::weak_ptr<GroupCall> call,
|
|
const std::vector<std::uint32_t> &ssrcs,
|
|
Fn<void(std::vector<tgcalls::MediaChannelDescription>&&)> done)
|
|
: _call(std::move(call))
|
|
, _ssrcs(ssrcs.begin(), ssrcs.end())
|
|
, _done(std::move(done)) {
|
|
}
|
|
|
|
auto GroupCall::MediaChannelDescriptionsTask::ssrcs() const
|
|
-> base::flat_set<uint32> {
|
|
return _ssrcs;
|
|
}
|
|
|
|
bool GroupCall::MediaChannelDescriptionsTask::finishWithAdding(
|
|
uint32 ssrc,
|
|
std::optional<tgcalls::MediaChannelDescription> description,
|
|
bool screen) {
|
|
Expects(_ssrcs.contains(ssrc));
|
|
|
|
using Type = tgcalls::MediaChannelDescription::Type;
|
|
_ssrcs.remove(ssrc);
|
|
if (!description) {
|
|
} else if (description->type == Type::Audio
|
|
|| (!screen && _cameraAdded.emplace(description->audioSsrc).second)
|
|
|| (screen && _screenAdded.emplace(description->audioSsrc).second)) {
|
|
_result.push_back(std::move(*description));
|
|
}
|
|
|
|
if (!_ssrcs.empty()) {
|
|
return false;
|
|
}
|
|
QMutexLocker lock(&_mutex);
|
|
if (_done) {
|
|
base::take(_done)(std::move(_result));
|
|
}
|
|
return true;
|
|
}
|
|
|
|
void GroupCall::MediaChannelDescriptionsTask::cancel() {
|
|
QMutexLocker lock(&_mutex);
|
|
if (!_done) {
|
|
return;
|
|
}
|
|
_done = nullptr;
|
|
lock.unlock();
|
|
|
|
if (_call) {
|
|
const auto that = this;
|
|
crl::on_main(_call, [weak = _call, that] {
|
|
if (const auto strong = weak.get()) {
|
|
strong->mediaChannelDescriptionsCancel(that);
|
|
}
|
|
});
|
|
}
|
|
}
|
|
|
|
not_null<PeerData*> GroupCall::TrackPeer(
|
|
const std::unique_ptr<VideoTrack> &track) {
|
|
return track->peer;
|
|
}
|
|
|
|
not_null<Webrtc::VideoTrack*> GroupCall::TrackPointer(
|
|
const std::unique_ptr<VideoTrack> &track) {
|
|
return &track->track;
|
|
}
|
|
|
|
rpl::producer<QSize> GroupCall::TrackSizeValue(
|
|
const std::unique_ptr<VideoTrack> &track) {
|
|
return track->trackSize.value();
|
|
}
|
|
|
|
GroupCall::GroupCall(
|
|
not_null<Delegate*> delegate,
|
|
Group::JoinInfo info,
|
|
const MTPInputGroupCall &inputCall)
|
|
: _delegate(delegate)
|
|
, _peer(info.peer)
|
|
, _history(_peer->owner().history(_peer))
|
|
, _api(&_peer->session().mtp())
|
|
, _joinAs(info.joinAs)
|
|
, _possibleJoinAs(std::move(info.possibleJoinAs))
|
|
, _joinHash(info.joinHash)
|
|
, _canManage(Data::CanManageGroupCallValue(_peer))
|
|
, _id(inputCall.c_inputGroupCall().vid().v)
|
|
, _scheduleDate(info.scheduleDate)
|
|
, _lastSpokeCheckTimer([=] { checkLastSpoke(); })
|
|
, _checkJoinedTimer([=] { checkJoined(); })
|
|
, _pushToTalkCancelTimer([=] { pushToTalkCancel(); })
|
|
, _connectingSoundTimer([=] { playConnectingSoundOnce(); })
|
|
, _mediaDevices(CreateMediaDevices()) {
|
|
_muted.value(
|
|
) | rpl::combine_previous(
|
|
) | rpl::start_with_next([=](MuteState previous, MuteState state) {
|
|
if (_instance) {
|
|
updateInstanceMuteState();
|
|
}
|
|
if (_joinState.ssrc
|
|
&& (!_initialMuteStateSent || state == MuteState::Active)) {
|
|
_initialMuteStateSent = true;
|
|
maybeSendMutedUpdate(previous);
|
|
}
|
|
}, _lifetime);
|
|
|
|
_instanceState.value(
|
|
) | rpl::filter([=] {
|
|
return _hadJoinedState;
|
|
}) | rpl::start_with_next([=](InstanceState state) {
|
|
if (state == InstanceState::Disconnected) {
|
|
playConnectingSound();
|
|
} else {
|
|
stopConnectingSound();
|
|
}
|
|
}, _lifetime);
|
|
|
|
checkGlobalShortcutAvailability();
|
|
|
|
if (const auto real = lookupReal()) {
|
|
subscribeToReal(real);
|
|
if (!canManage() && real->joinMuted()) {
|
|
_muted = MuteState::ForceMuted;
|
|
}
|
|
} else {
|
|
_peer->session().changes().peerFlagsValue(
|
|
_peer,
|
|
Data::PeerUpdate::Flag::GroupCall
|
|
) | rpl::map([=] {
|
|
return lookupReal();
|
|
}) | rpl::filter([](Data::GroupCall *real) {
|
|
return real != nullptr;
|
|
}) | rpl::map([](Data::GroupCall *real) {
|
|
return not_null{ real };
|
|
}) | rpl::take(
|
|
1
|
|
) | rpl::start_with_next([=](not_null<Data::GroupCall*> real) {
|
|
subscribeToReal(real);
|
|
_realChanges.fire_copy(real);
|
|
}, _lifetime);
|
|
}
|
|
|
|
setupMediaDevices();
|
|
setupOutgoingVideo();
|
|
|
|
if (_id) {
|
|
join(inputCall);
|
|
} else {
|
|
start(info.scheduleDate);
|
|
}
|
|
if (_scheduleDate) {
|
|
saveDefaultJoinAs(joinAs());
|
|
}
|
|
}
|
|
|
|
GroupCall::~GroupCall() {
|
|
destroyScreencast();
|
|
destroyController();
|
|
}
|
|
|
|
bool GroupCall::isSharingScreen() const {
|
|
return _isSharingScreen.current();
|
|
}
|
|
|
|
rpl::producer<bool> GroupCall::isSharingScreenValue() const {
|
|
return _isSharingScreen.value();
|
|
}
|
|
|
|
bool GroupCall::isScreenPaused() const {
|
|
return (_screenState.current() == Webrtc::VideoState::Paused);
|
|
}
|
|
|
|
const std::string &GroupCall::screenSharingEndpoint() const {
|
|
return isSharingScreen() ? _screenEndpoint : EmptyString();
|
|
}
|
|
|
|
bool GroupCall::isSharingCamera() const {
|
|
return _isSharingCamera.current();
|
|
}
|
|
|
|
rpl::producer<bool> GroupCall::isSharingCameraValue() const {
|
|
return _isSharingCamera.value();
|
|
}
|
|
|
|
bool GroupCall::isCameraPaused() const {
|
|
return (_cameraState.current() == Webrtc::VideoState::Paused);
|
|
}
|
|
|
|
const std::string &GroupCall::cameraSharingEndpoint() const {
|
|
return isSharingCamera() ? _cameraEndpoint : EmptyString();
|
|
}
|
|
|
|
QString GroupCall::screenSharingDeviceId() const {
|
|
return isSharingScreen() ? _screenDeviceId : QString();
|
|
}
|
|
|
|
bool GroupCall::screenSharingWithAudio() const {
|
|
return isSharingScreen() && _screenWithAudio;
|
|
}
|
|
|
|
bool GroupCall::mutedByAdmin() const {
|
|
const auto mute = muted();
|
|
return mute == MuteState::ForceMuted || mute == MuteState::RaisedHand;
|
|
}
|
|
|
|
bool GroupCall::canManage() const {
|
|
return _canManage.current();
|
|
}
|
|
|
|
rpl::producer<bool> GroupCall::canManageValue() const {
|
|
return _canManage.value();
|
|
}
|
|
|
|
void GroupCall::toggleVideo(bool active) {
|
|
if (!_instance || !_id) {
|
|
return;
|
|
}
|
|
_cameraState = active
|
|
? Webrtc::VideoState::Active
|
|
: Webrtc::VideoState::Inactive;
|
|
}
|
|
|
|
void GroupCall::toggleScreenSharing(
|
|
std::optional<QString> uniqueId,
|
|
bool withAudio) {
|
|
if (!_instance || !_id) {
|
|
return;
|
|
} else if (!uniqueId) {
|
|
_screenState = Webrtc::VideoState::Inactive;
|
|
return;
|
|
}
|
|
const auto changed = (_screenDeviceId != *uniqueId);
|
|
const auto wasSharing = isSharingScreen();
|
|
_screenDeviceId = *uniqueId;
|
|
_screenWithAudio = withAudio;
|
|
_screenState = Webrtc::VideoState::Active;
|
|
if (changed && wasSharing && isSharingScreen()) {
|
|
_screenCapture->switchToDevice(uniqueId->toStdString(), true);
|
|
}
|
|
if (_screenInstance) {
|
|
_screenInstance->setIsMuted(!withAudio);
|
|
}
|
|
}
|
|
|
|
bool GroupCall::hasVideoWithFrames() const {
|
|
return !_shownVideoTracks.empty();
|
|
}
|
|
|
|
rpl::producer<bool> GroupCall::hasVideoWithFramesValue() const {
|
|
return _videoStreamShownUpdates.events_starting_with(
|
|
VideoStateToggle()
|
|
) | rpl::map([=] {
|
|
return hasVideoWithFrames();
|
|
}) | rpl::distinct_until_changed();
|
|
}
|
|
|
|
void GroupCall::setScheduledDate(TimeId date) {
|
|
const auto was = _scheduleDate;
|
|
_scheduleDate = date;
|
|
if (was && !date) {
|
|
join(inputCall());
|
|
}
|
|
}
|
|
|
|
void GroupCall::subscribeToReal(not_null<Data::GroupCall*> real) {
|
|
real->scheduleDateValue(
|
|
) | rpl::start_with_next([=](TimeId date) {
|
|
setScheduledDate(date);
|
|
}, _lifetime);
|
|
|
|
// Postpone creating video tracks, so that we know if Panel
|
|
// supports OpenGL and we don't need ARGB32 frames at all.
|
|
Ui::PostponeCall(this, [=] {
|
|
if (const auto real = lookupReal()) {
|
|
real->participantsReloaded(
|
|
) | rpl::start_with_next([=] {
|
|
fillActiveVideoEndpoints();
|
|
}, _lifetime);
|
|
fillActiveVideoEndpoints();
|
|
}
|
|
});
|
|
|
|
using Update = Data::GroupCall::ParticipantUpdate;
|
|
real->participantUpdated(
|
|
) | rpl::start_with_next([=](const Update &data) {
|
|
const auto regularEndpoint = [&](const std::string &endpoint)
|
|
-> const std::string & {
|
|
return (endpoint.empty()
|
|
|| endpoint == _cameraEndpoint
|
|
|| endpoint == _screenEndpoint)
|
|
? EmptyString()
|
|
: endpoint;
|
|
};
|
|
|
|
const auto peer = data.was ? data.was->peer : data.now->peer;
|
|
if (peer == joinAs()) {
|
|
const auto working = data.now && data.now->videoJoined;
|
|
if (videoIsWorking() != working) {
|
|
fillActiveVideoEndpoints();
|
|
}
|
|
return;
|
|
}
|
|
const auto &wasCameraEndpoint = data.was
|
|
? regularEndpoint(GetCameraEndpoint(data.was->videoParams))
|
|
: EmptyString();
|
|
const auto &nowCameraEndpoint = data.now
|
|
? regularEndpoint(GetCameraEndpoint(data.now->videoParams))
|
|
: EmptyString();
|
|
const auto wasCameraPaused = !wasCameraEndpoint.empty()
|
|
&& IsCameraPaused(data.was->videoParams);
|
|
const auto nowCameraPaused = !nowCameraEndpoint.empty()
|
|
&& IsCameraPaused(data.now->videoParams);
|
|
if (wasCameraEndpoint != nowCameraEndpoint) {
|
|
markEndpointActive({
|
|
VideoEndpointType::Camera,
|
|
peer,
|
|
nowCameraEndpoint,
|
|
}, true, nowCameraPaused);
|
|
markEndpointActive({
|
|
VideoEndpointType::Camera,
|
|
peer,
|
|
wasCameraEndpoint,
|
|
}, false, false);
|
|
} else if (wasCameraPaused != nowCameraPaused) {
|
|
markTrackPaused({
|
|
VideoEndpointType::Camera,
|
|
peer,
|
|
nowCameraEndpoint,
|
|
}, nowCameraPaused);
|
|
}
|
|
const auto &wasScreenEndpoint = data.was
|
|
? regularEndpoint(data.was->screenEndpoint())
|
|
: EmptyString();
|
|
const auto &nowScreenEndpoint = data.now
|
|
? regularEndpoint(data.now->screenEndpoint())
|
|
: EmptyString();
|
|
const auto wasScreenPaused = !wasScreenEndpoint.empty()
|
|
&& IsScreenPaused(data.was->videoParams);
|
|
const auto nowScreenPaused = !nowScreenEndpoint.empty()
|
|
&& IsScreenPaused(data.now->videoParams);
|
|
if (wasScreenEndpoint != nowScreenEndpoint) {
|
|
markEndpointActive({
|
|
VideoEndpointType::Screen,
|
|
peer,
|
|
nowScreenEndpoint,
|
|
}, true, nowScreenPaused);
|
|
markEndpointActive({
|
|
VideoEndpointType::Screen,
|
|
peer,
|
|
wasScreenEndpoint,
|
|
}, false, false);
|
|
} else if (wasScreenPaused != nowScreenPaused) {
|
|
markTrackPaused({
|
|
VideoEndpointType::Screen,
|
|
peer,
|
|
wasScreenEndpoint,
|
|
}, nowScreenPaused);
|
|
}
|
|
}, _lifetime);
|
|
|
|
real->participantsResolved(
|
|
) | rpl::start_with_next([=](
|
|
not_null<const base::flat_map<
|
|
uint32,
|
|
Data::LastSpokeTimes>*> ssrcs) {
|
|
checkMediaChannelDescriptions([&](uint32 ssrc) {
|
|
return ssrcs->contains(ssrc);
|
|
});
|
|
}, _lifetime);
|
|
|
|
real->participantSpeaking(
|
|
) | rpl::filter([=] {
|
|
return _videoEndpointLarge.current();
|
|
}) | rpl::start_with_next([=](not_null<Data::GroupCallParticipant*> p) {
|
|
const auto now = crl::now();
|
|
if (_videoEndpointLarge.current().peer == p->peer) {
|
|
_videoLargeTillTime = std::max(
|
|
_videoLargeTillTime,
|
|
now + kFixSpeakingLargeVideoDuration);
|
|
return;
|
|
} else if (videoEndpointPinned() || _videoLargeTillTime > now) {
|
|
return;
|
|
}
|
|
using Type = VideoEndpointType;
|
|
const auto ¶ms = p->videoParams;
|
|
if (GetCameraEndpoint(params).empty()
|
|
&& GetScreenEndpoint(params).empty()) {
|
|
return;
|
|
}
|
|
const auto tryEndpoint = [&](Type type, const std::string &id) {
|
|
if (id.empty()) {
|
|
return false;
|
|
}
|
|
const auto endpoint = VideoEndpoint{ type, p->peer, id };
|
|
if (!shownVideoTracks().contains(endpoint)) {
|
|
return false;
|
|
}
|
|
setVideoEndpointLarge(endpoint);
|
|
return true;
|
|
};
|
|
if (tryEndpoint(Type::Screen, GetScreenEndpoint(params))
|
|
|| tryEndpoint(Type::Camera, GetCameraEndpoint(params))) {
|
|
_videoLargeTillTime = now + kFixSpeakingLargeVideoDuration;
|
|
}
|
|
}, _lifetime);
|
|
}
|
|
|
|
void GroupCall::checkGlobalShortcutAvailability() {
|
|
auto &settings = Core::App().settings();
|
|
if (!settings.groupCallPushToTalk()) {
|
|
return;
|
|
} else if (!base::GlobalShortcutsAllowed()) {
|
|
settings.setGroupCallPushToTalk(false);
|
|
Core::App().saveSettingsDelayed();
|
|
}
|
|
}
|
|
|
|
void GroupCall::setState(State state) {
|
|
const auto current = _state.current();
|
|
if (current == State::Failed) {
|
|
return;
|
|
} else if (current == State::Ended && state != State::Failed) {
|
|
return;
|
|
} else if (current == State::FailedHangingUp && state != State::Failed) {
|
|
return;
|
|
} else if (current == State::HangingUp
|
|
&& state != State::Ended
|
|
&& state != State::Failed) {
|
|
return;
|
|
}
|
|
if (current == state) {
|
|
return;
|
|
}
|
|
_state = state;
|
|
|
|
if (state == State::Joined) {
|
|
stopConnectingSound();
|
|
if (const auto call = _peer->groupCall(); call && call->id() == _id) {
|
|
call->setInCall();
|
|
}
|
|
}
|
|
|
|
if (false
|
|
|| state == State::Ended
|
|
|| state == State::Failed) {
|
|
// Destroy controller before destroying Call Panel,
|
|
// so that the panel hide animation is smooth.
|
|
destroyScreencast();
|
|
destroyController();
|
|
}
|
|
switch (state) {
|
|
case State::HangingUp:
|
|
case State::FailedHangingUp:
|
|
stopConnectingSound();
|
|
_delegate->groupCallPlaySound(Delegate::GroupCallSound::Ended);
|
|
break;
|
|
case State::Ended:
|
|
stopConnectingSound();
|
|
_delegate->groupCallFinished(this);
|
|
break;
|
|
case State::Failed:
|
|
stopConnectingSound();
|
|
_delegate->groupCallFailed(this);
|
|
break;
|
|
case State::Connecting:
|
|
if (!_checkJoinedTimer.isActive()) {
|
|
_checkJoinedTimer.callOnce(kCheckJoinedTimeout);
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
|
|
void GroupCall::playConnectingSound() {
|
|
const auto state = _state.current();
|
|
if (_connectingSoundTimer.isActive()
|
|
|| state == State::HangingUp
|
|
|| state == State::FailedHangingUp
|
|
|| state == State::Ended
|
|
|| state == State::Failed) {
|
|
return;
|
|
}
|
|
playConnectingSoundOnce();
|
|
_connectingSoundTimer.callEach(kPlayConnectingEach);
|
|
}
|
|
|
|
void GroupCall::stopConnectingSound() {
|
|
_connectingSoundTimer.cancel();
|
|
}
|
|
|
|
void GroupCall::playConnectingSoundOnce() {
|
|
_delegate->groupCallPlaySound(Delegate::GroupCallSound::Connecting);
|
|
}
|
|
|
|
bool GroupCall::showChooseJoinAs() const {
|
|
return (_possibleJoinAs.size() > 1)
|
|
|| (_possibleJoinAs.size() == 1
|
|
&& !_possibleJoinAs.front()->isSelf());
|
|
}
|
|
|
|
bool GroupCall::scheduleStartSubscribed() const {
|
|
if (const auto real = lookupReal()) {
|
|
return real->scheduleStartSubscribed();
|
|
}
|
|
return false;
|
|
}
|
|
|
|
Data::GroupCall *GroupCall::lookupReal() const {
|
|
const auto real = _peer->groupCall();
|
|
return (real && real->id() == _id) ? real : nullptr;
|
|
}
|
|
|
|
rpl::producer<not_null<Data::GroupCall*>> GroupCall::real() const {
|
|
if (const auto real = lookupReal()) {
|
|
return rpl::single(not_null{ real });
|
|
}
|
|
return _realChanges.events();
|
|
}
|
|
|
|
void GroupCall::start(TimeId scheduleDate) {
|
|
using Flag = MTPphone_CreateGroupCall::Flag;
|
|
_createRequestId = _api.request(MTPphone_CreateGroupCall(
|
|
MTP_flags(scheduleDate ? Flag::f_schedule_date : Flag(0)),
|
|
_peer->input,
|
|
MTP_int(base::RandomValue<int32>()),
|
|
MTPstring(), // title
|
|
MTP_int(scheduleDate)
|
|
)).done([=](const MTPUpdates &result) {
|
|
_acceptFields = true;
|
|
_peer->session().api().applyUpdates(result);
|
|
_acceptFields = false;
|
|
}).fail([=](const MTP::Error &error) {
|
|
LOG(("Call Error: Could not create, error: %1"
|
|
).arg(error.type()));
|
|
hangup();
|
|
}).send();
|
|
}
|
|
|
|
void GroupCall::join(const MTPInputGroupCall &inputCall) {
|
|
inputCall.match([&](const MTPDinputGroupCall &data) {
|
|
_id = data.vid().v;
|
|
_accessHash = data.vaccess_hash().v;
|
|
});
|
|
setState(_scheduleDate ? State::Waiting : State::Joining);
|
|
|
|
if (_scheduleDate) {
|
|
return;
|
|
}
|
|
rejoin();
|
|
|
|
using Update = Data::GroupCall::ParticipantUpdate;
|
|
const auto real = lookupReal();
|
|
Assert(real != nullptr);
|
|
real->participantUpdated(
|
|
) | rpl::filter([=](const Update &update) {
|
|
return (_instance != nullptr);
|
|
}) | rpl::start_with_next([=](const Update &update) {
|
|
if (!update.now) {
|
|
_instance->removeSsrcs({
|
|
update.was->ssrc,
|
|
GetAdditionalAudioSsrc(update.was->videoParams),
|
|
});
|
|
} else {
|
|
updateInstanceVolume(update.was, *update.now);
|
|
}
|
|
}, _lifetime);
|
|
|
|
_peer->session().updates().addActiveChat(
|
|
_peerStream.events_starting_with_copy(_peer));
|
|
SubscribeToMigration(_peer, _lifetime, [=](not_null<ChannelData*> group) {
|
|
_peer = group;
|
|
_canManage = Data::CanManageGroupCallValue(_peer);
|
|
_peerStream.fire_copy(group);
|
|
});
|
|
}
|
|
|
|
void GroupCall::setScreenEndpoint(std::string endpoint) {
|
|
if (_screenEndpoint == endpoint) {
|
|
return;
|
|
}
|
|
if (!_screenEndpoint.empty()) {
|
|
markEndpointActive({
|
|
VideoEndpointType::Screen,
|
|
joinAs(),
|
|
_screenEndpoint
|
|
}, false, false);
|
|
}
|
|
_screenEndpoint = std::move(endpoint);
|
|
if (_screenEndpoint.empty()) {
|
|
return;
|
|
}
|
|
if (isSharingScreen()) {
|
|
markEndpointActive({
|
|
VideoEndpointType::Screen,
|
|
joinAs(),
|
|
_screenEndpoint
|
|
}, true, isScreenPaused());
|
|
}
|
|
}
|
|
|
|
void GroupCall::setCameraEndpoint(std::string endpoint) {
|
|
if (_cameraEndpoint == endpoint) {
|
|
return;
|
|
}
|
|
if (!_cameraEndpoint.empty()) {
|
|
markEndpointActive({
|
|
VideoEndpointType::Camera,
|
|
joinAs(),
|
|
_cameraEndpoint
|
|
}, false, false);
|
|
}
|
|
_cameraEndpoint = std::move(endpoint);
|
|
if (_cameraEndpoint.empty()) {
|
|
return;
|
|
}
|
|
if (isSharingCamera()) {
|
|
markEndpointActive({
|
|
VideoEndpointType::Camera,
|
|
joinAs(),
|
|
_cameraEndpoint
|
|
}, true, isCameraPaused());
|
|
}
|
|
}
|
|
|
|
void GroupCall::addVideoOutput(
|
|
const std::string &endpoint,
|
|
SinkPointer sink) {
|
|
if (_cameraEndpoint == endpoint) {
|
|
if (auto strong = sink.data.lock()) {
|
|
_cameraCapture->setOutput(std::move(strong));
|
|
}
|
|
} else if (_screenEndpoint == endpoint) {
|
|
if (auto strong = sink.data.lock()) {
|
|
_screenCapture->setOutput(std::move(strong));
|
|
}
|
|
} else if (_instance) {
|
|
_instance->addIncomingVideoOutput(endpoint, std::move(sink.data));
|
|
} else {
|
|
_pendingVideoOutputs.emplace(endpoint, std::move(sink));
|
|
}
|
|
}
|
|
|
|
void GroupCall::markEndpointActive(
|
|
VideoEndpoint endpoint,
|
|
bool active,
|
|
bool paused) {
|
|
if (!endpoint) {
|
|
return;
|
|
}
|
|
const auto i = _activeVideoTracks.find(endpoint);
|
|
const auto changed = active
|
|
? (i == end(_activeVideoTracks))
|
|
: (i != end(_activeVideoTracks));
|
|
if (!changed) {
|
|
if (active) {
|
|
markTrackPaused(endpoint, paused);
|
|
}
|
|
return;
|
|
}
|
|
auto shown = false;
|
|
if (active) {
|
|
const auto i = _activeVideoTracks.emplace(
|
|
endpoint,
|
|
std::make_unique<VideoTrack>(
|
|
paused,
|
|
_requireARGB32,
|
|
endpoint.peer)).first;
|
|
const auto track = &i->second->track;
|
|
|
|
track->renderNextFrame(
|
|
) | rpl::start_with_next([=] {
|
|
const auto activeTrack = _activeVideoTracks[endpoint].get();
|
|
const auto size = track->frameSize();
|
|
if (size.isEmpty()) {
|
|
track->markFrameShown();
|
|
} else if (!activeTrack->shown) {
|
|
activeTrack->shown = true;
|
|
markTrackShown(endpoint, true);
|
|
}
|
|
activeTrack->trackSize = size;
|
|
}, i->second->lifetime);
|
|
|
|
const auto size = track->frameSize();
|
|
i->second->trackSize = size;
|
|
if (!size.isEmpty() || paused) {
|
|
i->second->shown = true;
|
|
shown = true;
|
|
} else {
|
|
track->stateValue(
|
|
) | rpl::filter([=](Webrtc::VideoState state) {
|
|
return (state == Webrtc::VideoState::Paused)
|
|
&& !_activeVideoTracks[endpoint]->shown;
|
|
}) | rpl::start_with_next([=] {
|
|
_activeVideoTracks[endpoint]->shown = true;
|
|
markTrackShown(endpoint, true);
|
|
}, i->second->lifetime);
|
|
}
|
|
addVideoOutput(i->first.id, { track->sink() });
|
|
} else {
|
|
if (_videoEndpointLarge.current() == endpoint) {
|
|
setVideoEndpointLarge({});
|
|
}
|
|
markTrackShown(endpoint, false);
|
|
markTrackPaused(endpoint, false);
|
|
_activeVideoTracks.erase(i);
|
|
}
|
|
updateRequestedVideoChannelsDelayed();
|
|
_videoStreamActiveUpdates.fire({ endpoint, active });
|
|
if (active) {
|
|
markTrackShown(endpoint, shown);
|
|
markTrackPaused(endpoint, paused);
|
|
}
|
|
}
|
|
|
|
void GroupCall::markTrackShown(const VideoEndpoint &endpoint, bool shown) {
|
|
const auto changed = shown
|
|
? _shownVideoTracks.emplace(endpoint).second
|
|
: _shownVideoTracks.remove(endpoint);
|
|
if (!changed) {
|
|
return;
|
|
}
|
|
_videoStreamShownUpdates.fire_copy({ endpoint, shown });
|
|
if (shown && endpoint.type == VideoEndpointType::Screen) {
|
|
crl::on_main(this, [=] {
|
|
if (_shownVideoTracks.contains(endpoint)) {
|
|
pinVideoEndpoint(endpoint);
|
|
}
|
|
});
|
|
}
|
|
}
|
|
|
|
void GroupCall::markTrackPaused(const VideoEndpoint &endpoint, bool paused) {
|
|
if (!endpoint) {
|
|
return;
|
|
}
|
|
|
|
const auto i = _activeVideoTracks.find(endpoint);
|
|
Assert(i != end(_activeVideoTracks));
|
|
|
|
i->second->track.setState(paused
|
|
? Webrtc::VideoState::Paused
|
|
: Webrtc::VideoState::Active);
|
|
}
|
|
|
|
void GroupCall::rejoin() {
|
|
rejoin(joinAs());
|
|
}
|
|
|
|
void GroupCall::rejoinWithHash(const QString &hash) {
|
|
if (!hash.isEmpty() && mutedByAdmin()) {
|
|
_joinHash = hash;
|
|
rejoin();
|
|
}
|
|
}
|
|
|
|
void GroupCall::setJoinAs(not_null<PeerData*> as) {
|
|
_joinAs = as;
|
|
if (const auto chat = _peer->asChat()) {
|
|
chat->setGroupCallDefaultJoinAs(joinAs()->id);
|
|
} else if (const auto channel = _peer->asChannel()) {
|
|
channel->setGroupCallDefaultJoinAs(joinAs()->id);
|
|
}
|
|
}
|
|
|
|
void GroupCall::saveDefaultJoinAs(not_null<PeerData*> as) {
|
|
setJoinAs(as);
|
|
_api.request(MTPphone_SaveDefaultGroupCallJoinAs(
|
|
_peer->input,
|
|
joinAs()->input
|
|
)).send();
|
|
}
|
|
|
|
void GroupCall::rejoin(not_null<PeerData*> as) {
|
|
if (state() != State::Joining
|
|
&& state() != State::Joined
|
|
&& state() != State::Connecting) {
|
|
return;
|
|
} else if (_joinState.action != JoinAction::None) {
|
|
return;
|
|
}
|
|
|
|
if (joinAs() != as) {
|
|
toggleVideo(false);
|
|
toggleScreenSharing(std::nullopt);
|
|
}
|
|
|
|
_joinState.action = JoinAction::Joining;
|
|
_joinState.ssrc = 0;
|
|
_initialMuteStateSent = false;
|
|
setState(State::Joining);
|
|
if (!tryCreateController()) {
|
|
setInstanceMode(InstanceMode::None);
|
|
}
|
|
applyMeInCallLocally();
|
|
LOG(("Call Info: Requesting join payload."));
|
|
|
|
setJoinAs(as);
|
|
|
|
const auto weak = base::make_weak(&_instanceGuard);
|
|
_instance->emitJoinPayload([=](tgcalls::GroupJoinPayload payload) {
|
|
crl::on_main(weak, [=, payload = std::move(payload)] {
|
|
if (state() != State::Joining) {
|
|
_joinState.finish();
|
|
checkNextJoinAction();
|
|
return;
|
|
}
|
|
const auto ssrc = payload.audioSsrc;
|
|
LOG(("Call Info: Join payload received, joining with ssrc: %1."
|
|
).arg(ssrc));
|
|
|
|
const auto json = QByteArray::fromStdString(payload.json);
|
|
const auto wasMuteState = muted();
|
|
const auto wasVideoStopped = !isSharingCamera();
|
|
using Flag = MTPphone_JoinGroupCall::Flag;
|
|
const auto flags = (wasMuteState != MuteState::Active
|
|
? Flag::f_muted
|
|
: Flag(0))
|
|
| (_joinHash.isEmpty()
|
|
? Flag(0)
|
|
: Flag::f_invite_hash)
|
|
| (wasVideoStopped
|
|
? Flag::f_video_stopped
|
|
: Flag(0));
|
|
_api.request(MTPphone_JoinGroupCall(
|
|
MTP_flags(flags),
|
|
inputCall(),
|
|
joinAs()->input,
|
|
MTP_string(_joinHash),
|
|
MTP_dataJSON(MTP_bytes(json))
|
|
)).done([=](
|
|
const MTPUpdates &updates,
|
|
const MTP::Response &response) {
|
|
_serverTimeMs = TimestampInMsFromMsgId(response.outerMsgId);
|
|
_serverTimeMsGotAt = crl::now();
|
|
|
|
_joinState.finish(ssrc);
|
|
_mySsrcs.emplace(ssrc);
|
|
|
|
setState((_instanceState.current()
|
|
== InstanceState::Disconnected)
|
|
? State::Connecting
|
|
: State::Joined);
|
|
applyMeInCallLocally();
|
|
maybeSendMutedUpdate(wasMuteState);
|
|
_peer->session().api().applyUpdates(updates);
|
|
applyQueuedSelfUpdates();
|
|
checkFirstTimeJoined();
|
|
_screenJoinState.nextActionPending = true;
|
|
checkNextJoinAction();
|
|
if (wasVideoStopped == isSharingCamera()) {
|
|
sendSelfUpdate(SendUpdateType::CameraStopped);
|
|
}
|
|
if (isCameraPaused()) {
|
|
sendSelfUpdate(SendUpdateType::CameraPaused);
|
|
}
|
|
sendPendingSelfUpdates();
|
|
}).fail([=](const MTP::Error &error) {
|
|
_joinState.finish();
|
|
|
|
const auto type = error.type();
|
|
LOG(("Call Error: Could not join, error: %1").arg(type));
|
|
|
|
if (type == u"GROUPCALL_SSRC_DUPLICATE_MUCH") {
|
|
rejoin();
|
|
return;
|
|
}
|
|
|
|
hangup();
|
|
Ui::ShowMultilineToast({
|
|
.text = { type == u"GROUPCALL_FORBIDDEN"_q
|
|
? tr::lng_group_not_accessible(tr::now)
|
|
: Lang::Hard::ServerError() },
|
|
});
|
|
}).send();
|
|
});
|
|
});
|
|
}
|
|
|
|
void GroupCall::checkNextJoinAction() {
|
|
if (_joinState.action != JoinAction::None) {
|
|
return;
|
|
} else if (_joinState.nextActionPending) {
|
|
_joinState.nextActionPending = false;
|
|
const auto state = _state.current();
|
|
if (state != State::HangingUp && state != State::FailedHangingUp) {
|
|
rejoin();
|
|
} else {
|
|
leave();
|
|
}
|
|
} else if (!_joinState.ssrc) {
|
|
rejoin();
|
|
} else if (_screenJoinState.action != JoinAction::None
|
|
|| !_screenJoinState.nextActionPending) {
|
|
return;
|
|
} else {
|
|
_screenJoinState.nextActionPending = false;
|
|
if (isSharingScreen()) {
|
|
rejoinPresentation();
|
|
} else {
|
|
leavePresentation();
|
|
}
|
|
}
|
|
}
|
|
|
|
void GroupCall::rejoinPresentation() {
|
|
if (!_joinState.ssrc
|
|
|| _screenJoinState.action == JoinAction::Joining
|
|
|| !isSharingScreen()) {
|
|
return;
|
|
} else if (_screenJoinState.action != JoinAction::None) {
|
|
_screenJoinState.nextActionPending = true;
|
|
return;
|
|
}
|
|
|
|
_screenJoinState.action = JoinAction::Joining;
|
|
_screenJoinState.ssrc = 0;
|
|
if (!tryCreateScreencast()) {
|
|
setScreenInstanceMode(InstanceMode::None);
|
|
}
|
|
LOG(("Call Info: Requesting join screen payload."));
|
|
|
|
const auto weak = base::make_weak(&_screenInstanceGuard);
|
|
_screenInstance->emitJoinPayload([=](tgcalls::GroupJoinPayload payload) {
|
|
crl::on_main(weak, [=, payload = std::move(payload)]{
|
|
if (!isSharingScreen() || !_joinState.ssrc) {
|
|
_screenJoinState.finish();
|
|
checkNextJoinAction();
|
|
return;
|
|
}
|
|
const auto withMainSsrc = _joinState.ssrc;
|
|
const auto ssrc = payload.audioSsrc;
|
|
LOG(("Call Info: Join screen payload received, ssrc: %1."
|
|
).arg(ssrc));
|
|
|
|
const auto json = QByteArray::fromStdString(payload.json);
|
|
_api.request(
|
|
MTPphone_JoinGroupCallPresentation(
|
|
inputCall(),
|
|
MTP_dataJSON(MTP_bytes(json)))
|
|
).done([=](const MTPUpdates &updates) {
|
|
_screenJoinState.finish(ssrc);
|
|
_mySsrcs.emplace(ssrc);
|
|
|
|
_peer->session().api().applyUpdates(updates);
|
|
checkNextJoinAction();
|
|
if (isScreenPaused()) {
|
|
sendSelfUpdate(SendUpdateType::ScreenPaused);
|
|
}
|
|
sendPendingSelfUpdates();
|
|
}).fail([=](const MTP::Error &error) {
|
|
_screenJoinState.finish();
|
|
|
|
const auto type = error.type();
|
|
if (type == u"GROUPCALL_SSRC_DUPLICATE_MUCH") {
|
|
_screenJoinState.nextActionPending = true;
|
|
checkNextJoinAction();
|
|
} else if (type == u"GROUPCALL_JOIN_MISSING"_q
|
|
|| type == u"GROUPCALL_FORBIDDEN"_q) {
|
|
if (_joinState.ssrc != withMainSsrc) {
|
|
// We've rejoined, rejoin presentation again.
|
|
_screenJoinState.nextActionPending = true;
|
|
checkNextJoinAction();
|
|
}
|
|
} else {
|
|
LOG(("Call Error: "
|
|
"Could not screen join, error: %1").arg(type));
|
|
_screenState = Webrtc::VideoState::Inactive;
|
|
_errors.fire_copy(mutedByAdmin()
|
|
? Error::MutedNoScreen
|
|
: Error::ScreenFailed);
|
|
}
|
|
}).send();
|
|
});
|
|
});
|
|
}
|
|
|
|
void GroupCall::leavePresentation() {
|
|
destroyScreencast();
|
|
if (!_screenJoinState.ssrc) {
|
|
setScreenEndpoint(std::string());
|
|
return;
|
|
} else if (_screenJoinState.action == JoinAction::Leaving) {
|
|
return;
|
|
} else if (_screenJoinState.action != JoinAction::None) {
|
|
_screenJoinState.nextActionPending = true;
|
|
return;
|
|
}
|
|
_api.request(
|
|
MTPphone_LeaveGroupCallPresentation(inputCall())
|
|
).done([=](const MTPUpdates &updates) {
|
|
_screenJoinState.finish();
|
|
|
|
_peer->session().api().applyUpdates(updates);
|
|
setScreenEndpoint(std::string());
|
|
checkNextJoinAction();
|
|
}).fail([=](const MTP::Error &error) {
|
|
_screenJoinState.finish();
|
|
|
|
const auto type = error.type();
|
|
LOG(("Call Error: "
|
|
"Could not screen leave, error: %1").arg(type));
|
|
setScreenEndpoint(std::string());
|
|
checkNextJoinAction();
|
|
}).send();
|
|
}
|
|
|
|
void GroupCall::applyMeInCallLocally() {
|
|
const auto real = lookupReal();
|
|
if (!real) {
|
|
return;
|
|
}
|
|
using Flag = MTPDgroupCallParticipant::Flag;
|
|
const auto participant = real->participantByPeer(joinAs());
|
|
const auto date = participant
|
|
? participant->date
|
|
: base::unixtime::now();
|
|
const auto lastActive = participant
|
|
? participant->lastActive
|
|
: TimeId(0);
|
|
const auto volume = participant
|
|
? participant->volume
|
|
: Group::kDefaultVolume;
|
|
const auto canSelfUnmute = !mutedByAdmin();
|
|
const auto raisedHandRating = (muted() != MuteState::RaisedHand)
|
|
? uint64(0)
|
|
: participant
|
|
? participant->raisedHandRating
|
|
: FindLocalRaisedHandRating(real->participants());
|
|
const auto flags = (canSelfUnmute ? Flag::f_can_self_unmute : Flag(0))
|
|
| (lastActive ? Flag::f_active_date : Flag(0))
|
|
| (_joinState.ssrc ? Flag(0) : Flag::f_left)
|
|
| (_videoIsWorking.current() ? Flag::f_video_joined : Flag(0))
|
|
| Flag::f_self
|
|
| Flag::f_volume // Without flag the volume is reset to 100%.
|
|
| Flag::f_volume_by_admin // Self volume can only be set by admin.
|
|
| ((muted() != MuteState::Active) ? Flag::f_muted : Flag(0))
|
|
| (raisedHandRating > 0 ? Flag::f_raise_hand_rating : Flag(0));
|
|
real->applyLocalUpdate(
|
|
MTP_updateGroupCallParticipants(
|
|
inputCall(),
|
|
MTP_vector<MTPGroupCallParticipant>(
|
|
1,
|
|
MTP_groupCallParticipant(
|
|
MTP_flags(flags),
|
|
peerToMTP(joinAs()->id),
|
|
MTP_int(date),
|
|
MTP_int(lastActive),
|
|
MTP_int(_joinState.ssrc),
|
|
MTP_int(volume),
|
|
MTPstring(), // Don't update about text in local updates.
|
|
MTP_long(raisedHandRating),
|
|
MTPGroupCallParticipantVideo(),
|
|
MTPGroupCallParticipantVideo())),
|
|
MTP_int(0)).c_updateGroupCallParticipants());
|
|
}
|
|
|
|
void GroupCall::applyParticipantLocally(
|
|
not_null<PeerData*> participantPeer,
|
|
bool mute,
|
|
std::optional<int> volume) {
|
|
const auto participant = LookupParticipant(_peer, _id, participantPeer);
|
|
if (!participant || !participant->ssrc) {
|
|
return;
|
|
}
|
|
const auto canManageCall = canManage();
|
|
const auto isMuted = participant->muted || (mute && canManageCall);
|
|
const auto canSelfUnmute = !canManageCall
|
|
? participant->canSelfUnmute
|
|
: (!mute || IsGroupCallAdmin(_peer, participantPeer));
|
|
const auto isMutedByYou = mute && !canManageCall;
|
|
using Flag = MTPDgroupCallParticipant::Flag;
|
|
const auto flags = (canSelfUnmute ? Flag::f_can_self_unmute : Flag(0))
|
|
| Flag::f_volume // Without flag the volume is reset to 100%.
|
|
| ((participant->applyVolumeFromMin && !volume)
|
|
? Flag::f_volume_by_admin
|
|
: Flag(0))
|
|
| (participant->videoJoined ? Flag::f_video_joined : Flag(0))
|
|
| (participant->lastActive ? Flag::f_active_date : Flag(0))
|
|
| (isMuted ? Flag::f_muted : Flag(0))
|
|
| (isMutedByYou ? Flag::f_muted_by_you : Flag(0))
|
|
| (participantPeer == joinAs() ? Flag::f_self : Flag(0))
|
|
| (participant->raisedHandRating
|
|
? Flag::f_raise_hand_rating
|
|
: Flag(0));
|
|
_peer->groupCall()->applyLocalUpdate(
|
|
MTP_updateGroupCallParticipants(
|
|
inputCall(),
|
|
MTP_vector<MTPGroupCallParticipant>(
|
|
1,
|
|
MTP_groupCallParticipant(
|
|
MTP_flags(flags),
|
|
peerToMTP(participantPeer->id),
|
|
MTP_int(participant->date),
|
|
MTP_int(participant->lastActive),
|
|
MTP_int(participant->ssrc),
|
|
MTP_int(volume.value_or(participant->volume)),
|
|
MTPstring(), // Don't update about text in local updates.
|
|
MTP_long(participant->raisedHandRating),
|
|
MTPGroupCallParticipantVideo(),
|
|
MTPGroupCallParticipantVideo())),
|
|
MTP_int(0)).c_updateGroupCallParticipants());
|
|
}
|
|
|
|
void GroupCall::hangup() {
|
|
finish(FinishType::Ended);
|
|
}
|
|
|
|
void GroupCall::discard() {
|
|
if (!_id) {
|
|
_api.request(_createRequestId).cancel();
|
|
hangup();
|
|
return;
|
|
}
|
|
_api.request(MTPphone_DiscardGroupCall(
|
|
inputCall()
|
|
)).done([=](const MTPUpdates &result) {
|
|
// Here 'this' could be destroyed by updates, so we set Ended after
|
|
// updates being handled, but in a guarded way.
|
|
crl::on_main(this, [=] { hangup(); });
|
|
_peer->session().api().applyUpdates(result);
|
|
}).fail([=](const MTP::Error &error) {
|
|
hangup();
|
|
}).send();
|
|
}
|
|
|
|
void GroupCall::rejoinAs(Group::JoinInfo info) {
|
|
_possibleJoinAs = std::move(info.possibleJoinAs);
|
|
if (info.joinAs == joinAs()) {
|
|
return;
|
|
}
|
|
const auto event = Group::RejoinEvent{
|
|
.wasJoinAs = joinAs(),
|
|
.nowJoinAs = info.joinAs,
|
|
};
|
|
if (_scheduleDate) {
|
|
saveDefaultJoinAs(info.joinAs);
|
|
} else {
|
|
setState(State::Joining);
|
|
rejoin(info.joinAs);
|
|
}
|
|
_rejoinEvents.fire_copy(event);
|
|
}
|
|
|
|
void GroupCall::finish(FinishType type) {
|
|
Expects(type != FinishType::None);
|
|
|
|
const auto finalState = (type == FinishType::Ended)
|
|
? State::Ended
|
|
: State::Failed;
|
|
const auto hangupState = (type == FinishType::Ended)
|
|
? State::HangingUp
|
|
: State::FailedHangingUp;
|
|
const auto state = _state.current();
|
|
if (state == State::HangingUp
|
|
|| state == State::FailedHangingUp
|
|
|| state == State::Ended
|
|
|| state == State::Failed) {
|
|
return;
|
|
} else if (_joinState.action == JoinAction::None && !_joinState.ssrc) {
|
|
setState(finalState);
|
|
return;
|
|
}
|
|
setState(hangupState);
|
|
_joinState.nextActionPending = true;
|
|
checkNextJoinAction();
|
|
}
|
|
|
|
void GroupCall::leave() {
|
|
Expects(_joinState.action == JoinAction::None);
|
|
|
|
_joinState.action = JoinAction::Leaving;
|
|
|
|
const auto finalState = (_state.current() == State::HangingUp)
|
|
? State::Ended
|
|
: State::Failed;
|
|
|
|
// We want to leave request still being sent and processed even if
|
|
// the call is already destroyed.
|
|
const auto session = &_peer->session();
|
|
const auto weak = base::make_weak(this);
|
|
session->api().request(MTPphone_LeaveGroupCall(
|
|
inputCall(),
|
|
MTP_int(base::take(_joinState.ssrc))
|
|
)).done([=](const MTPUpdates &result) {
|
|
// Here 'this' could be destroyed by updates, so we set Ended after
|
|
// updates being handled, but in a guarded way.
|
|
crl::on_main(weak, [=] { setState(finalState); });
|
|
session->api().applyUpdates(result);
|
|
}).fail(crl::guard(weak, [=](const MTP::Error &error) {
|
|
setState(finalState);
|
|
})).send();
|
|
}
|
|
|
|
void GroupCall::startScheduledNow() {
|
|
if (!lookupReal()) {
|
|
return;
|
|
}
|
|
_api.request(MTPphone_StartScheduledGroupCall(
|
|
inputCall()
|
|
)).done([=](const MTPUpdates &result) {
|
|
_peer->session().api().applyUpdates(result);
|
|
}).send();
|
|
}
|
|
|
|
void GroupCall::toggleScheduleStartSubscribed(bool subscribed) {
|
|
if (!lookupReal()) {
|
|
return;
|
|
}
|
|
_api.request(MTPphone_ToggleGroupCallStartSubscription(
|
|
inputCall(),
|
|
MTP_bool(subscribed)
|
|
)).done([=](const MTPUpdates &result) {
|
|
_peer->session().api().applyUpdates(result);
|
|
}).send();
|
|
}
|
|
|
|
void GroupCall::setNoiseSuppression(bool enabled) {
|
|
if (_instance) {
|
|
_instance->setIsNoiseSuppressionEnabled(enabled);
|
|
}
|
|
}
|
|
|
|
void GroupCall::addVideoOutput(
|
|
const std::string &endpoint,
|
|
not_null<Webrtc::VideoTrack*> track) {
|
|
addVideoOutput(endpoint, { track->sink() });
|
|
}
|
|
|
|
void GroupCall::setMuted(MuteState mute) {
|
|
const auto set = [=] {
|
|
const auto was = muted();
|
|
const auto wasSpeaking = (was == MuteState::Active)
|
|
|| (was == MuteState::PushToTalk);
|
|
const auto wasMuted = (was == MuteState::Muted)
|
|
|| (was == MuteState::PushToTalk);
|
|
const auto wasRaiseHand = (was == MuteState::RaisedHand);
|
|
_muted = mute;
|
|
const auto now = muted();
|
|
const auto nowSpeaking = (now == MuteState::Active)
|
|
|| (now == MuteState::PushToTalk);
|
|
const auto nowMuted = (now == MuteState::Muted)
|
|
|| (now == MuteState::PushToTalk);
|
|
const auto nowRaiseHand = (now == MuteState::RaisedHand);
|
|
if (wasMuted != nowMuted || wasRaiseHand != nowRaiseHand) {
|
|
applyMeInCallLocally();
|
|
}
|
|
if (mutedByAdmin()) {
|
|
toggleVideo(false);
|
|
toggleScreenSharing(std::nullopt);
|
|
}
|
|
if (wasSpeaking && !nowSpeaking && _joinState.ssrc) {
|
|
_levelUpdates.fire(LevelUpdate{
|
|
.ssrc = _joinState.ssrc,
|
|
.value = 0.f,
|
|
.voice = false,
|
|
.me = true,
|
|
});
|
|
}
|
|
};
|
|
if (mute == MuteState::Active || mute == MuteState::PushToTalk) {
|
|
_delegate->groupCallRequestPermissionsOrFail(crl::guard(this, set));
|
|
} else {
|
|
set();
|
|
}
|
|
}
|
|
|
|
void GroupCall::setMutedAndUpdate(MuteState mute) {
|
|
const auto was = muted();
|
|
|
|
// Active state is sent from _muted changes,
|
|
// because it may be set delayed, after permissions request, not now.
|
|
const auto send = _initialMuteStateSent && (mute != MuteState::Active);
|
|
setMuted(mute);
|
|
if (send) {
|
|
maybeSendMutedUpdate(was);
|
|
}
|
|
}
|
|
|
|
void GroupCall::handlePossibleCreateOrJoinResponse(
|
|
const MTPDupdateGroupCall &data) {
|
|
data.vcall().match([&](const MTPDgroupCall &data) {
|
|
handlePossibleCreateOrJoinResponse(data);
|
|
}, [&](const MTPDgroupCallDiscarded &data) {
|
|
handlePossibleDiscarded(data);
|
|
});
|
|
}
|
|
|
|
void GroupCall::handlePossibleCreateOrJoinResponse(
|
|
const MTPDgroupCall &data) {
|
|
if (_acceptFields) {
|
|
if (!_instance && !_id) {
|
|
const auto input = MTP_inputGroupCall(
|
|
data.vid(),
|
|
data.vaccess_hash());
|
|
const auto scheduleDate = data.vschedule_date().value_or_empty();
|
|
setScheduledDate(scheduleDate);
|
|
if (const auto chat = _peer->asChat()) {
|
|
chat->setGroupCall(input, scheduleDate);
|
|
} else if (const auto group = _peer->asChannel()) {
|
|
group->setGroupCall(input, scheduleDate);
|
|
} else {
|
|
Unexpected("Peer type in GroupCall::join.");
|
|
}
|
|
join(input);
|
|
}
|
|
return;
|
|
} else if (_id != data.vid().v || !_instance) {
|
|
return;
|
|
}
|
|
setScheduledDate(data.vschedule_date().value_or_empty());
|
|
if (const auto streamDcId = data.vstream_dc_id()) {
|
|
_broadcastDcId = MTP::BareDcId(streamDcId->v);
|
|
}
|
|
}
|
|
|
|
void GroupCall::handlePossibleCreateOrJoinResponse(
|
|
const MTPDupdateGroupCallConnection &data) {
|
|
if (data.is_presentation()) {
|
|
if (!_screenInstance) {
|
|
return;
|
|
}
|
|
setScreenInstanceMode(InstanceMode::Rtc);
|
|
data.vparams().match([&](const MTPDdataJSON &data) {
|
|
const auto json = data.vdata().v;
|
|
const auto response = ParseJoinResponse(json);
|
|
const auto endpoint = std::get_if<JoinVideoEndpoint>(&response);
|
|
if (endpoint) {
|
|
setScreenEndpoint(endpoint->id);
|
|
} else {
|
|
LOG(("Call Error: Bad response for 'presentation' flag."));
|
|
}
|
|
_screenInstance->setJoinResponsePayload(json.toStdString());
|
|
});
|
|
} else {
|
|
if (!_instance) {
|
|
return;
|
|
}
|
|
data.vparams().match([&](const MTPDdataJSON &data) {
|
|
const auto json = data.vdata().v;
|
|
const auto response = ParseJoinResponse(json);
|
|
const auto endpoint = std::get_if<JoinVideoEndpoint>(&response);
|
|
if (v::is<JoinBroadcastStream>(response)) {
|
|
if (!_broadcastDcId) {
|
|
LOG(("Api Error: Empty stream_dc_id in groupCall."));
|
|
_broadcastDcId = _peer->session().mtp().mainDcId();
|
|
}
|
|
setInstanceMode(InstanceMode::Stream);
|
|
} else {
|
|
setInstanceMode(InstanceMode::Rtc);
|
|
setCameraEndpoint(endpoint ? endpoint->id : std::string());
|
|
_instance->setJoinResponsePayload(json.toStdString());
|
|
}
|
|
updateRequestedVideoChannels();
|
|
checkMediaChannelDescriptions();
|
|
});
|
|
}
|
|
}
|
|
|
|
void GroupCall::handlePossibleDiscarded(const MTPDgroupCallDiscarded &data) {
|
|
if (data.vid().v == _id) {
|
|
LOG(("Call Info: Hangup after groupCallDiscarded."));
|
|
_joinState.finish();
|
|
hangup();
|
|
}
|
|
}
|
|
|
|
void GroupCall::checkMediaChannelDescriptions(
|
|
Fn<bool(uint32)> resolved) {
|
|
const auto real = lookupReal();
|
|
if (!real || (_instanceMode == InstanceMode::None)) {
|
|
return;
|
|
}
|
|
for (auto i = begin(_mediaChannelDescriptionses)
|
|
; i != end(_mediaChannelDescriptionses);) {
|
|
if (mediaChannelDescriptionsFill(i->get(), resolved)) {
|
|
i = _mediaChannelDescriptionses.erase(i);
|
|
} else {
|
|
++i;
|
|
}
|
|
}
|
|
if (!_unresolvedSsrcs.empty()) {
|
|
real->resolveParticipants(base::take(_unresolvedSsrcs));
|
|
}
|
|
}
|
|
|
|
void GroupCall::handleUpdate(const MTPUpdate &update) {
|
|
update.match([&](const MTPDupdateGroupCall &data) {
|
|
handleUpdate(data);
|
|
}, [&](const MTPDupdateGroupCallParticipants &data) {
|
|
handleUpdate(data);
|
|
}, [](const auto &) {
|
|
Unexpected("Type in Instance::applyGroupCallUpdateChecked.");
|
|
});
|
|
}
|
|
|
|
void GroupCall::handleUpdate(const MTPDupdateGroupCall &data) {
|
|
data.vcall().match([](const MTPDgroupCall &) {
|
|
}, [&](const MTPDgroupCallDiscarded &data) {
|
|
handlePossibleDiscarded(data);
|
|
});
|
|
}
|
|
|
|
void GroupCall::handleUpdate(const MTPDupdateGroupCallParticipants &data) {
|
|
const auto callId = data.vcall().match([](const auto &data) {
|
|
return data.vid().v;
|
|
});
|
|
if (_id != callId) {
|
|
return;
|
|
}
|
|
const auto state = _state.current();
|
|
const auto joined = (state == State::Joined)
|
|
|| (state == State::Connecting);
|
|
for (const auto &participant : data.vparticipants().v) {
|
|
participant.match([&](const MTPDgroupCallParticipant &data) {
|
|
const auto isSelf = data.is_self()
|
|
|| (data.is_min()
|
|
&& peerFromMTP(data.vpeer()) == joinAs()->id);
|
|
if (!isSelf) {
|
|
applyOtherParticipantUpdate(data);
|
|
} else if (joined) {
|
|
applySelfUpdate(data);
|
|
} else {
|
|
_queuedSelfUpdates.push_back(participant);
|
|
}
|
|
});
|
|
}
|
|
}
|
|
|
|
void GroupCall::applyQueuedSelfUpdates() {
|
|
const auto weak = base::make_weak(this);
|
|
while (weak
|
|
&& !_queuedSelfUpdates.empty()
|
|
&& (_state.current() == State::Joined
|
|
|| _state.current() == State::Connecting)) {
|
|
const auto update = _queuedSelfUpdates.front();
|
|
_queuedSelfUpdates.erase(_queuedSelfUpdates.begin());
|
|
update.match([&](const MTPDgroupCallParticipant &data) {
|
|
applySelfUpdate(data);
|
|
});
|
|
}
|
|
}
|
|
|
|
void GroupCall::applySelfUpdate(const MTPDgroupCallParticipant &data) {
|
|
if (data.is_left()) {
|
|
if (data.vsource().v == _joinState.ssrc) {
|
|
// I was removed from the call, rejoin.
|
|
LOG(("Call Info: "
|
|
"Rejoin after got 'left' with my ssrc."));
|
|
setState(State::Joining);
|
|
rejoin();
|
|
}
|
|
return;
|
|
} else if (data.vsource().v != _joinState.ssrc) {
|
|
if (!_mySsrcs.contains(data.vsource().v)) {
|
|
// I joined from another device, hangup.
|
|
LOG(("Call Info: "
|
|
"Hangup after '!left' with ssrc %1, my %2."
|
|
).arg(data.vsource().v
|
|
).arg(_joinState.ssrc));
|
|
_joinState.finish();
|
|
hangup();
|
|
} else {
|
|
LOG(("Call Info: "
|
|
"Some old 'self' with '!left' and ssrc %1, my %2."
|
|
).arg(data.vsource().v
|
|
).arg(_joinState.ssrc));
|
|
}
|
|
return;
|
|
}
|
|
if (data.is_muted() && !data.is_can_self_unmute()) {
|
|
setMuted(data.vraise_hand_rating().value_or_empty()
|
|
? MuteState::RaisedHand
|
|
: MuteState::ForceMuted);
|
|
} else if (_instanceMode == InstanceMode::Stream) {
|
|
LOG(("Call Info: Rejoin after unforcemute in stream mode."));
|
|
setState(State::Joining);
|
|
rejoin();
|
|
} else if (mutedByAdmin()) {
|
|
setMuted(MuteState::Muted);
|
|
if (!_instanceTransitioning) {
|
|
notifyAboutAllowedToSpeak();
|
|
}
|
|
} else if (data.is_muted() && muted() != MuteState::Muted) {
|
|
setMuted(MuteState::Muted);
|
|
}
|
|
}
|
|
|
|
void GroupCall::applyOtherParticipantUpdate(
|
|
const MTPDgroupCallParticipant &data) {
|
|
if (data.is_min()) {
|
|
// No real information about mutedByMe or my custom volume.
|
|
return;
|
|
}
|
|
const auto participantPeer = _peer->owner().peer(
|
|
peerFromMTP(data.vpeer()));
|
|
if (!LookupParticipant(_peer, _id, participantPeer)) {
|
|
return;
|
|
}
|
|
_otherParticipantStateValue.fire(Group::ParticipantState{
|
|
.peer = participantPeer,
|
|
.volume = data.vvolume().value_or_empty(),
|
|
.mutedByMe = data.is_muted_by_you(),
|
|
});
|
|
}
|
|
|
|
void GroupCall::setupMediaDevices() {
|
|
_mediaDevices->audioInputId(
|
|
) | rpl::start_with_next([=](QString id) {
|
|
_audioInputId = id;
|
|
if (_instance) {
|
|
_instance->setAudioInputDevice(id.toStdString());
|
|
}
|
|
}, _lifetime);
|
|
|
|
_mediaDevices->audioOutputId(
|
|
) | rpl::start_with_next([=](QString id) {
|
|
_audioOutputId = id;
|
|
if (_instance) {
|
|
_instance->setAudioOutputDevice(id.toStdString());
|
|
}
|
|
}, _lifetime);
|
|
|
|
_mediaDevices->videoInputId(
|
|
) | rpl::start_with_next([=](QString id) {
|
|
_cameraInputId = id;
|
|
if (_cameraCapture) {
|
|
_cameraCapture->switchToDevice(id.toStdString(), false);
|
|
}
|
|
}, _lifetime);
|
|
}
|
|
|
|
int GroupCall::activeVideoSendersCount() const {
|
|
auto result = 0;
|
|
for (const auto &[endpoint, track] : _activeVideoTracks) {
|
|
if (endpoint.type == VideoEndpointType::Camera) {
|
|
++result;
|
|
} else {
|
|
auto sharesCameraToo = false;
|
|
for (const auto &[other, _] : _activeVideoTracks) {
|
|
if (other.type == VideoEndpointType::Camera
|
|
&& other.peer == endpoint.peer) {
|
|
sharesCameraToo = true;
|
|
break;
|
|
}
|
|
}
|
|
if (!sharesCameraToo) {
|
|
++result;
|
|
}
|
|
}
|
|
}
|
|
return result;
|
|
}
|
|
|
|
bool GroupCall::emitShareCameraError() {
|
|
const auto emitError = [=](Error error) {
|
|
emitShareCameraError(error);
|
|
return true;
|
|
};
|
|
if (const auto real = lookupReal()
|
|
; real && activeVideoSendersCount() >= real->unmutedVideoLimit()) {
|
|
return emitError(Error::DisabledNoCamera);
|
|
} else if (!videoIsWorking()) {
|
|
return emitError(Error::DisabledNoCamera);
|
|
} else if (mutedByAdmin()) {
|
|
return emitError(Error::MutedNoCamera);
|
|
} else if (Webrtc::GetVideoInputList().empty()) {
|
|
return emitError(Error::NoCamera);
|
|
}
|
|
return false;
|
|
}
|
|
|
|
void GroupCall::emitShareCameraError(Error error) {
|
|
_cameraState = Webrtc::VideoState::Inactive;
|
|
if (error == Error::CameraFailed
|
|
&& Webrtc::GetVideoInputList().empty()) {
|
|
error = Error::NoCamera;
|
|
}
|
|
_errors.fire_copy(error);
|
|
}
|
|
|
|
bool GroupCall::emitShareScreenError() {
|
|
const auto emitError = [=](Error error) {
|
|
emitShareScreenError(error);
|
|
return true;
|
|
};
|
|
if (const auto real = lookupReal()
|
|
; real && activeVideoSendersCount() >= real->unmutedVideoLimit()) {
|
|
return emitError(Error::DisabledNoScreen);
|
|
} else if (!videoIsWorking()) {
|
|
return emitError(Error::DisabledNoScreen);
|
|
} else if (mutedByAdmin()) {
|
|
return emitError(Error::MutedNoScreen);
|
|
}
|
|
return false;
|
|
}
|
|
|
|
void GroupCall::emitShareScreenError(Error error) {
|
|
_screenState = Webrtc::VideoState::Inactive;
|
|
_errors.fire_copy(error);
|
|
}
|
|
|
|
void GroupCall::setupOutgoingVideo() {
|
|
using Webrtc::VideoState;
|
|
|
|
_cameraState.value(
|
|
) | rpl::combine_previous(
|
|
) | rpl::filter([=](VideoState previous, VideoState state) {
|
|
// Recursive entrance may happen if error happens when activating.
|
|
return (previous != state);
|
|
}) | rpl::start_with_next([=](VideoState previous, VideoState state) {
|
|
const auto wasActive = (previous != VideoState::Inactive);
|
|
const auto nowPaused = (state == VideoState::Paused);
|
|
const auto nowActive = (state != VideoState::Inactive);
|
|
if (wasActive == nowActive) {
|
|
Assert(wasActive && nowActive);
|
|
sendSelfUpdate(SendUpdateType::CameraPaused);
|
|
markTrackPaused({
|
|
VideoEndpointType::Camera,
|
|
joinAs(),
|
|
_cameraEndpoint
|
|
}, nowPaused);
|
|
return;
|
|
}
|
|
if (nowActive) {
|
|
if (emitShareCameraError()) {
|
|
return;
|
|
} else if (!_cameraCapture) {
|
|
_cameraCapture = _delegate->groupCallGetVideoCapture(
|
|
_cameraInputId);
|
|
if (!_cameraCapture) {
|
|
return emitShareCameraError(Error::CameraFailed);
|
|
}
|
|
const auto weak = base::make_weak(this);
|
|
_cameraCapture->setOnFatalError([=] {
|
|
crl::on_main(weak, [=] {
|
|
emitShareCameraError(Error::CameraFailed);
|
|
});
|
|
});
|
|
} else {
|
|
_cameraCapture->switchToDevice(
|
|
_cameraInputId.toStdString(),
|
|
false);
|
|
}
|
|
if (_instance) {
|
|
_instance->setVideoCapture(_cameraCapture);
|
|
}
|
|
_cameraCapture->setState(tgcalls::VideoState::Active);
|
|
} else if (_cameraCapture) {
|
|
_cameraCapture->setState(tgcalls::VideoState::Inactive);
|
|
}
|
|
_isSharingCamera = nowActive;
|
|
markEndpointActive({
|
|
VideoEndpointType::Camera,
|
|
joinAs(),
|
|
_cameraEndpoint
|
|
}, nowActive, nowPaused);
|
|
sendSelfUpdate(SendUpdateType::CameraStopped);
|
|
applyMeInCallLocally();
|
|
}, _lifetime);
|
|
|
|
_screenState.value(
|
|
) | rpl::combine_previous(
|
|
) | rpl::filter([=](VideoState previous, VideoState state) {
|
|
// Recursive entrance may happen if error happens when activating.
|
|
return (previous != state);
|
|
}) | rpl::start_with_next([=](VideoState previous, VideoState state) {
|
|
const auto wasActive = (previous != VideoState::Inactive);
|
|
const auto nowPaused = (state == VideoState::Paused);
|
|
const auto nowActive = (state != VideoState::Inactive);
|
|
if (wasActive == nowActive) {
|
|
Assert(wasActive && nowActive);
|
|
sendSelfUpdate(SendUpdateType::ScreenPaused);
|
|
markTrackPaused({
|
|
VideoEndpointType::Screen,
|
|
joinAs(),
|
|
_screenEndpoint
|
|
}, nowPaused);
|
|
return;
|
|
}
|
|
if (nowActive) {
|
|
if (emitShareScreenError()) {
|
|
return;
|
|
} else if (!_screenCapture) {
|
|
_screenCapture = std::shared_ptr<
|
|
tgcalls::VideoCaptureInterface
|
|
>(tgcalls::VideoCaptureInterface::Create(
|
|
tgcalls::StaticThreads::getThreads(),
|
|
_screenDeviceId.toStdString()));
|
|
if (!_screenCapture) {
|
|
return emitShareScreenError(Error::ScreenFailed);
|
|
}
|
|
const auto weak = base::make_weak(this);
|
|
_screenCapture->setOnFatalError([=] {
|
|
crl::on_main(weak, [=] {
|
|
emitShareScreenError(Error::ScreenFailed);
|
|
});
|
|
});
|
|
_screenCapture->setOnPause([=](bool paused) {
|
|
crl::on_main(weak, [=] {
|
|
if (isSharingScreen()) {
|
|
_screenState = paused
|
|
? VideoState::Paused
|
|
: VideoState::Active;
|
|
}
|
|
});
|
|
});
|
|
} else {
|
|
_screenCapture->switchToDevice(
|
|
_screenDeviceId.toStdString(),
|
|
true);
|
|
}
|
|
if (_screenInstance) {
|
|
_screenInstance->setVideoCapture(_screenCapture);
|
|
}
|
|
_screenCapture->setState(tgcalls::VideoState::Active);
|
|
} else if (_screenCapture) {
|
|
_screenCapture->setState(tgcalls::VideoState::Inactive);
|
|
}
|
|
_isSharingScreen = nowActive;
|
|
markEndpointActive({
|
|
VideoEndpointType::Screen,
|
|
joinAs(),
|
|
_screenEndpoint
|
|
}, nowActive, nowPaused);
|
|
_screenJoinState.nextActionPending = true;
|
|
checkNextJoinAction();
|
|
}, _lifetime);
|
|
}
|
|
|
|
void GroupCall::changeTitle(const QString &title) {
|
|
const auto real = lookupReal();
|
|
if (!real || real->title() == title) {
|
|
return;
|
|
}
|
|
|
|
_api.request(MTPphone_EditGroupCallTitle(
|
|
inputCall(),
|
|
MTP_string(title)
|
|
)).done([=](const MTPUpdates &result) {
|
|
_peer->session().api().applyUpdates(result);
|
|
_titleChanged.fire({});
|
|
}).fail([=](const MTP::Error &error) {
|
|
}).send();
|
|
}
|
|
|
|
void GroupCall::toggleRecording(
|
|
bool enabled,
|
|
const QString &title,
|
|
bool video,
|
|
bool videoPortrait) {
|
|
const auto real = lookupReal();
|
|
if (!real) {
|
|
return;
|
|
}
|
|
|
|
const auto already = (real->recordStartDate() != 0);
|
|
if (already == enabled) {
|
|
return;
|
|
}
|
|
|
|
if (!enabled) {
|
|
_recordingStoppedByMe = true;
|
|
}
|
|
using Flag = MTPphone_ToggleGroupCallRecord::Flag;
|
|
_api.request(MTPphone_ToggleGroupCallRecord(
|
|
MTP_flags((enabled ? Flag::f_start : Flag(0))
|
|
| (video ? Flag::f_video : Flag(0))
|
|
| (title.isEmpty() ? Flag(0) : Flag::f_title)),
|
|
inputCall(),
|
|
MTP_string(title),
|
|
MTP_bool(videoPortrait)
|
|
)).done([=](const MTPUpdates &result) {
|
|
_peer->session().api().applyUpdates(result);
|
|
_recordingStoppedByMe = false;
|
|
}).fail([=](const MTP::Error &error) {
|
|
_recordingStoppedByMe = false;
|
|
}).send();
|
|
}
|
|
|
|
bool GroupCall::tryCreateController() {
|
|
if (_instance) {
|
|
return false;
|
|
}
|
|
const auto &settings = Core::App().settings();
|
|
|
|
const auto weak = base::make_weak(&_instanceGuard);
|
|
const auto myLevel = std::make_shared<tgcalls::GroupLevelValue>();
|
|
tgcalls::GroupInstanceDescriptor descriptor = {
|
|
.threads = tgcalls::StaticThreads::getThreads(),
|
|
.config = tgcalls::GroupConfig{
|
|
},
|
|
.networkStateUpdated = [=](tgcalls::GroupNetworkState networkState) {
|
|
crl::on_main(weak, [=] { setInstanceConnected(networkState); });
|
|
},
|
|
.audioLevelsUpdated = [=](const tgcalls::GroupLevelsUpdate &data) {
|
|
const auto &updates = data.updates;
|
|
if (updates.empty()) {
|
|
return;
|
|
} else if (updates.size() == 1 && !updates.front().ssrc) {
|
|
const auto &value = updates.front().value;
|
|
// Don't send many 0 while we're muted.
|
|
if (myLevel->level == value.level
|
|
&& myLevel->voice == value.voice) {
|
|
return;
|
|
}
|
|
*myLevel = updates.front().value;
|
|
}
|
|
crl::on_main(weak, [=] { audioLevelsUpdated(data); });
|
|
},
|
|
.initialInputDeviceId = _audioInputId.toStdString(),
|
|
.initialOutputDeviceId = _audioOutputId.toStdString(),
|
|
.createAudioDeviceModule = Webrtc::AudioDeviceModuleCreator(
|
|
settings.callAudioBackend()),
|
|
.videoCapture = _cameraCapture,
|
|
.requestCurrentTime = [=, call = base::make_weak(this)](
|
|
std::function<void(int64_t)> done) {
|
|
auto result = std::make_shared<RequestCurrentTimeTask>(
|
|
call,
|
|
std::move(done));
|
|
crl::on_main(weak, [=] {
|
|
result->done(approximateServerTimeInMs());
|
|
});
|
|
return result;
|
|
},
|
|
.requestAudioBroadcastPart = [=, call = base::make_weak(this)](
|
|
int64_t time,
|
|
int64_t period,
|
|
std::function<void(tgcalls::BroadcastPart &&)> done) {
|
|
auto result = std::make_shared<LoadPartTask>(
|
|
call,
|
|
time,
|
|
period,
|
|
std::move(done));
|
|
crl::on_main(weak, [=]() mutable {
|
|
broadcastPartStart(std::move(result));
|
|
});
|
|
return result;
|
|
},
|
|
.requestVideoBroadcastPart = [=, call = base::make_weak(this)](
|
|
int64_t time,
|
|
int64_t period,
|
|
int32_t channel,
|
|
tgcalls::VideoChannelDescription::Quality quality,
|
|
std::function<void(tgcalls::BroadcastPart &&)> done) {
|
|
auto result = std::make_shared<LoadPartTask>(
|
|
call,
|
|
time,
|
|
period,
|
|
channel,
|
|
quality,
|
|
std::move(done));
|
|
crl::on_main(weak, [=]() mutable {
|
|
broadcastPartStart(std::move(result));
|
|
});
|
|
return result;
|
|
},
|
|
.videoContentType = tgcalls::VideoContentType::Generic,
|
|
.initialEnableNoiseSuppression
|
|
= settings.groupCallNoiseSuppression(),
|
|
.requestMediaChannelDescriptions = [=, call = base::make_weak(this)](
|
|
const std::vector<uint32_t> &ssrcs,
|
|
std::function<void(
|
|
std::vector<tgcalls::MediaChannelDescription> &&)> done) {
|
|
auto result = std::make_shared<MediaChannelDescriptionsTask>(
|
|
call,
|
|
ssrcs,
|
|
std::move(done));
|
|
crl::on_main(weak, [=]() mutable {
|
|
mediaChannelDescriptionsStart(std::move(result));
|
|
});
|
|
return result;
|
|
},
|
|
};
|
|
if (Logs::DebugEnabled()) {
|
|
auto callLogFolder = cWorkingDir() + qsl("DebugLogs");
|
|
auto callLogPath = callLogFolder + qsl("/last_group_call_log.txt");
|
|
auto callLogNative = QDir::toNativeSeparators(callLogPath);
|
|
#ifdef Q_OS_WIN
|
|
descriptor.config.logPath.data = callLogNative.toStdWString();
|
|
#else // Q_OS_WIN
|
|
const auto callLogUtf = QFile::encodeName(callLogNative);
|
|
descriptor.config.logPath.data.resize(callLogUtf.size());
|
|
ranges::copy(callLogUtf, descriptor.config.logPath.data.begin());
|
|
#endif // Q_OS_WIN
|
|
QFile(callLogPath).remove();
|
|
QDir().mkpath(callLogFolder);
|
|
}
|
|
|
|
LOG(("Call Info: Creating group instance"));
|
|
_instance = std::make_unique<tgcalls::GroupInstanceCustomImpl>(
|
|
std::move(descriptor));
|
|
|
|
updateInstanceMuteState();
|
|
updateInstanceVolumes();
|
|
for (auto &[endpoint, sink] : base::take(_pendingVideoOutputs)) {
|
|
_instance->addIncomingVideoOutput(endpoint, std::move(sink.data));
|
|
}
|
|
//raw->setAudioOutputDuckingEnabled(settings.callAudioDuckingEnabled());
|
|
return true;
|
|
}
|
|
|
|
bool GroupCall::tryCreateScreencast() {
|
|
if (_screenInstance) {
|
|
return false;
|
|
}
|
|
|
|
const auto weak = base::make_weak(&_screenInstanceGuard);
|
|
tgcalls::GroupInstanceDescriptor descriptor = {
|
|
.threads = tgcalls::StaticThreads::getThreads(),
|
|
.config = tgcalls::GroupConfig{
|
|
},
|
|
.networkStateUpdated = [=](tgcalls::GroupNetworkState networkState) {
|
|
crl::on_main(weak, [=] {
|
|
setScreenInstanceConnected(networkState);
|
|
});
|
|
},
|
|
.createAudioDeviceModule = Webrtc::LoopbackAudioDeviceModuleCreator(),
|
|
.videoCapture = _screenCapture,
|
|
.videoContentType = tgcalls::VideoContentType::Screencast,
|
|
};
|
|
|
|
LOG(("Call Info: Creating group screen instance"));
|
|
_screenInstance = std::make_unique<tgcalls::GroupInstanceCustomImpl>(
|
|
std::move(descriptor));
|
|
|
|
_screenInstance->setIsMuted(!_screenWithAudio);
|
|
|
|
return true;
|
|
}
|
|
|
|
void GroupCall::broadcastPartStart(std::shared_ptr<LoadPartTask> task) {
|
|
const auto raw = task.get();
|
|
const auto time = raw->time();
|
|
const auto scale = raw->scale();
|
|
const auto videoChannel = raw->videoChannel();
|
|
const auto videoQuality = raw->videoQuality();
|
|
const auto finish = [=](tgcalls::BroadcastPart &&part) {
|
|
raw->done(std::move(part));
|
|
_broadcastParts.erase(raw);
|
|
};
|
|
using Status = tgcalls::BroadcastPart::Status;
|
|
using Quality = tgcalls::VideoChannelDescription::Quality;
|
|
using Flag = MTPDinputGroupCallStream::Flag;
|
|
const auto requestId = _api.request(MTPupload_GetFile(
|
|
MTP_flags(0),
|
|
MTP_inputGroupCallStream(
|
|
MTP_flags(videoChannel
|
|
? (Flag::f_video_channel | Flag::f_video_quality)
|
|
: Flag(0)),
|
|
inputCall(),
|
|
MTP_long(time),
|
|
MTP_int(scale),
|
|
MTP_int(videoChannel),
|
|
MTP_int((videoQuality == Quality::Full)
|
|
? 2
|
|
: (videoQuality == Quality::Medium)
|
|
? 1
|
|
: 0)),
|
|
MTP_int(0),
|
|
MTP_int(128 * 1024)
|
|
)).done([=](
|
|
const MTPupload_File &result,
|
|
const MTP::Response &response) {
|
|
result.match([&](const MTPDupload_file &data) {
|
|
const auto size = data.vbytes().v.size();
|
|
auto bytes = std::vector<uint8_t>(size);
|
|
memcpy(bytes.data(), data.vbytes().v.constData(), size);
|
|
finish({
|
|
.timestampMilliseconds = time,
|
|
.responseTimestamp = TimestampFromMsgId(response.outerMsgId),
|
|
.status = Status::Success,
|
|
.data = std::move(bytes),
|
|
});
|
|
}, [&](const MTPDupload_fileCdnRedirect &data) {
|
|
LOG(("Voice Chat Stream Error: fileCdnRedirect received."));
|
|
finish({
|
|
.timestampMilliseconds = time,
|
|
.responseTimestamp = TimestampFromMsgId(response.outerMsgId),
|
|
.status = Status::ResyncNeeded,
|
|
});
|
|
});
|
|
}).fail([=](const MTP::Error &error, const MTP::Response &response) {
|
|
if (error.type() == u"GROUPCALL_JOIN_MISSING"_q
|
|
|| error.type() == u"GROUPCALL_FORBIDDEN"_q) {
|
|
for (const auto &[task, part] : _broadcastParts) {
|
|
_api.request(part.requestId).cancel();
|
|
}
|
|
setState(State::Joining);
|
|
rejoin();
|
|
return;
|
|
}
|
|
const auto status = (MTP::IsFloodError(error)
|
|
|| error.type() == u"TIME_TOO_BIG"_q)
|
|
? Status::NotReady
|
|
: Status::ResyncNeeded;
|
|
finish({
|
|
.timestampMilliseconds = time,
|
|
.responseTimestamp = TimestampFromMsgId(response.outerMsgId),
|
|
.status = status,
|
|
});
|
|
}).handleAllErrors().toDC(
|
|
MTP::groupCallStreamDcId(_broadcastDcId)
|
|
).send();
|
|
_broadcastParts.emplace(raw, LoadingPart{ std::move(task), requestId });
|
|
}
|
|
|
|
void GroupCall::broadcastPartCancel(not_null<LoadPartTask*> task) {
|
|
const auto i = _broadcastParts.find(task);
|
|
if (i != end(_broadcastParts)) {
|
|
_api.request(i->second.requestId).cancel();
|
|
_broadcastParts.erase(i);
|
|
}
|
|
}
|
|
|
|
void GroupCall::mediaChannelDescriptionsStart(
|
|
std::shared_ptr<MediaChannelDescriptionsTask> task) {
|
|
const auto real = lookupReal();
|
|
if (!real || (_instanceMode == InstanceMode::None)) {
|
|
for (const auto ssrc : task->ssrcs()) {
|
|
_unresolvedSsrcs.emplace(ssrc);
|
|
}
|
|
_mediaChannelDescriptionses.emplace(std::move(task));
|
|
return;
|
|
}
|
|
if (!mediaChannelDescriptionsFill(task.get())) {
|
|
_mediaChannelDescriptionses.emplace(std::move(task));
|
|
Assert(!_unresolvedSsrcs.empty());
|
|
}
|
|
if (!_unresolvedSsrcs.empty()) {
|
|
real->resolveParticipants(base::take(_unresolvedSsrcs));
|
|
}
|
|
}
|
|
|
|
bool GroupCall::mediaChannelDescriptionsFill(
|
|
not_null<MediaChannelDescriptionsTask*> task,
|
|
Fn<bool(uint32)> resolved) {
|
|
using Channel = tgcalls::MediaChannelDescription;
|
|
auto result = false;
|
|
const auto real = lookupReal();
|
|
Assert(real != nullptr);
|
|
for (const auto ssrc : task->ssrcs()) {
|
|
const auto add = [&](
|
|
std::optional<Channel> channel,
|
|
bool screen = false) {
|
|
if (task->finishWithAdding(ssrc, std::move(channel), screen)) {
|
|
result = true;
|
|
}
|
|
};
|
|
if (const auto byAudio = real->participantPeerByAudioSsrc(ssrc)) {
|
|
add(Channel{
|
|
.type = Channel::Type::Audio,
|
|
.audioSsrc = ssrc,
|
|
});
|
|
} else if (!resolved) {
|
|
_unresolvedSsrcs.emplace(ssrc);
|
|
} else if (resolved(ssrc)) {
|
|
add(std::nullopt);
|
|
}
|
|
}
|
|
return result;
|
|
}
|
|
|
|
void GroupCall::mediaChannelDescriptionsCancel(
|
|
not_null<MediaChannelDescriptionsTask*> task) {
|
|
const auto i = _mediaChannelDescriptionses.find(task.get());
|
|
if (i != end(_mediaChannelDescriptionses)) {
|
|
_mediaChannelDescriptionses.erase(i);
|
|
}
|
|
}
|
|
|
|
int64 GroupCall::approximateServerTimeInMs() const {
|
|
Expects(_serverTimeMs != 0);
|
|
|
|
return _serverTimeMs + (crl::now() - _serverTimeMsGotAt);
|
|
}
|
|
|
|
void GroupCall::updateRequestedVideoChannels() {
|
|
_requestedVideoChannelsUpdateScheduled = false;
|
|
const auto real = lookupReal();
|
|
if (!real || !_instance) {
|
|
return;
|
|
}
|
|
auto channels = std::vector<tgcalls::VideoChannelDescription>();
|
|
using Quality = tgcalls::VideoChannelDescription::Quality;
|
|
channels.reserve(_activeVideoTracks.size());
|
|
const auto &camera = cameraSharingEndpoint();
|
|
const auto &screen = screenSharingEndpoint();
|
|
auto mediums = 0;
|
|
auto fullcameras = 0;
|
|
auto fullscreencasts = 0;
|
|
for (const auto &[endpoint, video] : _activeVideoTracks) {
|
|
const auto &endpointId = endpoint.id;
|
|
if (endpointId == camera || endpointId == screen) {
|
|
continue;
|
|
}
|
|
const auto participant = real->participantByEndpoint(endpointId);
|
|
const auto params = (participant && participant->ssrc)
|
|
? participant->videoParams.get()
|
|
: nullptr;
|
|
if (!params) {
|
|
continue;
|
|
}
|
|
const auto min = (video->quality == Group::VideoQuality::Full
|
|
&& endpoint.type == VideoEndpointType::Screen)
|
|
? Quality::Full
|
|
: Quality::Thumbnail;
|
|
const auto max = (video->quality == Group::VideoQuality::Full)
|
|
? Quality::Full
|
|
: (video->quality == Group::VideoQuality::Medium
|
|
&& endpoint.type != VideoEndpointType::Screen)
|
|
? Quality::Medium
|
|
: Quality::Thumbnail;
|
|
if (max == Quality::Full) {
|
|
if (endpoint.type == VideoEndpointType::Screen) {
|
|
++fullscreencasts;
|
|
} else {
|
|
++fullcameras;
|
|
}
|
|
} else if (max == Quality::Medium) {
|
|
++mediums;
|
|
}
|
|
channels.push_back({
|
|
.audioSsrc = participant->ssrc,
|
|
.endpointId = endpointId,
|
|
.ssrcGroups = (params->camera.endpointId == endpointId
|
|
? params->camera.ssrcGroups
|
|
: params->screen.ssrcGroups),
|
|
.minQuality = min,
|
|
.maxQuality = max,
|
|
});
|
|
}
|
|
|
|
// We limit `count(Full) * kFullAsMediumsCount + count(medium)`.
|
|
//
|
|
// Try to preserve all qualities; If not
|
|
// Try to preserve all screencasts as Full and cameras as Medium; If not
|
|
// Try to preserve all screencasts as Full; If not
|
|
// Try to preserve all cameras as Medium;
|
|
const auto mediumsCount = mediums
|
|
+ (fullcameras + fullscreencasts) * kFullAsMediumsCount;
|
|
const auto downgradeSome = (mediumsCount > kMaxMediumQualities);
|
|
const auto downgradeAll = (fullscreencasts * kFullAsMediumsCount)
|
|
> kMaxMediumQualities;
|
|
if (downgradeSome) {
|
|
for (auto &channel : channels) {
|
|
if (channel.maxQuality == Quality::Full) {
|
|
const auto camera = (channel.minQuality != Quality::Full);
|
|
if (camera) {
|
|
channel.maxQuality = Quality::Medium;
|
|
} else if (downgradeAll) {
|
|
channel.maxQuality
|
|
= channel.minQuality
|
|
= Quality::Thumbnail;
|
|
--fullscreencasts;
|
|
}
|
|
}
|
|
}
|
|
mediums += fullcameras;
|
|
fullcameras = 0;
|
|
if (downgradeAll) {
|
|
fullscreencasts = 0;
|
|
}
|
|
}
|
|
if (mediums > kMaxMediumQualities) {
|
|
for (auto &channel : channels) {
|
|
if (channel.maxQuality == Quality::Medium) {
|
|
channel.maxQuality = Quality::Thumbnail;
|
|
}
|
|
}
|
|
}
|
|
_instance->setRequestedVideoChannels(std::move(channels));
|
|
}
|
|
|
|
void GroupCall::updateRequestedVideoChannelsDelayed() {
|
|
if (_requestedVideoChannelsUpdateScheduled) {
|
|
return;
|
|
}
|
|
_requestedVideoChannelsUpdateScheduled = true;
|
|
crl::on_main(this, [=] {
|
|
if (_requestedVideoChannelsUpdateScheduled) {
|
|
updateRequestedVideoChannels();
|
|
}
|
|
});
|
|
}
|
|
|
|
void GroupCall::fillActiveVideoEndpoints() {
|
|
const auto real = lookupReal();
|
|
Assert(real != nullptr);
|
|
|
|
const auto me = real->participantByPeer(joinAs());
|
|
if (me && me->videoJoined) {
|
|
_videoIsWorking = true;
|
|
} else {
|
|
_videoIsWorking = false;
|
|
toggleVideo(false);
|
|
toggleScreenSharing(std::nullopt);
|
|
}
|
|
|
|
const auto &large = _videoEndpointLarge.current();
|
|
auto largeFound = false;
|
|
auto endpoints = _activeVideoTracks | ranges::views::transform([](
|
|
const auto &pair) {
|
|
return pair.first;
|
|
});
|
|
auto removed = base::flat_set<VideoEndpoint>(
|
|
begin(endpoints),
|
|
end(endpoints));
|
|
const auto feedOne = [&](VideoEndpoint endpoint, bool paused) {
|
|
if (endpoint.empty()) {
|
|
return;
|
|
} else if (endpoint == large) {
|
|
largeFound = true;
|
|
}
|
|
if (removed.remove(endpoint)) {
|
|
markTrackPaused(endpoint, paused);
|
|
} else {
|
|
markEndpointActive(std::move(endpoint), true, paused);
|
|
}
|
|
};
|
|
using Type = VideoEndpointType;
|
|
for (const auto &participant : real->participants()) {
|
|
const auto camera = GetCameraEndpoint(participant.videoParams);
|
|
if (camera != _cameraEndpoint
|
|
&& camera != _screenEndpoint
|
|
&& participant.peer != joinAs()) {
|
|
const auto paused = IsCameraPaused(participant.videoParams);
|
|
feedOne({ Type::Camera, participant.peer, camera }, paused);
|
|
}
|
|
const auto screen = GetScreenEndpoint(participant.videoParams);
|
|
if (screen != _cameraEndpoint
|
|
&& screen != _screenEndpoint
|
|
&& participant.peer != joinAs()) {
|
|
const auto paused = IsScreenPaused(participant.videoParams);
|
|
feedOne({ Type::Screen, participant.peer, screen }, paused);
|
|
}
|
|
}
|
|
feedOne(
|
|
{ Type::Camera, joinAs(), cameraSharingEndpoint() },
|
|
isCameraPaused());
|
|
feedOne(
|
|
{ Type::Screen, joinAs(), screenSharingEndpoint() },
|
|
isScreenPaused());
|
|
if (large && !largeFound) {
|
|
setVideoEndpointLarge({});
|
|
}
|
|
for (const auto &endpoint : removed) {
|
|
markEndpointActive(endpoint, false, false);
|
|
}
|
|
updateRequestedVideoChannels();
|
|
}
|
|
|
|
void GroupCall::updateInstanceMuteState() {
|
|
Expects(_instance != nullptr);
|
|
|
|
const auto state = muted();
|
|
_instance->setIsMuted(state != MuteState::Active
|
|
&& state != MuteState::PushToTalk);
|
|
}
|
|
|
|
void GroupCall::updateInstanceVolumes() {
|
|
const auto real = lookupReal();
|
|
if (!real) {
|
|
return;
|
|
}
|
|
|
|
const auto &participants = real->participants();
|
|
for (const auto &participant : participants) {
|
|
updateInstanceVolume(std::nullopt, participant);
|
|
}
|
|
}
|
|
|
|
void GroupCall::updateInstanceVolume(
|
|
const std::optional<Data::GroupCallParticipant> &was,
|
|
const Data::GroupCallParticipant &now) {
|
|
const auto nonDefault = now.mutedByMe
|
|
|| (now.volume != Group::kDefaultVolume);
|
|
const auto volumeChanged = was
|
|
? (was->volume != now.volume || was->mutedByMe != now.mutedByMe)
|
|
: nonDefault;
|
|
const auto additionalSsrc = GetAdditionalAudioSsrc(now.videoParams);
|
|
const auto set = now.ssrc
|
|
&& (volumeChanged || (was && was->ssrc != now.ssrc));
|
|
const auto additionalSet = additionalSsrc
|
|
&& (volumeChanged
|
|
|| (was && (GetAdditionalAudioSsrc(was->videoParams)
|
|
!= additionalSsrc)));
|
|
const auto localVolume = now.mutedByMe
|
|
? 0.
|
|
: (now.volume / float64(Group::kDefaultVolume));
|
|
if (set) {
|
|
_instance->setVolume(now.ssrc, localVolume);
|
|
}
|
|
if (additionalSet) {
|
|
_instance->setVolume(additionalSsrc, localVolume);
|
|
}
|
|
}
|
|
|
|
void GroupCall::audioLevelsUpdated(const tgcalls::GroupLevelsUpdate &data) {
|
|
Expects(!data.updates.empty());
|
|
|
|
auto check = false;
|
|
auto checkNow = false;
|
|
const auto now = crl::now();
|
|
const auto meMuted = [&] {
|
|
const auto state = muted();
|
|
return (state != MuteState::Active)
|
|
&& (state != MuteState::PushToTalk);
|
|
};
|
|
for (const auto &[ssrcOrZero, value] : data.updates) {
|
|
const auto ssrc = ssrcOrZero ? ssrcOrZero : _joinState.ssrc;
|
|
if (!ssrc) {
|
|
continue;
|
|
}
|
|
const auto level = value.level;
|
|
const auto voice = value.voice;
|
|
const auto me = (ssrc == _joinState.ssrc);
|
|
const auto ignore = me && meMuted();
|
|
_levelUpdates.fire(LevelUpdate{
|
|
.ssrc = ssrc,
|
|
.value = ignore ? 0.f : level,
|
|
.voice = (!ignore && voice),
|
|
.me = me,
|
|
});
|
|
if (level <= kSpeakLevelThreshold) {
|
|
continue;
|
|
}
|
|
if (me
|
|
&& voice
|
|
&& (!_lastSendProgressUpdate
|
|
|| _lastSendProgressUpdate + kUpdateSendActionEach < now)) {
|
|
_lastSendProgressUpdate = now;
|
|
_peer->session().sendProgressManager().update(
|
|
_history,
|
|
Api::SendProgressType::Speaking);
|
|
}
|
|
|
|
check = true;
|
|
const auto i = _lastSpoke.find(ssrc);
|
|
if (i == _lastSpoke.end()) {
|
|
_lastSpoke.emplace(ssrc, Data::LastSpokeTimes{
|
|
.anything = now,
|
|
.voice = voice ? now : 0,
|
|
});
|
|
checkNow = true;
|
|
} else {
|
|
if ((i->second.anything + kCheckLastSpokeInterval / 3 <= now)
|
|
|| (voice
|
|
&& i->second.voice + kCheckLastSpokeInterval / 3 <= now)) {
|
|
checkNow = true;
|
|
}
|
|
i->second.anything = now;
|
|
if (voice) {
|
|
i->second.voice = now;
|
|
}
|
|
}
|
|
}
|
|
if (checkNow) {
|
|
checkLastSpoke();
|
|
} else if (check && !_lastSpokeCheckTimer.isActive()) {
|
|
_lastSpokeCheckTimer.callEach(kCheckLastSpokeInterval / 2);
|
|
}
|
|
}
|
|
|
|
void GroupCall::checkLastSpoke() {
|
|
const auto real = lookupReal();
|
|
if (!real) {
|
|
return;
|
|
}
|
|
|
|
constexpr auto kKeepInListFor = kCheckLastSpokeInterval * 2;
|
|
static_assert(Data::GroupCall::kSoundStatusKeptFor
|
|
<= kKeepInListFor - (kCheckLastSpokeInterval / 3));
|
|
|
|
auto hasRecent = false;
|
|
const auto now = crl::now();
|
|
auto list = base::take(_lastSpoke);
|
|
for (auto i = list.begin(); i != list.end();) {
|
|
const auto [ssrc, when] = *i;
|
|
if (when.anything + kKeepInListFor >= now) {
|
|
hasRecent = true;
|
|
++i;
|
|
} else {
|
|
i = list.erase(i);
|
|
}
|
|
|
|
// Ignore my levels from microphone if I'm already muted.
|
|
if (ssrc != _joinState.ssrc
|
|
|| muted() == MuteState::Active
|
|
|| muted() == MuteState::PushToTalk) {
|
|
real->applyLastSpoke(ssrc, when, now);
|
|
} else {
|
|
real->applyLastSpoke(ssrc, { crl::time(), crl::time() }, now);
|
|
}
|
|
}
|
|
_lastSpoke = std::move(list);
|
|
|
|
if (!hasRecent) {
|
|
_lastSpokeCheckTimer.cancel();
|
|
} else if (!_lastSpokeCheckTimer.isActive()) {
|
|
_lastSpokeCheckTimer.callEach(kCheckLastSpokeInterval / 3);
|
|
}
|
|
}
|
|
|
|
void GroupCall::checkJoined() {
|
|
if (state() != State::Connecting || !_id || !_joinState.ssrc) {
|
|
return;
|
|
}
|
|
auto sources = QVector<MTPint>(1, MTP_int(_joinState.ssrc));
|
|
if (_screenJoinState.ssrc) {
|
|
sources.push_back(MTP_int(_screenJoinState.ssrc));
|
|
}
|
|
_api.request(MTPphone_CheckGroupCall(
|
|
inputCall(),
|
|
MTP_vector<MTPint>(std::move(sources))
|
|
)).done([=](const MTPVector<MTPint> &result) {
|
|
if (!ranges::contains(result.v, MTP_int(_joinState.ssrc))) {
|
|
LOG(("Call Info: Rejoin after no my ssrc in checkGroupCall."));
|
|
_joinState.nextActionPending = true;
|
|
checkNextJoinAction();
|
|
} else {
|
|
if (state() == State::Connecting) {
|
|
_checkJoinedTimer.callOnce(kCheckJoinedTimeout);
|
|
}
|
|
if (_screenJoinState.ssrc
|
|
&& !ranges::contains(
|
|
result.v,
|
|
MTP_int(_screenJoinState.ssrc))) {
|
|
LOG(("Call Info: "
|
|
"Screen rejoin after _screenSsrc not found."));
|
|
_screenJoinState.nextActionPending = true;
|
|
checkNextJoinAction();
|
|
}
|
|
}
|
|
}).fail([=](const MTP::Error &error) {
|
|
LOG(("Call Info: Full rejoin after error '%1' in checkGroupCall."
|
|
).arg(error.type()));
|
|
rejoin();
|
|
}).send();
|
|
}
|
|
|
|
void GroupCall::setInstanceConnected(
|
|
tgcalls::GroupNetworkState networkState) {
|
|
const auto inTransit = networkState.isTransitioningFromBroadcastToRtc;
|
|
const auto instanceState = !networkState.isConnected
|
|
? InstanceState::Disconnected
|
|
: inTransit
|
|
? InstanceState::TransitionToRtc
|
|
: InstanceState::Connected;
|
|
const auto connected = (instanceState != InstanceState::Disconnected);
|
|
if (_instanceState.current() == instanceState
|
|
&& _instanceTransitioning == inTransit) {
|
|
return;
|
|
}
|
|
const auto nowCanSpeak = connected
|
|
&& _instanceTransitioning
|
|
&& !inTransit
|
|
&& (muted() == MuteState::Muted);
|
|
_instanceTransitioning = inTransit;
|
|
_instanceState = instanceState;
|
|
if (state() == State::Connecting && connected) {
|
|
setState(State::Joined);
|
|
} else if (state() == State::Joined && !connected) {
|
|
setState(State::Connecting);
|
|
}
|
|
if (nowCanSpeak) {
|
|
notifyAboutAllowedToSpeak();
|
|
}
|
|
if (!_hadJoinedState && state() == State::Joined) {
|
|
checkFirstTimeJoined();
|
|
}
|
|
}
|
|
|
|
void GroupCall::setScreenInstanceConnected(
|
|
tgcalls::GroupNetworkState networkState) {
|
|
const auto inTransit = networkState.isTransitioningFromBroadcastToRtc;
|
|
const auto screenInstanceState = !networkState.isConnected
|
|
? InstanceState::Disconnected
|
|
: inTransit
|
|
? InstanceState::TransitionToRtc
|
|
: InstanceState::Connected;
|
|
if (_screenInstanceState.current() == screenInstanceState) {
|
|
return;
|
|
}
|
|
_screenInstanceState = screenInstanceState;
|
|
}
|
|
|
|
void GroupCall::checkFirstTimeJoined() {
|
|
if (_hadJoinedState || state() != State::Joined) {
|
|
return;
|
|
}
|
|
_hadJoinedState = true;
|
|
applyGlobalShortcutChanges();
|
|
_delegate->groupCallPlaySound(Delegate::GroupCallSound::Started);
|
|
}
|
|
|
|
void GroupCall::notifyAboutAllowedToSpeak() {
|
|
if (!_hadJoinedState) {
|
|
return;
|
|
}
|
|
_delegate->groupCallPlaySound(
|
|
Delegate::GroupCallSound::AllowedToSpeak);
|
|
_allowedToSpeakNotifications.fire({});
|
|
}
|
|
|
|
void GroupCall::setInstanceMode(InstanceMode mode) {
|
|
Expects(_instance != nullptr);
|
|
|
|
_instanceMode = mode;
|
|
|
|
using Mode = tgcalls::GroupConnectionMode;
|
|
_instance->setConnectionMode([&] {
|
|
switch (_instanceMode) {
|
|
case InstanceMode::None: return Mode::GroupConnectionModeNone;
|
|
case InstanceMode::Rtc: return Mode::GroupConnectionModeRtc;
|
|
case InstanceMode::Stream: return Mode::GroupConnectionModeBroadcast;
|
|
}
|
|
Unexpected("Mode in GroupCall::setInstanceMode.");
|
|
}(), true);
|
|
}
|
|
|
|
void GroupCall::setScreenInstanceMode(InstanceMode mode) {
|
|
Expects(_screenInstance != nullptr);
|
|
|
|
_screenInstanceMode = mode;
|
|
|
|
using Mode = tgcalls::GroupConnectionMode;
|
|
_screenInstance->setConnectionMode([&] {
|
|
switch (_screenInstanceMode) {
|
|
case InstanceMode::None: return Mode::GroupConnectionModeNone;
|
|
case InstanceMode::Rtc: return Mode::GroupConnectionModeRtc;
|
|
case InstanceMode::Stream: return Mode::GroupConnectionModeBroadcast;
|
|
}
|
|
Unexpected("Mode in GroupCall::setInstanceMode.");
|
|
}(), true);
|
|
}
|
|
|
|
void GroupCall::maybeSendMutedUpdate(MuteState previous) {
|
|
// Send Active <-> !Active or ForceMuted <-> RaisedHand changes.
|
|
const auto now = muted();
|
|
if ((previous == MuteState::Active && now == MuteState::Muted)
|
|
|| (now == MuteState::Active
|
|
&& (previous == MuteState::Muted
|
|
|| previous == MuteState::PushToTalk))) {
|
|
sendSelfUpdate(SendUpdateType::Mute);
|
|
} else if ((now == MuteState::ForceMuted
|
|
&& previous == MuteState::RaisedHand)
|
|
|| (now == MuteState::RaisedHand
|
|
&& previous == MuteState::ForceMuted)) {
|
|
sendSelfUpdate(SendUpdateType::RaiseHand);
|
|
}
|
|
}
|
|
|
|
void GroupCall::sendPendingSelfUpdates() {
|
|
if ((state() != State::Connecting && state() != State::Joined)
|
|
|| _selfUpdateRequestId) {
|
|
return;
|
|
}
|
|
const auto updates = {
|
|
SendUpdateType::Mute,
|
|
SendUpdateType::RaiseHand,
|
|
SendUpdateType::CameraStopped,
|
|
SendUpdateType::CameraPaused,
|
|
SendUpdateType::ScreenPaused,
|
|
};
|
|
for (const auto type : updates) {
|
|
if (type == SendUpdateType::ScreenPaused
|
|
&& _screenJoinState.action != JoinAction::None) {
|
|
continue;
|
|
}
|
|
if (_pendingSelfUpdates & type) {
|
|
_pendingSelfUpdates &= ~type;
|
|
sendSelfUpdate(type);
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
|
|
void GroupCall::sendSelfUpdate(SendUpdateType type) {
|
|
if ((state() != State::Connecting && state() != State::Joined)
|
|
|| _selfUpdateRequestId) {
|
|
_pendingSelfUpdates |= type;
|
|
return;
|
|
}
|
|
using Flag = MTPphone_EditGroupCallParticipant::Flag;
|
|
_selfUpdateRequestId = _api.request(MTPphone_EditGroupCallParticipant(
|
|
MTP_flags((type == SendUpdateType::RaiseHand)
|
|
? Flag::f_raise_hand
|
|
: (type == SendUpdateType::CameraStopped)
|
|
? Flag::f_video_stopped
|
|
: (type == SendUpdateType::CameraPaused)
|
|
? Flag::f_video_paused
|
|
: (type == SendUpdateType::ScreenPaused)
|
|
? Flag::f_presentation_paused
|
|
: Flag::f_muted),
|
|
inputCall(),
|
|
joinAs()->input,
|
|
MTP_bool(muted() != MuteState::Active),
|
|
MTP_int(100000), // volume
|
|
MTP_bool(muted() == MuteState::RaisedHand),
|
|
MTP_bool(!isSharingCamera()),
|
|
MTP_bool(isCameraPaused()),
|
|
MTP_bool(isScreenPaused())
|
|
)).done([=](const MTPUpdates &result) {
|
|
_selfUpdateRequestId = 0;
|
|
_peer->session().api().applyUpdates(result);
|
|
sendPendingSelfUpdates();
|
|
}).fail([=](const MTP::Error &error) {
|
|
_selfUpdateRequestId = 0;
|
|
if (error.type() == u"GROUPCALL_FORBIDDEN"_q) {
|
|
LOG(("Call Info: Rejoin after error '%1' in editGroupCallMember."
|
|
).arg(error.type()));
|
|
rejoin();
|
|
}
|
|
}).send();
|
|
}
|
|
|
|
void GroupCall::pinVideoEndpoint(VideoEndpoint endpoint) {
|
|
_videoEndpointPinned = false;
|
|
if (endpoint) {
|
|
setVideoEndpointLarge(std::move(endpoint));
|
|
_videoEndpointPinned = true;
|
|
}
|
|
}
|
|
|
|
void GroupCall::showVideoEndpointLarge(VideoEndpoint endpoint) {
|
|
if (_videoEndpointLarge.current() == endpoint) {
|
|
return;
|
|
}
|
|
_videoEndpointPinned = false;
|
|
setVideoEndpointLarge(std::move(endpoint));
|
|
_videoLargeTillTime = crl::now() + kFixManualLargeVideoDuration;
|
|
}
|
|
|
|
void GroupCall::setVideoEndpointLarge(VideoEndpoint endpoint) {
|
|
if (!endpoint) {
|
|
_videoEndpointPinned = false;
|
|
}
|
|
_videoEndpointLarge = endpoint;
|
|
}
|
|
|
|
void GroupCall::requestVideoQuality(
|
|
const VideoEndpoint &endpoint,
|
|
Group::VideoQuality quality) {
|
|
if (!endpoint) {
|
|
return;
|
|
}
|
|
const auto i = _activeVideoTracks.find(endpoint);
|
|
if (i == end(_activeVideoTracks) || i->second->quality == quality) {
|
|
return;
|
|
}
|
|
i->second->quality = quality;
|
|
updateRequestedVideoChannelsDelayed();
|
|
}
|
|
|
|
void GroupCall::setCurrentAudioDevice(bool input, const QString &deviceId) {
|
|
if (input) {
|
|
_mediaDevices->switchToAudioInput(deviceId);
|
|
} else {
|
|
_mediaDevices->switchToAudioOutput(deviceId);
|
|
}
|
|
}
|
|
|
|
void GroupCall::toggleMute(const Group::MuteRequest &data) {
|
|
if (data.locallyOnly) {
|
|
applyParticipantLocally(data.peer, data.mute, std::nullopt);
|
|
} else {
|
|
editParticipant(data.peer, data.mute, std::nullopt);
|
|
}
|
|
}
|
|
|
|
void GroupCall::changeVolume(const Group::VolumeRequest &data) {
|
|
if (data.locallyOnly) {
|
|
applyParticipantLocally(data.peer, false, data.volume);
|
|
} else {
|
|
editParticipant(data.peer, false, data.volume);
|
|
}
|
|
}
|
|
|
|
void GroupCall::editParticipant(
|
|
not_null<PeerData*> participantPeer,
|
|
bool mute,
|
|
std::optional<int> volume) {
|
|
const auto participant = LookupParticipant(_peer, _id, participantPeer);
|
|
if (!participant) {
|
|
return;
|
|
}
|
|
applyParticipantLocally(participantPeer, mute, volume);
|
|
|
|
using Flag = MTPphone_EditGroupCallParticipant::Flag;
|
|
const auto flags = Flag::f_muted
|
|
| (volume.has_value() ? Flag::f_volume : Flag(0));
|
|
_api.request(MTPphone_EditGroupCallParticipant(
|
|
MTP_flags(flags),
|
|
inputCall(),
|
|
participantPeer->input,
|
|
MTP_bool(mute),
|
|
MTP_int(std::clamp(volume.value_or(0), 1, Group::kMaxVolume)),
|
|
MTPBool(), // raise_hand
|
|
MTPBool(), // video_muted
|
|
MTPBool(), // video_paused
|
|
MTPBool() // presentation_paused
|
|
)).done([=](const MTPUpdates &result) {
|
|
_peer->session().api().applyUpdates(result);
|
|
}).fail([=](const MTP::Error &error) {
|
|
if (error.type() == u"GROUPCALL_FORBIDDEN"_q) {
|
|
LOG(("Call Info: Rejoin after error '%1' in editGroupCallMember."
|
|
).arg(error.type()));
|
|
rejoin();
|
|
}
|
|
}).send();
|
|
}
|
|
|
|
std::variant<int, not_null<UserData*>> GroupCall::inviteUsers(
|
|
const std::vector<not_null<UserData*>> &users) {
|
|
const auto real = lookupReal();
|
|
if (!real) {
|
|
return 0;
|
|
}
|
|
const auto owner = &_peer->owner();
|
|
|
|
auto count = 0;
|
|
auto slice = QVector<MTPInputUser>();
|
|
auto result = std::variant<int, not_null<UserData*>>(0);
|
|
slice.reserve(kMaxInvitePerSlice);
|
|
const auto sendSlice = [&] {
|
|
count += slice.size();
|
|
_api.request(MTPphone_InviteToGroupCall(
|
|
inputCall(),
|
|
MTP_vector<MTPInputUser>(slice)
|
|
)).done([=](const MTPUpdates &result) {
|
|
_peer->session().api().applyUpdates(result);
|
|
}).send();
|
|
slice.clear();
|
|
};
|
|
for (const auto &user : users) {
|
|
if (!count && slice.empty()) {
|
|
result = user;
|
|
}
|
|
owner->registerInvitedToCallUser(_id, _peer, user);
|
|
slice.push_back(user->inputUser);
|
|
if (slice.size() == kMaxInvitePerSlice) {
|
|
sendSlice();
|
|
}
|
|
}
|
|
if (count != 0 || slice.size() != 1) {
|
|
result = int(count + slice.size());
|
|
}
|
|
if (!slice.empty()) {
|
|
sendSlice();
|
|
}
|
|
return result;
|
|
}
|
|
|
|
auto GroupCall::ensureGlobalShortcutManager()
|
|
-> std::shared_ptr<GlobalShortcutManager> {
|
|
if (!_shortcutManager) {
|
|
_shortcutManager = base::CreateGlobalShortcutManager();
|
|
}
|
|
return _shortcutManager;
|
|
}
|
|
|
|
void GroupCall::applyGlobalShortcutChanges() {
|
|
auto &settings = Core::App().settings();
|
|
if (!settings.groupCallPushToTalk()
|
|
|| settings.groupCallPushToTalkShortcut().isEmpty()
|
|
|| !base::GlobalShortcutsAvailable()
|
|
|| !base::GlobalShortcutsAllowed()) {
|
|
_shortcutManager = nullptr;
|
|
_pushToTalk = nullptr;
|
|
return;
|
|
}
|
|
ensureGlobalShortcutManager();
|
|
const auto shortcut = _shortcutManager->shortcutFromSerialized(
|
|
settings.groupCallPushToTalkShortcut());
|
|
if (!shortcut) {
|
|
settings.setGroupCallPushToTalkShortcut(QByteArray());
|
|
settings.setGroupCallPushToTalk(false);
|
|
Core::App().saveSettingsDelayed();
|
|
_shortcutManager = nullptr;
|
|
_pushToTalk = nullptr;
|
|
return;
|
|
}
|
|
if (_pushToTalk) {
|
|
if (shortcut->serialize() == _pushToTalk->serialize()) {
|
|
return;
|
|
}
|
|
_shortcutManager->stopWatching(_pushToTalk);
|
|
}
|
|
_pushToTalk = shortcut;
|
|
_shortcutManager->startWatching(_pushToTalk, [=](bool pressed) {
|
|
pushToTalk(
|
|
pressed,
|
|
Core::App().settings().groupCallPushToTalkDelay());
|
|
});
|
|
}
|
|
|
|
void GroupCall::pushToTalk(bool pressed, crl::time delay) {
|
|
if (mutedByAdmin() || muted() == MuteState::Active) {
|
|
return;
|
|
} else if (pressed) {
|
|
_pushToTalkCancelTimer.cancel();
|
|
setMuted(MuteState::PushToTalk);
|
|
} else if (delay) {
|
|
_pushToTalkCancelTimer.callOnce(delay);
|
|
} else {
|
|
pushToTalkCancel();
|
|
}
|
|
}
|
|
|
|
void GroupCall::pushToTalkCancel() {
|
|
_pushToTalkCancelTimer.cancel();
|
|
if (muted() == MuteState::PushToTalk) {
|
|
setMuted(MuteState::Muted);
|
|
}
|
|
}
|
|
|
|
void GroupCall::setNotRequireARGB32() {
|
|
_requireARGB32 = false;
|
|
}
|
|
|
|
auto GroupCall::otherParticipantStateValue() const
|
|
-> rpl::producer<Group::ParticipantState> {
|
|
return _otherParticipantStateValue.events();
|
|
}
|
|
|
|
MTPInputGroupCall GroupCall::inputCall() const {
|
|
Expects(_id != 0);
|
|
|
|
return MTP_inputGroupCall(
|
|
MTP_long(_id),
|
|
MTP_long(_accessHash));
|
|
}
|
|
|
|
void GroupCall::destroyController() {
|
|
if (_instance) {
|
|
DEBUG_LOG(("Call Info: Destroying call controller.."));
|
|
invalidate_weak_ptrs(&_instanceGuard);
|
|
|
|
crl::async([
|
|
instance = base::take(_instance),
|
|
done = _delegate->groupCallAddAsyncWaiter()
|
|
]() mutable {
|
|
instance = nullptr;
|
|
DEBUG_LOG(("Call Info: Call controller destroyed."));
|
|
done();
|
|
});
|
|
}
|
|
}
|
|
|
|
void GroupCall::destroyScreencast() {
|
|
if (_screenInstance) {
|
|
DEBUG_LOG(("Call Info: Destroying call screen controller.."));
|
|
invalidate_weak_ptrs(&_screenInstanceGuard);
|
|
crl::async([
|
|
instance = base::take(_screenInstance),
|
|
done = _delegate->groupCallAddAsyncWaiter()
|
|
]() mutable {
|
|
instance = nullptr;
|
|
DEBUG_LOG(("Call Info: Call screen controller destroyed."));
|
|
done();
|
|
});
|
|
}
|
|
}
|
|
|
|
} // namespace Calls
|