mirror of https://github.com/watcha-fr/synapse
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
2237 lines
88 KiB
2237 lines
88 KiB
# Copyright 2016-2020 The Matrix.org Foundation C.I.C.
|
|
# Copyright 2020 Sorunome
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
import abc
|
|
import logging
|
|
import random
|
|
from http import HTTPStatus
|
|
from typing import TYPE_CHECKING, Iterable, List, Optional, Set, Tuple
|
|
|
|
from synapse import types
|
|
from synapse.api.constants import (
|
|
AccountDataTypes,
|
|
EventContentFields,
|
|
EventTypes,
|
|
GuestAccess,
|
|
Membership,
|
|
)
|
|
from synapse.api.errors import (
|
|
AuthError,
|
|
Codes,
|
|
PartialStateConflictError,
|
|
ShadowBanError,
|
|
SynapseError,
|
|
)
|
|
from synapse.api.ratelimiting import Ratelimiter
|
|
from synapse.event_auth import get_named_level, get_power_level_event
|
|
from synapse.events import EventBase
|
|
from synapse.events.snapshot import EventContext
|
|
from synapse.handlers.pagination import PURGE_ROOM_ACTION_NAME
|
|
from synapse.handlers.profile import MAX_AVATAR_URL_LEN, MAX_DISPLAYNAME_LEN
|
|
from synapse.handlers.state_deltas import MatchChange, StateDeltasHandler
|
|
from synapse.handlers.worker_lock import NEW_EVENT_DURING_PURGE_LOCK_NAME
|
|
from synapse.logging import opentracing
|
|
from synapse.metrics import event_processing_positions
|
|
from synapse.metrics.background_process_metrics import run_as_background_process
|
|
from synapse.storage.databases.main.state_deltas import StateDelta
|
|
from synapse.types import (
|
|
JsonDict,
|
|
Requester,
|
|
RoomAlias,
|
|
RoomID,
|
|
StateMap,
|
|
UserID,
|
|
create_requester,
|
|
get_domain_from_id,
|
|
)
|
|
from synapse.types.state import StateFilter
|
|
from synapse.util.async_helpers import Linearizer
|
|
from synapse.util.distributor import user_left_room
|
|
|
|
if TYPE_CHECKING:
|
|
from synapse.server import HomeServer
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
class NoKnownServersError(SynapseError):
|
|
"""No server already resident to the room was provided to the join/knock operation."""
|
|
|
|
def __init__(self, msg: str = "No known servers"):
|
|
super().__init__(404, msg)
|
|
|
|
|
|
class RoomMemberHandler(metaclass=abc.ABCMeta):
|
|
# TODO(paul): This handler currently contains a messy conflation of
|
|
# low-level API that works on UserID objects and so on, and REST-level
|
|
# API that takes ID strings and returns pagination chunks. These concerns
|
|
# ought to be separated out a lot better.
|
|
|
|
def __init__(self, hs: "HomeServer"):
|
|
self.hs = hs
|
|
self.store = hs.get_datastores().main
|
|
self._storage_controllers = hs.get_storage_controllers()
|
|
self.auth = hs.get_auth()
|
|
self.state_handler = hs.get_state_handler()
|
|
self.config = hs.config
|
|
self._server_name = hs.hostname
|
|
|
|
self.federation_handler = hs.get_federation_handler()
|
|
self.directory_handler = hs.get_directory_handler()
|
|
self.identity_handler = hs.get_identity_handler()
|
|
self.registration_handler = hs.get_registration_handler()
|
|
self.profile_handler = hs.get_profile_handler()
|
|
self.event_creation_handler = hs.get_event_creation_handler()
|
|
self.account_data_handler = hs.get_account_data_handler()
|
|
self.event_auth_handler = hs.get_event_auth_handler()
|
|
self._worker_lock_handler = hs.get_worker_locks_handler()
|
|
|
|
self.member_linearizer: Linearizer = Linearizer(name="member")
|
|
self.member_as_limiter = Linearizer(max_count=10, name="member_as_limiter")
|
|
|
|
self.clock = hs.get_clock()
|
|
self._spam_checker_module_callbacks = hs.get_module_api_callbacks().spam_checker
|
|
self._third_party_event_rules = (
|
|
hs.get_module_api_callbacks().third_party_event_rules
|
|
)
|
|
self._server_notices_mxid = self.config.servernotices.server_notices_mxid
|
|
self._enable_lookup = hs.config.registration.enable_3pid_lookup
|
|
self.allow_per_room_profiles = self.config.server.allow_per_room_profiles
|
|
|
|
self._join_rate_limiter_local = Ratelimiter(
|
|
store=self.store,
|
|
clock=self.clock,
|
|
cfg=hs.config.ratelimiting.rc_joins_local,
|
|
)
|
|
# Tracks joins from local users to rooms this server isn't a member of.
|
|
# I.e. joins this server makes by requesting /make_join /send_join from
|
|
# another server.
|
|
self._join_rate_limiter_remote = Ratelimiter(
|
|
store=self.store,
|
|
clock=self.clock,
|
|
cfg=hs.config.ratelimiting.rc_joins_remote,
|
|
)
|
|
# TODO: find a better place to keep this Ratelimiter.
|
|
# It needs to be
|
|
# - written to by event persistence code
|
|
# - written to by something which can snoop on replication streams
|
|
# - read by the RoomMemberHandler to rate limit joins from local users
|
|
# - read by the FederationServer to rate limit make_joins and send_joins from
|
|
# other homeservers
|
|
# I wonder if a homeserver-wide collection of rate limiters might be cleaner?
|
|
self._join_rate_per_room_limiter = Ratelimiter(
|
|
store=self.store,
|
|
clock=self.clock,
|
|
cfg=hs.config.ratelimiting.rc_joins_per_room,
|
|
)
|
|
|
|
# Ratelimiter for invites, keyed by room (across all issuers, all
|
|
# recipients).
|
|
self._invites_per_room_limiter = Ratelimiter(
|
|
store=self.store,
|
|
clock=self.clock,
|
|
cfg=hs.config.ratelimiting.rc_invites_per_room,
|
|
)
|
|
|
|
# Ratelimiter for invites, keyed by recipient (across all rooms, all
|
|
# issuers).
|
|
self._invites_per_recipient_limiter = Ratelimiter(
|
|
store=self.store,
|
|
clock=self.clock,
|
|
cfg=hs.config.ratelimiting.rc_invites_per_user,
|
|
)
|
|
|
|
# Ratelimiter for invites, keyed by issuer (across all rooms, all
|
|
# recipients).
|
|
self._invites_per_issuer_limiter = Ratelimiter(
|
|
store=self.store,
|
|
clock=self.clock,
|
|
cfg=hs.config.ratelimiting.rc_invites_per_issuer,
|
|
)
|
|
|
|
self._third_party_invite_limiter = Ratelimiter(
|
|
store=self.store,
|
|
clock=self.clock,
|
|
cfg=hs.config.ratelimiting.rc_third_party_invite,
|
|
)
|
|
|
|
self.request_ratelimiter = hs.get_request_ratelimiter()
|
|
hs.get_notifier().add_new_join_in_room_callback(self._on_user_joined_room)
|
|
|
|
self._forgotten_room_retention_period = (
|
|
hs.config.server.forgotten_room_retention_period
|
|
)
|
|
|
|
def _on_user_joined_room(self, event_id: str, room_id: str) -> None:
|
|
"""Notify the rate limiter that a room join has occurred.
|
|
|
|
Use this to inform the RoomMemberHandler about joins that have either
|
|
- taken place on another homeserver, or
|
|
- on another worker in this homeserver.
|
|
Joins actioned by this worker should use the usual `ratelimit` method, which
|
|
checks the limit and increments the counter in one go.
|
|
"""
|
|
self._join_rate_per_room_limiter.record_action(requester=None, key=room_id)
|
|
|
|
@abc.abstractmethod
|
|
async def _remote_join(
|
|
self,
|
|
requester: Requester,
|
|
remote_room_hosts: List[str],
|
|
room_id: str,
|
|
user: UserID,
|
|
content: dict,
|
|
) -> Tuple[str, int]:
|
|
"""Try and join a room that this server is not in
|
|
|
|
Args:
|
|
requester: The user making the request, according to the access token.
|
|
remote_room_hosts: List of servers that can be used to join via.
|
|
room_id: Room that we are trying to join
|
|
user: User who is trying to join
|
|
content: A dict that should be used as the content of the join event.
|
|
|
|
Raises:
|
|
NoKnownServersError: if remote_room_hosts does not contain a server joined to
|
|
the room.
|
|
"""
|
|
raise NotImplementedError()
|
|
|
|
@abc.abstractmethod
|
|
async def remote_knock(
|
|
self,
|
|
requester: Requester,
|
|
remote_room_hosts: List[str],
|
|
room_id: str,
|
|
user: UserID,
|
|
content: dict,
|
|
) -> Tuple[str, int]:
|
|
"""Try and knock on a room that this server is not in
|
|
|
|
Args:
|
|
remote_room_hosts: List of servers that can be used to knock via.
|
|
room_id: Room that we are trying to knock on.
|
|
user: User who is trying to knock.
|
|
content: A dict that should be used as the content of the knock event.
|
|
"""
|
|
raise NotImplementedError()
|
|
|
|
@abc.abstractmethod
|
|
async def remote_reject_invite(
|
|
self,
|
|
invite_event_id: str,
|
|
txn_id: Optional[str],
|
|
requester: Requester,
|
|
content: JsonDict,
|
|
) -> Tuple[str, int]:
|
|
"""
|
|
Rejects an out-of-band invite we have received from a remote server
|
|
|
|
Args:
|
|
invite_event_id: ID of the invite to be rejected
|
|
txn_id: optional transaction ID supplied by the client
|
|
requester: user making the rejection request, according to the access token
|
|
content: additional content to include in the rejection event.
|
|
Normally an empty dict.
|
|
|
|
Returns:
|
|
event id, stream_id of the leave event
|
|
"""
|
|
raise NotImplementedError()
|
|
|
|
@abc.abstractmethod
|
|
async def remote_rescind_knock(
|
|
self,
|
|
knock_event_id: str,
|
|
txn_id: Optional[str],
|
|
requester: Requester,
|
|
content: JsonDict,
|
|
) -> Tuple[str, int]:
|
|
"""Rescind a local knock made on a remote room.
|
|
|
|
Args:
|
|
knock_event_id: The ID of the knock event to rescind.
|
|
txn_id: An optional transaction ID supplied by the client.
|
|
requester: The user making the request, according to the access token.
|
|
content: The content of the generated leave event.
|
|
|
|
Returns:
|
|
A tuple containing (event_id, stream_id of the leave event).
|
|
"""
|
|
raise NotImplementedError()
|
|
|
|
@abc.abstractmethod
|
|
async def _user_left_room(self, target: UserID, room_id: str) -> None:
|
|
"""Notifies distributor on master process that the user has left the
|
|
room.
|
|
|
|
Args:
|
|
target
|
|
room_id
|
|
"""
|
|
raise NotImplementedError()
|
|
|
|
async def forget(
|
|
self, user: UserID, room_id: str, do_not_schedule_purge: bool = False
|
|
) -> None:
|
|
user_id = user.to_string()
|
|
|
|
member = await self._storage_controllers.state.get_current_state_event(
|
|
room_id=room_id, event_type=EventTypes.Member, state_key=user_id
|
|
)
|
|
membership = member.membership if member else None
|
|
|
|
if membership is not None and membership not in [
|
|
Membership.LEAVE,
|
|
Membership.BAN,
|
|
]:
|
|
raise SynapseError(400, "User %s in room %s" % (user_id, room_id))
|
|
|
|
# In normal case this call is only required if `membership` is not `None`.
|
|
# But: After the last member had left the room, the background update
|
|
# `_background_remove_left_rooms` is deleting rows related to this room from
|
|
# the table `current_state_events` and `get_current_state_events` is `None`.
|
|
await self.store.forget(user_id, room_id)
|
|
|
|
# If everyone locally has left the room, then there is no reason for us to keep the
|
|
# room around and we automatically purge room after a little bit
|
|
if (
|
|
not do_not_schedule_purge
|
|
and self._forgotten_room_retention_period
|
|
and await self.store.is_locally_forgotten_room(room_id)
|
|
):
|
|
await self.hs.get_task_scheduler().schedule_task(
|
|
PURGE_ROOM_ACTION_NAME,
|
|
resource_id=room_id,
|
|
timestamp=self.clock.time_msec()
|
|
+ self._forgotten_room_retention_period,
|
|
)
|
|
|
|
async def ratelimit_multiple_invites(
|
|
self,
|
|
requester: Optional[Requester],
|
|
room_id: Optional[str],
|
|
n_invites: int,
|
|
update: bool = True,
|
|
) -> None:
|
|
"""Ratelimit more than one invite sent by the given requester in the given room.
|
|
|
|
Args:
|
|
requester: The requester sending the invites.
|
|
room_id: The room the invites are being sent in.
|
|
n_invites: The amount of invites to ratelimit for.
|
|
update: Whether to update the ratelimiter's cache.
|
|
|
|
Raises:
|
|
LimitExceededError: The requester can't send that many invites in the room.
|
|
"""
|
|
await self._invites_per_room_limiter.ratelimit(
|
|
requester,
|
|
room_id,
|
|
update=update,
|
|
n_actions=n_invites,
|
|
)
|
|
|
|
async def ratelimit_invite(
|
|
self,
|
|
requester: Optional[Requester],
|
|
room_id: Optional[str],
|
|
invitee_user_id: str,
|
|
) -> None:
|
|
"""Ratelimit invites by room and by target user.
|
|
|
|
If room ID is missing then we just rate limit by target user.
|
|
"""
|
|
if room_id:
|
|
await self._invites_per_room_limiter.ratelimit(requester, room_id)
|
|
|
|
await self._invites_per_recipient_limiter.ratelimit(requester, invitee_user_id)
|
|
if requester is not None:
|
|
await self._invites_per_issuer_limiter.ratelimit(requester)
|
|
|
|
async def _local_membership_update(
|
|
self,
|
|
requester: Requester,
|
|
target: UserID,
|
|
room_id: str,
|
|
membership: str,
|
|
allow_no_prev_events: bool = False,
|
|
prev_event_ids: Optional[List[str]] = None,
|
|
state_event_ids: Optional[List[str]] = None,
|
|
depth: Optional[int] = None,
|
|
txn_id: Optional[str] = None,
|
|
ratelimit: bool = True,
|
|
content: Optional[dict] = None,
|
|
require_consent: bool = True,
|
|
outlier: bool = False,
|
|
origin_server_ts: Optional[int] = None,
|
|
) -> Tuple[str, int]:
|
|
"""
|
|
Internal membership update function to get an existing event or create
|
|
and persist a new event for the new membership change.
|
|
|
|
Args:
|
|
requester: User requesting the membership change, i.e. the sender of the
|
|
desired membership event.
|
|
target: Use whose membership should change, i.e. the state_key of the
|
|
desired membership event.
|
|
room_id:
|
|
membership:
|
|
|
|
allow_no_prev_events: Whether to allow this event to be created an empty
|
|
list of prev_events. Normally this is prohibited just because most
|
|
events should have a prev_event and we should only use this in special
|
|
cases (previously useful for MSC2716).
|
|
prev_event_ids: The event IDs to use as the prev events
|
|
state_event_ids:
|
|
The full state at a given event. This was previously used particularly
|
|
by the MSC2716 /batch_send endpoint. This should normally be left as
|
|
None, which will cause the auth_event_ids to be calculated based on the
|
|
room state at the prev_events.
|
|
depth: Override the depth used to order the event in the DAG.
|
|
Should normally be set to None, which will cause the depth to be calculated
|
|
based on the prev_events.
|
|
|
|
txn_id:
|
|
ratelimit:
|
|
content:
|
|
require_consent:
|
|
|
|
outlier: Indicates whether the event is an `outlier`, i.e. if
|
|
it's from an arbitrary point and floating in the DAG as
|
|
opposed to being inline with the current DAG.
|
|
origin_server_ts: The origin_server_ts to use if a new event is created. Uses
|
|
the current timestamp if set to None.
|
|
|
|
Returns:
|
|
Tuple of event ID and stream ordering position
|
|
"""
|
|
user_id = target.to_string()
|
|
|
|
if content is None:
|
|
content = {}
|
|
|
|
content["membership"] = membership
|
|
if requester.is_guest:
|
|
content["kind"] = "guest"
|
|
|
|
# Check if we already have an event with a matching transaction ID. (We
|
|
# do this check just before we persist an event as well, but may as well
|
|
# do it up front for efficiency.)
|
|
if txn_id:
|
|
existing_event_id = (
|
|
await self.event_creation_handler.get_event_id_from_transaction(
|
|
requester, txn_id, room_id
|
|
)
|
|
)
|
|
if existing_event_id:
|
|
event_pos = await self.store.get_position_for_event(existing_event_id)
|
|
return existing_event_id, event_pos.stream
|
|
|
|
# Try several times, it could fail with PartialStateConflictError,
|
|
# in handle_new_client_event, cf comment in except block.
|
|
max_retries = 5
|
|
for i in range(max_retries):
|
|
try:
|
|
(
|
|
event,
|
|
unpersisted_context,
|
|
) = await self.event_creation_handler.create_event(
|
|
requester,
|
|
{
|
|
"type": EventTypes.Member,
|
|
"content": content,
|
|
"room_id": room_id,
|
|
"sender": requester.user.to_string(),
|
|
"state_key": user_id,
|
|
# For backwards compatibility:
|
|
"membership": membership,
|
|
"origin_server_ts": origin_server_ts,
|
|
},
|
|
txn_id=txn_id,
|
|
allow_no_prev_events=allow_no_prev_events,
|
|
prev_event_ids=prev_event_ids,
|
|
state_event_ids=state_event_ids,
|
|
depth=depth,
|
|
require_consent=require_consent,
|
|
outlier=outlier,
|
|
)
|
|
context = await unpersisted_context.persist(event)
|
|
prev_state_ids = await context.get_prev_state_ids(
|
|
StateFilter.from_types([(EventTypes.Member, user_id)])
|
|
)
|
|
|
|
prev_member_event_id = prev_state_ids.get(
|
|
(EventTypes.Member, user_id), None
|
|
)
|
|
|
|
with opentracing.start_active_span("handle_new_client_event"):
|
|
result_event = (
|
|
await self.event_creation_handler.handle_new_client_event(
|
|
requester,
|
|
events_and_context=[(event, context)],
|
|
extra_users=[target],
|
|
ratelimit=ratelimit,
|
|
)
|
|
)
|
|
|
|
if event.membership == Membership.LEAVE:
|
|
if prev_member_event_id:
|
|
prev_member_event = await self.store.get_event(
|
|
prev_member_event_id
|
|
)
|
|
if prev_member_event.membership == Membership.JOIN:
|
|
await self._user_left_room(target, room_id)
|
|
|
|
break
|
|
except PartialStateConflictError as e:
|
|
# Persisting couldn't happen because the room got un-partial stated
|
|
# in the meantime and context needs to be recomputed, so let's do so.
|
|
if i == max_retries - 1:
|
|
raise e
|
|
|
|
# we know it was persisted, so should have a stream ordering
|
|
assert result_event.internal_metadata.stream_ordering
|
|
return result_event.event_id, result_event.internal_metadata.stream_ordering
|
|
|
|
async def copy_room_tags_and_direct_to_room(
|
|
self, old_room_id: str, new_room_id: str, user_id: str
|
|
) -> None:
|
|
"""Copies the tags and direct room state from one room to another.
|
|
|
|
Args:
|
|
old_room_id: The room ID of the old room.
|
|
new_room_id: The room ID of the new room.
|
|
user_id: The user's ID.
|
|
"""
|
|
# Retrieve user account data for predecessor room
|
|
user_account_data = await self.store.get_global_account_data_for_user(user_id)
|
|
|
|
# Copy direct message state if applicable
|
|
direct_rooms = user_account_data.get(AccountDataTypes.DIRECT, {})
|
|
|
|
# Check which key this room is under
|
|
if isinstance(direct_rooms, dict):
|
|
for key, room_id_list in direct_rooms.items():
|
|
if old_room_id in room_id_list and new_room_id not in room_id_list:
|
|
# Add new room_id to this key
|
|
direct_rooms[key].append(new_room_id)
|
|
|
|
# Save back to user's m.direct account data
|
|
await self.account_data_handler.add_account_data_for_user(
|
|
user_id, AccountDataTypes.DIRECT, direct_rooms
|
|
)
|
|
break
|
|
|
|
# Copy room tags if applicable
|
|
room_tags = await self.store.get_tags_for_room(user_id, old_room_id)
|
|
|
|
# Copy each room tag to the new room
|
|
for tag, tag_content in room_tags.items():
|
|
await self.account_data_handler.add_tag_to_room(
|
|
user_id, new_room_id, tag, tag_content
|
|
)
|
|
|
|
async def update_membership(
|
|
self,
|
|
requester: Requester,
|
|
target: UserID,
|
|
room_id: str,
|
|
action: str,
|
|
txn_id: Optional[str] = None,
|
|
remote_room_hosts: Optional[List[str]] = None,
|
|
third_party_signed: Optional[dict] = None,
|
|
ratelimit: bool = True,
|
|
content: Optional[dict] = None,
|
|
new_room: bool = False,
|
|
require_consent: bool = True,
|
|
outlier: bool = False,
|
|
allow_no_prev_events: bool = False,
|
|
prev_event_ids: Optional[List[str]] = None,
|
|
state_event_ids: Optional[List[str]] = None,
|
|
depth: Optional[int] = None,
|
|
origin_server_ts: Optional[int] = None,
|
|
) -> Tuple[str, int]:
|
|
"""Update a user's membership in a room.
|
|
|
|
Params:
|
|
requester: The user who is performing the update.
|
|
target: The user whose membership is being updated.
|
|
room_id: The room ID whose membership is being updated.
|
|
action: The membership change, see synapse.api.constants.Membership.
|
|
txn_id: The transaction ID, if given.
|
|
remote_room_hosts: Remote servers to send the update to.
|
|
third_party_signed: Information from a 3PID invite.
|
|
ratelimit: Whether to rate limit the request.
|
|
content: The content of the created event.
|
|
new_room: Whether the membership update is happening in the context of a room
|
|
creation.
|
|
require_consent: Whether consent is required.
|
|
outlier: Indicates whether the event is an `outlier`, i.e. if
|
|
it's from an arbitrary point and floating in the DAG as
|
|
opposed to being inline with the current DAG.
|
|
allow_no_prev_events: Whether to allow this event to be created an empty
|
|
list of prev_events. Normally this is prohibited just because most
|
|
events should have a prev_event and we should only use this in special
|
|
cases (previously useful for MSC2716).
|
|
prev_event_ids: The event IDs to use as the prev events
|
|
state_event_ids:
|
|
The full state at a given event. This was previously used particularly
|
|
by the MSC2716 /batch_send endpoint. This should normally be left as
|
|
None, which will cause the auth_event_ids to be calculated based on the
|
|
room state at the prev_events.
|
|
depth: Override the depth used to order the event in the DAG.
|
|
Should normally be set to None, which will cause the depth to be calculated
|
|
based on the prev_events.
|
|
origin_server_ts: The origin_server_ts to use if a new event is created. Uses
|
|
the current timestamp if set to None.
|
|
|
|
Returns:
|
|
A tuple of the new event ID and stream ID.
|
|
|
|
Raises:
|
|
ShadowBanError if a shadow-banned requester attempts to send an invite.
|
|
"""
|
|
if ratelimit:
|
|
if action == Membership.JOIN:
|
|
# Only rate-limit if the user isn't already joined to the room, otherwise
|
|
# we'll end up blocking profile updates.
|
|
(
|
|
current_membership,
|
|
_,
|
|
) = await self.store.get_local_current_membership_for_user_in_room(
|
|
requester.user.to_string(),
|
|
room_id,
|
|
)
|
|
if current_membership != Membership.JOIN:
|
|
await self._join_rate_limiter_local.ratelimit(requester)
|
|
await self._join_rate_per_room_limiter.ratelimit(
|
|
requester, key=room_id, update=False
|
|
)
|
|
elif action == Membership.INVITE:
|
|
await self.ratelimit_invite(requester, room_id, target.to_string())
|
|
|
|
if action == Membership.INVITE and requester.shadow_banned:
|
|
# We randomly sleep a bit just to annoy the requester.
|
|
await self.clock.sleep(random.randint(1, 10))
|
|
raise ShadowBanError()
|
|
|
|
key = (room_id,)
|
|
|
|
as_id = object()
|
|
if requester.app_service:
|
|
as_id = requester.app_service.id
|
|
|
|
# We first linearise by the application service (to try to limit concurrent joins
|
|
# by application services), and then by room ID.
|
|
async with self.member_as_limiter.queue(as_id):
|
|
async with self.member_linearizer.queue(key):
|
|
async with self._worker_lock_handler.acquire_read_write_lock(
|
|
NEW_EVENT_DURING_PURGE_LOCK_NAME, room_id, write=False
|
|
):
|
|
with opentracing.start_active_span("update_membership_locked"):
|
|
result = await self.update_membership_locked(
|
|
requester,
|
|
target,
|
|
room_id,
|
|
action,
|
|
txn_id=txn_id,
|
|
remote_room_hosts=remote_room_hosts,
|
|
third_party_signed=third_party_signed,
|
|
ratelimit=ratelimit,
|
|
content=content,
|
|
new_room=new_room,
|
|
require_consent=require_consent,
|
|
outlier=outlier,
|
|
allow_no_prev_events=allow_no_prev_events,
|
|
prev_event_ids=prev_event_ids,
|
|
state_event_ids=state_event_ids,
|
|
depth=depth,
|
|
origin_server_ts=origin_server_ts,
|
|
)
|
|
|
|
return result
|
|
|
|
async def update_membership_locked(
|
|
self,
|
|
requester: Requester,
|
|
target: UserID,
|
|
room_id: str,
|
|
action: str,
|
|
txn_id: Optional[str] = None,
|
|
remote_room_hosts: Optional[List[str]] = None,
|
|
third_party_signed: Optional[dict] = None,
|
|
ratelimit: bool = True,
|
|
content: Optional[dict] = None,
|
|
new_room: bool = False,
|
|
require_consent: bool = True,
|
|
outlier: bool = False,
|
|
allow_no_prev_events: bool = False,
|
|
prev_event_ids: Optional[List[str]] = None,
|
|
state_event_ids: Optional[List[str]] = None,
|
|
depth: Optional[int] = None,
|
|
origin_server_ts: Optional[int] = None,
|
|
) -> Tuple[str, int]:
|
|
"""Helper for update_membership.
|
|
|
|
Assumes that the membership linearizer is already held for the room.
|
|
|
|
Args:
|
|
requester:
|
|
target:
|
|
room_id:
|
|
action:
|
|
txn_id:
|
|
remote_room_hosts:
|
|
third_party_signed:
|
|
ratelimit:
|
|
content:
|
|
new_room: Whether the membership update is happening in the context of a room
|
|
creation.
|
|
require_consent:
|
|
outlier: Indicates whether the event is an `outlier`, i.e. if
|
|
it's from an arbitrary point and floating in the DAG as
|
|
opposed to being inline with the current DAG.
|
|
allow_no_prev_events: Whether to allow this event to be created an empty
|
|
list of prev_events. Normally this is prohibited just because most
|
|
events should have a prev_event and we should only use this in special
|
|
cases (previously useful for MSC2716).
|
|
prev_event_ids: The event IDs to use as the prev events
|
|
state_event_ids:
|
|
The full state at a given event. This was previously used particularly
|
|
by the MSC2716 /batch_send endpoint. This should normally be left as
|
|
None, which will cause the auth_event_ids to be calculated based on the
|
|
room state at the prev_events.
|
|
depth: Override the depth used to order the event in the DAG.
|
|
Should normally be set to None, which will cause the depth to be calculated
|
|
based on the prev_events.
|
|
origin_server_ts: The origin_server_ts to use if a new event is created. Uses
|
|
the current timestamp if set to None.
|
|
|
|
Returns:
|
|
A tuple of the new event ID and stream ID.
|
|
"""
|
|
|
|
content_specified = bool(content)
|
|
if content is None:
|
|
content = {}
|
|
else:
|
|
# We do a copy here as we potentially change some keys
|
|
# later on.
|
|
content = dict(content)
|
|
|
|
# allow the server notices mxid to set room-level profile
|
|
is_requester_server_notices_user = (
|
|
self._server_notices_mxid is not None
|
|
and requester.user.to_string() == self._server_notices_mxid
|
|
)
|
|
|
|
if (
|
|
not self.allow_per_room_profiles and not is_requester_server_notices_user
|
|
) or requester.shadow_banned:
|
|
# Strip profile data, knowing that new profile data will be added to the
|
|
# event's content in event_creation_handler.create_event() using the target's
|
|
# global profile.
|
|
content.pop("displayname", None)
|
|
content.pop("avatar_url", None)
|
|
|
|
if len(content.get("displayname") or "") > MAX_DISPLAYNAME_LEN:
|
|
raise SynapseError(
|
|
400,
|
|
f"Displayname is too long (max {MAX_DISPLAYNAME_LEN})",
|
|
errcode=Codes.BAD_JSON,
|
|
)
|
|
|
|
if len(content.get("avatar_url") or "") > MAX_AVATAR_URL_LEN:
|
|
raise SynapseError(
|
|
400,
|
|
f"Avatar URL is too long (max {MAX_AVATAR_URL_LEN})",
|
|
errcode=Codes.BAD_JSON,
|
|
)
|
|
|
|
if "avatar_url" in content and content.get("avatar_url") is not None:
|
|
if not await self.profile_handler.check_avatar_size_and_mime_type(
|
|
content["avatar_url"],
|
|
):
|
|
raise SynapseError(403, "This avatar is not allowed", Codes.FORBIDDEN)
|
|
|
|
# The event content should *not* include the authorising user as
|
|
# it won't be properly signed. Strip it out since it might come
|
|
# back from a client updating a display name / avatar.
|
|
#
|
|
# This only applies to restricted rooms, but there should be no reason
|
|
# for a client to include it. Unconditionally remove it.
|
|
content.pop(EventContentFields.AUTHORISING_USER, None)
|
|
|
|
effective_membership_state = action
|
|
if action in ["kick", "unban"]:
|
|
effective_membership_state = "leave"
|
|
|
|
# if this is a join with a 3pid signature, we may need to turn a 3pid
|
|
# invite into a normal invite before we can handle the join.
|
|
if third_party_signed is not None:
|
|
await self.federation_handler.exchange_third_party_invite(
|
|
third_party_signed["sender"],
|
|
target.to_string(),
|
|
room_id,
|
|
third_party_signed,
|
|
)
|
|
|
|
if not remote_room_hosts:
|
|
remote_room_hosts = []
|
|
|
|
if effective_membership_state not in ("leave", "ban"):
|
|
is_blocked = await self.store.is_room_blocked(room_id)
|
|
if is_blocked:
|
|
raise SynapseError(403, "This room has been blocked on this server")
|
|
|
|
if effective_membership_state == Membership.INVITE:
|
|
target_id = target.to_string()
|
|
|
|
# block any attempts to invite the server notices mxid
|
|
if target_id == self._server_notices_mxid:
|
|
raise SynapseError(HTTPStatus.FORBIDDEN, "Cannot invite this user")
|
|
|
|
block_invite_result = None
|
|
|
|
if (
|
|
self._server_notices_mxid is not None
|
|
and requester.user.to_string() == self._server_notices_mxid
|
|
):
|
|
# allow the server notices mxid to send invites
|
|
is_requester_admin = True
|
|
|
|
else:
|
|
is_requester_admin = await self.auth.is_server_admin(requester)
|
|
|
|
if not is_requester_admin:
|
|
if self.config.server.block_non_admin_invites:
|
|
logger.info(
|
|
"Blocking invite: user is not admin and non-admin "
|
|
"invites disabled"
|
|
)
|
|
block_invite_result = (Codes.FORBIDDEN, {})
|
|
|
|
spam_check = await self._spam_checker_module_callbacks.user_may_invite(
|
|
requester.user.to_string(), target_id, room_id
|
|
)
|
|
if spam_check != self._spam_checker_module_callbacks.NOT_SPAM:
|
|
logger.info("Blocking invite due to spam checker")
|
|
block_invite_result = spam_check
|
|
|
|
if block_invite_result is not None:
|
|
raise SynapseError(
|
|
403,
|
|
"Invites have been disabled on this server",
|
|
errcode=block_invite_result[0],
|
|
additional_fields=block_invite_result[1],
|
|
)
|
|
|
|
# An empty prev_events list is allowed as long as the auth_event_ids are present
|
|
if prev_event_ids is not None:
|
|
return await self._local_membership_update(
|
|
requester=requester,
|
|
target=target,
|
|
room_id=room_id,
|
|
membership=effective_membership_state,
|
|
txn_id=txn_id,
|
|
ratelimit=ratelimit,
|
|
allow_no_prev_events=allow_no_prev_events,
|
|
prev_event_ids=prev_event_ids,
|
|
state_event_ids=state_event_ids,
|
|
depth=depth,
|
|
content=content,
|
|
require_consent=require_consent,
|
|
outlier=outlier,
|
|
origin_server_ts=origin_server_ts,
|
|
)
|
|
|
|
latest_event_ids = await self.store.get_prev_events_for_room(room_id)
|
|
|
|
is_partial_state_room = await self.store.is_partial_state_room(room_id)
|
|
partial_state_before_join = await self.state_handler.compute_state_after_events(
|
|
room_id, latest_event_ids, await_full_state=False
|
|
)
|
|
# `is_partial_state_room` also indicates whether `partial_state_before_join` is
|
|
# partial.
|
|
|
|
is_host_in_room = await self._is_host_in_room(partial_state_before_join)
|
|
|
|
# if we are not in the room, we won't have the current state
|
|
if is_host_in_room:
|
|
# TODO: Refactor into dictionary of explicitly allowed transitions
|
|
# between old and new state, with specific error messages for some
|
|
# transitions and generic otherwise
|
|
old_state_id = partial_state_before_join.get(
|
|
(EventTypes.Member, target.to_string())
|
|
)
|
|
|
|
if old_state_id:
|
|
old_state = await self.store.get_event(old_state_id, allow_none=True)
|
|
old_membership = (
|
|
old_state.content.get("membership") if old_state else None
|
|
)
|
|
if action == "unban" and old_membership != "ban":
|
|
raise SynapseError(
|
|
403,
|
|
"Cannot unban user who was not banned"
|
|
" (membership=%s)" % old_membership,
|
|
errcode=Codes.BAD_STATE,
|
|
)
|
|
if old_membership == "ban" and action not in ["ban", "unban", "leave"]:
|
|
raise SynapseError(
|
|
403,
|
|
"Cannot %s user who was banned" % (action,),
|
|
errcode=Codes.BAD_STATE,
|
|
)
|
|
|
|
if old_state:
|
|
same_content = content == old_state.content
|
|
same_membership = old_membership == effective_membership_state
|
|
same_sender = requester.user.to_string() == old_state.sender
|
|
if same_sender and same_membership and same_content:
|
|
# duplicate event.
|
|
# we know it was persisted, so must have a stream ordering.
|
|
assert old_state.internal_metadata.stream_ordering
|
|
return (
|
|
old_state.event_id,
|
|
old_state.internal_metadata.stream_ordering,
|
|
)
|
|
|
|
if old_membership in ["ban", "leave"] and action == "kick":
|
|
raise AuthError(403, "The target user is not in the room")
|
|
|
|
# we don't allow people to reject invites to the server notice
|
|
# room, but they can leave it once they are joined.
|
|
if (
|
|
old_membership == Membership.INVITE
|
|
and effective_membership_state == Membership.LEAVE
|
|
):
|
|
is_blocked = await self.store.is_server_notice_room(room_id)
|
|
if is_blocked:
|
|
raise SynapseError(
|
|
HTTPStatus.FORBIDDEN,
|
|
"You cannot reject this invite",
|
|
errcode=Codes.CANNOT_LEAVE_SERVER_NOTICE_ROOM,
|
|
)
|
|
else:
|
|
if action == "kick":
|
|
raise AuthError(403, "The target user is not in the room")
|
|
|
|
if effective_membership_state == Membership.JOIN:
|
|
if requester.is_guest:
|
|
guest_can_join = await self._can_guest_join(partial_state_before_join)
|
|
if not guest_can_join:
|
|
# This should be an auth check, but guests are a local concept,
|
|
# so don't really fit into the general auth process.
|
|
raise AuthError(403, "Guest access not allowed")
|
|
|
|
# Figure out whether the user is a server admin to determine whether they
|
|
# should be able to bypass the spam checker.
|
|
if (
|
|
self._server_notices_mxid is not None
|
|
and requester.user.to_string() == self._server_notices_mxid
|
|
):
|
|
# allow the server notices mxid to join rooms
|
|
bypass_spam_checker = True
|
|
|
|
else:
|
|
bypass_spam_checker = await self.auth.is_server_admin(requester)
|
|
|
|
inviter = await self._get_inviter(target.to_string(), room_id)
|
|
if (
|
|
not bypass_spam_checker
|
|
# We assume that if the spam checker allowed the user to create
|
|
# a room then they're allowed to join it.
|
|
and not new_room
|
|
):
|
|
spam_check = (
|
|
await self._spam_checker_module_callbacks.user_may_join_room(
|
|
target.to_string(), room_id, is_invited=inviter is not None
|
|
)
|
|
)
|
|
if spam_check != self._spam_checker_module_callbacks.NOT_SPAM:
|
|
raise SynapseError(
|
|
403,
|
|
"Not allowed to join this room",
|
|
errcode=spam_check[0],
|
|
additional_fields=spam_check[1],
|
|
)
|
|
|
|
# Check if a remote join should be performed.
|
|
remote_join, remote_room_hosts = await self._should_perform_remote_join(
|
|
target.to_string(),
|
|
room_id,
|
|
remote_room_hosts,
|
|
content,
|
|
is_partial_state_room,
|
|
is_host_in_room,
|
|
partial_state_before_join,
|
|
)
|
|
if remote_join:
|
|
if ratelimit:
|
|
await self._join_rate_limiter_remote.ratelimit(
|
|
requester,
|
|
)
|
|
await self._join_rate_per_room_limiter.ratelimit(
|
|
requester,
|
|
key=room_id,
|
|
update=False,
|
|
)
|
|
|
|
inviter = await self._get_inviter(target.to_string(), room_id)
|
|
if inviter and not self.hs.is_mine(inviter):
|
|
remote_room_hosts.append(inviter.domain)
|
|
|
|
content["membership"] = Membership.JOIN
|
|
|
|
try:
|
|
profile = self.profile_handler
|
|
if not content_specified:
|
|
content["displayname"] = await profile.get_displayname(target)
|
|
content["avatar_url"] = await profile.get_avatar_url(target)
|
|
except Exception as e:
|
|
logger.info(
|
|
"Failed to get profile information while processing remote join for %r: %s",
|
|
target,
|
|
e,
|
|
)
|
|
|
|
if requester.is_guest:
|
|
content["kind"] = "guest"
|
|
|
|
remote_join_response = await self._remote_join(
|
|
requester, remote_room_hosts, room_id, target, content
|
|
)
|
|
|
|
return remote_join_response
|
|
|
|
elif effective_membership_state == Membership.LEAVE:
|
|
if not is_host_in_room:
|
|
# Figure out the user's current membership state for the room
|
|
(
|
|
current_membership_type,
|
|
current_membership_event_id,
|
|
) = await self.store.get_local_current_membership_for_user_in_room(
|
|
target.to_string(), room_id
|
|
)
|
|
if not current_membership_type or not current_membership_event_id:
|
|
logger.info(
|
|
"%s sent a leave request to %s, but that is not an active room "
|
|
"on this server, or there is no pending invite or knock",
|
|
target,
|
|
room_id,
|
|
)
|
|
|
|
raise SynapseError(404, "Not a known room")
|
|
|
|
# perhaps we've been invited
|
|
if current_membership_type == Membership.INVITE:
|
|
invite = await self.store.get_event(current_membership_event_id)
|
|
logger.info(
|
|
"%s rejects invite to %s from %s",
|
|
target,
|
|
room_id,
|
|
invite.sender,
|
|
)
|
|
|
|
if not self.hs.is_mine_id(invite.sender):
|
|
# send the rejection to the inviter's HS (with fallback to
|
|
# local event)
|
|
return await self.remote_reject_invite(
|
|
invite.event_id,
|
|
txn_id,
|
|
requester,
|
|
content,
|
|
)
|
|
|
|
# the inviter was on our server, but has now left. Carry on
|
|
# with the normal rejection codepath, which will also send the
|
|
# rejection out to any other servers we believe are still in the room.
|
|
|
|
# thanks to overzealous cleaning up of event_forward_extremities in
|
|
# `delete_old_current_state_events`, it's possible to end up with no
|
|
# forward extremities here. If that happens, let's just hang the
|
|
# rejection off the invite event.
|
|
#
|
|
# see: https://github.com/matrix-org/synapse/issues/7139
|
|
if len(latest_event_ids) == 0:
|
|
latest_event_ids = [invite.event_id]
|
|
|
|
# or perhaps this is a remote room that a local user has knocked on
|
|
elif current_membership_type == Membership.KNOCK:
|
|
knock = await self.store.get_event(current_membership_event_id)
|
|
return await self.remote_rescind_knock(
|
|
knock.event_id, txn_id, requester, content
|
|
)
|
|
|
|
elif effective_membership_state == Membership.KNOCK:
|
|
if not is_host_in_room:
|
|
# The knock needs to be sent over federation instead
|
|
remote_room_hosts.append(get_domain_from_id(room_id))
|
|
|
|
content["membership"] = Membership.KNOCK
|
|
|
|
try:
|
|
profile = self.profile_handler
|
|
if "displayname" not in content:
|
|
content["displayname"] = await profile.get_displayname(target)
|
|
if "avatar_url" not in content:
|
|
content["avatar_url"] = await profile.get_avatar_url(target)
|
|
except Exception as e:
|
|
logger.info(
|
|
"Failed to get profile information while processing remote knock for %r: %s",
|
|
target,
|
|
e,
|
|
)
|
|
|
|
return await self.remote_knock(
|
|
requester, remote_room_hosts, room_id, target, content
|
|
)
|
|
|
|
return await self._local_membership_update(
|
|
requester=requester,
|
|
target=target,
|
|
room_id=room_id,
|
|
membership=effective_membership_state,
|
|
txn_id=txn_id,
|
|
ratelimit=ratelimit,
|
|
prev_event_ids=latest_event_ids,
|
|
state_event_ids=state_event_ids,
|
|
depth=depth,
|
|
content=content,
|
|
require_consent=require_consent,
|
|
outlier=outlier,
|
|
origin_server_ts=origin_server_ts,
|
|
)
|
|
|
|
async def _should_perform_remote_join(
|
|
self,
|
|
user_id: str,
|
|
room_id: str,
|
|
remote_room_hosts: List[str],
|
|
content: JsonDict,
|
|
is_partial_state_room: bool,
|
|
is_host_in_room: bool,
|
|
partial_state_before_join: StateMap[str],
|
|
) -> Tuple[bool, List[str]]:
|
|
"""
|
|
Check whether the server should do a remote join (as opposed to a local
|
|
join) for a user.
|
|
|
|
Generally a remote join is used if:
|
|
|
|
* The server is not yet in the room.
|
|
* The server is in the room, the room has restricted join rules, the user
|
|
is not joined or invited to the room, and the server does not have
|
|
another user who is capable of issuing invites.
|
|
|
|
Args:
|
|
user_id: The user joining the room.
|
|
room_id: The room being joined.
|
|
remote_room_hosts: A list of remote room hosts.
|
|
content: The content to use as the event body of the join. This may
|
|
be modified.
|
|
is_partial_state_room: `True` if the server currently doesn't hold the full
|
|
state of the room.
|
|
is_host_in_room: `True` if the host is in the room.
|
|
partial_state_before_join: The state before the join event (i.e. the
|
|
resolution of the states after its parent events). May be full or
|
|
partial state, depending on `is_partial_state_room`.
|
|
|
|
Returns:
|
|
A tuple of:
|
|
True if a remote join should be performed. False if the join can be
|
|
done locally.
|
|
|
|
A list of remote room hosts to use. This is an empty list if a
|
|
local join is to be done.
|
|
"""
|
|
# If the host isn't in the room, pass through the prospective hosts.
|
|
if not is_host_in_room:
|
|
return True, remote_room_hosts
|
|
|
|
prev_member_event_id = partial_state_before_join.get(
|
|
(EventTypes.Member, user_id), None
|
|
)
|
|
previous_membership = None
|
|
if prev_member_event_id:
|
|
prev_member_event = await self.store.get_event(prev_member_event_id)
|
|
previous_membership = prev_member_event.membership
|
|
|
|
# If we are not fully joined yet, and the target is not already in the room,
|
|
# let's do a remote join so another server with the full state can validate
|
|
# that the user has not been banned for example.
|
|
# We could just accept the join and wait for state res to resolve that later on
|
|
# but we would then leak room history to this person until then, which is pretty
|
|
# bad.
|
|
if is_partial_state_room and previous_membership != Membership.JOIN:
|
|
return True, remote_room_hosts
|
|
|
|
# If the host is in the room, but not one of the authorised hosts
|
|
# for restricted join rules, a remote join must be used.
|
|
room_version = await self.store.get_room_version(room_id)
|
|
|
|
# If restricted join rules are not being used, a local join can always
|
|
# be used.
|
|
if not await self.event_auth_handler.has_restricted_join_rules(
|
|
partial_state_before_join, room_version
|
|
):
|
|
return False, []
|
|
|
|
# If the user is invited to the room or already joined, the join
|
|
# event can always be issued locally.
|
|
if previous_membership in (Membership.JOIN, Membership.INVITE):
|
|
return False, []
|
|
|
|
# All the partial state cases are covered above. We have been given the full
|
|
# state of the room.
|
|
assert not is_partial_state_room
|
|
state_before_join = partial_state_before_join
|
|
|
|
# If the local host has a user who can issue invites, then a local
|
|
# join can be done.
|
|
#
|
|
# If not, generate a new list of remote hosts based on which
|
|
# can issue invites.
|
|
event_map = await self.store.get_events(state_before_join.values())
|
|
current_state = {
|
|
state_key: event_map[event_id]
|
|
for state_key, event_id in state_before_join.items()
|
|
}
|
|
allowed_servers = get_servers_from_users(
|
|
get_users_which_can_issue_invite(current_state)
|
|
)
|
|
|
|
# If the local server is not one of allowed servers, then a remote
|
|
# join must be done. Return the list of prospective servers based on
|
|
# which can issue invites.
|
|
if self.hs.hostname not in allowed_servers:
|
|
return True, list(allowed_servers)
|
|
|
|
# Ensure the member should be allowed access via membership in a room.
|
|
await self.event_auth_handler.check_restricted_join_rules(
|
|
state_before_join, room_version, user_id, previous_membership
|
|
)
|
|
|
|
# If this is going to be a local join, additional information must
|
|
# be included in the event content in order to efficiently validate
|
|
# the event.
|
|
content[
|
|
EventContentFields.AUTHORISING_USER
|
|
] = await self.event_auth_handler.get_user_which_could_invite(
|
|
room_id,
|
|
state_before_join,
|
|
)
|
|
|
|
return False, []
|
|
|
|
async def transfer_room_state_on_room_upgrade(
|
|
self, old_room_id: str, room_id: str
|
|
) -> None:
|
|
"""Upon our server becoming aware of an upgraded room, either by upgrading a room
|
|
ourselves or joining one, we can transfer over information from the previous room.
|
|
|
|
Copies user state (tags/push rules) for every local user that was in the old room, as
|
|
well as migrating the room directory state.
|
|
|
|
Args:
|
|
old_room_id: The ID of the old room
|
|
room_id: The ID of the new room
|
|
"""
|
|
logger.info("Transferring room state from %s to %s", old_room_id, room_id)
|
|
|
|
# Find all local users that were in the old room and copy over each user's state
|
|
local_users = await self.store.get_local_users_in_room(old_room_id)
|
|
await self.copy_user_state_on_room_upgrade(old_room_id, room_id, local_users)
|
|
|
|
# Add new room to the room directory if the old room was there
|
|
# Remove old room from the room directory
|
|
old_room = await self.store.get_room(old_room_id)
|
|
if old_room is not None and old_room["is_public"]:
|
|
await self.store.set_room_is_public(old_room_id, False)
|
|
await self.store.set_room_is_public(room_id, True)
|
|
|
|
# Transfer alias mappings in the room directory
|
|
await self.store.update_aliases_for_room(old_room_id, room_id)
|
|
|
|
async def copy_user_state_on_room_upgrade(
|
|
self, old_room_id: str, new_room_id: str, user_ids: Iterable[str]
|
|
) -> None:
|
|
"""Copy user-specific information when they join a new room when that new room is the
|
|
result of a room upgrade
|
|
|
|
Args:
|
|
old_room_id: The ID of upgraded room
|
|
new_room_id: The ID of the new room
|
|
user_ids: User IDs to copy state for
|
|
"""
|
|
|
|
logger.debug(
|
|
"Copying over room tags and push rules from %s to %s for users %s",
|
|
old_room_id,
|
|
new_room_id,
|
|
user_ids,
|
|
)
|
|
|
|
for user_id in user_ids:
|
|
try:
|
|
# It is an upgraded room. Copy over old tags
|
|
await self.copy_room_tags_and_direct_to_room(
|
|
old_room_id, new_room_id, user_id
|
|
)
|
|
# Copy over push rules
|
|
await self.store.copy_push_rules_from_room_to_room_for_user(
|
|
old_room_id, new_room_id, user_id
|
|
)
|
|
except Exception:
|
|
logger.exception(
|
|
"Error copying tags and/or push rules from rooms %s to %s for user %s. "
|
|
"Skipping...",
|
|
old_room_id,
|
|
new_room_id,
|
|
user_id,
|
|
)
|
|
continue
|
|
|
|
async def send_membership_event(
|
|
self,
|
|
requester: Optional[Requester],
|
|
event: EventBase,
|
|
context: EventContext,
|
|
ratelimit: bool = True,
|
|
) -> None:
|
|
"""
|
|
Change the membership status of a user in a room.
|
|
|
|
Args:
|
|
requester: The local user who requested the membership
|
|
event. If None, certain checks, like whether this homeserver can
|
|
act as the sender, will be skipped.
|
|
event: The membership event.
|
|
context: The context of the event.
|
|
ratelimit: Whether to rate limit this request.
|
|
Raises:
|
|
SynapseError if there was a problem changing the membership.
|
|
PartialStateConflictError: if attempting to persist a partial state event in
|
|
a room that has been un-partial stated.
|
|
"""
|
|
target_user = UserID.from_string(event.state_key)
|
|
room_id = event.room_id
|
|
|
|
if requester is not None:
|
|
sender = UserID.from_string(event.sender)
|
|
assert (
|
|
sender == requester.user
|
|
), "Sender (%s) must be same as requester (%s)" % (sender, requester.user)
|
|
assert self.hs.is_mine(sender), "Sender must be our own: %s" % (sender,)
|
|
else:
|
|
requester = types.create_requester(target_user)
|
|
|
|
prev_state_ids = await context.get_prev_state_ids(
|
|
StateFilter.from_types([(EventTypes.GuestAccess, "")])
|
|
)
|
|
if event.membership == Membership.JOIN:
|
|
if requester.is_guest:
|
|
guest_can_join = await self._can_guest_join(prev_state_ids)
|
|
if not guest_can_join:
|
|
# This should be an auth check, but guests are a local concept,
|
|
# so don't really fit into the general auth process.
|
|
raise AuthError(403, "Guest access not allowed")
|
|
|
|
if event.membership not in (Membership.LEAVE, Membership.BAN):
|
|
is_blocked = await self.store.is_room_blocked(room_id)
|
|
if is_blocked:
|
|
raise SynapseError(403, "This room has been blocked on this server")
|
|
|
|
event = await self.event_creation_handler.handle_new_client_event(
|
|
requester,
|
|
events_and_context=[(event, context)],
|
|
extra_users=[target_user],
|
|
ratelimit=ratelimit,
|
|
)
|
|
|
|
if event.membership == Membership.LEAVE:
|
|
prev_state_ids = await context.get_prev_state_ids(
|
|
StateFilter.from_types([(EventTypes.Member, event.state_key)])
|
|
)
|
|
prev_member_event_id = prev_state_ids.get(
|
|
(EventTypes.Member, event.state_key), None
|
|
)
|
|
|
|
if prev_member_event_id:
|
|
prev_member_event = await self.store.get_event(prev_member_event_id)
|
|
if prev_member_event.membership == Membership.JOIN:
|
|
await self._user_left_room(target_user, room_id)
|
|
|
|
async def _can_guest_join(self, partial_current_state_ids: StateMap[str]) -> bool:
|
|
"""
|
|
Returns whether a guest can join a room based on its current state.
|
|
|
|
Args:
|
|
partial_current_state_ids: The current state of the room. May be full or
|
|
partial state.
|
|
"""
|
|
guest_access_id = partial_current_state_ids.get(
|
|
(EventTypes.GuestAccess, ""), None
|
|
)
|
|
if not guest_access_id:
|
|
return False
|
|
|
|
guest_access = await self.store.get_event(guest_access_id)
|
|
|
|
return bool(
|
|
guest_access
|
|
and guest_access.content
|
|
and guest_access.content.get(EventContentFields.GUEST_ACCESS)
|
|
== GuestAccess.CAN_JOIN
|
|
)
|
|
|
|
async def kick_guest_users(self, current_state: Iterable[EventBase]) -> None:
|
|
"""Kick any local guest users from the room.
|
|
|
|
This is called when the room state changes from guests allowed to not-allowed.
|
|
|
|
Params:
|
|
current_state: the current state of the room. We will iterate this to look
|
|
for guest users to kick.
|
|
"""
|
|
for member_event in current_state:
|
|
try:
|
|
if member_event.type != EventTypes.Member:
|
|
continue
|
|
|
|
if not self.hs.is_mine_id(member_event.state_key):
|
|
continue
|
|
|
|
if member_event.content["membership"] not in {
|
|
Membership.JOIN,
|
|
Membership.INVITE,
|
|
}:
|
|
continue
|
|
|
|
if (
|
|
"kind" not in member_event.content
|
|
or member_event.content["kind"] != "guest"
|
|
):
|
|
continue
|
|
|
|
# We make the user choose to leave, rather than have the
|
|
# event-sender kick them. This is partially because we don't
|
|
# need to worry about power levels, and partially because guest
|
|
# users are a concept which doesn't hugely work over federation,
|
|
# and having homeservers have their own users leave keeps more
|
|
# of that decision-making and control local to the guest-having
|
|
# homeserver.
|
|
target_user = UserID.from_string(member_event.state_key)
|
|
requester = create_requester(
|
|
target_user, is_guest=True, authenticated_entity=self._server_name
|
|
)
|
|
handler = self.hs.get_room_member_handler()
|
|
await handler.update_membership(
|
|
requester,
|
|
target_user,
|
|
member_event.room_id,
|
|
"leave",
|
|
ratelimit=False,
|
|
require_consent=False,
|
|
)
|
|
except Exception as e:
|
|
logger.exception("Error kicking guest user: %s" % (e,))
|
|
|
|
async def lookup_room_alias(
|
|
self, room_alias: RoomAlias
|
|
) -> Tuple[RoomID, List[str]]:
|
|
"""
|
|
Get the room ID associated with a room alias.
|
|
|
|
Args:
|
|
room_alias: The alias to look up.
|
|
Returns:
|
|
A tuple of:
|
|
The room ID as a RoomID object.
|
|
Hosts likely to be participating in the room ([str]).
|
|
Raises:
|
|
SynapseError if room alias could not be found.
|
|
"""
|
|
directory_handler = self.directory_handler
|
|
mapping = await directory_handler.get_association(room_alias)
|
|
|
|
if not mapping:
|
|
raise SynapseError(404, "No such room alias")
|
|
|
|
room_id = mapping["room_id"]
|
|
servers = mapping["servers"]
|
|
|
|
# put the server which owns the alias at the front of the server list.
|
|
if room_alias.domain in servers:
|
|
servers.remove(room_alias.domain)
|
|
servers.insert(0, room_alias.domain)
|
|
|
|
return RoomID.from_string(room_id), servers
|
|
|
|
async def _get_inviter(self, user_id: str, room_id: str) -> Optional[UserID]:
|
|
invite = await self.store.get_invite_for_local_user_in_room(
|
|
user_id=user_id, room_id=room_id
|
|
)
|
|
if invite:
|
|
return UserID.from_string(invite.sender)
|
|
return None
|
|
|
|
async def do_3pid_invite(
|
|
self,
|
|
room_id: str,
|
|
inviter: UserID,
|
|
medium: str,
|
|
address: str,
|
|
id_server: str,
|
|
requester: Requester,
|
|
txn_id: Optional[str],
|
|
id_access_token: str,
|
|
prev_event_ids: Optional[List[str]] = None,
|
|
depth: Optional[int] = None,
|
|
) -> Tuple[str, int]:
|
|
"""Invite a 3PID to a room.
|
|
|
|
Args:
|
|
room_id: The room to invite the 3PID to.
|
|
inviter: The user sending the invite.
|
|
medium: The 3PID's medium.
|
|
address: The 3PID's address.
|
|
id_server: The identity server to use.
|
|
requester: The user making the request.
|
|
txn_id: The transaction ID this is part of, or None if this is not
|
|
part of a transaction.
|
|
id_access_token: Identity server access token.
|
|
depth: Override the depth used to order the event in the DAG.
|
|
prev_event_ids: The event IDs to use as the prev events
|
|
Should normally be set to None, which will cause the depth to be calculated
|
|
based on the prev_events.
|
|
|
|
Returns:
|
|
Tuple of event ID and stream ordering position
|
|
|
|
Raises:
|
|
ShadowBanError if the requester has been shadow-banned.
|
|
"""
|
|
if self.config.server.block_non_admin_invites:
|
|
is_requester_admin = await self.auth.is_server_admin(requester)
|
|
if not is_requester_admin:
|
|
raise SynapseError(
|
|
403, "Invites have been disabled on this server", Codes.FORBIDDEN
|
|
)
|
|
|
|
if requester.shadow_banned:
|
|
# We randomly sleep a bit just to annoy the requester.
|
|
await self.clock.sleep(random.randint(1, 10))
|
|
raise ShadowBanError()
|
|
|
|
# We need to rate limit *before* we send out any 3PID invites, so we
|
|
# can't just rely on the standard ratelimiting of events.
|
|
await self._third_party_invite_limiter.ratelimit(requester)
|
|
|
|
can_invite = await self._third_party_event_rules.check_threepid_can_be_invited(
|
|
medium, address, room_id
|
|
)
|
|
if not can_invite:
|
|
raise SynapseError(
|
|
403,
|
|
"This third-party identifier can not be invited in this room",
|
|
Codes.FORBIDDEN,
|
|
)
|
|
|
|
if not self._enable_lookup:
|
|
raise SynapseError(
|
|
403, "Looking up third-party identifiers is denied from this server"
|
|
)
|
|
|
|
invitee = await self.identity_handler.lookup_3pid(
|
|
id_server, medium, address, id_access_token
|
|
)
|
|
|
|
if invitee:
|
|
# Note that update_membership with an action of "invite" can raise
|
|
# a ShadowBanError, but this was done above already.
|
|
# We don't check the invite against the spamchecker(s) here (through
|
|
# user_may_invite) because we'll do it further down the line anyway (in
|
|
# update_membership_locked).
|
|
event_id, stream_id = await self.update_membership(
|
|
requester, UserID.from_string(invitee), room_id, "invite", txn_id=txn_id
|
|
)
|
|
else:
|
|
# Check if the spamchecker(s) allow this invite to go through.
|
|
spam_check = (
|
|
await self._spam_checker_module_callbacks.user_may_send_3pid_invite(
|
|
inviter_userid=requester.user.to_string(),
|
|
medium=medium,
|
|
address=address,
|
|
room_id=room_id,
|
|
)
|
|
)
|
|
if spam_check != self._spam_checker_module_callbacks.NOT_SPAM:
|
|
raise SynapseError(
|
|
403,
|
|
"Cannot send threepid invite",
|
|
errcode=spam_check[0],
|
|
additional_fields=spam_check[1],
|
|
)
|
|
|
|
event, stream_id = await self._make_and_store_3pid_invite(
|
|
requester,
|
|
id_server,
|
|
medium,
|
|
address,
|
|
room_id,
|
|
inviter,
|
|
txn_id=txn_id,
|
|
id_access_token=id_access_token,
|
|
prev_event_ids=prev_event_ids,
|
|
depth=depth,
|
|
)
|
|
event_id = event.event_id
|
|
|
|
return event_id, stream_id
|
|
|
|
async def _make_and_store_3pid_invite(
|
|
self,
|
|
requester: Requester,
|
|
id_server: str,
|
|
medium: str,
|
|
address: str,
|
|
room_id: str,
|
|
user: UserID,
|
|
txn_id: Optional[str],
|
|
id_access_token: str,
|
|
prev_event_ids: Optional[List[str]] = None,
|
|
depth: Optional[int] = None,
|
|
) -> Tuple[EventBase, int]:
|
|
room_state = await self._storage_controllers.state.get_current_state(
|
|
room_id,
|
|
StateFilter.from_types(
|
|
[
|
|
(EventTypes.Member, user.to_string()),
|
|
(EventTypes.CanonicalAlias, ""),
|
|
(EventTypes.Name, ""),
|
|
(EventTypes.Create, ""),
|
|
(EventTypes.JoinRules, ""),
|
|
(EventTypes.RoomAvatar, ""),
|
|
]
|
|
),
|
|
)
|
|
|
|
inviter_display_name = ""
|
|
inviter_avatar_url = ""
|
|
member_event = room_state.get((EventTypes.Member, user.to_string()))
|
|
if member_event:
|
|
inviter_display_name = member_event.content.get("displayname", "")
|
|
inviter_avatar_url = member_event.content.get("avatar_url", "")
|
|
|
|
# if user has no display name, default to their MXID
|
|
if not inviter_display_name:
|
|
inviter_display_name = user.to_string()
|
|
|
|
canonical_room_alias = ""
|
|
canonical_alias_event = room_state.get((EventTypes.CanonicalAlias, ""))
|
|
if canonical_alias_event:
|
|
canonical_room_alias = canonical_alias_event.content.get("alias", "")
|
|
|
|
room_name = ""
|
|
room_name_event = room_state.get((EventTypes.Name, ""))
|
|
if room_name_event:
|
|
room_name = room_name_event.content.get("name", "")
|
|
|
|
room_type = None
|
|
room_create_event = room_state.get((EventTypes.Create, ""))
|
|
if room_create_event:
|
|
room_type = room_create_event.content.get(EventContentFields.ROOM_TYPE)
|
|
|
|
room_join_rules = ""
|
|
join_rules_event = room_state.get((EventTypes.JoinRules, ""))
|
|
if join_rules_event:
|
|
room_join_rules = join_rules_event.content.get("join_rule", "")
|
|
|
|
room_avatar_url = ""
|
|
room_avatar_event = room_state.get((EventTypes.RoomAvatar, ""))
|
|
if room_avatar_event:
|
|
room_avatar_url = room_avatar_event.content.get("url", "")
|
|
|
|
(
|
|
token,
|
|
public_keys,
|
|
fallback_public_key,
|
|
display_name,
|
|
) = await self.identity_handler.ask_id_server_for_third_party_invite(
|
|
requester=requester,
|
|
id_server=id_server,
|
|
medium=medium,
|
|
address=address,
|
|
room_id=room_id,
|
|
inviter_user_id=user.to_string(),
|
|
room_alias=canonical_room_alias,
|
|
room_avatar_url=room_avatar_url,
|
|
room_join_rules=room_join_rules,
|
|
room_name=room_name,
|
|
room_type=room_type,
|
|
inviter_display_name=inviter_display_name,
|
|
inviter_avatar_url=inviter_avatar_url,
|
|
id_access_token=id_access_token,
|
|
)
|
|
|
|
(
|
|
event,
|
|
stream_id,
|
|
) = await self.event_creation_handler.create_and_send_nonmember_event(
|
|
requester,
|
|
{
|
|
"type": EventTypes.ThirdPartyInvite,
|
|
"content": {
|
|
"display_name": display_name,
|
|
"public_keys": public_keys,
|
|
# For backwards compatibility:
|
|
"key_validity_url": fallback_public_key["key_validity_url"],
|
|
"public_key": fallback_public_key["public_key"],
|
|
},
|
|
"room_id": room_id,
|
|
"sender": user.to_string(),
|
|
"state_key": token,
|
|
},
|
|
ratelimit=False,
|
|
txn_id=txn_id,
|
|
prev_event_ids=prev_event_ids,
|
|
depth=depth,
|
|
)
|
|
return event, stream_id
|
|
|
|
async def _is_host_in_room(self, partial_current_state_ids: StateMap[str]) -> bool:
|
|
"""Returns whether the homeserver is in the room based on its current state.
|
|
|
|
Args:
|
|
partial_current_state_ids: The current state of the room. May be full or
|
|
partial state.
|
|
"""
|
|
# Have we just created the room, and is this about to be the very
|
|
# first member event?
|
|
create_event_id = partial_current_state_ids.get(("m.room.create", ""))
|
|
if len(partial_current_state_ids) == 1 and create_event_id:
|
|
# We can only get here if we're in the process of creating the room
|
|
return True
|
|
|
|
for etype, state_key in partial_current_state_ids:
|
|
if etype != EventTypes.Member or not self.hs.is_mine_id(state_key):
|
|
continue
|
|
|
|
event_id = partial_current_state_ids[(etype, state_key)]
|
|
event = await self.store.get_event(event_id, allow_none=True)
|
|
if not event:
|
|
continue
|
|
|
|
if event.membership == Membership.JOIN:
|
|
return True
|
|
|
|
return False
|
|
|
|
|
|
class RoomMemberMasterHandler(RoomMemberHandler):
|
|
def __init__(self, hs: "HomeServer"):
|
|
super().__init__(hs)
|
|
|
|
self.distributor = hs.get_distributor()
|
|
self.distributor.declare("user_left_room")
|
|
|
|
async def _is_remote_room_too_complex(
|
|
self, room_id: str, remote_room_hosts: List[str]
|
|
) -> Optional[bool]:
|
|
"""
|
|
Check if complexity of a remote room is too great.
|
|
|
|
Args:
|
|
room_id
|
|
remote_room_hosts
|
|
|
|
Returns: bool of whether the complexity is too great, or None
|
|
if unable to be fetched
|
|
"""
|
|
max_complexity = self.hs.config.server.limit_remote_rooms.complexity
|
|
complexity = await self.federation_handler.get_room_complexity(
|
|
remote_room_hosts, room_id
|
|
)
|
|
|
|
if complexity:
|
|
return complexity["v1"] > max_complexity
|
|
return None
|
|
|
|
async def _is_local_room_too_complex(self, room_id: str) -> bool:
|
|
"""
|
|
Check if the complexity of a local room is too great.
|
|
|
|
Args:
|
|
room_id: The room ID to check for complexity.
|
|
"""
|
|
max_complexity = self.hs.config.server.limit_remote_rooms.complexity
|
|
complexity = await self.store.get_room_complexity(room_id)
|
|
|
|
return complexity["v1"] > max_complexity
|
|
|
|
async def _remote_join(
|
|
self,
|
|
requester: Requester,
|
|
remote_room_hosts: List[str],
|
|
room_id: str,
|
|
user: UserID,
|
|
content: dict,
|
|
) -> Tuple[str, int]:
|
|
"""Implements RoomMemberHandler._remote_join"""
|
|
# filter ourselves out of remote_room_hosts: do_invite_join ignores it
|
|
# and if it is the only entry we'd like to return a 404 rather than a
|
|
# 500.
|
|
remote_room_hosts = [
|
|
host for host in remote_room_hosts if host != self.hs.hostname
|
|
]
|
|
|
|
if len(remote_room_hosts) == 0:
|
|
raise NoKnownServersError(
|
|
"Can't join remote room because no servers "
|
|
"that are in the room have been provided.",
|
|
)
|
|
|
|
check_complexity = self.hs.config.server.limit_remote_rooms.enabled
|
|
if (
|
|
check_complexity
|
|
and self.hs.config.server.limit_remote_rooms.admins_can_join
|
|
):
|
|
check_complexity = not await self.store.is_server_admin(user)
|
|
|
|
if check_complexity:
|
|
# Fetch the room complexity
|
|
too_complex = await self._is_remote_room_too_complex(
|
|
room_id, remote_room_hosts
|
|
)
|
|
if too_complex is True:
|
|
raise SynapseError(
|
|
code=400,
|
|
msg=self.hs.config.server.limit_remote_rooms.complexity_error,
|
|
errcode=Codes.RESOURCE_LIMIT_EXCEEDED,
|
|
)
|
|
|
|
# We don't do an auth check if we are doing an invite
|
|
# join dance for now, since we're kinda implicitly checking
|
|
# that we are allowed to join when we decide whether or not we
|
|
# need to do the invite/join dance.
|
|
event_id, stream_id = await self.federation_handler.do_invite_join(
|
|
remote_room_hosts, room_id, user.to_string(), content
|
|
)
|
|
|
|
# Check the room we just joined wasn't too large, if we didn't fetch the
|
|
# complexity of it before.
|
|
if check_complexity:
|
|
if too_complex is False:
|
|
# We checked, and we're under the limit.
|
|
return event_id, stream_id
|
|
|
|
# Check again, but with the local state events
|
|
too_complex = await self._is_local_room_too_complex(room_id)
|
|
|
|
if too_complex is False:
|
|
# We're under the limit.
|
|
return event_id, stream_id
|
|
|
|
# The room is too large. Leave.
|
|
requester = types.create_requester(
|
|
user, authenticated_entity=self._server_name
|
|
)
|
|
await self.update_membership(
|
|
requester=requester, target=user, room_id=room_id, action="leave"
|
|
)
|
|
raise SynapseError(
|
|
code=400,
|
|
msg=self.hs.config.server.limit_remote_rooms.complexity_error,
|
|
errcode=Codes.RESOURCE_LIMIT_EXCEEDED,
|
|
)
|
|
|
|
return event_id, stream_id
|
|
|
|
async def remote_reject_invite(
|
|
self,
|
|
invite_event_id: str,
|
|
txn_id: Optional[str],
|
|
requester: Requester,
|
|
content: JsonDict,
|
|
) -> Tuple[str, int]:
|
|
"""
|
|
Rejects an out-of-band invite received from a remote user
|
|
|
|
Implements RoomMemberHandler.remote_reject_invite
|
|
"""
|
|
invite_event = await self.store.get_event(invite_event_id)
|
|
room_id = invite_event.room_id
|
|
target_user = invite_event.state_key
|
|
|
|
# first of all, try doing a rejection via the inviting server
|
|
fed_handler = self.federation_handler
|
|
try:
|
|
inviter_id = UserID.from_string(invite_event.sender)
|
|
event, stream_id = await fed_handler.do_remotely_reject_invite(
|
|
[inviter_id.domain], room_id, target_user, content=content
|
|
)
|
|
return event.event_id, stream_id
|
|
except Exception as e:
|
|
# if we were unable to reject the invite, we will generate our own
|
|
# leave event.
|
|
#
|
|
# The 'except' clause is very broad, but we need to
|
|
# capture everything from DNS failures upwards
|
|
#
|
|
logger.warning("Failed to reject invite: %s", e)
|
|
|
|
return await self._generate_local_out_of_band_leave(
|
|
invite_event, txn_id, requester, content
|
|
)
|
|
|
|
async def remote_rescind_knock(
|
|
self,
|
|
knock_event_id: str,
|
|
txn_id: Optional[str],
|
|
requester: Requester,
|
|
content: JsonDict,
|
|
) -> Tuple[str, int]:
|
|
"""
|
|
Rescinds a local knock made on a remote room
|
|
|
|
Args:
|
|
knock_event_id: The ID of the knock event to rescind.
|
|
txn_id: The transaction ID to use.
|
|
requester: The originator of the request.
|
|
content: The content of the leave event.
|
|
|
|
Implements RoomMemberHandler.remote_rescind_knock
|
|
"""
|
|
# TODO: We don't yet support rescinding knocks over federation
|
|
# as we don't know which homeserver to send it to. An obvious
|
|
# candidate is the remote homeserver we originally knocked through,
|
|
# however we don't currently store that information.
|
|
|
|
# Just rescind the knock locally
|
|
knock_event = await self.store.get_event(knock_event_id)
|
|
return await self._generate_local_out_of_band_leave(
|
|
knock_event, txn_id, requester, content
|
|
)
|
|
|
|
async def _generate_local_out_of_band_leave(
|
|
self,
|
|
previous_membership_event: EventBase,
|
|
txn_id: Optional[str],
|
|
requester: Requester,
|
|
content: JsonDict,
|
|
) -> Tuple[str, int]:
|
|
"""Generate a local leave event for a room
|
|
|
|
This can be called after we e.g fail to reject an invite via a remote server.
|
|
It generates an out-of-band membership event locally.
|
|
|
|
Args:
|
|
previous_membership_event: the previous membership event for this user
|
|
txn_id: optional transaction ID supplied by the client
|
|
requester: user making the request, according to the access token
|
|
content: additional content to include in the leave event.
|
|
Normally an empty dict.
|
|
|
|
Returns:
|
|
A tuple containing (event_id, stream_id of the leave event)
|
|
"""
|
|
room_id = previous_membership_event.room_id
|
|
target_user = previous_membership_event.state_key
|
|
|
|
content["membership"] = Membership.LEAVE
|
|
|
|
event_dict = {
|
|
"type": EventTypes.Member,
|
|
"room_id": room_id,
|
|
"sender": target_user,
|
|
"content": content,
|
|
"state_key": target_user,
|
|
}
|
|
|
|
# the auth events for the new event are the same as that of the previous event, plus
|
|
# the event itself.
|
|
#
|
|
# the prev_events consist solely of the previous membership event.
|
|
prev_event_ids = [previous_membership_event.event_id]
|
|
auth_event_ids = (
|
|
list(previous_membership_event.auth_event_ids()) + prev_event_ids
|
|
)
|
|
|
|
# Try several times, it could fail with PartialStateConflictError
|
|
# in handle_new_client_event, cf comment in except block.
|
|
max_retries = 5
|
|
for i in range(max_retries):
|
|
try:
|
|
(
|
|
event,
|
|
unpersisted_context,
|
|
) = await self.event_creation_handler.create_event(
|
|
requester,
|
|
event_dict,
|
|
txn_id=txn_id,
|
|
prev_event_ids=prev_event_ids,
|
|
auth_event_ids=auth_event_ids,
|
|
outlier=True,
|
|
)
|
|
context = await unpersisted_context.persist(event)
|
|
event.internal_metadata.out_of_band_membership = True
|
|
|
|
result_event = (
|
|
await self.event_creation_handler.handle_new_client_event(
|
|
requester,
|
|
events_and_context=[(event, context)],
|
|
extra_users=[UserID.from_string(target_user)],
|
|
)
|
|
)
|
|
|
|
break
|
|
except PartialStateConflictError as e:
|
|
# Persisting couldn't happen because the room got un-partial stated
|
|
# in the meantime and context needs to be recomputed, so let's do so.
|
|
if i == max_retries - 1:
|
|
raise e
|
|
|
|
# we know it was persisted, so must have a stream ordering
|
|
assert result_event.internal_metadata.stream_ordering
|
|
|
|
return result_event.event_id, result_event.internal_metadata.stream_ordering
|
|
|
|
async def remote_knock(
|
|
self,
|
|
requester: Requester,
|
|
remote_room_hosts: List[str],
|
|
room_id: str,
|
|
user: UserID,
|
|
content: dict,
|
|
) -> Tuple[str, int]:
|
|
"""Sends a knock to a room. Attempts to do so via one remote out of a given list.
|
|
|
|
Args:
|
|
remote_room_hosts: A list of homeservers to try knocking through.
|
|
room_id: The ID of the room to knock on.
|
|
user: The user to knock on behalf of.
|
|
content: The content of the knock event.
|
|
|
|
Returns:
|
|
A tuple of (event ID, stream ID).
|
|
"""
|
|
# filter ourselves out of remote_room_hosts
|
|
remote_room_hosts = [
|
|
host for host in remote_room_hosts if host != self.hs.hostname
|
|
]
|
|
|
|
if len(remote_room_hosts) == 0:
|
|
raise NoKnownServersError()
|
|
|
|
return await self.federation_handler.do_knock(
|
|
remote_room_hosts, room_id, user.to_string(), content=content
|
|
)
|
|
|
|
async def _user_left_room(self, target: UserID, room_id: str) -> None:
|
|
"""Implements RoomMemberHandler._user_left_room"""
|
|
user_left_room(self.distributor, target, room_id)
|
|
|
|
|
|
class RoomForgetterHandler(StateDeltasHandler):
|
|
"""Forgets rooms when they are left, when enabled in the homeserver config.
|
|
|
|
For the purposes of this feature, kicks, bans and "leaves" via state resolution
|
|
weirdness are all considered to be leaves.
|
|
|
|
Derived from `StatsHandler` and `UserDirectoryHandler`.
|
|
"""
|
|
|
|
def __init__(self, hs: "HomeServer"):
|
|
super().__init__(hs)
|
|
|
|
self._hs = hs
|
|
self._store = hs.get_datastores().main
|
|
self._storage_controllers = hs.get_storage_controllers()
|
|
self._clock = hs.get_clock()
|
|
self._notifier = hs.get_notifier()
|
|
self._room_member_handler = hs.get_room_member_handler()
|
|
|
|
# The current position in the current_state_delta stream
|
|
self.pos: Optional[int] = None
|
|
|
|
# Guard to ensure we only process deltas one at a time
|
|
self._is_processing = False
|
|
|
|
if hs.config.worker.run_background_tasks:
|
|
self._notifier.add_replication_callback(self.notify_new_event)
|
|
|
|
# We kick this off to pick up outstanding work from before the last restart.
|
|
self._clock.call_later(0, self.notify_new_event)
|
|
|
|
def notify_new_event(self) -> None:
|
|
"""Called when there may be more deltas to process"""
|
|
if self._is_processing:
|
|
return
|
|
|
|
self._is_processing = True
|
|
|
|
async def process() -> None:
|
|
try:
|
|
await self._unsafe_process()
|
|
finally:
|
|
self._is_processing = False
|
|
|
|
run_as_background_process("room_forgetter.notify_new_event", process)
|
|
|
|
async def _unsafe_process(self) -> None:
|
|
# If self.pos is None then means we haven't fetched it from DB
|
|
if self.pos is None:
|
|
self.pos = await self._store.get_room_forgetter_stream_pos()
|
|
room_max_stream_ordering = self._store.get_room_max_stream_ordering()
|
|
if self.pos > room_max_stream_ordering:
|
|
# apparently, we've processed more events than exist in the database!
|
|
# this can happen if events are removed with history purge or similar.
|
|
logger.warning(
|
|
"Event stream ordering appears to have gone backwards (%i -> %i): "
|
|
"rewinding room forgetter processor",
|
|
self.pos,
|
|
room_max_stream_ordering,
|
|
)
|
|
self.pos = room_max_stream_ordering
|
|
|
|
if not self._hs.config.room.forget_on_leave:
|
|
# Update the processing position, so that if the server admin turns the
|
|
# feature on at a later date, we don't decide to forget every room that
|
|
# has ever been left in the past.
|
|
self.pos = self._store.get_room_max_stream_ordering()
|
|
await self._store.update_room_forgetter_stream_pos(self.pos)
|
|
return
|
|
|
|
# Loop round handling deltas until we're up to date
|
|
|
|
while True:
|
|
# Be sure to read the max stream_ordering *before* checking if there are any outstanding
|
|
# deltas, since there is otherwise a chance that we could miss updates which arrive
|
|
# after we check the deltas.
|
|
room_max_stream_ordering = self._store.get_room_max_stream_ordering()
|
|
if self.pos == room_max_stream_ordering:
|
|
break
|
|
|
|
logger.debug(
|
|
"Processing room forgetting %s->%s", self.pos, room_max_stream_ordering
|
|
)
|
|
(
|
|
max_pos,
|
|
deltas,
|
|
) = await self._storage_controllers.state.get_current_state_deltas(
|
|
self.pos, room_max_stream_ordering
|
|
)
|
|
|
|
logger.debug("Handling %d state deltas", len(deltas))
|
|
await self._handle_deltas(deltas)
|
|
|
|
self.pos = max_pos
|
|
|
|
# Expose current event processing position to prometheus
|
|
event_processing_positions.labels("room_forgetter").set(max_pos)
|
|
|
|
await self._store.update_room_forgetter_stream_pos(max_pos)
|
|
|
|
async def _handle_deltas(self, deltas: List[StateDelta]) -> None:
|
|
"""Called with the state deltas to process"""
|
|
for delta in deltas:
|
|
if delta.event_type != EventTypes.Member:
|
|
continue
|
|
|
|
if not self._hs.is_mine_id(delta.state_key):
|
|
continue
|
|
|
|
change = await self._get_key_change(
|
|
delta.prev_event_id,
|
|
delta.event_id,
|
|
key_name="membership",
|
|
public_value=Membership.JOIN,
|
|
)
|
|
is_leave = change is MatchChange.now_false
|
|
|
|
if is_leave:
|
|
try:
|
|
await self._room_member_handler.forget(
|
|
UserID.from_string(delta.state_key), delta.room_id
|
|
)
|
|
except SynapseError as e:
|
|
if e.code == 400:
|
|
# The user is back in the room.
|
|
pass
|
|
else:
|
|
raise
|
|
|
|
|
|
def get_users_which_can_issue_invite(auth_events: StateMap[EventBase]) -> List[str]:
|
|
"""
|
|
Return the list of users which can issue invites.
|
|
|
|
This is done by exploring the joined users and comparing their power levels
|
|
to the necessyar power level to issue an invite.
|
|
|
|
Args:
|
|
auth_events: state in force at this point in the room
|
|
|
|
Returns:
|
|
The users which can issue invites.
|
|
"""
|
|
invite_level = get_named_level(auth_events, "invite", 0)
|
|
users_default_level = get_named_level(auth_events, "users_default", 0)
|
|
power_level_event = get_power_level_event(auth_events)
|
|
|
|
# Custom power-levels for users.
|
|
if power_level_event:
|
|
users = power_level_event.content.get("users", {})
|
|
else:
|
|
users = {}
|
|
|
|
result = []
|
|
|
|
# Check which members are able to invite by ensuring they're joined and have
|
|
# the necessary power level.
|
|
for (event_type, state_key), event in auth_events.items():
|
|
if event_type != EventTypes.Member:
|
|
continue
|
|
|
|
if event.membership != Membership.JOIN:
|
|
continue
|
|
|
|
# Check if the user has a custom power level.
|
|
if users.get(state_key, users_default_level) >= invite_level:
|
|
result.append(state_key)
|
|
|
|
return result
|
|
|
|
|
|
def get_servers_from_users(users: List[str]) -> Set[str]:
|
|
"""
|
|
Resolve a list of users into their servers.
|
|
|
|
Args:
|
|
users: A list of users.
|
|
|
|
Returns:
|
|
A set of servers.
|
|
"""
|
|
servers = set()
|
|
for user in users:
|
|
try:
|
|
servers.add(get_domain_from_id(user))
|
|
except SynapseError:
|
|
pass
|
|
return servers
|
|
|