Return immutable objects for cachedList decorators (#16350)
This commit is contained in:
parent
5a66ff2f5c
commit
d7c89c5908
|
@ -0,0 +1 @@
|
|||
Improve type hints.
|
|
@ -23,7 +23,7 @@ from netaddr import IPSet
|
|||
|
||||
from synapse.api.constants import EventTypes
|
||||
from synapse.events import EventBase
|
||||
from synapse.types import DeviceListUpdates, JsonDict, UserID
|
||||
from synapse.types import DeviceListUpdates, JsonDict, JsonMapping, UserID
|
||||
from synapse.util.caches.descriptors import _CacheContext, cached
|
||||
|
||||
if TYPE_CHECKING:
|
||||
|
@ -379,8 +379,8 @@ class AppServiceTransaction:
|
|||
service: ApplicationService,
|
||||
id: int,
|
||||
events: Sequence[EventBase],
|
||||
ephemeral: List[JsonDict],
|
||||
to_device_messages: List[JsonDict],
|
||||
ephemeral: List[JsonMapping],
|
||||
to_device_messages: List[JsonMapping],
|
||||
one_time_keys_count: TransactionOneTimeKeysCount,
|
||||
unused_fallback_keys: TransactionUnusedFallbackKeys,
|
||||
device_list_summary: DeviceListUpdates,
|
||||
|
|
|
@ -41,7 +41,7 @@ from synapse.events import EventBase
|
|||
from synapse.events.utils import SerializeEventConfig, serialize_event
|
||||
from synapse.http.client import SimpleHttpClient, is_unknown_endpoint
|
||||
from synapse.logging import opentracing
|
||||
from synapse.types import DeviceListUpdates, JsonDict, ThirdPartyInstanceID
|
||||
from synapse.types import DeviceListUpdates, JsonDict, JsonMapping, ThirdPartyInstanceID
|
||||
from synapse.util.caches.response_cache import ResponseCache
|
||||
|
||||
if TYPE_CHECKING:
|
||||
|
@ -306,8 +306,8 @@ class ApplicationServiceApi(SimpleHttpClient):
|
|||
self,
|
||||
service: "ApplicationService",
|
||||
events: Sequence[EventBase],
|
||||
ephemeral: List[JsonDict],
|
||||
to_device_messages: List[JsonDict],
|
||||
ephemeral: List[JsonMapping],
|
||||
to_device_messages: List[JsonMapping],
|
||||
one_time_keys_count: TransactionOneTimeKeysCount,
|
||||
unused_fallback_keys: TransactionUnusedFallbackKeys,
|
||||
device_list_summary: DeviceListUpdates,
|
||||
|
|
|
@ -73,7 +73,7 @@ from synapse.events import EventBase
|
|||
from synapse.logging.context import run_in_background
|
||||
from synapse.metrics.background_process_metrics import run_as_background_process
|
||||
from synapse.storage.databases.main import DataStore
|
||||
from synapse.types import DeviceListUpdates, JsonDict
|
||||
from synapse.types import DeviceListUpdates, JsonMapping
|
||||
from synapse.util import Clock
|
||||
|
||||
if TYPE_CHECKING:
|
||||
|
@ -121,8 +121,8 @@ class ApplicationServiceScheduler:
|
|||
self,
|
||||
appservice: ApplicationService,
|
||||
events: Optional[Collection[EventBase]] = None,
|
||||
ephemeral: Optional[Collection[JsonDict]] = None,
|
||||
to_device_messages: Optional[Collection[JsonDict]] = None,
|
||||
ephemeral: Optional[Collection[JsonMapping]] = None,
|
||||
to_device_messages: Optional[Collection[JsonMapping]] = None,
|
||||
device_list_summary: Optional[DeviceListUpdates] = None,
|
||||
) -> None:
|
||||
"""
|
||||
|
@ -180,9 +180,9 @@ class _ServiceQueuer:
|
|||
# dict of {service_id: [events]}
|
||||
self.queued_events: Dict[str, List[EventBase]] = {}
|
||||
# dict of {service_id: [events]}
|
||||
self.queued_ephemeral: Dict[str, List[JsonDict]] = {}
|
||||
self.queued_ephemeral: Dict[str, List[JsonMapping]] = {}
|
||||
# dict of {service_id: [to_device_message_json]}
|
||||
self.queued_to_device_messages: Dict[str, List[JsonDict]] = {}
|
||||
self.queued_to_device_messages: Dict[str, List[JsonMapping]] = {}
|
||||
# dict of {service_id: [device_list_summary]}
|
||||
self.queued_device_list_summaries: Dict[str, List[DeviceListUpdates]] = {}
|
||||
|
||||
|
@ -293,8 +293,8 @@ class _ServiceQueuer:
|
|||
self,
|
||||
service: ApplicationService,
|
||||
events: Iterable[EventBase],
|
||||
ephemerals: Iterable[JsonDict],
|
||||
to_device_messages: Iterable[JsonDict],
|
||||
ephemerals: Iterable[JsonMapping],
|
||||
to_device_messages: Iterable[JsonMapping],
|
||||
) -> Tuple[TransactionOneTimeKeysCount, TransactionUnusedFallbackKeys]:
|
||||
"""
|
||||
Given a list of the events, ephemeral messages and to-device messages,
|
||||
|
@ -364,8 +364,8 @@ class _TransactionController:
|
|||
self,
|
||||
service: ApplicationService,
|
||||
events: Sequence[EventBase],
|
||||
ephemeral: Optional[List[JsonDict]] = None,
|
||||
to_device_messages: Optional[List[JsonDict]] = None,
|
||||
ephemeral: Optional[List[JsonMapping]] = None,
|
||||
to_device_messages: Optional[List[JsonMapping]] = None,
|
||||
one_time_keys_count: Optional[TransactionOneTimeKeysCount] = None,
|
||||
unused_fallback_keys: Optional[TransactionUnusedFallbackKeys] = None,
|
||||
device_list_summary: Optional[DeviceListUpdates] = None,
|
||||
|
|
|
@ -46,6 +46,7 @@ from synapse.storage.databases.main.directory import RoomAliasMapping
|
|||
from synapse.types import (
|
||||
DeviceListUpdates,
|
||||
JsonDict,
|
||||
JsonMapping,
|
||||
RoomAlias,
|
||||
RoomStreamToken,
|
||||
StreamKeyType,
|
||||
|
@ -397,7 +398,7 @@ class ApplicationServicesHandler:
|
|||
|
||||
async def _handle_typing(
|
||||
self, service: ApplicationService, new_token: int
|
||||
) -> List[JsonDict]:
|
||||
) -> List[JsonMapping]:
|
||||
"""
|
||||
Return the typing events since the given stream token that the given application
|
||||
service should receive.
|
||||
|
@ -432,7 +433,7 @@ class ApplicationServicesHandler:
|
|||
|
||||
async def _handle_receipts(
|
||||
self, service: ApplicationService, new_token: int
|
||||
) -> List[JsonDict]:
|
||||
) -> List[JsonMapping]:
|
||||
"""
|
||||
Return the latest read receipts that the given application service should receive.
|
||||
|
||||
|
@ -471,7 +472,7 @@ class ApplicationServicesHandler:
|
|||
service: ApplicationService,
|
||||
users: Collection[Union[str, UserID]],
|
||||
new_token: Optional[int],
|
||||
) -> List[JsonDict]:
|
||||
) -> List[JsonMapping]:
|
||||
"""
|
||||
Return the latest presence updates that the given application service should receive.
|
||||
|
||||
|
@ -491,7 +492,7 @@ class ApplicationServicesHandler:
|
|||
A list of json dictionaries containing data derived from the presence events
|
||||
that should be sent to the given application service.
|
||||
"""
|
||||
events: List[JsonDict] = []
|
||||
events: List[JsonMapping] = []
|
||||
presence_source = self.event_sources.sources.presence
|
||||
from_key = await self.store.get_type_stream_id_for_appservice(
|
||||
service, "presence"
|
||||
|
|
|
@ -14,7 +14,7 @@
|
|||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
import logging
|
||||
from typing import TYPE_CHECKING, Any, Dict, Iterable, List, Mapping, Optional, Tuple
|
||||
from typing import TYPE_CHECKING, Dict, Iterable, List, Mapping, Optional, Tuple
|
||||
|
||||
import attr
|
||||
from canonicaljson import encode_canonical_json
|
||||
|
@ -31,6 +31,7 @@ from synapse.logging.context import make_deferred_yieldable, run_in_background
|
|||
from synapse.logging.opentracing import log_kv, set_tag, tag_args, trace
|
||||
from synapse.types import (
|
||||
JsonDict,
|
||||
JsonMapping,
|
||||
UserID,
|
||||
get_domain_from_id,
|
||||
get_verify_key_from_cross_signing_key,
|
||||
|
@ -272,11 +273,7 @@ class E2eKeysHandler:
|
|||
delay_cancellation=True,
|
||||
)
|
||||
|
||||
ret = {"device_keys": results, "failures": failures}
|
||||
|
||||
ret.update(cross_signing_keys)
|
||||
|
||||
return ret
|
||||
return {"device_keys": results, "failures": failures, **cross_signing_keys}
|
||||
|
||||
@trace
|
||||
async def _query_devices_for_destination(
|
||||
|
@ -408,7 +405,7 @@ class E2eKeysHandler:
|
|||
@cancellable
|
||||
async def get_cross_signing_keys_from_cache(
|
||||
self, query: Iterable[str], from_user_id: Optional[str]
|
||||
) -> Dict[str, Dict[str, dict]]:
|
||||
) -> Dict[str, Dict[str, JsonMapping]]:
|
||||
"""Get cross-signing keys for users from the database
|
||||
|
||||
Args:
|
||||
|
@ -551,16 +548,13 @@ class E2eKeysHandler:
|
|||
self.config.federation.allow_device_name_lookup_over_federation
|
||||
),
|
||||
)
|
||||
ret = {"device_keys": res}
|
||||
|
||||
# add in the cross-signing keys
|
||||
cross_signing_keys = await self.get_cross_signing_keys_from_cache(
|
||||
device_keys_query, None
|
||||
)
|
||||
|
||||
ret.update(cross_signing_keys)
|
||||
|
||||
return ret
|
||||
return {"device_keys": res, **cross_signing_keys}
|
||||
|
||||
async def claim_local_one_time_keys(
|
||||
self,
|
||||
|
@ -1127,7 +1121,7 @@ class E2eKeysHandler:
|
|||
user_id: str,
|
||||
master_key_id: str,
|
||||
signed_master_key: JsonDict,
|
||||
stored_master_key: JsonDict,
|
||||
stored_master_key: JsonMapping,
|
||||
devices: Dict[str, Dict[str, JsonDict]],
|
||||
) -> List["SignatureListItem"]:
|
||||
"""Check signatures of a user's master key made by their devices.
|
||||
|
@ -1278,7 +1272,7 @@ class E2eKeysHandler:
|
|||
|
||||
async def _get_e2e_cross_signing_verify_key(
|
||||
self, user_id: str, key_type: str, from_user_id: Optional[str] = None
|
||||
) -> Tuple[JsonDict, str, VerifyKey]:
|
||||
) -> Tuple[JsonMapping, str, VerifyKey]:
|
||||
"""Fetch locally or remotely query for a cross-signing public key.
|
||||
|
||||
First, attempt to fetch the cross-signing public key from storage.
|
||||
|
@ -1333,7 +1327,7 @@ class E2eKeysHandler:
|
|||
self,
|
||||
user: UserID,
|
||||
desired_key_type: str,
|
||||
) -> Optional[Tuple[Dict[str, Any], str, VerifyKey]]:
|
||||
) -> Optional[Tuple[JsonMapping, str, VerifyKey]]:
|
||||
"""Queries cross-signing keys for a remote user and saves them to the database
|
||||
|
||||
Only the key specified by `key_type` will be returned, while all retrieved keys
|
||||
|
@ -1474,7 +1468,7 @@ def _check_device_signature(
|
|||
user_id: str,
|
||||
verify_key: VerifyKey,
|
||||
signed_device: JsonDict,
|
||||
stored_device: JsonDict,
|
||||
stored_device: JsonMapping,
|
||||
) -> None:
|
||||
"""Check that a signature on a device or cross-signing key is correct and
|
||||
matches the copy of the device/key that we have stored. Throws an
|
||||
|
|
|
@ -32,6 +32,7 @@ from synapse.storage.roommember import RoomsForUser
|
|||
from synapse.streams.config import PaginationConfig
|
||||
from synapse.types import (
|
||||
JsonDict,
|
||||
JsonMapping,
|
||||
Requester,
|
||||
RoomStreamToken,
|
||||
StreamKeyType,
|
||||
|
@ -454,7 +455,7 @@ class InitialSyncHandler:
|
|||
for s in states
|
||||
]
|
||||
|
||||
async def get_receipts() -> List[JsonDict]:
|
||||
async def get_receipts() -> List[JsonMapping]:
|
||||
receipts = await self.store.get_linearized_receipts_for_room(
|
||||
room_id, to_key=now_token.receipt_key
|
||||
)
|
||||
|
|
|
@ -19,6 +19,7 @@ from synapse.appservice import ApplicationService
|
|||
from synapse.streams import EventSource
|
||||
from synapse.types import (
|
||||
JsonDict,
|
||||
JsonMapping,
|
||||
ReadReceipt,
|
||||
StreamKeyType,
|
||||
UserID,
|
||||
|
@ -204,15 +205,15 @@ class ReceiptsHandler:
|
|||
await self.federation_sender.send_read_receipt(receipt)
|
||||
|
||||
|
||||
class ReceiptEventSource(EventSource[int, JsonDict]):
|
||||
class ReceiptEventSource(EventSource[int, JsonMapping]):
|
||||
def __init__(self, hs: "HomeServer"):
|
||||
self.store = hs.get_datastores().main
|
||||
self.config = hs.config
|
||||
|
||||
@staticmethod
|
||||
def filter_out_private_receipts(
|
||||
rooms: Sequence[JsonDict], user_id: str
|
||||
) -> List[JsonDict]:
|
||||
rooms: Sequence[JsonMapping], user_id: str
|
||||
) -> List[JsonMapping]:
|
||||
"""
|
||||
Filters a list of serialized receipts (as returned by /sync and /initialSync)
|
||||
and removes private read receipts of other users.
|
||||
|
@ -229,7 +230,7 @@ class ReceiptEventSource(EventSource[int, JsonDict]):
|
|||
The same as rooms, but filtered.
|
||||
"""
|
||||
|
||||
result = []
|
||||
result: List[JsonMapping] = []
|
||||
|
||||
# Iterate through each room's receipt content.
|
||||
for room in rooms:
|
||||
|
@ -282,7 +283,7 @@ class ReceiptEventSource(EventSource[int, JsonDict]):
|
|||
room_ids: Iterable[str],
|
||||
is_guest: bool,
|
||||
explicit_room_id: Optional[str] = None,
|
||||
) -> Tuple[List[JsonDict], int]:
|
||||
) -> Tuple[List[JsonMapping], int]:
|
||||
from_key = int(from_key)
|
||||
to_key = self.get_current_key()
|
||||
|
||||
|
@ -301,7 +302,7 @@ class ReceiptEventSource(EventSource[int, JsonDict]):
|
|||
|
||||
async def get_new_events_as(
|
||||
self, from_key: int, to_key: int, service: ApplicationService
|
||||
) -> Tuple[List[JsonDict], int]:
|
||||
) -> Tuple[List[JsonMapping], int]:
|
||||
"""Returns a set of new read receipt events that an appservice
|
||||
may be interested in.
|
||||
|
||||
|
|
|
@ -235,7 +235,7 @@ class SyncResult:
|
|||
archived: List[ArchivedSyncResult]
|
||||
to_device: List[JsonDict]
|
||||
device_lists: DeviceListUpdates
|
||||
device_one_time_keys_count: JsonDict
|
||||
device_one_time_keys_count: JsonMapping
|
||||
device_unused_fallback_key_types: List[str]
|
||||
|
||||
def __bool__(self) -> bool:
|
||||
|
@ -1558,7 +1558,7 @@ class SyncHandler:
|
|||
|
||||
logger.debug("Fetching OTK data")
|
||||
device_id = sync_config.device_id
|
||||
one_time_keys_count: JsonDict = {}
|
||||
one_time_keys_count: JsonMapping = {}
|
||||
unused_fallback_key_types: List[str] = []
|
||||
if device_id:
|
||||
# TODO: We should have a way to let clients differentiate between the states of:
|
||||
|
|
|
@ -26,7 +26,14 @@ from synapse.metrics.background_process_metrics import (
|
|||
)
|
||||
from synapse.replication.tcp.streams import TypingStream
|
||||
from synapse.streams import EventSource
|
||||
from synapse.types import JsonDict, Requester, StrCollection, StreamKeyType, UserID
|
||||
from synapse.types import (
|
||||
JsonDict,
|
||||
JsonMapping,
|
||||
Requester,
|
||||
StrCollection,
|
||||
StreamKeyType,
|
||||
UserID,
|
||||
)
|
||||
from synapse.util.caches.stream_change_cache import StreamChangeCache
|
||||
from synapse.util.metrics import Measure
|
||||
from synapse.util.retryutils import filter_destinations_by_retry_limiter
|
||||
|
@ -487,7 +494,7 @@ class TypingWriterHandler(FollowerTypingHandler):
|
|||
raise Exception("Typing writer instance got typing info over replication")
|
||||
|
||||
|
||||
class TypingNotificationEventSource(EventSource[int, JsonDict]):
|
||||
class TypingNotificationEventSource(EventSource[int, JsonMapping]):
|
||||
def __init__(self, hs: "HomeServer"):
|
||||
self._main_store = hs.get_datastores().main
|
||||
self.clock = hs.get_clock()
|
||||
|
@ -497,7 +504,7 @@ class TypingNotificationEventSource(EventSource[int, JsonDict]):
|
|||
#
|
||||
self.get_typing_handler = hs.get_typing_handler
|
||||
|
||||
def _make_event_for(self, room_id: str) -> JsonDict:
|
||||
def _make_event_for(self, room_id: str) -> JsonMapping:
|
||||
typing = self.get_typing_handler()._room_typing[room_id]
|
||||
return {
|
||||
"type": EduTypes.TYPING,
|
||||
|
@ -507,7 +514,7 @@ class TypingNotificationEventSource(EventSource[int, JsonDict]):
|
|||
|
||||
async def get_new_events_as(
|
||||
self, from_key: int, service: ApplicationService
|
||||
) -> Tuple[List[JsonDict], int]:
|
||||
) -> Tuple[List[JsonMapping], int]:
|
||||
"""Returns a set of new typing events that an appservice
|
||||
may be interested in.
|
||||
|
||||
|
@ -551,7 +558,7 @@ class TypingNotificationEventSource(EventSource[int, JsonDict]):
|
|||
room_ids: Iterable[str],
|
||||
is_guest: bool,
|
||||
explicit_room_id: Optional[str] = None,
|
||||
) -> Tuple[List[JsonDict], int]:
|
||||
) -> Tuple[List[JsonMapping], int]:
|
||||
with Measure(self.clock, "typing.get_new_events"):
|
||||
from_key = int(from_key)
|
||||
handler = self.get_typing_handler()
|
||||
|
|
|
@ -131,7 +131,7 @@ class BulkPushRuleEvaluator:
|
|||
async def _get_rules_for_event(
|
||||
self,
|
||||
event: EventBase,
|
||||
) -> Dict[str, FilteredPushRules]:
|
||||
) -> Mapping[str, FilteredPushRules]:
|
||||
"""Get the push rules for all users who may need to be notified about
|
||||
the event.
|
||||
|
||||
|
|
|
@ -45,7 +45,7 @@ from synapse.storage.databases.main.events_worker import EventsWorkerStore
|
|||
from synapse.storage.databases.main.roommember import RoomMemberWorkerStore
|
||||
from synapse.storage.types import Cursor
|
||||
from synapse.storage.util.sequence import build_sequence_generator
|
||||
from synapse.types import DeviceListUpdates, JsonDict
|
||||
from synapse.types import DeviceListUpdates, JsonMapping
|
||||
from synapse.util import json_encoder
|
||||
from synapse.util.caches.descriptors import _CacheContext, cached
|
||||
|
||||
|
@ -268,8 +268,8 @@ class ApplicationServiceTransactionWorkerStore(
|
|||
self,
|
||||
service: ApplicationService,
|
||||
events: Sequence[EventBase],
|
||||
ephemeral: List[JsonDict],
|
||||
to_device_messages: List[JsonDict],
|
||||
ephemeral: List[JsonMapping],
|
||||
to_device_messages: List[JsonMapping],
|
||||
one_time_keys_count: TransactionOneTimeKeysCount,
|
||||
unused_fallback_keys: TransactionUnusedFallbackKeys,
|
||||
device_list_summary: DeviceListUpdates,
|
||||
|
|
|
@ -55,7 +55,12 @@ from synapse.storage.util.id_generators import (
|
|||
AbstractStreamIdGenerator,
|
||||
StreamIdGenerator,
|
||||
)
|
||||
from synapse.types import JsonDict, StrCollection, get_verify_key_from_cross_signing_key
|
||||
from synapse.types import (
|
||||
JsonDict,
|
||||
JsonMapping,
|
||||
StrCollection,
|
||||
get_verify_key_from_cross_signing_key,
|
||||
)
|
||||
from synapse.util import json_decoder, json_encoder
|
||||
from synapse.util.caches.descriptors import cached, cachedList
|
||||
from synapse.util.caches.lrucache import LruCache
|
||||
|
@ -746,7 +751,7 @@ class DeviceWorkerStore(RoomMemberWorkerStore, EndToEndKeyWorkerStore):
|
|||
@cancellable
|
||||
async def get_user_devices_from_cache(
|
||||
self, user_ids: Set[str], user_and_device_ids: List[Tuple[str, str]]
|
||||
) -> Tuple[Set[str], Dict[str, Mapping[str, JsonDict]]]:
|
||||
) -> Tuple[Set[str], Dict[str, Mapping[str, JsonMapping]]]:
|
||||
"""Get the devices (and keys if any) for remote users from the cache.
|
||||
|
||||
Args:
|
||||
|
@ -766,13 +771,13 @@ class DeviceWorkerStore(RoomMemberWorkerStore, EndToEndKeyWorkerStore):
|
|||
user_ids_not_in_cache = unique_user_ids - user_ids_in_cache
|
||||
|
||||
# First fetch all the users which all devices are to be returned.
|
||||
results: Dict[str, Mapping[str, JsonDict]] = {}
|
||||
results: Dict[str, Mapping[str, JsonMapping]] = {}
|
||||
for user_id in user_ids:
|
||||
if user_id in user_ids_in_cache:
|
||||
results[user_id] = await self.get_cached_devices_for_user(user_id)
|
||||
# Then fetch all device-specific requests, but skip users we've already
|
||||
# fetched all devices for.
|
||||
device_specific_results: Dict[str, Dict[str, JsonDict]] = {}
|
||||
device_specific_results: Dict[str, Dict[str, JsonMapping]] = {}
|
||||
for user_id, device_id in user_and_device_ids:
|
||||
if user_id in user_ids_in_cache and user_id not in user_ids:
|
||||
device = await self._get_cached_user_device(user_id, device_id)
|
||||
|
@ -801,7 +806,9 @@ class DeviceWorkerStore(RoomMemberWorkerStore, EndToEndKeyWorkerStore):
|
|||
return user_ids_in_cache
|
||||
|
||||
@cached(num_args=2, tree=True)
|
||||
async def _get_cached_user_device(self, user_id: str, device_id: str) -> JsonDict:
|
||||
async def _get_cached_user_device(
|
||||
self, user_id: str, device_id: str
|
||||
) -> JsonMapping:
|
||||
content = await self.db_pool.simple_select_one_onecol(
|
||||
table="device_lists_remote_cache",
|
||||
keyvalues={"user_id": user_id, "device_id": device_id},
|
||||
|
@ -811,7 +818,9 @@ class DeviceWorkerStore(RoomMemberWorkerStore, EndToEndKeyWorkerStore):
|
|||
return db_to_json(content)
|
||||
|
||||
@cached()
|
||||
async def get_cached_devices_for_user(self, user_id: str) -> Mapping[str, JsonDict]:
|
||||
async def get_cached_devices_for_user(
|
||||
self, user_id: str
|
||||
) -> Mapping[str, JsonMapping]:
|
||||
devices = await self.db_pool.simple_select_list(
|
||||
table="device_lists_remote_cache",
|
||||
keyvalues={"user_id": user_id},
|
||||
|
@ -1042,7 +1051,7 @@ class DeviceWorkerStore(RoomMemberWorkerStore, EndToEndKeyWorkerStore):
|
|||
)
|
||||
async def get_device_list_last_stream_id_for_remotes(
|
||||
self, user_ids: Iterable[str]
|
||||
) -> Dict[str, Optional[str]]:
|
||||
) -> Mapping[str, Optional[str]]:
|
||||
rows = await self.db_pool.simple_select_many_batch(
|
||||
table="device_lists_remote_extremeties",
|
||||
column="user_id",
|
||||
|
|
|
@ -52,7 +52,7 @@ from synapse.storage.database import (
|
|||
from synapse.storage.databases.main.cache import CacheInvalidationWorkerStore
|
||||
from synapse.storage.engines import PostgresEngine
|
||||
from synapse.storage.util.id_generators import StreamIdGenerator
|
||||
from synapse.types import JsonDict
|
||||
from synapse.types import JsonDict, JsonMapping
|
||||
from synapse.util import json_decoder, json_encoder
|
||||
from synapse.util.caches.descriptors import cached, cachedList
|
||||
from synapse.util.cancellation import cancellable
|
||||
|
@ -125,7 +125,7 @@ class EndToEndKeyWorkerStore(EndToEndKeyBackgroundStore, CacheInvalidationWorker
|
|||
|
||||
async def get_e2e_device_keys_for_federation_query(
|
||||
self, user_id: str
|
||||
) -> Tuple[int, List[JsonDict]]:
|
||||
) -> Tuple[int, Sequence[JsonMapping]]:
|
||||
"""Get all devices (with any device keys) for a user
|
||||
|
||||
Returns:
|
||||
|
@ -174,7 +174,7 @@ class EndToEndKeyWorkerStore(EndToEndKeyBackgroundStore, CacheInvalidationWorker
|
|||
@cached(iterable=True)
|
||||
async def _get_e2e_device_keys_for_federation_query_inner(
|
||||
self, user_id: str
|
||||
) -> List[JsonDict]:
|
||||
) -> Sequence[JsonMapping]:
|
||||
"""Get all devices (with any device keys) for a user"""
|
||||
|
||||
devices = await self.get_e2e_device_keys_and_signatures([(user_id, None)])
|
||||
|
@ -578,7 +578,7 @@ class EndToEndKeyWorkerStore(EndToEndKeyBackgroundStore, CacheInvalidationWorker
|
|||
@cached(max_entries=10000)
|
||||
async def count_e2e_one_time_keys(
|
||||
self, user_id: str, device_id: str
|
||||
) -> Dict[str, int]:
|
||||
) -> Mapping[str, int]:
|
||||
"""Count the number of one time keys the server has for a device
|
||||
Returns:
|
||||
A mapping from algorithm to number of keys for that algorithm.
|
||||
|
@ -812,7 +812,7 @@ class EndToEndKeyWorkerStore(EndToEndKeyBackgroundStore, CacheInvalidationWorker
|
|||
|
||||
async def get_e2e_cross_signing_key(
|
||||
self, user_id: str, key_type: str, from_user_id: Optional[str] = None
|
||||
) -> Optional[JsonDict]:
|
||||
) -> Optional[JsonMapping]:
|
||||
"""Returns a user's cross-signing key.
|
||||
|
||||
Args:
|
||||
|
@ -833,7 +833,9 @@ class EndToEndKeyWorkerStore(EndToEndKeyBackgroundStore, CacheInvalidationWorker
|
|||
return user_keys.get(key_type)
|
||||
|
||||
@cached(num_args=1)
|
||||
def _get_bare_e2e_cross_signing_keys(self, user_id: str) -> Mapping[str, JsonDict]:
|
||||
def _get_bare_e2e_cross_signing_keys(
|
||||
self, user_id: str
|
||||
) -> Mapping[str, JsonMapping]:
|
||||
"""Dummy function. Only used to make a cache for
|
||||
_get_bare_e2e_cross_signing_keys_bulk.
|
||||
"""
|
||||
|
@ -846,7 +848,7 @@ class EndToEndKeyWorkerStore(EndToEndKeyBackgroundStore, CacheInvalidationWorker
|
|||
)
|
||||
async def _get_bare_e2e_cross_signing_keys_bulk(
|
||||
self, user_ids: Iterable[str]
|
||||
) -> Dict[str, Optional[Mapping[str, JsonDict]]]:
|
||||
) -> Mapping[str, Optional[Mapping[str, JsonMapping]]]:
|
||||
"""Returns the cross-signing keys for a set of users. The output of this
|
||||
function should be passed to _get_e2e_cross_signing_signatures_txn if
|
||||
the signatures for the calling user need to be fetched.
|
||||
|
@ -860,15 +862,12 @@ class EndToEndKeyWorkerStore(EndToEndKeyBackgroundStore, CacheInvalidationWorker
|
|||
their user ID will map to None.
|
||||
|
||||
"""
|
||||
result = await self.db_pool.runInteraction(
|
||||
return await self.db_pool.runInteraction(
|
||||
"get_bare_e2e_cross_signing_keys_bulk",
|
||||
self._get_bare_e2e_cross_signing_keys_bulk_txn,
|
||||
user_ids,
|
||||
)
|
||||
|
||||
# The `Optional` comes from the `@cachedList` decorator.
|
||||
return cast(Dict[str, Optional[Mapping[str, JsonDict]]], result)
|
||||
|
||||
def _get_bare_e2e_cross_signing_keys_bulk_txn(
|
||||
self,
|
||||
txn: LoggingTransaction,
|
||||
|
@ -1026,7 +1025,7 @@ class EndToEndKeyWorkerStore(EndToEndKeyBackgroundStore, CacheInvalidationWorker
|
|||
@cancellable
|
||||
async def get_e2e_cross_signing_keys_bulk(
|
||||
self, user_ids: List[str], from_user_id: Optional[str] = None
|
||||
) -> Dict[str, Optional[Mapping[str, JsonDict]]]:
|
||||
) -> Mapping[str, Optional[Mapping[str, JsonMapping]]]:
|
||||
"""Returns the cross-signing keys for a set of users.
|
||||
|
||||
Args:
|
||||
|
@ -1043,7 +1042,7 @@ class EndToEndKeyWorkerStore(EndToEndKeyBackgroundStore, CacheInvalidationWorker
|
|||
|
||||
if from_user_id:
|
||||
result = cast(
|
||||
Dict[str, Optional[Mapping[str, JsonDict]]],
|
||||
Dict[str, Optional[Mapping[str, JsonMapping]]],
|
||||
await self.db_pool.runInteraction(
|
||||
"get_e2e_cross_signing_signatures",
|
||||
self._get_e2e_cross_signing_signatures_txn,
|
||||
|
|
|
@ -24,6 +24,7 @@ from typing import (
|
|||
Dict,
|
||||
Iterable,
|
||||
List,
|
||||
Mapping,
|
||||
MutableMapping,
|
||||
Optional,
|
||||
Set,
|
||||
|
@ -1633,7 +1634,7 @@ class EventsWorkerStore(SQLBaseStore):
|
|||
self,
|
||||
room_id: str,
|
||||
event_ids: Collection[str],
|
||||
) -> Dict[str, bool]:
|
||||
) -> Mapping[str, bool]:
|
||||
"""Helper for have_seen_events
|
||||
|
||||
Returns:
|
||||
|
@ -2325,7 +2326,7 @@ class EventsWorkerStore(SQLBaseStore):
|
|||
@cachedList(cached_method_name="is_partial_state_event", list_name="event_ids")
|
||||
async def get_partial_state_events(
|
||||
self, event_ids: Collection[str]
|
||||
) -> Dict[str, bool]:
|
||||
) -> Mapping[str, bool]:
|
||||
"""Checks which of the given events have partial state
|
||||
|
||||
Args:
|
||||
|
|
|
@ -16,7 +16,7 @@
|
|||
import itertools
|
||||
import json
|
||||
import logging
|
||||
from typing import Dict, Iterable, Optional, Tuple
|
||||
from typing import Dict, Iterable, Mapping, Optional, Tuple
|
||||
|
||||
from canonicaljson import encode_canonical_json
|
||||
from signedjson.key import decode_verify_key_bytes
|
||||
|
@ -130,7 +130,7 @@ class KeyStore(CacheInvalidationWorkerStore):
|
|||
)
|
||||
async def get_server_keys_json(
|
||||
self, server_name_and_key_ids: Iterable[Tuple[str, str]]
|
||||
) -> Dict[Tuple[str, str], FetchKeyResult]:
|
||||
) -> Mapping[Tuple[str, str], FetchKeyResult]:
|
||||
"""
|
||||
Args:
|
||||
server_name_and_key_ids:
|
||||
|
@ -200,7 +200,7 @@ class KeyStore(CacheInvalidationWorkerStore):
|
|||
)
|
||||
async def get_server_keys_json_for_remote(
|
||||
self, server_name: str, key_ids: Iterable[str]
|
||||
) -> Dict[str, Optional[FetchKeyResultForRemote]]:
|
||||
) -> Mapping[str, Optional[FetchKeyResultForRemote]]:
|
||||
"""Fetch the cached keys for the given server/key IDs.
|
||||
|
||||
If we have multiple entries for a given key ID, returns the most recent.
|
||||
|
|
|
@ -11,7 +11,17 @@
|
|||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
from typing import TYPE_CHECKING, Any, Dict, Iterable, List, Optional, Tuple, cast
|
||||
from typing import (
|
||||
TYPE_CHECKING,
|
||||
Any,
|
||||
Dict,
|
||||
Iterable,
|
||||
List,
|
||||
Mapping,
|
||||
Optional,
|
||||
Tuple,
|
||||
cast,
|
||||
)
|
||||
|
||||
from synapse.api.presence import PresenceState, UserPresenceState
|
||||
from synapse.replication.tcp.streams import PresenceStream
|
||||
|
@ -249,7 +259,7 @@ class PresenceStore(PresenceBackgroundUpdateStore, CacheInvalidationWorkerStore)
|
|||
)
|
||||
async def get_presence_for_users(
|
||||
self, user_ids: Iterable[str]
|
||||
) -> Dict[str, UserPresenceState]:
|
||||
) -> Mapping[str, UserPresenceState]:
|
||||
rows = await self.db_pool.simple_select_many_batch(
|
||||
table="presence_stream",
|
||||
column="user_id",
|
||||
|
|
|
@ -216,7 +216,7 @@ class PushRulesWorkerStore(
|
|||
@cachedList(cached_method_name="get_push_rules_for_user", list_name="user_ids")
|
||||
async def bulk_get_push_rules(
|
||||
self, user_ids: Collection[str]
|
||||
) -> Dict[str, FilteredPushRules]:
|
||||
) -> Mapping[str, FilteredPushRules]:
|
||||
if not user_ids:
|
||||
return {}
|
||||
|
||||
|
|
|
@ -43,7 +43,7 @@ from synapse.storage.util.id_generators import (
|
|||
MultiWriterIdGenerator,
|
||||
StreamIdGenerator,
|
||||
)
|
||||
from synapse.types import JsonDict
|
||||
from synapse.types import JsonDict, JsonMapping
|
||||
from synapse.util import json_encoder
|
||||
from synapse.util.caches.descriptors import cached, cachedList
|
||||
from synapse.util.caches.stream_change_cache import StreamChangeCache
|
||||
|
@ -218,7 +218,7 @@ class ReceiptsWorkerStore(SQLBaseStore):
|
|||
@cached()
|
||||
async def _get_receipts_for_user_with_orderings(
|
||||
self, user_id: str, receipt_type: str
|
||||
) -> JsonDict:
|
||||
) -> JsonMapping:
|
||||
"""
|
||||
Fetch receipts for all rooms that the given user is joined to.
|
||||
|
||||
|
@ -258,7 +258,7 @@ class ReceiptsWorkerStore(SQLBaseStore):
|
|||
|
||||
async def get_linearized_receipts_for_rooms(
|
||||
self, room_ids: Iterable[str], to_key: int, from_key: Optional[int] = None
|
||||
) -> List[dict]:
|
||||
) -> List[JsonMapping]:
|
||||
"""Get receipts for multiple rooms for sending to clients.
|
||||
|
||||
Args:
|
||||
|
@ -287,7 +287,7 @@ class ReceiptsWorkerStore(SQLBaseStore):
|
|||
|
||||
async def get_linearized_receipts_for_room(
|
||||
self, room_id: str, to_key: int, from_key: Optional[int] = None
|
||||
) -> Sequence[JsonDict]:
|
||||
) -> Sequence[JsonMapping]:
|
||||
"""Get receipts for a single room for sending to clients.
|
||||
|
||||
Args:
|
||||
|
@ -310,7 +310,7 @@ class ReceiptsWorkerStore(SQLBaseStore):
|
|||
@cached(tree=True)
|
||||
async def _get_linearized_receipts_for_room(
|
||||
self, room_id: str, to_key: int, from_key: Optional[int] = None
|
||||
) -> Sequence[JsonDict]:
|
||||
) -> Sequence[JsonMapping]:
|
||||
"""See get_linearized_receipts_for_room"""
|
||||
|
||||
def f(txn: LoggingTransaction) -> List[Dict[str, Any]]:
|
||||
|
@ -353,7 +353,7 @@ class ReceiptsWorkerStore(SQLBaseStore):
|
|||
)
|
||||
async def _get_linearized_receipts_for_rooms(
|
||||
self, room_ids: Collection[str], to_key: int, from_key: Optional[int] = None
|
||||
) -> Dict[str, Sequence[JsonDict]]:
|
||||
) -> Mapping[str, Sequence[JsonMapping]]:
|
||||
if not room_ids:
|
||||
return {}
|
||||
|
||||
|
@ -415,7 +415,7 @@ class ReceiptsWorkerStore(SQLBaseStore):
|
|||
)
|
||||
async def get_linearized_receipts_for_all_rooms(
|
||||
self, to_key: int, from_key: Optional[int] = None
|
||||
) -> Mapping[str, JsonDict]:
|
||||
) -> Mapping[str, JsonMapping]:
|
||||
"""Get receipts for all rooms between two stream_ids, up
|
||||
to a limit of the latest 100 read receipts.
|
||||
|
||||
|
|
|
@ -519,7 +519,7 @@ class RelationsWorkerStore(SQLBaseStore):
|
|||
@cachedList(cached_method_name="get_applicable_edit", list_name="event_ids")
|
||||
async def get_applicable_edits(
|
||||
self, event_ids: Collection[str]
|
||||
) -> Dict[str, Optional[EventBase]]:
|
||||
) -> Mapping[str, Optional[EventBase]]:
|
||||
"""Get the most recent edit (if any) that has happened for the given
|
||||
events.
|
||||
|
||||
|
@ -605,7 +605,7 @@ class RelationsWorkerStore(SQLBaseStore):
|
|||
@cachedList(cached_method_name="get_thread_summary", list_name="event_ids")
|
||||
async def get_thread_summaries(
|
||||
self, event_ids: Collection[str]
|
||||
) -> Dict[str, Optional[Tuple[int, EventBase]]]:
|
||||
) -> Mapping[str, Optional[Tuple[int, EventBase]]]:
|
||||
"""Get the number of threaded replies and the latest reply (if any) for the given events.
|
||||
|
||||
Args:
|
||||
|
@ -779,7 +779,7 @@ class RelationsWorkerStore(SQLBaseStore):
|
|||
@cachedList(cached_method_name="get_thread_participated", list_name="event_ids")
|
||||
async def get_threads_participated(
|
||||
self, event_ids: Collection[str], user_id: str
|
||||
) -> Dict[str, bool]:
|
||||
) -> Mapping[str, bool]:
|
||||
"""Get whether the requesting user participated in the given threads.
|
||||
|
||||
This is separate from get_thread_summaries since that can be cached across
|
||||
|
|
|
@ -191,7 +191,7 @@ class RoomMemberWorkerStore(EventsWorkerStore, CacheInvalidationWorkerStore):
|
|||
)
|
||||
async def get_subset_users_in_room_with_profiles(
|
||||
self, room_id: str, user_ids: Collection[str]
|
||||
) -> Dict[str, ProfileInfo]:
|
||||
) -> Mapping[str, ProfileInfo]:
|
||||
"""Get a mapping from user ID to profile information for a list of users
|
||||
in a given room.
|
||||
|
||||
|
@ -676,7 +676,7 @@ class RoomMemberWorkerStore(EventsWorkerStore, CacheInvalidationWorkerStore):
|
|||
)
|
||||
async def _get_rooms_for_users(
|
||||
self, user_ids: Collection[str]
|
||||
) -> Dict[str, FrozenSet[str]]:
|
||||
) -> Mapping[str, FrozenSet[str]]:
|
||||
"""A batched version of `get_rooms_for_user`.
|
||||
|
||||
Returns:
|
||||
|
@ -881,7 +881,7 @@ class RoomMemberWorkerStore(EventsWorkerStore, CacheInvalidationWorkerStore):
|
|||
)
|
||||
async def _get_user_ids_from_membership_event_ids(
|
||||
self, event_ids: Iterable[str]
|
||||
) -> Dict[str, Optional[str]]:
|
||||
) -> Mapping[str, Optional[str]]:
|
||||
"""For given set of member event_ids check if they point to a join
|
||||
event.
|
||||
|
||||
|
@ -1191,7 +1191,7 @@ class RoomMemberWorkerStore(EventsWorkerStore, CacheInvalidationWorkerStore):
|
|||
)
|
||||
async def get_membership_from_event_ids(
|
||||
self, member_event_ids: Iterable[str]
|
||||
) -> Dict[str, Optional[EventIdMembership]]:
|
||||
) -> Mapping[str, Optional[EventIdMembership]]:
|
||||
"""Get user_id and membership of a set of event IDs.
|
||||
|
||||
Returns:
|
||||
|
|
|
@ -14,7 +14,17 @@
|
|||
# limitations under the License.
|
||||
import collections.abc
|
||||
import logging
|
||||
from typing import TYPE_CHECKING, Any, Collection, Dict, Iterable, Optional, Set, Tuple
|
||||
from typing import (
|
||||
TYPE_CHECKING,
|
||||
Any,
|
||||
Collection,
|
||||
Dict,
|
||||
Iterable,
|
||||
Mapping,
|
||||
Optional,
|
||||
Set,
|
||||
Tuple,
|
||||
)
|
||||
|
||||
import attr
|
||||
|
||||
|
@ -372,7 +382,7 @@ class StateGroupWorkerStore(EventsWorkerStore, SQLBaseStore):
|
|||
)
|
||||
async def _get_state_group_for_events(
|
||||
self, event_ids: Collection[str]
|
||||
) -> Dict[str, int]:
|
||||
) -> Mapping[str, int]:
|
||||
"""Returns mapping event_id -> state_group.
|
||||
|
||||
Raises:
|
||||
|
|
|
@ -14,7 +14,7 @@
|
|||
|
||||
import logging
|
||||
from enum import Enum
|
||||
from typing import TYPE_CHECKING, Dict, Iterable, List, Optional, Tuple, cast
|
||||
from typing import TYPE_CHECKING, Iterable, List, Mapping, Optional, Tuple, cast
|
||||
|
||||
import attr
|
||||
from canonicaljson import encode_canonical_json
|
||||
|
@ -210,7 +210,7 @@ class TransactionWorkerStore(CacheInvalidationWorkerStore):
|
|||
)
|
||||
async def get_destination_retry_timings_batch(
|
||||
self, destinations: StrCollection
|
||||
) -> Dict[str, Optional[DestinationRetryTimings]]:
|
||||
) -> Mapping[str, Optional[DestinationRetryTimings]]:
|
||||
rows = await self.db_pool.simple_select_many_batch(
|
||||
table="destinations",
|
||||
iterable=destinations,
|
||||
|
|
|
@ -12,7 +12,7 @@
|
|||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
from typing import Dict, Iterable
|
||||
from typing import Iterable, Mapping
|
||||
|
||||
from synapse.storage.database import LoggingTransaction
|
||||
from synapse.storage.databases.main import CacheInvalidationWorkerStore
|
||||
|
@ -40,7 +40,7 @@ class UserErasureWorkerStore(CacheInvalidationWorkerStore):
|
|||
return bool(result)
|
||||
|
||||
@cachedList(cached_method_name="is_user_erased", list_name="user_ids")
|
||||
async def are_users_erased(self, user_ids: Iterable[str]) -> Dict[str, bool]:
|
||||
async def are_users_erased(self, user_ids: Iterable[str]) -> Mapping[str, bool]:
|
||||
"""
|
||||
Checks which users in a list have requested erasure
|
||||
|
||||
|
|
Loading…
Reference in New Issue