Remove support for aggregate room lists
This commit is contained in:
parent
4f181f361d
commit
4131381123
|
@ -29,7 +29,6 @@ class ServerConfig(Config):
|
||||||
self.user_agent_suffix = config.get("user_agent_suffix")
|
self.user_agent_suffix = config.get("user_agent_suffix")
|
||||||
self.use_frozen_dicts = config.get("use_frozen_dicts", False)
|
self.use_frozen_dicts = config.get("use_frozen_dicts", False)
|
||||||
self.public_baseurl = config.get("public_baseurl")
|
self.public_baseurl = config.get("public_baseurl")
|
||||||
self.secondary_directory_servers = config.get("secondary_directory_servers", [])
|
|
||||||
|
|
||||||
if self.public_baseurl is not None:
|
if self.public_baseurl is not None:
|
||||||
if self.public_baseurl[-1] != '/':
|
if self.public_baseurl[-1] != '/':
|
||||||
|
@ -142,14 +141,6 @@ class ServerConfig(Config):
|
||||||
# The GC threshold parameters to pass to `gc.set_threshold`, if defined
|
# The GC threshold parameters to pass to `gc.set_threshold`, if defined
|
||||||
# gc_thresholds: [700, 10, 10]
|
# gc_thresholds: [700, 10, 10]
|
||||||
|
|
||||||
# A list of other Home Servers to fetch the public room directory from
|
|
||||||
# and include in the public room directory of this home server
|
|
||||||
# This is a temporary stopgap solution to populate new server with a
|
|
||||||
# list of rooms until there exists a good solution of a decentralized
|
|
||||||
# room directory.
|
|
||||||
# secondary_directory_servers:
|
|
||||||
# - matrix.org
|
|
||||||
|
|
||||||
# List of ports that Synapse should listen on, their purpose and their
|
# List of ports that Synapse should listen on, their purpose and their
|
||||||
# configuration.
|
# configuration.
|
||||||
listeners:
|
listeners:
|
||||||
|
|
|
@ -39,12 +39,6 @@ class RoomListHandler(BaseHandler):
|
||||||
def __init__(self, hs):
|
def __init__(self, hs):
|
||||||
super(RoomListHandler, self).__init__(hs)
|
super(RoomListHandler, self).__init__(hs)
|
||||||
self.response_cache = ResponseCache(hs)
|
self.response_cache = ResponseCache(hs)
|
||||||
self.remote_list_request_cache = ResponseCache(hs)
|
|
||||||
self.remote_list_cache = {}
|
|
||||||
self.fetch_looping_call = hs.get_clock().looping_call(
|
|
||||||
self.fetch_all_remote_lists, REMOTE_ROOM_LIST_POLL_INTERVAL
|
|
||||||
)
|
|
||||||
self.fetch_all_remote_lists()
|
|
||||||
|
|
||||||
def get_local_public_room_list(self, limit=None, next_batch=None):
|
def get_local_public_room_list(self, limit=None, next_batch=None):
|
||||||
result = self.response_cache.get((limit, next_batch))
|
result = self.response_cache.get((limit, next_batch))
|
||||||
|
@ -219,14 +213,6 @@ class RoomListHandler(BaseHandler):
|
||||||
"chunk": results,
|
"chunk": results,
|
||||||
})
|
})
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
|
||||||
def fetch_all_remote_lists(self):
|
|
||||||
deferred = self.hs.get_replication_layer().get_public_rooms(
|
|
||||||
self.hs.config.secondary_directory_servers
|
|
||||||
)
|
|
||||||
self.remote_list_request_cache.set((), deferred)
|
|
||||||
self.remote_list_cache = yield deferred
|
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def get_remote_public_room_list(self, server_name):
|
def get_remote_public_room_list(self, server_name):
|
||||||
res = yield self.hs.get_replication_layer().get_public_rooms(
|
res = yield self.hs.get_replication_layer().get_public_rooms(
|
||||||
|
@ -237,47 +223,6 @@ class RoomListHandler(BaseHandler):
|
||||||
raise SynapseError(404, "Server not found")
|
raise SynapseError(404, "Server not found")
|
||||||
defer.returnValue(res[server_name])
|
defer.returnValue(res[server_name])
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
|
||||||
def get_aggregated_public_room_list(self):
|
|
||||||
"""
|
|
||||||
Get the public room list from this server and the servers
|
|
||||||
specified in the secondary_directory_servers config option.
|
|
||||||
XXX: Pagination...
|
|
||||||
"""
|
|
||||||
# We return the results from out cache which is updated by a looping call,
|
|
||||||
# unless we're missing a cache entry, in which case wait for the result
|
|
||||||
# of the fetch if there's one in progress. If not, omit that server.
|
|
||||||
wait = False
|
|
||||||
for s in self.hs.config.secondary_directory_servers:
|
|
||||||
if s not in self.remote_list_cache:
|
|
||||||
logger.warn("No cached room list from %s: waiting for fetch", s)
|
|
||||||
wait = True
|
|
||||||
break
|
|
||||||
|
|
||||||
if wait and self.remote_list_request_cache.get(()):
|
|
||||||
yield self.remote_list_request_cache.get(())
|
|
||||||
|
|
||||||
public_rooms = yield self.get_local_public_room_list()
|
|
||||||
|
|
||||||
# keep track of which room IDs we've seen so we can de-dup
|
|
||||||
room_ids = set()
|
|
||||||
|
|
||||||
# tag all the ones in our list with our server name.
|
|
||||||
# Also add the them to the de-deping set
|
|
||||||
for room in public_rooms['chunk']:
|
|
||||||
room["server_name"] = self.hs.hostname
|
|
||||||
room_ids.add(room["room_id"])
|
|
||||||
|
|
||||||
# Now add the results from federation
|
|
||||||
for server_name, server_result in self.remote_list_cache.items():
|
|
||||||
for room in server_result["chunk"]:
|
|
||||||
if room["room_id"] not in room_ids:
|
|
||||||
room["server_name"] = server_name
|
|
||||||
public_rooms["chunk"].append(room)
|
|
||||||
room_ids.add(room["room_id"])
|
|
||||||
|
|
||||||
defer.returnValue(public_rooms)
|
|
||||||
|
|
||||||
|
|
||||||
class RoomListNextBatch(namedtuple("RoomListNextBatch", (
|
class RoomListNextBatch(namedtuple("RoomListNextBatch", (
|
||||||
"stream_ordering", # stream_ordering of the first public room list
|
"stream_ordering", # stream_ordering of the first public room list
|
||||||
|
|
|
@ -321,7 +321,7 @@ class PublicRoomListRestServlet(ClientV1RestServlet):
|
||||||
if server:
|
if server:
|
||||||
data = yield handler.get_remote_public_room_list(server)
|
data = yield handler.get_remote_public_room_list(server)
|
||||||
else:
|
else:
|
||||||
data = yield handler.get_aggregated_public_room_list()
|
data = yield handler.get_local_public_room_list()
|
||||||
|
|
||||||
defer.returnValue((200, data))
|
defer.returnValue((200, data))
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue