Start implementing the non-incremental sync portion of the v2 /sync API
parent
3186c5bdbc
commit
436513068d
|
@ -89,7 +89,7 @@ def prune_event(event):
|
|||
return type(event)(allowed_fields)
|
||||
|
||||
|
||||
def serialize_event(e, time_now_ms, client_event=True):
|
||||
def serialize_event(e, time_now_ms, client_event=True, strip_ids=False):
|
||||
# FIXME(erikj): To handle the case of presence events and the like
|
||||
if not isinstance(e, EventBase):
|
||||
return e
|
||||
|
@ -138,4 +138,8 @@ def serialize_event(e, time_now_ms, client_event=True):
|
|||
d.pop("unsigned", None)
|
||||
d.pop("origin", None)
|
||||
|
||||
if strip_ids:
|
||||
d.pop("room_id", None)
|
||||
d.pop("event_id", None)
|
||||
|
||||
return d
|
||||
|
|
|
@ -26,6 +26,7 @@ from .presence import PresenceHandler
|
|||
from .directory import DirectoryHandler
|
||||
from .typing import TypingNotificationHandler
|
||||
from .admin import AdminHandler
|
||||
from .sync import SyncHandler
|
||||
|
||||
|
||||
class Handlers(object):
|
||||
|
@ -51,3 +52,4 @@ class Handlers(object):
|
|||
self.directory_handler = DirectoryHandler(hs)
|
||||
self.typing_notification_handler = TypingNotificationHandler(hs)
|
||||
self.admin_handler = AdminHandler(hs)
|
||||
self.sync_handler = SyncHandler(hs)
|
||||
|
|
|
@ -1,26 +1,49 @@
|
|||
# -*- coding: utf-8 -*-
|
||||
# Copyright 2015 OpenMarket Ltd
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
from ._base import BaseHandler
|
||||
|
||||
from synapse.streams.config import PaginationConfig
|
||||
from synapse.api.constants import Membership
|
||||
|
||||
from twisted.internet import defer
|
||||
|
||||
import collections
|
||||
import logging
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
SyncConfig = collections.namedtuple("SyncConfig", [
|
||||
"user",
|
||||
"device",
|
||||
"since",
|
||||
"limit",
|
||||
"gap",
|
||||
"sort"
|
||||
"backfill"
|
||||
"sort",
|
||||
"backfill",
|
||||
"filter",
|
||||
)
|
||||
])
|
||||
|
||||
|
||||
RoomSyncResult = collections.namedtuple("RoomSyncResult", [
|
||||
"room_id",
|
||||
"limited",
|
||||
"published",
|
||||
"prev_batch",
|
||||
"events",
|
||||
"events", # dict of event
|
||||
"state",
|
||||
"event_map",
|
||||
"prev_batch",
|
||||
])
|
||||
|
||||
|
||||
|
@ -41,10 +64,11 @@ class SyncHandler(BaseHandler):
|
|||
def __init__(self, hs):
|
||||
super(SyncHandler, self).__init__(hs)
|
||||
self.event_sources = hs.get_event_sources()
|
||||
self.clock = hs.get_clock()
|
||||
|
||||
def wait_for_sync_for_user(self, sync_config, since_token=None, timeout=0):
|
||||
if timeout == 0:
|
||||
return self.current_sync_for_user(sync_config, since)
|
||||
return self.current_sync_for_user(sync_config, since_token)
|
||||
else:
|
||||
def current_sync_callback(since_token):
|
||||
return self.current_sync_for_user(
|
||||
|
@ -53,58 +77,71 @@ class SyncHandler(BaseHandler):
|
|||
return self.notifier.wait_for_events(
|
||||
sync_config.filter, since_token, current_sync_callback
|
||||
)
|
||||
defer.returnValue(result)
|
||||
|
||||
def current_sync_for_user(self, sync_config, since_token=None):
|
||||
if since_token is None:
|
||||
return self.inital_sync(sync_config)
|
||||
return self.initial_sync(sync_config)
|
||||
else:
|
||||
return self.incremental_sync(sync_config)
|
||||
|
||||
@defer.inlineCallbacks
|
||||
def initial_sync(self, sync_config):
|
||||
if sync_config.sort == "timeline,desc":
|
||||
# TODO(mjark): Handle going through events in reverse order?.
|
||||
# What does "most recent events" mean when applying the limits mean
|
||||
# in this case?
|
||||
raise NotImplementedError()
|
||||
|
||||
now_token = yield self.event_sources.get_current_token()
|
||||
|
||||
presence_stream = self.event_sources.sources["presence"]
|
||||
# TODO (markjh): This looks wrong, shouldn't we be getting the presence
|
||||
# TODO (mjark): This looks wrong, shouldn't we be getting the presence
|
||||
# UP to the present rather than after the present?
|
||||
pagination_config = PaginationConfig(from_token=now_token)
|
||||
presence, _ = yield presence_stream.get_pagination_rows(
|
||||
user, pagination_config.get_source_config("presence"), None
|
||||
user=sync_config.user,
|
||||
pagination_config=pagination_config.get_source_config("presence"),
|
||||
key=None
|
||||
)
|
||||
room_list = yield self.store.get_rooms_for_user_where_membership_is(
|
||||
user_id=user_id,
|
||||
user_id=sync_config.user.to_string(),
|
||||
membership_list=[Membership.INVITE, Membership.JOIN]
|
||||
)
|
||||
|
||||
# TODO (markjh): Does public mean "published"?
|
||||
# TODO (mjark): Does public mean "published"?
|
||||
published_rooms = yield self.store.get_rooms(is_public=True)
|
||||
published_room_ids = set(r["room_id"] for r in public_rooms)
|
||||
published_room_ids = set(r["room_id"] for r in published_rooms)
|
||||
|
||||
rooms = []
|
||||
for event in room_list:
|
||||
|
||||
messages, token = yield self.store.get_recent_events_for_room(
|
||||
#TODO (mjark): Apply the event filter in sync_config.
|
||||
recent_events, token = yield self.store.get_recent_events_for_room(
|
||||
event.room_id,
|
||||
limit=sync_config.limit,
|
||||
end_token=now_token.room_key,
|
||||
)
|
||||
prev_batch_token = now_token.copy_and_replace("room_key", token[0])
|
||||
current_state = yield self.state_handler.get_current_state(
|
||||
current_state_events = yield self.state_handler.get_current_state(
|
||||
event.room_id
|
||||
)
|
||||
|
||||
rooms.append(RoomSyncResult(
|
||||
room_id=event.room_id,
|
||||
published=event.room_id in published_room_ids,
|
||||
events=recent_events,
|
||||
prev_batch=prev_batch_token,
|
||||
state=current_state_events,
|
||||
limited=True,
|
||||
))
|
||||
|
||||
|
||||
|
||||
defer.returnValue(SyncResult(
|
||||
public_user_data=presence,
|
||||
private_user_data=[],
|
||||
rooms=rooms,
|
||||
next_batch=now_token,
|
||||
))
|
||||
|
||||
|
||||
@defer.inlineCallbacks
|
||||
def incremental_sync(self, sync_config):
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
pass
|
||||
|
|
|
@ -1,5 +1,5 @@
|
|||
# -*- coding: utf-8 -*-
|
||||
# Copyright 2014, 2015 OpenMarket Ltd
|
||||
# Copyright 2015 OpenMarket Ltd
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
|
@ -16,6 +16,9 @@
|
|||
from twisted.internet import defer
|
||||
|
||||
from synapse.http.servlet import RestServlet
|
||||
from synapse.handlers.sync import SyncConfig
|
||||
from synapse.types import StreamToken
|
||||
from synapse.events.utils import serialize_event
|
||||
from ._base import client_v2_pattern
|
||||
|
||||
import logging
|
||||
|
@ -73,14 +76,15 @@ class SyncRestServlet(RestServlet):
|
|||
def __init__(self, hs):
|
||||
super(SyncRestServlet, self).__init__()
|
||||
self.auth = hs.get_auth()
|
||||
#self.sync_handler = hs.get_handlers().sync_hanlder
|
||||
self.sync_handler = hs.get_handlers().sync_handler
|
||||
self.clock = hs.get_clock()
|
||||
|
||||
@defer.inlineCallbacks
|
||||
def on_GET(self, request):
|
||||
user = yield self.auth.get_user_by_req(request)
|
||||
|
||||
timeout = self.parse_integer(request, "timeout", default=0)
|
||||
limit = self.parse_integer(request, "limit", default=None)
|
||||
limit = self.parse_integer(request, "limit", required=True)
|
||||
gap = self.parse_boolean(request, "gap", default=True)
|
||||
sort = self.parse_string(
|
||||
request, "sort", default="timeline,asc",
|
||||
|
@ -91,7 +95,7 @@ class SyncRestServlet(RestServlet):
|
|||
request, "set_presence", default="online",
|
||||
allowed_values=self.ALLOWED_PRESENCE
|
||||
)
|
||||
backfill = self.parse_boolean(request, "backfill", default=True)
|
||||
backfill = self.parse_boolean(request, "backfill", default=False)
|
||||
filter_id = self.parse_string(request, "filter", default=None)
|
||||
|
||||
logger.info(
|
||||
|
@ -108,36 +112,81 @@ class SyncRestServlet(RestServlet):
|
|||
# if filter.matches(event):
|
||||
# # stuff
|
||||
|
||||
# if timeout != 0:
|
||||
# register for updates from the event stream
|
||||
sync_config = SyncConfig(
|
||||
user=user,
|
||||
device="TODO", # TODO(mjark) Get the device_id from access_token
|
||||
gap=gap,
|
||||
limit=limit,
|
||||
sort=sort,
|
||||
backfill=backfill,
|
||||
filter="TODO", # TODO(mjark) Add the filter to the config.
|
||||
)
|
||||
|
||||
#rooms = []
|
||||
|
||||
if gap:
|
||||
pass
|
||||
# now_stream_token = get_current_stream_token
|
||||
# for room_id in get_rooms_for_user(user, filter=filter):
|
||||
# state, events, start, end, limited, published = updates_for_room(
|
||||
# from=since, to=now_stream_token, limit=limit,
|
||||
# anchor_to_start=False
|
||||
# )
|
||||
# rooms[room_id] = (state, events, start, limited, published)
|
||||
# next_stream_token = now.
|
||||
if since is not None:
|
||||
since_token = StreamToken.from_string(since)
|
||||
else:
|
||||
pass
|
||||
# now_stream_token = get_current_stream_token
|
||||
# for room_id in get_rooms_for_user(user, filter=filter)
|
||||
# state, events, start, end, limited, published = updates_for_room(
|
||||
# from=since, to=now_stream_token, limit=limit,
|
||||
# anchor_to_start=False
|
||||
# )
|
||||
# next_stream_token = min(next_stream_token, end)
|
||||
since_token = None
|
||||
|
||||
sync_result = yield self.sync_handler.wait_for_sync_for_user(
|
||||
sync_config, since_token=since_token, timeout=timeout
|
||||
)
|
||||
|
||||
response_content = {}
|
||||
time_now = self.clock.time_msec()
|
||||
|
||||
response_content = {
|
||||
"public_user_data": self.encode_events(
|
||||
sync_result.public_user_data, filter, time_now
|
||||
),
|
||||
"private_user_data": self.encode_events(
|
||||
sync_result.private_user_data, filter, time_now
|
||||
),
|
||||
"rooms": self.encode_rooms(sync_result.rooms, filter, time_now),
|
||||
"next_batch": sync_result.next_batch.to_string(),
|
||||
}
|
||||
|
||||
defer.returnValue((200, response_content))
|
||||
|
||||
def encode_events(self, events, filter, time_now):
|
||||
return [self.encode_event(event, filter, time_now) for event in events]
|
||||
|
||||
@staticmethod
|
||||
def encode_event(event, filter, time_now):
|
||||
# TODO(mjark): Respect formatting requirements in the filter.
|
||||
return serialize_event(event, time_now)
|
||||
|
||||
def encode_rooms(self, rooms, filter, time_now):
|
||||
return [self.encode_room(room, filter, time_now) for room in rooms]
|
||||
|
||||
@staticmethod
|
||||
def encode_room(room, filter, time_now):
|
||||
event_map = {}
|
||||
state_event_ids = []
|
||||
recent_event_ids = []
|
||||
for event in room.state:
|
||||
# TODO(mjark): Respect formatting requirements in the filter.
|
||||
event_map[event.event_id] = serialize_event(
|
||||
event, time_now, strip_ids=True
|
||||
)
|
||||
state_event_ids.append(event.event_id)
|
||||
|
||||
for event in room.events:
|
||||
# TODO(mjark): Respect formatting requirements in the filter.
|
||||
event_map[event.event_id] = serialize_event(
|
||||
event, time_now, strip_ids=True
|
||||
)
|
||||
recent_event_ids.append(event.event_id)
|
||||
return {
|
||||
"room_id": room.room_id,
|
||||
"event_map": event_map,
|
||||
"events": {
|
||||
"batch": recent_event_ids,
|
||||
"prev_batch": room.prev_batch.to_string(),
|
||||
},
|
||||
"state": state_event_ids,
|
||||
"limited": room.limited,
|
||||
"published": room.published,
|
||||
}
|
||||
|
||||
|
||||
def register_servlets(hs, http_server):
|
||||
SyncRestServlet(hs).register(http_server)
|
||||
|
|
Loading…
Reference in New Issue