2014-11-19 19:20:59 +01:00
|
|
|
# -*- coding: utf-8 -*-
|
2016-01-07 05:26:29 +01:00
|
|
|
# Copyright 2014-2016 OpenMarket Ltd
|
2014-11-19 19:20:59 +01:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
2015-09-22 19:14:15 +02:00
|
|
|
from ._base import SQLBaseStore
|
2014-11-19 19:20:59 +01:00
|
|
|
from twisted.internet import defer
|
|
|
|
|
2015-08-24 17:17:38 +02:00
|
|
|
from canonicaljson import encode_canonical_json
|
2015-04-16 12:01:09 +02:00
|
|
|
|
2016-06-01 12:08:45 +02:00
|
|
|
from synapse.util.caches.descriptors import cachedInlineCallbacks, cachedList
|
2016-04-06 16:42:15 +02:00
|
|
|
|
2014-11-19 19:20:59 +01:00
|
|
|
import logging
|
2015-04-29 18:13:51 +02:00
|
|
|
import simplejson as json
|
2015-04-29 20:41:14 +02:00
|
|
|
import types
|
2014-11-19 19:20:59 +01:00
|
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2014-12-03 14:37:02 +01:00
|
|
|
|
2014-11-19 19:20:59 +01:00
|
|
|
class PusherStore(SQLBaseStore):
|
2015-04-29 20:41:14 +02:00
|
|
|
def _decode_pushers_rows(self, rows):
|
|
|
|
for r in rows:
|
|
|
|
dataJson = r['data']
|
|
|
|
r['data'] = None
|
|
|
|
try:
|
|
|
|
if isinstance(dataJson, types.BufferType):
|
|
|
|
dataJson = str(dataJson).decode("UTF8")
|
|
|
|
|
|
|
|
r['data'] = json.loads(dataJson)
|
|
|
|
except Exception as e:
|
|
|
|
logger.warn(
|
|
|
|
"Invalid JSON in data for pusher %d: %s, %s",
|
|
|
|
r['id'], dataJson, e.message,
|
|
|
|
)
|
|
|
|
pass
|
|
|
|
|
|
|
|
if isinstance(r['pushkey'], types.BufferType):
|
|
|
|
r['pushkey'] = str(r['pushkey']).decode("UTF8")
|
|
|
|
|
|
|
|
return rows
|
|
|
|
|
2016-04-08 16:29:59 +02:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def user_has_pusher(self, user_id):
|
|
|
|
ret = yield self._simple_select_one_onecol(
|
|
|
|
"pushers", {"user_name": user_id}, "id", allow_none=True
|
|
|
|
)
|
|
|
|
defer.returnValue(ret is not None)
|
|
|
|
|
2015-03-25 20:06:22 +01:00
|
|
|
def get_pushers_by_app_id_and_pushkey(self, app_id, pushkey):
|
2016-04-12 14:47:17 +02:00
|
|
|
return self.get_pushers_by({
|
|
|
|
"app_id": app_id,
|
|
|
|
"pushkey": pushkey,
|
|
|
|
})
|
2014-12-18 15:49:22 +01:00
|
|
|
|
2016-04-12 14:35:08 +02:00
|
|
|
def get_pushers_by_user_id(self, user_id):
|
2016-04-12 14:47:17 +02:00
|
|
|
return self.get_pushers_by({
|
|
|
|
"user_name": user_id,
|
|
|
|
})
|
2016-04-11 19:00:03 +02:00
|
|
|
|
2016-04-12 14:47:17 +02:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_pushers_by(self, keyvalues):
|
|
|
|
ret = yield self._simple_select_list(
|
|
|
|
"pushers", keyvalues,
|
|
|
|
[
|
|
|
|
"id",
|
|
|
|
"user_name",
|
|
|
|
"access_token",
|
|
|
|
"profile_tag",
|
|
|
|
"kind",
|
|
|
|
"app_id",
|
|
|
|
"app_display_name",
|
|
|
|
"device_display_name",
|
|
|
|
"pushkey",
|
|
|
|
"ts",
|
|
|
|
"lang",
|
|
|
|
"data",
|
|
|
|
"last_stream_ordering",
|
|
|
|
"last_success",
|
|
|
|
"failing_since",
|
2016-04-12 14:55:32 +02:00
|
|
|
], desc="get_pushers_by"
|
2016-04-11 19:00:03 +02:00
|
|
|
)
|
2016-04-12 14:47:17 +02:00
|
|
|
defer.returnValue(self._decode_pushers_rows(ret))
|
2016-04-11 19:00:03 +02:00
|
|
|
|
2014-12-18 15:49:22 +01:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_all_pushers(self):
|
2015-04-29 19:03:42 +02:00
|
|
|
def get_pushers(txn):
|
|
|
|
txn.execute("SELECT * FROM pushers")
|
|
|
|
rows = self.cursor_to_dict(txn)
|
2014-12-18 15:49:22 +01:00
|
|
|
|
2015-04-29 20:41:14 +02:00
|
|
|
return self._decode_pushers_rows(rows)
|
2015-04-29 19:07:36 +02:00
|
|
|
|
2015-04-29 19:04:35 +02:00
|
|
|
rows = yield self.runInteraction("get_all_pushers", get_pushers)
|
2015-04-28 14:39:42 +02:00
|
|
|
defer.returnValue(rows)
|
2014-11-19 19:20:59 +01:00
|
|
|
|
2016-03-15 18:01:43 +01:00
|
|
|
def get_pushers_stream_token(self):
|
2016-04-01 14:29:05 +02:00
|
|
|
return self._pushers_id_gen.get_current_token()
|
2016-03-15 18:01:43 +01:00
|
|
|
|
|
|
|
def get_all_updated_pushers(self, last_id, current_id, limit):
|
2016-04-27 12:54:13 +02:00
|
|
|
if last_id == current_id:
|
|
|
|
return defer.succeed(([], []))
|
|
|
|
|
2016-03-15 18:01:43 +01:00
|
|
|
def get_all_updated_pushers_txn(txn):
|
|
|
|
sql = (
|
|
|
|
"SELECT id, user_name, access_token, profile_tag, kind,"
|
|
|
|
" app_id, app_display_name, device_display_name, pushkey, ts,"
|
|
|
|
" lang, data"
|
|
|
|
" FROM pushers"
|
|
|
|
" WHERE ? < id AND id <= ?"
|
|
|
|
" ORDER BY id ASC LIMIT ?"
|
|
|
|
)
|
|
|
|
txn.execute(sql, (last_id, current_id, limit))
|
|
|
|
updated = txn.fetchall()
|
|
|
|
|
|
|
|
sql = (
|
|
|
|
"SELECT stream_id, user_id, app_id, pushkey"
|
|
|
|
" FROM deleted_pushers"
|
|
|
|
" WHERE ? < stream_id AND stream_id <= ?"
|
|
|
|
" ORDER BY stream_id ASC LIMIT ?"
|
|
|
|
)
|
|
|
|
txn.execute(sql, (last_id, current_id, limit))
|
|
|
|
deleted = txn.fetchall()
|
|
|
|
|
|
|
|
return (updated, deleted)
|
|
|
|
return self.runInteraction(
|
|
|
|
"get_all_updated_pushers", get_all_updated_pushers_txn
|
|
|
|
)
|
|
|
|
|
2016-06-01 12:34:06 +02:00
|
|
|
@cachedInlineCallbacks(lru=True, num_args=1, max_entries=15000)
|
2016-06-01 12:08:45 +02:00
|
|
|
def get_if_user_has_pusher(self, user_id):
|
2016-04-06 16:42:15 +02:00
|
|
|
result = yield self._simple_select_many_batch(
|
2016-06-01 12:08:45 +02:00
|
|
|
table='pushers',
|
|
|
|
keyvalues={
|
|
|
|
'user_name': 'user_id',
|
|
|
|
},
|
|
|
|
retcol='user_name',
|
|
|
|
desc='get_if_user_has_pusher',
|
|
|
|
allow_none=True,
|
|
|
|
)
|
|
|
|
|
|
|
|
defer.returnValue(bool(result))
|
|
|
|
|
|
|
|
@cachedList(cached_method_name="get_if_user_has_pusher",
|
|
|
|
list_name="user_ids", num_args=1, inlineCallbacks=True)
|
|
|
|
def get_if_users_have_pushers(self, user_ids):
|
|
|
|
rows = yield self._simple_select_many_batch(
|
2016-04-14 12:36:23 +02:00
|
|
|
table='pushers',
|
|
|
|
column='user_name',
|
2016-06-01 12:08:45 +02:00
|
|
|
iterable=user_ids,
|
2016-04-14 12:36:23 +02:00
|
|
|
retcols=['user_name'],
|
2016-06-01 12:08:45 +02:00
|
|
|
desc='get_if_users_have_pushers'
|
2016-04-06 16:42:15 +02:00
|
|
|
)
|
|
|
|
|
2016-06-01 12:08:45 +02:00
|
|
|
result = {user_id: False for user_id in user_ids}
|
|
|
|
result.update({r['user_name']: True for r in rows})
|
|
|
|
|
|
|
|
defer.returnValue(result)
|
2016-04-06 16:42:15 +02:00
|
|
|
|
2014-11-19 19:20:59 +01:00
|
|
|
@defer.inlineCallbacks
|
2016-02-18 17:05:13 +01:00
|
|
|
def add_pusher(self, user_id, access_token, kind, app_id,
|
2015-01-13 20:48:37 +01:00
|
|
|
app_display_name, device_display_name,
|
2016-04-06 16:42:15 +02:00
|
|
|
pushkey, pushkey_ts, lang, data, last_stream_ordering,
|
|
|
|
profile_tag=""):
|
2016-04-07 18:38:48 +02:00
|
|
|
with self._pushers_id_gen.get_next() as stream_id:
|
|
|
|
def f(txn):
|
2016-05-13 12:25:02 +02:00
|
|
|
newly_inserted = self._simple_upsert_txn(
|
2016-04-06 16:42:15 +02:00
|
|
|
txn,
|
|
|
|
"pushers",
|
2016-04-07 18:45:01 +02:00
|
|
|
{
|
|
|
|
"app_id": app_id,
|
|
|
|
"pushkey": pushkey,
|
|
|
|
"user_name": user_id,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
"access_token": access_token,
|
|
|
|
"kind": kind,
|
|
|
|
"app_display_name": app_display_name,
|
|
|
|
"device_display_name": device_display_name,
|
|
|
|
"ts": pushkey_ts,
|
|
|
|
"lang": lang,
|
|
|
|
"data": encode_canonical_json(data),
|
|
|
|
"last_stream_ordering": last_stream_ordering,
|
|
|
|
"profile_tag": profile_tag,
|
|
|
|
"id": stream_id,
|
|
|
|
},
|
2016-04-06 16:42:15 +02:00
|
|
|
)
|
2016-05-13 12:25:02 +02:00
|
|
|
if newly_inserted:
|
2016-06-01 12:08:45 +02:00
|
|
|
# get_if_user_has_pusher only cares if the user has
|
2016-05-13 12:25:02 +02:00
|
|
|
# at least *one* pusher.
|
2016-06-01 12:08:45 +02:00
|
|
|
txn.call_after(self.get_if_user_has_pusher.invalidate, (user_id,))
|
2016-05-13 12:25:02 +02:00
|
|
|
|
|
|
|
yield self.runInteraction("add_pusher", f)
|
2014-11-19 19:20:59 +01:00
|
|
|
|
2015-01-13 20:48:37 +01:00
|
|
|
@defer.inlineCallbacks
|
2016-01-13 14:08:59 +01:00
|
|
|
def delete_pusher_by_app_id_pushkey_user_id(self, app_id, pushkey, user_id):
|
2016-03-15 18:01:43 +01:00
|
|
|
def delete_pusher_txn(txn, stream_id):
|
2016-06-01 12:08:45 +02:00
|
|
|
txn.call_after(self.get_if_user_has_pusher.invalidate, (user_id,))
|
2016-05-13 12:25:02 +02:00
|
|
|
|
2016-03-15 18:47:36 +01:00
|
|
|
self._simple_delete_one_txn(
|
2016-03-15 18:01:43 +01:00
|
|
|
txn,
|
|
|
|
"pushers",
|
|
|
|
{"app_id": app_id, "pushkey": pushkey, "user_name": user_id}
|
|
|
|
)
|
|
|
|
self._simple_upsert_txn(
|
|
|
|
txn,
|
|
|
|
"deleted_pushers",
|
|
|
|
{"app_id": app_id, "pushkey": pushkey, "user_id": user_id},
|
2016-03-15 18:47:36 +01:00
|
|
|
{"stream_id": stream_id},
|
2016-03-15 18:01:43 +01:00
|
|
|
)
|
2016-05-13 12:25:02 +02:00
|
|
|
|
2016-03-15 18:01:43 +01:00
|
|
|
with self._pushers_id_gen.get_next() as stream_id:
|
|
|
|
yield self.runInteraction(
|
|
|
|
"delete_pusher", delete_pusher_txn, stream_id
|
|
|
|
)
|
2015-01-13 20:48:37 +01:00
|
|
|
|
2014-11-19 19:20:59 +01:00
|
|
|
@defer.inlineCallbacks
|
2016-04-06 16:42:15 +02:00
|
|
|
def update_pusher_last_stream_ordering(self, app_id, pushkey, user_id,
|
|
|
|
last_stream_ordering):
|
2014-12-03 14:37:02 +01:00
|
|
|
yield self._simple_update_one(
|
2016-01-12 15:41:26 +01:00
|
|
|
"pushers",
|
2016-01-13 14:08:59 +01:00
|
|
|
{'app_id': app_id, 'pushkey': pushkey, 'user_name': user_id},
|
2016-04-06 16:42:15 +02:00
|
|
|
{'last_stream_ordering': last_stream_ordering},
|
|
|
|
desc="update_pusher_last_stream_ordering",
|
2014-11-19 19:20:59 +01:00
|
|
|
)
|
|
|
|
|
2014-11-21 13:21:00 +01:00
|
|
|
@defer.inlineCallbacks
|
2016-04-06 16:42:15 +02:00
|
|
|
def update_pusher_last_stream_ordering_and_success(self, app_id, pushkey,
|
|
|
|
user_id,
|
|
|
|
last_stream_ordering,
|
|
|
|
last_success):
|
2014-12-03 14:37:02 +01:00
|
|
|
yield self._simple_update_one(
|
2016-01-12 15:41:26 +01:00
|
|
|
"pushers",
|
2016-01-13 14:08:59 +01:00
|
|
|
{'app_id': app_id, 'pushkey': pushkey, 'user_name': user_id},
|
2016-04-06 16:42:15 +02:00
|
|
|
{
|
|
|
|
'last_stream_ordering': last_stream_ordering,
|
|
|
|
'last_success': last_success
|
|
|
|
},
|
|
|
|
desc="update_pusher_last_stream_ordering_and_success",
|
2014-11-21 13:21:00 +01:00
|
|
|
)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2016-01-13 14:08:59 +01:00
|
|
|
def update_pusher_failing_since(self, app_id, pushkey, user_id,
|
2015-03-25 20:06:22 +01:00
|
|
|
failing_since):
|
2014-12-03 14:37:02 +01:00
|
|
|
yield self._simple_update_one(
|
2016-01-12 15:41:26 +01:00
|
|
|
"pushers",
|
2016-01-13 14:08:59 +01:00
|
|
|
{'app_id': app_id, 'pushkey': pushkey, 'user_name': user_id},
|
2015-03-20 16:59:18 +01:00
|
|
|
{'failing_since': failing_since},
|
|
|
|
desc="update_pusher_failing_since",
|
2014-11-21 13:21:00 +01:00
|
|
|
)
|
2016-04-19 15:24:36 +02:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_throttle_params_by_room(self, pusher_id):
|
|
|
|
res = yield self._simple_select_list(
|
|
|
|
"pusher_throttle",
|
|
|
|
{"pusher": pusher_id},
|
|
|
|
["room_id", "last_sent_ts", "throttle_ms"],
|
|
|
|
desc="get_throttle_params_by_room"
|
|
|
|
)
|
|
|
|
|
|
|
|
params_by_room = {}
|
|
|
|
for row in res:
|
|
|
|
params_by_room[row["room_id"]] = {
|
|
|
|
"last_sent_ts": row["last_sent_ts"],
|
|
|
|
"throttle_ms": row["throttle_ms"]
|
|
|
|
}
|
|
|
|
|
|
|
|
defer.returnValue(params_by_room)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def set_throttle_params(self, pusher_id, room_id, params):
|
|
|
|
yield self._simple_upsert(
|
|
|
|
"pusher_throttle",
|
|
|
|
{"pusher": pusher_id, "room_id": room_id},
|
|
|
|
params,
|
|
|
|
desc="set_throttle_params"
|
2016-04-20 14:02:01 +02:00
|
|
|
)
|