MatrixSynapse/synapse/storage/_base.py

859 lines
29 KiB
Python
Raw Normal View History

2014-08-12 16:10:52 +02:00
# -*- coding: utf-8 -*-
2016-01-07 05:26:29 +01:00
# Copyright 2014-2016 OpenMarket Ltd
2014-08-12 16:10:52 +02:00
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import logging
from synapse.api.errors import StoreError
2016-02-04 11:22:44 +01:00
from synapse.util.logcontext import LoggingContext, PreserveLoggingContext
from synapse.util.caches.dictionary_cache import DictionaryCache
from synapse.util.caches.descriptors import Cache
2016-03-23 17:13:05 +01:00
from synapse.util.caches import intern_dict
import synapse.metrics
2014-08-12 16:10:52 +02:00
from twisted.internet import defer
2014-08-12 16:10:52 +02:00
import sys
2014-10-28 11:34:05 +01:00
import time
import threading
import os
CACHE_SIZE_FACTOR = float(os.environ.get("SYNAPSE_CACHE_FACTOR", 0.1))
2014-08-12 16:10:52 +02:00
logger = logging.getLogger(__name__)
sql_logger = logging.getLogger("synapse.storage.SQL")
transaction_logger = logging.getLogger("synapse.storage.txn")
perf_logger = logging.getLogger("synapse.storage.TIME")
metrics = synapse.metrics.get_metrics_for("synapse.storage")
sql_scheduling_timer = metrics.register_distribution("schedule_time")
sql_query_timer = metrics.register_distribution("query_time", labels=["verb"])
sql_txn_timer = metrics.register_distribution("transaction_time", labels=["desc"])
2015-08-07 19:14:49 +02:00
class LoggingTransaction(object):
"""An object that almost-transparently proxies for the 'txn' object
passed to the constructor. Adds logging and metrics to the .execute()
method."""
__slots__ = ["txn", "name", "database_engine", "after_callbacks"]
def __init__(self, txn, name, database_engine, after_callbacks):
object.__setattr__(self, "txn", txn)
object.__setattr__(self, "name", name)
object.__setattr__(self, "database_engine", database_engine)
object.__setattr__(self, "after_callbacks", after_callbacks)
def call_after(self, callback, *args):
2015-05-05 18:45:11 +02:00
"""Call the given callback on the main twisted thread after the
transaction has finished. Used to invalidate the caches on the
correct thread.
"""
self.after_callbacks.append((callback, args))
2014-10-28 11:53:11 +01:00
def __getattr__(self, name):
return getattr(self.txn, name)
def __setattr__(self, name, value):
2014-10-28 11:53:11 +01:00
setattr(self.txn, name, value)
2015-05-05 16:13:25 +02:00
def execute(self, sql, *args):
self._do_execute(self.txn.execute, sql, *args)
def executemany(self, sql, *args):
self._do_execute(self.txn.executemany, sql, *args)
def _do_execute(self, func, sql, *args):
# TODO(paul): Maybe use 'info' and 'debug' for values?
sql_logger.debug("[SQL] {%s} %s", self.name, sql)
sql = self.database_engine.convert_param_style(sql)
2015-05-05 16:13:25 +02:00
if args:
2015-04-02 11:06:22 +02:00
try:
sql_logger.debug(
2015-05-05 16:13:25 +02:00
"[SQL values] {%s} %r",
self.name, args[0]
)
2015-04-02 11:06:22 +02:00
except:
# Don't let logging failures stop SQL from working
pass
start = time.time() * 1000
2015-04-02 11:06:22 +02:00
2014-10-28 11:34:05 +01:00
try:
2015-05-05 16:13:25 +02:00
return func(
sql, *args
2015-04-08 14:11:28 +02:00
)
except Exception as e:
2015-04-08 14:11:28 +02:00
logger.debug("[SQL FAIL] {%s} %s", self.name, e)
raise
2014-10-28 11:34:05 +01:00
finally:
msecs = (time.time() * 1000) - start
sql_logger.debug("[SQL time] {%s} %f", self.name, msecs)
sql_query_timer.inc_by(msecs, sql.split()[0])
2014-08-12 16:10:52 +02:00
class PerformanceCounters(object):
def __init__(self):
self.current_counters = {}
self.previous_counters = {}
def update(self, key, start_time, end_time=None):
if end_time is None:
2015-02-10 17:30:48 +01:00
end_time = time.time() * 1000
duration = end_time - start_time
count, cum_time = self.current_counters.get(key, (0, 0))
count += 1
cum_time += duration
self.current_counters[key] = (count, cum_time)
return end_time
def interval(self, interval_duration, limit=3):
counters = []
for name, (count, cum_time) in self.current_counters.items():
prev_count, prev_time = self.previous_counters.get(name, (0, 0))
counters.append((
(cum_time - prev_time) / interval_duration,
count - prev_count,
name
))
self.previous_counters = dict(self.current_counters)
counters.sort(reverse=True)
top_n_counters = ", ".join(
"%s(%d): %.3f%%" % (name, count, 100 * ratio)
2015-02-10 15:54:07 +01:00
for ratio, count, name in counters[:limit]
)
return top_n_counters
2014-08-12 16:10:52 +02:00
class SQLBaseStore(object):
_TXN_ID = 0
2014-08-12 16:10:52 +02:00
def __init__(self, hs):
self.hs = hs
2014-08-12 16:10:52 +02:00
self._db_pool = hs.get_db_pool()
self._clock = hs.get_clock()
2014-08-12 16:10:52 +02:00
self._previous_txn_total_time = 0
self._current_txn_total_time = 0
self._previous_loop_ts = 0
# TODO(paul): These can eventually be removed once the metrics code
# is running in mainline, and we have some nice monitoring frontends
# to watch it
self._txn_perf_counters = PerformanceCounters()
self._get_event_counters = PerformanceCounters()
self._get_event_cache = Cache("*getEvent*", keylen=3, lru=True,
max_entries=hs.config.event_cache_size)
self._state_group_cache = DictionaryCache(
"*stateGroupCache*", 2000 * CACHE_SIZE_FACTOR
)
2015-05-15 12:35:04 +02:00
self._event_fetch_lock = threading.Condition()
self._event_fetch_list = []
2015-05-14 16:40:21 +02:00
self._event_fetch_ongoing = 0
2015-05-15 11:54:04 +02:00
self._pending_ds = []
self.database_engine = hs.database_engine
def start_profiling(self):
self._previous_loop_ts = self._clock.time_msec()
def loop():
curr = self._current_txn_total_time
prev = self._previous_txn_total_time
self._previous_txn_total_time = curr
time_now = self._clock.time_msec()
time_then = self._previous_loop_ts
self._previous_loop_ts = time_now
2016-02-02 18:18:50 +01:00
ratio = (curr - prev) / (time_now - time_then)
top_three_counters = self._txn_perf_counters.interval(
time_now - time_then, limit=3
)
top_3_event_counters = self._get_event_counters.interval(
time_now - time_then, limit=3
)
perf_logger.info(
2015-02-10 15:54:07 +01:00
"Total database time: %.3f%% {%s} {%s}",
ratio * 100, top_three_counters, top_3_event_counters
)
self._clock.looping_call(loop, 10000)
def _new_transaction(self, conn, desc, after_callbacks, logging_context,
func, *args, **kwargs):
2015-05-14 17:54:35 +02:00
start = time.time() * 1000
txn_id = self._TXN_ID
# We don't really need these to be unique, so lets stop it from
# growing really large.
self._TXN_ID = (self._TXN_ID + 1) % (sys.maxint - 1)
name = "%s-%x" % (desc, txn_id, )
transaction_logger.debug("[TXN START] {%s}", name)
try:
i = 0
N = 5
while True:
try:
txn = conn.cursor()
txn = LoggingTransaction(
txn, name, self.database_engine, after_callbacks
)
2015-05-15 11:54:04 +02:00
r = func(txn, *args, **kwargs)
conn.commit()
return r
2015-05-14 17:54:35 +02:00
except self.database_engine.module.OperationalError as e:
# This can happen if the database disappears mid
# transaction.
logger.warn(
"[TXN OPERROR] {%s} %s %d/%d",
name, e, i, N
)
if i < N:
i += 1
try:
conn.rollback()
except self.database_engine.module.Error as e1:
logger.warn(
"[TXN EROLL] {%s} %s",
name, e1,
)
continue
raise
except self.database_engine.module.DatabaseError as e:
if self.database_engine.is_deadlock(e):
logger.warn("[TXN DEADLOCK] {%s} %d/%d", name, i, N)
if i < N:
i += 1
try:
conn.rollback()
except self.database_engine.module.Error as e1:
logger.warn(
"[TXN EROLL] {%s} %s",
name, e1,
)
continue
raise
except Exception as e:
logger.debug("[TXN FAIL] {%s} %s", name, e)
raise
finally:
end = time.time() * 1000
duration = end - start
if logging_context is not None:
logging_context.add_database_transaction(duration)
2015-05-14 17:54:35 +02:00
transaction_logger.debug("[TXN END] {%s} %f", name, duration)
self._current_txn_total_time += duration
self._txn_perf_counters.update(desc, start, end)
sql_txn_timer.inc_by(duration, desc)
2014-10-30 02:21:33 +01:00
@defer.inlineCallbacks
def runInteraction(self, desc, func, *args, **kwargs):
"""Wraps the .runInteraction() method on the underlying db_pool."""
2014-10-30 02:21:33 +01:00
current_context = LoggingContext.current_context()
2014-11-20 18:26:36 +01:00
start_time = time.time() * 1000
after_callbacks = []
2015-04-08 14:11:28 +02:00
def inner_func(conn, *args, **kwargs):
2014-10-30 02:21:33 +01:00
with LoggingContext("runInteraction") as context:
2015-05-14 17:54:35 +02:00
sql_scheduling_timer.inc_by(time.time() * 1000 - start_time)
if self.database_engine.is_connection_closed(conn):
logger.debug("Reconnecting closed database connection")
conn.reconnect()
2014-10-30 02:21:33 +01:00
current_context.copy_to(context)
2015-05-15 11:54:04 +02:00
return self._new_transaction(
conn, desc, after_callbacks, current_context,
func, *args, **kwargs
2015-05-15 11:54:04 +02:00
)
2016-02-04 11:22:44 +01:00
with PreserveLoggingContext():
result = yield self._db_pool.runWithConnection(
inner_func, *args, **kwargs
)
2015-05-14 17:54:35 +02:00
for after_callback, after_args in after_callbacks:
after_callback(*after_args)
defer.returnValue(result)
2015-05-14 17:54:35 +02:00
@defer.inlineCallbacks
def runWithConnection(self, func, *args, **kwargs):
"""Wraps the .runInteraction() method on the underlying db_pool."""
current_context = LoggingContext.current_context()
start_time = time.time() * 1000
def inner_func(conn, *args, **kwargs):
with LoggingContext("runWithConnection") as context:
sql_scheduling_timer.inc_by(time.time() * 1000 - start_time)
2015-05-14 17:54:35 +02:00
if self.database_engine.is_connection_closed(conn):
logger.debug("Reconnecting closed database connection")
conn.reconnect()
current_context.copy_to(context)
2015-05-14 17:54:35 +02:00
return func(conn, *args, **kwargs)
2016-02-04 11:22:44 +01:00
with PreserveLoggingContext():
result = yield self._db_pool.runWithConnection(
inner_func, *args, **kwargs
)
2014-10-30 02:21:33 +01:00
defer.returnValue(result)
@staticmethod
def cursor_to_dict(cursor):
2014-08-12 16:10:52 +02:00
"""Converts a SQL cursor into an list of dicts.
Args:
cursor : The DBAPI cursor which has executed a query.
Returns:
A list of dicts where the key is the column header.
"""
col_headers = list(column[0] for column in cursor.description)
results = list(
2016-03-23 17:13:05 +01:00
intern_dict(dict(zip(col_headers, row))) for row in cursor.fetchall()
2014-08-12 16:10:52 +02:00
)
return results
2015-03-11 18:19:17 +01:00
def _execute(self, desc, decoder, query, *args):
2014-08-12 16:10:52 +02:00
"""Runs a single query for a result set.
Args:
decoder - The function which can resolve the cursor results to
something meaningful.
query - The query string to execute
*args - Query args.
Returns:
The result of decoder(results)
"""
def interaction(txn):
txn.execute(query, args)
if decoder:
return decoder(txn)
else:
return txn.fetchall()
2015-03-11 18:19:17 +01:00
return self.runInteraction(desc, interaction)
2014-08-12 16:10:52 +02:00
# "Simple" SQL API methods that operate on a single table with no JOINs,
# no complex WHERE clauses, just a dict of values for columns.
@defer.inlineCallbacks
2015-04-07 13:06:01 +02:00
def _simple_insert(self, table, values, or_ignore=False,
desc="_simple_insert"):
2014-08-12 16:10:52 +02:00
"""Executes an INSERT query on the named table.
Args:
table : string giving the table name
values : dict of new column names and values for them
"""
try:
yield self.runInteraction(
desc,
self._simple_insert_txn, table, values,
)
except self.database_engine.module.IntegrityError:
# We have to do or_ignore flag at this layer, since we can't reuse
# a cursor after we receive an error from the db.
if not or_ignore:
raise
@staticmethod
def _simple_insert_txn(txn, table, values):
2015-05-05 16:13:25 +02:00
keys, vals = zip(*values.items())
sql = "INSERT INTO %s (%s) VALUES(%s)" % (
2014-08-12 16:10:52 +02:00
table,
2015-05-05 16:13:25 +02:00
", ".join(k for k in keys),
", ".join("?" for _ in keys)
2014-08-12 16:10:52 +02:00
)
2015-05-05 16:13:25 +02:00
txn.execute(sql, vals)
@staticmethod
def _simple_insert_many_txn(txn, table, values):
2015-05-05 16:13:25 +02:00
if not values:
return
# This is a *slight* abomination to get a list of tuples of key names
# and a list of tuples of value names.
#
# i.e. [{"a": 1, "b": 2}, {"c": 3, "d": 4}]
# => [("a", "b",), ("c", "d",)] and [(1, 2,), (3, 4,)]
#
# The sort is to ensure that we don't rely on dictionary iteration
# order.
2015-05-05 16:13:25 +02:00
keys, vals = zip(*[
zip(
*(sorted(i.items(), key=lambda kv: kv[0]))
)
for i in values
if i
])
for k in keys:
if k != keys[0]:
raise RuntimeError(
"All items must have the same keys"
)
sql = "INSERT INTO %s (%s) VALUES(%s)" % (
table,
", ".join(k for k in keys[0]),
", ".join("?" for _ in keys[0])
)
2015-05-05 16:13:25 +02:00
txn.executemany(sql, vals)
2014-08-12 16:10:52 +02:00
2015-04-15 17:24:14 +02:00
def _simple_upsert(self, table, keyvalues, values,
2015-05-01 11:46:48 +02:00
insertion_values={}, desc="_simple_upsert", lock=True):
"""
2015-01-28 15:44:41 +01:00
Args:
table (str): The table to upsert into
keyvalues (dict): The unique key tables and their new values
values (dict): The nonunique columns and their new values
2015-04-15 17:24:14 +02:00
insertion_values (dict): key/values to use when inserting
2016-05-13 12:25:02 +02:00
Returns:
Deferred(bool): True if a new entry was created, False if an
2016-05-13 15:42:25 +02:00
existing one was updated.
"""
return self.runInteraction(
desc,
2015-04-15 17:24:14 +02:00
self._simple_upsert_txn, table, keyvalues, values, insertion_values,
2015-05-01 11:46:48 +02:00
lock
)
2015-05-01 11:46:48 +02:00
def _simple_upsert_txn(self, txn, table, keyvalues, values, insertion_values={},
lock=True):
# We need to lock the table :(, unless we're *really* careful
if lock:
self.database_engine.lock_table(txn, table)
# Try to update
sql = "UPDATE %s SET %s WHERE %s" % (
table,
", ".join("%s = ?" % (k,) for k in values),
" AND ".join("%s = ?" % (k,) for k in keyvalues)
)
sqlargs = values.values() + keyvalues.values()
logger.debug(
"[SQL] %s Args=%s",
sql, sqlargs,
)
txn.execute(sql, sqlargs)
if txn.rowcount == 0:
# We didn't update and rows so insert a new one
allvalues = {}
allvalues.update(keyvalues)
allvalues.update(values)
2015-04-15 17:24:14 +02:00
allvalues.update(insertion_values)
sql = "INSERT INTO %s (%s) VALUES (%s)" % (
table,
", ".join(k for k in allvalues),
", ".join("?" for _ in allvalues)
)
logger.debug(
"[SQL] %s Args=%s",
sql, keyvalues.values(),
)
txn.execute(sql, allvalues.values())
2016-05-13 12:25:02 +02:00
return True
else:
return False
2014-08-12 16:10:52 +02:00
def _simple_select_one(self, table, keyvalues, retcols,
2015-03-20 15:59:48 +01:00
allow_none=False, desc="_simple_select_one"):
2014-08-12 16:10:52 +02:00
"""Executes a SELECT query on the named table, which is expected to
return a single row, returning a single column from it.
Args:
table : string giving the table name
keyvalues : dict of column names and values to select the row with
retcols : list of strings giving the names of the columns to return
allow_none : If true, return None instead of failing if the SELECT
statement returns no rows
"""
2015-03-20 15:59:48 +01:00
return self.runInteraction(
desc,
self._simple_select_one_txn,
table, keyvalues, retcols, allow_none,
2014-08-12 16:10:52 +02:00
)
def _simple_select_one_onecol(self, table, keyvalues, retcol,
allow_none=False,
desc="_simple_select_one_onecol"):
2014-08-12 16:10:52 +02:00
"""Executes a SELECT query on the named table, which is expected to
2015-10-09 16:48:31 +02:00
return a single row, returning a single column from it.
2014-08-12 16:10:52 +02:00
Args:
table : string giving the table name
keyvalues : dict of column names and values to select the row with
retcol : string giving the name of the column to return
"""
return self.runInteraction(
desc,
self._simple_select_one_onecol_txn,
table, keyvalues, retcol, allow_none=allow_none,
)
@classmethod
def _simple_select_one_onecol_txn(cls, txn, table, keyvalues, retcol,
allow_none=False):
ret = cls._simple_select_onecol_txn(
txn,
2014-08-12 16:10:52 +02:00
table=table,
keyvalues=keyvalues,
retcol=retcol,
2014-08-12 16:10:52 +02:00
)
if ret:
return ret[0]
2014-08-12 16:10:52 +02:00
else:
if allow_none:
return None
else:
raise StoreError(404, "No row found")
@staticmethod
def _simple_select_onecol_txn(txn, table, keyvalues, retcol):
sql = (
"SELECT %(retcol)s FROM %(table)s WHERE %(where)s"
) % {
"retcol": retcol,
"table": table,
"where": " AND ".join("%s = ?" % k for k in keyvalues.keys()),
}
txn.execute(sql, keyvalues.values())
return [r[0] for r in txn.fetchall()]
2014-08-12 16:10:52 +02:00
def _simple_select_onecol(self, table, keyvalues, retcol,
desc="_simple_select_onecol"):
2014-08-12 16:10:52 +02:00
"""Executes a SELECT query on the named table, which returns a list
comprising of the values of the named column from the selected rows.
Args:
table (str): table name
keyvalues (dict): column names and values to select the rows with
retcol (str): column whos value we wish to retrieve.
Returns:
Deferred: Results in a list
"""
return self.runInteraction(
desc,
self._simple_select_onecol_txn,
table, keyvalues, retcol
)
2014-08-12 16:10:52 +02:00
def _simple_select_list(self, table, keyvalues, retcols,
desc="_simple_select_list"):
2014-08-12 16:10:52 +02:00
"""Executes a SELECT query on the named table, which may return zero or
more rows, returning the result as a list of dicts.
Args:
table : string giving the table name
keyvalues : dict of column names and values to select the rows with,
or None to not apply a WHERE clause.
2014-08-12 16:10:52 +02:00
retcols : list of strings giving the names of the columns to return
"""
return self.runInteraction(
desc,
self._simple_select_list_txn,
table, keyvalues, retcols
)
2014-08-12 16:10:52 +02:00
@classmethod
def _simple_select_list_txn(cls, txn, table, keyvalues, retcols):
2014-08-12 16:10:52 +02:00
"""Executes a SELECT query on the named table, which may return zero or
more rows, returning the result as a list of dicts.
Args:
txn : Transaction object
2014-08-12 16:10:52 +02:00
table : string giving the table name
keyvalues : dict of column names and values to select the rows with
retcols : list of strings giving the names of the columns to return
"""
if keyvalues:
sql = "SELECT %s FROM %s WHERE %s" % (
", ".join(retcols),
table,
" AND ".join("%s = ?" % (k, ) for k in keyvalues)
)
txn.execute(sql, keyvalues.values())
else:
sql = "SELECT %s FROM %s" % (
", ".join(retcols),
table
)
txn.execute(sql)
2014-08-12 16:10:52 +02:00
return cls.cursor_to_dict(txn)
2014-08-12 16:10:52 +02:00
2016-01-25 14:36:02 +01:00
@defer.inlineCallbacks
def _simple_select_many_batch(self, table, column, iterable, retcols,
keyvalues={}, desc="_simple_select_many_batch",
batch_size=100):
"""Executes a SELECT query on the named table, which may return zero or
more rows, returning the result as a list of dicts.
Filters rows by if value of `column` is in `iterable`.
Args:
table : string giving the table name
column : column name to test for inclusion against `iterable`
iterable : list
keyvalues : dict of column names and values to select the rows with
retcols : list of strings giving the names of the columns to return
"""
results = []
2016-01-25 16:59:29 +01:00
if not iterable:
defer.returnValue(results)
2016-02-02 18:18:50 +01:00
chunks = [
iterable[i:i + batch_size]
for i in xrange(0, len(iterable), batch_size)
]
2016-01-25 14:36:02 +01:00
for chunk in chunks:
rows = yield self.runInteraction(
desc,
self._simple_select_many_txn,
table, column, chunk, keyvalues, retcols
)
results.extend(rows)
defer.returnValue(results)
@classmethod
def _simple_select_many_txn(cls, txn, table, column, iterable, keyvalues, retcols):
2016-01-25 14:36:02 +01:00
"""Executes a SELECT query on the named table, which may return zero or
more rows, returning the result as a list of dicts.
Filters rows by if value of `column` is in `iterable`.
Args:
txn : Transaction object
table : string giving the table name
column : column name to test for inclusion against `iterable`
iterable : list
keyvalues : dict of column names and values to select the rows with
retcols : list of strings giving the names of the columns to return
"""
2016-01-25 16:59:29 +01:00
if not iterable:
return []
2016-01-25 14:36:02 +01:00
sql = "SELECT %s FROM %s" % (", ".join(retcols), table)
clauses = []
values = []
clauses.append(
"%s IN (%s)" % (column, ",".join("?" for _ in iterable))
)
values.extend(iterable)
for key, value in keyvalues.items():
clauses.append("%s = ?" % (key,))
values.append(value)
if clauses:
sql = "%s WHERE %s" % (
sql,
" AND ".join(clauses),
)
txn.execute(sql, values)
return cls.cursor_to_dict(txn)
2016-01-25 14:36:02 +01:00
2014-08-12 16:10:52 +02:00
def _simple_update_one(self, table, keyvalues, updatevalues,
2015-03-20 15:59:48 +01:00
desc="_simple_update_one"):
2014-08-12 16:10:52 +02:00
"""Executes an UPDATE query on the named table, setting new values for
columns in a row matching the key values.
Args:
table : string giving the table name
keyvalues : dict of column names and values to select the row with
updatevalues : dict giving column names and values to update
retcols : optional list of column names to return
If present, retcols gives a list of column names on which to perform
a SELECT statement *before* performing the UPDATE statement. The values
of these will be returned in a dict.
These are performed within the same transaction, allowing an atomic
get-and-set. This can be used to implement compare-and-set by putting
the update column in the 'keyvalues' dict as well.
"""
2015-03-20 15:59:48 +01:00
return self.runInteraction(
desc,
self._simple_update_one_txn,
table, keyvalues, updatevalues,
)
2014-08-12 16:10:52 +02:00
@staticmethod
def _simple_update_one_txn(txn, table, keyvalues, updatevalues):
2015-03-20 15:59:48 +01:00
update_sql = "UPDATE %s SET %s WHERE %s" % (
table,
", ".join("%s = ?" % (k,) for k in updatevalues),
" AND ".join("%s = ?" % (k,) for k in keyvalues)
)
2014-08-12 16:10:52 +02:00
2015-03-20 15:59:48 +01:00
txn.execute(
update_sql,
updatevalues.values() + keyvalues.values()
)
if txn.rowcount == 0:
raise StoreError(404, "No row found")
if txn.rowcount > 1:
raise StoreError(500, "More than one row matched")
@staticmethod
def _simple_select_one_txn(txn, table, keyvalues, retcols,
2015-03-20 17:03:25 +01:00
allow_none=False):
2015-03-24 17:19:01 +01:00
select_sql = "SELECT %s FROM %s WHERE %s" % (
2015-03-20 15:59:48 +01:00
", ".join(retcols),
table,
" AND ".join("%s = ?" % (k,) for k in keyvalues)
2015-03-20 15:59:48 +01:00
)
txn.execute(select_sql, keyvalues.values())
row = txn.fetchone()
if not row:
if allow_none:
return None
raise StoreError(404, "No row found")
if txn.rowcount > 1:
raise StoreError(500, "More than one row matched")
return dict(zip(retcols, row))
2014-08-12 16:10:52 +02:00
def _simple_delete_one(self, table, keyvalues, desc="_simple_delete_one"):
2014-08-12 16:10:52 +02:00
"""Executes a DELETE query on the named table, expecting to delete a
single row.
2016-03-01 14:35:37 +01:00
Args:
table : string giving the table name
keyvalues : dict of column names and values to select the row with
"""
return self.runInteraction(
desc, self._simple_delete_one_txn, table, keyvalues
)
@staticmethod
def _simple_delete_one_txn(txn, table, keyvalues):
"""Executes a DELETE query on the named table, expecting to delete a
single row.
2014-08-12 16:10:52 +02:00
Args:
table : string giving the table name
keyvalues : dict of column names and values to select the row with
"""
sql = "DELETE FROM %s WHERE %s" % (
table,
" AND ".join("%s = ?" % (k, ) for k in keyvalues)
2014-08-12 16:10:52 +02:00
)
2016-03-01 14:35:37 +01:00
txn.execute(sql, keyvalues.values())
if txn.rowcount == 0:
raise StoreError(404, "No row found")
if txn.rowcount > 1:
raise StoreError(500, "more than one row matched")
@staticmethod
def _simple_delete_txn(txn, table, keyvalues):
sql = "DELETE FROM %s WHERE %s" % (
table,
" AND ".join("%s = ?" % (k, ) for k in keyvalues)
)
return txn.execute(sql, keyvalues.values())
2014-08-12 16:10:52 +02:00
def _get_cache_dict(self, db_conn, table, entity_column, stream_column,
max_value):
# Fetch a mapping of room_id -> max stream position for "recent" rooms.
# It doesn't really matter how many we get, the StreamChangeCache will
# do the right thing to ensure it respects the max size of cache.
sql = (
"SELECT %(entity)s, MAX(%(stream)s) FROM %(table)s"
" WHERE %(stream)s > ? - 100000"
" GROUP BY %(entity)s"
) % {
"table": table,
"entity": entity_column,
"stream": stream_column,
}
sql = self.database_engine.convert_param_style(sql)
txn = db_conn.cursor()
txn.execute(sql, (int(max_value),))
rows = txn.fetchall()
txn.close()
cache = {
row[0]: int(row[1])
for row in rows
}
if cache:
min_val = min(cache.values())
else:
min_val = max_value
return cache, min_val
2015-03-20 14:52:56 +01:00
class _RollbackButIsFineException(Exception):
""" This exception is used to rollback a transaction without implying
something went wrong.
"""
pass