125 lines
		
	
	
		
			4.3 KiB
		
	
	
	
		
			Python
		
	
	
			
		
		
	
	
			125 lines
		
	
	
		
			4.3 KiB
		
	
	
	
		
			Python
		
	
	
| # Copyright 2014-2016 OpenMarket Ltd
 | |
| # Copyright 2017-2018 New Vector Ltd
 | |
| # Copyright 2019 The Matrix.org Foundation C.I.C.
 | |
| #
 | |
| # Licensed under the Apache License, Version 2.0 (the "License");
 | |
| # you may not use this file except in compliance with the License.
 | |
| # You may obtain a copy of the License at
 | |
| #
 | |
| #     http://www.apache.org/licenses/LICENSE-2.0
 | |
| #
 | |
| # Unless required by applicable law or agreed to in writing, software
 | |
| # distributed under the License is distributed on an "AS IS" BASIS,
 | |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| # See the License for the specific language governing permissions and
 | |
| # limitations under the License.
 | |
| import logging
 | |
| from abc import ABCMeta
 | |
| from typing import TYPE_CHECKING, Any, Collection, Iterable, Optional, Union
 | |
| 
 | |
| from synapse.storage.database import LoggingTransaction  # noqa: F401
 | |
| from synapse.storage.database import make_in_list_sql_clause  # noqa: F401
 | |
| from synapse.storage.database import DatabasePool
 | |
| from synapse.storage.types import Connection
 | |
| from synapse.types import get_domain_from_id
 | |
| from synapse.util import json_decoder
 | |
| 
 | |
| if TYPE_CHECKING:
 | |
|     from synapse.server import HomeServer
 | |
| 
 | |
| logger = logging.getLogger(__name__)
 | |
| 
 | |
| 
 | |
| # some of our subclasses have abstract methods, so we use the ABCMeta metaclass.
 | |
| class SQLBaseStore(metaclass=ABCMeta):
 | |
|     """Base class for data stores that holds helper functions.
 | |
| 
 | |
|     Note that multiple instances of this class will exist as there will be one
 | |
|     per data store (and not one per physical database).
 | |
|     """
 | |
| 
 | |
|     def __init__(self, database: DatabasePool, db_conn: Connection, hs: "HomeServer"):
 | |
|         self.hs = hs
 | |
|         self._clock = hs.get_clock()
 | |
|         self.database_engine = database.engine
 | |
|         self.db_pool = database
 | |
| 
 | |
|     def process_replication_rows(
 | |
|         self,
 | |
|         stream_name: str,
 | |
|         instance_name: str,
 | |
|         token: int,
 | |
|         rows: Iterable[Any],
 | |
|     ) -> None:
 | |
|         pass
 | |
| 
 | |
|     def _invalidate_state_caches(
 | |
|         self, room_id: str, members_changed: Iterable[str]
 | |
|     ) -> None:
 | |
|         """Invalidates caches that are based on the current state, but does
 | |
|         not stream invalidations down replication.
 | |
| 
 | |
|         Args:
 | |
|             room_id: Room where state changed
 | |
|             members_changed: The user_ids of members that have changed
 | |
|         """
 | |
|         for host in {get_domain_from_id(u) for u in members_changed}:
 | |
|             self._attempt_to_invalidate_cache("is_host_joined", (room_id, host))
 | |
| 
 | |
|         self._attempt_to_invalidate_cache("get_users_in_room", (room_id,))
 | |
|         self._attempt_to_invalidate_cache("get_users_in_room_with_profiles", (room_id,))
 | |
|         self._attempt_to_invalidate_cache("get_room_summary", (room_id,))
 | |
|         self._attempt_to_invalidate_cache("get_current_state_ids", (room_id,))
 | |
| 
 | |
|     def _attempt_to_invalidate_cache(
 | |
|         self, cache_name: str, key: Optional[Collection[Any]]
 | |
|     ) -> None:
 | |
|         """Attempts to invalidate the cache of the given name, ignoring if the
 | |
|         cache doesn't exist. Mainly used for invalidating caches on workers,
 | |
|         where they may not have the cache.
 | |
| 
 | |
|         Args:
 | |
|             cache_name
 | |
|             key: Entry to invalidate. If None then invalidates the entire
 | |
|                 cache.
 | |
|         """
 | |
| 
 | |
|         try:
 | |
|             cache = getattr(self, cache_name)
 | |
|         except AttributeError:
 | |
|             # We probably haven't pulled in the cache in this worker,
 | |
|             # which is fine.
 | |
|             return
 | |
| 
 | |
|         if key is None:
 | |
|             cache.invalidate_all()
 | |
|         else:
 | |
|             cache.invalidate(tuple(key))
 | |
| 
 | |
| 
 | |
| def db_to_json(db_content: Union[memoryview, bytes, bytearray, str]) -> Any:
 | |
|     """
 | |
|     Take some data from a database row and return a JSON-decoded object.
 | |
| 
 | |
|     Args:
 | |
|         db_content: The JSON-encoded contents from the database.
 | |
| 
 | |
|     Returns:
 | |
|         The object decoded from JSON.
 | |
|     """
 | |
|     # psycopg2 on Python 3 returns memoryview objects, which we need to
 | |
|     # cast to bytes to decode
 | |
|     if isinstance(db_content, memoryview):
 | |
|         db_content = db_content.tobytes()
 | |
| 
 | |
|     # Decode it to a Unicode string before feeding it to the JSON decoder, since
 | |
|     # it only supports handling strings
 | |
|     if isinstance(db_content, (bytes, bytearray)):
 | |
|         db_content = db_content.decode("utf8")
 | |
| 
 | |
|     try:
 | |
|         return json_decoder.decode(db_content)
 | |
|     except Exception:
 | |
|         logging.warning("Tried to decode '%r' as JSON and failed", db_content)
 | |
|         raise
 |