123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170 |
- # Copyright 2014-2016 OpenMarket Ltd
- # Copyright 2017-2018 New Vector Ltd
- # Copyright 2019 The Matrix.org Foundation C.I.C.
- #
- # Licensed under the Apache License, Version 2.0 (the "License");
- # you may not use this file except in compliance with the License.
- # You may obtain a copy of the License at
- #
- # http://www.apache.org/licenses/LICENSE-2.0
- #
- # Unless required by applicable law or agreed to in writing, software
- # distributed under the License is distributed on an "AS IS" BASIS,
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- # See the License for the specific language governing permissions and
- # limitations under the License.
- import logging
- from abc import ABCMeta
- from typing import TYPE_CHECKING, Any, Collection, Dict, Iterable, Optional, Union
- from synapse.storage.database import make_in_list_sql_clause # noqa: F401; noqa: F401
- from synapse.storage.database import DatabasePool, LoggingDatabaseConnection
- from synapse.types import get_domain_from_id
- from synapse.util import json_decoder
- from synapse.util.caches.descriptors import CachedFunction
- if TYPE_CHECKING:
- from synapse.server import HomeServer
- logger = logging.getLogger(__name__)
- # some of our subclasses have abstract methods, so we use the ABCMeta metaclass.
- class SQLBaseStore(metaclass=ABCMeta):
- """Base class for data stores that holds helper functions.
- Note that multiple instances of this class will exist as there will be one
- per data store (and not one per physical database).
- """
- def __init__(
- self,
- database: DatabasePool,
- db_conn: LoggingDatabaseConnection,
- hs: "HomeServer",
- ):
- self.hs = hs
- self._clock = hs.get_clock()
- self.database_engine = database.engine
- self.db_pool = database
- self.external_cached_functions: Dict[str, CachedFunction] = {}
- def process_replication_rows( # noqa: B027 (no-op by design)
- self,
- stream_name: str,
- instance_name: str,
- token: int,
- rows: Iterable[Any],
- ) -> None:
- pass
- def _invalidate_state_caches(
- self, room_id: str, members_changed: Collection[str]
- ) -> None:
- """Invalidates caches that are based on the current state, but does
- not stream invalidations down replication.
- Args:
- room_id: Room where state changed
- members_changed: The user_ids of members that have changed
- """
- # If there were any membership changes, purge the appropriate caches.
- for host in {get_domain_from_id(u) for u in members_changed}:
- self._attempt_to_invalidate_cache("is_host_joined", (room_id, host))
- if members_changed:
- self._attempt_to_invalidate_cache("get_users_in_room", (room_id,))
- self._attempt_to_invalidate_cache("get_current_hosts_in_room", (room_id,))
- self._attempt_to_invalidate_cache(
- "get_users_in_room_with_profiles", (room_id,)
- )
- self._attempt_to_invalidate_cache(
- "get_number_joined_users_in_room", (room_id,)
- )
- self._attempt_to_invalidate_cache("get_local_users_in_room", (room_id,))
- # There's no easy way of invalidating this cache for just the users
- # that have changed, so we just clear the entire thing.
- self._attempt_to_invalidate_cache("does_pair_of_users_share_a_room", None)
- for user_id in members_changed:
- self._attempt_to_invalidate_cache(
- "get_user_in_room_with_profile", (room_id, user_id)
- )
- self._attempt_to_invalidate_cache(
- "get_rooms_for_user_with_stream_ordering", (user_id,)
- )
- self._attempt_to_invalidate_cache("get_rooms_for_user", (user_id,))
- # Purge other caches based on room state.
- self._attempt_to_invalidate_cache("get_room_summary", (room_id,))
- self._attempt_to_invalidate_cache("get_partial_current_state_ids", (room_id,))
- def _attempt_to_invalidate_cache(
- self, cache_name: str, key: Optional[Collection[Any]]
- ) -> bool:
- """Attempts to invalidate the cache of the given name, ignoring if the
- cache doesn't exist. Mainly used for invalidating caches on workers,
- where they may not have the cache.
- Note that this function does not invalidate any remote caches, only the
- local in-memory ones. Any remote invalidation must be performed before
- calling this.
- Args:
- cache_name
- key: Entry to invalidate. If None then invalidates the entire
- cache.
- """
- try:
- cache = getattr(self, cache_name)
- except AttributeError:
- # Check if an externally defined module cache has been registered
- cache = self.external_cached_functions.get(cache_name)
- if not cache:
- # We probably haven't pulled in the cache in this worker,
- # which is fine.
- return False
- if key is None:
- cache.invalidate_all()
- else:
- # Prefer any local-only invalidation method. Invalidating any non-local
- # cache must be be done before this.
- invalidate_method = getattr(cache, "invalidate_local", cache.invalidate)
- invalidate_method(tuple(key))
- return True
- def register_external_cached_function(
- self, cache_name: str, func: CachedFunction
- ) -> None:
- self.external_cached_functions[cache_name] = func
- def db_to_json(db_content: Union[memoryview, bytes, bytearray, str]) -> Any:
- """
- Take some data from a database row and return a JSON-decoded object.
- Args:
- db_content: The JSON-encoded contents from the database.
- Returns:
- The object decoded from JSON.
- """
- # psycopg2 on Python 3 returns memoryview objects, which we need to
- # cast to bytes to decode
- if isinstance(db_content, memoryview):
- db_content = db_content.tobytes()
- # Decode it to a Unicode string before feeding it to the JSON decoder, since
- # it only supports handling strings
- if isinstance(db_content, (bytes, bytearray)):
- db_content = db_content.decode("utf8")
- try:
- return json_decoder.decode(db_content)
- except Exception:
- logging.warning("Tried to decode '%r' as JSON and failed", db_content)
- raise
|