123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124 |
- # Copyright 2014-2016 OpenMarket Ltd
- # Copyright 2017-2018 New Vector Ltd
- # Copyright 2019 The Matrix.org Foundation C.I.C.
- #
- # Licensed under the Apache License, Version 2.0 (the "License");
- # you may not use this file except in compliance with the License.
- # You may obtain a copy of the License at
- #
- # http://www.apache.org/licenses/LICENSE-2.0
- #
- # Unless required by applicable law or agreed to in writing, software
- # distributed under the License is distributed on an "AS IS" BASIS,
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- # See the License for the specific language governing permissions and
- # limitations under the License.
- import logging
- from abc import ABCMeta
- from typing import TYPE_CHECKING, Any, Collection, Iterable, Optional, Union
- from synapse.storage.database import LoggingTransaction # noqa: F401
- from synapse.storage.database import make_in_list_sql_clause # noqa: F401
- from synapse.storage.database import DatabasePool
- from synapse.storage.types import Connection
- from synapse.types import StreamToken, get_domain_from_id
- from synapse.util import json_decoder
- if TYPE_CHECKING:
- from synapse.server import HomeServer
- logger = logging.getLogger(__name__)
- # some of our subclasses have abstract methods, so we use the ABCMeta metaclass.
- class SQLBaseStore(metaclass=ABCMeta):
- """Base class for data stores that holds helper functions.
- Note that multiple instances of this class will exist as there will be one
- per data store (and not one per physical database).
- """
- def __init__(self, database: DatabasePool, db_conn: Connection, hs: "HomeServer"):
- self.hs = hs
- self._clock = hs.get_clock()
- self.database_engine = database.engine
- self.db_pool = database
- def process_replication_rows(
- self,
- stream_name: str,
- instance_name: str,
- token: StreamToken,
- rows: Iterable[Any],
- ) -> None:
- pass
- def _invalidate_state_caches(
- self, room_id: str, members_changed: Iterable[str]
- ) -> None:
- """Invalidates caches that are based on the current state, but does
- not stream invalidations down replication.
- Args:
- room_id: Room where state changed
- members_changed: The user_ids of members that have changed
- """
- for host in {get_domain_from_id(u) for u in members_changed}:
- self._attempt_to_invalidate_cache("is_host_joined", (room_id, host))
- self._attempt_to_invalidate_cache("get_users_in_room", (room_id,))
- self._attempt_to_invalidate_cache("get_users_in_room_with_profiles", (room_id,))
- self._attempt_to_invalidate_cache("get_room_summary", (room_id,))
- self._attempt_to_invalidate_cache("get_current_state_ids", (room_id,))
- def _attempt_to_invalidate_cache(
- self, cache_name: str, key: Optional[Collection[Any]]
- ) -> None:
- """Attempts to invalidate the cache of the given name, ignoring if the
- cache doesn't exist. Mainly used for invalidating caches on workers,
- where they may not have the cache.
- Args:
- cache_name
- key: Entry to invalidate. If None then invalidates the entire
- cache.
- """
- try:
- cache = getattr(self, cache_name)
- except AttributeError:
- # We probably haven't pulled in the cache in this worker,
- # which is fine.
- return
- if key is None:
- cache.invalidate_all()
- else:
- cache.invalidate(tuple(key))
- def db_to_json(db_content: Union[memoryview, bytes, bytearray, str]) -> Any:
- """
- Take some data from a database row and return a JSON-decoded object.
- Args:
- db_content: The JSON-encoded contents from the database.
- Returns:
- The object decoded from JSON.
- """
- # psycopg2 on Python 3 returns memoryview objects, which we need to
- # cast to bytes to decode
- if isinstance(db_content, memoryview):
- db_content = db_content.tobytes()
- # Decode it to a Unicode string before feeding it to the JSON decoder, since
- # it only supports handling strings
- if isinstance(db_content, (bytes, bytearray)):
- db_content = db_content.decode("utf8")
- try:
- return json_decoder.decode(db_content)
- except Exception:
- logging.warning("Tried to decode '%r' as JSON and failed", db_content)
- raise
|