123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675 |
- # -*- coding: utf-8 -*-
- # Copyright 2014-2016 OpenMarket Ltd
- # Copyright 2018-2019 New Vector Ltd
- #
- # Licensed under the Apache License, Version 2.0 (the "License");
- # you may not use this file except in compliance with the License.
- # You may obtain a copy of the License at
- #
- # http://www.apache.org/licenses/LICENSE-2.0
- #
- # Unless required by applicable law or agreed to in writing, software
- # distributed under the License is distributed on an "AS IS" BASIS,
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- # See the License for the specific language governing permissions and
- # limitations under the License.
- import atexit
- import hashlib
- import os
- import time
- import uuid
- import warnings
- from inspect import getcallargs
- from mock import Mock, patch
- from six.moves.urllib import parse as urlparse
- from twisted.internet import defer, reactor
- from synapse.api.constants import EventTypes
- from synapse.api.errors import CodeMessageException, cs_error
- from synapse.api.room_versions import RoomVersions
- from synapse.config.homeserver import HomeServerConfig
- from synapse.config.server import DEFAULT_ROOM_VERSION
- from synapse.federation.transport import server as federation_server
- from synapse.http.server import HttpServer
- from synapse.logging.context import LoggingContext
- from synapse.server import HomeServer
- from synapse.storage import DataStore
- from synapse.storage.engines import PostgresEngine, create_engine
- from synapse.storage.prepare_database import (
- _get_or_create_schema_state,
- _setup_new_database,
- prepare_database,
- )
- from synapse.util.ratelimitutils import FederationRateLimiter
- # set this to True to run the tests against postgres instead of sqlite.
- #
- # When running under postgres, we first create a base database with the name
- # POSTGRES_BASE_DB and update it to the current schema. Then, for each test case, we
- # create another unique database, using the base database as a template.
- USE_POSTGRES_FOR_TESTS = os.environ.get("SYNAPSE_POSTGRES", False)
- LEAVE_DB = os.environ.get("SYNAPSE_LEAVE_DB", False)
- POSTGRES_USER = os.environ.get("SYNAPSE_POSTGRES_USER", None)
- POSTGRES_HOST = os.environ.get("SYNAPSE_POSTGRES_HOST", None)
- POSTGRES_PASSWORD = os.environ.get("SYNAPSE_POSTGRES_PASSWORD", None)
- POSTGRES_BASE_DB = "_synapse_unit_tests_base_%s" % (os.getpid(),)
- # the dbname we will connect to in order to create the base database.
- POSTGRES_DBNAME_FOR_INITIAL_CREATE = "postgres"
- def setupdb():
- # If we're using PostgreSQL, set up the db once
- if USE_POSTGRES_FOR_TESTS:
- # create a PostgresEngine
- db_engine = create_engine({"name": "psycopg2", "args": {}})
- # connect to postgres to create the base database.
- db_conn = db_engine.module.connect(
- user=POSTGRES_USER,
- host=POSTGRES_HOST,
- password=POSTGRES_PASSWORD,
- dbname=POSTGRES_DBNAME_FOR_INITIAL_CREATE,
- )
- db_conn.autocommit = True
- cur = db_conn.cursor()
- cur.execute("DROP DATABASE IF EXISTS %s;" % (POSTGRES_BASE_DB,))
- cur.execute("CREATE DATABASE %s;" % (POSTGRES_BASE_DB,))
- cur.close()
- db_conn.close()
- # Set up in the db
- db_conn = db_engine.module.connect(
- database=POSTGRES_BASE_DB,
- user=POSTGRES_USER,
- host=POSTGRES_HOST,
- password=POSTGRES_PASSWORD,
- )
- cur = db_conn.cursor()
- _get_or_create_schema_state(cur, db_engine)
- _setup_new_database(cur, db_engine)
- db_conn.commit()
- cur.close()
- db_conn.close()
- def _cleanup():
- db_conn = db_engine.module.connect(
- user=POSTGRES_USER,
- host=POSTGRES_HOST,
- password=POSTGRES_PASSWORD,
- dbname=POSTGRES_DBNAME_FOR_INITIAL_CREATE,
- )
- db_conn.autocommit = True
- cur = db_conn.cursor()
- cur.execute("DROP DATABASE IF EXISTS %s;" % (POSTGRES_BASE_DB,))
- cur.close()
- db_conn.close()
- atexit.register(_cleanup)
- def default_config(name, parse=False):
- """
- Create a reasonable test config.
- """
- config_dict = {
- "server_name": name,
- "media_store_path": "media",
- "uploads_path": "uploads",
- # the test signing key is just an arbitrary ed25519 key to keep the config
- # parser happy
- "signing_key": "ed25519 a_lPym qvioDNmfExFBRPgdTU+wtFYKq4JfwFRv7sYVgWvmgJg",
- "event_cache_size": 1,
- "enable_registration": True,
- "enable_registration_captcha": False,
- "macaroon_secret_key": "not even a little secret",
- "trusted_third_party_id_servers": [],
- "room_invite_state_types": [],
- "password_providers": [],
- "worker_replication_url": "",
- "worker_app": None,
- "block_non_admin_invites": False,
- "federation_domain_whitelist": None,
- "filter_timeline_limit": 5000,
- "user_directory_search_all_users": False,
- "user_consent_server_notice_content": None,
- "block_events_without_consent_error": None,
- "user_consent_at_registration": False,
- "user_consent_policy_name": "Privacy Policy",
- "media_storage_providers": [],
- "autocreate_auto_join_rooms": True,
- "auto_join_rooms": [],
- "limit_usage_by_mau": False,
- "hs_disabled": False,
- "hs_disabled_message": "",
- "hs_disabled_limit_type": "",
- "max_mau_value": 50,
- "mau_trial_days": 0,
- "mau_stats_only": False,
- "mau_limits_reserved_threepids": [],
- "admin_contact": None,
- "rc_message": {"per_second": 10000, "burst_count": 10000},
- "rc_registration": {"per_second": 10000, "burst_count": 10000},
- "rc_login": {
- "address": {"per_second": 10000, "burst_count": 10000},
- "account": {"per_second": 10000, "burst_count": 10000},
- "failed_attempts": {"per_second": 10000, "burst_count": 10000},
- },
- "saml2_enabled": False,
- "public_baseurl": None,
- "default_identity_server": None,
- "key_refresh_interval": 24 * 60 * 60 * 1000,
- "old_signing_keys": {},
- "tls_fingerprints": [],
- "use_frozen_dicts": False,
- # We need a sane default_room_version, otherwise attempts to create
- # rooms will fail.
- "default_room_version": DEFAULT_ROOM_VERSION,
- # disable user directory updates, because they get done in the
- # background, which upsets the test runner.
- "update_user_directory": False,
- }
- if parse:
- config = HomeServerConfig()
- config.parse_config_dict(config_dict, "", "")
- return config
- return config_dict
- class TestHomeServer(HomeServer):
- DATASTORE_CLASS = DataStore
- @defer.inlineCallbacks
- def setup_test_homeserver(
- cleanup_func,
- name="test",
- datastore=None,
- config=None,
- reactor=None,
- homeserverToUse=TestHomeServer,
- **kargs
- ):
- """
- Setup a homeserver suitable for running tests against. Keyword arguments
- are passed to the Homeserver constructor.
- If no datastore is supplied, one is created and given to the homeserver.
- Args:
- cleanup_func : The function used to register a cleanup routine for
- after the test.
- Calling this method directly is deprecated: you should instead derive from
- HomeserverTestCase.
- """
- if reactor is None:
- from twisted.internet import reactor
- if config is None:
- config = default_config(name, parse=True)
- config.ldap_enabled = False
- if "clock" not in kargs:
- kargs["clock"] = MockClock()
- if USE_POSTGRES_FOR_TESTS:
- test_db = "synapse_test_%s" % uuid.uuid4().hex
- config.database_config = {
- "name": "psycopg2",
- "args": {
- "database": test_db,
- "host": POSTGRES_HOST,
- "password": POSTGRES_PASSWORD,
- "user": POSTGRES_USER,
- "cp_min": 1,
- "cp_max": 5,
- },
- }
- else:
- config.database_config = {
- "name": "sqlite3",
- "args": {"database": ":memory:", "cp_min": 1, "cp_max": 1},
- }
- db_engine = create_engine(config.database_config)
- # Create the database before we actually try and connect to it, based off
- # the template database we generate in setupdb()
- if datastore is None and isinstance(db_engine, PostgresEngine):
- db_conn = db_engine.module.connect(
- database=POSTGRES_BASE_DB,
- user=POSTGRES_USER,
- host=POSTGRES_HOST,
- password=POSTGRES_PASSWORD,
- )
- db_conn.autocommit = True
- cur = db_conn.cursor()
- cur.execute("DROP DATABASE IF EXISTS %s;" % (test_db,))
- cur.execute(
- "CREATE DATABASE %s WITH TEMPLATE %s;" % (test_db, POSTGRES_BASE_DB)
- )
- cur.close()
- db_conn.close()
- # we need to configure the connection pool to run the on_new_connection
- # function, so that we can test code that uses custom sqlite functions
- # (like rank).
- config.database_config["args"]["cp_openfun"] = db_engine.on_new_connection
- if datastore is None:
- hs = homeserverToUse(
- name,
- config=config,
- db_config=config.database_config,
- version_string="Synapse/tests",
- database_engine=db_engine,
- tls_server_context_factory=Mock(),
- tls_client_options_factory=Mock(),
- reactor=reactor,
- **kargs
- )
- # Prepare the DB on SQLite -- PostgreSQL is a copy of an already up to
- # date db
- if not isinstance(db_engine, PostgresEngine):
- db_conn = hs.get_db_conn()
- yield prepare_database(db_conn, db_engine, config)
- db_conn.commit()
- db_conn.close()
- else:
- # We need to do cleanup on PostgreSQL
- def cleanup():
- import psycopg2
- # Close all the db pools
- hs.get_db_pool().close()
- dropped = False
- # Drop the test database
- db_conn = db_engine.module.connect(
- database=POSTGRES_BASE_DB,
- user=POSTGRES_USER,
- host=POSTGRES_HOST,
- password=POSTGRES_PASSWORD,
- )
- db_conn.autocommit = True
- cur = db_conn.cursor()
- # Try a few times to drop the DB. Some things may hold on to the
- # database for a few more seconds due to flakiness, preventing
- # us from dropping it when the test is over. If we can't drop
- # it, warn and move on.
- for x in range(5):
- try:
- cur.execute("DROP DATABASE IF EXISTS %s;" % (test_db,))
- db_conn.commit()
- dropped = True
- except psycopg2.OperationalError as e:
- warnings.warn(
- "Couldn't drop old db: " + str(e), category=UserWarning
- )
- time.sleep(0.5)
- cur.close()
- db_conn.close()
- if not dropped:
- warnings.warn("Failed to drop old DB.", category=UserWarning)
- if not LEAVE_DB:
- # Register the cleanup hook
- cleanup_func(cleanup)
- hs.setup()
- if homeserverToUse.__name__ == "TestHomeServer":
- hs.setup_master()
- else:
- hs = homeserverToUse(
- name,
- db_pool=None,
- datastore=datastore,
- config=config,
- version_string="Synapse/tests",
- database_engine=db_engine,
- tls_server_context_factory=Mock(),
- tls_client_options_factory=Mock(),
- reactor=reactor,
- **kargs
- )
- # bcrypt is far too slow to be doing in unit tests
- # Need to let the HS build an auth handler and then mess with it
- # because AuthHandler's constructor requires the HS, so we can't make one
- # beforehand and pass it in to the HS's constructor (chicken / egg)
- hs.get_auth_handler().hash = lambda p: hashlib.md5(p.encode("utf8")).hexdigest()
- hs.get_auth_handler().validate_hash = (
- lambda p, h: hashlib.md5(p.encode("utf8")).hexdigest() == h
- )
- fed = kargs.get("resource_for_federation", None)
- if fed:
- register_federation_servlets(hs, fed)
- return hs
- def register_federation_servlets(hs, resource):
- federation_server.register_servlets(
- hs,
- resource=resource,
- authenticator=federation_server.Authenticator(hs),
- ratelimiter=FederationRateLimiter(
- hs.get_clock(), config=hs.config.rc_federation
- ),
- )
- def get_mock_call_args(pattern_func, mock_func):
- """ Return the arguments the mock function was called with interpreted
- by the pattern functions argument list.
- """
- invoked_args, invoked_kargs = mock_func.call_args
- return getcallargs(pattern_func, *invoked_args, **invoked_kargs)
- def mock_getRawHeaders(headers=None):
- headers = headers if headers is not None else {}
- def getRawHeaders(name, default=None):
- return headers.get(name, default)
- return getRawHeaders
- # This is a mock /resource/ not an entire server
- class MockHttpResource(HttpServer):
- def __init__(self, prefix=""):
- self.callbacks = [] # 3-tuple of method/pattern/function
- self.prefix = prefix
- def trigger_get(self, path):
- return self.trigger(b"GET", path, None)
- @patch("twisted.web.http.Request")
- @defer.inlineCallbacks
- def trigger(
- self, http_method, path, content, mock_request, federation_auth_origin=None
- ):
- """ Fire an HTTP event.
- Args:
- http_method : The HTTP method
- path : The HTTP path
- content : The HTTP body
- mock_request : Mocked request to pass to the event so it can get
- content.
- federation_auth_origin (bytes|None): domain to authenticate as, for federation
- Returns:
- A tuple of (code, response)
- Raises:
- KeyError If no event is found which will handle the path.
- """
- path = self.prefix + path
- # annoyingly we return a twisted http request which has chained calls
- # to get at the http content, hence mock it here.
- mock_content = Mock()
- config = {"read.return_value": content}
- mock_content.configure_mock(**config)
- mock_request.content = mock_content
- mock_request.method = http_method.encode("ascii")
- mock_request.uri = path.encode("ascii")
- mock_request.getClientIP.return_value = "-"
- headers = {}
- if federation_auth_origin is not None:
- headers[b"Authorization"] = [
- b"X-Matrix origin=%s,key=,sig=" % (federation_auth_origin,)
- ]
- mock_request.requestHeaders.getRawHeaders = mock_getRawHeaders(headers)
- # return the right path if the event requires it
- mock_request.path = path
- # add in query params to the right place
- try:
- mock_request.args = urlparse.parse_qs(path.split("?")[1])
- mock_request.path = path.split("?")[0]
- path = mock_request.path
- except Exception:
- pass
- if isinstance(path, bytes):
- path = path.decode("utf8")
- for (method, pattern, func) in self.callbacks:
- if http_method != method:
- continue
- matcher = pattern.match(path)
- if matcher:
- try:
- args = [urlparse.unquote(u) for u in matcher.groups()]
- (code, response) = yield func(mock_request, *args)
- return code, response
- except CodeMessageException as e:
- return (e.code, cs_error(e.msg, code=e.errcode))
- raise KeyError("No event can handle %s" % path)
- def register_paths(self, method, path_patterns, callback, servlet_name):
- for path_pattern in path_patterns:
- self.callbacks.append((method, path_pattern, callback))
- class MockKey(object):
- alg = "mock_alg"
- version = "mock_version"
- signature = b"\x9a\x87$"
- @property
- def verify_key(self):
- return self
- def sign(self, message):
- return self
- def verify(self, message, sig):
- assert sig == b"\x9a\x87$"
- def encode(self):
- return b"<fake_encoded_key>"
- class MockClock(object):
- now = 1000
- def __init__(self):
- # list of lists of [absolute_time, callback, expired] in no particular
- # order
- self.timers = []
- self.loopers = []
- def time(self):
- return self.now
- def time_msec(self):
- return self.time() * 1000
- def call_later(self, delay, callback, *args, **kwargs):
- current_context = LoggingContext.current_context()
- def wrapped_callback():
- LoggingContext.thread_local.current_context = current_context
- callback(*args, **kwargs)
- t = [self.now + delay, wrapped_callback, False]
- self.timers.append(t)
- return t
- def looping_call(self, function, interval):
- self.loopers.append([function, interval / 1000.0, self.now])
- def cancel_call_later(self, timer, ignore_errs=False):
- if timer[2]:
- if not ignore_errs:
- raise Exception("Cannot cancel an expired timer")
- timer[2] = True
- self.timers = [t for t in self.timers if t != timer]
- # For unit testing
- def advance_time(self, secs):
- self.now += secs
- timers = self.timers
- self.timers = []
- for t in timers:
- time, callback, expired = t
- if expired:
- raise Exception("Timer already expired")
- if self.now >= time:
- t[2] = True
- callback()
- else:
- self.timers.append(t)
- for looped in self.loopers:
- func, interval, last = looped
- if last + interval < self.now:
- func()
- looped[2] = self.now
- def advance_time_msec(self, ms):
- self.advance_time(ms / 1000.0)
- def time_bound_deferred(self, d, *args, **kwargs):
- # We don't bother timing things out for now.
- return d
- def _format_call(args, kwargs):
- return ", ".join(
- ["%r" % (a) for a in args] + ["%s=%r" % (k, v) for k, v in kwargs.items()]
- )
- class DeferredMockCallable(object):
- """A callable instance that stores a set of pending call expectations and
- return values for them. It allows a unit test to assert that the given set
- of function calls are eventually made, by awaiting on them to be called.
- """
- def __init__(self):
- self.expectations = []
- self.calls = []
- def __call__(self, *args, **kwargs):
- self.calls.append((args, kwargs))
- if not self.expectations:
- raise ValueError(
- "%r has no pending calls to handle call(%s)"
- % (self, _format_call(args, kwargs))
- )
- for (call, result, d) in self.expectations:
- if args == call[1] and kwargs == call[2]:
- d.callback(None)
- return result
- failure = AssertionError(
- "Was not expecting call(%s)" % (_format_call(args, kwargs))
- )
- for _, _, d in self.expectations:
- try:
- d.errback(failure)
- except Exception:
- pass
- raise failure
- def expect_call_and_return(self, call, result):
- self.expectations.append((call, result, defer.Deferred()))
- @defer.inlineCallbacks
- def await_calls(self, timeout=1000):
- deferred = defer.DeferredList(
- [d for _, _, d in self.expectations], fireOnOneErrback=True
- )
- timer = reactor.callLater(
- timeout / 1000,
- deferred.errback,
- AssertionError(
- "%d pending calls left: %s"
- % (
- len([e for e in self.expectations if not e[2].called]),
- [e for e in self.expectations if not e[2].called],
- )
- ),
- )
- yield deferred
- timer.cancel()
- self.calls = []
- def assert_had_no_calls(self):
- if self.calls:
- calls = self.calls
- self.calls = []
- raise AssertionError(
- "Expected not to received any calls, got:\n"
- + "\n".join(["call(%s)" % _format_call(c[0], c[1]) for c in calls])
- )
- @defer.inlineCallbacks
- def create_room(hs, room_id, creator_id):
- """Creates and persist a creation event for the given room
- Args:
- hs
- room_id (str)
- creator_id (str)
- """
- store = hs.get_datastore()
- event_builder_factory = hs.get_event_builder_factory()
- event_creation_handler = hs.get_event_creation_handler()
- builder = event_builder_factory.for_room_version(
- RoomVersions.V1,
- {
- "type": EventTypes.Create,
- "state_key": "",
- "sender": creator_id,
- "room_id": room_id,
- "content": {},
- },
- )
- event, context = yield event_creation_handler.create_new_client_event(builder)
- yield store.persist_event(event, context)
|