123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657 |
- #!/usr/bin/env python
- # -*- coding: utf-8 -*-
- # Copyright 2014-2016 OpenMarket Ltd
- # Copyright 2019 New Vector Ltd
- #
- # Licensed under the Apache License, Version 2.0 (the "License");
- # you may not use this file except in compliance with the License.
- # You may obtain a copy of the License at
- #
- # http://www.apache.org/licenses/LICENSE-2.0
- #
- # Unless required by applicable law or agreed to in writing, software
- # distributed under the License is distributed on an "AS IS" BASIS,
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- # See the License for the specific language governing permissions and
- # limitations under the License.
- from __future__ import print_function
- import gc
- import logging
- import os
- import sys
- from six import iteritems
- import psutil
- from prometheus_client import Gauge
- from twisted.application import service
- from twisted.internet import defer, reactor
- from twisted.python.failure import Failure
- from twisted.web.resource import EncodingResourceWrapper, NoResource
- from twisted.web.server import GzipEncoderFactory
- from twisted.web.static import File
- import synapse
- import synapse.config.logger
- from synapse import events
- from synapse.api.urls import (
- CONTENT_REPO_PREFIX,
- FEDERATION_PREFIX,
- LEGACY_MEDIA_PREFIX,
- MEDIA_PREFIX,
- SERVER_KEY_V2_PREFIX,
- STATIC_PREFIX,
- WEB_CLIENT_PREFIX,
- )
- from synapse.app import _base
- from synapse.app._base import listen_ssl, listen_tcp, quit_with_error
- from synapse.config._base import ConfigError
- from synapse.config.homeserver import HomeServerConfig
- from synapse.federation.transport.server import TransportLayerServer
- from synapse.http.additional_resource import AdditionalResource
- from synapse.http.server import RootRedirect
- from synapse.http.site import SynapseSite
- from synapse.logging.context import LoggingContext
- from synapse.metrics import METRICS_PREFIX, MetricsResource, RegistryProxy
- from synapse.metrics.background_process_metrics import run_as_background_process
- from synapse.module_api import ModuleApi
- from synapse.python_dependencies import check_requirements
- from synapse.replication.http import REPLICATION_PREFIX, ReplicationRestResource
- from synapse.replication.tcp.resource import ReplicationStreamProtocolFactory
- from synapse.rest import ClientRestResource
- from synapse.rest.admin import AdminRestResource
- from synapse.rest.key.v2 import KeyApiV2Resource
- from synapse.rest.media.v0.content_repository import ContentRepoResource
- from synapse.rest.well_known import WellKnownResource
- from synapse.server import HomeServer
- from synapse.storage import DataStore, are_all_users_on_domain
- from synapse.storage.engines import IncorrectDatabaseSetup, create_engine
- from synapse.storage.prepare_database import UpgradeDatabaseException, prepare_database
- from synapse.util.caches import CACHE_SIZE_FACTOR
- from synapse.util.httpresourcetree import create_resource_tree
- from synapse.util.manhole import manhole
- from synapse.util.module_loader import load_module
- from synapse.util.rlimit import change_resource_limit
- from synapse.util.versionstring import get_version_string
- logger = logging.getLogger("synapse.app.homeserver")
- def gz_wrap(r):
- return EncodingResourceWrapper(r, [GzipEncoderFactory()])
- class SynapseHomeServer(HomeServer):
- DATASTORE_CLASS = DataStore
- def _listener_http(self, config, listener_config):
- port = listener_config["port"]
- bind_addresses = listener_config["bind_addresses"]
- tls = listener_config.get("tls", False)
- site_tag = listener_config.get("tag", port)
- resources = {}
- for res in listener_config["resources"]:
- for name in res["names"]:
- if name == "openid" and "federation" in res["names"]:
- # Skip loading openid resource if federation is defined
- # since federation resource will include openid
- continue
- resources.update(
- self._configure_named_resource(name, res.get("compress", False))
- )
- additional_resources = listener_config.get("additional_resources", {})
- logger.debug("Configuring additional resources: %r", additional_resources)
- module_api = ModuleApi(self, self.get_auth_handler())
- for path, resmodule in additional_resources.items():
- handler_cls, config = load_module(resmodule)
- handler = handler_cls(config, module_api)
- resources[path] = AdditionalResource(self, handler.handle_request)
- # try to find something useful to redirect '/' to
- if WEB_CLIENT_PREFIX in resources:
- root_resource = RootRedirect(WEB_CLIENT_PREFIX)
- elif STATIC_PREFIX in resources:
- root_resource = RootRedirect(STATIC_PREFIX)
- else:
- root_resource = NoResource()
- root_resource = create_resource_tree(resources, root_resource)
- if tls:
- ports = listen_ssl(
- bind_addresses,
- port,
- SynapseSite(
- "synapse.access.https.%s" % (site_tag,),
- site_tag,
- listener_config,
- root_resource,
- self.version_string,
- ),
- self.tls_server_context_factory,
- reactor=self.get_reactor(),
- )
- logger.info("Synapse now listening on TCP port %d (TLS)", port)
- else:
- ports = listen_tcp(
- bind_addresses,
- port,
- SynapseSite(
- "synapse.access.http.%s" % (site_tag,),
- site_tag,
- listener_config,
- root_resource,
- self.version_string,
- ),
- reactor=self.get_reactor(),
- )
- logger.info("Synapse now listening on TCP port %d", port)
- return ports
- def _configure_named_resource(self, name, compress=False):
- """Build a resource map for a named resource
- Args:
- name (str): named resource: one of "client", "federation", etc
- compress (bool): whether to enable gzip compression for this
- resource
- Returns:
- dict[str, Resource]: map from path to HTTP resource
- """
- resources = {}
- if name == "client":
- client_resource = ClientRestResource(self)
- if compress:
- client_resource = gz_wrap(client_resource)
- resources.update(
- {
- "/_matrix/client/api/v1": client_resource,
- "/_matrix/client/r0": client_resource,
- "/_matrix/client/unstable": client_resource,
- "/_matrix/client/v2_alpha": client_resource,
- "/_matrix/client/versions": client_resource,
- "/.well-known/matrix/client": WellKnownResource(self),
- "/_synapse/admin": AdminRestResource(self),
- }
- )
- if self.get_config().saml2_enabled:
- from synapse.rest.saml2 import SAML2Resource
- resources["/_matrix/saml2"] = SAML2Resource(self)
- if name == "consent":
- from synapse.rest.consent.consent_resource import ConsentResource
- consent_resource = ConsentResource(self)
- if compress:
- consent_resource = gz_wrap(consent_resource)
- resources.update({"/_matrix/consent": consent_resource})
- if name == "federation":
- resources.update({FEDERATION_PREFIX: TransportLayerServer(self)})
- if name == "openid":
- resources.update(
- {
- FEDERATION_PREFIX: TransportLayerServer(
- self, servlet_groups=["openid"]
- )
- }
- )
- if name in ["static", "client"]:
- resources.update(
- {
- STATIC_PREFIX: File(
- os.path.join(os.path.dirname(synapse.__file__), "static")
- )
- }
- )
- if name in ["media", "federation", "client"]:
- if self.get_config().enable_media_repo:
- media_repo = self.get_media_repository_resource()
- resources.update(
- {
- MEDIA_PREFIX: media_repo,
- LEGACY_MEDIA_PREFIX: media_repo,
- CONTENT_REPO_PREFIX: ContentRepoResource(
- self, self.config.uploads_path
- ),
- }
- )
- elif name == "media":
- raise ConfigError(
- "'media' resource conflicts with enable_media_repo=False"
- )
- if name in ["keys", "federation"]:
- resources[SERVER_KEY_V2_PREFIX] = KeyApiV2Resource(self)
- if name == "webclient":
- webclient_path = self.get_config().web_client_location
- if webclient_path is None:
- logger.warning(
- "Not enabling webclient resource, as web_client_location is unset."
- )
- else:
- # GZip is disabled here due to
- # https://twistedmatrix.com/trac/ticket/7678
- resources[WEB_CLIENT_PREFIX] = File(webclient_path)
- if name == "metrics" and self.get_config().enable_metrics:
- resources[METRICS_PREFIX] = MetricsResource(RegistryProxy)
- if name == "replication":
- resources[REPLICATION_PREFIX] = ReplicationRestResource(self)
- return resources
- def start_listening(self, listeners):
- config = self.get_config()
- for listener in listeners:
- if listener["type"] == "http":
- self._listening_services.extend(self._listener_http(config, listener))
- elif listener["type"] == "manhole":
- listen_tcp(
- listener["bind_addresses"],
- listener["port"],
- manhole(
- username="matrix", password="rabbithole", globals={"hs": self}
- ),
- )
- elif listener["type"] == "replication":
- services = listen_tcp(
- listener["bind_addresses"],
- listener["port"],
- ReplicationStreamProtocolFactory(self),
- )
- for s in services:
- reactor.addSystemEventTrigger("before", "shutdown", s.stopListening)
- elif listener["type"] == "metrics":
- if not self.get_config().enable_metrics:
- logger.warn(
- (
- "Metrics listener configured, but "
- "enable_metrics is not True!"
- )
- )
- else:
- _base.listen_metrics(listener["bind_addresses"], listener["port"])
- else:
- logger.warn("Unrecognized listener type: %s", listener["type"])
- def run_startup_checks(self, db_conn, database_engine):
- all_users_native = are_all_users_on_domain(
- db_conn.cursor(), database_engine, self.hostname
- )
- if not all_users_native:
- quit_with_error(
- "Found users in database not native to %s!\n"
- "You cannot changed a synapse server_name after it's been configured"
- % (self.hostname,)
- )
- try:
- database_engine.check_database(db_conn.cursor())
- except IncorrectDatabaseSetup as e:
- quit_with_error(str(e))
- # Gauges to expose monthly active user control metrics
- current_mau_gauge = Gauge("synapse_admin_mau:current", "Current MAU")
- max_mau_gauge = Gauge("synapse_admin_mau:max", "MAU Limit")
- registered_reserved_users_mau_gauge = Gauge(
- "synapse_admin_mau:registered_reserved_users",
- "Registered users with reserved threepids",
- )
- def setup(config_options):
- """
- Args:
- config_options_options: The options passed to Synapse. Usually
- `sys.argv[1:]`.
- Returns:
- HomeServer
- """
- try:
- config = HomeServerConfig.load_or_generate_config(
- "Synapse Homeserver", config_options
- )
- except ConfigError as e:
- sys.stderr.write("\n" + str(e) + "\n")
- sys.exit(1)
- if not config:
- # If a config isn't returned, and an exception isn't raised, we're just
- # generating config files and shouldn't try to continue.
- sys.exit(0)
- synapse.config.logger.setup_logging(config, use_worker_options=False)
- events.USE_FROZEN_DICTS = config.use_frozen_dicts
- database_engine = create_engine(config.database_config)
- config.database_config["args"]["cp_openfun"] = database_engine.on_new_connection
- hs = SynapseHomeServer(
- config.server_name,
- db_config=config.database_config,
- config=config,
- version_string="Synapse/" + get_version_string(synapse),
- database_engine=database_engine,
- )
- logger.info("Preparing database: %s...", config.database_config["name"])
- try:
- with hs.get_db_conn(run_new_connection=False) as db_conn:
- prepare_database(db_conn, database_engine, config=config)
- database_engine.on_new_connection(db_conn)
- hs.run_startup_checks(db_conn, database_engine)
- db_conn.commit()
- except UpgradeDatabaseException:
- sys.stderr.write(
- "\nFailed to upgrade database.\n"
- "Have you checked for version specific instructions in"
- " UPGRADES.rst?\n"
- )
- sys.exit(1)
- logger.info("Database prepared in %s.", config.database_config["name"])
- hs.setup()
- hs.setup_master()
- @defer.inlineCallbacks
- def do_acme():
- """
- Reprovision an ACME certificate, if it's required.
- Returns:
- Deferred[bool]: Whether the cert has been updated.
- """
- acme = hs.get_acme_handler()
- # Check how long the certificate is active for.
- cert_days_remaining = hs.config.is_disk_cert_valid(allow_self_signed=False)
- # We want to reprovision if cert_days_remaining is None (meaning no
- # certificate exists), or the days remaining number it returns
- # is less than our re-registration threshold.
- provision = False
- if (
- cert_days_remaining is None
- or cert_days_remaining < hs.config.acme_reprovision_threshold
- ):
- provision = True
- if provision:
- yield acme.provision_certificate()
- return provision
- @defer.inlineCallbacks
- def reprovision_acme():
- """
- Provision a certificate from ACME, if required, and reload the TLS
- certificate if it's renewed.
- """
- reprovisioned = yield do_acme()
- if reprovisioned:
- _base.refresh_certificate(hs)
- @defer.inlineCallbacks
- def start():
- try:
- # Run the ACME provisioning code, if it's enabled.
- if hs.config.acme_enabled:
- acme = hs.get_acme_handler()
- # Start up the webservices which we will respond to ACME
- # challenges with, and then provision.
- yield acme.start_listening()
- yield do_acme()
- # Check if it needs to be reprovisioned every day.
- hs.get_clock().looping_call(reprovision_acme, 24 * 60 * 60 * 1000)
- _base.start(hs, config.listeners)
- hs.get_pusherpool().start()
- hs.get_datastore().start_doing_background_updates()
- except Exception:
- # Print the exception and bail out.
- print("Error during startup:", file=sys.stderr)
- # this gives better tracebacks than traceback.print_exc()
- Failure().printTraceback(file=sys.stderr)
- if reactor.running:
- reactor.stop()
- sys.exit(1)
- reactor.addSystemEventTrigger("before", "startup", start)
- return hs
- class SynapseService(service.Service):
- """
- A twisted Service class that will start synapse. Used to run synapse
- via twistd and a .tac.
- """
- def __init__(self, config):
- self.config = config
- def startService(self):
- hs = setup(self.config)
- change_resource_limit(hs.config.soft_file_limit)
- if hs.config.gc_thresholds:
- gc.set_threshold(*hs.config.gc_thresholds)
- def stopService(self):
- return self._port.stopListening()
- def run(hs):
- PROFILE_SYNAPSE = False
- if PROFILE_SYNAPSE:
- def profile(func):
- from cProfile import Profile
- from threading import current_thread
- def profiled(*args, **kargs):
- profile = Profile()
- profile.enable()
- func(*args, **kargs)
- profile.disable()
- ident = current_thread().ident
- profile.dump_stats(
- "/tmp/%s.%s.%i.pstat" % (hs.hostname, func.__name__, ident)
- )
- return profiled
- from twisted.python.threadpool import ThreadPool
- ThreadPool._worker = profile(ThreadPool._worker)
- reactor.run = profile(reactor.run)
- clock = hs.get_clock()
- start_time = clock.time()
- stats = {}
- # Contains the list of processes we will be monitoring
- # currently either 0 or 1
- stats_process = []
- def start_phone_stats_home():
- return run_as_background_process("phone_stats_home", phone_stats_home)
- @defer.inlineCallbacks
- def phone_stats_home():
- logger.info("Gathering stats for reporting")
- now = int(hs.get_clock().time())
- uptime = int(now - start_time)
- if uptime < 0:
- uptime = 0
- stats["homeserver"] = hs.config.server_name
- stats["server_context"] = hs.config.server_context
- stats["timestamp"] = now
- stats["uptime_seconds"] = uptime
- version = sys.version_info
- stats["python_version"] = "{}.{}.{}".format(
- version.major, version.minor, version.micro
- )
- stats["total_users"] = yield hs.get_datastore().count_all_users()
- total_nonbridged_users = yield hs.get_datastore().count_nonbridged_users()
- stats["total_nonbridged_users"] = total_nonbridged_users
- daily_user_type_results = yield hs.get_datastore().count_daily_user_type()
- for name, count in iteritems(daily_user_type_results):
- stats["daily_user_type_" + name] = count
- room_count = yield hs.get_datastore().get_room_count()
- stats["total_room_count"] = room_count
- stats["daily_active_users"] = yield hs.get_datastore().count_daily_users()
- stats["monthly_active_users"] = yield hs.get_datastore().count_monthly_users()
- stats[
- "daily_active_rooms"
- ] = yield hs.get_datastore().count_daily_active_rooms()
- stats["daily_messages"] = yield hs.get_datastore().count_daily_messages()
- r30_results = yield hs.get_datastore().count_r30_users()
- for name, count in iteritems(r30_results):
- stats["r30_users_" + name] = count
- daily_sent_messages = yield hs.get_datastore().count_daily_sent_messages()
- stats["daily_sent_messages"] = daily_sent_messages
- stats["cache_factor"] = CACHE_SIZE_FACTOR
- stats["event_cache_size"] = hs.config.event_cache_size
- if len(stats_process) > 0:
- stats["memory_rss"] = 0
- stats["cpu_average"] = 0
- for process in stats_process:
- stats["memory_rss"] += process.memory_info().rss
- stats["cpu_average"] += int(process.cpu_percent(interval=None))
- stats["database_engine"] = hs.get_datastore().database_engine_name
- stats["database_server_version"] = hs.get_datastore().get_server_version()
- logger.info("Reporting stats to matrix.org: %s" % (stats,))
- try:
- yield hs.get_simple_http_client().put_json(
- "https://matrix.org/report-usage-stats/push", stats
- )
- except Exception as e:
- logger.warn("Error reporting stats: %s", e)
- def performance_stats_init():
- try:
- process = psutil.Process()
- # Ensure we can fetch both, and make the initial request for cpu_percent
- # so the next request will use this as the initial point.
- process.memory_info().rss
- process.cpu_percent(interval=None)
- logger.info("report_stats can use psutil")
- stats_process.append(process)
- except (AttributeError):
- logger.warning("Unable to read memory/cpu stats. Disabling reporting.")
- def generate_user_daily_visit_stats():
- return run_as_background_process(
- "generate_user_daily_visits", hs.get_datastore().generate_user_daily_visits
- )
- # Rather than update on per session basis, batch up the requests.
- # If you increase the loop period, the accuracy of user_daily_visits
- # table will decrease
- clock.looping_call(generate_user_daily_visit_stats, 5 * 60 * 1000)
- # monthly active user limiting functionality
- def reap_monthly_active_users():
- return run_as_background_process(
- "reap_monthly_active_users", hs.get_datastore().reap_monthly_active_users
- )
- clock.looping_call(reap_monthly_active_users, 1000 * 60 * 60)
- reap_monthly_active_users()
- @defer.inlineCallbacks
- def generate_monthly_active_users():
- current_mau_count = 0
- reserved_count = 0
- store = hs.get_datastore()
- if hs.config.limit_usage_by_mau or hs.config.mau_stats_only:
- current_mau_count = yield store.get_monthly_active_count()
- reserved_count = yield store.get_registered_reserved_users_count()
- current_mau_gauge.set(float(current_mau_count))
- registered_reserved_users_mau_gauge.set(float(reserved_count))
- max_mau_gauge.set(float(hs.config.max_mau_value))
- def start_generate_monthly_active_users():
- return run_as_background_process(
- "generate_monthly_active_users", generate_monthly_active_users
- )
- start_generate_monthly_active_users()
- if hs.config.limit_usage_by_mau or hs.config.mau_stats_only:
- clock.looping_call(start_generate_monthly_active_users, 5 * 60 * 1000)
- # End of monthly active user settings
- if hs.config.report_stats:
- logger.info("Scheduling stats reporting for 3 hour intervals")
- clock.looping_call(start_phone_stats_home, 3 * 60 * 60 * 1000)
- # We need to defer this init for the cases that we daemonize
- # otherwise the process ID we get is that of the non-daemon process
- clock.call_later(0, performance_stats_init)
- # We wait 5 minutes to send the first set of stats as the server can
- # be quite busy the first few minutes
- clock.call_later(5 * 60, start_phone_stats_home)
- _base.start_reactor(
- "synapse-homeserver",
- soft_file_limit=hs.config.soft_file_limit,
- gc_thresholds=hs.config.gc_thresholds,
- pid_file=hs.config.pid_file,
- daemonize=hs.config.daemonize,
- print_pidfile=hs.config.print_pidfile,
- logger=logger,
- )
- def main():
- with LoggingContext("main"):
- # check base requirements
- check_requirements()
- hs = setup(sys.argv[1:])
- run(hs)
- if __name__ == "__main__":
- main()
|