123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439 |
- import json
- import logging
- from io import BytesIO
- from six import text_type
- import attr
- from zope.interface import implementer
- from twisted.internet import address, threads, udp
- from twisted.internet._resolver import HostResolution
- from twisted.internet.address import IPv4Address
- from twisted.internet.defer import Deferred
- from twisted.internet.error import DNSLookupError
- from twisted.internet.interfaces import IReactorPluggableNameResolver
- from twisted.python.failure import Failure
- from twisted.test.proto_helpers import MemoryReactorClock
- from twisted.web.http import unquote
- from twisted.web.http_headers import Headers
- from synapse.http.site import SynapseRequest
- from synapse.util import Clock
- from tests.utils import setup_test_homeserver as _sth
- logger = logging.getLogger(__name__)
- class TimedOutException(Exception):
- """
- A web query timed out.
- """
- @attr.s
- class FakeChannel(object):
- """
- A fake Twisted Web Channel (the part that interfaces with the
- wire).
- """
- _reactor = attr.ib()
- result = attr.ib(default=attr.Factory(dict))
- _producer = None
- @property
- def json_body(self):
- if not self.result:
- raise Exception("No result yet.")
- return json.loads(self.result["body"].decode('utf8'))
- @property
- def code(self):
- if not self.result:
- raise Exception("No result yet.")
- return int(self.result["code"])
- @property
- def headers(self):
- if not self.result:
- raise Exception("No result yet.")
- h = Headers()
- for i in self.result["headers"]:
- h.addRawHeader(*i)
- return h
- def writeHeaders(self, version, code, reason, headers):
- self.result["version"] = version
- self.result["code"] = code
- self.result["reason"] = reason
- self.result["headers"] = headers
- def write(self, content):
- assert isinstance(content, bytes), "Should be bytes! " + repr(content)
- if "body" not in self.result:
- self.result["body"] = b""
- self.result["body"] += content
- def registerProducer(self, producer, streaming):
- self._producer = producer
- self.producerStreaming = streaming
- def _produce():
- if self._producer:
- self._producer.resumeProducing()
- self._reactor.callLater(0.1, _produce)
- if not streaming:
- self._reactor.callLater(0.0, _produce)
- def unregisterProducer(self):
- if self._producer is None:
- return
- self._producer = None
- def requestDone(self, _self):
- self.result["done"] = True
- def getPeer(self):
- # We give an address so that getClientIP returns a non null entry,
- # causing us to record the MAU
- return address.IPv4Address("TCP", "127.0.0.1", 3423)
- def getHost(self):
- return None
- @property
- def transport(self):
- return self
- class FakeSite:
- """
- A fake Twisted Web Site, with mocks of the extra things that
- Synapse adds.
- """
- server_version_string = b"1"
- site_tag = "test"
- @property
- def access_logger(self):
- class FakeLogger:
- def info(self, *args, **kwargs):
- pass
- return FakeLogger()
- def make_request(
- reactor,
- method,
- path,
- content=b"",
- access_token=None,
- request=SynapseRequest,
- shorthand=True,
- ):
- """
- Make a web request using the given method and path, feed it the
- content, and return the Request and the Channel underneath.
- Args:
- method (bytes/unicode): The HTTP request method ("verb").
- path (bytes/unicode): The HTTP path, suitably URL encoded (e.g.
- escaped UTF-8 & spaces and such).
- content (bytes or dict): The body of the request. JSON-encoded, if
- a dict.
- shorthand: Whether to try and be helpful and prefix the given URL
- with the usual REST API path, if it doesn't contain it.
- Returns:
- A synapse.http.site.SynapseRequest.
- """
- if not isinstance(method, bytes):
- method = method.encode('ascii')
- if not isinstance(path, bytes):
- path = path.encode('ascii')
- # Decorate it to be the full path, if we're using shorthand
- if shorthand and not path.startswith(b"/_matrix"):
- path = b"/_matrix/client/r0/" + path
- path = path.replace(b"//", b"/")
- if not path.startswith(b"/"):
- path = b"/" + path
- if isinstance(content, text_type):
- content = content.encode('utf8')
- site = FakeSite()
- channel = FakeChannel(reactor)
- req = request(site, channel)
- req.process = lambda: b""
- req.content = BytesIO(content)
- req.postpath = list(map(unquote, path[1:].split(b'/')))
- if access_token:
- req.requestHeaders.addRawHeader(
- b"Authorization", b"Bearer " + access_token.encode('ascii')
- )
- if content:
- req.requestHeaders.addRawHeader(b"Content-Type", b"application/json")
- req.requestReceived(method, path, b"1.1")
- return req, channel
- def wait_until_result(clock, request, timeout=100):
- """
- Wait until the request is finished.
- """
- clock.run()
- x = 0
- while not request.finished:
- # If there's a producer, tell it to resume producing so we get content
- if request._channel._producer:
- request._channel._producer.resumeProducing()
- x += 1
- if x > timeout:
- raise TimedOutException("Timed out waiting for request to finish.")
- clock.advance(0.1)
- def render(request, resource, clock):
- request.render(resource)
- wait_until_result(clock, request)
- @implementer(IReactorPluggableNameResolver)
- class ThreadedMemoryReactorClock(MemoryReactorClock):
- """
- A MemoryReactorClock that supports callFromThread.
- """
- def __init__(self):
- self._udp = []
- self.lookups = {}
- class Resolver(object):
- def resolveHostName(
- _self,
- resolutionReceiver,
- hostName,
- portNumber=0,
- addressTypes=None,
- transportSemantics='TCP',
- ):
- resolution = HostResolution(hostName)
- resolutionReceiver.resolutionBegan(resolution)
- if hostName not in self.lookups:
- raise DNSLookupError("OH NO")
- resolutionReceiver.addressResolved(
- IPv4Address('TCP', self.lookups[hostName], portNumber)
- )
- resolutionReceiver.resolutionComplete()
- return resolution
- self.nameResolver = Resolver()
- super(ThreadedMemoryReactorClock, self).__init__()
- def listenUDP(self, port, protocol, interface='', maxPacketSize=8196):
- p = udp.Port(port, protocol, interface, maxPacketSize, self)
- p.startListening()
- self._udp.append(p)
- return p
- def callFromThread(self, callback, *args, **kwargs):
- """
- Make the callback fire in the next reactor iteration.
- """
- d = Deferred()
- d.addCallback(lambda x: callback(*args, **kwargs))
- self.callLater(0, d.callback, True)
- return d
- def setup_test_homeserver(cleanup_func, *args, **kwargs):
- """
- Set up a synchronous test server, driven by the reactor used by
- the homeserver.
- """
- d = _sth(cleanup_func, *args, **kwargs).result
- if isinstance(d, Failure):
- d.raiseException()
- # Make the thread pool synchronous.
- clock = d.get_clock()
- pool = d.get_db_pool()
- def runWithConnection(func, *args, **kwargs):
- return threads.deferToThreadPool(
- pool._reactor,
- pool.threadpool,
- pool._runWithConnection,
- func,
- *args,
- **kwargs
- )
- def runInteraction(interaction, *args, **kwargs):
- return threads.deferToThreadPool(
- pool._reactor,
- pool.threadpool,
- pool._runInteraction,
- interaction,
- *args,
- **kwargs
- )
- pool.runWithConnection = runWithConnection
- pool.runInteraction = runInteraction
- class ThreadPool:
- """
- Threadless thread pool.
- """
- def start(self):
- pass
- def stop(self):
- pass
- def callInThreadWithCallback(self, onResult, function, *args, **kwargs):
- def _(res):
- if isinstance(res, Failure):
- onResult(False, res)
- else:
- onResult(True, res)
- d = Deferred()
- d.addCallback(lambda x: function(*args, **kwargs))
- d.addBoth(_)
- clock._reactor.callLater(0, d.callback, True)
- return d
- clock.threadpool = ThreadPool()
- pool.threadpool = ThreadPool()
- pool.running = True
- return d
- def get_clock():
- clock = ThreadedMemoryReactorClock()
- hs_clock = Clock(clock)
- return (clock, hs_clock)
- @attr.s(cmp=False)
- class FakeTransport(object):
- """
- A twisted.internet.interfaces.ITransport implementation which sends all its data
- straight into an IProtocol object: it exists to connect two IProtocols together.
- To use it, instantiate it with the receiving IProtocol, and then pass it to the
- sending IProtocol's makeConnection method:
- server = HTTPChannel()
- client.makeConnection(FakeTransport(server, self.reactor))
- If you want bidirectional communication, you'll need two instances.
- """
- other = attr.ib()
- """The Protocol object which will receive any data written to this transport.
- :type: twisted.internet.interfaces.IProtocol
- """
- _reactor = attr.ib()
- """Test reactor
- :type: twisted.internet.interfaces.IReactorTime
- """
- disconnecting = False
- buffer = attr.ib(default=b'')
- producer = attr.ib(default=None)
- def getPeer(self):
- return None
- def getHost(self):
- return None
- def loseConnection(self):
- self.disconnecting = True
- def abortConnection(self):
- self.disconnecting = True
- def pauseProducing(self):
- if not self.producer:
- return
- self.producer.pauseProducing()
- def resumeProducing(self):
- if not self.producer:
- return
- self.producer.resumeProducing()
- def unregisterProducer(self):
- if not self.producer:
- return
- self.producer = None
- def registerProducer(self, producer, streaming):
- self.producer = producer
- self.producerStreaming = streaming
- def _produce():
- d = self.producer.resumeProducing()
- d.addCallback(lambda x: self._reactor.callLater(0.1, _produce))
- if not streaming:
- self._reactor.callLater(0.0, _produce)
- def write(self, byt):
- self.buffer = self.buffer + byt
- def _write():
- if not self.buffer:
- # nothing to do. Don't write empty buffers: it upsets the
- # TLSMemoryBIOProtocol
- return
- if getattr(self.other, "transport") is not None:
- self.other.dataReceived(self.buffer)
- self.buffer = b""
- return
- self._reactor.callLater(0.0, _write)
- # always actually do the write asynchronously. Some protocols (notably the
- # TLSMemoryBIOProtocol) get very confused if a read comes back while they are
- # still doing a write. Doing a callLater here breaks the cycle.
- self._reactor.callLater(0.0, _write)
- def writeSequence(self, seq):
- for x in seq:
- self.write(x)
|