python tornado TCPserver异步协程实例

文章目录

[隐藏]

  • 项目所用知识点
  • tornado tcpserver源码抛析
  • TCPServer执行过程
  • 代码实例

内容预览:

  • 它的值包括四种:AF_UNIX,AF_INET,AF_INET6和AF_UNSPEC~
  • # 相反,我们使用listen backlog作为我们可以合理接受的连接数的~
  • if errno_from_exception(e) == errno.ECONNABORTED: continue raise ca…~
项目所用知识点
  • tornado
  • socket
  • tcpserver
  • 协程
  • 异步
tornado tcpserver源码抛析

在tornado的tcpserver文件中,实现了TCPServer这个类,他是一个单线程的,非阻塞的tcp 服务。

为了与上层协议(在tornado中就是HTTPServer)交互,TCPServer提供了一个接口:handle_stream, 要求其子类必需实现该方法,该方法就是主要用来处理应用层逻辑的。

我们可以通过下面代码倒入模块查看源码

from tornado.tcpserver import TCPServer  

源码中给了好多解释,先把源码注释贴进来

class TCPServer(object):        ‘’‘        1.要想用TCPserver,我给你提供你一个接口handle_stream,子类中必须要有这个方法        2.他已经给我们举出了例子        3.往下他给咱们介绍了几种启动方法,而我用的listen()方法启动看起来简单明了        ’‘’        r"""A non-blocking, single-threaded TCP server.            To use `TCPServer`, define a subclass which overrides the `handle_stream`        method. For example, a simple echo server could be defined like this::              from tornado.tcpserver import TCPServer          from tornado.iostream import StreamClosedError          from tornado import gen              class EchoServer(TCPServer):              @gen.coroutine              def handle_stream(self, stream, address):                  while True:                      try:                          data = yield stream.read_until(b"n")                          yield stream.write(data)                      except StreamClosedError:                          break            To make this server serve SSL traffic, send the ``ssl_options`` keyword        argument with an `ssl.SSLContext` object. For compatibility with older        versions of Python ``ssl_options`` may also be a dictionary of keyword        arguments for the `ssl.wrap_socket` method.::               ssl_ctx = ssl.create_default_context(ssl.Purpose.CLIENT_AUTH)           ssl_ctx.load_cert_chain(os.path.join(data_dir, "mydomain.crt"),                                   os.path.join(data_dir, "mydomain.key"))           TCPServer(ssl_options=ssl_ctx)            `TCPServer` initialization follows one of three patterns:            1. `listen`: simple single-process::                    server = TCPServer()                server.listen(8888)                IOLoop.current().start()            2. `bind`/`start`: simple multi-process::                    server = TCPServer()                server.bind(8888)                server.start(0)  # Forks multiple sub-processes                IOLoop.current().start()               When using this interface, an `.IOLoop` must *not* be passed           to the `TCPServer` constructor.  `start` will always start           the server on the default singleton `.IOLoop`.            3. `add_sockets`: advanced multi-process::                    sockets = bind_sockets(8888)                tornado.process.fork_processes(0)                server = TCPServer()                server.add_sockets(sockets)                IOLoop.current().start()               The `add_sockets` interface is more complicated, but it can be           used with `tornado.process.fork_processes` to give you more           flexibility in when the fork happens.  `add_sockets` can           also be used in single-process servers if you want to create           your listening sockets in some way other than           `~tornado.netutil.bind_sockets`.            .. versionadded:: 3.1           The ``max_buffer_size`` argument.            .. versionchanged:: 5.0           The ``io_loop`` argument has been removed.        """  

自己仔细看该类的其他方法

    def listen(self, port, address=""):            """Starts accepting connections on the given port.                This method may be called more than once to listen on multiple ports.            `listen` takes effect immediately; it is not necessary to call            `TCPServer.start` afterwards.  It is, however, necessary to start            the `.IOLoop`.            """            sockets = bind_sockets(port, address=address)            self.add_sockets(sockets)            def add_sockets(self, sockets):            """Makes this server start accepting connections on the given sockets.                The ``sockets`` parameter is a list of socket objects such as            those returned by `~tornado.netutil.bind_sockets`.            `add_sockets` is typically used in combination with that            method and `tornado.process.fork_processes` to provide greater            control over the initialization of a multi-process server.            """            for sock in sockets:                self._sockets[sock.fileno()] = sock                self._handlers[sock.fileno()] = add_accept_handler(                    sock, self._handle_connection)            def add_socket(self, socket):            """Singular version of `add_sockets`.  Takes a single socket object."""            self.add_sockets([socket])            def bind(self, port, address=None, family=socket.AF_UNSPEC, backlog=128,                 reuse_port=False):            """Binds this server to the given port on the given address.                To start the server, call `start`. If you want to run this server            in a single process, you can call `listen` as a shortcut to the            sequence of `bind` and `start` calls.                Address may be either an IP address or hostname.  If it's a hostname,            the server will listen on all IP addresses associated with the            name.  Address may be an empty string or None to listen on all            available interfaces.  Family may be set to either `socket.AF_INET`            or `socket.AF_INET6` to restrict to IPv4 or IPv6 addresses, otherwise            both will be used if available.                The ``backlog`` argument has the same meaning as for            `socket.listen <socket.socket.listen>`. The ``reuse_port`` argument            has the same meaning as for `.bind_sockets`.                This method may be called multiple times prior to `start` to listen            on multiple ports or interfaces.                .. versionchanged:: 4.4               Added the ``reuse_port`` argument.            """            sockets = bind_sockets(port, address=address, family=family,                                   backlog=backlog, reuse_port=reuse_port)            if self._started:                self.add_sockets(sockets)            else:                self._pending_sockets.extend(sockets)            def start(self, num_processes=1):            """Starts this server in the `.IOLoop`.                By default, we run the server in this process and do not fork any            additional child process.                If num_processes is ``None`` or <= 0, we detect the number of cores            available on this machine and fork that number of child            processes. If num_processes is given and > 1, we fork that            specific number of sub-processes.                Since we use processes and not threads, there is no shared memory            between any server code.                Note that multiple processes are not compatible with the autoreload            module (or the ``autoreload=True`` option to `tornado.web.Application`            which defaults to True when ``debug=True``).            When using multiple processes, no IOLoops can be created or            referenced until after the call to ``TCPServer.start(n)``.            """            assert not self._started            self._started = True            if num_processes != 1:                process.fork_processes(num_processes)            sockets = self._pending_sockets            self._pending_sockets = []            self.add_sockets(sockets)            def stop(self):            """Stops listening for new connections.                Requests currently in progress may still continue after the            server is stopped.            """            if self._stopped:                return            self._stopped = True            for fd, sock in self._sockets.items():                assert sock.fileno() == fd                # Unregister socket from IOLoop                self._handlers.pop(fd)()                sock.close()            def handle_stream(self, stream, address):            """Override to handle a new `.IOStream` from an incoming connection.                This method may be a coroutine; if so any exceptions it raises            asynchronously will be logged. Accepting of incoming connections            will not be blocked by this coroutine.                If this `TCPServer` is configured for SSL, ``handle_stream``            may be called before the SSL handshake has completed. Use            `.SSLIOStream.wait_for_handshake` if you need to verify the client's            certificate or use NPN/ALPN.                .. versionchanged:: 4.2               Added the option for this method to be a coroutine.            """            raise NotImplementedError()            def _handle_connection(self, connection, address):            if self.ssl_options is not None:                assert ssl, "Python 2.6+ and OpenSSL required for SSL"                try:                    connection = ssl_wrap_socket(connection,                                                 self.ssl_options,                                                 server_side=True,                                                 do_handshake_on_connect=False)                except ssl.SSLError as err:                    if err.args[0] == ssl.SSL_ERROR_EOF:                        return connection.close()                    else:                        raise                except socket.error as err:                    # If the connection is closed immediately after it is created                    # (as in a port scan), we can get one of several errors.                    # wrap_socket makes an internal call to getpeername,                    # which may return either EINVAL (Mac OS X) or ENOTCONN                    # (Linux).  If it returns ENOTCONN, this error is                    # silently swallowed by the ssl module, so we need to                    # catch another error later on (AttributeError in                    # SSLIOStream._do_ssl_handshake).                    # To test this behavior, try nmap with the -sT flag.                    # https://github.com/tornadoweb/tornado/pull/750                    if errno_from_exception(err) in (errno.ECONNABORTED, errno.EINVAL):                        return connection.close()                    else:                        raise            try:                if self.ssl_options is not None:                    stream = SSLIOStream(connection,                                         max_buffer_size=self.max_buffer_size,                                         read_chunk_size=self.read_chunk_size)                else:                    stream = IOStream(connection,                                      max_buffer_size=self.max_buffer_size,                                      read_chunk_size=self.read_chunk_size)                    future = self.handle_stream(stream, address)                if future is not None:                    IOLoop.current().add_future(gen.convert_yielded(future),                                                lambda f: f.result())            except Exception:                app_log.error("Error in connection callback", exc_info=True)  

通过方法名就能看出来,而且开头已经给出实例怎么去用,所以这个就不一一解释了,我自己的用法如下

from tornado.tcpserver import TCPServer    from tornado.iostream import IOStream, StreamClosedError    from tornado import gen    from tornado.ioloop import IOLoop    import struct        class ProxyServer(TCPServer):        def __init__(self, *args, **kwargs):            super(ProxyServer, self).__init__(*args, **kwargs)            self.devices = dict()            @gen.coroutine        def handle_stream(self, stream, address):              pass            if __name__ == "__main__":        server = ProxyServer()        server.listen(1234)        IOLoop.current().start()  

具体步骤来分析 一下

TCPServer执行过程

1.server = ProxyServer()创建tcpserver对象,该步骤仅仅做了一个初始化操作

def __init__(self, io_loop=None, ssl_options=None, max_buffer_size=None, read_chunk_size=None):            self.io_loop = io_loop            self.ssl_options = ssl_options            self._sockets = {}  # fd -> socket object    用来存储文件描述符与socket对象的映射关系            self._pending_sockets = []            self._started = False            self.max_buffer_size = max_buffer_size    # 最大缓冲长度            self.read_chunk_size = read_chunk_size    # 每次读的chunk大小                # 校验ssl选项.            if self.ssl_options is not None and isinstance(self.ssl_options, dict):                if 'certfile' not in self.ssl_options:                    raise KeyError('missing key "certfile" in ssl_options')                    if not os.path.exists(self.ssl_options['certfile']):                    raise ValueError('certfile "%s" does not exist' % self.ssl_options['certfile'])                if ('keyfile' in self.ssl_options and not os.path.exists(self.ssl_options['keyfile'])):                    raise ValueError('keyfile "%s" does not exist' % self.ssl_options['keyfile'])  

2.想都不要想肯定是开启socket

步骤是执行server.listen(1234)的时候,

    def listen(self, port, address=""):            """Starts accepting connections on the given port.                This method may be called more than once to listen on multiple ports.            `listen` takes effect immediately; it is not necessary to call            `TCPServer.start` afterwards.  It is, however, necessary to start            the `.IOLoop`.            """            sockets = bind_sockets(port, address=address)            self.add_sockets(sockets)  

3.看下listen里面有调用bind_sockets()方法,来看下该方法

def bind_sockets(port, address=None, family=socket.AF_UNSPEC, backlog=_DEFAULT_BACKLOG, flags=None, reuse_port=False):        if reuse_port and not hasattr(socket, "SO_REUSEPORT"):            raise ValueError("the platform doesn't support SO_REUSEPORT")            sockets = []        if address == "":            address = None        # address family参数指定调用者期待返回的套接口地址结构的类型。它的值包括四种:AF_UNIX,AF_INET,AF_INET6和AF_UNSPEC。        # AF_UNIX用于同一台机器上的进程间通信        # 如果指定AF_INET,那么函数就不能返回任何IPV6相关的地址信息;如果仅指定了AF_INET6,则就不能返回任何IPV4地址信息。        # AF_UNSPEC则意味着函数返回的是适用于指定主机名和服务名且适合任何协议族的地址。        # 如果某个主机既有AAAA记录(IPV6)地址,同时又有A记录(IPV4)地址,那么AAAA记录将作为sockaddr_in6结构返回,而A记录则作为sockaddr_in结构返回        if not socket.has_ipv6 and family == socket.AF_UNSPEC: # 如果系统不支持ipv6            family = socket.AF_INET        if flags is None:            flags = socket.AI_PASSIVE        bound_port = None        for res in set(socket.getaddrinfo(address, port, family, socket.SOCK_STREAM, 0, flags)):            af, socktype, proto, canonname, sockaddr = res            if (sys.platform == 'darwin' and address == 'localhost' and af == socket.AF_INET6 and sockaddr[3] != 0):                # Mac OS X在“localhost”的getaddrinfo结果中包含一个链接本地地址fe80 :: 1%lo0。                # 但是,防火墙不了解这是一个本地地址,并且会提示访问。 所以跳过这些地址。                continue            try:                sock = socket.socket(af, socktype, proto)            except socket.error as e:                # 如果协议不支持该地址                if errno_from_exception(e) == errno.EAFNOSUPPORT:                    continue                raise            # 为 fd 设置 FD_CLOEXEC 标识            set_close_exec(sock.fileno())            if os.name != 'nt': # 非windows                sock.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)            if reuse_port:                sock.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEPORT, 1)            if af == socket.AF_INET6:                # On linux, ipv6 sockets accept ipv4 too by default,                # but this makes it impossible to bind to both                # 0.0.0.0 in ipv4 and :: in ipv6.  On other systems,                # separate sockets *must* be used to listen for both ipv4                # and ipv6.  For consistency, always disable ipv4 on our                # ipv6 sockets and use a separate ipv4 socket when needed.                #                # Python 2.x on windows doesn't have IPPROTO_IPV6.                if hasattr(socket, "IPPROTO_IPV6"):                    sock.setsockopt(socket.IPPROTO_IPV6, socket.IPV6_V6ONLY, 1)                # 自动端口分配,端口=None            # 应该绑定在IPv4和IPv6地址上的同一个端口上            host, requested_port = sockaddr[:2]            if requested_port == 0 and bound_port is not None:                sockaddr = tuple([host, bound_port] + list(sockaddr[2:]))            # 设置socket为非阻塞            sock.setblocking(0)            sock.bind(sockaddr)            bound_port = sock.getsockname()[1]            sock.listen(backlog)            sockets.append(sock)        return sockets  

4.接下来执行的是add_sockets()方法

def add_sockets(self, sockets):            if self.io_loop is None:                self.io_loop = IOLoop.current()    # 获取IOLoop实例对象                for sock in sockets:                self._sockets[sock.fileno()] = sock                add_accept_handler(sock, self._handle_connection, io_loop=self.io_loop)  

可以看到里面调用了add_accept_handler方法,来我们进去看看该方法干啥了

5.探析add_accept_handler方法

def add_accept_handler(sock, callback, io_loop=None):        if io_loop is None: # 获取IOLoop实例对象            io_loop = IOLoop.current()            def accept_handler(fd, events):            # 我们处理回调时可能会有许多的连接等待建立; 为了防止其他任务的饥饿,我们必须限制我们一次接受的连接数。            # 理想情况下,我们接受在处理回调过程中等待的连接数,但此可能会对负载产生不利影响。            # 相反,我们使用listen backlog作为我们可以合理接受的连接数的。            for i in xrange(_DEFAULT_BACKLOG): # _DEFAULT_BACKLOG默认为128                try:                    connection, address = sock.accept()                except socket.error as e:                    # _ERRNO_WOULDBLOCK 与EAGAIN相同,表示再尝试一下,很多情况下是因为资源不足,或者条件未达成                    # 当某个子进程与客户端建立了连接,其他子进程再次尝试与该客户端建立连接时就会产生该错误                    if errno_from_exception(e) in _ERRNO_WOULDBLOCK:                        return                    # ECONNABORTED表示有一个连接,在他处于等待被服务端accept的时候主动关闭了。                    if errno_from_exception(e) == errno.ECONNABORTED:                        continue                    raise                callback(connection, address)        io_loop.add_handler(sock, accept_handler, IOLoop.READ) # 为socket注册handler:当发生READ事件时运行accept_handler函数。  

欣欣然我们来到了最后一步

6.IOLoop.current().start(),然我们看下源码

def start(self):            try:                while True:                    callbacks = self._callbacks                    self._callbacks = []                    due_timeouts = []                    # 将时间已到的定时任务放置到due_timeouts中,过程省略                    for callback in callbacks:          # 执行callback                        self._run_callback(callback)                    for timeout in due_timeouts:        # 执行定时任务                        if timeout.callback is not None:                            self._run_callback(timeout.callback)                    callbacks = callback = due_timeouts = timeout = None    # 释放内存                    # 根据情况设置poll_timeout的值,过程省略                    if not self._running:    # 终止ioloop运行时,在执行完了callback后结束循环                        breaktry:                        event_pairs = self._impl.poll(poll_timeout)                    except Exception as e:                        if errno_from_exception(e) == errno.EINTR:  # 系统调用被信号处理函数中断,进行下一次循环                            continue                        else:                            raise                    self._events.update(event_pairs)                    while self._events:                        fd, events = self._events.popitem()             # 获取一个fd以及对应事件                        try:                            fd_obj, handler_func = self._handlers[fd]   # 获取该fd对应的事件处理函数                            handler_func(fd_obj, events)                # 运行该事件处理函数                        except (OSError, IOError) as e:                            if errno_from_exception(e) == errno.EPIPE:     # 当客户端关闭连接时会产生EPIPE错误                                pass                            # 其他异常处理已经省略                    fd_obj = handler_func = None       # 释放内存空间  

这一步想了解更多去参考这篇文章http://www.cnblogs.com/MnCu8261/p/6730691.html

代码实例

目前公司有这么一个需求,iphonex–server–ue4,面对两个客户端,达到iphonex把数据给ue4,server起一个代理作用,需求大概就是这样,具体实现代码如下

from tornado.tcpserver import TCPServer    from tornado.iostream import IOStream, StreamClosedError    from tornado import gen    from tornado.ioloop import IOLoop    import struct        class ProxyServer(TCPServer):        def __init__(self, *args, **kwargs):            super(ProxyServer, self).__init__(*args, **kwargs)            self.devices = dict()            @gen.coroutine        def handle_stream(self, stream, address):            device = yield stream.read_bytes(1)            if device == b"x0a":                self.handle_iphonex_stream(stream, address)            elif device == b"x0b":                self.handle_ue4_stream(stream, address)            else:                print("protocol error.")            @gen.coroutine        def handle_iphonex_stream(self, stream, address):            yield stream.write(b"x00")            print("iphonex")                # uuid            rlen = yield stream.read_bytes(4)            rlen = struct.unpack(">I", rlen)[0]            uuid = yield stream.read_bytes(rlen)            uuid = uuid.decode()            yield stream.write(b"x00")            print(uuid)                # keys            rlen = yield stream.read_bytes(4)            rlen = struct.unpack(">I", rlen)[0]            keys = yield stream.read_bytes(rlen)            keys = keys.decode()            yield stream.write(b"x00")            print(keys)                # save            self.devices[uuid] = {'keys': keys}                # data            keys = keys.split(',')            fmt = "%df" % len(keys)                while True:                try:                    data = yield stream.read_bytes(struct.calcsize(fmt))                except StreamClosedError:                    print 'iphonex is closed'                    break                pdata = struct.unpack(fmt, data)                print(pdata)                    ue4stream = self.devices[uuid].get('ue4')                if ue4stream:                    try:                        yield ue4stream.write(data)                    except Exception as e:                        self.devices[uuid]['ue4'] = None                        print('request for %s closed' % uuid)            @gen.coroutine        def handle_ue4_stream(self, stream, address):            yield stream.write(b"x00")            print("ue4")                # uuid            rlen = yield stream.read_bytes(4)            rlen = struct.unpack(">I", rlen)[0]            uuid = yield stream.read_bytes(rlen)            uuid = uuid.decode()            print(uuid)                if self.devices.get(uuid):                yield stream.write(b"x00")            else:                yield stream.write(b"x01")                raise Exception                # send keys            keys = self.devices[uuid].get('keys')            stream.write(struct.pack(">I", len(keys)))            stream.write(keys.encode())                valid = yield stream.read_bytes(1)            if valid == b'x  

原文出处:01hai -> https://www.01hai.com/note/av125426

本站所发布的一切资源仅限用于学习和研究目的;不得将上述内容用于商业或者非法用途,否则,一切后果请用户自负。本站信息来自网络,版权争议与本站无关。您必须在下载后的24个小时之内,从您的电脑中彻底删除上述内容。如果您喜欢该程序,请支持正版软件,购买注册,得到更好的正版服务。如果侵犯你的利益,请发送邮箱到 [email protected],我们会很快的为您处理。