[Zodb-checkins] CVS: Packages/ZEO/zrpc - trigger.py:1.14.20.2

Tim Peters tim.one at comcast.net
Mon Aug 1 17:22:25 EDT 2005


Update of /cvs-repository/Packages/ZEO/zrpc
In directory cvs.zope.org:/tmp/cvs-serv30839/ZEO/zrpc

Modified Files:
      Tag: Zope-2_7-branch
	trigger.py 
Log Message:
Replace trigger.py with ZODB 3.4's version.

This simplifies the code, and worms around suspected race bugs in
Microsoft's socket implementation.


=== Packages/ZEO/zrpc/trigger.py 1.14.20.1 => 1.14.20.2 ===
--- Packages/ZEO/zrpc/trigger.py:1.14.20.1	Tue Sep 30 23:03:10 2003
+++ Packages/ZEO/zrpc/trigger.py	Mon Aug  1 17:21:54 2005
@@ -1,10 +1,10 @@
 ##############################################################################
 #
-# Copyright (c) 2001, 2002 Zope Corporation and Contributors.
+# Copyright (c) 2001-2005 Zope Corporation and Contributors.
 # All Rights Reserved.
 #
 # This software is subject to the provisions of the Zope Public License,
-# Version 2.0 (ZPL).  A copy of the ZPL should accompany this distribution.
+# Version 2.1 (ZPL).  A copy of the ZPL should accompany this distribution.
 # THIS SOFTWARE IS PROVIDED "AS IS" AND ANY AND ALL EXPRESS OR IMPLIED
 # WARRANTIES ARE DISCLAIMED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
 # WARRANTIES OF TITLE, MERCHANTABILITY, AGAINST INFRINGEMENT, AND FITNESS
@@ -18,200 +18,202 @@
 import thread
 import errno
 
-if os.name == 'posix':
-
-    class trigger(asyncore.file_dispatcher):
+from ZODB.utils import positive_id
 
-        "Wake up a call to select() running in the main thread"
+# Original comments follow; they're hard to follow in the context of
+# ZEO's use of triggers.  TODO:  rewrite from a ZEO perspective.
 
-        # This is useful in a context where you are using Medusa's I/O
-        # subsystem to deliver data, but the data is generated by another
-        # thread.  Normally, if Medusa is in the middle of a call to
-        # select(), new output data generated by another thread will have
-        # to sit until the call to select() either times out or returns.
-        # If the trigger is 'pulled' by another thread, it should immediately
-        # generate a READ event on the trigger object, which will force the
-        # select() invocation to return.
-
-        # A common use for this facility: letting Medusa manage I/O for a
-        # large number of connections; but routing each request through a
-        # thread chosen from a fixed-size thread pool.  When a thread is
-        # acquired, a transaction is performed, but output data is
-        # accumulated into buffers that will be emptied more efficiently
-        # by Medusa. [picture a server that can process database queries
-        # rapidly, but doesn't want to tie up threads waiting to send data
-        # to low-bandwidth connections]
-
-        # The other major feature provided by this class is the ability to
-        # move work back into the main thread: if you call pull_trigger()
-        # with a thunk argument, when select() wakes up and receives the
-        # event it will call your thunk from within that thread.  The main
-        # purpose of this is to remove the need to wrap thread locks around
-        # Medusa's data structures, which normally do not need them.  [To see
-        # why this is true, imagine this scenario: A thread tries to push some
-        # new data onto a channel's outgoing data queue at the same time that
-        # the main thread is trying to remove some]
+# Wake up a call to select() running in the main thread.
+#
+# This is useful in a context where you are using Medusa's I/O
+# subsystem to deliver data, but the data is generated by another
+# thread.  Normally, if Medusa is in the middle of a call to
+# select(), new output data generated by another thread will have
+# to sit until the call to select() either times out or returns.
+# If the trigger is 'pulled' by another thread, it should immediately
+# generate a READ event on the trigger object, which will force the
+# select() invocation to return.
+#
+# A common use for this facility: letting Medusa manage I/O for a
+# large number of connections; but routing each request through a
+# thread chosen from a fixed-size thread pool.  When a thread is
+# acquired, a transaction is performed, but output data is
+# accumulated into buffers that will be emptied more efficiently
+# by Medusa. [picture a server that can process database queries
+# rapidly, but doesn't want to tie up threads waiting to send data
+# to low-bandwidth connections]
+#
+# The other major feature provided by this class is the ability to
+# move work back into the main thread: if you call pull_trigger()
+# with a thunk argument, when select() wakes up and receives the
+# event it will call your thunk from within that thread.  The main
+# purpose of this is to remove the need to wrap thread locks around
+# Medusa's data structures, which normally do not need them.  [To see
+# why this is true, imagine this scenario: A thread tries to push some
+# new data onto a channel's outgoing data queue at the same time that
+# the main thread is trying to remove some]
+
+class _triggerbase(object):
+    """OS-independent base class for OS-dependent trigger class."""
+
+    kind = None  # subclass must set to "pipe" or "loopback"; used by repr
+
+    def __init__(self):
+        self._closed = False
+
+        # `lock` protects the `thunks` list from being traversed and
+        # appended to simultaneously.
+        self.lock = thread.allocate_lock()
+
+        # List of no-argument callbacks to invoke when the trigger is
+        # pulled.  These run in the thread running the asyncore mainloop,
+        # regardless of which thread pulls the trigger.
+        self.thunks = []
+
+    def readable(self):
+        return 1
+
+    def writable(self):
+        return 0
+
+    def handle_connect(self):
+        pass
+
+    def handle_close(self):
+        self.close()
+
+    # Override the asyncore close() method, because it doesn't know about
+    # (so can't close) all the gimmicks we have open.  Subclass must
+    # supply a _close() method to do platform-specific closing work.  _close()
+    # will be called iff we're not already closed.
+    def close(self):
+        if not self._closed:
+            self._closed = True
+            self.del_channel()
+            self._close()  # subclass does OS-specific stuff
 
-        def __init__(self):
-            r, w = self._fds = os.pipe()
-            self.trigger = w
-            asyncore.file_dispatcher.__init__(self, r)
-            self.lock = thread.allocate_lock()
-            self.thunks = []
-            self._closed = 0
+    def _close(self):    # see close() above; subclass must supply
+        raise NotImplementedError
 
-        # Override the asyncore close() method, because it seems that
-        # it would only close the r file descriptor and not w.  The
-        # constructor calls file_dispatcher.__init__ and passes r,
-        # which would get stored in a file_wrapper and get closed by
-        # the default close.  But that would leave w open...
-
-        def close(self):
-            if not self._closed:
-                self._closed = 1
-                self.del_channel()
-                for fd in self._fds:
-                    os.close(fd)
-                self._fds = []
-
-        def __repr__(self):
-            return '<select-trigger (pipe) at %x>' % id(self)
-
-        def readable(self):
-            return 1
-
-        def writable(self):
-            return 0
-
-        def handle_connect(self):
-            pass
-
-        def handle_close(self):
-            self.close()
-
-        def pull_trigger(self, thunk=None):
-            if thunk:
-                self.lock.acquire()
-                try:
-                    self.thunks.append(thunk)
-                finally:
-                    self.lock.release()
-            os.write(self.trigger, 'x')
-
-        def handle_read(self):
-            try:
-                self.recv(8192)
-            except socket.error:
-                return
+    def pull_trigger(self, thunk=None):
+        if thunk:
             self.lock.acquire()
             try:
-                for thunk in self.thunks:
-                    try:
-                        thunk()
-                    except:
-                        nil, t, v, tbinfo = asyncore.compact_traceback()
-                        print ('exception in trigger thunk:'
-                               ' (%s:%s %s)' % (t, v, tbinfo))
-                self.thunks = []
+                self.thunks.append(thunk)
             finally:
                 self.lock.release()
+        self._physical_pull()
 
-else:
+    # Subclass must supply _physical_pull, which does whatever the OS
+    # needs to do to provoke the "write" end of the trigger.
+    def _physical_pull(self):
+        raise NotImplementedError
+
+    def handle_read(self):
+        try:
+            self.recv(8192)
+        except socket.error:
+            return
+        self.lock.acquire()
+        try:
+            for thunk in self.thunks:
+                try:
+                    thunk()
+                except:
+                    nil, t, v, tbinfo = asyncore.compact_traceback()
+                    print ('exception in trigger thunk:'
+                           ' (%s:%s %s)' % (t, v, tbinfo))
+            self.thunks = []
+        finally:
+            self.lock.release()
 
-    # XXX Should define a base class that has the common methods and
-    # then put the platform-specific in a subclass named trigger.
+    def __repr__(self):
+        return '<select-trigger (%s) at %x>' % (self.kind, positive_id(self))
 
-    # win32-safe version
+if os.name == 'posix':
 
-    HOST = '127.0.0.1'
-    MINPORT = 19950
-    NPORTS = 50
+    class trigger(_triggerbase, asyncore.file_dispatcher):
+        kind = "pipe"
 
-    class trigger(asyncore.dispatcher):
+        def __init__(self):
+            _triggerbase.__init__(self)
+            r, self.trigger = self._fds = os.pipe()
+            asyncore.file_dispatcher.__init__(self, r)
 
-        portoffset = 0
+        def _close(self):
+            for fd in self._fds:
+                os.close(fd)
+            self._fds = []
 
-        def __init__(self):
-            a = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
-            w = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
+        def _physical_pull(self):
+            os.write(self.trigger, 'x')
 
-            # set TCP_NODELAY to true to avoid buffering
-            w.setsockopt(socket.IPPROTO_TCP, 1, 1)
+else:
+    # Windows version; uses just sockets, because a pipe isn't select'able
+    # on Windows.
 
-            # tricky: get a pair of connected sockets
-            for i in range(NPORTS):
-                trigger.portoffset = (trigger.portoffset + 1) % NPORTS
-                port = MINPORT + trigger.portoffset
-                address = (HOST, port)
-                try:
-                    a.bind(address)
-                except socket.error:
-                    continue
-                else:
-                    break
-            else:
-                raise RuntimeError, 'Cannot bind trigger!'
+    class trigger(_triggerbase, asyncore.dispatcher):
+        kind = "loopback"
 
-            a.listen(1)
-            w.setblocking(0)
-            try:
-                w.connect(address)
-            except:
-                pass
-            r, addr = a.accept()
+        def __init__(self):
+            _triggerbase.__init__(self)
+
+            # Get a pair of connected sockets.  The trigger is the 'w'
+            # end of the pair, which is connected to 'r'.  'r' is put
+            # in the asyncore socket map.  "pulling the trigger" then
+            # means writing something on w, which will wake up r.
+
+            w = socket.socket()
+            # Disable buffering -- pulling the trigger sends 1 byte,
+            # and we want that sent immediately, to wake up asyncore's
+            # select() ASAP.
+            w.setsockopt(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1)
+
+            count = 0
+            while 1:
+               count += 1
+               # Bind to a local port; for efficiency, let the OS pick
+               # a free port for us.
+               # Unfortunately, stress tests showed that we may not
+               # be able to connect to that port ("Address already in
+               # use") despite that the OS picked it.  This appears
+               # to be a race bug in the Windows socket implementation.
+               # So we loop until a connect() succeeds (almost always
+               # on the first try).  See the long thread at
+               # http://mail.zope.org/pipermail/zope/2005-July/160433.html
+               # for hideous details.
+               a = socket.socket()
+               a.bind(("127.0.0.1", 0))
+               connect_address = a.getsockname()  # assigned (host, port) pair
+               a.listen(1)
+               try:
+                   w.connect(connect_address)
+                   break    # success
+               except socket.error, detail:
+                   if detail[0] != errno.WSAEADDRINUSE:
+                       # "Address already in use" is the only error
+                       # I've seen on two WinXP Pro SP2 boxes, under
+                       # Pythons 2.3.5 and 2.4.1.
+                       raise
+                   # (10048, 'Address already in use')
+                   # assert count <= 2 # never triggered in Tim's tests
+                   if count >= 10:  # I've never seen it go above 2
+                       a.close()
+                       w.close()
+                       raise BindError("Cannot bind trigger!")
+                   # Close `a` and try again.  Note:  I originally put a short
+                   # sleep() here, but it didn't appear to help or hurt.
+                   a.close()
+
+            r, addr = a.accept()  # r becomes asyncore's (self.)socket
             a.close()
-            w.setblocking(1)
             self.trigger = w
-
             asyncore.dispatcher.__init__(self, r)
-            self.lock = thread.allocate_lock()
-            self.thunks = []
-            self._trigger_connected = 0
-            self._closed = 0
 
-        def close(self):
-            if not self._closed:
-                self._closed = 1
-                self.del_channel()
-                # self.socket is a, self.trigger is w from __init__
-                self.socket.close()
-                self.trigger.close()
-
-        def __repr__(self):
-            return '<select-trigger (loopback) at %x>' % id(self)
-
-        def readable(self):
-            return 1
-
-        def writable(self):
-            return 0
-
-        def handle_connect(self):
-            pass
-
-        def pull_trigger(self, thunk=None):
-            if thunk:
-                self.lock.acquire()
-                try:
-                    self.thunks.append(thunk)
-                finally:
-                    self.lock.release()
-            self.trigger.send('x')
+        def _close(self):
+            # self.socket is r, and self.trigger is w, from __init__
+            self.socket.close()
+            self.trigger.close()
 
-        def handle_read(self):
-            try:
-                self.recv(8192)
-            except socket.error:
-                return
-            self.lock.acquire()
-            try:
-                for thunk in self.thunks:
-                    try:
-                        thunk()
-                    except:
-                        nil, t, v, tbinfo = asyncore.compact_traceback()
-                        print ('exception in trigger thunk:'
-                               ' (%s:%s %s)' % (t, v, tbinfo))
-                self.thunks = []
-            finally:
-                self.lock.release()
+        def _physical_pull(self):
+            self.trigger.send('x')



More information about the Zodb-checkins mailing list