OpenStack Networking (Neutron)
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

67 lines
2.6 KiB

# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import threading
import eventlet
class BatchNotifier(object):
def __init__(self, batch_interval, callback):
self._pending_events = eventlet.Queue()
self.callback = callback
self.batch_interval = batch_interval
self._mutex = threading.Lock()
def queue_event(self, event):
"""Called to queue sending an event with the next batch of events.
Sending events individually, as they occur, has been problematic as it
can result in a flood of sends. Previously, there was a loopingcall
thread that would send batched events on a periodic interval. However,
maintaining a persistent thread in the loopingcall was also
This replaces the loopingcall with a mechanism that creates a
short-lived thread on demand whenever an event is queued. That thread
will check if the lock is released, send all queued events and then
sleep for 'batch_interval' seconds. If at the end of this sleep time,
other threads have added new events to the event queue, the same thread
will process them.
At the same time, other threads will be able to add new events to the
queue and will spawn new "synced_send" threads to process them. But if
the mutex is locked, the spawned thread will end immediately.
:param event: the event that occurred.
if not event:
def synced_send():
if not self._mutex.locked():
with self._mutex:
while not self._pending_events.empty():
# sleeping after send while holding the lock allows
# subsequent events to batch up
def _notify(self):
batched_events = []
while not self._pending_events.empty():