1185 lines
		
	
	
		
			36 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			1185 lines
		
	
	
		
			36 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| # Copyright 2013-2016 DataStax, Inc.
 | |
| #
 | |
| # Licensed under the Apache License, Version 2.0 (the "License");
 | |
| # you may not use this file except in compliance with the License.
 | |
| # You may obtain a copy of the License at
 | |
| #
 | |
| # http://www.apache.org/licenses/LICENSE-2.0
 | |
| #
 | |
| # Unless required by applicable law or agreed to in writing, software
 | |
| # distributed under the License is distributed on an "AS IS" BASIS,
 | |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| # See the License for the specific language governing permissions and
 | |
| # limitations under the License.
 | |
| 
 | |
| from __future__ import with_statement
 | |
| import calendar
 | |
| import datetime
 | |
| import random
 | |
| import six
 | |
| import uuid
 | |
| import sys
 | |
| 
 | |
| DATETIME_EPOC = datetime.datetime(1970, 1, 1)
 | |
| 
 | |
| assert sys.byteorder in ('little', 'big')
 | |
| is_little_endian = sys.byteorder == 'little'
 | |
| 
 | |
| def datetime_from_timestamp(timestamp):
 | |
|     """
 | |
|     Creates a timezone-agnostic datetime from timestamp (in seconds) in a consistent manner.
 | |
|     Works around a Windows issue with large negative timestamps (PYTHON-119),
 | |
|     and rounding differences in Python 3.4 (PYTHON-340).
 | |
| 
 | |
|     :param timestamp: a unix timestamp, in seconds
 | |
|     """
 | |
|     dt = DATETIME_EPOC + datetime.timedelta(seconds=timestamp)
 | |
|     return dt
 | |
| 
 | |
| 
 | |
| def unix_time_from_uuid1(uuid_arg):
 | |
|     """
 | |
|     Converts a version 1 :class:`uuid.UUID` to a timestamp with the same precision
 | |
|     as :meth:`time.time()` returns.  This is useful for examining the
 | |
|     results of queries returning a v1 :class:`~uuid.UUID`.
 | |
| 
 | |
|     :param uuid_arg: a version 1 :class:`~uuid.UUID`
 | |
|     """
 | |
|     return (uuid_arg.time - 0x01B21DD213814000) / 1e7
 | |
| 
 | |
| 
 | |
| def datetime_from_uuid1(uuid_arg):
 | |
|     """
 | |
|     Creates a timezone-agnostic datetime from the timestamp in the
 | |
|     specified type-1 UUID.
 | |
| 
 | |
|     :param uuid_arg: a version 1 :class:`~uuid.UUID`
 | |
|     """
 | |
|     return datetime_from_timestamp(unix_time_from_uuid1(uuid_arg))
 | |
| 
 | |
| 
 | |
| def min_uuid_from_time(timestamp):
 | |
|     """
 | |
|     Generates the minimum TimeUUID (type 1) for a given timestamp, as compared by Cassandra.
 | |
| 
 | |
|     See :func:`uuid_from_time` for argument and return types.
 | |
|     """
 | |
|     return uuid_from_time(timestamp, 0x808080808080, 0x80)  # Cassandra does byte-wise comparison; fill with min signed bytes (0x80 = -128)
 | |
| 
 | |
| 
 | |
| def max_uuid_from_time(timestamp):
 | |
|     """
 | |
|     Generates the maximum TimeUUID (type 1) for a given timestamp, as compared by Cassandra.
 | |
| 
 | |
|     See :func:`uuid_from_time` for argument and return types.
 | |
|     """
 | |
|     return uuid_from_time(timestamp, 0x7f7f7f7f7f7f, 0x3f7f)  # Max signed bytes (0x7f = 127)
 | |
| 
 | |
| 
 | |
| def uuid_from_time(time_arg, node=None, clock_seq=None):
 | |
|     """
 | |
|     Converts a datetime or timestamp to a type 1 :class:`uuid.UUID`.
 | |
| 
 | |
|     :param time_arg:
 | |
|       The time to use for the timestamp portion of the UUID.
 | |
|       This can either be a :class:`datetime` object or a timestamp
 | |
|       in seconds (as returned from :meth:`time.time()`).
 | |
|     :type datetime: :class:`datetime` or timestamp
 | |
| 
 | |
|     :param node:
 | |
|       None integer for the UUID (up to 48 bits). If not specified, this
 | |
|       field is randomized.
 | |
|     :type node: long
 | |
| 
 | |
|     :param clock_seq:
 | |
|       Clock sequence field for the UUID (up to 14 bits). If not specified,
 | |
|       a random sequence is generated.
 | |
|     :type clock_seq: int
 | |
| 
 | |
|     :rtype: :class:`uuid.UUID`
 | |
| 
 | |
|     """
 | |
|     if hasattr(time_arg, 'utctimetuple'):
 | |
|         seconds = int(calendar.timegm(time_arg.utctimetuple()))
 | |
|         microseconds = (seconds * 1e6) + time_arg.time().microsecond
 | |
|     else:
 | |
|         microseconds = int(time_arg * 1e6)
 | |
| 
 | |
|     # 0x01b21dd213814000 is the number of 100-ns intervals between the
 | |
|     # UUID epoch 1582-10-15 00:00:00 and the Unix epoch 1970-01-01 00:00:00.
 | |
|     intervals = int(microseconds * 10) + 0x01b21dd213814000
 | |
| 
 | |
|     time_low = intervals & 0xffffffff
 | |
|     time_mid = (intervals >> 32) & 0xffff
 | |
|     time_hi_version = (intervals >> 48) & 0x0fff
 | |
| 
 | |
|     if clock_seq is None:
 | |
|         clock_seq = random.getrandbits(14)
 | |
|     else:
 | |
|         if clock_seq > 0x3fff:
 | |
|             raise ValueError('clock_seq is out of range (need a 14-bit value)')
 | |
| 
 | |
|     clock_seq_low = clock_seq & 0xff
 | |
|     clock_seq_hi_variant = 0x80 | ((clock_seq >> 8) & 0x3f)
 | |
| 
 | |
|     if node is None:
 | |
|         node = random.getrandbits(48)
 | |
| 
 | |
|     return uuid.UUID(fields=(time_low, time_mid, time_hi_version,
 | |
|                              clock_seq_hi_variant, clock_seq_low, node), version=1)
 | |
| 
 | |
| LOWEST_TIME_UUID = uuid.UUID('00000000-0000-1000-8080-808080808080')
 | |
| """ The lowest possible TimeUUID, as sorted by Cassandra. """
 | |
| 
 | |
| HIGHEST_TIME_UUID = uuid.UUID('ffffffff-ffff-1fff-bf7f-7f7f7f7f7f7f')
 | |
| """ The highest possible TimeUUID, as sorted by Cassandra. """
 | |
| 
 | |
| 
 | |
| try:
 | |
|     from collections import OrderedDict
 | |
| except ImportError:
 | |
|     # OrderedDict from Python 2.7+
 | |
| 
 | |
|     # Copyright (c) 2009 Raymond Hettinger
 | |
|     #
 | |
|     # Permission is hereby granted, free of charge, to any person
 | |
|     # obtaining a copy of this software and associated documentation files
 | |
|     # (the "Software"), to deal in the Software without restriction,
 | |
|     # including without limitation the rights to use, copy, modify, merge,
 | |
|     # publish, distribute, sublicense, and/or sell copies of the Software,
 | |
|     # and to permit persons to whom the Software is furnished to do so,
 | |
|     # subject to the following conditions:
 | |
|     #
 | |
|     #     The above copyright notice and this permission notice shall be
 | |
|     #     included in all copies or substantial portions of the Software.
 | |
|     #
 | |
|     #     THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
 | |
|     #     EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
 | |
|     #     OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
 | |
|     #     NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
 | |
|     #     HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
 | |
|     #     WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
 | |
|     #     FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
 | |
|     #     OTHER DEALINGS IN THE SOFTWARE.
 | |
|     from UserDict import DictMixin
 | |
| 
 | |
|     class OrderedDict(dict, DictMixin):  # noqa
 | |
|         """ A dictionary which maintains the insertion order of keys. """
 | |
| 
 | |
|         def __init__(self, *args, **kwds):
 | |
|             """ A dictionary which maintains the insertion order of keys. """
 | |
| 
 | |
|             if len(args) > 1:
 | |
|                 raise TypeError('expected at most 1 arguments, got %d' % len(args))
 | |
|             try:
 | |
|                 self.__end
 | |
|             except AttributeError:
 | |
|                 self.clear()
 | |
|             self.update(*args, **kwds)
 | |
| 
 | |
|         def clear(self):
 | |
|             self.__end = end = []
 | |
|             end += [None, end, end]         # sentinel node for doubly linked list
 | |
|             self.__map = {}                 # key --> [key, prev, next]
 | |
|             dict.clear(self)
 | |
| 
 | |
|         def __setitem__(self, key, value):
 | |
|             if key not in self:
 | |
|                 end = self.__end
 | |
|                 curr = end[1]
 | |
|                 curr[2] = end[1] = self.__map[key] = [key, curr, end]
 | |
|             dict.__setitem__(self, key, value)
 | |
| 
 | |
|         def __delitem__(self, key):
 | |
|             dict.__delitem__(self, key)
 | |
|             key, prev, next = self.__map.pop(key)
 | |
|             prev[2] = next
 | |
|             next[1] = prev
 | |
| 
 | |
|         def __iter__(self):
 | |
|             end = self.__end
 | |
|             curr = end[2]
 | |
|             while curr is not end:
 | |
|                 yield curr[0]
 | |
|                 curr = curr[2]
 | |
| 
 | |
|         def __reversed__(self):
 | |
|             end = self.__end
 | |
|             curr = end[1]
 | |
|             while curr is not end:
 | |
|                 yield curr[0]
 | |
|                 curr = curr[1]
 | |
| 
 | |
|         def popitem(self, last=True):
 | |
|             if not self:
 | |
|                 raise KeyError('dictionary is empty')
 | |
|             if last:
 | |
|                 key = next(reversed(self))
 | |
|             else:
 | |
|                 key = next(iter(self))
 | |
|             value = self.pop(key)
 | |
|             return key, value
 | |
| 
 | |
|         def __reduce__(self):
 | |
|             items = [[k, self[k]] for k in self]
 | |
|             tmp = self.__map, self.__end
 | |
|             del self.__map, self.__end
 | |
|             inst_dict = vars(self).copy()
 | |
|             self.__map, self.__end = tmp
 | |
|             if inst_dict:
 | |
|                 return (self.__class__, (items,), inst_dict)
 | |
|             return self.__class__, (items,)
 | |
| 
 | |
|         def keys(self):
 | |
|             return list(self)
 | |
| 
 | |
|         setdefault = DictMixin.setdefault
 | |
|         update = DictMixin.update
 | |
|         pop = DictMixin.pop
 | |
|         values = DictMixin.values
 | |
|         items = DictMixin.items
 | |
|         iterkeys = DictMixin.iterkeys
 | |
|         itervalues = DictMixin.itervalues
 | |
|         iteritems = DictMixin.iteritems
 | |
| 
 | |
|         def __repr__(self):
 | |
|             if not self:
 | |
|                 return '%s()' % (self.__class__.__name__,)
 | |
|             return '%s(%r)' % (self.__class__.__name__, self.items())
 | |
| 
 | |
|         def copy(self):
 | |
|             return self.__class__(self)
 | |
| 
 | |
|         @classmethod
 | |
|         def fromkeys(cls, iterable, value=None):
 | |
|             d = cls()
 | |
|             for key in iterable:
 | |
|                 d[key] = value
 | |
|             return d
 | |
| 
 | |
|         def __eq__(self, other):
 | |
|             if isinstance(other, OrderedDict):
 | |
|                 if len(self) != len(other):
 | |
|                     return False
 | |
|                 for p, q in zip(self.items(), other.items()):
 | |
|                     if p != q:
 | |
|                         return False
 | |
|                 return True
 | |
|             return dict.__eq__(self, other)
 | |
| 
 | |
|         def __ne__(self, other):
 | |
|             return not self == other
 | |
| 
 | |
| 
 | |
| # WeakSet from Python 2.7+ (https://code.google.com/p/weakrefset)
 | |
| 
 | |
| from _weakref import ref
 | |
| 
 | |
| 
 | |
| class _IterationGuard(object):
 | |
|     # This context manager registers itself in the current iterators of the
 | |
|     # weak container, such as to delay all removals until the context manager
 | |
|     # exits.
 | |
|     # This technique should be relatively thread-safe (since sets are).
 | |
| 
 | |
|     def __init__(self, weakcontainer):
 | |
|         # Don't create cycles
 | |
|         self.weakcontainer = ref(weakcontainer)
 | |
| 
 | |
|     def __enter__(self):
 | |
|         w = self.weakcontainer()
 | |
|         if w is not None:
 | |
|             w._iterating.add(self)
 | |
|         return self
 | |
| 
 | |
|     def __exit__(self, e, t, b):
 | |
|         w = self.weakcontainer()
 | |
|         if w is not None:
 | |
|             s = w._iterating
 | |
|             s.remove(self)
 | |
|             if not s:
 | |
|                 w._commit_removals()
 | |
| 
 | |
| 
 | |
| class WeakSet(object):
 | |
|     def __init__(self, data=None):
 | |
|         self.data = set()
 | |
| 
 | |
|         def _remove(item, selfref=ref(self)):
 | |
|             self = selfref()
 | |
|             if self is not None:
 | |
|                 if self._iterating:
 | |
|                     self._pending_removals.append(item)
 | |
|                 else:
 | |
|                     self.data.discard(item)
 | |
| 
 | |
|         self._remove = _remove
 | |
|         # A list of keys to be removed
 | |
|         self._pending_removals = []
 | |
|         self._iterating = set()
 | |
|         if data is not None:
 | |
|             self.update(data)
 | |
| 
 | |
|     def _commit_removals(self):
 | |
|         l = self._pending_removals
 | |
|         discard = self.data.discard
 | |
|         while l:
 | |
|             discard(l.pop())
 | |
| 
 | |
|     def __iter__(self):
 | |
|         with _IterationGuard(self):
 | |
|             for itemref in self.data:
 | |
|                 item = itemref()
 | |
|                 if item is not None:
 | |
|                     yield item
 | |
| 
 | |
|     def __len__(self):
 | |
|         return sum(x() is not None for x in self.data)
 | |
| 
 | |
|     def __contains__(self, item):
 | |
|         return ref(item) in self.data
 | |
| 
 | |
|     def __reduce__(self):
 | |
|         return (self.__class__, (list(self),),
 | |
|                 getattr(self, '__dict__', None))
 | |
| 
 | |
|     __hash__ = None
 | |
| 
 | |
|     def add(self, item):
 | |
|         if self._pending_removals:
 | |
|             self._commit_removals()
 | |
|         self.data.add(ref(item, self._remove))
 | |
| 
 | |
|     def clear(self):
 | |
|         if self._pending_removals:
 | |
|             self._commit_removals()
 | |
|         self.data.clear()
 | |
| 
 | |
|     def copy(self):
 | |
|         return self.__class__(self)
 | |
| 
 | |
|     def pop(self):
 | |
|         if self._pending_removals:
 | |
|             self._commit_removals()
 | |
|         while True:
 | |
|             try:
 | |
|                 itemref = self.data.pop()
 | |
|             except KeyError:
 | |
|                 raise KeyError('pop from empty WeakSet')
 | |
|             item = itemref()
 | |
|             if item is not None:
 | |
|                 return item
 | |
| 
 | |
|     def remove(self, item):
 | |
|         if self._pending_removals:
 | |
|             self._commit_removals()
 | |
|         self.data.remove(ref(item))
 | |
| 
 | |
|     def discard(self, item):
 | |
|         if self._pending_removals:
 | |
|             self._commit_removals()
 | |
|         self.data.discard(ref(item))
 | |
| 
 | |
|     def update(self, other):
 | |
|         if self._pending_removals:
 | |
|             self._commit_removals()
 | |
|         if isinstance(other, self.__class__):
 | |
|             self.data.update(other.data)
 | |
|         else:
 | |
|             for element in other:
 | |
|                 self.add(element)
 | |
| 
 | |
|     def __ior__(self, other):
 | |
|         self.update(other)
 | |
|         return self
 | |
| 
 | |
|     # Helper functions for simple delegating methods.
 | |
|     def _apply(self, other, method):
 | |
|         if not isinstance(other, self.__class__):
 | |
|             other = self.__class__(other)
 | |
|         newdata = method(other.data)
 | |
|         newset = self.__class__()
 | |
|         newset.data = newdata
 | |
|         return newset
 | |
| 
 | |
|     def difference(self, other):
 | |
|         return self._apply(other, self.data.difference)
 | |
|     __sub__ = difference
 | |
| 
 | |
|     def difference_update(self, other):
 | |
|         if self._pending_removals:
 | |
|             self._commit_removals()
 | |
|         if self is other:
 | |
|             self.data.clear()
 | |
|         else:
 | |
|             self.data.difference_update(ref(item) for item in other)
 | |
| 
 | |
|     def __isub__(self, other):
 | |
|         if self._pending_removals:
 | |
|             self._commit_removals()
 | |
|         if self is other:
 | |
|             self.data.clear()
 | |
|         else:
 | |
|             self.data.difference_update(ref(item) for item in other)
 | |
|         return self
 | |
| 
 | |
|     def intersection(self, other):
 | |
|         return self._apply(other, self.data.intersection)
 | |
|     __and__ = intersection
 | |
| 
 | |
|     def intersection_update(self, other):
 | |
|         if self._pending_removals:
 | |
|             self._commit_removals()
 | |
|         self.data.intersection_update(ref(item) for item in other)
 | |
| 
 | |
|     def __iand__(self, other):
 | |
|         if self._pending_removals:
 | |
|             self._commit_removals()
 | |
|         self.data.intersection_update(ref(item) for item in other)
 | |
|         return self
 | |
| 
 | |
|     def issubset(self, other):
 | |
|         return self.data.issubset(ref(item) for item in other)
 | |
|     __lt__ = issubset
 | |
| 
 | |
|     def __le__(self, other):
 | |
|         return self.data <= set(ref(item) for item in other)
 | |
| 
 | |
|     def issuperset(self, other):
 | |
|         return self.data.issuperset(ref(item) for item in other)
 | |
|     __gt__ = issuperset
 | |
| 
 | |
|     def __ge__(self, other):
 | |
|         return self.data >= set(ref(item) for item in other)
 | |
| 
 | |
|     def __eq__(self, other):
 | |
|         if not isinstance(other, self.__class__):
 | |
|             return NotImplemented
 | |
|         return self.data == set(ref(item) for item in other)
 | |
| 
 | |
|     def symmetric_difference(self, other):
 | |
|         return self._apply(other, self.data.symmetric_difference)
 | |
|     __xor__ = symmetric_difference
 | |
| 
 | |
|     def symmetric_difference_update(self, other):
 | |
|         if self._pending_removals:
 | |
|             self._commit_removals()
 | |
|         if self is other:
 | |
|             self.data.clear()
 | |
|         else:
 | |
|             self.data.symmetric_difference_update(ref(item) for item in other)
 | |
| 
 | |
|     def __ixor__(self, other):
 | |
|         if self._pending_removals:
 | |
|             self._commit_removals()
 | |
|         if self is other:
 | |
|             self.data.clear()
 | |
|         else:
 | |
|             self.data.symmetric_difference_update(ref(item) for item in other)
 | |
|         return self
 | |
| 
 | |
|     def union(self, other):
 | |
|         return self._apply(other, self.data.union)
 | |
|     __or__ = union
 | |
| 
 | |
|     def isdisjoint(self, other):
 | |
|         return len(self.intersection(other)) == 0
 | |
| 
 | |
| 
 | |
| from bisect import bisect_left
 | |
| 
 | |
| 
 | |
| class SortedSet(object):
 | |
|     '''
 | |
|     A sorted set based on sorted list
 | |
| 
 | |
|     A sorted set implementation is used in this case because it does not
 | |
|     require its elements to be immutable/hashable.
 | |
| 
 | |
|     #Not implemented: update functions, inplace operators
 | |
|     '''
 | |
| 
 | |
|     def __init__(self, iterable=()):
 | |
|         self._items = []
 | |
|         self.update(iterable)
 | |
| 
 | |
|     def __len__(self):
 | |
|         return len(self._items)
 | |
| 
 | |
|     def __getitem__(self, i):
 | |
|         return self._items[i]
 | |
| 
 | |
|     def __iter__(self):
 | |
|         return iter(self._items)
 | |
| 
 | |
|     def __reversed__(self):
 | |
|         return reversed(self._items)
 | |
| 
 | |
|     def __repr__(self):
 | |
|         return '%s(%r)' % (
 | |
|             self.__class__.__name__,
 | |
|             self._items)
 | |
| 
 | |
|     def __reduce__(self):
 | |
|         return self.__class__, (self._items,)
 | |
| 
 | |
|     def __eq__(self, other):
 | |
|         if isinstance(other, self.__class__):
 | |
|             return self._items == other._items
 | |
|         else:
 | |
|             try:
 | |
|                 return len(other) == len(self._items) and all(item in self for item in other)
 | |
|             except TypeError:
 | |
|                 return NotImplemented
 | |
| 
 | |
|     def __ne__(self, other):
 | |
|         if isinstance(other, self.__class__):
 | |
|             return self._items != other._items
 | |
|         else:
 | |
|             try:
 | |
|                 return len(other) != len(self._items) or any(item not in self for item in other)
 | |
|             except TypeError:
 | |
|                 return NotImplemented
 | |
| 
 | |
|     def __le__(self, other):
 | |
|         return self.issubset(other)
 | |
| 
 | |
|     def __lt__(self, other):
 | |
|         return len(other) > len(self._items) and self.issubset(other)
 | |
| 
 | |
|     def __ge__(self, other):
 | |
|         return self.issuperset(other)
 | |
| 
 | |
|     def __gt__(self, other):
 | |
|         return len(self._items) > len(other) and self.issuperset(other)
 | |
| 
 | |
|     def __and__(self, other):
 | |
|         return self._intersect(other)
 | |
|     __rand__ = __and__
 | |
| 
 | |
|     def __iand__(self, other):
 | |
|         isect = self._intersect(other)
 | |
|         self._items = isect._items
 | |
|         return self
 | |
| 
 | |
|     def __or__(self, other):
 | |
|         return self.union(other)
 | |
|     __ror__ = __or__
 | |
| 
 | |
|     def __ior__(self, other):
 | |
|         union = self.union(other)
 | |
|         self._items = union._items
 | |
|         return self
 | |
| 
 | |
|     def __sub__(self, other):
 | |
|         return self._diff(other)
 | |
| 
 | |
|     def __rsub__(self, other):
 | |
|         return sortedset(other) - self
 | |
| 
 | |
|     def __isub__(self, other):
 | |
|         diff = self._diff(other)
 | |
|         self._items = diff._items
 | |
|         return self
 | |
| 
 | |
|     def __xor__(self, other):
 | |
|         return self.symmetric_difference(other)
 | |
|     __rxor__ = __xor__
 | |
| 
 | |
|     def __ixor__(self, other):
 | |
|         sym_diff = self.symmetric_difference(other)
 | |
|         self._items = sym_diff._items
 | |
|         return self
 | |
| 
 | |
|     def __contains__(self, item):
 | |
|         i = bisect_left(self._items, item)
 | |
|         return i < len(self._items) and self._items[i] == item
 | |
| 
 | |
|     def __delitem__(self, i):
 | |
|         del self._items[i]
 | |
| 
 | |
|     def __delslice__(self, i, j):
 | |
|         del self._items[i:j]
 | |
| 
 | |
|     def add(self, item):
 | |
|         i = bisect_left(self._items, item)
 | |
|         if i < len(self._items):
 | |
|             if self._items[i] != item:
 | |
|                 self._items.insert(i, item)
 | |
|         else:
 | |
|             self._items.append(item)
 | |
| 
 | |
|     def update(self, iterable):
 | |
|         for i in iterable:
 | |
|             self.add(i)
 | |
| 
 | |
|     def clear(self):
 | |
|         del self._items[:]
 | |
| 
 | |
|     def copy(self):
 | |
|         new = sortedset()
 | |
|         new._items = list(self._items)
 | |
|         return new
 | |
| 
 | |
|     def isdisjoint(self, other):
 | |
|         return len(self._intersect(other)) == 0
 | |
| 
 | |
|     def issubset(self, other):
 | |
|         return len(self._intersect(other)) == len(self._items)
 | |
| 
 | |
|     def issuperset(self, other):
 | |
|         return len(self._intersect(other)) == len(other)
 | |
| 
 | |
|     def pop(self):
 | |
|         if not self._items:
 | |
|             raise KeyError("pop from empty set")
 | |
|         return self._items.pop()
 | |
| 
 | |
|     def remove(self, item):
 | |
|         i = bisect_left(self._items, item)
 | |
|         if i < len(self._items):
 | |
|             if self._items[i] == item:
 | |
|                 self._items.pop(i)
 | |
|                 return
 | |
|         raise KeyError('%r' % item)
 | |
| 
 | |
|     def union(self, *others):
 | |
|         union = sortedset()
 | |
|         union._items = list(self._items)
 | |
|         for other in others:
 | |
|             if isinstance(other, self.__class__):
 | |
|                 i = 0
 | |
|                 for item in other._items:
 | |
|                     i = bisect_left(union._items, item, i)
 | |
|                     if i < len(union._items):
 | |
|                         if item != union._items[i]:
 | |
|                             union._items.insert(i, item)
 | |
|                     else:
 | |
|                         union._items.append(item)
 | |
|             else:
 | |
|                 for item in other:
 | |
|                     union.add(item)
 | |
|         return union
 | |
| 
 | |
|     def intersection(self, *others):
 | |
|         isect = self.copy()
 | |
|         for other in others:
 | |
|             isect = isect._intersect(other)
 | |
|             if not isect:
 | |
|                 break
 | |
|         return isect
 | |
| 
 | |
|     def difference(self, *others):
 | |
|         diff = self.copy()
 | |
|         for other in others:
 | |
|             diff = diff._diff(other)
 | |
|             if not diff:
 | |
|                 break
 | |
|         return diff
 | |
| 
 | |
|     def symmetric_difference(self, other):
 | |
|         diff_self_other = self._diff(other)
 | |
|         diff_other_self = other.difference(self)
 | |
|         return diff_self_other.union(diff_other_self)
 | |
| 
 | |
|     def _diff(self, other):
 | |
|         diff = sortedset()
 | |
|         if isinstance(other, self.__class__):
 | |
|             i = 0
 | |
|             for item in self._items:
 | |
|                 i = bisect_left(other._items, item, i)
 | |
|                 if i < len(other._items):
 | |
|                     if item != other._items[i]:
 | |
|                         diff._items.append(item)
 | |
|                 else:
 | |
|                     diff._items.append(item)
 | |
|         else:
 | |
|             for item in self._items:
 | |
|                 if item not in other:
 | |
|                     diff.add(item)
 | |
|         return diff
 | |
| 
 | |
|     def _intersect(self, other):
 | |
|         isect = sortedset()
 | |
|         if isinstance(other, self.__class__):
 | |
|             i = 0
 | |
|             for item in self._items:
 | |
|                 i = bisect_left(other._items, item, i)
 | |
|                 if i < len(other._items):
 | |
|                     if item == other._items[i]:
 | |
|                         isect._items.append(item)
 | |
|                 else:
 | |
|                     break
 | |
|         else:
 | |
|             for item in self._items:
 | |
|                 if item in other:
 | |
|                     isect.add(item)
 | |
|         return isect
 | |
| 
 | |
| sortedset = SortedSet  # backwards-compatibility
 | |
| 
 | |
| 
 | |
| from collections import Mapping
 | |
| from six.moves import cPickle
 | |
| 
 | |
| 
 | |
| class OrderedMap(Mapping):
 | |
|     '''
 | |
|     An ordered map that accepts non-hashable types for keys. It also maintains the
 | |
|     insertion order of items, behaving as OrderedDict in that regard. These maps
 | |
|     are constructed and read just as normal mapping types, exept that they may
 | |
|     contain arbitrary collections and other non-hashable items as keys::
 | |
| 
 | |
|         >>> od = OrderedMap([({'one': 1, 'two': 2}, 'value'),
 | |
|         ...                  ({'three': 3, 'four': 4}, 'value2')])
 | |
|         >>> list(od.keys())
 | |
|         [{'two': 2, 'one': 1}, {'three': 3, 'four': 4}]
 | |
|         >>> list(od.values())
 | |
|         ['value', 'value2']
 | |
| 
 | |
|     These constructs are needed to support nested collections in Cassandra 2.1.3+,
 | |
|     where frozen collections can be specified as parameters to others\*::
 | |
| 
 | |
|         CREATE TABLE example (
 | |
|             ...
 | |
|             value map<frozen<map<int, int>>, double>
 | |
|             ...
 | |
|         )
 | |
| 
 | |
|     This class derives from the (immutable) Mapping API. Objects in these maps
 | |
|     are not intended be modified.
 | |
| 
 | |
|     \* Note: Because of the way Cassandra encodes nested types, when using the
 | |
|     driver with nested collections, :attr:`~.Cluster.protocol_version` must be 3
 | |
|     or higher.
 | |
| 
 | |
|     '''
 | |
| 
 | |
|     def __init__(self, *args, **kwargs):
 | |
|         if len(args) > 1:
 | |
|             raise TypeError('expected at most 1 arguments, got %d' % len(args))
 | |
| 
 | |
|         self._items = []
 | |
|         self._index = {}
 | |
|         if args:
 | |
|             e = args[0]
 | |
|             if callable(getattr(e, 'keys', None)):
 | |
|                 for k in e.keys():
 | |
|                     self._insert(k, e[k])
 | |
|             else:
 | |
|                 for k, v in e:
 | |
|                     self._insert(k, v)
 | |
| 
 | |
|         for k, v in six.iteritems(kwargs):
 | |
|             self._insert(k, v)
 | |
| 
 | |
|     def _insert(self, key, value):
 | |
|         flat_key = self._serialize_key(key)
 | |
|         i = self._index.get(flat_key, -1)
 | |
|         if i >= 0:
 | |
|             self._items[i] = (key, value)
 | |
|         else:
 | |
|             self._items.append((key, value))
 | |
|             self._index[flat_key] = len(self._items) - 1
 | |
| 
 | |
|     __setitem__ = _insert
 | |
| 
 | |
|     def __getitem__(self, key):
 | |
|         try:
 | |
|             index = self._index[self._serialize_key(key)]
 | |
|             return self._items[index][1]
 | |
|         except KeyError:
 | |
|             raise KeyError(str(key))
 | |
| 
 | |
|     def __delitem__(self, key):
 | |
|         # not efficient -- for convenience only
 | |
|         try:
 | |
|             index = self._index.pop(self._serialize_key(key))
 | |
|             self._index = dict((k, i if i < index else i - 1) for k, i in self._index.items())
 | |
|             self._items.pop(index)
 | |
|         except KeyError:
 | |
|             raise KeyError(str(key))
 | |
| 
 | |
|     def __iter__(self):
 | |
|         for i in self._items:
 | |
|             yield i[0]
 | |
| 
 | |
|     def __len__(self):
 | |
|         return len(self._items)
 | |
| 
 | |
|     def __eq__(self, other):
 | |
|         if isinstance(other, OrderedMap):
 | |
|             return self._items == other._items
 | |
|         try:
 | |
|             d = dict(other)
 | |
|             return len(d) == len(self._items) and all(i[1] == d[i[0]] for i in self._items)
 | |
|         except KeyError:
 | |
|             return False
 | |
|         except TypeError:
 | |
|             pass
 | |
|         return NotImplemented
 | |
| 
 | |
|     def __repr__(self):
 | |
|         return '%s([%s])' % (
 | |
|             self.__class__.__name__,
 | |
|             ', '.join("(%r, %r)" % (k, v) for k, v in self._items))
 | |
| 
 | |
|     def __str__(self):
 | |
|         return '{%s}' % ', '.join("%r: %r" % (k, v) for k, v in self._items)
 | |
| 
 | |
|     def popitem(self):
 | |
|         try:
 | |
|             kv = self._items.pop()
 | |
|             del self._index[self._serialize_key(kv[0])]
 | |
|             return kv
 | |
|         except IndexError:
 | |
|             raise KeyError()
 | |
| 
 | |
|     def _serialize_key(self, key):
 | |
|         return cPickle.dumps(key)
 | |
| 
 | |
| 
 | |
| class OrderedMapSerializedKey(OrderedMap):
 | |
| 
 | |
|     def __init__(self, cass_type, protocol_version):
 | |
|         super(OrderedMapSerializedKey, self).__init__()
 | |
|         self.cass_key_type = cass_type
 | |
|         self.protocol_version = protocol_version
 | |
| 
 | |
|     def _insert_unchecked(self, key, flat_key, value):
 | |
|         self._items.append((key, value))
 | |
|         self._index[flat_key] = len(self._items) - 1
 | |
| 
 | |
|     def _serialize_key(self, key):
 | |
|         return self.cass_key_type.serialize(key, self.protocol_version)
 | |
| 
 | |
| 
 | |
| import datetime
 | |
| import time
 | |
| 
 | |
| if six.PY3:
 | |
|     long = int
 | |
| 
 | |
| 
 | |
| class Time(object):
 | |
|     '''
 | |
|     Idealized time, independent of day.
 | |
| 
 | |
|     Up to nanosecond resolution
 | |
|     '''
 | |
| 
 | |
|     MICRO = 1000
 | |
|     MILLI = 1000 * MICRO
 | |
|     SECOND = 1000 * MILLI
 | |
|     MINUTE = 60 * SECOND
 | |
|     HOUR = 60 * MINUTE
 | |
|     DAY = 24 * HOUR
 | |
| 
 | |
|     nanosecond_time = 0
 | |
| 
 | |
|     def __init__(self, value):
 | |
|         """
 | |
|         Initializer value can be:
 | |
| 
 | |
|             - integer_type: absolute nanoseconds in the day
 | |
|             - datetime.time: built-in time
 | |
|             - string_type: a string time of the form "HH:MM:SS[.mmmuuunnn]"
 | |
|         """
 | |
|         if isinstance(value, six.integer_types):
 | |
|             self._from_timestamp(value)
 | |
|         elif isinstance(value, datetime.time):
 | |
|             self._from_time(value)
 | |
|         elif isinstance(value, six.string_types):
 | |
|             self._from_timestring(value)
 | |
|         else:
 | |
|             raise TypeError('Time arguments must be a whole number, datetime.time, or string')
 | |
| 
 | |
|     @property
 | |
|     def hour(self):
 | |
|         """
 | |
|         The hour component of this time (0-23)
 | |
|         """
 | |
|         return self.nanosecond_time // Time.HOUR
 | |
| 
 | |
|     @property
 | |
|     def minute(self):
 | |
|         """
 | |
|         The minute component of this time (0-59)
 | |
|         """
 | |
|         minutes = self.nanosecond_time // Time.MINUTE
 | |
|         return minutes % 60
 | |
| 
 | |
|     @property
 | |
|     def second(self):
 | |
|         """
 | |
|         The second component of this time (0-59)
 | |
|         """
 | |
|         seconds = self.nanosecond_time // Time.SECOND
 | |
|         return seconds % 60
 | |
| 
 | |
|     @property
 | |
|     def nanosecond(self):
 | |
|         """
 | |
|         The fractional seconds component of the time, in nanoseconds
 | |
|         """
 | |
|         return self.nanosecond_time % Time.SECOND
 | |
| 
 | |
|     def _from_timestamp(self, t):
 | |
|         if t >= Time.DAY:
 | |
|             raise ValueError("value must be less than number of nanoseconds in a day (%d)" % Time.DAY)
 | |
|         self.nanosecond_time = t
 | |
| 
 | |
|     def _from_timestring(self, s):
 | |
|         try:
 | |
|             parts = s.split('.')
 | |
|             base_time = time.strptime(parts[0], "%H:%M:%S")
 | |
|             self.nanosecond_time = (base_time.tm_hour * Time.HOUR +
 | |
|                                     base_time.tm_min * Time.MINUTE +
 | |
|                                     base_time.tm_sec * Time.SECOND)
 | |
| 
 | |
|             if len(parts) > 1:
 | |
|                 # right pad to 9 digits
 | |
|                 nano_time_str = parts[1] + "0" * (9 - len(parts[1]))
 | |
|                 self.nanosecond_time += int(nano_time_str)
 | |
| 
 | |
|         except ValueError:
 | |
|             raise ValueError("can't interpret %r as a time" % (s,))
 | |
| 
 | |
|     def _from_time(self, t):
 | |
|         self.nanosecond_time = (t.hour * Time.HOUR +
 | |
|                                 t.minute * Time.MINUTE +
 | |
|                                 t.second * Time.SECOND +
 | |
|                                 t.microsecond * Time.MICRO)
 | |
| 
 | |
|     def __hash__(self):
 | |
|         return self.nanosecond_time
 | |
| 
 | |
|     def __eq__(self, other):
 | |
|         if isinstance(other, Time):
 | |
|             return self.nanosecond_time == other.nanosecond_time
 | |
| 
 | |
|         if isinstance(other, six.integer_types):
 | |
|             return self.nanosecond_time == other
 | |
| 
 | |
|         return self.nanosecond_time % Time.MICRO == 0 and \
 | |
|             datetime.time(hour=self.hour, minute=self.minute, second=self.second,
 | |
|                           microsecond=self.nanosecond // Time.MICRO) == other
 | |
| 
 | |
|     def __lt__(self, other):
 | |
|         return self.nanosecond_time < other.nanosecond_time
 | |
| 
 | |
|     def __repr__(self):
 | |
|         return "Time(%s)" % self.nanosecond_time
 | |
| 
 | |
|     def __str__(self):
 | |
|         return "%02d:%02d:%02d.%09d" % (self.hour, self.minute,
 | |
|                                         self.second, self.nanosecond)
 | |
| 
 | |
| 
 | |
| class Date(object):
 | |
|     '''
 | |
|     Idealized date: year, month, day
 | |
| 
 | |
|     Offers wider year range than datetime.date. For Dates that cannot be represented
 | |
|     as a datetime.date (because datetime.MINYEAR, datetime.MAXYEAR), this type falls back
 | |
|     to printing days_from_epoch offset.
 | |
|     '''
 | |
| 
 | |
|     MINUTE = 60
 | |
|     HOUR = 60 * MINUTE
 | |
|     DAY = 24 * HOUR
 | |
| 
 | |
|     date_format = "%Y-%m-%d"
 | |
| 
 | |
|     days_from_epoch = 0
 | |
| 
 | |
|     def __init__(self, value):
 | |
|         """
 | |
|         Initializer value can be:
 | |
| 
 | |
|             - integer_type: absolute days from epoch (1970, 1, 1). Can be negative.
 | |
|             - datetime.date: built-in date
 | |
|             - string_type: a string time of the form "yyyy-mm-dd"
 | |
|         """
 | |
|         if isinstance(value, six.integer_types):
 | |
|             self.days_from_epoch = value
 | |
|         elif isinstance(value, (datetime.date, datetime.datetime)):
 | |
|             self._from_timetuple(value.timetuple())
 | |
|         elif isinstance(value, six.string_types):
 | |
|             self._from_datestring(value)
 | |
|         else:
 | |
|             raise TypeError('Date arguments must be a whole number, datetime.date, or string')
 | |
| 
 | |
|     @property
 | |
|     def seconds(self):
 | |
|         """
 | |
|         Absolute seconds from epoch (can be negative)
 | |
|         """
 | |
|         return self.days_from_epoch * Date.DAY
 | |
| 
 | |
|     def date(self):
 | |
|         """
 | |
|         Return a built-in datetime.date for Dates falling in the years [datetime.MINYEAR, datetime.MAXYEAR]
 | |
| 
 | |
|         ValueError is raised for Dates outside this range.
 | |
|         """
 | |
|         try:
 | |
|             dt = datetime_from_timestamp(self.seconds)
 | |
|             return datetime.date(dt.year, dt.month, dt.day)
 | |
|         except Exception:
 | |
|             raise ValueError("%r exceeds ranges for built-in datetime.date" % self)
 | |
| 
 | |
|     def _from_timetuple(self, t):
 | |
|         self.days_from_epoch = calendar.timegm(t) // Date.DAY
 | |
| 
 | |
|     def _from_datestring(self, s):
 | |
|         if s[0] == '+':
 | |
|             s = s[1:]
 | |
|         dt = datetime.datetime.strptime(s, self.date_format)
 | |
|         self._from_timetuple(dt.timetuple())
 | |
| 
 | |
|     def __hash__(self):
 | |
|         return self.days_from_epoch
 | |
| 
 | |
|     def __eq__(self, other):
 | |
|         if isinstance(other, Date):
 | |
|             return self.days_from_epoch == other.days_from_epoch
 | |
| 
 | |
|         if isinstance(other, six.integer_types):
 | |
|             return self.days_from_epoch == other
 | |
| 
 | |
|         try:
 | |
|             return self.date() == other
 | |
|         except Exception:
 | |
|             return False
 | |
| 
 | |
|     def __lt__(self, other):
 | |
|         return self.days_from_epoch < other.days_from_epoch
 | |
| 
 | |
|     def __repr__(self):
 | |
|         return "Date(%s)" % self.days_from_epoch
 | |
| 
 | |
|     def __str__(self):
 | |
|         try:
 | |
|             dt = datetime_from_timestamp(self.seconds)
 | |
|             return "%04d-%02d-%02d" % (dt.year, dt.month, dt.day)
 | |
|         except:
 | |
|             # If we overflow datetime.[MIN|MAX]
 | |
|             return str(self.days_from_epoch)
 | |
| 
 | |
| import socket
 | |
| if hasattr(socket, 'inet_pton'):
 | |
|     inet_pton = socket.inet_pton
 | |
|     inet_ntop = socket.inet_ntop
 | |
| else:
 | |
|     """
 | |
|     Windows doesn't have socket.inet_pton and socket.inet_ntop until Python 3.4
 | |
|     This is an alternative impl using ctypes, based on this win_inet_pton project:
 | |
|     https://github.com/hickeroar/win_inet_pton
 | |
|     """
 | |
|     import ctypes
 | |
| 
 | |
|     class sockaddr(ctypes.Structure):
 | |
|         """
 | |
|         Shared struct for ipv4 and ipv6.
 | |
| 
 | |
|         https://msdn.microsoft.com/en-us/library/windows/desktop/ms740496(v=vs.85).aspx
 | |
| 
 | |
|         ``__pad1`` always covers the port.
 | |
| 
 | |
|         When being used for ``sockaddr_in6``, ``ipv4_addr`` actually covers ``sin6_flowinfo``, resulting
 | |
|         in proper alignment for ``ipv6_addr``.
 | |
|         """
 | |
|         _fields_ = [("sa_family", ctypes.c_short),
 | |
|                     ("__pad1", ctypes.c_ushort),
 | |
|                     ("ipv4_addr", ctypes.c_byte * 4),
 | |
|                     ("ipv6_addr", ctypes.c_byte * 16),
 | |
|                     ("__pad2", ctypes.c_ulong)]
 | |
| 
 | |
|     if hasattr(ctypes, 'windll'):
 | |
|         WSAStringToAddressA = ctypes.windll.ws2_32.WSAStringToAddressA
 | |
|         WSAAddressToStringA = ctypes.windll.ws2_32.WSAAddressToStringA
 | |
|     else:
 | |
|         def not_windows(*args):
 | |
|             raise OSError("IPv6 addresses cannot be handled on Windows. "
 | |
|                             "Missing ctypes.windll")
 | |
|         WSAStringToAddressA = not_windows
 | |
|         WSAAddressToStringA = not_windows
 | |
| 
 | |
|     def inet_pton(address_family, ip_string):
 | |
|         if address_family == socket.AF_INET:
 | |
|             return socket.inet_aton(ip_string)
 | |
| 
 | |
|         addr = sockaddr()
 | |
|         addr.sa_family = address_family
 | |
|         addr_size = ctypes.c_int(ctypes.sizeof(addr))
 | |
| 
 | |
|         if WSAStringToAddressA(
 | |
|                 ip_string,
 | |
|                 address_family,
 | |
|                 None,
 | |
|                 ctypes.byref(addr),
 | |
|                 ctypes.byref(addr_size)
 | |
|         ) != 0:
 | |
|             raise socket.error(ctypes.FormatError())
 | |
| 
 | |
|         if address_family == socket.AF_INET6:
 | |
|             return ctypes.string_at(addr.ipv6_addr, 16)
 | |
| 
 | |
|         raise socket.error('unknown address family')
 | |
| 
 | |
|     def inet_ntop(address_family, packed_ip):
 | |
|         if address_family == socket.AF_INET:
 | |
|             return socket.inet_ntoa(packed_ip)
 | |
| 
 | |
|         addr = sockaddr()
 | |
|         addr.sa_family = address_family
 | |
|         addr_size = ctypes.c_int(ctypes.sizeof(addr))
 | |
|         ip_string = ctypes.create_string_buffer(128)
 | |
|         ip_string_size = ctypes.c_int(ctypes.sizeof(ip_string))
 | |
| 
 | |
|         if address_family == socket.AF_INET6:
 | |
|             if len(packed_ip) != ctypes.sizeof(addr.ipv6_addr):
 | |
|                 raise socket.error('packed IP wrong length for inet_ntoa')
 | |
|             ctypes.memmove(addr.ipv6_addr, packed_ip, 16)
 | |
|         else:
 | |
|             raise socket.error('unknown address family')
 | |
| 
 | |
|         if WSAAddressToStringA(
 | |
|                 ctypes.byref(addr),
 | |
|                 addr_size,
 | |
|                 None,
 | |
|                 ip_string,
 | |
|                 ctypes.byref(ip_string_size)
 | |
|         ) != 0:
 | |
|             raise socket.error(ctypes.FormatError())
 | |
| 
 | |
|         return ip_string[:ip_string_size.value - 1]
 | |
| 
 | |
| 
 | |
| import keyword
 | |
| 
 | |
| 
 | |
| # similar to collections.namedtuple, reproduced here because Python 2.6 did not have the rename logic
 | |
| def _positional_rename_invalid_identifiers(field_names):
 | |
|     names_out = list(field_names)
 | |
|     for index, name in enumerate(field_names):
 | |
|         if (not all(c.isalnum() or c == '_' for c in name)
 | |
|             or keyword.iskeyword(name)
 | |
|             or not name
 | |
|             or name[0].isdigit()
 | |
|             or name.startswith('_')):
 | |
|             names_out[index] = 'field_%d_' % index
 | |
|     return names_out
 | |
| 
 | |
| 
 | |
| def _sanitize_identifiers(field_names):
 | |
|     names_out = _positional_rename_invalid_identifiers(field_names)
 | |
|     if len(names_out) != len(set(names_out)):
 | |
|         observed_names = set()
 | |
|         for index, name in enumerate(names_out):
 | |
|             while names_out[index] in observed_names:
 | |
|                 names_out[index] = "%s_" % (names_out[index],)
 | |
|             observed_names.add(names_out[index])
 | |
|     return names_out
 | 
