 ae18522ef2
			
		
	
	ae18522ef2
	
	
	
		
			
			* Write flags as a 32-bit unsigned int, not int. * Forgot to import the uint32_pack function from the marshal package.
		
			
				
	
	
		
			1343 lines
		
	
	
		
			41 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			1343 lines
		
	
	
		
			41 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| # Copyright 2013-2016 DataStax, Inc.
 | |
| #
 | |
| # Licensed under the Apache License, Version 2.0 (the "License");
 | |
| # you may not use this file except in compliance with the License.
 | |
| # You may obtain a copy of the License at
 | |
| #
 | |
| # http://www.apache.org/licenses/LICENSE-2.0
 | |
| #
 | |
| # Unless required by applicable law or agreed to in writing, software
 | |
| # distributed under the License is distributed on an "AS IS" BASIS,
 | |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| # See the License for the specific language governing permissions and
 | |
| # limitations under the License.
 | |
| 
 | |
| from __future__ import absolute_import  # to enable import io from stdlib
 | |
| from collections import namedtuple
 | |
| import logging
 | |
| import socket
 | |
| from uuid import UUID
 | |
| 
 | |
| import six
 | |
| from six.moves import range
 | |
| import io
 | |
| 
 | |
| from cassandra import type_codes, DriverException
 | |
| from cassandra import (Unavailable, WriteTimeout, ReadTimeout,
 | |
|                        WriteFailure, ReadFailure, FunctionFailure,
 | |
|                        AlreadyExists, InvalidRequest, Unauthorized,
 | |
|                        UnsupportedOperation, UserFunctionDescriptor,
 | |
|                        UserAggregateDescriptor, SchemaTargetType)
 | |
| from cassandra.marshal import (int32_pack, int32_unpack, uint16_pack, uint16_unpack,
 | |
|                                int8_pack, int8_unpack, uint64_pack, header_pack,
 | |
|                                v3_header_pack, uint32_pack)
 | |
| from cassandra.cqltypes import (AsciiType, BytesType, BooleanType,
 | |
|                                 CounterColumnType, DateType, DecimalType,
 | |
|                                 DoubleType, FloatType, Int32Type,
 | |
|                                 InetAddressType, IntegerType, ListType,
 | |
|                                 LongType, MapType, SetType, TimeUUIDType,
 | |
|                                 UTF8Type, VarcharType, UUIDType, UserType,
 | |
|                                 TupleType, lookup_casstype, SimpleDateType,
 | |
|                                 TimeType, ByteType, ShortType)
 | |
| from cassandra.policies import WriteType
 | |
| from cassandra.cython_deps import HAVE_CYTHON, HAVE_NUMPY
 | |
| from cassandra import util
 | |
| 
 | |
| log = logging.getLogger(__name__)
 | |
| 
 | |
| 
 | |
| class NotSupportedError(Exception):
 | |
|     pass
 | |
| 
 | |
| 
 | |
| class InternalError(Exception):
 | |
|     pass
 | |
| 
 | |
| ColumnMetadata = namedtuple("ColumnMetadata", ['keyspace_name', 'table_name', 'name', 'type'])
 | |
| 
 | |
| MIN_SUPPORTED_VERSION = 1
 | |
| MAX_SUPPORTED_VERSION = 5
 | |
| 
 | |
| HEADER_DIRECTION_TO_CLIENT = 0x80
 | |
| HEADER_DIRECTION_MASK = 0x80
 | |
| 
 | |
| COMPRESSED_FLAG = 0x01
 | |
| TRACING_FLAG = 0x02
 | |
| CUSTOM_PAYLOAD_FLAG = 0x04
 | |
| WARNING_FLAG = 0x08
 | |
| USE_BETA_FLAG = 0x10
 | |
| USE_BETA_MASK = ~USE_BETA_FLAG
 | |
| 
 | |
| _message_types_by_opcode = {}
 | |
| 
 | |
| _UNSET_VALUE = object()
 | |
| 
 | |
| 
 | |
| def register_class(cls):
 | |
|     _message_types_by_opcode[cls.opcode] = cls
 | |
| 
 | |
| 
 | |
| def get_registered_classes():
 | |
|     return _message_types_by_opcode.copy()
 | |
| 
 | |
| 
 | |
| class _RegisterMessageType(type):
 | |
|     def __init__(cls, name, bases, dct):
 | |
|         if not name.startswith('_'):
 | |
|             register_class(cls)
 | |
| 
 | |
| 
 | |
| @six.add_metaclass(_RegisterMessageType)
 | |
| class _MessageType(object):
 | |
| 
 | |
|     tracing = False
 | |
|     custom_payload = None
 | |
|     warnings = None
 | |
| 
 | |
|     def update_custom_payload(self, other):
 | |
|         if other:
 | |
|             if not self.custom_payload:
 | |
|                 self.custom_payload = {}
 | |
|             self.custom_payload.update(other)
 | |
|             if len(self.custom_payload) > 65535:
 | |
|                 raise ValueError("Custom payload map exceeds max count allowed by protocol (65535)")
 | |
| 
 | |
|     def __repr__(self):
 | |
|         return '<%s(%s)>' % (self.__class__.__name__, ', '.join('%s=%r' % i for i in _get_params(self)))
 | |
| 
 | |
| 
 | |
| def _get_params(message_obj):
 | |
|     base_attrs = dir(_MessageType)
 | |
|     return (
 | |
|         (n, a) for n, a in message_obj.__dict__.items()
 | |
|         if n not in base_attrs and not n.startswith('_') and not callable(a)
 | |
|     )
 | |
| 
 | |
| 
 | |
| error_classes = {}
 | |
| 
 | |
| 
 | |
| class ErrorMessage(_MessageType, Exception):
 | |
|     opcode = 0x00
 | |
|     name = 'ERROR'
 | |
|     summary = 'Unknown'
 | |
| 
 | |
|     def __init__(self, code, message, info):
 | |
|         self.code = code
 | |
|         self.message = message
 | |
|         self.info = info
 | |
| 
 | |
|     @classmethod
 | |
|     def recv_body(cls, f, protocol_version, *args):
 | |
|         code = read_int(f)
 | |
|         msg = read_string(f)
 | |
|         subcls = error_classes.get(code, cls)
 | |
|         extra_info = subcls.recv_error_info(f, protocol_version)
 | |
|         return subcls(code=code, message=msg, info=extra_info)
 | |
| 
 | |
|     def summary_msg(self):
 | |
|         msg = 'Error from server: code=%04x [%s] message="%s"' \
 | |
|               % (self.code, self.summary, self.message)
 | |
|         if six.PY2 and isinstance(msg, six.text_type):
 | |
|             msg = msg.encode('utf-8')
 | |
|         return msg
 | |
| 
 | |
|     def __str__(self):
 | |
|         return '<%s>' % self.summary_msg()
 | |
|     __repr__ = __str__
 | |
| 
 | |
|     @staticmethod
 | |
|     def recv_error_info(f, protocol_version):
 | |
|         pass
 | |
| 
 | |
|     def to_exception(self):
 | |
|         return self
 | |
| 
 | |
| 
 | |
| class ErrorMessageSubclass(_RegisterMessageType):
 | |
|     def __init__(cls, name, bases, dct):
 | |
|         if cls.error_code is not None:  # Server has an error code of 0.
 | |
|             error_classes[cls.error_code] = cls
 | |
| 
 | |
| 
 | |
| @six.add_metaclass(ErrorMessageSubclass)
 | |
| class ErrorMessageSub(ErrorMessage):
 | |
|     error_code = None
 | |
| 
 | |
| 
 | |
| class RequestExecutionException(ErrorMessageSub):
 | |
|     pass
 | |
| 
 | |
| 
 | |
| class RequestValidationException(ErrorMessageSub):
 | |
|     pass
 | |
| 
 | |
| 
 | |
| class ServerError(ErrorMessageSub):
 | |
|     summary = 'Server error'
 | |
|     error_code = 0x0000
 | |
| 
 | |
| 
 | |
| class ProtocolException(ErrorMessageSub):
 | |
|     summary = 'Protocol error'
 | |
|     error_code = 0x000A
 | |
| 
 | |
| 
 | |
| class BadCredentials(ErrorMessageSub):
 | |
|     summary = 'Bad credentials'
 | |
|     error_code = 0x0100
 | |
| 
 | |
| 
 | |
| class UnavailableErrorMessage(RequestExecutionException):
 | |
|     summary = 'Unavailable exception'
 | |
|     error_code = 0x1000
 | |
| 
 | |
|     @staticmethod
 | |
|     def recv_error_info(f, protocol_version):
 | |
|         return {
 | |
|             'consistency': read_consistency_level(f),
 | |
|             'required_replicas': read_int(f),
 | |
|             'alive_replicas': read_int(f),
 | |
|         }
 | |
| 
 | |
|     def to_exception(self):
 | |
|         return Unavailable(self.summary_msg(), **self.info)
 | |
| 
 | |
| 
 | |
| class OverloadedErrorMessage(RequestExecutionException):
 | |
|     summary = 'Coordinator node overloaded'
 | |
|     error_code = 0x1001
 | |
| 
 | |
| 
 | |
| class IsBootstrappingErrorMessage(RequestExecutionException):
 | |
|     summary = 'Coordinator node is bootstrapping'
 | |
|     error_code = 0x1002
 | |
| 
 | |
| 
 | |
| class TruncateError(RequestExecutionException):
 | |
|     summary = 'Error during truncate'
 | |
|     error_code = 0x1003
 | |
| 
 | |
| 
 | |
| class WriteTimeoutErrorMessage(RequestExecutionException):
 | |
|     summary = "Coordinator node timed out waiting for replica nodes' responses"
 | |
|     error_code = 0x1100
 | |
| 
 | |
|     @staticmethod
 | |
|     def recv_error_info(f, protocol_version):
 | |
|         return {
 | |
|             'consistency': read_consistency_level(f),
 | |
|             'received_responses': read_int(f),
 | |
|             'required_responses': read_int(f),
 | |
|             'write_type': WriteType.name_to_value[read_string(f)],
 | |
|         }
 | |
| 
 | |
|     def to_exception(self):
 | |
|         return WriteTimeout(self.summary_msg(), **self.info)
 | |
| 
 | |
| 
 | |
| class ReadTimeoutErrorMessage(RequestExecutionException):
 | |
|     summary = "Coordinator node timed out waiting for replica nodes' responses"
 | |
|     error_code = 0x1200
 | |
| 
 | |
|     @staticmethod
 | |
|     def recv_error_info(f, protocol_version):
 | |
|         return {
 | |
|             'consistency': read_consistency_level(f),
 | |
|             'received_responses': read_int(f),
 | |
|             'required_responses': read_int(f),
 | |
|             'data_retrieved': bool(read_byte(f)),
 | |
|         }
 | |
| 
 | |
|     def to_exception(self):
 | |
|         return ReadTimeout(self.summary_msg(), **self.info)
 | |
| 
 | |
| 
 | |
| class ReadFailureMessage(RequestExecutionException):
 | |
|     summary = "Replica(s) failed to execute read"
 | |
|     error_code = 0x1300
 | |
| 
 | |
|     @staticmethod
 | |
|     def recv_error_info(f, protocol_version):
 | |
|         consistency = read_consistency_level(f)
 | |
|         received_responses = read_int(f)
 | |
|         required_responses = read_int(f)
 | |
| 
 | |
|         if protocol_version >= 5:
 | |
|             error_code_map = read_error_code_map(f)
 | |
|             failures = len(error_code_map)
 | |
|         else:
 | |
|             error_code_map = None
 | |
|             failures = read_int(f)
 | |
| 
 | |
|         data_retrieved = bool(read_byte(f))
 | |
| 
 | |
|         return {
 | |
|             'consistency': consistency,
 | |
|             'received_responses': received_responses,
 | |
|             'required_responses': required_responses,
 | |
|             'failures': failures,
 | |
|             'error_code_map': error_code_map,
 | |
|             'data_retrieved': data_retrieved
 | |
|         }
 | |
| 
 | |
|     def to_exception(self):
 | |
|         return ReadFailure(self.summary_msg(), **self.info)
 | |
| 
 | |
| 
 | |
| class FunctionFailureMessage(RequestExecutionException):
 | |
|     summary = "User Defined Function failure"
 | |
|     error_code = 0x1400
 | |
| 
 | |
|     @staticmethod
 | |
|     def recv_error_info(f, protocol_version):
 | |
|         return {
 | |
|             'keyspace': read_string(f),
 | |
|             'function': read_string(f),
 | |
|             'arg_types': [read_string(f) for _ in range(read_short(f))],
 | |
|         }
 | |
| 
 | |
|     def to_exception(self):
 | |
|         return FunctionFailure(self.summary_msg(), **self.info)
 | |
| 
 | |
| 
 | |
| class WriteFailureMessage(RequestExecutionException):
 | |
|     summary = "Replica(s) failed to execute write"
 | |
|     error_code = 0x1500
 | |
| 
 | |
|     @staticmethod
 | |
|     def recv_error_info(f, protocol_version):
 | |
|         consistency = read_consistency_level(f)
 | |
|         received_responses = read_int(f)
 | |
|         required_responses = read_int(f)
 | |
| 
 | |
|         if protocol_version >= 5:
 | |
|             error_code_map = read_error_code_map(f)
 | |
|             failures = len(error_code_map)
 | |
|         else:
 | |
|             error_code_map = None
 | |
|             failures = read_int(f)
 | |
| 
 | |
|         write_type = WriteType.name_to_value[read_string(f)]
 | |
| 
 | |
|         return {
 | |
|             'consistency': consistency,
 | |
|             'received_responses': received_responses,
 | |
|             'required_responses': required_responses,
 | |
|             'failures': failures,
 | |
|             'error_code_map': error_code_map,
 | |
|             'write_type': write_type
 | |
|         }
 | |
| 
 | |
|     def to_exception(self):
 | |
|         return WriteFailure(self.summary_msg(), **self.info)
 | |
| 
 | |
| 
 | |
| class SyntaxException(RequestValidationException):
 | |
|     summary = 'Syntax error in CQL query'
 | |
|     error_code = 0x2000
 | |
| 
 | |
| 
 | |
| class UnauthorizedErrorMessage(RequestValidationException):
 | |
|     summary = 'Unauthorized'
 | |
|     error_code = 0x2100
 | |
| 
 | |
|     def to_exception(self):
 | |
|         return Unauthorized(self.summary_msg())
 | |
| 
 | |
| 
 | |
| class InvalidRequestException(RequestValidationException):
 | |
|     summary = 'Invalid query'
 | |
|     error_code = 0x2200
 | |
| 
 | |
|     def to_exception(self):
 | |
|         return InvalidRequest(self.summary_msg())
 | |
| 
 | |
| 
 | |
| class ConfigurationException(RequestValidationException):
 | |
|     summary = 'Query invalid because of configuration issue'
 | |
|     error_code = 0x2300
 | |
| 
 | |
| 
 | |
| class PreparedQueryNotFound(RequestValidationException):
 | |
|     summary = 'Matching prepared statement not found on this node'
 | |
|     error_code = 0x2500
 | |
| 
 | |
|     @staticmethod
 | |
|     def recv_error_info(f, protocol_version):
 | |
|         # return the query ID
 | |
|         return read_binary_string(f)
 | |
| 
 | |
| 
 | |
| class AlreadyExistsException(ConfigurationException):
 | |
|     summary = 'Item already exists'
 | |
|     error_code = 0x2400
 | |
| 
 | |
|     @staticmethod
 | |
|     def recv_error_info(f, protocol_version):
 | |
|         return {
 | |
|             'keyspace': read_string(f),
 | |
|             'table': read_string(f),
 | |
|         }
 | |
| 
 | |
|     def to_exception(self):
 | |
|         return AlreadyExists(**self.info)
 | |
| 
 | |
| 
 | |
| class StartupMessage(_MessageType):
 | |
|     opcode = 0x01
 | |
|     name = 'STARTUP'
 | |
| 
 | |
|     KNOWN_OPTION_KEYS = set((
 | |
|         'CQL_VERSION',
 | |
|         'COMPRESSION',
 | |
|     ))
 | |
| 
 | |
|     def __init__(self, cqlversion, options):
 | |
|         self.cqlversion = cqlversion
 | |
|         self.options = options
 | |
| 
 | |
|     def send_body(self, f, protocol_version):
 | |
|         optmap = self.options.copy()
 | |
|         optmap['CQL_VERSION'] = self.cqlversion
 | |
|         write_stringmap(f, optmap)
 | |
| 
 | |
| 
 | |
| class ReadyMessage(_MessageType):
 | |
|     opcode = 0x02
 | |
|     name = 'READY'
 | |
| 
 | |
|     @classmethod
 | |
|     def recv_body(cls, *args):
 | |
|         return cls()
 | |
| 
 | |
| 
 | |
| class AuthenticateMessage(_MessageType):
 | |
|     opcode = 0x03
 | |
|     name = 'AUTHENTICATE'
 | |
| 
 | |
|     def __init__(self, authenticator):
 | |
|         self.authenticator = authenticator
 | |
| 
 | |
|     @classmethod
 | |
|     def recv_body(cls, f, *args):
 | |
|         authname = read_string(f)
 | |
|         return cls(authenticator=authname)
 | |
| 
 | |
| 
 | |
| class CredentialsMessage(_MessageType):
 | |
|     opcode = 0x04
 | |
|     name = 'CREDENTIALS'
 | |
| 
 | |
|     def __init__(self, creds):
 | |
|         self.creds = creds
 | |
| 
 | |
|     def send_body(self, f, protocol_version):
 | |
|         if protocol_version > 1:
 | |
|             raise UnsupportedOperation(
 | |
|                 "Credentials-based authentication is not supported with "
 | |
|                 "protocol version 2 or higher.  Use the SASL authentication "
 | |
|                 "mechanism instead.")
 | |
|         write_short(f, len(self.creds))
 | |
|         for credkey, credval in self.creds.items():
 | |
|             write_string(f, credkey)
 | |
|             write_string(f, credval)
 | |
| 
 | |
| 
 | |
| class AuthChallengeMessage(_MessageType):
 | |
|     opcode = 0x0E
 | |
|     name = 'AUTH_CHALLENGE'
 | |
| 
 | |
|     def __init__(self, challenge):
 | |
|         self.challenge = challenge
 | |
| 
 | |
|     @classmethod
 | |
|     def recv_body(cls, f, *args):
 | |
|         return cls(read_binary_longstring(f))
 | |
| 
 | |
| 
 | |
| class AuthResponseMessage(_MessageType):
 | |
|     opcode = 0x0F
 | |
|     name = 'AUTH_RESPONSE'
 | |
| 
 | |
|     def __init__(self, response):
 | |
|         self.response = response
 | |
| 
 | |
|     def send_body(self, f, protocol_version):
 | |
|         write_longstring(f, self.response)
 | |
| 
 | |
| 
 | |
| class AuthSuccessMessage(_MessageType):
 | |
|     opcode = 0x10
 | |
|     name = 'AUTH_SUCCESS'
 | |
| 
 | |
|     def __init__(self, token):
 | |
|         self.token = token
 | |
| 
 | |
|     @classmethod
 | |
|     def recv_body(cls, f, *args):
 | |
|         return cls(read_longstring(f))
 | |
| 
 | |
| 
 | |
| class OptionsMessage(_MessageType):
 | |
|     opcode = 0x05
 | |
|     name = 'OPTIONS'
 | |
| 
 | |
|     def send_body(self, f, protocol_version):
 | |
|         pass
 | |
| 
 | |
| 
 | |
| class SupportedMessage(_MessageType):
 | |
|     opcode = 0x06
 | |
|     name = 'SUPPORTED'
 | |
| 
 | |
|     def __init__(self, cql_versions, options):
 | |
|         self.cql_versions = cql_versions
 | |
|         self.options = options
 | |
| 
 | |
|     @classmethod
 | |
|     def recv_body(cls, f, *args):
 | |
|         options = read_stringmultimap(f)
 | |
|         cql_versions = options.pop('CQL_VERSION')
 | |
|         return cls(cql_versions=cql_versions, options=options)
 | |
| 
 | |
| 
 | |
| # used for QueryMessage and ExecuteMessage
 | |
| _VALUES_FLAG = 0x01
 | |
| _SKIP_METADATA_FLAG = 0x02
 | |
| _PAGE_SIZE_FLAG = 0x04
 | |
| _WITH_PAGING_STATE_FLAG = 0x08
 | |
| _WITH_SERIAL_CONSISTENCY_FLAG = 0x10
 | |
| _PROTOCOL_TIMESTAMP = 0x20
 | |
| 
 | |
| 
 | |
| class QueryMessage(_MessageType):
 | |
|     opcode = 0x07
 | |
|     name = 'QUERY'
 | |
| 
 | |
|     def __init__(self, query, consistency_level, serial_consistency_level=None,
 | |
|                  fetch_size=None, paging_state=None, timestamp=None):
 | |
|         self.query = query
 | |
|         self.consistency_level = consistency_level
 | |
|         self.serial_consistency_level = serial_consistency_level
 | |
|         self.fetch_size = fetch_size
 | |
|         self.paging_state = paging_state
 | |
|         self.timestamp = timestamp
 | |
|         self._query_params = None  # only used internally. May be set to a list of native-encoded values to have them sent with the request.
 | |
| 
 | |
|     def send_body(self, f, protocol_version):
 | |
|         write_longstring(f, self.query)
 | |
|         write_consistency_level(f, self.consistency_level)
 | |
|         flags = 0x00
 | |
|         if self._query_params is not None:
 | |
|             flags |= _VALUES_FLAG  # also v2+, but we're only setting params internally right now
 | |
| 
 | |
|         if self.serial_consistency_level:
 | |
|             if protocol_version >= 2:
 | |
|                 flags |= _WITH_SERIAL_CONSISTENCY_FLAG
 | |
|             else:
 | |
|                 raise UnsupportedOperation(
 | |
|                     "Serial consistency levels require the use of protocol version "
 | |
|                     "2 or higher. Consider setting Cluster.protocol_version to 2 "
 | |
|                     "to support serial consistency levels.")
 | |
| 
 | |
|         if self.fetch_size:
 | |
|             if protocol_version >= 2:
 | |
|                 flags |= _PAGE_SIZE_FLAG
 | |
|             else:
 | |
|                 raise UnsupportedOperation(
 | |
|                     "Automatic query paging may only be used with protocol version "
 | |
|                     "2 or higher. Consider setting Cluster.protocol_version to 2.")
 | |
| 
 | |
|         if self.paging_state:
 | |
|             if protocol_version >= 2:
 | |
|                 flags |= _WITH_PAGING_STATE_FLAG
 | |
|             else:
 | |
|                 raise UnsupportedOperation(
 | |
|                     "Automatic query paging may only be used with protocol version "
 | |
|                     "2 or higher. Consider setting Cluster.protocol_version to 2.")
 | |
| 
 | |
|         if self.timestamp is not None:
 | |
|             flags |= _PROTOCOL_TIMESTAMP
 | |
| 
 | |
|         if protocol_version >= 5:
 | |
|             write_uint(f, flags)
 | |
|         else:
 | |
|             write_byte(f, flags)
 | |
| 
 | |
|         if self._query_params is not None:
 | |
|             write_short(f, len(self._query_params))
 | |
|             for param in self._query_params:
 | |
|                 write_value(f, param)
 | |
| 
 | |
|         if self.fetch_size:
 | |
|             write_int(f, self.fetch_size)
 | |
|         if self.paging_state:
 | |
|             write_longstring(f, self.paging_state)
 | |
|         if self.serial_consistency_level:
 | |
|             write_consistency_level(f, self.serial_consistency_level)
 | |
|         if self.timestamp is not None:
 | |
|             write_long(f, self.timestamp)
 | |
| 
 | |
| 
 | |
| CUSTOM_TYPE = object()
 | |
| 
 | |
| RESULT_KIND_VOID = 0x0001
 | |
| RESULT_KIND_ROWS = 0x0002
 | |
| RESULT_KIND_SET_KEYSPACE = 0x0003
 | |
| RESULT_KIND_PREPARED = 0x0004
 | |
| RESULT_KIND_SCHEMA_CHANGE = 0x0005
 | |
| 
 | |
| 
 | |
| class ResultMessage(_MessageType):
 | |
|     opcode = 0x08
 | |
|     name = 'RESULT'
 | |
| 
 | |
|     kind = None
 | |
|     results = None
 | |
|     paging_state = None
 | |
| 
 | |
|     # Names match type name in module scope. Most are imported from cassandra.cqltypes (except CUSTOM_TYPE)
 | |
|     type_codes = _cqltypes_by_code = dict((v, globals()[k]) for k, v in type_codes.__dict__.items() if not k.startswith('_'))
 | |
| 
 | |
|     _FLAGS_GLOBAL_TABLES_SPEC = 0x0001
 | |
|     _HAS_MORE_PAGES_FLAG = 0x0002
 | |
|     _NO_METADATA_FLAG = 0x0004
 | |
| 
 | |
|     def __init__(self, kind, results, paging_state=None, col_types=None):
 | |
|         self.kind = kind
 | |
|         self.results = results
 | |
|         self.paging_state = paging_state
 | |
|         self.col_types = col_types
 | |
| 
 | |
|     @classmethod
 | |
|     def recv_body(cls, f, protocol_version, user_type_map, result_metadata):
 | |
|         kind = read_int(f)
 | |
|         paging_state = None
 | |
|         col_types = None
 | |
|         if kind == RESULT_KIND_VOID:
 | |
|             results = None
 | |
|         elif kind == RESULT_KIND_ROWS:
 | |
|             paging_state, col_types, results = cls.recv_results_rows(
 | |
|                 f, protocol_version, user_type_map, result_metadata)
 | |
|         elif kind == RESULT_KIND_SET_KEYSPACE:
 | |
|             ksname = read_string(f)
 | |
|             results = ksname
 | |
|         elif kind == RESULT_KIND_PREPARED:
 | |
|             results = cls.recv_results_prepared(f, protocol_version, user_type_map)
 | |
|         elif kind == RESULT_KIND_SCHEMA_CHANGE:
 | |
|             results = cls.recv_results_schema_change(f, protocol_version)
 | |
|         else:
 | |
|             raise DriverException("Unknown RESULT kind: %d" % kind)
 | |
|         return cls(kind, results, paging_state, col_types)
 | |
| 
 | |
|     @classmethod
 | |
|     def recv_results_rows(cls, f, protocol_version, user_type_map, result_metadata):
 | |
|         paging_state, column_metadata = cls.recv_results_metadata(f, user_type_map)
 | |
|         column_metadata = column_metadata or result_metadata
 | |
|         rowcount = read_int(f)
 | |
|         rows = [cls.recv_row(f, len(column_metadata)) for _ in range(rowcount)]
 | |
|         colnames = [c[2] for c in column_metadata]
 | |
|         coltypes = [c[3] for c in column_metadata]
 | |
|         try:
 | |
|             parsed_rows = [
 | |
|                 tuple(ctype.from_binary(val, protocol_version)
 | |
|                       for ctype, val in zip(coltypes, row))
 | |
|                 for row in rows]
 | |
|         except Exception:
 | |
|             for row in rows:
 | |
|                 for i in range(len(row)):
 | |
|                     try:
 | |
|                         coltypes[i].from_binary(row[i], protocol_version)
 | |
|                     except Exception as e:
 | |
|                         raise DriverException('Failed decoding result column "%s" of type %s: %s' % (colnames[i],
 | |
|                                                                                                      coltypes[i].cql_parameterized_type(),
 | |
|                                                                                                      str(e)))
 | |
|         return paging_state, coltypes, (colnames, parsed_rows)
 | |
| 
 | |
|     @classmethod
 | |
|     def recv_results_prepared(cls, f, protocol_version, user_type_map):
 | |
|         query_id = read_binary_string(f)
 | |
|         bind_metadata, pk_indexes, result_metadata = cls.recv_prepared_metadata(f, protocol_version, user_type_map)
 | |
|         return query_id, bind_metadata, pk_indexes, result_metadata
 | |
| 
 | |
|     @classmethod
 | |
|     def recv_results_metadata(cls, f, user_type_map):
 | |
|         flags = read_int(f)
 | |
|         colcount = read_int(f)
 | |
| 
 | |
|         if flags & cls._HAS_MORE_PAGES_FLAG:
 | |
|             paging_state = read_binary_longstring(f)
 | |
|         else:
 | |
|             paging_state = None
 | |
| 
 | |
|         no_meta = bool(flags & cls._NO_METADATA_FLAG)
 | |
|         if no_meta:
 | |
|             return paging_state, []
 | |
| 
 | |
|         glob_tblspec = bool(flags & cls._FLAGS_GLOBAL_TABLES_SPEC)
 | |
|         if glob_tblspec:
 | |
|             ksname = read_string(f)
 | |
|             cfname = read_string(f)
 | |
|         column_metadata = []
 | |
|         for _ in range(colcount):
 | |
|             if glob_tblspec:
 | |
|                 colksname = ksname
 | |
|                 colcfname = cfname
 | |
|             else:
 | |
|                 colksname = read_string(f)
 | |
|                 colcfname = read_string(f)
 | |
|             colname = read_string(f)
 | |
|             coltype = cls.read_type(f, user_type_map)
 | |
|             column_metadata.append((colksname, colcfname, colname, coltype))
 | |
|         return paging_state, column_metadata
 | |
| 
 | |
|     @classmethod
 | |
|     def recv_prepared_metadata(cls, f, protocol_version, user_type_map):
 | |
|         flags = read_int(f)
 | |
|         colcount = read_int(f)
 | |
|         pk_indexes = None
 | |
|         if protocol_version >= 4:
 | |
|             num_pk_indexes = read_int(f)
 | |
|             pk_indexes = [read_short(f) for _ in range(num_pk_indexes)]
 | |
| 
 | |
|         glob_tblspec = bool(flags & cls._FLAGS_GLOBAL_TABLES_SPEC)
 | |
|         if glob_tblspec:
 | |
|             ksname = read_string(f)
 | |
|             cfname = read_string(f)
 | |
|         bind_metadata = []
 | |
|         for _ in range(colcount):
 | |
|             if glob_tblspec:
 | |
|                 colksname = ksname
 | |
|                 colcfname = cfname
 | |
|             else:
 | |
|                 colksname = read_string(f)
 | |
|                 colcfname = read_string(f)
 | |
|             colname = read_string(f)
 | |
|             coltype = cls.read_type(f, user_type_map)
 | |
|             bind_metadata.append(ColumnMetadata(colksname, colcfname, colname, coltype))
 | |
| 
 | |
|         if protocol_version >= 2:
 | |
|             _, result_metadata = cls.recv_results_metadata(f, user_type_map)
 | |
|             return bind_metadata, pk_indexes, result_metadata
 | |
|         else:
 | |
|             return bind_metadata, pk_indexes, None
 | |
| 
 | |
|     @classmethod
 | |
|     def recv_results_schema_change(cls, f, protocol_version):
 | |
|         return EventMessage.recv_schema_change(f, protocol_version)
 | |
| 
 | |
|     @classmethod
 | |
|     def read_type(cls, f, user_type_map):
 | |
|         optid = read_short(f)
 | |
|         try:
 | |
|             typeclass = cls.type_codes[optid]
 | |
|         except KeyError:
 | |
|             raise NotSupportedError("Unknown data type code 0x%04x. Have to skip"
 | |
|                                     " entire result set." % (optid,))
 | |
|         if typeclass in (ListType, SetType):
 | |
|             subtype = cls.read_type(f, user_type_map)
 | |
|             typeclass = typeclass.apply_parameters((subtype,))
 | |
|         elif typeclass == MapType:
 | |
|             keysubtype = cls.read_type(f, user_type_map)
 | |
|             valsubtype = cls.read_type(f, user_type_map)
 | |
|             typeclass = typeclass.apply_parameters((keysubtype, valsubtype))
 | |
|         elif typeclass == TupleType:
 | |
|             num_items = read_short(f)
 | |
|             types = tuple(cls.read_type(f, user_type_map) for _ in range(num_items))
 | |
|             typeclass = typeclass.apply_parameters(types)
 | |
|         elif typeclass == UserType:
 | |
|             ks = read_string(f)
 | |
|             udt_name = read_string(f)
 | |
|             num_fields = read_short(f)
 | |
|             names, types = zip(*((read_string(f), cls.read_type(f, user_type_map))
 | |
|                                  for _ in range(num_fields)))
 | |
|             specialized_type = typeclass.make_udt_class(ks, udt_name, names, types)
 | |
|             specialized_type.mapped_class = user_type_map.get(ks, {}).get(udt_name)
 | |
|             typeclass = specialized_type
 | |
|         elif typeclass == CUSTOM_TYPE:
 | |
|             classname = read_string(f)
 | |
|             typeclass = lookup_casstype(classname)
 | |
| 
 | |
|         return typeclass
 | |
| 
 | |
|     @staticmethod
 | |
|     def recv_row(f, colcount):
 | |
|         return [read_value(f) for _ in range(colcount)]
 | |
| 
 | |
| 
 | |
| class PrepareMessage(_MessageType):
 | |
|     opcode = 0x09
 | |
|     name = 'PREPARE'
 | |
| 
 | |
|     def __init__(self, query):
 | |
|         self.query = query
 | |
| 
 | |
|     def send_body(self, f, protocol_version):
 | |
|         write_longstring(f, self.query)
 | |
|         if protocol_version >= 5:
 | |
|             # Write the flags byte; with 0 value for now, but this should change in PYTHON-678
 | |
|             write_uint(f, 0)
 | |
| 
 | |
| 
 | |
| class ExecuteMessage(_MessageType):
 | |
|     opcode = 0x0A
 | |
|     name = 'EXECUTE'
 | |
| 
 | |
|     def __init__(self, query_id, query_params, consistency_level,
 | |
|                  serial_consistency_level=None, fetch_size=None,
 | |
|                  paging_state=None, timestamp=None, skip_meta=False):
 | |
|         self.query_id = query_id
 | |
|         self.query_params = query_params
 | |
|         self.consistency_level = consistency_level
 | |
|         self.serial_consistency_level = serial_consistency_level
 | |
|         self.fetch_size = fetch_size
 | |
|         self.paging_state = paging_state
 | |
|         self.timestamp = timestamp
 | |
|         self.skip_meta = skip_meta
 | |
| 
 | |
|     def send_body(self, f, protocol_version):
 | |
|         write_string(f, self.query_id)
 | |
|         if protocol_version == 1:
 | |
|             if self.serial_consistency_level:
 | |
|                 raise UnsupportedOperation(
 | |
|                     "Serial consistency levels require the use of protocol version "
 | |
|                     "2 or higher. Consider setting Cluster.protocol_version to 2 "
 | |
|                     "to support serial consistency levels.")
 | |
|             if self.fetch_size or self.paging_state:
 | |
|                 raise UnsupportedOperation(
 | |
|                     "Automatic query paging may only be used with protocol version "
 | |
|                     "2 or higher. Consider setting Cluster.protocol_version to 2.")
 | |
|             write_short(f, len(self.query_params))
 | |
|             for param in self.query_params:
 | |
|                 write_value(f, param)
 | |
|             write_consistency_level(f, self.consistency_level)
 | |
|         else:
 | |
|             write_consistency_level(f, self.consistency_level)
 | |
|             flags = _VALUES_FLAG
 | |
|             if self.serial_consistency_level:
 | |
|                 flags |= _WITH_SERIAL_CONSISTENCY_FLAG
 | |
|             if self.fetch_size:
 | |
|                 flags |= _PAGE_SIZE_FLAG
 | |
|             if self.paging_state:
 | |
|                 flags |= _WITH_PAGING_STATE_FLAG
 | |
|             if self.timestamp is not None:
 | |
|                 if protocol_version >= 3:
 | |
|                     flags |= _PROTOCOL_TIMESTAMP
 | |
|                 else:
 | |
|                     raise UnsupportedOperation(
 | |
|                         "Protocol-level timestamps may only be used with protocol version "
 | |
|                         "3 or higher. Consider setting Cluster.protocol_version to 3.")
 | |
|             if self.skip_meta:
 | |
|                 flags |= _SKIP_METADATA_FLAG
 | |
| 
 | |
|             if protocol_version >= 5:
 | |
|                 write_uint(f, flags)
 | |
|             else:
 | |
|                 write_byte(f, flags)
 | |
| 
 | |
|             write_short(f, len(self.query_params))
 | |
|             for param in self.query_params:
 | |
|                 write_value(f, param)
 | |
|             if self.fetch_size:
 | |
|                 write_int(f, self.fetch_size)
 | |
|             if self.paging_state:
 | |
|                 write_longstring(f, self.paging_state)
 | |
|             if self.serial_consistency_level:
 | |
|                 write_consistency_level(f, self.serial_consistency_level)
 | |
|             if self.timestamp is not None:
 | |
|                 write_long(f, self.timestamp)
 | |
| 
 | |
| 
 | |
| 
 | |
| class BatchMessage(_MessageType):
 | |
|     opcode = 0x0D
 | |
|     name = 'BATCH'
 | |
| 
 | |
|     def __init__(self, batch_type, queries, consistency_level,
 | |
|                  serial_consistency_level=None, timestamp=None):
 | |
|         self.batch_type = batch_type
 | |
|         self.queries = queries
 | |
|         self.consistency_level = consistency_level
 | |
|         self.serial_consistency_level = serial_consistency_level
 | |
|         self.timestamp = timestamp
 | |
| 
 | |
|     def send_body(self, f, protocol_version):
 | |
|         write_byte(f, self.batch_type.value)
 | |
|         write_short(f, len(self.queries))
 | |
|         for prepared, string_or_query_id, params in self.queries:
 | |
|             if not prepared:
 | |
|                 write_byte(f, 0)
 | |
|                 write_longstring(f, string_or_query_id)
 | |
|             else:
 | |
|                 write_byte(f, 1)
 | |
|                 write_short(f, len(string_or_query_id))
 | |
|                 f.write(string_or_query_id)
 | |
|             write_short(f, len(params))
 | |
|             for param in params:
 | |
|                 write_value(f, param)
 | |
| 
 | |
|         write_consistency_level(f, self.consistency_level)
 | |
|         if protocol_version >= 3:
 | |
|             flags = 0
 | |
|             if self.serial_consistency_level:
 | |
|                 flags |= _WITH_SERIAL_CONSISTENCY_FLAG
 | |
|             if self.timestamp is not None:
 | |
|                 flags |= _PROTOCOL_TIMESTAMP
 | |
| 
 | |
|             if protocol_version >= 5:
 | |
|                 write_int(f, flags)
 | |
|             else:
 | |
|                 write_byte(f, flags)
 | |
| 
 | |
|             if self.serial_consistency_level:
 | |
|                 write_consistency_level(f, self.serial_consistency_level)
 | |
|             if self.timestamp is not None:
 | |
|                 write_long(f, self.timestamp)
 | |
| 
 | |
| 
 | |
| known_event_types = frozenset((
 | |
|     'TOPOLOGY_CHANGE',
 | |
|     'STATUS_CHANGE',
 | |
|     'SCHEMA_CHANGE'
 | |
| ))
 | |
| 
 | |
| 
 | |
| class RegisterMessage(_MessageType):
 | |
|     opcode = 0x0B
 | |
|     name = 'REGISTER'
 | |
| 
 | |
|     def __init__(self, event_list):
 | |
|         self.event_list = event_list
 | |
| 
 | |
|     def send_body(self, f, protocol_version):
 | |
|         write_stringlist(f, self.event_list)
 | |
| 
 | |
| 
 | |
| class EventMessage(_MessageType):
 | |
|     opcode = 0x0C
 | |
|     name = 'EVENT'
 | |
| 
 | |
|     def __init__(self, event_type, event_args):
 | |
|         self.event_type = event_type
 | |
|         self.event_args = event_args
 | |
| 
 | |
|     @classmethod
 | |
|     def recv_body(cls, f, protocol_version, *args):
 | |
|         event_type = read_string(f).upper()
 | |
|         if event_type in known_event_types:
 | |
|             read_method = getattr(cls, 'recv_' + event_type.lower())
 | |
|             return cls(event_type=event_type, event_args=read_method(f, protocol_version))
 | |
|         raise NotSupportedError('Unknown event type %r' % event_type)
 | |
| 
 | |
|     @classmethod
 | |
|     def recv_topology_change(cls, f, protocol_version):
 | |
|         # "NEW_NODE" or "REMOVED_NODE"
 | |
|         change_type = read_string(f)
 | |
|         address = read_inet(f)
 | |
|         return dict(change_type=change_type, address=address)
 | |
| 
 | |
|     @classmethod
 | |
|     def recv_status_change(cls, f, protocol_version):
 | |
|         # "UP" or "DOWN"
 | |
|         change_type = read_string(f)
 | |
|         address = read_inet(f)
 | |
|         return dict(change_type=change_type, address=address)
 | |
| 
 | |
|     @classmethod
 | |
|     def recv_schema_change(cls, f, protocol_version):
 | |
|         # "CREATED", "DROPPED", or "UPDATED"
 | |
|         change_type = read_string(f)
 | |
|         if protocol_version >= 3:
 | |
|             target = read_string(f)
 | |
|             keyspace = read_string(f)
 | |
|             event = {'target_type': target, 'change_type': change_type, 'keyspace': keyspace}
 | |
|             if target != SchemaTargetType.KEYSPACE:
 | |
|                 target_name = read_string(f)
 | |
|                 if target == SchemaTargetType.FUNCTION:
 | |
|                     event['function'] = UserFunctionDescriptor(target_name, [read_string(f) for _ in range(read_short(f))])
 | |
|                 elif target == SchemaTargetType.AGGREGATE:
 | |
|                     event['aggregate'] = UserAggregateDescriptor(target_name, [read_string(f) for _ in range(read_short(f))])
 | |
|                 else:
 | |
|                     event[target.lower()] = target_name
 | |
|         else:
 | |
|             keyspace = read_string(f)
 | |
|             table = read_string(f)
 | |
|             if table:
 | |
|                 event = {'target_type': SchemaTargetType.TABLE, 'change_type': change_type, 'keyspace': keyspace, 'table': table}
 | |
|             else:
 | |
|                 event = {'target_type': SchemaTargetType.KEYSPACE, 'change_type': change_type, 'keyspace': keyspace}
 | |
|         return event
 | |
| 
 | |
| 
 | |
| class _ProtocolHandler(object):
 | |
|     """
 | |
|     _ProtocolHander handles encoding and decoding messages.
 | |
| 
 | |
|     This class can be specialized to compose Handlers which implement alternative
 | |
|     result decoding or type deserialization. Class definitions are passed to :class:`cassandra.cluster.Cluster`
 | |
|     on initialization.
 | |
| 
 | |
|     Contracted class methods are :meth:`_ProtocolHandler.encode_message` and :meth:`_ProtocolHandler.decode_message`.
 | |
|     """
 | |
| 
 | |
|     message_types_by_opcode = _message_types_by_opcode.copy()
 | |
|     """
 | |
|     Default mapping of opcode to Message implementation. The default ``decode_message`` implementation uses
 | |
|     this to instantiate a message and populate using ``recv_body``. This mapping can be updated to inject specialized
 | |
|     result decoding implementations.
 | |
|     """
 | |
| 
 | |
|     @classmethod
 | |
|     def encode_message(cls, msg, stream_id, protocol_version, compressor, allow_beta_protocol_version):
 | |
|         """
 | |
|         Encodes a message using the specified frame parameters, and compressor
 | |
| 
 | |
|         :param msg: the message, typically of cassandra.protocol._MessageType, generated by the driver
 | |
|         :param stream_id: protocol stream id for the frame header
 | |
|         :param protocol_version: version for the frame header, and used encoding contents
 | |
|         :param compressor: optional compression function to be used on the body
 | |
|         """
 | |
|         flags = 0
 | |
|         body = io.BytesIO()
 | |
|         if msg.custom_payload:
 | |
|             if protocol_version < 4:
 | |
|                 raise UnsupportedOperation("Custom key/value payloads can only be used with protocol version 4 or higher")
 | |
|             flags |= CUSTOM_PAYLOAD_FLAG
 | |
|             write_bytesmap(body, msg.custom_payload)
 | |
|         msg.send_body(body, protocol_version)
 | |
|         body = body.getvalue()
 | |
| 
 | |
|         if compressor and len(body) > 0:
 | |
|             body = compressor(body)
 | |
|             flags |= COMPRESSED_FLAG
 | |
| 
 | |
|         if msg.tracing:
 | |
|             flags |= TRACING_FLAG
 | |
| 
 | |
|         if allow_beta_protocol_version:
 | |
|             flags |= USE_BETA_FLAG
 | |
| 
 | |
|         buff = io.BytesIO()
 | |
|         cls._write_header(buff, protocol_version, flags, stream_id, msg.opcode, len(body))
 | |
|         buff.write(body)
 | |
| 
 | |
|         return buff.getvalue()
 | |
| 
 | |
|     @staticmethod
 | |
|     def _write_header(f, version, flags, stream_id, opcode, length):
 | |
|         """
 | |
|         Write a CQL protocol frame header.
 | |
|         """
 | |
|         pack = v3_header_pack if version >= 3 else header_pack
 | |
|         f.write(pack(version, flags, stream_id, opcode))
 | |
|         write_int(f, length)
 | |
| 
 | |
|     @classmethod
 | |
|     def decode_message(cls, protocol_version, user_type_map, stream_id, flags, opcode, body,
 | |
|                        decompressor, result_metadata):
 | |
|         """
 | |
|         Decodes a native protocol message body
 | |
| 
 | |
|         :param protocol_version: version to use decoding contents
 | |
|         :param user_type_map: map[keyspace name] = map[type name] = custom type to instantiate when deserializing this type
 | |
|         :param stream_id: native protocol stream id from the frame header
 | |
|         :param flags: native protocol flags bitmap from the header
 | |
|         :param opcode: native protocol opcode from the header
 | |
|         :param body: frame body
 | |
|         :param decompressor: optional decompression function to inflate the body
 | |
|         :return: a message decoded from the body and frame attributes
 | |
|         """
 | |
|         if flags & COMPRESSED_FLAG:
 | |
|             if decompressor is None:
 | |
|                 raise RuntimeError("No de-compressor available for compressed frame!")
 | |
|             body = decompressor(body)
 | |
|             flags ^= COMPRESSED_FLAG
 | |
| 
 | |
|         body = io.BytesIO(body)
 | |
|         if flags & TRACING_FLAG:
 | |
|             trace_id = UUID(bytes=body.read(16))
 | |
|             flags ^= TRACING_FLAG
 | |
|         else:
 | |
|             trace_id = None
 | |
| 
 | |
|         if flags & WARNING_FLAG:
 | |
|             warnings = read_stringlist(body)
 | |
|             flags ^= WARNING_FLAG
 | |
|         else:
 | |
|             warnings = None
 | |
| 
 | |
|         if flags & CUSTOM_PAYLOAD_FLAG:
 | |
|             custom_payload = read_bytesmap(body)
 | |
|             flags ^= CUSTOM_PAYLOAD_FLAG
 | |
|         else:
 | |
|             custom_payload = None
 | |
| 
 | |
|         flags &= USE_BETA_MASK # will only be set if we asserted it in connection estabishment
 | |
| 
 | |
|         if flags:
 | |
|             log.warning("Unknown protocol flags set: %02x. May cause problems.", flags)
 | |
| 
 | |
|         msg_class = cls.message_types_by_opcode[opcode]
 | |
|         msg = msg_class.recv_body(body, protocol_version, user_type_map, result_metadata)
 | |
|         msg.stream_id = stream_id
 | |
|         msg.trace_id = trace_id
 | |
|         msg.custom_payload = custom_payload
 | |
|         msg.warnings = warnings
 | |
| 
 | |
|         if msg.warnings:
 | |
|             for w in msg.warnings:
 | |
|                 log.warning("Server warning: %s", w)
 | |
| 
 | |
|         return msg
 | |
| 
 | |
| def cython_protocol_handler(colparser):
 | |
|     """
 | |
|     Given a column parser to deserialize ResultMessages, return a suitable
 | |
|     Cython-based protocol handler.
 | |
| 
 | |
|     There are three Cython-based protocol handlers:
 | |
| 
 | |
|         - obj_parser.ListParser
 | |
|             decodes result messages into a list of tuples
 | |
| 
 | |
|         - obj_parser.LazyParser
 | |
|             decodes result messages lazily by returning an iterator
 | |
| 
 | |
|         - numpy_parser.NumPyParser
 | |
|             decodes result messages into NumPy arrays
 | |
| 
 | |
|     The default is to use obj_parser.ListParser
 | |
|     """
 | |
|     from cassandra.row_parser import make_recv_results_rows
 | |
| 
 | |
|     class FastResultMessage(ResultMessage):
 | |
|         """
 | |
|         Cython version of Result Message that has a faster implementation of
 | |
|         recv_results_row.
 | |
|         """
 | |
|         # type_codes = ResultMessage.type_codes.copy()
 | |
|         code_to_type = dict((v, k) for k, v in ResultMessage.type_codes.items())
 | |
|         recv_results_rows = classmethod(make_recv_results_rows(colparser))
 | |
| 
 | |
|     class CythonProtocolHandler(_ProtocolHandler):
 | |
|         """
 | |
|         Use FastResultMessage to decode query result message messages.
 | |
|         """
 | |
| 
 | |
|         my_opcodes = _ProtocolHandler.message_types_by_opcode.copy()
 | |
|         my_opcodes[FastResultMessage.opcode] = FastResultMessage
 | |
|         message_types_by_opcode = my_opcodes
 | |
| 
 | |
|         col_parser = colparser
 | |
| 
 | |
|     return CythonProtocolHandler
 | |
| 
 | |
| 
 | |
| if HAVE_CYTHON:
 | |
|     from cassandra.obj_parser import ListParser, LazyParser
 | |
|     ProtocolHandler = cython_protocol_handler(ListParser())
 | |
|     LazyProtocolHandler = cython_protocol_handler(LazyParser())
 | |
| else:
 | |
|     # Use Python-based ProtocolHandler
 | |
|     ProtocolHandler = _ProtocolHandler
 | |
|     LazyProtocolHandler = None
 | |
| 
 | |
| 
 | |
| if HAVE_CYTHON and HAVE_NUMPY:
 | |
|     from cassandra.numpy_parser import NumpyParser
 | |
|     NumpyProtocolHandler = cython_protocol_handler(NumpyParser())
 | |
| else:
 | |
|     NumpyProtocolHandler = None
 | |
| 
 | |
| 
 | |
| def read_byte(f):
 | |
|     return int8_unpack(f.read(1))
 | |
| 
 | |
| 
 | |
| def write_byte(f, b):
 | |
|     f.write(int8_pack(b))
 | |
| 
 | |
| 
 | |
| def read_int(f):
 | |
|     return int32_unpack(f.read(4))
 | |
| 
 | |
| 
 | |
| def write_int(f, i):
 | |
|     f.write(int32_pack(i))
 | |
| 
 | |
| 
 | |
| def write_uint(f, i):
 | |
|     f.write(uint32_pack(i))
 | |
| 
 | |
| 
 | |
| def write_long(f, i):
 | |
|     f.write(uint64_pack(i))
 | |
| 
 | |
| 
 | |
| def read_short(f):
 | |
|     return uint16_unpack(f.read(2))
 | |
| 
 | |
| 
 | |
| def write_short(f, s):
 | |
|     f.write(uint16_pack(s))
 | |
| 
 | |
| 
 | |
| def read_consistency_level(f):
 | |
|     return read_short(f)
 | |
| 
 | |
| 
 | |
| def write_consistency_level(f, cl):
 | |
|     write_short(f, cl)
 | |
| 
 | |
| 
 | |
| def read_string(f):
 | |
|     size = read_short(f)
 | |
|     contents = f.read(size)
 | |
|     return contents.decode('utf8')
 | |
| 
 | |
| 
 | |
| def read_binary_string(f):
 | |
|     size = read_short(f)
 | |
|     contents = f.read(size)
 | |
|     return contents
 | |
| 
 | |
| 
 | |
| def write_string(f, s):
 | |
|     if isinstance(s, six.text_type):
 | |
|         s = s.encode('utf8')
 | |
|     write_short(f, len(s))
 | |
|     f.write(s)
 | |
| 
 | |
| 
 | |
| def read_binary_longstring(f):
 | |
|     size = read_int(f)
 | |
|     contents = f.read(size)
 | |
|     return contents
 | |
| 
 | |
| 
 | |
| def read_longstring(f):
 | |
|     return read_binary_longstring(f).decode('utf8')
 | |
| 
 | |
| 
 | |
| def write_longstring(f, s):
 | |
|     if isinstance(s, six.text_type):
 | |
|         s = s.encode('utf8')
 | |
|     write_int(f, len(s))
 | |
|     f.write(s)
 | |
| 
 | |
| 
 | |
| def read_stringlist(f):
 | |
|     numstrs = read_short(f)
 | |
|     return [read_string(f) for _ in range(numstrs)]
 | |
| 
 | |
| 
 | |
| def write_stringlist(f, stringlist):
 | |
|     write_short(f, len(stringlist))
 | |
|     for s in stringlist:
 | |
|         write_string(f, s)
 | |
| 
 | |
| 
 | |
| def read_stringmap(f):
 | |
|     numpairs = read_short(f)
 | |
|     strmap = {}
 | |
|     for _ in range(numpairs):
 | |
|         k = read_string(f)
 | |
|         strmap[k] = read_string(f)
 | |
|     return strmap
 | |
| 
 | |
| 
 | |
| def write_stringmap(f, strmap):
 | |
|     write_short(f, len(strmap))
 | |
|     for k, v in strmap.items():
 | |
|         write_string(f, k)
 | |
|         write_string(f, v)
 | |
| 
 | |
| 
 | |
| def read_bytesmap(f):
 | |
|     numpairs = read_short(f)
 | |
|     bytesmap = {}
 | |
|     for _ in range(numpairs):
 | |
|         k = read_string(f)
 | |
|         bytesmap[k] = read_value(f)
 | |
|     return bytesmap
 | |
| 
 | |
| 
 | |
| def write_bytesmap(f, bytesmap):
 | |
|     write_short(f, len(bytesmap))
 | |
|     for k, v in bytesmap.items():
 | |
|         write_string(f, k)
 | |
|         write_value(f, v)
 | |
| 
 | |
| 
 | |
| def read_stringmultimap(f):
 | |
|     numkeys = read_short(f)
 | |
|     strmmap = {}
 | |
|     for _ in range(numkeys):
 | |
|         k = read_string(f)
 | |
|         strmmap[k] = read_stringlist(f)
 | |
|     return strmmap
 | |
| 
 | |
| 
 | |
| def write_stringmultimap(f, strmmap):
 | |
|     write_short(f, len(strmmap))
 | |
|     for k, v in strmmap.items():
 | |
|         write_string(f, k)
 | |
|         write_stringlist(f, v)
 | |
| 
 | |
| 
 | |
| def read_error_code_map(f):
 | |
|     numpairs = read_int(f)
 | |
|     error_code_map = {}
 | |
|     for _ in range(numpairs):
 | |
|         endpoint = read_inet_addr_only(f)
 | |
|         error_code_map[endpoint] = read_short(f)
 | |
|     return error_code_map
 | |
| 
 | |
| 
 | |
| def read_value(f):
 | |
|     size = read_int(f)
 | |
|     if size < 0:
 | |
|         return None
 | |
|     return f.read(size)
 | |
| 
 | |
| 
 | |
| def write_value(f, v):
 | |
|     if v is None:
 | |
|         write_int(f, -1)
 | |
|     elif v is _UNSET_VALUE:
 | |
|         write_int(f, -2)
 | |
|     else:
 | |
|         write_int(f, len(v))
 | |
|         f.write(v)
 | |
| 
 | |
| 
 | |
| def read_inet_addr_only(f):
 | |
|     size = read_byte(f)
 | |
|     addrbytes = f.read(size)
 | |
|     if size == 4:
 | |
|         addrfam = socket.AF_INET
 | |
|     elif size == 16:
 | |
|         addrfam = socket.AF_INET6
 | |
|     else:
 | |
|         raise InternalError("bad inet address: %r" % (addrbytes,))
 | |
|     return util.inet_ntop(addrfam, addrbytes)
 | |
| 
 | |
| 
 | |
| def read_inet(f):
 | |
|     addr = read_inet_addr_only(f)
 | |
|     port = read_int(f)
 | |
|     return (addr, port)
 | |
| 
 | |
| 
 | |
| def write_inet(f, addrtuple):
 | |
|     addr, port = addrtuple
 | |
|     if ':' in addr:
 | |
|         addrfam = socket.AF_INET6
 | |
|     else:
 | |
|         addrfam = socket.AF_INET
 | |
|     addrbytes = util.inet_pton(addrfam, addr)
 | |
|     write_byte(f, len(addrbytes))
 | |
|     f.write(addrbytes)
 | |
|     write_int(f, port)
 |