877 lines
		
	
	
		
			33 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			877 lines
		
	
	
		
			33 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| # Copyright 2013-2014 DataStax, Inc.
 | |
| #
 | |
| # Licensed under the Apache License, Version 2.0 (the "License");
 | |
| # you may not use this file except in compliance with the License.
 | |
| # You may obtain a copy of the License at
 | |
| #
 | |
| # http://www.apache.org/licenses/LICENSE-2.0
 | |
| #
 | |
| # Unless required by applicable law or agreed to in writing, software
 | |
| # distributed under the License is distributed on an "AS IS" BASIS,
 | |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| # See the License for the specific language governing permissions and
 | |
| # limitations under the License.
 | |
| 
 | |
| try:
 | |
|     import unittest2 as unittest
 | |
| except ImportError:
 | |
|     import unittest # noqa
 | |
| 
 | |
| from itertools import islice, cycle
 | |
| from mock import Mock
 | |
| from random import randint
 | |
| import six
 | |
| import sys
 | |
| import struct
 | |
| from threading import Thread
 | |
| 
 | |
| from cassandra import ConsistencyLevel
 | |
| from cassandra.cluster import Cluster
 | |
| from cassandra.metadata import Metadata
 | |
| from cassandra.policies import (RoundRobinPolicy, DCAwareRoundRobinPolicy,
 | |
|                                 TokenAwarePolicy, SimpleConvictionPolicy,
 | |
|                                 HostDistance, ExponentialReconnectionPolicy,
 | |
|                                 RetryPolicy, WriteType,
 | |
|                                 DowngradingConsistencyRetryPolicy, ConstantReconnectionPolicy,
 | |
|                                 LoadBalancingPolicy, ConvictionPolicy, ReconnectionPolicy, FallthroughRetryPolicy)
 | |
| from cassandra.pool import Host
 | |
| from cassandra.query import Statement
 | |
| 
 | |
| from six.moves import xrange
 | |
| 
 | |
| 
 | |
| class LoadBalancingPolicyTest(unittest.TestCase):
 | |
|     def test_non_implemented(self):
 | |
|         """
 | |
|         Code coverage for interface-style base class
 | |
|         """
 | |
| 
 | |
|         policy = LoadBalancingPolicy()
 | |
|         host = Host("ip1", SimpleConvictionPolicy)
 | |
|         host.set_location_info("dc1", "rack1")
 | |
| 
 | |
|         self.assertRaises(NotImplementedError, policy.distance, host)
 | |
|         self.assertRaises(NotImplementedError, policy.populate, None, host)
 | |
|         self.assertRaises(NotImplementedError, policy.make_query_plan)
 | |
|         self.assertRaises(NotImplementedError, policy.on_up, host)
 | |
|         self.assertRaises(NotImplementedError, policy.on_down, host)
 | |
|         self.assertRaises(NotImplementedError, policy.on_add, host)
 | |
|         self.assertRaises(NotImplementedError, policy.on_remove, host)
 | |
| 
 | |
|     def test_instance_check(self):
 | |
|         self.assertRaises(TypeError, Cluster, load_balancing_policy=RoundRobinPolicy)
 | |
| 
 | |
| 
 | |
| class RoundRobinPolicyTest(unittest.TestCase):
 | |
| 
 | |
|     def test_basic(self):
 | |
|         hosts = [0, 1, 2, 3]
 | |
|         policy = RoundRobinPolicy()
 | |
|         policy.populate(None, hosts)
 | |
|         qplan = list(policy.make_query_plan())
 | |
|         self.assertEqual(sorted(qplan), hosts)
 | |
| 
 | |
|     def test_multiple_query_plans(self):
 | |
|         hosts = [0, 1, 2, 3]
 | |
|         policy = RoundRobinPolicy()
 | |
|         policy.populate(None, hosts)
 | |
|         for i in xrange(20):
 | |
|             qplan = list(policy.make_query_plan())
 | |
|             self.assertEqual(sorted(qplan), hosts)
 | |
| 
 | |
|     def test_single_host(self):
 | |
|         policy = RoundRobinPolicy()
 | |
|         policy.populate(None, [0])
 | |
|         qplan = list(policy.make_query_plan())
 | |
|         self.assertEqual(qplan, [0])
 | |
| 
 | |
|     def test_status_updates(self):
 | |
|         hosts = [0, 1, 2, 3]
 | |
|         policy = RoundRobinPolicy()
 | |
|         policy.populate(None, hosts)
 | |
|         policy.on_down(0)
 | |
|         policy.on_remove(1)
 | |
|         policy.on_up(4)
 | |
|         policy.on_add(5)
 | |
|         qplan = list(policy.make_query_plan())
 | |
|         self.assertEqual(sorted(qplan), [2, 3, 4, 5])
 | |
| 
 | |
|     def test_thread_safety(self):
 | |
|         hosts = range(100)
 | |
|         policy = RoundRobinPolicy()
 | |
|         policy.populate(None, hosts)
 | |
| 
 | |
|         def check_query_plan():
 | |
|             for i in range(100):
 | |
|                 qplan = list(policy.make_query_plan())
 | |
|                 self.assertEqual(sorted(qplan), hosts)
 | |
| 
 | |
|         threads = [Thread(target=check_query_plan) for i in range(4)]
 | |
|         map(lambda t: t.start(), threads)
 | |
|         map(lambda t: t.join(), threads)
 | |
| 
 | |
|     def test_thread_safety_during_modification(self):
 | |
|         hosts = range(100)
 | |
|         policy = RoundRobinPolicy()
 | |
|         policy.populate(None, hosts)
 | |
| 
 | |
|         errors = []
 | |
| 
 | |
|         def check_query_plan():
 | |
|             try:
 | |
|                 for i in xrange(100):
 | |
|                     list(policy.make_query_plan())
 | |
|             except Exception as exc:
 | |
|                 errors.append(exc)
 | |
| 
 | |
|         def host_up():
 | |
|             for i in xrange(1000):
 | |
|                 policy.on_up(randint(0, 99))
 | |
| 
 | |
|         def host_down():
 | |
|             for i in xrange(1000):
 | |
|                 policy.on_down(randint(0, 99))
 | |
| 
 | |
|         threads = []
 | |
|         for i in range(5):
 | |
|             threads.append(Thread(target=check_query_plan))
 | |
|             threads.append(Thread(target=host_up))
 | |
|             threads.append(Thread(target=host_down))
 | |
| 
 | |
|         # make the GIL switch after every instruction, maximizing
 | |
|         # the chace of race conditions
 | |
|         if six.PY2:
 | |
|             original_interval = sys.getcheckinterval()
 | |
|         else:
 | |
|             original_interval = sys.getswitchinterval()
 | |
| 
 | |
|         try:
 | |
|             if six.PY2:
 | |
|                 sys.setcheckinterval(0)
 | |
|             else:
 | |
|                 sys.setswitchinterval(0.0001)
 | |
|             map(lambda t: t.start(), threads)
 | |
|             map(lambda t: t.join(), threads)
 | |
|         finally:
 | |
|             if six.PY2:
 | |
|                 sys.setcheckinterval(original_interval)
 | |
|             else:
 | |
|                 sys.setswitchinterval(original_interval)
 | |
| 
 | |
|         if errors:
 | |
|             self.fail("Saw errors: %s" % (errors,))
 | |
| 
 | |
|     def test_no_live_nodes(self):
 | |
|         """
 | |
|         Ensure query plan for a downed cluster will execute without errors
 | |
|         """
 | |
|         hosts = [0, 1, 2, 3]
 | |
|         policy = RoundRobinPolicy()
 | |
|         policy.populate(None, hosts)
 | |
| 
 | |
|         for i in range(4):
 | |
|             policy.on_down(i)
 | |
| 
 | |
|         qplan = list(policy.make_query_plan())
 | |
|         self.assertEqual(qplan, [])
 | |
| 
 | |
| 
 | |
| class DCAwareRoundRobinPolicyTest(unittest.TestCase):
 | |
| 
 | |
|     def test_no_remote(self):
 | |
|         hosts = []
 | |
|         for i in range(4):
 | |
|             h = Host(i, SimpleConvictionPolicy)
 | |
|             h.set_location_info("dc1", "rack1")
 | |
|             hosts.append(h)
 | |
| 
 | |
|         policy = DCAwareRoundRobinPolicy("dc1")
 | |
|         policy.populate(None, hosts)
 | |
|         qplan = list(policy.make_query_plan())
 | |
|         self.assertEqual(sorted(qplan), sorted(hosts))
 | |
| 
 | |
|     def test_with_remotes(self):
 | |
|         hosts = [Host(i, SimpleConvictionPolicy) for i in range(4)]
 | |
|         for h in hosts[:2]:
 | |
|             h.set_location_info("dc1", "rack1")
 | |
|         for h in hosts[2:]:
 | |
|             h.set_location_info("dc2", "rack1")
 | |
| 
 | |
|         local_hosts = set(h for h in hosts if h.datacenter == "dc1")
 | |
|         remote_hosts = set(h for h in hosts if h.datacenter != "dc1")
 | |
| 
 | |
|         # allow all of the remote hosts to be used
 | |
|         policy = DCAwareRoundRobinPolicy("dc1", used_hosts_per_remote_dc=2)
 | |
|         policy.populate(Mock(), hosts)
 | |
|         qplan = list(policy.make_query_plan())
 | |
|         self.assertEqual(set(qplan[:2]), local_hosts)
 | |
|         self.assertEqual(set(qplan[2:]), remote_hosts)
 | |
| 
 | |
|         # allow only one of the remote hosts to be used
 | |
|         policy = DCAwareRoundRobinPolicy("dc1", used_hosts_per_remote_dc=1)
 | |
|         policy.populate(Mock(), hosts)
 | |
|         qplan = list(policy.make_query_plan())
 | |
|         self.assertEqual(set(qplan[:2]), local_hosts)
 | |
| 
 | |
|         used_remotes = set(qplan[2:])
 | |
|         self.assertEqual(1, len(used_remotes))
 | |
|         self.assertIn(qplan[2], remote_hosts)
 | |
| 
 | |
|         # allow no remote hosts to be used
 | |
|         policy = DCAwareRoundRobinPolicy("dc1", used_hosts_per_remote_dc=0)
 | |
|         policy.populate(Mock(), hosts)
 | |
|         qplan = list(policy.make_query_plan())
 | |
|         self.assertEqual(2, len(qplan))
 | |
|         self.assertEqual(local_hosts, set(qplan))
 | |
| 
 | |
|     def test_get_distance(self):
 | |
|         policy = DCAwareRoundRobinPolicy("dc1", used_hosts_per_remote_dc=0)
 | |
|         host = Host("ip1", SimpleConvictionPolicy)
 | |
|         host.set_location_info("dc1", "rack1")
 | |
|         policy.populate(Mock(), [host])
 | |
| 
 | |
|         self.assertEqual(policy.distance(host), HostDistance.LOCAL)
 | |
| 
 | |
|         # used_hosts_per_remote_dc is set to 0, so ignore it
 | |
|         remote_host = Host("ip2", SimpleConvictionPolicy)
 | |
|         remote_host.set_location_info("dc2", "rack1")
 | |
|         self.assertEqual(policy.distance(remote_host), HostDistance.IGNORED)
 | |
| 
 | |
|         # dc2 isn't registered in the policy's live_hosts dict
 | |
|         policy.used_hosts_per_remote_dc = 1
 | |
|         self.assertEqual(policy.distance(remote_host), HostDistance.IGNORED)
 | |
| 
 | |
|         # make sure the policy has both dcs registered
 | |
|         policy.populate(Mock(), [host, remote_host])
 | |
|         self.assertEqual(policy.distance(remote_host), HostDistance.REMOTE)
 | |
| 
 | |
|         # since used_hosts_per_remote_dc is set to 1, only the first
 | |
|         # remote host in dc2 will be REMOTE, the rest are IGNORED
 | |
|         second_remote_host = Host("ip3", SimpleConvictionPolicy)
 | |
|         second_remote_host.set_location_info("dc2", "rack1")
 | |
|         policy.populate(Mock(), [host, remote_host, second_remote_host])
 | |
|         distances = set([policy.distance(remote_host), policy.distance(second_remote_host)])
 | |
|         self.assertEqual(distances, set([HostDistance.REMOTE, HostDistance.IGNORED]))
 | |
| 
 | |
|     def test_status_updates(self):
 | |
|         hosts = [Host(i, SimpleConvictionPolicy) for i in range(4)]
 | |
|         for h in hosts[:2]:
 | |
|             h.set_location_info("dc1", "rack1")
 | |
|         for h in hosts[2:]:
 | |
|             h.set_location_info("dc2", "rack1")
 | |
| 
 | |
|         policy = DCAwareRoundRobinPolicy("dc1", used_hosts_per_remote_dc=1)
 | |
|         policy.populate(Mock(), hosts)
 | |
|         policy.on_down(hosts[0])
 | |
|         policy.on_remove(hosts[2])
 | |
| 
 | |
|         new_local_host = Host(4, SimpleConvictionPolicy)
 | |
|         new_local_host.set_location_info("dc1", "rack1")
 | |
|         policy.on_up(new_local_host)
 | |
| 
 | |
|         new_remote_host = Host(5, SimpleConvictionPolicy)
 | |
|         new_remote_host.set_location_info("dc9000", "rack1")
 | |
|         policy.on_add(new_remote_host)
 | |
| 
 | |
|         # we now have two local hosts and two remote hosts in separate dcs
 | |
|         qplan = list(policy.make_query_plan())
 | |
|         self.assertEqual(set(qplan[:2]), set([hosts[1], new_local_host]))
 | |
|         self.assertEqual(set(qplan[2:]), set([hosts[3], new_remote_host]))
 | |
| 
 | |
|         # since we have hosts in dc9000, the distance shouldn't be IGNORED
 | |
|         self.assertEqual(policy.distance(new_remote_host), HostDistance.REMOTE)
 | |
| 
 | |
|         policy.on_down(new_local_host)
 | |
|         policy.on_down(hosts[1])
 | |
|         qplan = list(policy.make_query_plan())
 | |
|         self.assertEqual(set(qplan), set([hosts[3], new_remote_host]))
 | |
| 
 | |
|         policy.on_down(new_remote_host)
 | |
|         policy.on_down(hosts[3])
 | |
|         qplan = list(policy.make_query_plan())
 | |
|         self.assertEqual(qplan, [])
 | |
| 
 | |
|     def test_no_live_nodes(self):
 | |
|         """
 | |
|         Ensure query plan for a downed cluster will execute without errors
 | |
|         """
 | |
| 
 | |
|         hosts = []
 | |
|         for i in range(4):
 | |
|             h = Host(i, SimpleConvictionPolicy)
 | |
|             h.set_location_info("dc1", "rack1")
 | |
|             hosts.append(h)
 | |
| 
 | |
|         policy = DCAwareRoundRobinPolicy("dc1", used_hosts_per_remote_dc=1)
 | |
|         policy.populate(Mock(), hosts)
 | |
| 
 | |
|         for host in hosts:
 | |
|             policy.on_down(host)
 | |
| 
 | |
|         qplan = list(policy.make_query_plan())
 | |
|         self.assertEqual(qplan, [])
 | |
| 
 | |
|     def test_no_nodes(self):
 | |
|         """
 | |
|         Ensure query plan for an empty cluster will execute without errors
 | |
|         """
 | |
| 
 | |
|         policy = DCAwareRoundRobinPolicy("dc1", used_hosts_per_remote_dc=1)
 | |
|         policy.populate(None, [])
 | |
| 
 | |
|         qplan = list(policy.make_query_plan())
 | |
|         self.assertEqual(qplan, [])
 | |
| 
 | |
|     def test_default_dc(self):
 | |
|         host_local = Host(1, SimpleConvictionPolicy, 'local')
 | |
|         host_remote = Host(2, SimpleConvictionPolicy, 'remote')
 | |
|         host_none = Host(1, SimpleConvictionPolicy)
 | |
| 
 | |
|         # contact point is '1'
 | |
|         cluster = Mock(contact_points=[1])
 | |
| 
 | |
|         # contact DC first
 | |
|         policy = DCAwareRoundRobinPolicy()
 | |
|         policy.populate(cluster, [host_none])
 | |
|         self.assertFalse(policy.local_dc)
 | |
|         policy.on_add(host_local)
 | |
|         policy.on_add(host_remote)
 | |
|         self.assertNotEqual(policy.local_dc, host_remote.datacenter)
 | |
|         self.assertEqual(policy.local_dc, host_local.datacenter)
 | |
| 
 | |
|         # contact DC second
 | |
|         policy = DCAwareRoundRobinPolicy()
 | |
|         policy.populate(cluster, [host_none])
 | |
|         self.assertFalse(policy.local_dc)
 | |
|         policy.on_add(host_remote)
 | |
|         policy.on_add(host_local)
 | |
|         self.assertNotEqual(policy.local_dc, host_remote.datacenter)
 | |
|         self.assertEqual(policy.local_dc, host_local.datacenter)
 | |
| 
 | |
|         # no DC
 | |
|         policy = DCAwareRoundRobinPolicy()
 | |
|         policy.populate(cluster, [host_none])
 | |
|         self.assertFalse(policy.local_dc)
 | |
|         policy.on_add(host_none)
 | |
|         self.assertFalse(policy.local_dc)
 | |
| 
 | |
|         # only other DC
 | |
|         policy = DCAwareRoundRobinPolicy()
 | |
|         policy.populate(cluster, [host_none])
 | |
|         self.assertFalse(policy.local_dc)
 | |
|         policy.on_add(host_remote)
 | |
|         self.assertFalse(policy.local_dc)
 | |
| 
 | |
| class TokenAwarePolicyTest(unittest.TestCase):
 | |
| 
 | |
|     def test_wrap_round_robin(self):
 | |
|         cluster = Mock(spec=Cluster)
 | |
|         cluster.metadata = Mock(spec=Metadata)
 | |
|         hosts = [Host(str(i), SimpleConvictionPolicy) for i in range(4)]
 | |
|         for host in hosts:
 | |
|             host.set_up()
 | |
| 
 | |
|         def get_replicas(keyspace, packed_key):
 | |
|             index = struct.unpack('>i', packed_key)[0]
 | |
|             return list(islice(cycle(hosts), index, index + 2))
 | |
| 
 | |
|         cluster.metadata.get_replicas.side_effect = get_replicas
 | |
| 
 | |
|         policy = TokenAwarePolicy(RoundRobinPolicy())
 | |
|         policy.populate(cluster, hosts)
 | |
| 
 | |
|         for i in range(4):
 | |
|             query = Statement(routing_key=struct.pack('>i', i))
 | |
|             qplan = list(policy.make_query_plan(None, query))
 | |
| 
 | |
|             replicas = get_replicas(None, struct.pack('>i', i))
 | |
|             other = set(h for h in hosts if h not in replicas)
 | |
|             self.assertEqual(replicas, qplan[:2])
 | |
|             self.assertEqual(other, set(qplan[2:]))
 | |
| 
 | |
|         # Should use the secondary policy
 | |
|         for i in range(4):
 | |
|             qplan = list(policy.make_query_plan())
 | |
| 
 | |
|             self.assertEqual(set(qplan), set(hosts))
 | |
| 
 | |
|     def test_wrap_dc_aware(self):
 | |
|         cluster = Mock(spec=Cluster)
 | |
|         cluster.metadata = Mock(spec=Metadata)
 | |
|         hosts = [Host(str(i), SimpleConvictionPolicy) for i in range(4)]
 | |
|         for host in hosts:
 | |
|             host.set_up()
 | |
|         for h in hosts[:2]:
 | |
|             h.set_location_info("dc1", "rack1")
 | |
|         for h in hosts[2:]:
 | |
|             h.set_location_info("dc2", "rack1")
 | |
| 
 | |
|         def get_replicas(keyspace, packed_key):
 | |
|             index = struct.unpack('>i', packed_key)[0]
 | |
|             # return one node from each DC
 | |
|             if index % 2 == 0:
 | |
|                 return [hosts[0], hosts[2]]
 | |
|             else:
 | |
|                 return [hosts[1], hosts[3]]
 | |
| 
 | |
|         cluster.metadata.get_replicas.side_effect = get_replicas
 | |
| 
 | |
|         policy = TokenAwarePolicy(DCAwareRoundRobinPolicy("dc1", used_hosts_per_remote_dc=1))
 | |
|         policy.populate(cluster, hosts)
 | |
| 
 | |
|         for i in range(4):
 | |
|             query = Statement(routing_key=struct.pack('>i', i))
 | |
|             qplan = list(policy.make_query_plan(None, query))
 | |
|             replicas = get_replicas(None, struct.pack('>i', i))
 | |
| 
 | |
|             # first should be the only local replica
 | |
|             self.assertIn(qplan[0], replicas)
 | |
|             self.assertEqual(qplan[0].datacenter, "dc1")
 | |
| 
 | |
|             # then the local non-replica
 | |
|             self.assertNotIn(qplan[1], replicas)
 | |
|             self.assertEqual(qplan[1].datacenter, "dc1")
 | |
| 
 | |
|             # then one of the remotes (used_hosts_per_remote_dc is 1, so we
 | |
|             # shouldn't see two remotes)
 | |
|             self.assertEqual(qplan[2].datacenter, "dc2")
 | |
|             self.assertEqual(3, len(qplan))
 | |
| 
 | |
|     class FakeCluster:
 | |
|         def __init__(self):
 | |
|             self.metadata = Mock(spec=Metadata)
 | |
| 
 | |
|     def test_get_distance(self):
 | |
|         """
 | |
|         Same test as DCAwareRoundRobinPolicyTest.test_get_distance()
 | |
|         Except a FakeCluster is needed for the metadata variable and
 | |
|         policy.child_policy is needed to change child policy settings
 | |
|         """
 | |
| 
 | |
|         policy = TokenAwarePolicy(DCAwareRoundRobinPolicy("dc1", used_hosts_per_remote_dc=0))
 | |
|         host = Host("ip1", SimpleConvictionPolicy)
 | |
|         host.set_location_info("dc1", "rack1")
 | |
| 
 | |
|         policy.populate(self.FakeCluster(), [host])
 | |
| 
 | |
|         self.assertEqual(policy.distance(host), HostDistance.LOCAL)
 | |
| 
 | |
|         # used_hosts_per_remote_dc is set to 0, so ignore it
 | |
|         remote_host = Host("ip2", SimpleConvictionPolicy)
 | |
|         remote_host.set_location_info("dc2", "rack1")
 | |
|         self.assertEqual(policy.distance(remote_host), HostDistance.IGNORED)
 | |
| 
 | |
|         # dc2 isn't registered in the policy's live_hosts dict
 | |
|         policy._child_policy.used_hosts_per_remote_dc = 1
 | |
|         self.assertEqual(policy.distance(remote_host), HostDistance.IGNORED)
 | |
| 
 | |
|         # make sure the policy has both dcs registered
 | |
|         policy.populate(self.FakeCluster(), [host, remote_host])
 | |
|         self.assertEqual(policy.distance(remote_host), HostDistance.REMOTE)
 | |
| 
 | |
|         # since used_hosts_per_remote_dc is set to 1, only the first
 | |
|         # remote host in dc2 will be REMOTE, the rest are IGNORED
 | |
|         second_remote_host = Host("ip3", SimpleConvictionPolicy)
 | |
|         second_remote_host.set_location_info("dc2", "rack1")
 | |
|         policy.populate(self.FakeCluster(), [host, remote_host, second_remote_host])
 | |
|         distances = set([policy.distance(remote_host), policy.distance(second_remote_host)])
 | |
|         self.assertEqual(distances, set([HostDistance.REMOTE, HostDistance.IGNORED]))
 | |
| 
 | |
|     def test_status_updates(self):
 | |
|         """
 | |
|         Same test as DCAwareRoundRobinPolicyTest.test_status_updates()
 | |
|         """
 | |
| 
 | |
|         hosts = [Host(i, SimpleConvictionPolicy) for i in range(4)]
 | |
|         for h in hosts[:2]:
 | |
|             h.set_location_info("dc1", "rack1")
 | |
|         for h in hosts[2:]:
 | |
|             h.set_location_info("dc2", "rack1")
 | |
| 
 | |
|         policy = TokenAwarePolicy(DCAwareRoundRobinPolicy("dc1", used_hosts_per_remote_dc=1))
 | |
|         policy.populate(self.FakeCluster(), hosts)
 | |
|         policy.on_down(hosts[0])
 | |
|         policy.on_remove(hosts[2])
 | |
| 
 | |
|         new_local_host = Host(4, SimpleConvictionPolicy)
 | |
|         new_local_host.set_location_info("dc1", "rack1")
 | |
|         policy.on_up(new_local_host)
 | |
| 
 | |
|         new_remote_host = Host(5, SimpleConvictionPolicy)
 | |
|         new_remote_host.set_location_info("dc9000", "rack1")
 | |
|         policy.on_add(new_remote_host)
 | |
| 
 | |
|         # we now have two local hosts and two remote hosts in separate dcs
 | |
|         qplan = list(policy.make_query_plan())
 | |
|         self.assertEqual(set(qplan[:2]), set([hosts[1], new_local_host]))
 | |
|         self.assertEqual(set(qplan[2:]), set([hosts[3], new_remote_host]))
 | |
| 
 | |
|         # since we have hosts in dc9000, the distance shouldn't be IGNORED
 | |
|         self.assertEqual(policy.distance(new_remote_host), HostDistance.REMOTE)
 | |
| 
 | |
|         policy.on_down(new_local_host)
 | |
|         policy.on_down(hosts[1])
 | |
|         qplan = list(policy.make_query_plan())
 | |
|         self.assertEqual(set(qplan), set([hosts[3], new_remote_host]))
 | |
| 
 | |
|         policy.on_down(new_remote_host)
 | |
|         policy.on_down(hosts[3])
 | |
|         qplan = list(policy.make_query_plan())
 | |
|         self.assertEqual(qplan, [])
 | |
| 
 | |
| 
 | |
| class ConvictionPolicyTest(unittest.TestCase):
 | |
|     def test_not_implemented(self):
 | |
|         """
 | |
|         Code coverage for interface-style base class
 | |
|         """
 | |
| 
 | |
|         conviction_policy = ConvictionPolicy(1)
 | |
|         self.assertRaises(NotImplementedError, conviction_policy.add_failure, 1)
 | |
|         self.assertRaises(NotImplementedError, conviction_policy.reset)
 | |
| 
 | |
| 
 | |
| class SimpleConvictionPolicyTest(unittest.TestCase):
 | |
|     def test_basic_responses(self):
 | |
|         """
 | |
|         Code coverage for SimpleConvictionPolicy
 | |
|         """
 | |
| 
 | |
|         conviction_policy = SimpleConvictionPolicy(1)
 | |
|         self.assertEqual(conviction_policy.add_failure(1), True)
 | |
|         self.assertEqual(conviction_policy.reset(), None)
 | |
| 
 | |
| 
 | |
| class ReconnectionPolicyTest(unittest.TestCase):
 | |
|     def test_basic_responses(self):
 | |
|         """
 | |
|         Code coverage for interface-style base class
 | |
|         """
 | |
| 
 | |
|         policy = ReconnectionPolicy()
 | |
|         self.assertRaises(NotImplementedError, policy.new_schedule)
 | |
| 
 | |
| 
 | |
| class ConstantReconnectionPolicyTest(unittest.TestCase):
 | |
| 
 | |
|     def test_bad_vals(self):
 | |
|         """
 | |
|         Test initialization values
 | |
|         """
 | |
| 
 | |
|         self.assertRaises(ValueError, ConstantReconnectionPolicy, -1, 0)
 | |
| 
 | |
|     def test_schedule(self):
 | |
|         """
 | |
|         Test ConstantReconnectionPolicy schedule
 | |
|         """
 | |
| 
 | |
|         delay = 2
 | |
|         max_attempts = 100
 | |
|         policy = ConstantReconnectionPolicy(delay=delay, max_attempts=max_attempts)
 | |
|         schedule = list(policy.new_schedule())
 | |
|         self.assertEqual(len(schedule), max_attempts)
 | |
|         for i, delay in enumerate(schedule):
 | |
|             self.assertEqual(delay, delay)
 | |
| 
 | |
|     def test_schedule_negative_max_attempts(self):
 | |
|         """
 | |
|         Test how negative max_attempts are handled
 | |
|         """
 | |
| 
 | |
|         delay = 2
 | |
|         max_attempts = -100
 | |
| 
 | |
|         try:
 | |
|             ConstantReconnectionPolicy(delay=delay, max_attempts=max_attempts)
 | |
|             self.fail('max_attempts should throw ValueError when negative')
 | |
|         except ValueError:
 | |
|             pass
 | |
| 
 | |
| 
 | |
| class ExponentialReconnectionPolicyTest(unittest.TestCase):
 | |
| 
 | |
|     def test_bad_vals(self):
 | |
|         self.assertRaises(ValueError, ExponentialReconnectionPolicy, -1, 0)
 | |
|         self.assertRaises(ValueError, ExponentialReconnectionPolicy, 0, -1)
 | |
|         self.assertRaises(ValueError, ExponentialReconnectionPolicy, 9000, 1)
 | |
| 
 | |
|     def test_schedule(self):
 | |
|         policy = ExponentialReconnectionPolicy(base_delay=2, max_delay=100)
 | |
|         schedule = list(policy.new_schedule())
 | |
|         self.assertEqual(len(schedule), 64)
 | |
|         for i, delay in enumerate(schedule):
 | |
|             if i == 0:
 | |
|                 self.assertEqual(delay, 2)
 | |
|             elif i < 6:
 | |
|                 self.assertEqual(delay, schedule[i - 1] * 2)
 | |
|             else:
 | |
|                 self.assertEqual(delay, 100)
 | |
| 
 | |
| ONE = ConsistencyLevel.ONE
 | |
| 
 | |
| class RetryPolicyTest(unittest.TestCase):
 | |
| 
 | |
|     def test_read_timeout(self):
 | |
|         policy = RetryPolicy()
 | |
| 
 | |
|         # if this is the second or greater attempt, rethrow
 | |
|         retry, consistency = policy.on_read_timeout(
 | |
|             query=None, consistency=ONE, required_responses=1, received_responses=2,
 | |
|             data_retrieved=True, retry_num=1)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         # if we didn't get enough responses, rethrow
 | |
|         retry, consistency = policy.on_read_timeout(
 | |
|             query=None, consistency=ONE, required_responses=2, received_responses=1,
 | |
|             data_retrieved=True, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         # if we got enough responses, but also got a data response, rethrow
 | |
|         retry, consistency = policy.on_read_timeout(
 | |
|             query=None, consistency=ONE, required_responses=2, received_responses=2,
 | |
|             data_retrieved=True, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         # we got enough responses but no data response, so retry
 | |
|         retry, consistency = policy.on_read_timeout(
 | |
|             query=None, consistency=ONE, required_responses=2, received_responses=2,
 | |
|             data_retrieved=False, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETRY)
 | |
|         self.assertEqual(consistency, ONE)
 | |
| 
 | |
|     def test_write_timeout(self):
 | |
|         policy = RetryPolicy()
 | |
| 
 | |
|         # if this is the second or greater attempt, rethrow
 | |
|         retry, consistency = policy.on_write_timeout(
 | |
|             query=None, consistency=ONE, write_type=WriteType.SIMPLE,
 | |
|             required_responses=1, received_responses=2, retry_num=1)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         # if it's not a BATCH_LOG write, don't retry it
 | |
|         retry, consistency = policy.on_write_timeout(
 | |
|             query=None, consistency=ONE, write_type=WriteType.SIMPLE,
 | |
|             required_responses=1, received_responses=2, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         # retry BATCH_LOG writes regardless of received responses
 | |
|         retry, consistency = policy.on_write_timeout(
 | |
|             query=None, consistency=ONE, write_type=WriteType.BATCH_LOG,
 | |
|             required_responses=10000, received_responses=1, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETRY)
 | |
|         self.assertEqual(consistency, ONE)
 | |
| 
 | |
|     def test_unavailable(self):
 | |
|         """
 | |
|         Use the same tests for test_write_timeout, but ensure they only RETHROW
 | |
|         """
 | |
|         policy = RetryPolicy()
 | |
| 
 | |
|         retry, consistency = policy.on_unavailable(
 | |
|             query=None, consistency=ONE,
 | |
|             required_replicas=1, alive_replicas=2, retry_num=1)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         retry, consistency = policy.on_unavailable(
 | |
|             query=None, consistency=ONE,
 | |
|             required_replicas=1, alive_replicas=2, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         retry, consistency = policy.on_unavailable(
 | |
|             query=None, consistency=ONE,
 | |
|             required_replicas=10000, alive_replicas=1, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
| 
 | |
| class FallthroughRetryPolicyTest(unittest.TestCase):
 | |
| 
 | |
|     """
 | |
|     Use the same tests for test_write_timeout, but ensure they only RETHROW
 | |
|     """
 | |
| 
 | |
|     def test_read_timeout(self):
 | |
|         policy = FallthroughRetryPolicy()
 | |
| 
 | |
|         retry, consistency = policy.on_read_timeout(
 | |
|             query=None, consistency=ONE, required_responses=1, received_responses=2,
 | |
|             data_retrieved=True, retry_num=1)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         retry, consistency = policy.on_read_timeout(
 | |
|             query=None, consistency=ONE, required_responses=2, received_responses=1,
 | |
|             data_retrieved=True, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         retry, consistency = policy.on_read_timeout(
 | |
|             query=None, consistency=ONE, required_responses=2, received_responses=2,
 | |
|             data_retrieved=True, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         retry, consistency = policy.on_read_timeout(
 | |
|             query=None, consistency=ONE, required_responses=2, received_responses=2,
 | |
|             data_retrieved=False, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|     def test_write_timeout(self):
 | |
|         policy = FallthroughRetryPolicy()
 | |
| 
 | |
|         retry, consistency = policy.on_write_timeout(
 | |
|             query=None, consistency=ONE, write_type=WriteType.SIMPLE,
 | |
|             required_responses=1, received_responses=2, retry_num=1)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         retry, consistency = policy.on_write_timeout(
 | |
|             query=None, consistency=ONE, write_type=WriteType.SIMPLE,
 | |
|             required_responses=1, received_responses=2, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         retry, consistency = policy.on_write_timeout(
 | |
|             query=None, consistency=ONE, write_type=WriteType.BATCH_LOG,
 | |
|             required_responses=10000, received_responses=1, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|     def test_unavailable(self):
 | |
|         policy = FallthroughRetryPolicy()
 | |
| 
 | |
|         retry, consistency = policy.on_unavailable(
 | |
|             query=None, consistency=ONE,
 | |
|             required_replicas=1, alive_replicas=2, retry_num=1)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         retry, consistency = policy.on_unavailable(
 | |
|             query=None, consistency=ONE,
 | |
|             required_replicas=1, alive_replicas=2, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         retry, consistency = policy.on_unavailable(
 | |
|             query=None, consistency=ONE,
 | |
|             required_replicas=10000, alive_replicas=1, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
| 
 | |
| class DowngradingConsistencyRetryPolicyTest(unittest.TestCase):
 | |
| 
 | |
|     def test_read_timeout(self):
 | |
|         policy = DowngradingConsistencyRetryPolicy()
 | |
| 
 | |
|         # if this is the second or greater attempt, rethrow
 | |
|         retry, consistency = policy.on_read_timeout(
 | |
|             query=None, consistency=ONE, required_responses=1, received_responses=2,
 | |
|             data_retrieved=True, retry_num=1)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         # if we didn't get enough responses, retry at a lower consistency
 | |
|         retry, consistency = policy.on_read_timeout(
 | |
|             query=None, consistency=ONE, required_responses=4, received_responses=3,
 | |
|             data_retrieved=True, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETRY)
 | |
|         self.assertEqual(consistency, ConsistencyLevel.THREE)
 | |
| 
 | |
|         # if we didn't get enough responses, retry at a lower consistency
 | |
|         retry, consistency = policy.on_read_timeout(
 | |
|             query=None, consistency=ONE, required_responses=3, received_responses=2,
 | |
|             data_retrieved=True, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETRY)
 | |
|         self.assertEqual(consistency, ConsistencyLevel.TWO)
 | |
| 
 | |
|         # retry consistency level goes down based on the # of recv'd responses
 | |
|         retry, consistency = policy.on_read_timeout(
 | |
|             query=None, consistency=ONE, required_responses=3, received_responses=1,
 | |
|             data_retrieved=True, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETRY)
 | |
|         self.assertEqual(consistency, ConsistencyLevel.ONE)
 | |
| 
 | |
|         # if we got no responses, rethrow
 | |
|         retry, consistency = policy.on_read_timeout(
 | |
|             query=None, consistency=ONE, required_responses=3, received_responses=0,
 | |
|             data_retrieved=True, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         # if we got enough response but no data, retry
 | |
|         retry, consistency = policy.on_read_timeout(
 | |
|             query=None, consistency=ONE, required_responses=3, received_responses=3,
 | |
|             data_retrieved=False, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETRY)
 | |
|         self.assertEqual(consistency, ONE)
 | |
| 
 | |
|         # if we got enough responses, but also got a data response, rethrow
 | |
|         retry, consistency = policy.on_read_timeout(
 | |
|             query=None, consistency=ONE, required_responses=2, received_responses=2,
 | |
|             data_retrieved=True, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|     def test_write_timeout(self):
 | |
|         policy = DowngradingConsistencyRetryPolicy()
 | |
| 
 | |
|         # if this is the second or greater attempt, rethrow
 | |
|         retry, consistency = policy.on_write_timeout(
 | |
|             query=None, consistency=ONE, write_type=WriteType.SIMPLE,
 | |
|             required_responses=1, received_responses=2, retry_num=1)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         # ignore failures on these types of writes
 | |
|         for write_type in (WriteType.SIMPLE, WriteType.BATCH, WriteType.COUNTER):
 | |
|             retry, consistency = policy.on_write_timeout(
 | |
|                 query=None, consistency=ONE, write_type=write_type,
 | |
|                 required_responses=1, received_responses=2, retry_num=0)
 | |
|             self.assertEqual(retry, RetryPolicy.IGNORE)
 | |
| 
 | |
|         # downgrade consistency level on unlogged batch writes
 | |
|         retry, consistency = policy.on_write_timeout(
 | |
|             query=None, consistency=ONE, write_type=WriteType.UNLOGGED_BATCH,
 | |
|             required_responses=3, received_responses=1, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETRY)
 | |
|         self.assertEqual(consistency, ConsistencyLevel.ONE)
 | |
| 
 | |
|         # retry batch log writes at the same consistency level
 | |
|         retry, consistency = policy.on_write_timeout(
 | |
|             query=None, consistency=ONE, write_type=WriteType.BATCH_LOG,
 | |
|             required_responses=3, received_responses=1, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETRY)
 | |
|         self.assertEqual(consistency, ONE)
 | |
| 
 | |
|         # timeout on an unknown write_type
 | |
|         retry, consistency = policy.on_write_timeout(
 | |
|             query=None, consistency=ONE, write_type=None,
 | |
|             required_responses=1, received_responses=2, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|     def test_unavailable(self):
 | |
|         policy = DowngradingConsistencyRetryPolicy()
 | |
| 
 | |
|         # if this is the second or greater attempt, rethrow
 | |
|         retry, consistency = policy.on_unavailable(
 | |
|             query=None, consistency=ONE, required_replicas=3, alive_replicas=1, retry_num=1)
 | |
|         self.assertEqual(retry, RetryPolicy.RETHROW)
 | |
|         self.assertEqual(consistency, None)
 | |
| 
 | |
|         # downgrade consistency on unavailable exceptions
 | |
|         retry, consistency = policy.on_unavailable(
 | |
|             query=None, consistency=ONE, required_replicas=3, alive_replicas=1, retry_num=0)
 | |
|         self.assertEqual(retry, RetryPolicy.RETRY)
 | |
|         self.assertEqual(consistency, ConsistencyLevel.ONE)
 | 
