Class | Name | Status | Type | Time(s) |
8_cythonno_x86_64_1_8 | test_network_topology_strategy | Success | | 229.567 |
8_cythonno_x86_64_1_8 | test_disk_balance_bootstrap | Success | | 104.656 |
8_cythonno_x86_64_1_8 | test_throttled_partition_update | Success | | 512.460 |
8_cythonno_x86_64_1_8 | test_resume_failed_replace | Success | | 169.834 |
8_cythonno_x86_64_1_8 | test_initial_empty_repair_tables | Skipped | hangs CI
/home/cassandra/cassandra-dtest/repair_tests/repair_test.py:1416: hangs CI | 0.001 |
8_cythonno_x86_64_2_8 | test_network_topology_strategy_each_quorum | Success | | 200.718 |
8_cythonno_x86_64_2_8 | test_add_dc_after_mv_simple_replication | Success | | 106.612 |
8_cythonno_x86_64_2_8 | test_really_complex_repair | Success | | 280.186 |
8_cythonno_x86_64_2_8 | test_restart_failed_replace_with_reset_resume_state | Success | | 206.689 |
8_cythonno_x86_64_2_8 | test_repair_parent_table | Success | | 93.043 |
8_cythonno_x86_64_3_8 | test_simple_strategy_users | Success | | 155.693 |
8_cythonno_x86_64_3_8 | test_add_dc_after_mv_network_replication | Success | | 110.675 |
8_cythonno_x86_64_3_8 | test_replace_stopped_node | Success | | 186.002 |
8_cythonno_x86_64_3_8 | test_restart_failed_replace | Success | | 262.469 |
8_cythonno_x86_64_3_8 | test_repair_table | Success | | 81.422 |
8_cythonno_x86_64_4_8 | test_simple_strategy_each_quorum_users | Success | | 88.215 |
8_cythonno_x86_64_4_8 | test_add_node_after_mv | Success | | 88.430 |
8_cythonno_x86_64_4_8 | test_replace_shutdown_node | Success | | 187.929 |
8_cythonno_x86_64_4_8 | test_rf_collapse_gossiping_property_file_snitch_multi_dc | Skipped | 5.1 > 5.0.x
/home/cassandra/cassandra-dtest/conftest.py:468: 5.1 > 5.0.x | 0.454 |
8_cythonno_x86_64_5_8 | test_network_topology_strategy_users | Success | | 260.268 |
8_cythonno_x86_64_5_8 | test_add_node_after_wide_mv_with_range_deletions | Success | | 117.894 |
8_cythonno_x86_64_5_8 | test_replace_stopped_node_same_address | Success | | 198.055 |
8_cythonno_x86_64_5_8 | test_rf_expand_gossiping_property_file_snitch_multi_dc | Skipped | 5.1 > 5.0.x
/home/cassandra/cassandra-dtest/conftest.py:468: 5.1 > 5.0.x | 0.275 |
8_cythonno_x86_64_6_8 | test_network_topology_strategy_each_quorum_users | Success | | 95.032 |
8_cythonno_x86_64_6_8 | test_add_node_after_very_wide_mv | Success | | 276.067 |
8_cythonno_x86_64_6_8 | test_replace_first_boot | Success | | 192.407 |
8_cythonno_x86_64_6_8 | test_multidatacenter_local_quorum | Success | | 144.511 |
8_cythonno_x86_64_7_8 | test_network_topology_strategy_counters | Success | | 120.958 |
8_cythonno_x86_64_7_8 | test_add_write_survey_node_after_mv | Success | | 76.316 |
8_cythonno_x86_64_7_8 | test_replace_active_node | Success | | 90.947 |
8_cythonno_x86_64_7_8 | test_stop_decommission_too_few_replicas_multi_dc | Success | | 59.806 |
8_cythonno_x86_64_8_8 | test_network_topology_strategy_each_quorum_counters | Success | | 78.154 |
8_cythonno_x86_64_8_8 | test_complex_repair | Success | | 278.768 |
8_cythonno_x86_64_8_8 | test_replace_nonexistent_node | Success | | 83.336 |
8_cythonno_x86_64_8_8 | test_sstable_marking_not_intersecting_all_ranges | Success | | 83.510 |
8_cythonno_x86_64_1_8 | test_network_topology_strategy | Success | | 281.392 |
8_cythonno_x86_64_1_8 | test_disk_balance_bootstrap | Success | | 134.303 |
8_cythonno_x86_64_1_8 | test_throttled_partition_update | Success | | 528.657 |
8_cythonno_x86_64_1_8 | test_resume_failed_replace | Success | | 173.310 |
8_cythonno_x86_64_1_8 | test_initial_empty_repair_tables | Skipped | hangs CI
/home/cassandra/cassandra-dtest/repair_tests/repair_test.py:1416: hangs CI | 0.001 |
8_cythonno_x86_64_2_8 | test_network_topology_strategy_each_quorum | Success | | 176.777 |
8_cythonno_x86_64_2_8 | test_add_dc_after_mv_simple_replication | Success | | 94.022 |
8_cythonno_x86_64_2_8 | test_really_complex_repair | Success | | 257.900 |
8_cythonno_x86_64_2_8 | test_restart_failed_replace_with_reset_resume_state | Success | | 199.804 |
8_cythonno_x86_64_2_8 | test_repair_parent_table | Success | | 89.170 |
8_cythonno_x86_64_3_8 | test_simple_strategy_users | Success | | 208.074 |
8_cythonno_x86_64_3_8 | test_add_dc_after_mv_network_replication | Success | | 113.464 |
8_cythonno_x86_64_3_8 | test_replace_stopped_node | Success | | 190.798 |
8_cythonno_x86_64_3_8 | test_restart_failed_replace | Success | | 266.000 |
8_cythonno_x86_64_3_8 | test_repair_table | Success | | 82.904 |
8_cythonno_x86_64_4_8 | test_simple_strategy_each_quorum_users | Success | | 65.130 |
8_cythonno_x86_64_4_8 | test_add_node_after_mv | Success | | 67.965 |
8_cythonno_x86_64_4_8 | test_replace_shutdown_node | Success | | 171.310 |
8_cythonno_x86_64_4_8 | test_rf_collapse_gossiping_property_file_snitch_multi_dc | Skipped | 5.1 > 5.0.x
/home/cassandra/cassandra-dtest/conftest.py:468: 5.1 > 5.0.x | 0.188 |
8_cythonno_x86_64_5_8 | test_network_topology_strategy_users | Success | | 243.066 |
8_cythonno_x86_64_5_8 | test_add_node_after_wide_mv_with_range_deletions | Success | | 106.425 |
8_cythonno_x86_64_5_8 | test_replace_stopped_node_same_address | Success | | 184.474 |
8_cythonno_x86_64_5_8 | test_rf_expand_gossiping_property_file_snitch_multi_dc | Skipped | 5.1 > 5.0.x
/home/cassandra/cassandra-dtest/conftest.py:468: 5.1 > 5.0.x | 0.756 |
8_cythonno_x86_64_6_8 | test_network_topology_strategy_each_quorum_users | Success | | 100.170 |
8_cythonno_x86_64_6_8 | test_add_node_after_very_wide_mv | Success | | 294.198 |
8_cythonno_x86_64_6_8 | test_replace_first_boot | Success | | 186.584 |
8_cythonno_x86_64_6_8 | test_multidatacenter_local_quorum | Failure | AssertionError: too many reads hit the degraded node: 3
assert 3 <= 1
self = <snitch_test.TestDynamicEndpointSnitch object at 0x7efdff5ef040>
@pytest.mark.resource_intensive
@since('3.10')
def test_multidatacenter_local_quorum(self):
'''
@jira_ticket CASSANDRA-13074
If we do only local datacenters reads in a multidatacenter DES setup,
DES should take effect and route around a degraded node
'''
def no_cross_dc(scores, cross_dc_nodes):
return all('/' + k.address() not in scores for k in cross_dc_nodes)
def snitchable(scores_before, scores_after, needed_nodes):
return all('/' + k.address() in scores_before and '/' + k.address()
in scores_after for k in needed_nodes)
cluster = self.cluster
cluster.populate([3, 3])
coordinator_node, healthy_node, degraded_node, node4, node5, node6 = cluster.nodelist()
# increase DES reset/update interval so we clear any cross-DC startup reads faster
cluster.set_configuration_options(values={'dynamic_snitch_reset_interval_in_ms': 10000,
'dynamic_snitch_update_interval_in_ms': 50,
'phi_convict_threshold': 12})
# Delay reads on the degraded node by 50 milliseconds
degraded_node.start(jvm_args=['-Dcassandra.test.read_iteration_delay_ms=50',
'-Dcassandra.allow_unsafe_join=true'])
cluster.start()
des = make_mbean('db', type='DynamicEndpointSnitch')
read_stage = make_mbean('metrics', type='ThreadPools', path='request',
scope='ReadStage', name='CompletedTasks')
session = self.patient_exclusive_cql_connection(coordinator_node)
session.execute("CREATE KEYSPACE snitchtestks WITH replication = {'class': 'NetworkTopologyStrategy', 'dc1': 3, 'dc2': 3}")
session.execute("CREATE TABLE snitchtestks.tbl1 (key int PRIMARY KEY) WITH speculative_retry = 'NONE'")
read_stmt = session.prepare("SELECT * FROM snitchtestks.tbl1 where key = ?")
read_stmt.consistency_level = ConsistencyLevel.LOCAL_QUORUM
insert_stmt = session.prepare("INSERT INTO snitchtestks.tbl1 (key) VALUES (?)")
insert_stmt.consistency_level = ConsistencyLevel.ALL
with JolokiaAgent(coordinator_node) as jmx:
with JolokiaAgent(degraded_node) as bad_jmx:
for x in range(0, 150):
session.execute(insert_stmt, [x])
cleared = False
# Wait for a snitch reset in case any earlier
# startup process populated cross-DC read timings
while not cleared:
scores = jmx.read_attribute(des, 'Scores')
cleared = ('/127.0.0.1' in scores and (len(scores) == 1)) or not scores
snitchable_count = 0
for x in range(0, 150):
degraded_reads_before = bad_jmx.read_attribute(read_stage, 'Value')
scores_before = jmx.read_attribute(des, 'Scores')
assert no_cross_dc(scores_before, [node4, node5, node6]), "Cross DC scores were present: " + str(scores_before)
future = session.execute_async(read_stmt, [x])
future.result()
scores_after = jmx.read_attribute(des, 'Scores')
assert no_cross_dc(scores_after, [node4, node5, node6]), "Cross DC scores were present: " + str(scores_after)
if snitchable(scores_before, scores_after,
[coordinator_node, healthy_node, degraded_node]):
snitchable_count = snitchable_count + 1
# If the DES correctly routed the read around the degraded node,
# it shouldn't have another completed read request in metrics,
# unless there was one right after a reset.
degraded_reads = bad_jmx.read_attribute(read_stage, 'Value')
difference = abs(degraded_reads_before - degraded_reads)
> assert difference <= 1, "too many reads hit the degraded node: %s" % difference
E AssertionError: too many reads hit the degraded node: 3
E assert 3 <= 1
snitch_test.py:235: AssertionError | 139.928 |
8_cythonno_x86_64_7_8 | test_network_topology_strategy_counters | Success | | 117.610 |
8_cythonno_x86_64_7_8 | test_add_write_survey_node_after_mv | Success | | 67.373 |
8_cythonno_x86_64_7_8 | test_replace_active_node | Success | | 84.289 |
8_cythonno_x86_64_7_8 | test_stop_decommission_too_few_replicas_multi_dc | Success | | 51.598 |
8_cythonno_x86_64_8_8 | test_network_topology_strategy_each_quorum_counters | Success | | 96.651 |
8_cythonno_x86_64_8_8 | test_complex_repair | Success | | 301.955 |
8_cythonno_x86_64_8_8 | test_replace_nonexistent_node | Success | | 83.609 |
8_cythonno_x86_64_8_8 | test_sstable_marking_not_intersecting_all_ranges | Success | | 82.887 |