add 'string' to all firstrecord_timestamp and lastrecord_timestamp to fix bug
Closes-Bug:#1603529 Change-Id: I5585fc5c376e3220d79d22be3394e3b6ad0e6214
This commit is contained in:
parent
3bf98a894b
commit
c498564929
|
@ -80,11 +80,13 @@ class InsertComponent(Component):
|
||||||
|
|
||||||
value_meta_part = {"record_count": instance_usage_dict.get(
|
value_meta_part = {"record_count": instance_usage_dict.get(
|
||||||
"record_count", 0),
|
"record_count", 0),
|
||||||
"firstrecord_timestamp": instance_usage_dict.get(
|
"firstrecord_timestamp_string":
|
||||||
"firstrecord_timestamp",
|
instance_usage_dict.get(
|
||||||
|
"firstrecord_timestamp_string",
|
||||||
Component.DEFAULT_UNAVAILABLE_VALUE),
|
Component.DEFAULT_UNAVAILABLE_VALUE),
|
||||||
"lastrecord_timestamp": instance_usage_dict.get(
|
"lastrecord_timestamp_string":
|
||||||
"lastrecord_timestamp",
|
instance_usage_dict.get(
|
||||||
|
"lastrecord_timestamp_string",
|
||||||
Component.DEFAULT_UNAVAILABLE_VALUE)}
|
Component.DEFAULT_UNAVAILABLE_VALUE)}
|
||||||
|
|
||||||
metric_part = {"name": instance_usage_dict.get(
|
metric_part = {"name": instance_usage_dict.get(
|
||||||
|
@ -117,9 +119,9 @@ class InsertComponent(Component):
|
||||||
"aggregated_metric_name":
|
"aggregated_metric_name":
|
||||||
row.aggregated_metric_name,
|
row.aggregated_metric_name,
|
||||||
"quantity": row.quantity,
|
"quantity": row.quantity,
|
||||||
"firstrecord_timestamp":
|
"firstrecord_timestamp_string":
|
||||||
row.firstrecord_timestamp_string,
|
row.firstrecord_timestamp_string,
|
||||||
"lastrecord_timestamp":
|
"lastrecord_timestamp_string":
|
||||||
row.lastrecord_timestamp_string,
|
row.lastrecord_timestamp_string,
|
||||||
"record_count": row.record_count,
|
"record_count": row.record_count,
|
||||||
"service_group": row.service_group,
|
"service_group": row.service_group,
|
||||||
|
@ -153,9 +155,9 @@ class InsertComponent(Component):
|
||||||
"aggregated_metric_name":
|
"aggregated_metric_name":
|
||||||
row.aggregated_metric_name,
|
row.aggregated_metric_name,
|
||||||
"quantity": row.quantity,
|
"quantity": row.quantity,
|
||||||
"firstrecord_timestamp":
|
"firstrecord_timestamp_string":
|
||||||
row.firstrecord_timestamp_string,
|
row.firstrecord_timestamp_string,
|
||||||
"lastrecord_timestamp":
|
"lastrecord_timestamp_string":
|
||||||
row.lastrecord_timestamp_string,
|
row.lastrecord_timestamp_string,
|
||||||
"firstrecord_timestamp_unix":
|
"firstrecord_timestamp_unix":
|
||||||
row.firstrecord_timestamp_unix,
|
row.firstrecord_timestamp_unix,
|
||||||
|
|
|
@ -148,10 +148,10 @@ class SparkTest(SparkContextTest):
|
||||||
.get('record_count'))
|
.get('record_count'))
|
||||||
self.assertEqual('2016-01-20 16:40:00',
|
self.assertEqual('2016-01-20 16:40:00',
|
||||||
total_mb_agg_metric.get('metric').get('value_meta')
|
total_mb_agg_metric.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:46',
|
self.assertEqual('2016-01-20 16:40:46',
|
||||||
total_mb_agg_metric.get('metric').get('value_meta')
|
total_mb_agg_metric.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify mem.usable_mb_agg metrics
|
# Verify mem.usable_mb_agg metrics
|
||||||
usable_mb_agg_metric = [
|
usable_mb_agg_metric = [
|
||||||
|
@ -180,10 +180,10 @@ class SparkTest(SparkContextTest):
|
||||||
.get('record_count'))
|
.get('record_count'))
|
||||||
self.assertEqual('2016-01-20 16:40:00',
|
self.assertEqual('2016-01-20 16:40:00',
|
||||||
usable_mb_agg_metric.get('metric').get('value_meta')
|
usable_mb_agg_metric.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:46',
|
self.assertEqual('2016-01-20 16:40:46',
|
||||||
usable_mb_agg_metric.get('metric').get('value_meta')
|
usable_mb_agg_metric.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify vcpus_agg metrics for all projects
|
# Verify vcpus_agg metrics for all projects
|
||||||
vcpus_agg_metric = [
|
vcpus_agg_metric = [
|
||||||
|
@ -219,11 +219,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-01-20 16:40:00',
|
self.assertEqual('2016-01-20 16:40:00',
|
||||||
vcpus_agg_metric
|
vcpus_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:46',
|
self.assertEqual('2016-01-20 16:40:46',
|
||||||
vcpus_agg_metric
|
vcpus_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify vcpus_agg metrics for 8647fd5030b04a799b0411cc38c4102d
|
# Verify vcpus_agg metrics for 8647fd5030b04a799b0411cc38c4102d
|
||||||
# project
|
# project
|
||||||
|
@ -260,11 +260,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-01-20 16:40:00',
|
self.assertEqual('2016-01-20 16:40:00',
|
||||||
vcpus_agg_metric
|
vcpus_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:42',
|
self.assertEqual('2016-01-20 16:40:42',
|
||||||
vcpus_agg_metric
|
vcpus_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify vcpus_agg metrics for 9647fd5030b04a799b0411cc38c4102d
|
# Verify vcpus_agg metrics for 9647fd5030b04a799b0411cc38c4102d
|
||||||
# project
|
# project
|
||||||
|
@ -301,11 +301,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-01-20 16:40:05',
|
self.assertEqual('2016-01-20 16:40:05',
|
||||||
vcpus_agg_metric
|
vcpus_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:46',
|
self.assertEqual('2016-01-20 16:40:46',
|
||||||
vcpus_agg_metric
|
vcpus_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify vm.mem.total_mb_agg metrics for all projects
|
# Verify vm.mem.total_mb_agg metrics for all projects
|
||||||
vm_mem_total_mb_agg_metric = [
|
vm_mem_total_mb_agg_metric = [
|
||||||
|
@ -341,11 +341,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-06-07 16:27:54',
|
self.assertEqual('2016-06-07 16:27:54',
|
||||||
vm_mem_total_mb_agg_metric
|
vm_mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-06-07 16:30:54',
|
self.assertEqual('2016-06-07 16:30:54',
|
||||||
vm_mem_total_mb_agg_metric
|
vm_mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify vm.mem.total_mb_agg metrics for the 1 project
|
# Verify vm.mem.total_mb_agg metrics for the 1 project
|
||||||
vm_mem_total_mb_agg_metric = [
|
vm_mem_total_mb_agg_metric = [
|
||||||
|
@ -380,11 +380,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-06-07 16:27:54',
|
self.assertEqual('2016-06-07 16:27:54',
|
||||||
vm_mem_total_mb_agg_metric
|
vm_mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-06-07 16:29:54',
|
self.assertEqual('2016-06-07 16:29:54',
|
||||||
vm_mem_total_mb_agg_metric
|
vm_mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify vm.mem.total_mb_agg metrics for the 2 project
|
# Verify vm.mem.total_mb_agg metrics for the 2 project
|
||||||
vm_mem_total_mb_agg_metric = [
|
vm_mem_total_mb_agg_metric = [
|
||||||
|
@ -419,11 +419,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-06-07 16:28:54',
|
self.assertEqual('2016-06-07 16:28:54',
|
||||||
vm_mem_total_mb_agg_metric
|
vm_mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-06-07 16:30:54',
|
self.assertEqual('2016-06-07 16:30:54',
|
||||||
vm_mem_total_mb_agg_metric
|
vm_mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify nova.vm.disk.total_allocated_gb_agg metrics
|
# Verify nova.vm.disk.total_allocated_gb_agg metrics
|
||||||
total_allocated_disk_agg_metric = [
|
total_allocated_disk_agg_metric = [
|
||||||
|
@ -454,10 +454,12 @@ class SparkTest(SparkContextTest):
|
||||||
.get('value_meta').get('record_count'))
|
.get('value_meta').get('record_count'))
|
||||||
self.assertEqual('2016-05-17 15:14:08',
|
self.assertEqual('2016-05-17 15:14:08',
|
||||||
total_allocated_disk_agg_metric.get('metric')
|
total_allocated_disk_agg_metric.get('metric')
|
||||||
.get('value_meta').get('firstrecord_timestamp'))
|
.get('value_meta')
|
||||||
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-05-17 15:14:44',
|
self.assertEqual('2016-05-17 15:14:44',
|
||||||
total_allocated_disk_agg_metric.get('metric')
|
total_allocated_disk_agg_metric.get('metric')
|
||||||
.get('value_meta').get('lastrecord_timestamp'))
|
.get('value_meta')
|
||||||
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify vm.disk.allocation_agg metrics for all projects
|
# Verify vm.disk.allocation_agg metrics for all projects
|
||||||
vm_disk_allocation_agg_metric = [
|
vm_disk_allocation_agg_metric = [
|
||||||
|
@ -493,11 +495,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-01-20 16:40:00',
|
self.assertEqual('2016-01-20 16:40:00',
|
||||||
vm_disk_allocation_agg_metric
|
vm_disk_allocation_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:46',
|
self.assertEqual('2016-01-20 16:40:46',
|
||||||
vm_disk_allocation_agg_metric
|
vm_disk_allocation_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify vm.disk.allocation_agg metrics for
|
# Verify vm.disk.allocation_agg metrics for
|
||||||
# 5f681592f7084c5fbcd4e8a20a4fef15 project
|
# 5f681592f7084c5fbcd4e8a20a4fef15 project
|
||||||
|
@ -534,11 +536,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-01-20 16:40:00',
|
self.assertEqual('2016-01-20 16:40:00',
|
||||||
vm_disk_allocation_agg_metric
|
vm_disk_allocation_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:40',
|
self.assertEqual('2016-01-20 16:40:40',
|
||||||
vm_disk_allocation_agg_metric
|
vm_disk_allocation_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify vm.disk.allocation_agg metrics for
|
# Verify vm.disk.allocation_agg metrics for
|
||||||
# 6f681592f7084c5fbcd4e8a20a4fef15 project
|
# 6f681592f7084c5fbcd4e8a20a4fef15 project
|
||||||
|
@ -575,11 +577,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-01-20 16:40:00',
|
self.assertEqual('2016-01-20 16:40:00',
|
||||||
vm_disk_allocation_agg_metric
|
vm_disk_allocation_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:46',
|
self.assertEqual('2016-01-20 16:40:46',
|
||||||
vm_disk_allocation_agg_metric
|
vm_disk_allocation_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify vm.cpu.utilization_perc_agg metrics for
|
# Verify vm.cpu.utilization_perc_agg metrics for
|
||||||
# 817331145b804dc9a7accb6edfb0674d project
|
# 817331145b804dc9a7accb6edfb0674d project
|
||||||
|
@ -613,10 +615,12 @@ class SparkTest(SparkContextTest):
|
||||||
.get('value_meta').get('record_count'))
|
.get('value_meta').get('record_count'))
|
||||||
self.assertEqual('2016-05-26 17:31:01',
|
self.assertEqual('2016-05-26 17:31:01',
|
||||||
vm_cpu_util_perc_agg_metric.get('metric')
|
vm_cpu_util_perc_agg_metric.get('metric')
|
||||||
.get('value_meta').get('firstrecord_timestamp'))
|
.get('value_meta')
|
||||||
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-05-26 17:31:36',
|
self.assertEqual('2016-05-26 17:31:36',
|
||||||
vm_cpu_util_perc_agg_metric.get('metric')
|
vm_cpu_util_perc_agg_metric.get('metric')
|
||||||
.get('value_meta').get('lastrecord_timestamp'))
|
.get('value_meta')
|
||||||
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify vm.cpu.utilization_perc_agg metrics for
|
# Verify vm.cpu.utilization_perc_agg metrics for
|
||||||
# 5d0e49bdc4534bb4b65909228aa040da project
|
# 5d0e49bdc4534bb4b65909228aa040da project
|
||||||
|
@ -650,10 +654,12 @@ class SparkTest(SparkContextTest):
|
||||||
.get('value_meta').get('record_count'))
|
.get('value_meta').get('record_count'))
|
||||||
self.assertEqual('2016-05-26 17:30:27',
|
self.assertEqual('2016-05-26 17:30:27',
|
||||||
vm_cpu_util_perc_agg_metric.get('metric')
|
vm_cpu_util_perc_agg_metric.get('metric')
|
||||||
.get('value_meta').get('firstrecord_timestamp'))
|
.get('value_meta')
|
||||||
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-05-26 17:30:27',
|
self.assertEqual('2016-05-26 17:30:27',
|
||||||
vm_cpu_util_perc_agg_metric.get('metric')
|
vm_cpu_util_perc_agg_metric.get('metric')
|
||||||
.get('value_meta').get('lastrecord_timestamp'))
|
.get('value_meta')
|
||||||
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify disk.total_space_mb_agg metrics
|
# Verify disk.total_space_mb_agg metrics
|
||||||
disk_total_space_agg_metric = [
|
disk_total_space_agg_metric = [
|
||||||
|
@ -687,11 +693,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-06-01 21:09:21',
|
self.assertEqual('2016-06-01 21:09:21',
|
||||||
disk_total_space_agg_metric
|
disk_total_space_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-06-01 21:09:21',
|
self.assertEqual('2016-06-01 21:09:21',
|
||||||
disk_total_space_agg_metric
|
disk_total_space_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify disk.total_used_space_mb_agg metrics
|
# Verify disk.total_used_space_mb_agg metrics
|
||||||
disk_total_used_agg_metric = [
|
disk_total_used_agg_metric = [
|
||||||
|
@ -725,11 +731,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-06-01 21:09:21',
|
self.assertEqual('2016-06-01 21:09:21',
|
||||||
disk_total_used_agg_metric
|
disk_total_used_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-06-01 21:09:21',
|
self.assertEqual('2016-06-01 21:09:21',
|
||||||
disk_total_used_agg_metric
|
disk_total_used_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify cpu.utilized_logical_cores_agg metrics for all hosts
|
# Verify cpu.utilized_logical_cores_agg metrics for all hosts
|
||||||
cpu_util_cores_agg_metric = [
|
cpu_util_cores_agg_metric = [
|
||||||
|
@ -766,11 +772,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-03-07 16:09:23',
|
self.assertEqual('2016-03-07 16:09:23',
|
||||||
cpu_util_cores_agg_metric
|
cpu_util_cores_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-03-07 16:10:38',
|
self.assertEqual('2016-03-07 16:10:38',
|
||||||
cpu_util_cores_agg_metric
|
cpu_util_cores_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify cpu.utilized_logical_cores_agg metrics for
|
# Verify cpu.utilized_logical_cores_agg metrics for
|
||||||
# test-cp1-comp0294-mgmt host
|
# test-cp1-comp0294-mgmt host
|
||||||
|
@ -808,11 +814,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-03-07 16:09:23',
|
self.assertEqual('2016-03-07 16:09:23',
|
||||||
cpu_util_cores_agg_metric
|
cpu_util_cores_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-03-07 16:10:38',
|
self.assertEqual('2016-03-07 16:10:38',
|
||||||
cpu_util_cores_agg_metric
|
cpu_util_cores_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify vm.cpu.utilization_perc_agg metrics for
|
# Verify vm.cpu.utilization_perc_agg metrics for
|
||||||
# test-cp1-comp0037-mgmt host
|
# test-cp1-comp0037-mgmt host
|
||||||
|
@ -850,11 +856,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-03-07 16:09:23',
|
self.assertEqual('2016-03-07 16:09:23',
|
||||||
cpu_util_cores_agg_metric
|
cpu_util_cores_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-03-07 16:10:38',
|
self.assertEqual('2016-03-07 16:10:38',
|
||||||
cpu_util_cores_agg_metric
|
cpu_util_cores_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify vm.mem.used_mb_agg metrics for all projects
|
# Verify vm.mem.used_mb_agg metrics for all projects
|
||||||
vm_mem_used_mb_agg_metric = [
|
vm_mem_used_mb_agg_metric = [
|
||||||
|
@ -890,11 +896,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-06-07 16:27:54',
|
self.assertEqual('2016-06-07 16:27:54',
|
||||||
vm_mem_used_mb_agg_metric
|
vm_mem_used_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-06-07 16:30:54',
|
self.assertEqual('2016-06-07 16:30:54',
|
||||||
vm_mem_used_mb_agg_metric
|
vm_mem_used_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify vm.mem.used_mb_agg metrics for the 1 project
|
# Verify vm.mem.used_mb_agg metrics for the 1 project
|
||||||
vm_mem_used_mb_agg_metric = [
|
vm_mem_used_mb_agg_metric = [
|
||||||
|
@ -929,11 +935,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-06-07 16:27:54',
|
self.assertEqual('2016-06-07 16:27:54',
|
||||||
vm_mem_used_mb_agg_metric
|
vm_mem_used_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-06-07 16:29:54',
|
self.assertEqual('2016-06-07 16:29:54',
|
||||||
vm_mem_used_mb_agg_metric
|
vm_mem_used_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify vm.mem.used_mb_agg metrics for the 2 project
|
# Verify vm.mem.used_mb_agg metrics for the 2 project
|
||||||
vm_mem_used_mb_agg_metric = [
|
vm_mem_used_mb_agg_metric = [
|
||||||
|
@ -968,11 +974,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-06-07 16:28:54',
|
self.assertEqual('2016-06-07 16:28:54',
|
||||||
vm_mem_used_mb_agg_metric
|
vm_mem_used_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-06-07 16:30:54',
|
self.assertEqual('2016-06-07 16:30:54',
|
||||||
vm_mem_used_mb_agg_metric
|
vm_mem_used_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify swiftlm.diskusage.val.size_agg for all hosts
|
# Verify swiftlm.diskusage.val.size_agg for all hosts
|
||||||
used_swift_agg_metric = [
|
used_swift_agg_metric = [
|
||||||
|
@ -998,10 +1004,10 @@ class SparkTest(SparkContextTest):
|
||||||
.get('record_count'))
|
.get('record_count'))
|
||||||
self.assertEqual('2016-06-10 20:27:01',
|
self.assertEqual('2016-06-10 20:27:01',
|
||||||
used_swift_agg_metric.get('metric').get('value_meta')
|
used_swift_agg_metric.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-06-10 20:27:01',
|
self.assertEqual('2016-06-10 20:27:01',
|
||||||
used_swift_agg_metric.get('metric').get('value_meta')
|
used_swift_agg_metric.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify swiftlm.diskusage.val.size_agg for host a
|
# Verify swiftlm.diskusage.val.size_agg for host a
|
||||||
used_swift_agg_metric = [
|
used_swift_agg_metric = [
|
||||||
|
@ -1027,10 +1033,10 @@ class SparkTest(SparkContextTest):
|
||||||
.get('record_count'))
|
.get('record_count'))
|
||||||
self.assertEqual('2016-06-10 20:27:01',
|
self.assertEqual('2016-06-10 20:27:01',
|
||||||
used_swift_agg_metric.get('metric').get('value_meta')
|
used_swift_agg_metric.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-06-10 20:27:01',
|
self.assertEqual('2016-06-10 20:27:01',
|
||||||
used_swift_agg_metric.get('metric').get('value_meta')
|
used_swift_agg_metric.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify swiftlm.diskusage.val.size_agg for host b
|
# Verify swiftlm.diskusage.val.size_agg for host b
|
||||||
used_swift_agg_metric = [
|
used_swift_agg_metric = [
|
||||||
|
@ -1056,10 +1062,10 @@ class SparkTest(SparkContextTest):
|
||||||
.get('record_count'))
|
.get('record_count'))
|
||||||
self.assertEqual('2016-06-10 20:27:01',
|
self.assertEqual('2016-06-10 20:27:01',
|
||||||
used_swift_agg_metric.get('metric').get('value_meta')
|
used_swift_agg_metric.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-06-10 20:27:01',
|
self.assertEqual('2016-06-10 20:27:01',
|
||||||
used_swift_agg_metric.get('metric').get('value_meta')
|
used_swift_agg_metric.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify swiftlm.diskusage.val.avail_agg for all hosts
|
# Verify swiftlm.diskusage.val.avail_agg for all hosts
|
||||||
avail_swift_agg_metric = [
|
avail_swift_agg_metric = [
|
||||||
|
@ -1085,10 +1091,10 @@ class SparkTest(SparkContextTest):
|
||||||
.get('record_count'))
|
.get('record_count'))
|
||||||
self.assertEqual('2016-06-10 20:27:01',
|
self.assertEqual('2016-06-10 20:27:01',
|
||||||
avail_swift_agg_metric.get('metric').get('value_meta')
|
avail_swift_agg_metric.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-06-10 20:27:01',
|
self.assertEqual('2016-06-10 20:27:01',
|
||||||
avail_swift_agg_metric.get('metric').get('value_meta')
|
avail_swift_agg_metric.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify swiftlm.diskusage.val.avail_agg for host a
|
# Verify swiftlm.diskusage.val.avail_agg for host a
|
||||||
avail_swift_agg_metric = [
|
avail_swift_agg_metric = [
|
||||||
|
@ -1114,10 +1120,10 @@ class SparkTest(SparkContextTest):
|
||||||
.get('record_count'))
|
.get('record_count'))
|
||||||
self.assertEqual('2016-06-10 20:27:01',
|
self.assertEqual('2016-06-10 20:27:01',
|
||||||
avail_swift_agg_metric.get('metric').get('value_meta')
|
avail_swift_agg_metric.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-06-10 20:27:01',
|
self.assertEqual('2016-06-10 20:27:01',
|
||||||
avail_swift_agg_metric.get('metric').get('value_meta')
|
avail_swift_agg_metric.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify swiftlm.diskusage.val.avail_agg for host b
|
# Verify swiftlm.diskusage.val.avail_agg for host b
|
||||||
avail_swift_agg_metric = [
|
avail_swift_agg_metric = [
|
||||||
|
@ -1143,10 +1149,10 @@ class SparkTest(SparkContextTest):
|
||||||
.get('record_count'))
|
.get('record_count'))
|
||||||
self.assertEqual('2016-06-10 20:27:01',
|
self.assertEqual('2016-06-10 20:27:01',
|
||||||
avail_swift_agg_metric.get('metric').get('value_meta')
|
avail_swift_agg_metric.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-06-10 20:27:01',
|
self.assertEqual('2016-06-10 20:27:01',
|
||||||
avail_swift_agg_metric.get('metric').get('value_meta')
|
avail_swift_agg_metric.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify swiftlm.diskusage.rate_agg metrics
|
# Verify swiftlm.diskusage.rate_agg metrics
|
||||||
diskusage_rate_agg_metric = [
|
diskusage_rate_agg_metric = [
|
||||||
|
@ -1174,10 +1180,12 @@ class SparkTest(SparkContextTest):
|
||||||
.get('value_meta').get('record_count'))
|
.get('value_meta').get('record_count'))
|
||||||
self.assertEqual('2016-06-10 20:27:01',
|
self.assertEqual('2016-06-10 20:27:01',
|
||||||
diskusage_rate_agg_metric.get('metric')
|
diskusage_rate_agg_metric.get('metric')
|
||||||
.get('value_meta').get('firstrecord_timestamp'))
|
.get('value_meta')
|
||||||
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-06-10 20:27:01',
|
self.assertEqual('2016-06-10 20:27:01',
|
||||||
diskusage_rate_agg_metric.get('metric')
|
diskusage_rate_agg_metric.get('metric')
|
||||||
.get('value_meta').get('lastrecord_timestamp'))
|
.get('value_meta')
|
||||||
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
|
|
||||||
def simple_count_transform(rdd):
|
def simple_count_transform(rdd):
|
||||||
|
|
|
@ -90,11 +90,13 @@ class TestPreHourlyProcessorAgg(SparkContextTest):
|
||||||
self.assertEqual("2016-06-20 11:49:44",
|
self.assertEqual("2016-06-20 11:49:44",
|
||||||
mem_total_mb_agg_metric
|
mem_total_mb_agg_metric
|
||||||
.get("metric")
|
.get("metric")
|
||||||
.get('value_meta').get('lastrecord_timestamp'))
|
.get('value_meta')
|
||||||
|
.get('lastrecord_timestamp_string'))
|
||||||
self.assertEqual("2016-06-20 11:24:59",
|
self.assertEqual("2016-06-20 11:24:59",
|
||||||
mem_total_mb_agg_metric
|
mem_total_mb_agg_metric
|
||||||
.get("metric")
|
.get("metric")
|
||||||
.get('value_meta').get('firstrecord_timestamp'))
|
.get('value_meta')
|
||||||
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual(60.0,
|
self.assertEqual(60.0,
|
||||||
mem_total_mb_agg_metric
|
mem_total_mb_agg_metric
|
||||||
.get("metric")
|
.get("metric")
|
||||||
|
@ -114,11 +116,13 @@ class TestPreHourlyProcessorAgg(SparkContextTest):
|
||||||
self.assertEqual("2016-06-20 11:49:44",
|
self.assertEqual("2016-06-20 11:49:44",
|
||||||
mem_usable_mb_agg_metric
|
mem_usable_mb_agg_metric
|
||||||
.get("metric")
|
.get("metric")
|
||||||
.get('value_meta').get('lastrecord_timestamp'))
|
.get('value_meta')
|
||||||
|
.get('lastrecord_timestamp_string'))
|
||||||
self.assertEqual("2016-06-20 11:24:59",
|
self.assertEqual("2016-06-20 11:24:59",
|
||||||
mem_usable_mb_agg_metric
|
mem_usable_mb_agg_metric
|
||||||
.get("metric")
|
.get("metric")
|
||||||
.get('value_meta').get('firstrecord_timestamp'))
|
.get('value_meta')
|
||||||
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual(60.0,
|
self.assertEqual(60.0,
|
||||||
mem_usable_mb_agg_metric
|
mem_usable_mb_agg_metric
|
||||||
.get("metric")
|
.get("metric")
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
('',{"service_group":"all","geolocation":"all","aggregated_metric_name":"mem.total_mb_agg","record_count":20.0,"user_id":"all","zone":"all","firstrecord_timestamp":"2016-06-20 11:24:59","tenant_id":"all","region":"all","usage_hour":"11","usage_date":"2016-06-20","processing_meta":{"metric_id":"mem_total_all"},"resource_uuid":"all","aggregation_period":"hourly","host":"all","lastrecord_timestamp":"2016-06-20 11:29:44","firstrecord_timestamp_unix":1466421899.0,"service_id":"all","project_id":"all","usage_minute":"all","lastrecord_timestamp_unix":1466422184.0,"quantity":16049.0})
|
('',{"service_group":"all","geolocation":"all","aggregated_metric_name":"mem.total_mb_agg","record_count":20.0,"user_id":"all","zone":"all","firstrecord_timestamp_string":"2016-06-20 11:24:59","tenant_id":"all","region":"all","usage_hour":"11","usage_date":"2016-06-20","processing_meta":{"metric_id":"mem_total_all"},"resource_uuid":"all","aggregation_period":"hourly","host":"all","lastrecord_timestamp_string":"2016-06-20 11:29:44","firstrecord_timestamp_unix":1466421899.0,"service_id":"all","project_id":"all","usage_minute":"all","lastrecord_timestamp_unix":1466422184.0,"quantity":16049.0})
|
||||||
('',{"service_group":"all","geolocation":"all","aggregated_metric_name":"mem.total_mb_agg","record_count":20.0,"user_id":"all","zone":"all","firstrecord_timestamp":"2016-06-20 11:29:44","tenant_id":"all","region":"all","usage_hour":"11","usage_date":"2016-06-20","processing_meta":{"metric_id":"mem_total_all"},"resource_uuid":"all","aggregation_period":"hourly","host":"all","lastrecord_timestamp":"2016-06-20 11:39:44","firstrecord_timestamp_unix":1466421899.0,"service_id":"all","project_id":"all","usage_minute":"all","lastrecord_timestamp_unix":1466422784.0,"quantity":16049.0})
|
('',{"service_group":"all","geolocation":"all","aggregated_metric_name":"mem.total_mb_agg","record_count":20.0,"user_id":"all","zone":"all","firstrecord_timestamp_string":"2016-06-20 11:29:44","tenant_id":"all","region":"all","usage_hour":"11","usage_date":"2016-06-20","processing_meta":{"metric_id":"mem_total_all"},"resource_uuid":"all","aggregation_period":"hourly","host":"all","lastrecord_timestamp_string":"2016-06-20 11:39:44","firstrecord_timestamp_unix":1466421899.0,"service_id":"all","project_id":"all","usage_minute":"all","lastrecord_timestamp_unix":1466422784.0,"quantity":16049.0})
|
||||||
('',{"service_group":"all","geolocation":"all","aggregated_metric_name":"mem.total_mb_agg","record_count":20.0,"user_id":"all","zone":"all","firstrecord_timestamp":"2016-06-20 11:39:44","tenant_id":"all","region":"all","usage_hour":"11","usage_date":"2016-06-20","processing_meta":{"metric_id":"mem_total_all"},"resource_uuid":"all","aggregation_period":"hourly","host":"all","lastrecord_timestamp":"2016-06-20 11:49:44","firstrecord_timestamp_unix":1466421899.0,"service_id":"all","project_id":"all","usage_minute":"all","lastrecord_timestamp_unix":1466423384.0,"quantity":16049.0})
|
('',{"service_group":"all","geolocation":"all","aggregated_metric_name":"mem.total_mb_agg","record_count":20.0,"user_id":"all","zone":"all","firstrecord_timestamp_string":"2016-06-20 11:39:44","tenant_id":"all","region":"all","usage_hour":"11","usage_date":"2016-06-20","processing_meta":{"metric_id":"mem_total_all"},"resource_uuid":"all","aggregation_period":"hourly","host":"all","lastrecord_timestamp_string":"2016-06-20 11:49:44","firstrecord_timestamp_unix":1466421899.0,"service_id":"all","project_id":"all","usage_minute":"all","lastrecord_timestamp_unix":1466423384.0,"quantity":16049.0})
|
||||||
('',{"service_group":"all","geolocation":"all","aggregated_metric_name":"mem.usable_mb_agg","record_count":20.0,"user_id":"all","zone":"all","firstrecord_timestamp":"2016-06-20 11:24:59","tenant_id":"all","region":"all","usage_hour":"11","usage_date":"2016-06-20","processing_meta":{"metric_id":"mem_usable_all"},"resource_uuid":"all","aggregation_period":"hourly","host":"all","lastrecord_timestamp":"2016-06-20 11:29:44","firstrecord_timestamp_unix":1466421899.0,"service_id":"all","project_id":"all","usage_minute":"all","lastrecord_timestamp_unix":1466422184.0,"quantity":10283.1})
|
('',{"service_group":"all","geolocation":"all","aggregated_metric_name":"mem.usable_mb_agg","record_count":20.0,"user_id":"all","zone":"all","firstrecord_timestamp_string":"2016-06-20 11:24:59","tenant_id":"all","region":"all","usage_hour":"11","usage_date":"2016-06-20","processing_meta":{"metric_id":"mem_usable_all"},"resource_uuid":"all","aggregation_period":"hourly","host":"all","lastrecord_timestamp_string":"2016-06-20 11:29:44","firstrecord_timestamp_unix":1466421899.0,"service_id":"all","project_id":"all","usage_minute":"all","lastrecord_timestamp_unix":1466422184.0,"quantity":10283.1})
|
||||||
('',{"service_group":"all","geolocation":"all","aggregated_metric_name":"mem.usable_mb_agg","record_count":20.0,"user_id":"all","zone":"all","firstrecord_timestamp":"2016-06-20 11:29:44","tenant_id":"all","region":"all","usage_hour":"11","usage_date":"2016-06-20","processing_meta":{"metric_id":"mem_usable_all"},"resource_uuid":"all","aggregation_period":"hourly","host":"all","lastrecord_timestamp":"2016-06-20 11:39:44","firstrecord_timestamp_unix":1466421899.0,"service_id":"all","project_id":"all","usage_minute":"all","lastrecord_timestamp_unix":1466422784.0,"quantity":10283.1})
|
('',{"service_group":"all","geolocation":"all","aggregated_metric_name":"mem.usable_mb_agg","record_count":20.0,"user_id":"all","zone":"all","firstrecord_timestamp_string":"2016-06-20 11:29:44","tenant_id":"all","region":"all","usage_hour":"11","usage_date":"2016-06-20","processing_meta":{"metric_id":"mem_usable_all"},"resource_uuid":"all","aggregation_period":"hourly","host":"all","lastrecord_timestamp_string":"2016-06-20 11:39:44","firstrecord_timestamp_unix":1466421899.0,"service_id":"all","project_id":"all","usage_minute":"all","lastrecord_timestamp_unix":1466422784.0,"quantity":10283.1})
|
||||||
('',{"service_group":"all","geolocation":"all","aggregated_metric_name":"mem.usable_mb_agg","record_count":20.0,"user_id":"all","zone":"all","firstrecord_timestamp":"2016-06-20 11:39:44","tenant_id":"all","region":"all","usage_hour":"11","usage_date":"2016-06-20","processing_meta":{"metric_id":"mem_usable_all"},"resource_uuid":"all","aggregation_period":"hourly","host":"all","lastrecord_timestamp":"2016-06-20 11:49:44","firstrecord_timestamp_unix":1466421899.0,"service_id":"all","project_id":"all","usage_minute":"all","lastrecord_timestamp_unix":1466423384.0,"quantity":10283.1})
|
('',{"service_group":"all","geolocation":"all","aggregated_metric_name":"mem.usable_mb_agg","record_count":20.0,"user_id":"all","zone":"all","firstrecord_timestamp_string":"2016-06-20 11:39:44","tenant_id":"all","region":"all","usage_hour":"11","usage_date":"2016-06-20","processing_meta":{"metric_id":"mem_usable_all"},"resource_uuid":"all","aggregation_period":"hourly","host":"all","lastrecord_timestamp_string":"2016-06-20 11:49:44","firstrecord_timestamp_unix":1466421899.0,"service_id":"all","project_id":"all","usage_minute":"all","lastrecord_timestamp_unix":1466423384.0,"quantity":10283.1})
|
||||||
|
|
|
@ -228,11 +228,11 @@ class TestFetchQuantityAgg(SparkContextTest):
|
||||||
self.assertEqual('2016-01-20 16:40:00',
|
self.assertEqual('2016-01-20 16:40:00',
|
||||||
mem_total_mb_agg_metric
|
mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:46',
|
self.assertEqual('2016-01-20 16:40:46',
|
||||||
mem_total_mb_agg_metric
|
mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
||||||
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
||||||
|
@ -335,11 +335,11 @@ class TestFetchQuantityAgg(SparkContextTest):
|
||||||
self.assertEqual('2016-01-20 16:40:00',
|
self.assertEqual('2016-01-20 16:40:00',
|
||||||
mem_total_mb_agg_metric
|
mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:46',
|
self.assertEqual('2016-01-20 16:40:46',
|
||||||
mem_total_mb_agg_metric
|
mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
||||||
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
||||||
|
@ -441,11 +441,11 @@ class TestFetchQuantityAgg(SparkContextTest):
|
||||||
self.assertEqual('2016-01-20 16:40:00',
|
self.assertEqual('2016-01-20 16:40:00',
|
||||||
mem_total_mb_agg_metric
|
mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:46',
|
self.assertEqual('2016-01-20 16:40:46',
|
||||||
mem_total_mb_agg_metric
|
mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
||||||
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
||||||
|
@ -548,11 +548,11 @@ class TestFetchQuantityAgg(SparkContextTest):
|
||||||
self.assertEqual('2016-01-20 16:40:00',
|
self.assertEqual('2016-01-20 16:40:00',
|
||||||
mem_total_mb_agg_metric
|
mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:46',
|
self.assertEqual('2016-01-20 16:40:46',
|
||||||
mem_total_mb_agg_metric
|
mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
||||||
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
||||||
|
@ -655,11 +655,11 @@ class TestFetchQuantityAgg(SparkContextTest):
|
||||||
self.assertEqual('2016-01-20 16:40:00',
|
self.assertEqual('2016-01-20 16:40:00',
|
||||||
mem_total_mb_agg_metric
|
mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:46',
|
self.assertEqual('2016-01-20 16:40:46',
|
||||||
mem_total_mb_agg_metric
|
mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
||||||
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
||||||
|
@ -762,11 +762,11 @@ class TestFetchQuantityAgg(SparkContextTest):
|
||||||
self.assertEqual('2016-01-20 16:40:00',
|
self.assertEqual('2016-01-20 16:40:00',
|
||||||
mem_total_mb_agg_metric
|
mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:46',
|
self.assertEqual('2016-01-20 16:40:46',
|
||||||
mem_total_mb_agg_metric
|
mem_total_mb_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
||||||
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
||||||
|
|
|
@ -211,11 +211,11 @@ class TestFetchQuantityUtilAgg(SparkContextTest):
|
||||||
self.assertEqual('2016-03-07 16:09:23',
|
self.assertEqual('2016-03-07 16:09:23',
|
||||||
utilized_cpu_logical_agg_metric.get(
|
utilized_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-03-07 16:10:38',
|
self.assertEqual('2016-03-07 16:10:38',
|
||||||
utilized_cpu_logical_agg_metric.get(
|
utilized_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
||||||
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
||||||
|
@ -314,11 +314,11 @@ class TestFetchQuantityUtilAgg(SparkContextTest):
|
||||||
self.assertEqual('2016-03-07 16:09:23',
|
self.assertEqual('2016-03-07 16:09:23',
|
||||||
utilized_cpu_logical_agg_metric.get(
|
utilized_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-03-07 16:10:38',
|
self.assertEqual('2016-03-07 16:10:38',
|
||||||
utilized_cpu_logical_agg_metric.get(
|
utilized_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
||||||
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
||||||
|
@ -417,11 +417,11 @@ class TestFetchQuantityUtilAgg(SparkContextTest):
|
||||||
self.assertEqual('2016-03-07 16:09:23',
|
self.assertEqual('2016-03-07 16:09:23',
|
||||||
utilized_cpu_logical_agg_metric.get(
|
utilized_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-03-07 16:10:38',
|
self.assertEqual('2016-03-07 16:10:38',
|
||||||
utilized_cpu_logical_agg_metric.get(
|
utilized_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
@mock.patch('monasca_transform.data_driven_specs.data_driven_specs_repo.'
|
||||||
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
'DataDrivenSpecsRepoFactory.get_data_driven_specs_repo')
|
||||||
|
|
|
@ -125,11 +125,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-03-07 16:09:23',
|
self.assertEqual('2016-03-07 16:09:23',
|
||||||
total_cpu_logical_agg_metric.get(
|
total_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-03-07 16:10:38',
|
self.assertEqual('2016-03-07 16:10:38',
|
||||||
total_cpu_logical_agg_metric.get(
|
total_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify cpu.total_logical_cores_agg for test-cp1-comp0333-mgmt host
|
# Verify cpu.total_logical_cores_agg for test-cp1-comp0333-mgmt host
|
||||||
total_cpu_logical_agg_metric = [
|
total_cpu_logical_agg_metric = [
|
||||||
|
@ -164,11 +164,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-03-07 16:09:23',
|
self.assertEqual('2016-03-07 16:09:23',
|
||||||
total_cpu_logical_agg_metric.get(
|
total_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-03-07 16:10:38',
|
self.assertEqual('2016-03-07 16:10:38',
|
||||||
total_cpu_logical_agg_metric.get(
|
total_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify cpu.total_logical_cores_agg for test-cp1-comp0027-mgmt host
|
# Verify cpu.total_logical_cores_agg for test-cp1-comp0027-mgmt host
|
||||||
total_cpu_logical_agg_metric = [
|
total_cpu_logical_agg_metric = [
|
||||||
|
@ -203,11 +203,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-03-07 16:09:23',
|
self.assertEqual('2016-03-07 16:09:23',
|
||||||
total_cpu_logical_agg_metric.get(
|
total_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-03-07 16:10:38',
|
self.assertEqual('2016-03-07 16:10:38',
|
||||||
total_cpu_logical_agg_metric.get(
|
total_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify cpu.utilized_logical_cores_agg for all hosts
|
# Verify cpu.utilized_logical_cores_agg for all hosts
|
||||||
utilized_cpu_logical_agg_metric = [
|
utilized_cpu_logical_agg_metric = [
|
||||||
|
@ -243,11 +243,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-03-07 16:09:23',
|
self.assertEqual('2016-03-07 16:09:23',
|
||||||
utilized_cpu_logical_agg_metric.get(
|
utilized_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-03-07 16:10:38',
|
self.assertEqual('2016-03-07 16:10:38',
|
||||||
utilized_cpu_logical_agg_metric.get(
|
utilized_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify cpu.utilized_logical_cores_agg for the
|
# Verify cpu.utilized_logical_cores_agg for the
|
||||||
# test-cp1-comp0333-mgmt host
|
# test-cp1-comp0333-mgmt host
|
||||||
|
@ -284,11 +284,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-03-07 16:09:23',
|
self.assertEqual('2016-03-07 16:09:23',
|
||||||
utilized_cpu_logical_agg_metric.get(
|
utilized_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-03-07 16:10:38',
|
self.assertEqual('2016-03-07 16:10:38',
|
||||||
utilized_cpu_logical_agg_metric.get(
|
utilized_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
# Verify cpu.utilized_logical_cores_agg for the
|
# Verify cpu.utilized_logical_cores_agg for the
|
||||||
# test-cp1-comp0027-mgmt host
|
# test-cp1-comp0027-mgmt host
|
||||||
|
@ -325,11 +325,11 @@ class SparkTest(SparkContextTest):
|
||||||
self.assertEqual('2016-03-07 16:09:23',
|
self.assertEqual('2016-03-07 16:09:23',
|
||||||
utilized_cpu_logical_agg_metric.get(
|
utilized_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-03-07 16:10:38',
|
self.assertEqual('2016-03-07 16:10:38',
|
||||||
utilized_cpu_logical_agg_metric.get(
|
utilized_cpu_logical_agg_metric.get(
|
||||||
'metric').get('value_meta')
|
'metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
|
|
||||||
def simple_count_transform(rdd):
|
def simple_count_transform(rdd):
|
||||||
|
|
|
@ -178,11 +178,11 @@ class TestVmCpuAllocatedAgg(SparkContextTest):
|
||||||
self.assertEqual('2016-01-20 16:40:05',
|
self.assertEqual('2016-01-20 16:40:05',
|
||||||
vcpus_agg_metric
|
vcpus_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:46',
|
self.assertEqual('2016-01-20 16:40:46',
|
||||||
vcpus_agg_metric
|
vcpus_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
vcpus_agg_metric = [
|
vcpus_agg_metric = [
|
||||||
value for value in metrics
|
value for value in metrics
|
||||||
|
@ -217,11 +217,11 @@ class TestVmCpuAllocatedAgg(SparkContextTest):
|
||||||
self.assertEqual('2016-01-20 16:40:00',
|
self.assertEqual('2016-01-20 16:40:00',
|
||||||
vcpus_agg_metric
|
vcpus_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:42',
|
self.assertEqual('2016-01-20 16:40:42',
|
||||||
vcpus_agg_metric
|
vcpus_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
def get_pre_transform_specs_json_by_all(self):
|
def get_pre_transform_specs_json_by_all(self):
|
||||||
"""get pre_transform_specs driver table info."""
|
"""get pre_transform_specs driver table info."""
|
||||||
|
@ -353,11 +353,11 @@ class TestVmCpuAllocatedAgg(SparkContextTest):
|
||||||
self.assertEqual('2016-01-20 16:40:00',
|
self.assertEqual('2016-01-20 16:40:00',
|
||||||
vcpus_agg_metric
|
vcpus_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('firstrecord_timestamp'))
|
.get('firstrecord_timestamp_string'))
|
||||||
self.assertEqual('2016-01-20 16:40:46',
|
self.assertEqual('2016-01-20 16:40:46',
|
||||||
vcpus_agg_metric
|
vcpus_agg_metric
|
||||||
.get('metric').get('value_meta')
|
.get('metric').get('value_meta')
|
||||||
.get('lastrecord_timestamp'))
|
.get('lastrecord_timestamp_string'))
|
||||||
|
|
||||||
|
|
||||||
def simple_count_transform(rdd):
|
def simple_count_transform(rdd):
|
||||||
|
|
Loading…
Reference in New Issue