Merge "[Sahara] Split EDP context"

This commit is contained in:
Jenkins 2015-10-28 17:08:57 +00:00 committed by Gerrit Code Review
commit a298b40f96
13 changed files with 286 additions and 129 deletions

View File

@ -0,0 +1,89 @@
# Copyright 2015: Mirantis Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from rally.common.i18n import _
from rally.common import log as logging
from rally.common import utils as rutils
from rally import consts
from rally import exceptions
from rally import osclients
from rally.plugins.openstack.context.cleanup import manager as resource_manager
from rally.task import context
LOG = logging.getLogger(__name__)
@context.configure(name="sahara_data_sources", order=443)
class SaharaDataSources(context.Context):
"""Context class for setting up Data Sources for an EDP job."""
CONFIG_SCHEMA = {
"type": "object",
"$schema": consts.JSON_SCHEMA,
"properties": {
"input_type": {
"enum": ["swift", "hdfs"],
},
"input_url": {
"type": "string",
},
"output_type": {
"enum": ["swift", "hdfs"],
},
"output_url_prefix": {
"type": "string",
}
},
"additionalProperties": False,
"required": ["input_type", "input_url",
"output_type", "output_url_prefix"]
}
@rutils.log_task_wrapper(LOG.info,
_("Enter context: `Sahara Data Sources`"))
def setup(self):
self.context["sahara_output_conf"] = {
"output_type": self.config["output_type"],
"output_url_prefix": self.config["output_url_prefix"]
}
for user, tenant_id in rutils.iterate_per_tenants(
self.context["users"]):
clients = osclients.Clients(user["endpoint"])
sahara = clients.sahara()
self.setup_inputs(sahara, tenant_id, self.config["input_type"],
self.config["input_url"])
def setup_inputs(self, sahara, tenant_id, input_type, input_url):
if input_type == "swift":
raise exceptions.RallyException(
_("Swift Data Sources are not implemented yet"))
# Todo(nkonovalov): Add swift credentials parameters and data upload
input_ds = sahara.data_sources.create(
name=self.generate_random_name(),
description="",
data_source_type=input_type,
url=input_url)
self.context["tenants"][tenant_id]["sahara_input"] = input_ds.id
@rutils.log_task_wrapper(LOG.info, _("Exit context: `Sahara EDP`"))
def cleanup(self):
resources = ["job_executions", "jobs", "data_sources"]
# TODO(boris-42): Delete only resources created by this context
resource_manager.cleanup(
names=["sahara.%s" % res for res in resources],
users=self.context.get("users", []))

View File

@ -28,26 +28,14 @@ from rally.task import context
LOG = logging.getLogger(__name__)
@context.configure(name="sahara_edp", order=442)
class SaharaEDP(context.Context):
"""Context class for setting up the environment for an EDP job."""
@context.configure(name="sahara_job_binaries", order=442)
class SaharaJobBinaries(context.Context):
"""Context class for setting up Job Binaries for an EDP job."""
CONFIG_SCHEMA = {
"type": "object",
"$schema": consts.JSON_SCHEMA,
"properties": {
"input_type": {
"enum": ["swift", "hdfs"],
},
"input_url": {
"type": "string",
},
"output_type": {
"enum": ["swift", "hdfs"],
},
"output_url_prefix": {
"type": "string",
},
"mains": {
"type": "array",
"items": {
@ -81,24 +69,16 @@ class SaharaEDP(context.Context):
}
}
},
"additionalProperties": False,
"required": ["input_type", "input_url",
"output_type", "output_url_prefix"]
"additionalProperties": False
}
@rutils.log_task_wrapper(LOG.info, _("Enter context: `Sahara EDP`"))
def setup(self):
self.context["sahara_output_conf"] = {
"output_type": self.config["output_type"],
"output_url_prefix": self.config["output_url_prefix"]
}
self.context["sahara_mains"] = {}
self.context["sahara_libs"] = {}
# This cache will hold the downloaded libs content to prevent repeated
# downloads for each tenant
lib_cache = {}
input_type = self.config["input_type"]
input_url = self.config["input_url"]
mains = self.config.get("mains", [])
libs = self.config.get("libs", [])
@rutils.log_task_wrapper(LOG.info,
_("Enter context: `Sahara Job Binaries`"))
def setup(self):
for user, tenant_id in rutils.iterate_per_tenants(
self.context["users"]):
@ -106,12 +86,10 @@ class SaharaEDP(context.Context):
clients = osclients.Clients(user["endpoint"])
sahara = clients.sahara()
self.setup_inputs(sahara, tenant_id, input_type, input_url)
self.context["tenants"][tenant_id]["sahara_mains"] = []
self.context["tenants"][tenant_id]["sahara_libs"] = []
for main in mains:
for main in self.config.get("mains", []):
self.download_and_save_lib(
sahara=sahara,
lib_type="sahara_mains",
@ -119,7 +97,7 @@ class SaharaEDP(context.Context):
download_url=main["download_url"],
tenant_id=tenant_id)
for lib in libs:
for lib in self.config.get("libs", []):
self.download_and_save_lib(
sahara=sahara,
lib_type="sahara_libs",
@ -142,7 +120,11 @@ class SaharaEDP(context.Context):
def download_and_save_lib(self, sahara, lib_type, name, download_url,
tenant_id):
lib_data = requests.get(download_url).content
if download_url not in self.lib_cache:
lib_data = requests.get(download_url).content
self.lib_cache[download_url] = lib_data
else:
lib_data = self.lib_cache[download_url]
job_binary_internal = sahara.job_binary_internals.create(
name=name,
@ -156,10 +138,10 @@ class SaharaEDP(context.Context):
self.context["tenants"][tenant_id][lib_type].append(job_binary.id)
@rutils.log_task_wrapper(LOG.info, _("Exit context: `Sahara EDP`"))
@rutils.log_task_wrapper(LOG.info,
_("Exit context: `Sahara Job Binaries`"))
def cleanup(self):
resources = ["job_executions", "jobs", "job_binary_internals",
"job_binaries", "data_sources"]
resources = ["job_binary_internals", "job_binaries"]
# TODO(boris-42): Delete only resources created by this context
resource_manager.cleanup(

View File

@ -26,8 +26,8 @@ class SaharaJob(utils.SaharaScenario):
"""Benchmark scenarios for Sahara jobs."""
@validation.required_services(consts.Service.SAHARA)
@validation.required_contexts("users", "sahara_image", "sahara_edp",
"sahara_cluster")
@validation.required_contexts("users", "sahara_image",
"sahara_job_binaries", "sahara_cluster")
@scenario.configure(context={"cleanup": ["sahara"]})
def create_launch_job(self, job_type, configs, job_idx=0):
"""Create and execute a Sahara EDP Job.
@ -69,8 +69,8 @@ class SaharaJob(utils.SaharaScenario):
job_idx=job_idx)
@validation.required_services(consts.Service.SAHARA)
@validation.required_contexts("users", "sahara_image", "sahara_edp",
"sahara_cluster")
@validation.required_contexts("users", "sahara_image",
"sahara_job_binaries", "sahara_cluster")
@scenario.configure(context={"cleanup": ["sahara"]})
def create_launch_job_sequence(self, jobs):
"""Create and execute a sequence of the Sahara EDP Jobs.
@ -86,8 +86,8 @@ class SaharaJob(utils.SaharaScenario):
self.create_launch_job(job["job_type"], job["configs"], idx)
@validation.required_services(consts.Service.SAHARA)
@validation.required_contexts("users", "sahara_image", "sahara_edp",
"sahara_cluster")
@validation.required_contexts("users", "sahara_image",
"sahara_job_binaries", "sahara_cluster")
@scenario.configure(context={"cleanup": ["sahara"]})
def create_launch_job_sequence_with_scaling(self, jobs, deltas):
"""Create and execute Sahara EDP Jobs on a scaling Cluster.

View File

@ -38,11 +38,7 @@
"plugin_name": "vanilla",
"hadoop_version": "1.2.1"
},
"sahara_edp": {
"input_type": "hdfs",
"output_type": "hdfs",
"input_url": "/",
"output_url_prefix": "/out_",
"sahara_job_binaries": {
"libs": [
{
"name": "tests.jar",

View File

@ -38,11 +38,7 @@
username: "ubuntu"
plugin_name: "vanilla"
hadoop_version: "1.2.1"
sahara_edp:
input_type: "hdfs"
output_type: "hdfs"
input_url: "/"
output_url_prefix: "/out_"
sahara_job_binaries:
libs:
-
name: "tests.jar"

View File

@ -38,11 +38,7 @@
"plugin_name": "vanilla",
"hadoop_version": "1.2.1"
},
"sahara_edp": {
"input_type": "hdfs",
"output_type": "hdfs",
"input_url": "/",
"output_url_prefix": "/out_",
"sahara_job_binaries": {
"libs": [
{
"name": "tests.jar",

View File

@ -41,11 +41,7 @@
username: "ubuntu"
plugin_name: "vanilla"
hadoop_version: "1.2.1"
sahara_edp:
input_type: "hdfs"
output_type: "hdfs"
input_url: "/"
output_url_prefix: "/out_"
sahara_job_binaries:
libs:
-
name: "tests.jar"

View File

@ -26,11 +26,7 @@
"plugin_name": "vanilla",
"hadoop_version": "1.2.1"
},
"sahara_edp": {
"input_type": "hdfs",
"output_type": "hdfs",
"input_url": "/",
"output_url_prefix": "/out_",
"sahara_job_binaries": {
"libs": [
{
"name": "examples.jar",

View File

@ -22,11 +22,7 @@
username: "ubuntu"
plugin_name: "vanilla"
hadoop_version: "1.2.1"
sahara_edp:
input_type: "hdfs"
output_type: "hdfs"
input_url: "/"
output_url_prefix: "/out_"
sahara_job_binaries:
libs:
-
name: "examples.jar"

View File

@ -21,11 +21,7 @@
"plugin_name": "vanilla",
"hadoop_version": "1.2.1"
},
"sahara_edp": {
"input_type": "hdfs",
"output_type": "hdfs",
"input_url": "/",
"output_url_prefix": "/out_",
"sahara_job_binaries": {
"mains": [
{
"name": "example.pig",
@ -39,6 +35,12 @@
}
]
},
"sahara_data_sources": {
"input_type": "hdfs",
"output_type": "hdfs",
"input_url": "/",
"output_url_prefix": "/out_"
},
"sahara_cluster": {
"flavor_id": "2",
"workers_count": 3,

View File

@ -17,11 +17,7 @@
username: "ubuntu"
plugin_name: "vanilla"
hadoop_version: "1.2.1"
sahara_edp:
input_type: "hdfs"
output_type: "hdfs"
input_url: "/"
output_url_prefix: "/out_"
sahara_job_binaries:
mains:
-
name: "example.pig"
@ -30,6 +26,11 @@
-
name: "udf.jar"
download_url: "https://github.com/openstack/sahara/blob/master/etc/edp-examples/pig-job/udf.jar?raw=true"
sahara_data_sources:
input_type: "hdfs"
output_type: "hdfs"
input_url: "/"
output_url_prefix: "/out_"
sahara_cluster:
flavor_id: "2"
workers_count: 3

View File

@ -14,16 +14,17 @@
import mock
from rally.plugins.openstack.context.sahara import sahara_edp
from rally.plugins.openstack.context.sahara import sahara_data_sources
from tests.unit import test
CTX = "rally.plugins.openstack.context.sahara"
class SaharaEDPTestCase(test.TestCase):
class SaharaDataSourcesTestCase(test.ScenarioTestCase):
def setUp(self):
super(SaharaEDPTestCase, self).setUp()
super(SaharaDataSourcesTestCase, self).setUp()
self.tenants_num = 2
self.users_per_tenant = 2
self.users = self.tenants_num * self.users_per_tenant
@ -44,55 +45,35 @@ class SaharaEDPTestCase(test.TestCase):
for j in range(self.tenants_num)
for i in range(self.users_per_tenant)]
@property
def context_without_edp_keys(self):
context = test.get_test_context()
context.update({
self.context.update({
"config": {
"users": {
"tenants": self.tenants_num,
"users_per_tenant": self.users_per_tenant,
},
"sahara_edp": {
"sahara_data_sources": {
"input_type": "hdfs",
"output_type": "hdfs",
"input_url": "hdfs://test_host/",
"output_url_prefix": "hdfs://test_host/out_",
"libs": [
{
"name": "test.jar",
"download_url": "http://example.com/test.jar"
}
]
"output_url_prefix": "hdfs://test_host/out_"
},
},
"admin": {"endpoint": mock.MagicMock()},
"users": self.users_key,
"tenants": self.tenants
})
return context
@mock.patch("%s.sahara_edp.resource_manager.cleanup" % CTX)
@mock.patch("%s.sahara_edp.requests" % CTX)
@mock.patch("%s.sahara_edp.osclients" % CTX)
def test_setup_and_cleanup(self, mock_osclients, mock_requests,
mock_cleanup):
@mock.patch("%s.sahara_data_sources.resource_manager.cleanup" % CTX)
@mock.patch("%s.sahara_data_sources.osclients" % CTX)
def test_setup_and_cleanup(self, mock_osclients, mock_cleanup):
mock_sahara = mock_osclients.Clients(mock.MagicMock()).sahara()
mock_sahara.data_sources.create.return_value = mock.MagicMock(id=42)
mock_sahara.job_binary_internals.create.return_value = (
mock.MagicMock(id=42))
mock_requests.get().content = "test_binary"
ctx = self.context_without_edp_keys
sahara_ctx = sahara_edp.SaharaEDP(ctx)
sahara_ctx = sahara_data_sources.SaharaDataSources(self.context)
sahara_ctx.generate_random_name = mock.Mock()
input_ds_crete_calls = []
download_calls = []
job_binary_internals_calls = []
job_binaries_calls = []
for i in range(self.tenants_num):
input_ds_crete_calls.append(mock.call(
@ -100,28 +81,14 @@ class SaharaEDPTestCase(test.TestCase):
description="",
data_source_type="hdfs",
url="hdfs://test_host/"))
download_calls.append(mock.call("http://example.com/test.jar"))
job_binary_internals_calls.append(mock.call(
name="test.jar",
data="test_binary"))
job_binaries_calls.append(mock.call(
name="test.jar",
url="internal-db://42",
description="",
extra={}))
sahara_ctx.setup()
mock_sahara.data_sources.create.assert_has_calls(input_ds_crete_calls)
mock_requests.get.assert_has_calls(download_calls)
mock_sahara.job_binary_internals.create.assert_has_calls(
job_binary_internals_calls)
mock_sahara.job_binaries.create.assert_has_calls(job_binaries_calls)
sahara_ctx.cleanup()
mock_cleanup.assert_called_once_with(
names=["sahara.job_executions", "sahara.jobs",
"sahara.job_binary_internals", "sahara.job_binaries",
"sahara.data_sources"],
users=ctx["users"])
users=self.context["users"])

View File

@ -0,0 +1,140 @@
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import mock
from rally.plugins.openstack.context.sahara import sahara_job_binaries
from tests.unit import test
CTX = "rally.plugins.openstack.context.sahara"
class SaharaJobBinariesTestCase(test.ScenarioTestCase):
def setUp(self):
super(SaharaJobBinariesTestCase, self).setUp()
self.tenants_num = 2
self.users_per_tenant = 2
self.users = self.tenants_num * self.users_per_tenant
self.task = mock.MagicMock()
self.tenants = {}
self.users_key = []
for i in range(self.tenants_num):
self.tenants[str(i)] = {"id": str(i), "name": str(i),
"sahara_image": "42"}
for j in range(self.users_per_tenant):
self.users_key.append({"id": "%s_%s" % (str(i), str(j)),
"tenant_id": str(i),
"endpoint": "endpoint"})
self.user_key = [{"id": i, "tenant_id": j, "endpoint": "endpoint"}
for j in range(self.tenants_num)
for i in range(self.users_per_tenant)]
self.context.update({
"config": {
"users": {
"tenants": self.tenants_num,
"users_per_tenant": self.users_per_tenant,
},
"sahara_job_binaries": {
"libs": [
{
"name": "test.jar",
"download_url": "http://example.com/test.jar"
}
],
"mains": [
{
"name": "test.jar",
"download_url": "http://example.com/test.jar"
}
]
},
},
"admin": {"endpoint": mock.MagicMock()},
"task": mock.MagicMock(),
"users": self.users_key,
"tenants": self.tenants
})
@mock.patch("%s.sahara_job_binaries.resource_manager.cleanup" % CTX)
@mock.patch(("%s.sahara_job_binaries.SaharaJobBinaries."
"download_and_save_lib") % CTX)
@mock.patch("%s.sahara_job_binaries.osclients" % CTX)
def test_setup_and_cleanup(
self,
mock_osclients,
mock_sahara_job_binaries_download_and_save_lib,
mock_cleanup):
mock_sahara = mock_osclients.Clients(mock.MagicMock()).sahara()
sahara_ctx = sahara_job_binaries.SaharaJobBinaries(self.context)
download_calls = []
for i in range(self.tenants_num):
download_calls.append(mock.call(
sahara=mock_sahara,
lib_type="sahara_mains",
name="test.jar",
download_url="http://example.com/test.jar",
tenant_id=str(i)))
download_calls.append(mock.call(
sahara=mock_sahara,
lib_type="sahara_libs",
name="test.jar",
download_url="http://example.com/test.jar",
tenant_id=str(i)))
sahara_ctx.setup()
(mock_sahara_job_binaries_download_and_save_lib.
assert_has_calls(download_calls))
sahara_ctx.cleanup()
mock_cleanup.assert_called_once_with(
names=["sahara.job_binary_internals", "sahara.job_binaries"],
users=self.context["users"])
@mock.patch("%s.sahara_job_binaries.requests" % CTX)
@mock.patch("%s.sahara_job_binaries.osclients" % CTX)
def test_download_and_save_lib(self, mock_osclients, mock_requests):
mock_requests.get.content.return_value = "some_binary_content"
mock_sahara = mock_osclients.Clients(mock.MagicMock()).sahara()
mock_sahara.job_binary_internals.create.return_value = (
mock.MagicMock(id=42))
sahara_ctx = sahara_job_binaries.SaharaJobBinaries(self.context)
sahara_ctx.context["tenants"]["0"]["sahara_mains"] = []
sahara_ctx.context["tenants"]["0"]["sahara_libs"] = []
sahara_ctx.download_and_save_lib(sahara=mock_sahara,
lib_type="sahara_mains",
name="test_binary",
download_url="http://somewhere",
tenant_id="0")
sahara_ctx.download_and_save_lib(sahara=mock_sahara,
lib_type="sahara_libs",
name="test_binary_2",
download_url="http://somewhere",
tenant_id="0")
mock_requests.get.assert_called_once_with("http://somewhere")