Plugins splitted from sahara core

Change-Id: I43e0beec6508f93a436a150749bfa23571986b9d
This commit is contained in:
Telles Nobrega 2019-01-02 12:34:13 -03:00
parent 31aa7c01f2
commit 40fd2aa34e
160 changed files with 581 additions and 508 deletions

View File

@ -1,3 +1,3 @@
[DEFAULT] [DEFAULT]
test_path=./sahara/tests/unit test_path=./sahara_plugin_mapr/tests/unit
top_dir=./ top_dir=./

35
README.rst Normal file
View File

@ -0,0 +1,35 @@
========================
Team and repository tags
========================
.. image:: https://governance.openstack.org/tc/badges/sahara.svg
:target: https://governance.openstack.org/tc/reference/tags/index.html
.. Change things from this point on
OpenStack Data Processing ("Sahara") project
============================================
Sahara at wiki.openstack.org: https://wiki.openstack.org/wiki/Sahara
Storyboard project: https://storyboard.openstack.org/#!/project/935
Sahara docs site: https://docs.openstack.org/sahara/latest/
Roadmap: https://wiki.openstack.org/wiki/Sahara/Roadmap
Quickstart guide: https://docs.openstack.org/sahara/latest/user/quickstart.html
How to participate: https://docs.openstack.org/sahara/latest/contributor/how-to-participate.html
Source: https://git.openstack.org/cgit/openstack/sahara
Bugs and feature requests: https://storyboard.openstack.org/#!/project/935
Release notes: https://docs.openstack.org/releasenotes/sahara/
License
-------
Apache License Version 2.0 http://www.apache.org/licenses/LICENSE-2.0

9
doc/requirements.txt Normal file
View File

@ -0,0 +1,9 @@
# The order of packages is significant, because pip processes them in the order
# of appearance. Changing the order has an impact on the overall integration
# process, which may cause wedges in the gate later.
openstackdocstheme>=1.18.1 # Apache-2.0
os-api-ref>=1.4.0 # Apache-2.0
reno>=2.5.0 # Apache-2.0
sphinx!=1.6.6,!=1.6.7,>=1.6.2 # BSD
sphinxcontrib-httpdomain>=1.3.0 # BSD
whereto>=0.3.0 # Apache-2.0

162
lower-constraints.txt Normal file
View File

@ -0,0 +1,162 @@
alabaster==0.7.10
alembic==0.8.10
amqp==2.2.2
appdirs==1.4.3
asn1crypto==0.24.0
astroid==1.3.8
Babel==2.3.4
bandit==1.1.0
bashate==0.5.1
bcrypt==3.1.4
botocore==1.5.1
cachetools==2.0.1
castellan==0.16.0
certifi==2018.1.18
cffi==1.11.5
chardet==3.0.4
click==6.7
cliff==2.11.0
cmd2==0.8.1
contextlib2==0.5.5
coverage==4.0
cryptography==2.1.4
debtcollector==1.19.0
decorator==4.2.1
deprecation==2.0
doc8==0.6.0
docutils==0.14
dogpile.cache==0.6.5
dulwich==0.19.0
enum-compat==0.0.2
eventlet==0.18.2
extras==1.0.0
fasteners==0.14.1
fixtures==3.0.0
flake8==2.6.2
Flask==1.0.2
future==0.16.0
futurist==1.6.0
gitdb2==2.0.3
GitPython==2.1.8
greenlet==0.4.13
hacking==1.1.0
idna==2.6
imagesize==1.0.0
iso8601==0.1.11
itsdangerous==0.24
Jinja2==2.10
jmespath==0.9.3
jsonpatch==1.21
jsonpointer==2.0
jsonschema==2.6.0
keystoneauth1==3.4.0
keystonemiddleware==4.17.0
kombu==4.1.0
linecache2==1.0.0
logilab-common==1.4.1
Mako==1.0.7
MarkupSafe==1.0
mccabe==0.2.1
mock==2.0.0
monotonic==1.4
mox3==0.25.0
msgpack==0.5.6
munch==2.2.0
netaddr==0.7.19
netifaces==0.10.6
openstackdocstheme==1.18.1
openstacksdk==0.12.0
os-api-ref==1.4.0
os-client-config==1.29.0
os-service-types==1.2.0
osc-lib==1.10.0
oslo.cache==1.29.0
oslo.concurrency==3.26.0
oslo.config==5.2.0
oslo.context==2.19.2
oslo.db==4.27.0
oslo.i18n==3.15.3
oslo.log==3.36.0
oslo.messaging==5.29.0
oslo.middleware==3.31.0
oslo.policy==1.30.0
oslo.rootwrap==5.8.0
oslo.serialization==2.18.0
oslo.service==1.24.0
oslo.upgradecheck==0.1.0
oslo.utils==3.33.0
oslotest==3.2.0
packaging==17.1
paramiko==2.0.0
Paste==2.0.3
PasteDeploy==1.5.2
pbr==2.0.0
pika-pool==0.1.3
pika==0.10.0
prettytable==0.7.2
psycopg2==2.6.2
pyasn1==0.4.2
pycadf==2.7.0
pycparser==2.18
pycodestyle==2.4.0
pyflakes==0.8.1
Pygments==2.2.0
pyinotify==0.9.6
pylint==1.4.5
PyMySQL==0.7.6
PyNaCl==1.2.1
pyOpenSSL==17.5.0
pyparsing==2.2.0
pyperclip==1.6.0
python-barbicanclient==4.6.0
python-cinderclient==3.3.0
python-dateutil==2.7.0
python-editor==1.0.3
python-glanceclient==2.8.0
python-heatclient==1.10.0
python-keystoneclient==3.8.0
python-manilaclient==1.16.0
python-mimeparse==1.6.0
python-neutronclient==6.7.0
python-novaclient==9.1.0
python-openstackclient==3.14.0
python-saharaclient==1.4.0
python-subunit==1.2.0
python-swiftclient==3.2.0
pytz==2018.3
PyYAML==3.12
reno==2.5.0
repoze.lru==0.7
requests==2.14.2
requestsexceptions==1.4.0
restructuredtext-lint==1.1.3
rfc3986==1.1.0
Routes==2.4.1
simplejson==3.13.2
six==1.10.0
smmap2==2.0.3
snowballstemmer==1.2.1
Sphinx==1.6.2
sphinxcontrib-httpdomain==1.3.0
sphinxcontrib-websupport==1.0.1
sqlalchemy-migrate==0.11.0
SQLAlchemy==1.0.10
sqlparse==0.2.4
statsd==3.2.2
stestr==1.0.0
stevedore==1.20.0
Tempita==0.5.2
tenacity==4.9.0
testresources==2.0.0
testscenarios==0.4
testtools==2.2.0
tooz==1.58.0
traceback2==1.4.0
unittest2==1.1.0
urllib3==1.22
vine==1.1.4
voluptuous==0.11.1
warlock==1.3.0
WebOb==1.7.1
Werkzeug==0.14.1
wrapt==1.10.11

View File

@ -39,6 +39,7 @@ python-swiftclient>=3.2.0 # Apache-2.0
python-neutronclient>=6.7.0 # Apache-2.0 python-neutronclient>=6.7.0 # Apache-2.0
python-heatclient>=1.10.0 # Apache-2.0 python-heatclient>=1.10.0 # Apache-2.0
python-glanceclient>=2.8.0 # Apache-2.0 python-glanceclient>=2.8.0 # Apache-2.0
sahara
six>=1.10.0 # MIT six>=1.10.0 # MIT
stevedore>=1.20.0 # Apache-2.0 stevedore>=1.20.0 # Apache-2.0
SQLAlchemy!=1.1.5,!=1.1.6,!=1.1.7,!=1.1.8,>=1.0.10 # MIT SQLAlchemy!=1.1.5,!=1.1.6,!=1.1.7,!=1.1.8,>=1.0.10 # MIT

View File

@ -18,24 +18,23 @@ import abc
from oslo_log import log as logging from oslo_log import log as logging
import six import six
from sahara import conductor from sahara.plugins import conductor
from sahara import context from sahara.plugins import context
from sahara.i18n import _ import sahara.plugins.utils as utils
import sahara.plugins.mapr.abstract.configurer as ac from sahara_plugin_mapr.i18n import _
from sahara.plugins.mapr.domain import distro as d import sahara_plugin_mapr.plugins.mapr.abstract.configurer as ac
from sahara.plugins.mapr.domain import service as srvc from sahara_plugin_mapr.plugins.mapr.domain import distro as d
import sahara.plugins.mapr.services.management.management as mng from sahara_plugin_mapr.plugins.mapr.domain import service as srvc
import sahara.plugins.mapr.services.mapreduce.mapreduce as mr import sahara_plugin_mapr.plugins.mapr.services.management.management as mng
from sahara.plugins.mapr.services.maprfs import maprfs import sahara_plugin_mapr.plugins.mapr.services.mapreduce.mapreduce as mr
from sahara.plugins.mapr.services.mysql import mysql from sahara_plugin_mapr.plugins.mapr.services.maprfs import maprfs
import sahara.plugins.mapr.services.yarn.yarn as yarn from sahara_plugin_mapr.plugins.mapr.services.mysql import mysql
from sahara.plugins.mapr.util import event_log as el import sahara_plugin_mapr.plugins.mapr.services.yarn.yarn as yarn
import sahara.plugins.mapr.util.general as util from sahara_plugin_mapr.plugins.mapr.util import event_log as el
import sahara.plugins.mapr.util.password_utils as pu import sahara_plugin_mapr.plugins.mapr.util.general as util
import sahara.utils.files as files import sahara_plugin_mapr.plugins.mapr.util.password_utils as pu
LOG = logging.getLogger(__name__) LOG = logging.getLogger(__name__)
conductor = conductor.API
_JAVA_HOME = '/usr/java/jdk1.7.0_51' _JAVA_HOME = '/usr/java/jdk1.7.0_51'
_CONFIGURE_SH_TIMEOUT = 600 _CONFIGURE_SH_TIMEOUT = 600
@ -173,7 +172,7 @@ class BaseConfigurer(ac.AbstractConfigurer):
topology_map = "\n".join(topology_map) + "\n" topology_map = "\n".join(topology_map) + "\n"
data_path = "%s/topology.data" % cluster_context.mapr_home data_path = "%s/topology.data" % cluster_context.mapr_home
script = files.get_file_text(_TOPO_SCRIPT) script = utils.get_file_text(_TOPO_SCRIPT, 'sahara_plugin_mapr')
script_path = '%s/topology.sh' % cluster_context.mapr_home script_path = '%s/topology.sh' % cluster_context.mapr_home
@el.provision_event() @el.provision_event()
@ -310,7 +309,7 @@ class BaseConfigurer(ac.AbstractConfigurer):
'port': mysql.MySQL.MYSQL_SERVER_PORT, 'port': mysql.MySQL.MYSQL_SERVER_PORT,
}) })
with context.ThreadGroup() as tg: with context.PluginsThreadGroup() as tg:
for instance in instances: for instance in instances:
tg.spawn('configure-sh-%s' % instance.id, tg.spawn('configure-sh-%s' % instance.id,
self._configure_sh_instance, cluster_context, self._configure_sh_instance, cluster_context,

View File

@ -17,21 +17,20 @@ import collections
from oslo_config import cfg from oslo_config import cfg
import sahara.exceptions as e import sahara.plugins.exceptions as e
from sahara.i18n import _ from sahara.plugins import topology_helper as th
import sahara.plugins.mapr.abstract.cluster_context as cc
import sahara.plugins.mapr.domain.configuration_file as bcf
import sahara.plugins.mapr.domain.distro as distro
import sahara.plugins.mapr.services.management.management as mng
import sahara.plugins.mapr.services.maprfs.maprfs as mfs
import sahara.plugins.mapr.services.oozie.oozie as oozie
from sahara.plugins.mapr.services.swift import swift
import sahara.plugins.mapr.services.yarn.yarn as yarn
import sahara.plugins.mapr.util.general as g
import sahara.plugins.mapr.util.service_utils as su
import sahara.plugins.utils as u import sahara.plugins.utils as u
from sahara.topology import topology_helper as th from sahara_plugin_mapr.i18n import _
import sahara.utils.configs as sahara_configs import sahara_plugin_mapr.plugins.mapr.abstract.cluster_context as cc
import sahara_plugin_mapr.plugins.mapr.domain.configuration_file as bcf
import sahara_plugin_mapr.plugins.mapr.domain.distro as distro
import sahara_plugin_mapr.plugins.mapr.services.management.management as mng
import sahara_plugin_mapr.plugins.mapr.services.maprfs.maprfs as mfs
import sahara_plugin_mapr.plugins.mapr.services.oozie.oozie as oozie
from sahara_plugin_mapr.plugins.mapr.services.swift import swift
import sahara_plugin_mapr.plugins.mapr.services.yarn.yarn as yarn
import sahara_plugin_mapr.plugins.mapr.util.general as g
import sahara_plugin_mapr.plugins.mapr.util.service_utils as su
CONF = cfg.CONF CONF = cfg.CONF
CONF.import_opt("enable_data_locality", "sahara.topology.topology_helper") CONF.import_opt("enable_data_locality", "sahara.topology.topology_helper")
@ -275,12 +274,12 @@ class BaseClusterContext(cc.AbstractClusterContext):
def get_service(self, node_process): def get_service(self, node_process):
ui_name = self.get_service_name_by_node_process(node_process) ui_name = self.get_service_name_by_node_process(node_process)
if ui_name is None: if ui_name is None:
raise e.InvalidDataException( raise e.PluginInvalidDataException(
_('Service not found in services list')) _('Service not found in services list'))
version = self.get_chosen_service_version(ui_name) version = self.get_chosen_service_version(ui_name)
service = self._find_service_instance(ui_name, version) service = self._find_service_instance(ui_name, version)
if service is None: if service is None:
raise e.InvalidDataException(_('Can not map service')) raise e.PluginInvalidDataException(_('Can not map service'))
return service return service
def _find_service_instance(self, ui_name, version): def _find_service_instance(self, ui_name, version):
@ -417,7 +416,7 @@ class BaseClusterContext(cc.AbstractClusterContext):
services = self.get_cluster_services(node_group) services = self.get_cluster_services(node_group)
user_configs = node_group.configuration() user_configs = node_group.configuration()
default_configs = self.get_services_configs_dict(services) default_configs = self.get_services_configs_dict(services)
return sahara_configs.merge_configs(default_configs, user_configs) return u.merge_configs(default_configs, user_configs)
def get_config_files(self, node_group): def get_config_files(self, node_group):
services = self.get_cluster_services(node_group) services = self.get_cluster_services(node_group)

View File

@ -13,9 +13,9 @@
# under the License. # under the License.
import sahara.plugins.mapr.abstract.cluster_validator as v import sahara_plugin_mapr.plugins.mapr.abstract.cluster_validator as v
import sahara.plugins.mapr.util.validation_utils as vu import sahara_plugin_mapr.plugins.mapr.util.validation_utils as vu
import sahara.plugins.mapr.versions.version_handler_factory as vhf import sahara_plugin_mapr.plugins.mapr.versions.version_handler_factory as vhf
class BaseValidator(v.AbstractValidator): class BaseValidator(v.AbstractValidator):

View File

@ -15,15 +15,14 @@
import os import os
from sahara import context from sahara.plugins import context
import sahara.plugins.mapr.util.maprfs_helper as mfs from sahara.plugins import edp
import sahara.plugins.mapr.versions.version_handler_factory as vhf
from sahara.service.edp.job_binaries import manager as jb_manager from sahara.service.edp.job_binaries import manager as jb_manager
import sahara.service.edp.oozie.engine as e import sahara_plugin_mapr.plugins.mapr.util.maprfs_helper as mfs
from sahara.utils import edp import sahara_plugin_mapr.plugins.mapr.versions.version_handler_factory as vhf
class MapROozieJobEngine(e.OozieJobEngine): class MapROozieJobEngine(edp.PluginsOozieJobEngine):
def __init__(self, cluster): def __init__(self, cluster):
super(MapROozieJobEngine, self).__init__(cluster) super(MapROozieJobEngine, self).__init__(cluster)
self.cluster_context = self._get_cluster_context(self.cluster) self.cluster_context = self._get_cluster_context(self.cluster)

View File

@ -15,12 +15,12 @@
import functools import functools
from sahara.i18n import _ from sahara.plugins import health_check_base
import sahara.plugins.mapr.abstract.health_checker as hc from sahara_plugin_mapr.i18n import _
from sahara.plugins.mapr.domain import node_process as np import sahara_plugin_mapr.plugins.mapr.abstract.health_checker as hc
from sahara.plugins.mapr.services.management import management from sahara_plugin_mapr.plugins.mapr.domain import node_process as np
from sahara.plugins.mapr.services.spark import spark from sahara_plugin_mapr.plugins.mapr.services.management import management
from sahara.service.health import health_check_base from sahara_plugin_mapr.plugins.mapr.services.spark import spark
class BaseHealthChecker(hc.AbstractHealthChecker): class BaseHealthChecker(hc.AbstractHealthChecker):

View File

@ -19,14 +19,14 @@ from oslo_log import log as logging
from oslo_serialization import jsonutils as json from oslo_serialization import jsonutils as json
from oslo_utils import timeutils from oslo_utils import timeutils
from sahara import context from sahara.plugins import context
from sahara.i18n import _
import sahara.plugins.exceptions as ex import sahara.plugins.exceptions as ex
import sahara.plugins.mapr.abstract.node_manager as s from sahara.plugins import utils
import sahara.plugins.mapr.services.management.management as mng from sahara_plugin_mapr.i18n import _
import sahara.plugins.mapr.services.maprfs.maprfs as mfs import sahara_plugin_mapr.plugins.mapr.abstract.node_manager as s
import sahara.plugins.mapr.util.event_log as el import sahara_plugin_mapr.plugins.mapr.services.management.management as mng
from sahara.utils import cluster_progress_ops as cpo import sahara_plugin_mapr.plugins.mapr.services.maprfs.maprfs as mfs
import sahara_plugin_mapr.plugins.mapr.util.event_log as el
LOG = logging.getLogger(__name__) LOG = logging.getLogger(__name__)
@ -80,14 +80,15 @@ class BaseNodeManager(s.AbstractNodeManager):
others = filter( others = filter(
lambda i: not cluster_context.check_for_process(i, mfs.CLDB), lambda i: not cluster_context.check_for_process(i, mfs.CLDB),
instances) instances)
cpo.add_provisioning_step(cluster_context.cluster.id, utils.add_provisioning_step(cluster_context.cluster.id,
_("Start ZooKeepers nodes"), len(zookeepers)) _("Start ZooKeepers nodes"),
len(zookeepers))
self._start_zk_nodes(zookeepers) self._start_zk_nodes(zookeepers)
cpo.add_provisioning_step(cluster_context.cluster.id, utils.add_provisioning_step(cluster_context.cluster.id,
_("Start CLDB nodes"), len(cldbs)) _("Start CLDB nodes"), len(cldbs))
self._start_cldb_nodes(cldbs) self._start_cldb_nodes(cldbs)
if others: if others:
cpo.add_provisioning_step(cluster_context.cluster.id, utils.add_provisioning_step(cluster_context.cluster.id,
_("Start non-CLDB nodes"), _("Start non-CLDB nodes"),
len(list(others))) len(list(others)))
self._start_non_cldb_nodes(others) self._start_non_cldb_nodes(others)
@ -96,10 +97,11 @@ class BaseNodeManager(s.AbstractNodeManager):
def stop(self, cluster_context, instances=None): def stop(self, cluster_context, instances=None):
instances = instances or cluster_context.get_instances() instances = instances or cluster_context.get_instances()
zookeepers = cluster_context.filter_instances(instances, mng.ZOOKEEPER) zookeepers = cluster_context.filter_instances(instances, mng.ZOOKEEPER)
cpo.add_provisioning_step(cluster_context.cluster.id, utils.add_provisioning_step(cluster_context.cluster.id,
_("Stop ZooKeepers nodes"), len(zookeepers)) _("Stop ZooKeepers nodes"),
len(zookeepers))
self._stop_zk_nodes(zookeepers) self._stop_zk_nodes(zookeepers)
cpo.add_provisioning_step(cluster_context.cluster.id, utils.add_provisioning_step(cluster_context.cluster.id,
_("Stop Warden nodes"), len(instances)) _("Stop Warden nodes"), len(instances))
self._stop_warden_on_nodes(instances) self._stop_warden_on_nodes(instances)
@ -133,13 +135,13 @@ class BaseNodeManager(s.AbstractNodeManager):
raise ex.HadoopProvisionError(_("CLDB failed to start")) raise ex.HadoopProvisionError(_("CLDB failed to start"))
def _start_nodes(self, instances, sys_service): def _start_nodes(self, instances, sys_service):
with context.ThreadGroup() as tg: with context.PluginsThreadGroup() as tg:
for instance in instances: for instance in instances:
tg.spawn('start-%s-%s' % (sys_service, instance.id), tg.spawn('start-%s-%s' % (sys_service, instance.id),
self._start_service, instance, sys_service) self._start_service, instance, sys_service)
def _stop_nodes(self, instances, sys_service): def _stop_nodes(self, instances, sys_service):
with context.ThreadGroup() as tg: with context.PluginsThreadGroup() as tg:
for instance in instances: for instance in instances:
tg.spawn('stop-%s-%s' % (sys_service, instance.id), tg.spawn('stop-%s-%s' % (sys_service, instance.id),
self._stop_service, instance, sys_service) self._stop_service, instance, sys_service)

View File

@ -15,23 +15,23 @@
import collections as c import collections as c
from sahara.i18n import _
import sahara.plugins.mapr.abstract.version_handler as vh
import sahara.plugins.mapr.base.base_cluster_configurer as base_conf
import sahara.plugins.mapr.base.base_cluster_validator as bv
import sahara.plugins.mapr.base.base_edp_engine as edp
import sahara.plugins.mapr.base.base_health_checker as health
import sahara.plugins.mapr.base.base_node_manager as bs
from sahara.plugins.mapr import images
import sahara.plugins.mapr.util.general as util
import sahara.plugins.provisioning as p import sahara.plugins.provisioning as p
import sahara.plugins.utils as u import sahara.plugins.utils as u
from sahara_plugin_mapr.i18n import _
import sahara_plugin_mapr.plugins.mapr.abstract.version_handler as vh
import sahara_plugin_mapr.plugins.mapr.base.base_cluster_configurer as b_conf
import sahara_plugin_mapr.plugins.mapr.base.base_cluster_validator as bv
import sahara_plugin_mapr.plugins.mapr.base.base_edp_engine as edp
import sahara_plugin_mapr.plugins.mapr.base.base_health_checker as health
import sahara_plugin_mapr.plugins.mapr.base.base_node_manager as bs
from sahara_plugin_mapr.plugins.mapr import images
import sahara_plugin_mapr.plugins.mapr.util.general as util
class BaseVersionHandler(vh.AbstractVersionHandler): class BaseVersionHandler(vh.AbstractVersionHandler):
def __init__(self): def __init__(self):
self._validator = bv.BaseValidator() self._validator = bv.BaseValidator()
self._configurer = base_conf.BaseConfigurer() self._configurer = b_conf.BaseConfigurer()
self._health_checker = health.BaseHealthChecker() self._health_checker = health.BaseHealthChecker()
self._node_manager = bs.BaseNodeManager() self._node_manager = bs.BaseNodeManager()
self._version = None self._version = None

View File

@ -20,9 +20,9 @@ import re
import jinja2 as j2 import jinja2 as j2
import six import six
import sahara.exceptions as e import sahara.plugins.exceptions as e
from sahara.i18n import _ import sahara.plugins.utils as utils
import sahara.utils.xmlutils as xml from sahara_plugin_mapr.i18n import _
@six.add_metaclass(abc.ABCMeta) @six.add_metaclass(abc.ABCMeta)
@ -88,12 +88,12 @@ class HadoopXML(BaseConfigurationFile):
super(HadoopXML, self).__init__(file_name) super(HadoopXML, self).__init__(file_name)
def parse(self, content): def parse(self, content):
configs = xml.parse_hadoop_xml_with_name_and_value(content) configs = utils.parse_hadoop_xml_with_name_and_value(content)
for cfg in configs: for cfg in configs:
self.add_property(cfg["name"], cfg["value"]) self.add_property(cfg["name"], cfg["value"])
def render(self): def render(self):
return xml.create_hadoop_xml(self._config_dict) return utils.create_hadoop_xml(self._config_dict)
class RawFile(BaseConfigurationFile): class RawFile(BaseConfigurationFile):
@ -141,7 +141,8 @@ class TemplateFile(BaseConfigurationFile):
if template: if template:
return template.render(arg_dict) return template.render(arg_dict)
else: else:
raise e.InvalidDataException(_('Template object must be defined')) raise e.PluginsInvalidDataException(
_('Template object must be defined'))
def render(self): def render(self):
return self._j2_render(self._template, self._config_dict) return self._j2_render(self._template, self._config_dict)

View File

@ -15,9 +15,9 @@
import oslo_serialization.jsonutils as json import oslo_serialization.jsonutils as json
from sahara.i18n import _ from sahara.plugins import utils as plugin_utils
import sahara.plugins.mapr.util.general as util from sahara_plugin_mapr.i18n import _
from sahara.utils import poll_utils as polls import sahara_plugin_mapr.plugins.mapr.util.general as util
WARDEN_MANAGED_CMD = ('sudo -u mapr maprcli node services' WARDEN_MANAGED_CMD = ('sudo -u mapr maprcli node services'
@ -77,7 +77,7 @@ class NodeProcess(object):
'instance': instance.instance_name, 'instance': instance.instance_name,
'status': status.name, 'status': status.name,
} }
return polls.poll( return plugin_utils.poll(
get_status=lambda: self.status(instance) == status, get_status=lambda: self.status(instance) == status,
operation_name=operation_name.format(**args), operation_name=operation_name.format(**args),
timeout=timeout, timeout=timeout,

View File

@ -17,15 +17,14 @@ from oslo_log import log as logging
from oslo_serialization import jsonutils as json from oslo_serialization import jsonutils as json
import six import six
import sahara.exceptions as e
from sahara.i18n import _
import sahara.plugins.exceptions as ex import sahara.plugins.exceptions as ex
from sahara.plugins.mapr.util import commands as cmd
from sahara.plugins.mapr.util import event_log as el
from sahara.plugins.mapr.util import general as g
from sahara.plugins.mapr.util import service_utils as su
import sahara.plugins.provisioning as p import sahara.plugins.provisioning as p
from sahara.utils import files from sahara.plugins import utils
from sahara_plugin_mapr.i18n import _
from sahara_plugin_mapr.plugins.mapr.util import commands as cmd
from sahara_plugin_mapr.plugins.mapr.util import event_log as el
from sahara_plugin_mapr.plugins.mapr.util import general as g
from sahara_plugin_mapr.plugins.mapr.util import service_utils as su
LOG = logging.getLogger(__name__) LOG = logging.getLogger(__name__)
@ -160,7 +159,7 @@ class Service(object):
return {self.ui_name: result} return {self.ui_name: result}
def _load_config_file(self, file_path=None): def _load_config_file(self, file_path=None):
return json.loads(files.get_file_text(file_path)) return json.loads(utils.get_file_text(file_path, 'sahara_plugin_mapr'))
def get_config_files(self, cluster_context, configs, instance=None): def get_config_files(self, cluster_context, configs, instance=None):
return [] return []
@ -180,7 +179,7 @@ class Service(object):
raise ex.HadoopProvisionError(_("Config missing 'name'")) raise ex.HadoopProvisionError(_("Config missing 'name'"))
if conf_value is None: if conf_value is None:
raise e.InvalidDataException( raise ex.PluginInvalidDataException(
_("Config '%s' missing 'value'") % conf_name) _("Config '%s' missing 'value'") % conf_name)
if high_priority or item.get('priority', 2) == 1: if high_priority or item.get('priority', 2) == 1:

View File

@ -19,7 +19,8 @@ from sahara.plugins import utils as plugin_utils
_validator = images.SaharaImageValidator.from_yaml( _validator = images.SaharaImageValidator.from_yaml(
'plugins/mapr/resources/images/image.yaml', 'plugins/mapr/resources/images/image.yaml',
resource_roots=['plugins/mapr/resources/images']) resource_roots=['plugins/mapr/resources/images'],
package='sahara_plugin_mapr')
def get_image_arguments(): def get_image_arguments():

View File

@ -13,9 +13,9 @@
# under the License. # under the License.
from sahara.i18n import _
import sahara.plugins.mapr.versions.version_handler_factory as vhf
import sahara.plugins.provisioning as p import sahara.plugins.provisioning as p
from sahara_plugin_mapr.i18n import _
import sahara_plugin_mapr.plugins.mapr.versions.version_handler_factory as vhf
class MapRPlugin(p.ProvisioningPluginBase): class MapRPlugin(p.ProvisioningPluginBase):

View File

@ -13,10 +13,10 @@
# under the License. # under the License.
import sahara.plugins.mapr.domain.node_process as np import sahara_plugin_mapr.plugins.mapr.domain.node_process as np
import sahara.plugins.mapr.domain.service as s import sahara_plugin_mapr.plugins.mapr.domain.service as s
import sahara.plugins.mapr.util.commands as cmd import sahara_plugin_mapr.plugins.mapr.util.commands as cmd
import sahara.plugins.mapr.util.validation_utils as vu import sahara_plugin_mapr.plugins.mapr.util.validation_utils as vu
DRILL = np.NodeProcess( DRILL = np.NodeProcess(
name='drill-bits', name='drill-bits',

View File

@ -13,9 +13,9 @@
# under the License. # under the License.
import sahara.plugins.mapr.domain.node_process as np import sahara_plugin_mapr.plugins.mapr.domain.node_process as np
import sahara.plugins.mapr.domain.service as s import sahara_plugin_mapr.plugins.mapr.domain.service as s
import sahara.plugins.mapr.util.validation_utils as vu import sahara_plugin_mapr.plugins.mapr.util.validation_utils as vu
FLUME = np.NodeProcess( FLUME = np.NodeProcess(

View File

@ -13,10 +13,10 @@
# under the License. # under the License.
import sahara.plugins.mapr.domain.configuration_file as bcf import sahara_plugin_mapr.plugins.mapr.domain.configuration_file as bcf
import sahara.plugins.mapr.domain.node_process as np import sahara_plugin_mapr.plugins.mapr.domain.node_process as np
import sahara.plugins.mapr.domain.service as s import sahara_plugin_mapr.plugins.mapr.domain.service as s
import sahara.plugins.mapr.util.validation_utils as vu import sahara_plugin_mapr.plugins.mapr.util.validation_utils as vu
HBASE_MASTER = np.NodeProcess( HBASE_MASTER = np.NodeProcess(

View File

@ -16,13 +16,13 @@
from oslo_log import log as logging from oslo_log import log as logging
from sahara.i18n import _ from sahara.i18n import _
import sahara.plugins.mapr.domain.configuration_file as bcf
import sahara.plugins.mapr.domain.node_process as np
import sahara.plugins.mapr.domain.service as s
import sahara.plugins.mapr.services.sentry.sentry as sentry
import sahara.plugins.mapr.util.event_log as el
import sahara.plugins.mapr.util.validation_utils as vu
import sahara.utils.files as files import sahara.utils.files as files
import sahara_plugin_mapr.plugins.mapr.domain.configuration_file as bcf
import sahara_plugin_mapr.plugins.mapr.domain.node_process as np
import sahara_plugin_mapr.plugins.mapr.domain.service as s
import sahara_plugin_mapr.plugins.mapr.services.sentry.sentry as sentry
import sahara_plugin_mapr.plugins.mapr.util.event_log as el
import sahara_plugin_mapr.plugins.mapr.util.validation_utils as vu
LOG = logging.getLogger(__name__) LOG = logging.getLogger(__name__)
@ -58,7 +58,8 @@ class Hive(s.Service):
hive_site.remote_path = self.conf_dir(cluster_context) hive_site.remote_path = self.conf_dir(cluster_context)
if instance: if instance:
hive_site.fetch(instance) hive_site.fetch(instance)
hive_site.parse(files.get_file_text(hive_default)) hive_site.parse(files.get_file_text(hive_default,
'sahara_plugin_mapr'))
hive_site.add_properties(self._get_hive_site_props(cluster_context)) hive_site.add_properties(self._get_hive_site_props(cluster_context))
sentry_host = cluster_context.get_instance(sentry.SENTRY) sentry_host = cluster_context.get_instance(sentry.SENTRY)
if sentry_host: if sentry_host:
@ -74,19 +75,21 @@ class Hive(s.Service):
'sentry-default.xml' 'sentry-default.xml'
sentry_db = \ sentry_db = \
'plugins/mapr/services/hive/resources/sentry-db.xml' 'plugins/mapr/services/hive/resources/sentry-db.xml'
hive_site.parse(files.get_file_text(sentry_default)) hive_site.parse(files.get_file_text(sentry_default,
'sahara_plugin_mapr'))
hive_site.add_property('hive.sentry.conf.url', hive_site.add_property('hive.sentry.conf.url',
'file://%s/sentry-site.xml' % 'file://%s/sentry-site.xml' %
sentry_service.conf_dir( sentry_service.conf_dir(
cluster_context)) cluster_context))
if sentry_mode == sentry.DB_STORAGE_SENTRY_MODE: if sentry_mode == sentry.DB_STORAGE_SENTRY_MODE:
hive_site.parse(files.get_file_text(sentry_db)) hive_site.parse(files.get_file_text(sentry_db,
'sahara_plugin_mapr'))
return [hive_site] return [hive_site]
def _get_hive_site_props(self, cluster_context): def _get_hive_site_props(self, cluster_context):
# Import here to resolve circular dependency # Import here to resolve circular dependency
from sahara.plugins.mapr.services.mysql import mysql from sahara_plugin_mapr.plugins.mapr.services.mysql import mysql
zookeepers = cluster_context.get_zookeeper_nodes_ip() zookeepers = cluster_context.get_zookeeper_nodes_ip()
metastore_specs = mysql.MySQL.METASTORE_SPECS metastore_specs = mysql.MySQL.METASTORE_SPECS
@ -104,7 +107,7 @@ class Hive(s.Service):
def _get_jdbc_uri(self, cluster_context): def _get_jdbc_uri(self, cluster_context):
# Import here to resolve circular dependency # Import here to resolve circular dependency
from sahara.plugins.mapr.services.mysql import mysql from sahara_plugin_mapr.plugins.mapr.services.mysql import mysql
jdbc_uri = ('jdbc:mysql://%(db_host)s:%(db_port)s/%(db_name)s?' jdbc_uri = ('jdbc:mysql://%(db_host)s:%(db_port)s/%(db_name)s?'
'createDatabaseIfNotExist=true') 'createDatabaseIfNotExist=true')
@ -122,7 +125,7 @@ class Hive(s.Service):
def post_start(self, cluster_context, instances): def post_start(self, cluster_context, instances):
# Import here to resolve circular dependency # Import here to resolve circular dependency
import sahara.plugins.mapr.services.maprfs.maprfs as mfs import sahara_plugin_mapr.plugins.mapr.services.maprfs.maprfs as mfs
create_path = lambda p: 'sudo -u mapr hadoop fs -mkdir %s' % p create_path = lambda p: 'sudo -u mapr hadoop fs -mkdir %s' % p
check_path = 'sudo -u mapr hadoop fs -ls %s' check_path = 'sudo -u mapr hadoop fs -ls %s'

View File

@ -13,9 +13,9 @@
# under the License. # under the License.
import sahara.plugins.mapr.domain.node_process as np import sahara_plugin_mapr.plugins.mapr.domain.node_process as np
import sahara.plugins.mapr.domain.service as s import sahara_plugin_mapr.plugins.mapr.domain.service as s
import sahara.plugins.mapr.util.validation_utils as vu import sahara_plugin_mapr.plugins.mapr.util.validation_utils as vu
HTTP_FS = np.NodeProcess( HTTP_FS = np.NodeProcess(

View File

@ -19,27 +19,27 @@ import string
from oslo_log import log as logging from oslo_log import log as logging
from sahara.i18n import _
import sahara.plugins.mapr.domain.configuration_file as bcf
import sahara.plugins.mapr.domain.node_process as np
import sahara.plugins.mapr.domain.service as s
import sahara.plugins.mapr.services.hbase.hbase as hbase
import sahara.plugins.mapr.services.hive.hive as hive
import sahara.plugins.mapr.services.httpfs.httpfs as httpfs
import sahara.plugins.mapr.services.impala.impala as impala
import sahara.plugins.mapr.services.mapreduce.mapreduce as mr
import sahara.plugins.mapr.services.mysql.mysql as mysql
import sahara.plugins.mapr.services.oozie.oozie as oozie
import sahara.plugins.mapr.services.sentry.sentry as sentry
import sahara.plugins.mapr.services.spark.spark as spark
import sahara.plugins.mapr.services.sqoop.sqoop2 as sqoop
import sahara.plugins.mapr.services.yarn.yarn as yarn
import sahara.plugins.mapr.util.event_log as el
import sahara.plugins.mapr.util.general as g
import sahara.plugins.mapr.util.password_utils as pu
import sahara.plugins.mapr.util.validation_utils as vu
import sahara.plugins.provisioning as p import sahara.plugins.provisioning as p
import sahara.utils.files as files import sahara.plugins.utils as utils
from sahara_plugin_mapr.i18n import _
import sahara_plugin_mapr.plugins.mapr.domain.configuration_file as bcf
import sahara_plugin_mapr.plugins.mapr.domain.node_process as np
import sahara_plugin_mapr.plugins.mapr.domain.service as s
import sahara_plugin_mapr.plugins.mapr.services.hbase.hbase as hbase
import sahara_plugin_mapr.plugins.mapr.services.hive.hive as hive
import sahara_plugin_mapr.plugins.mapr.services.httpfs.httpfs as httpfs
import sahara_plugin_mapr.plugins.mapr.services.impala.impala as impala
import sahara_plugin_mapr.plugins.mapr.services.mapreduce.mapreduce as mr
import sahara_plugin_mapr.plugins.mapr.services.mysql.mysql as mysql
import sahara_plugin_mapr.plugins.mapr.services.oozie.oozie as oozie
import sahara_plugin_mapr.plugins.mapr.services.sentry.sentry as sentry
import sahara_plugin_mapr.plugins.mapr.services.spark.spark as spark
import sahara_plugin_mapr.plugins.mapr.services.sqoop.sqoop2 as sqoop
import sahara_plugin_mapr.plugins.mapr.services.yarn.yarn as yarn
import sahara_plugin_mapr.plugins.mapr.util.event_log as el
import sahara_plugin_mapr.plugins.mapr.util.general as g
import sahara_plugin_mapr.plugins.mapr.util.password_utils as pu
import sahara_plugin_mapr.plugins.mapr.util.validation_utils as vu
LOG = logging.getLogger(__name__) LOG = logging.getLogger(__name__)
@ -102,7 +102,8 @@ class Hue(s.Service):
# hue.ini # hue.ini
hue_ini = bcf.TemplateFile("hue.ini") hue_ini = bcf.TemplateFile("hue.ini")
hue_ini.remote_path = self.conf_dir(cluster_context) hue_ini.remote_path = self.conf_dir(cluster_context)
hue_ini.parse(files.get_file_text(template % self.version)) hue_ini.parse(utils.get_file_text(template % self.version,
'sahara_plugin_mapr'))
hue_ini.add_properties(self._get_hue_ini_props(cluster_context)) hue_ini.add_properties(self._get_hue_ini_props(cluster_context))
hue_ini.add_property("thrift_version", hue_ini.add_property("thrift_version",
configs[self.THRIFT_VERSION.name]) configs[self.THRIFT_VERSION.name])
@ -112,7 +113,8 @@ class Hue(s.Service):
'resources/hue_sh_%s.template' 'resources/hue_sh_%s.template'
hue_sh = bcf.TemplateFile("hue.sh") hue_sh = bcf.TemplateFile("hue.sh")
hue_sh.remote_path = self.home_dir(cluster_context) + '/bin' hue_sh.remote_path = self.home_dir(cluster_context) + '/bin'
hue_sh.parse(files.get_file_text(hue_sh_template % self.version)) hue_sh.parse(utils.get_file_text(hue_sh_template % self.version,
'sahara_plugin_mapr'))
hue_sh.add_property('hadoop_version', cluster_context.hadoop_version) hue_sh.add_property('hadoop_version', cluster_context.hadoop_version)
hue_sh.mode = 777 hue_sh.mode = 777
@ -339,14 +341,16 @@ class HueLivyV310(s.Service):
livy_conf_template = 'plugins/mapr/services/hue/' \ livy_conf_template = 'plugins/mapr/services/hue/' \
'resources/livy_conf_%s.template' 'resources/livy_conf_%s.template'
livy_conf = bcf.TemplateFile("livy.conf") livy_conf = bcf.TemplateFile("livy.conf")
livy_conf.parse(files.get_file_text(livy_conf_template % self.version)) livy_conf.parse(utils.get_file_text(livy_conf_template % self.version,
'sahara_plugin_mapr'))
livy_conf.remote_path = self.home_dir(cluster_context) + '/conf' livy_conf.remote_path = self.home_dir(cluster_context) + '/conf'
livy_sh_template = 'plugins/mapr/services/hue/' \ livy_sh_template = 'plugins/mapr/services/hue/' \
'resources/livy_sh_%s.template' 'resources/livy_sh_%s.template'
livy_sh = bcf.TemplateFile("livy-env.sh") livy_sh = bcf.TemplateFile("livy-env.sh")
livy_sh.remote_path = self.home_dir(cluster_context) + '/conf' livy_sh.remote_path = self.home_dir(cluster_context) + '/conf'
livy_sh.parse(files.get_file_text(livy_sh_template % self.version)) livy_sh.parse(utils.get_file_text(livy_sh_template % self.version,
'sahara_plugin_mapr'))
livy_sh.add_property('hadoop_version', cluster_context.hadoop_version) livy_sh.add_property('hadoop_version', cluster_context.hadoop_version)
livy_sh.add_property('spark_version', spark.SparkOnYarnV201().version) livy_sh.add_property('spark_version', spark.SparkOnYarnV201().version)
livy_sh.mode = 777 livy_sh.mode = 777

View File

@ -13,15 +13,15 @@
# under the License. # under the License.
import sahara.plugins.mapr.domain.configuration_file as bcf import sahara.plugins.utils as utils
import sahara.plugins.mapr.domain.node_process as np import sahara_plugin_mapr.plugins.mapr.domain.configuration_file as bcf
import sahara.plugins.mapr.domain.service as s import sahara_plugin_mapr.plugins.mapr.domain.node_process as np
import sahara.plugins.mapr.services.hbase.hbase as hbase import sahara_plugin_mapr.plugins.mapr.domain.service as s
import sahara.plugins.mapr.services.hive.hive as hive import sahara_plugin_mapr.plugins.mapr.services.hbase.hbase as hbase
import sahara.plugins.mapr.services.sentry.sentry as sentry import sahara_plugin_mapr.plugins.mapr.services.hive.hive as hive
import sahara.plugins.mapr.util.general as g import sahara_plugin_mapr.plugins.mapr.services.sentry.sentry as sentry
import sahara.plugins.mapr.util.validation_utils as vu import sahara_plugin_mapr.plugins.mapr.util.general as g
import sahara.utils.files as files import sahara_plugin_mapr.plugins.mapr.util.validation_utils as vu
IMPALA_SERVER = np.NodeProcess( IMPALA_SERVER = np.NodeProcess(
name='impalaserver', name='impalaserver',
@ -82,7 +82,7 @@ class Impala(s.Service):
impala_env.remote_path = self.conf_dir(cluster_context) impala_env.remote_path = self.conf_dir(cluster_context)
if instance: if instance:
impala_env.fetch(instance) impala_env.fetch(instance)
impala_env.parse(files.get_file_text(defaults)) impala_env.parse(utils.get_file_text(defaults, 'sahara_plugin_mapr'))
impala_env.add_properties(self._get_impala_env_props(cluster_context)) impala_env.add_properties(self._get_impala_env_props(cluster_context))
sentry_host = cluster_context.get_instance(sentry.SENTRY) sentry_host = cluster_context.get_instance(sentry.SENTRY)
if sentry_host: if sentry_host:

View File

@ -13,9 +13,9 @@
# under the License. # under the License.
import sahara.plugins.mapr.domain.node_process as np import sahara_plugin_mapr.plugins.mapr.domain.node_process as np
import sahara.plugins.mapr.domain.service as s import sahara_plugin_mapr.plugins.mapr.domain.service as s
import sahara.plugins.mapr.util.validation_utils as vu import sahara_plugin_mapr.plugins.mapr.util.validation_utils as vu
KAFKA = np.NodeProcess( KAFKA = np.NodeProcess(
name='kafka', name='kafka',

View File

@ -13,9 +13,9 @@
# under the License. # under the License.
import sahara.plugins.mapr.domain.node_process as np import sahara_plugin_mapr.plugins.mapr.domain.node_process as np
import sahara.plugins.mapr.domain.service as s import sahara_plugin_mapr.plugins.mapr.domain.service as s
import sahara.plugins.mapr.util.validation_utils as vu import sahara_plugin_mapr.plugins.mapr.util.validation_utils as vu
MAHOUT = np.NodeProcess( MAHOUT = np.NodeProcess(
name='mahout', name='mahout',

Some files were not shown because too many files have changed in this diff Show More