a6df0c1dfc
Implements: blueprint mapr-refactor Change-Id: Ib4385845a00aab2506dddccb9e6a6f3297a5622d
87 lines
3.2 KiB
Python
87 lines
3.2 KiB
Python
# Copyright (c) 2015, MapR Technologies
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
|
|
import os
|
|
|
|
import sahara.plugins.mapr.util.maprfs_helper as mfs
|
|
import sahara.plugins.mapr.versions.version_handler_factory as vhf
|
|
import sahara.service.edp.binary_retrievers.dispatch as d
|
|
import sahara.service.edp.oozie.engine as e
|
|
from sahara.utils import edp
|
|
|
|
|
|
class MapROozieJobEngine(e.OozieJobEngine):
|
|
def __init__(self, cluster):
|
|
super(MapROozieJobEngine, self).__init__(cluster)
|
|
self.ctx = self._get_cluster_context(self.cluster)
|
|
|
|
hdfs_user = 'mapr'
|
|
|
|
def get_hdfs_user(self):
|
|
return MapROozieJobEngine.hdfs_user
|
|
|
|
def create_hdfs_dir(self, remote, dir_name):
|
|
mfs.create_maprfs4_dir(remote, dir_name, self.get_hdfs_user())
|
|
|
|
def _upload_workflow_file(self, where, job_dir, wf_xml, hdfs_user):
|
|
f_name = 'workflow.xml'
|
|
with where.remote() as r:
|
|
mfs.put_file_to_maprfs(r, wf_xml, f_name, job_dir, hdfs_user)
|
|
return os.path.join(job_dir, f_name)
|
|
|
|
def _upload_job_files_to_hdfs(self, where, job_dir, job, configs,
|
|
proxy_configs=None):
|
|
mains = job.mains or []
|
|
libs = job.libs or []
|
|
builtin_libs = edp.get_builtin_binaries(job, configs)
|
|
uploaded_paths = []
|
|
hdfs_user = self.get_hdfs_user()
|
|
lib_dir = job_dir + '/lib'
|
|
|
|
with where.remote() as r:
|
|
for m in mains:
|
|
raw_data = d.get_raw_binary(m, proxy_configs)
|
|
mfs.put_file_to_maprfs(r, raw_data, m.name, job_dir, hdfs_user)
|
|
uploaded_paths.append(os.path.join(job_dir, m.name))
|
|
if len(libs) > 0:
|
|
self.create_hdfs_dir(r, lib_dir)
|
|
for l in libs:
|
|
raw_data = d.get_raw_binary(l, proxy_configs)
|
|
mfs.put_file_to_maprfs(r, raw_data, l.name, lib_dir,
|
|
hdfs_user)
|
|
uploaded_paths.append(os.path.join(lib_dir, l.name))
|
|
for lib in builtin_libs:
|
|
mfs.put_file_to_maprfs(r, lib['raw'], lib['name'], lib_dir,
|
|
hdfs_user)
|
|
uploaded_paths.append(lib_dir + '/' + lib['name'])
|
|
return uploaded_paths
|
|
|
|
def get_name_node_uri(self, cluster):
|
|
return self.ctx.name_node_uri
|
|
|
|
def get_oozie_server_uri(self, cluster):
|
|
return self.ctx.oozie_server_uri
|
|
|
|
def get_oozie_server(self, cluster):
|
|
return self.ctx.oozie_server
|
|
|
|
def get_resource_manager_uri(self, cluster):
|
|
return self.ctx.resource_manager_uri
|
|
|
|
def _get_cluster_context(self, cluster):
|
|
h_version = cluster.hadoop_version
|
|
v_handler = vhf.VersionHandlerFactory.get().get_handler(h_version)
|
|
return v_handler.get_context(cluster)
|