Updating dashboard user guide doc for icehouse. * New HDFS data source * "Internal database" name change * Relaunch functionality * Remove reference to "job binary internal" Partial-Implements: blueprint update-docs-icehouse Change-Id: I87d2c48ebd37aa682582f9b1c6f8d763e207ddaf
7.8 KiB
Sahara UI User Guide
This guide assumes that you already have sahara-api and the Sahara Dashboard configured and running. If you require assistance with that, please see the installation guides.
Launching a cluster via the Sahara Dashboard
Registering an Image
- Navigate to the "Sahara" tab in the dashboard, then click on the "Image Registry" panel.
- From that page, click on the "Register Image" button at the top right.
- Choose the image that you'd like to register as a Hadoop Image
- Enter the username of the cloud-init user on the image.
- Click on the tags that you want to add to the image. (A version ie: 1.2.1 and a type ie: vanilla are required for cluster functionality)
- Click the "Done" button to finish the registration.
Create Node Group Templates
- Navigate to the "Sahara" tab in the dashboard, then click on the "Node Group Templates" panel.
- From that page, click on the "Create Template" button at the top right.
- Choose your desired Plugin name and Version from the dropdowns and click "Create".
- Give your Node Group Template a name (description is optional)
- Choose a flavor for this template (based on your CPU/memory/disk needs)
- Choose the storage location for your instance, this can be either "Ephemeral Drive" or "Cinder Volume". If you choose "Cinder Volume", you will need to add additional configuration.
- Choose which processes should be run for any instances that are spawned from this Node Group Template.
- Click on the "Create" button to finish creating your Node Group Template.
Create a Cluster Template
- Navigate to the "Sahara" tab in the dashboard, then click on the "Cluster Templates" panel.
- From that page, click on the "Create Template" button at the top right.
- Choose your desired Plugin name and Version from the dropdowns and click "Create".
- Under the "Details" tab, you must give your template a name.
- Under the "Node Groups" tab, you should add one or more nodes that can be based on one or more templates.
- To do this, start by choosing a Node Group Template from the dropdown and click the "+" button.
- You can adjust the number of nodes to be spawned for this node group via the text box or the "-" and "+" buttons.
- Repeat these steps if you need nodes from additional node group templates.
- Optionally, you can adjust your configuration further by using the "General Parameters", "HDFS Parameters" and "MapReduce Parameters" tabs.
- Click on the "Create" button to finish creating your Cluster Template.
Launching a Cluster
- Navigate to the "Sahara" tab in the dashboard, then click on the "Clusters" panel.
- Click on the "Launch Cluster" button at the top right.
- Choose your desired Plugin name and Version from the dropdowns and click "Create".
- Give your cluster a name. (required)
- Choose which cluster template should be used for your cluster.
- Choose the image that should be used for your cluster (if you do not see any options here, see Registering an Image above).
- Optionally choose a keypair that can be used to authenticate to your cluster instances.
- Click on the "Create" button to start your cluster.
- Your cluster's status will display on the Clusters table.
- It will likely take several minutes to reach the "Active" state.
Scaling a Cluster
- From the Sahara/Clusters page, click on the "Scale Cluster" button of the row that contains the cluster that you want to scale.
- You can adjust the numbers of instances for existing Node Group Templates.
- You can also add a new Node Group Template and choose a number of instances to launch.
- This can be done by selecting your desired Node Group Template from the dropdown and clicking the "+" button.
- Your new Node Group will appear below and you can adjust the number of instances via the text box or the +/- buttons.
- To confirm the scaling settings and trigger the spawning/deletion of instances, click on "Scale".
Elastic Data Processing (EDP)
Data Sources
Data Sources are where the input and output from your jobs are housed.
- From the Sahara/Data Sources page, click on the "Create Data Source" button at the top right.
- Give your Data Source a name.
- Enter the URL to the Data Source.
- For a Swift object, the url will look like <container>.sahara/<path> (ie: mycontainer.sahara/inputfile). The "swift://" is automatically added for you.
- For an HDFS object, the url will look like <host>/<path> (ie: myhost/user/hadoop/inputfile). The "hdfs://" is automatically added for you.
- Enter the username and password for the Data Source.
- Enter an optional description.
- Click on "Create".
- Repeat for additional Data Sources.
Job Binaries
Job Binaries are where you define/upload the source code (mains and libraries) for your job.
- From the Sahara/Job Binaries page, click on the "Create Job Binary" button at the top right.
- Give your Job Binary a name (this can be different than the actual filename).
- Choose the type of storage for your Job Binary.
- For "Swift", you will need to enter the URL of your binary (<container>.sahara/<path>) as well as the username and password.
- For "Internal database", you can choose from "Create a script" or "Upload a new file".
- Enter an optional description.
- Click on "Create".
- Repeat for additional Job Binaries
Jobs
Jobs are where you define the type of job you'd like to run as well as which "Job Binaries" are required.
- From the Sahara/Jobs page, click on the "Create Job" button at the top right.
- Give your Job a name.
- Choose the type of job you'd like to run (Pig, Hive, MapReduce, Streaming MapReduce, Java Action)
- Choose the main binary from the dropdown (not applicable for MapReduce or Java Action).
- Enter an optional description for your Job.
- Optionally, click on the "Libs" tab and add one or more libraries that are required for your job. Each library must be defined as a Job Binary.
- Click on "Create".
Job Executions
Job Executions are what you get by "Launching" a job. You can monitor the status of your job to see when it has completed its run.
- From the Sahara/Jobs page, find the row that contains the job you want to launch and click on the "Launch Job" button at the right side of that row.
- Choose the cluster (already running--see Launching a Cluster above) on which you would like the job to run.
- Choose the Input and Output Data Sources (Data Sources defined above).
- If additional configuration is required, click on the "Configure" tab.
- Additional configuration properties can be defined by clicking on the "Add" button.
- An example configuration entry might be mapred.mapper.class for the Name and org.apache.oozie.example.SampleMapper for the Value.
- Click on "Launch". To monitor the status of your job, you can navigate to the Sahara/Job Executions panel.
- You can relaunch a Job Execution from the Job Executions page by using the "Relaunch on New Cluster" or "Relaunch on Existing Cluster" links.
- Relaunch on New Cluster will take you through the forms to start a new cluster before letting you specify input/output Data Sources and job configuration.
- Relaunch on Existing Cluster will prompt you for input/output Data Sources as well as allow you to change job configuration before launching the job.
Additional Notes
1) Throughout the Sahara UI, you will find that if you try to delete an object that you will not be able to delete it if another object depends on it. An example of this would be trying to delete a Job that has an existing Job Execution. In order to be able to delete that job, you would first need to delete any Job Executions that relate to that job.