68999ce9cb
We add a directory v_5_4_0 in cdh plugin, so that we have a choice 5.4.0 when cdh plugin is chosen. Implements: blueprint cdh-5-4-support Change-Id: I05d2082b589eb031b0ce1b5a253dc62a75912e1b
368 lines
21 KiB
JSON
368 lines
21 KiB
JSON
[
|
|
{
|
|
"desc": "The health test thresholds for monitoring of free space on the filesystem that contains this role's log directory. Specified as a percentage of the capacity on that filesystem. This setting is not used if a Log Directory Free Space Monitoring Absolute Thresholds setting is configured.",
|
|
"display_name": "Log Directory Free Space Monitoring Percentage Thresholds",
|
|
"name": "log_directory_free_space_percentage_thresholds",
|
|
"value": "{\"critical\":\"never\",\"warning\":\"never\"}"
|
|
},
|
|
{
|
|
"desc": "Kerberos principal used by the Impala Llama ApplicationMaster roles.",
|
|
"display_name": "Role-Specific Kerberos Principal",
|
|
"name": "kerberos_role_princ_name",
|
|
"value": "llama"
|
|
},
|
|
{
|
|
"desc": "Whether Llama should cache allocated resources on release.",
|
|
"display_name": "Enable Resource Caching",
|
|
"name": "llama_am_cache_enabled",
|
|
"value": "true"
|
|
},
|
|
{
|
|
"desc": "Whether or not periodic stacks collection is enabled.",
|
|
"display_name": "Stacks Collection Enabled",
|
|
"name": "stacks_collection_enabled",
|
|
"value": "false"
|
|
},
|
|
{
|
|
"desc": "Directory where Llama will place its log files.",
|
|
"display_name": "Llama Log Directory",
|
|
"name": "llama_log_dir",
|
|
"value": "/var/log/impala-llama"
|
|
},
|
|
{
|
|
"desc": "Queues Llama ApplicationMaster should connect to at start up.",
|
|
"display_name": "Core Queues",
|
|
"name": "llama_am_core_queues",
|
|
"value": ""
|
|
},
|
|
{
|
|
"desc": "Maximum amount of time the backed off reservations will be in 'backed off' state. The actual amount time is a random value between the minimum and the maximum.",
|
|
"display_name": "Anti-Deadlock Maximum Delay",
|
|
"name": "llama_am_gang_anti_deadlock_max_delay_ms",
|
|
"value": "30000"
|
|
},
|
|
{
|
|
"desc": "The health test thresholds for monitoring of free space on the filesystem that contains this role's heap dump directory. Specified as a percentage of the capacity on that filesystem. This setting is not used if a Heap Dump Directory Free Space Monitoring Absolute Thresholds setting is configured.",
|
|
"display_name": "Heap Dump Directory Free Space Monitoring Percentage Thresholds",
|
|
"name": "heap_dump_directory_free_space_percentage_thresholds",
|
|
"value": "{\"critical\":\"never\",\"warning\":\"never\"}"
|
|
},
|
|
{
|
|
"desc": "ACL for Impala ApplicationMaster clients. The ACL is a comma-separated list of user and group names. The user and group list is separated by a blank. For e.g. \"alice,bob users,wheel\". A special value of \"*\" means all users are allowed. These take effect only if security is enabled.",
|
|
"display_name": "Client ACLs",
|
|
"name": "llama_am_server_thrift_client_acl",
|
|
"value": "*"
|
|
},
|
|
{
|
|
"desc": "For advanced use only, a string to be inserted into <strong>llama-site.xml</strong> for this role only.",
|
|
"display_name": "Impala Llama ApplicationMaster Advanced Configuration Snippet (Safety Valve) for llama-site.xml",
|
|
"name": "llama_config_valve",
|
|
"value": null
|
|
},
|
|
{
|
|
"desc": "When computing the overall Impala Llama ApplicationMaster health, consider the host's health.",
|
|
"display_name": "Impala Llama ApplicationMaster Host Health Test",
|
|
"name": "llama_host_health_enabled",
|
|
"value": "true"
|
|
},
|
|
{
|
|
"desc": "Port on which the Llama ApplicationMaster listens to HTTP requests.",
|
|
"display_name": "Llama HTTP Port",
|
|
"name": "llama_http_port",
|
|
"value": "15001"
|
|
},
|
|
{
|
|
"desc": "Minimum number of threads used by the Llama ApplicationMaster auxiliary service uses for serving client requests.",
|
|
"display_name": "Thrift Server Minimum Threads",
|
|
"name": "llama_am_server_thrift_server_min_threads",
|
|
"value": "10"
|
|
},
|
|
{
|
|
"desc": "The method used to collect stacks. The jstack option involves periodically running the jstack command against the role's daemon process. The servlet method is available for those roles that have an HTTP server endpoint exposing the current stacks traces of all threads. When the servlet method is selected, that HTTP endpoint is periodically scraped.",
|
|
"display_name": "Stacks Collection Method",
|
|
"name": "stacks_collection_method",
|
|
"value": "jstack"
|
|
},
|
|
{
|
|
"desc": "Enter an XML string that will be inserted verbatim into the Fair Scheduler allocations file. Overrides the configuration set using the Pools configuration UI. This configuration only has effect on Impala versions 1.3 or greater.",
|
|
"display_name": "Fair Scheduler XML Advanced Configuration Snippet (Safety Valve)",
|
|
"name": "llama_fair_scheduler_safety_valve",
|
|
"value": null
|
|
},
|
|
{
|
|
"desc": "When set, this role's process is automatically (and transparently) restarted in the event of an unexpected failure.",
|
|
"display_name": "Automatically Restart Process",
|
|
"name": "process_auto_restart",
|
|
"value": "true"
|
|
},
|
|
{
|
|
"desc": "<p>The configured triggers for this role. This is a JSON formatted list of triggers. These triggers are evaluated as part as the health system. Every trigger expression is parsed, and if the trigger condition is met, the list of actions provided in the trigger expression is executed.</p><p>Each trigger has all of the following fields:</p><ul><li><code>triggerName</code> <strong>(mandatory)</strong> - The name of the trigger. This value must be unique for the specific role. </li><li><code>triggerExpression</code> <strong>(mandatory)</strong> - A tsquery expression representing the trigger. </li><li><code>streamThreshold</code> <strong>(optional)</strong> - The maximum number of streams that can satisfy a condition of a trigger before the condition fires. By default set to 0, and any stream returned causes the condition to fire. </li><li><code>enabled</code> <strong> (optional)</strong> - By default set to 'true'. If set to 'false', the trigger will not be evaluated.</li><li><code>expressionEditorConfig</code> <strong> (optional)</strong> - Metadata for the trigger editor. If present, the trigger should only be edited from the Edit Trigger page; editing the trigger here may lead to inconsistencies.</li></ul></p><p>For example, the following JSON formatted trigger configured for a DataNode fires if the DataNode has more than 1500 file-descriptors opened:</p><p><pre>[{\"triggerName\": \"sample-trigger\",\n \"triggerExpression\": \"IF (SELECT fd_open WHERE roleName=$ROLENAME and last(fd_open) > 1500) DO health:bad\",\n \"streamThreshold\": 0, \"enabled\": \"true\"}]</pre></p><p>See the trigger rules documentation for more details on how to write triggers using tsquery.</p><p>The JSON format is evolving and may change in the future and, as a result, backward compatibility is not guaranteed between releases at this time.</p>",
|
|
"display_name": "Role Triggers",
|
|
"name": "role_triggers",
|
|
"value": "[]"
|
|
},
|
|
{
|
|
"desc": "Client notification retry interval, in milliseconds.",
|
|
"display_name": "Client Notification Retry Interval",
|
|
"name": "llama_am_server_thrift_client_notifier_retry_interval_ms",
|
|
"value": "5000"
|
|
},
|
|
{
|
|
"desc": "Soft memory limit to assign to this role, enforced by the Linux kernel. When the limit is reached, the kernel will reclaim pages charged to the process if and only if the host is facing memory pressure. If reclaiming fails, the kernel may kill the process. Both anonymous as well as page cache pages contribute to the limit. Use a value of -1 B to specify no limit. By default processes not managed by Cloudera Manager will have no limit.",
|
|
"display_name": "Cgroup Memory Soft Limit",
|
|
"name": "rm_memory_soft_limit",
|
|
"value": "-1"
|
|
},
|
|
{
|
|
"desc": "The group that the Llama processes should run as.",
|
|
"display_name": "Llama System Group",
|
|
"name": "process_groupname",
|
|
"value": "llama"
|
|
},
|
|
{
|
|
"desc": "Timeout policy for resources being cached.",
|
|
"display_name": "Resource Caching Idle Timeout",
|
|
"name": "llama_am_cache_eviction_timeout_policy_idle_timeout_ms",
|
|
"value": "30000"
|
|
},
|
|
{
|
|
"desc": "Port on which the Llama ApplicationMaster listens to administrative requests on its administrative Thrift interface.",
|
|
"display_name": "Llama Thrift Admin Port",
|
|
"name": "llama_am_server_thrift_admin_address",
|
|
"value": "15002"
|
|
},
|
|
{
|
|
"desc": "Maximum size in bytes for the Java Process heap memory. Passed to Java -Xmx.",
|
|
"display_name": "Java Heap Size of Impala Llama ApplicationMaster in Bytes",
|
|
"name": "llama_java_heapsize",
|
|
"value": "268435456"
|
|
},
|
|
{
|
|
"desc": "Interval of time without any new allocation that will trigger the Impala ApplicationMaster anti-deadlock logic.",
|
|
"display_name": "Anti-Deadlock No Allocation Limit Interval",
|
|
"name": "llama_am_gang_anti_deadlock_no_allocation_limit_ms",
|
|
"value": "30000"
|
|
},
|
|
{
|
|
"desc": "The health test thresholds on the swap memory usage of the process.",
|
|
"display_name": "Process Swap Memory Thresholds",
|
|
"name": "process_swap_memory_thresholds",
|
|
"value": "{\"critical\":\"never\",\"warning\":\"any\"}"
|
|
},
|
|
{
|
|
"desc": "The period to review when computing unexpected exits.",
|
|
"display_name": "Unexpected Exits Monitoring Period",
|
|
"name": "unexpected_exits_window",
|
|
"value": "5"
|
|
},
|
|
{
|
|
"desc": "Maximum number of threads used by the Llama ApplicationMaster auxiliary service uses for serving client requests.",
|
|
"display_name": "Thrift Server Maximum Threads",
|
|
"name": "llama_am_server_thrift_server_max_threads",
|
|
"value": "50"
|
|
},
|
|
{
|
|
"desc": "For advanced use only, key-value pairs (one on each line) to be inserted into a role's environment. Applies to configurations of this role except client configuration.",
|
|
"display_name": "Impala Llama ApplicationMaster Environment Advanced Configuration Snippet (Safety Valve)",
|
|
"name": "LLAMA_role_env_safety_valve",
|
|
"value": null
|
|
},
|
|
{
|
|
"desc": "The amount of stacks data that is retained. After the retention limit is reached, the oldest data is deleted.",
|
|
"display_name": "Stacks Collection Data Retention",
|
|
"name": "stacks_collection_data_retention",
|
|
"value": "104857600"
|
|
},
|
|
{
|
|
"desc": "Hard memory limit to assign to this role, enforced by the Linux kernel. When the limit is reached, the kernel will reclaim pages charged to the process. If reclaiming fails, the kernel may kill the process. Both anonymous as well as page cache pages contribute to the limit. Use a value of -1 B to specify no limit. By default processes not managed by Cloudera Manager will have no limit.",
|
|
"display_name": "Cgroup Memory Hard Limit",
|
|
"name": "rm_memory_hard_limit",
|
|
"value": "-1"
|
|
},
|
|
{
|
|
"desc": "The health test thresholds for monitoring of free space on the filesystem that contains this role's heap dump directory.",
|
|
"display_name": "Heap Dump Directory Free Space Monitoring Absolute Thresholds",
|
|
"name": "heap_dump_directory_free_space_absolute_thresholds",
|
|
"value": "{\"critical\":\"5.36870912E9\",\"warning\":\"1.073741824E10\"}"
|
|
},
|
|
{
|
|
"desc": "Time in milliseconds after which Llama will discard its AM for a queue that has been empty of reservations. Does not apply to queues specified with the Core Queues property.",
|
|
"display_name": "Queue Expiration Age",
|
|
"name": "llama_am_queue_expire_ms",
|
|
"value": "300000"
|
|
},
|
|
{
|
|
"desc": "For advanced use only, a string to be inserted into <strong>log4j.properties</strong> for this role only.",
|
|
"display_name": "Impala Llama ApplicationMaster Logging Advanced Configuration Snippet (Safety Valve)",
|
|
"name": "log4j_safety_valve",
|
|
"value": null
|
|
},
|
|
{
|
|
"desc": "Minimum amount of time the backed off reservations will be in 'backed off' state. The actual amount time is a random value between the minimum and the maximum.",
|
|
"display_name": "Anti-Deadlock Minimum Delay",
|
|
"name": "llama_am_gang_anti_deadlock_min_delay_ms",
|
|
"value": "10000"
|
|
},
|
|
{
|
|
"desc": "The health test thresholds for monitoring of free space on the filesystem that contains this role's log directory.",
|
|
"display_name": "Log Directory Free Space Monitoring Absolute Thresholds",
|
|
"name": "log_directory_free_space_absolute_thresholds",
|
|
"value": "{\"critical\":\"5.36870912E9\",\"warning\":\"1.073741824E10\"}"
|
|
},
|
|
{
|
|
"desc": "The health test thresholds of the number of file descriptors used. Specified as a percentage of file descriptor limit.",
|
|
"display_name": "File Descriptor Monitoring Thresholds",
|
|
"name": "llama_fd_thresholds",
|
|
"value": "{\"critical\":\"70.0\",\"warning\":\"50.0\"}"
|
|
},
|
|
{
|
|
"desc": "If enabled, the Impala Llama ApplicationMaster binds to the wildcard address (\"0.0.0.0\") on all of its ports.",
|
|
"display_name": "Bind Impala Llama ApplicationMaster to Wildcard Address",
|
|
"name": "llama_bind_wildcard",
|
|
"value": "false"
|
|
},
|
|
{
|
|
"desc": "If configured, overrides the process soft and hard rlimits (also called ulimits) for file descriptors to the configured value.",
|
|
"display_name": "Maximum Process File Descriptors",
|
|
"name": "rlimit_fds",
|
|
"value": null
|
|
},
|
|
{
|
|
"desc": "When set, Cloudera Manager will send alerts when the health of this role reaches the threshold specified by the EventServer setting eventserver_health_events_alert_threshold",
|
|
"display_name": "Enable Health Alerts for this Role",
|
|
"name": "enable_alerts",
|
|
"value": "true"
|
|
},
|
|
{
|
|
"desc": "Weight for the read I/O requests issued by this role. The greater the weight, the higher the priority of the requests when the host experiences I/O contention. Must be between 100 and 1000. Defaults to 1000 for processes not managed by Cloudera Manager.",
|
|
"display_name": "Cgroup I/O Weight",
|
|
"name": "rm_io_weight",
|
|
"value": "500"
|
|
},
|
|
{
|
|
"desc": "The maximum size, in megabytes, per log file for Impala Llama ApplicationMaster logs. Typically used by log4j or logback.",
|
|
"display_name": "Impala Llama ApplicationMaster Max Log Size",
|
|
"name": "max_log_size",
|
|
"value": "200"
|
|
},
|
|
{
|
|
"desc": "When set, a SIGKILL signal is sent to the role process when java.lang.OutOfMemoryError is thrown.",
|
|
"display_name": "Kill When Out of Memory",
|
|
"name": "oom_sigkill_enabled",
|
|
"value": "true"
|
|
},
|
|
{
|
|
"desc": "When set, generates heap dump file when java.lang.OutOfMemoryError is thrown.",
|
|
"display_name": "Dump Heap When Out of Memory",
|
|
"name": "oom_heap_dump_enabled",
|
|
"value": "false"
|
|
},
|
|
{
|
|
"desc": "These arguments will be passed as part of the Java command line. Commonly, garbage collection flags or extra debugging flags would be passed here.",
|
|
"display_name": "Java Configuration Options for Llama Server",
|
|
"name": "llama_java_opts",
|
|
"value": ""
|
|
},
|
|
{
|
|
"desc": "The health test thresholds for unexpected exits encountered within a recent period specified by the unexpected_exits_window configuration for the role.",
|
|
"display_name": "Unexpected Exits Thresholds",
|
|
"name": "unexpected_exits_thresholds",
|
|
"value": "{\"critical\":\"any\",\"warning\":\"never\"}"
|
|
},
|
|
{
|
|
"desc": "Llama ApplicationMaster heartbeat interval, in milliseconds. On each heartbeat the ApplicationMaster submits new reservations to YARN ResourceManager and gets updates from it.",
|
|
"display_name": "AM Heartbeat Interval",
|
|
"name": "llama_am_server_thrift_client_notifier_heartbeat_ms",
|
|
"value": "5000"
|
|
},
|
|
{
|
|
"desc": "The user that the Llama process should run as.",
|
|
"display_name": "Llama System User",
|
|
"name": "process_username",
|
|
"value": "llama"
|
|
},
|
|
{
|
|
"desc": "When set, Cloudera Manager will send alerts when this entity's configuration changes.",
|
|
"display_name": "Enable Configuration Change Alerts",
|
|
"name": "enable_config_alerts",
|
|
"value": "false"
|
|
},
|
|
{
|
|
"desc": "Percentage of resources that will be backed off by the Impala ApplicationMaster anti-deadlock logic. Random reservations will be backed off until the percentage of backed off resources reaches this percentage.",
|
|
"display_name": "Anti-Deadlock Backoff Percentage",
|
|
"name": "llama_am_gang_anti_deadlock_backoff_percent",
|
|
"value": "30"
|
|
},
|
|
{
|
|
"desc": "Maximum number of retries for a client notification. After the maximum number of client notification retries has been reached without success the client is considered lost and all its reservations are released. A successful client notification resets the retries count.",
|
|
"display_name": "Maximum Client Notification Retries",
|
|
"name": "llama_am_server_thrift_client_notifier_max_retries",
|
|
"value": "5"
|
|
},
|
|
{
|
|
"desc": "Port on which the Llama ApplicationMaster serves its Thrift interface.",
|
|
"display_name": "Llama Thrift Port",
|
|
"name": "llama_port",
|
|
"value": "15000"
|
|
},
|
|
{
|
|
"desc": "Whether to break resource requests into smaller requests of standard size before the resource cache. The sizes are taken from Yarn settings Container Memory Increment and Container Virtual CPU Cores Increment.",
|
|
"display_name": "Enable Resource Cache Normalization",
|
|
"name": "llama_am_resource_normalizing_enabled",
|
|
"value": "true"
|
|
},
|
|
{
|
|
"desc": "The maximum number of rolled log files to keep for Impala Llama ApplicationMaster logs. Typically used by log4j or logback.",
|
|
"display_name": "Impala Llama ApplicationMaster Maximum Log File Backups",
|
|
"name": "max_log_backup_index",
|
|
"value": "10"
|
|
},
|
|
{
|
|
"desc": "Socket time, in milliseconds, used Llama ApplicationMaster auxiliary service for all its server and client Thrift connections.",
|
|
"display_name": "Thrift Transport Timeout",
|
|
"name": "llama_am_server_thrift_transport_timeout_ms",
|
|
"value": "60000"
|
|
},
|
|
{
|
|
"desc": "Enables the health test that the Impala Llama ApplicationMaster's process state is consistent with the role configuration",
|
|
"display_name": "Impala Llama ApplicationMaster Process Health Test",
|
|
"name": "llama_scm_health_enabled",
|
|
"value": "true"
|
|
},
|
|
{
|
|
"desc": "Number of CPU shares to assign to this role. The greater the number of shares, the larger the share of the host's CPUs that will be given to this role when the host experiences CPU contention. Must be between 2 and 262144. Defaults to 1024 for processes not managed by Cloudera Manager.",
|
|
"display_name": "Cgroup CPU Shares",
|
|
"name": "rm_cpu_shares",
|
|
"value": "1024"
|
|
},
|
|
{
|
|
"desc": "Path to directory where heap dumps are generated when java.lang.OutOfMemoryError error is thrown. This directory is automatically created if it does not exist. If this directory already exists, role user must have write access to this directory. If this directory is shared among multiple roles, it should have 1777 permissions. The heap dump files are created with 600 permissions and are owned by the role user. The amount of free space in this directory should be greater than the maximum Java Process heap size configured for this role.",
|
|
"display_name": "Heap Dump Directory",
|
|
"name": "oom_heap_dump_dir",
|
|
"value": "/tmp"
|
|
},
|
|
{
|
|
"desc": "ACL for Impala ApplicationMaster admins. The ACL is a comma-separated list of user and group names. The user and group list is separated by a blank. For e.g. \"alice,bob users,wheel\". A special value of \"*\" means all users are allowed. These take effect only if security is enabled.",
|
|
"display_name": "Administrative Interface ACLs",
|
|
"name": "llama_am_server_thrift_admin_acl",
|
|
"value": "*"
|
|
},
|
|
{
|
|
"desc": "The frequency with which stacks are collected.",
|
|
"display_name": "Stacks Collection Frequency",
|
|
"name": "stacks_collection_frequency",
|
|
"value": "5.0"
|
|
},
|
|
{
|
|
"desc": "The directory in which stacks logs are placed. If not set, stacks are logged into a <span class='code'>stacks</span> subdirectory of the role's log directory.",
|
|
"display_name": "Stacks Collection Directory",
|
|
"name": "stacks_collection_directory",
|
|
"value": null
|
|
},
|
|
{
|
|
"desc": "The minimum log level for Impala Llama ApplicationMaster logs",
|
|
"display_name": "Impala Llama ApplicationMaster Logging Threshold",
|
|
"name": "log_threshold",
|
|
"value": "INFO"
|
|
}
|
|
] |