Update Config reference for nova
Partial-Bug: #1407581 Closes-Bug: 1478544 Closes-Bug: 1483231 Closes-Bug: 1483192 Closes-Bug: 1489638 Change-Id: I46b990567343dbeed4b99c06c1c006555f3ea645
This commit is contained in:
parent
6c50b80c36
commit
be70453021
@ -23,7 +23,7 @@
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>api_rate_limit</option> = <replaceable>False</replaceable></td>
|
||||
<td>(BoolOpt) Whether to use per-user rate limiting for the api. This option is only used by v2 api. Rate limiting is removed from v3 api.</td>
|
||||
<td>(BoolOpt) Whether to use per-user rate limiting for the api. This option is only used by v2 api. Rate limiting is removed from v2.1 api.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>client_socket_timeout</option> = <replaceable>900</replaceable></td>
|
||||
@ -63,11 +63,11 @@
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>osapi_compute_ext_list</option> = <replaceable></replaceable></td>
|
||||
<td>(ListOpt) Specify list of extensions to load when using osapi_compute_extension option with nova.api.openstack.compute.contrib.select_extensions</td>
|
||||
<td>(ListOpt) DEPRECATED: Specify list of extensions to load when using osapi_compute_extension option with nova.api.openstack.compute.legacy_v2.contrib.select_extensions This option will be removed in the near future. After that point you have to run all of the API.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>osapi_compute_extension</option> = <replaceable>['nova.api.openstack.compute.contrib.standard_extensions']</replaceable></td>
|
||||
<td>(MultiStrOpt) osapi compute extension to load</td>
|
||||
<td><option>osapi_compute_extension</option> = <replaceable>['nova.api.openstack.compute.legacy_v2.contrib.standard_extensions']</replaceable></td>
|
||||
<td>(MultiStrOpt) osapi compute extension to load. This option will be removed in the near future. After that point you have to run all of the API.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>osapi_compute_link_prefix</option> = <replaceable>None</replaceable></td>
|
||||
@ -89,6 +89,10 @@
|
||||
<td><option>osapi_hide_server_address_states</option> = <replaceable>building</replaceable></td>
|
||||
<td>(ListOpt) List of instance states that should hide network info</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>secure_proxy_ssl_header</option> = <replaceable>None</replaceable></td>
|
||||
<td>(StrOpt) The HTTP header used to determine the scheme for the original request, even if it was removed by an SSL terminating proxy. Typical value is "HTTP_X_FORWARDED_PROTO".</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>servicegroup_driver</option> = <replaceable>db</replaceable></td>
|
||||
<td>(StrOpt) The driver for servicegroup service (valid options are: db, zk, mc)</td>
|
||||
|
@ -19,11 +19,9 @@
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>auth_strategy</option> = <replaceable>keystone</replaceable></td>
|
||||
<td>(StrOpt) The strategy to use for auth: keystone, noauth (deprecated), or
|
||||
noauth2. Both noauth and noauth2 are designed for testing only, as
|
||||
they do no actual credential checking. noauth provides administrative
|
||||
credentials regardless of the passed in user, noauth2 only does if
|
||||
'admin' is specified as the username.</td>
|
||||
<td>(StrOpt) The strategy to use for auth: keystone or noauth2. noauth2 is designed for
|
||||
testing only, as it does no actual credential checking. noauth2 provides
|
||||
administrative credentials only if 'admin' is specified as the username.</td>
|
||||
</tr>
|
||||
</tbody>
|
||||
</table>
|
||||
|
@ -33,6 +33,10 @@
|
||||
<td><option>consoleauth_topic</option> = <replaceable>consoleauth</replaceable></td>
|
||||
<td>(StrOpt) The topic console auth proxy nodes listen on</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>executor_thread_pool_size</option> = <replaceable>64</replaceable></td>
|
||||
<td>(IntOpt) Size of executor thread pool.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>host</option> = <replaceable>localhost</replaceable></td>
|
||||
<td>(StrOpt) Name of this node. This can be an opaque identifier. It is not necessarily a hostname, FQDN, or IP address. However, the node name must be valid within an AMQP key, and if using ZeroMQ, a valid hostname, FQDN, or IP address</td>
|
||||
@ -54,7 +58,7 @@
|
||||
<td>(StrOpt) If set, send compute.instance.update notifications on instance state changes. Valid values are None for no notifications, "vm_state" for notifications on VM state changes, or "vm_and_task_state" for notifications on VM and task state changes.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>pybasedir</option> = <replaceable>/usr/lib/python/site-packages/nova</replaceable></td>
|
||||
<td><option>pybasedir</option> = <replaceable>/home/sakaia/work/openstack-doc-tools/autogenerate_config_docs/sources/nova</replaceable></td>
|
||||
<td>(StrOpt) Directory where the nova python module is installed</td>
|
||||
</tr>
|
||||
<tr>
|
||||
@ -77,6 +81,10 @@
|
||||
<td><option>tempdir</option> = <replaceable>None</replaceable></td>
|
||||
<td>(StrOpt) Explicitly specify the temporary working directory</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>use_rootwrap_daemon</option> = <replaceable>False</replaceable></td>
|
||||
<td>(BoolOpt) Start and use a daemon that can run the commands that need to be run with root privileges. This option is usually enabled on nodes that run nova compute processes</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th colspan="2">[keystone_authtoken]</th>
|
||||
</tr>
|
||||
|
@ -31,7 +31,7 @@
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>compute_monitors</option> = <replaceable></replaceable></td>
|
||||
<td>(ListOpt) A list of monitors that can be used for getting compute metrics. You can use the alias/name from the setuptools entry points for nova.compute.monitors.* namespaces.</td>
|
||||
<td>(ListOpt) A list of monitors that can be used for getting compute metrics. You can use the alias/name from the setuptools entry points for nova.compute.monitors.* namespaces. If no namespace is supplied, the "cpu." namespace is assumed for backwards-compatibility. An example value that would enable both the CPU and NUMA memory bandwidth monitors that used the virt driver variant: ["cpu.virt_driver", "numa_mem_bw.virt_driver"]</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>compute_resources</option> = <replaceable>vcpu</replaceable></td>
|
||||
@ -135,7 +135,7 @@
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>shelved_offload_time</option> = <replaceable>0</replaceable></td>
|
||||
<td>(IntOpt) Time in seconds before a shelved instance is eligible for removing from a host. -1 never offload, 0 offload when shelved</td>
|
||||
<td>(IntOpt) Time in seconds before a shelved instance is eligible for removing from a host. -1 never offload, 0 offload immediately when shelved</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>shelved_poll_interval</option> = <replaceable>3600</replaceable></td>
|
||||
|
@ -17,6 +17,10 @@
|
||||
<tr>
|
||||
<th colspan="2">[DEFAULT]</th>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>console_allowed_origins</option> = <replaceable></replaceable></td>
|
||||
<td>(ListOpt) Allowed Origin header hostnames for access to console proxy servers</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>console_public_hostname</option> = <replaceable>localhost</replaceable></td>
|
||||
<td>(StrOpt) Publicly visible name for this console host</td>
|
||||
@ -29,6 +33,17 @@
|
||||
<td><option>consoleauth_manager</option> = <replaceable>nova.consoleauth.manager.ConsoleAuthManager</replaceable></td>
|
||||
<td>(StrOpt) Manager for console auth</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th colspan="2">[mks]</th>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>enabled</option> = <replaceable>False</replaceable></td>
|
||||
<td>(BoolOpt) Enable MKS related features</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>mksproxy_base_url</option> = <replaceable>http://127.0.0.1:6090/</replaceable></td>
|
||||
<td>(StrOpt) Location of MKS web console proxy, in the form "http://127.0.0.1:6090/"</td>
|
||||
</tr>
|
||||
</tbody>
|
||||
</table>
|
||||
</para>
|
||||
|
@ -14,13 +14,6 @@
|
||||
</tr>
|
||||
</thead>
|
||||
<tbody>
|
||||
<tr>
|
||||
<th colspan="2">[DEFAULT]</th>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>backdoor_port</option> = <replaceable>None</replaceable></td>
|
||||
<td>(StrOpt) Enable eventlet backdoor. Acceptable values are 0, <port>, and <start>:<end>, where 0 results in listening on a random tcp port number; <port> results in listening on the specified port number (and not enabling backdoor if that port is in use); and <start>:<end> results in listening on the smallest unused port number within the specified range of port numbers. The chosen port is displayed in the service's log file.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th colspan="2">[guestfs]</th>
|
||||
</tr>
|
||||
|
@ -45,6 +45,14 @@
|
||||
<td><option>mounted_disk_query_retry_interval</option> = <replaceable>5</replaceable></td>
|
||||
<td>(IntOpt) Interval between checks for a mounted iSCSI disk, in seconds.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>power_state_check_timeframe</option> = <replaceable>60</replaceable></td>
|
||||
<td>(IntOpt) The timeframe to be checked for instance power state changes.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>power_state_event_polling_interval</option> = <replaceable>2</replaceable></td>
|
||||
<td>(IntOpt) Instance power state change event polling frequency.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>qemu_img_cmd</option> = <replaceable>qemu-img.exe</replaceable></td>
|
||||
<td>(StrOpt) Path of qemu-img command which is used to convert between different image types</td>
|
||||
|
@ -43,7 +43,7 @@
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>api_max_retries</option> = <replaceable>60</replaceable></td>
|
||||
<td>(IntOpt) How many retries when a request does conflict.</td>
|
||||
<td>(IntOpt) How many retries when a request does conflict. If <= 0, only try once, no retries.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>api_retry_interval</option> = <replaceable>2</replaceable></td>
|
||||
|
@ -21,17 +21,41 @@
|
||||
<td><option>live_migration_retry_count</option> = <replaceable>30</replaceable></td>
|
||||
<td>(IntOpt) Number of 1 second retries needed in live_migration</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>max_concurrent_live_migrations</option> = <replaceable>1</replaceable></td>
|
||||
<td>(IntOpt) Maximum number of live migrations to run concurrently. This limit is enforced to avoid outbound live migrations overwhelming the host/network and causing failures. It is not recommended that you change this unless you are very sure that doing so is safe and stable in your environment.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th colspan="2">[libvirt]</th>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>live_migration_bandwidth</option> = <replaceable>0</replaceable></td>
|
||||
<td>(IntOpt) Maximum bandwidth to be used during migration, in Mbps</td>
|
||||
<td>(IntOpt) Maximum bandwidth(in MiB/s) to be used during migration. If set to 0, will choose a suitable default. Some hypervisors do not support this feature and will return an error if bandwidth is not 0. Please refer to the libvirt documentation for further details</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>live_migration_completion_timeout</option> = <replaceable>800</replaceable></td>
|
||||
<td>(IntOpt) Time to wait, in seconds, for migration to successfully complete transferring data before aborting the operation. Value is per GiB of guest RAM + disk to be transferred, with lower bound of a minimum of 2 GiB. Should usually be larger than downtime delay * downtime steps. Set to 0 to disable timeouts.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>live_migration_downtime</option> = <replaceable>500</replaceable></td>
|
||||
<td>(IntOpt) Maximum permitted downtime, in milliseconds, for live migration switchover. Will be rounded up to a minimum of 100ms. Use a large value if guest liveness is unimportant.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>live_migration_downtime_delay</option> = <replaceable>75</replaceable></td>
|
||||
<td>(IntOpt) Time to wait, in seconds, between each step increase of the migration downtime. Minimum delay is 10 seconds. Value is per GiB of guest RAM + disk to be transferred, with lower bound of a minimum of 2 GiB per device</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>live_migration_downtime_steps</option> = <replaceable>10</replaceable></td>
|
||||
<td>(IntOpt) Number of incremental steps to reach max downtime value. Will be rounded up to a minimum of 3 steps</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>live_migration_flag</option> = <replaceable>VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER, VIR_MIGRATE_LIVE, VIR_MIGRATE_TUNNELLED</replaceable></td>
|
||||
<td>(StrOpt) Migration flags to be set for live migration</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>live_migration_progress_timeout</option> = <replaceable>150</replaceable></td>
|
||||
<td>(IntOpt) Time to wait, in seconds, for migration to make forward progress in transferring data before aborting the operation. Set to 0 to disable timeouts.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>live_migration_uri</option> = <replaceable>qemu+tcp://%s/system</replaceable></td>
|
||||
<td>(StrOpt) Migration target URI (any included "%s" is replaced with the migration target hostname)</td>
|
||||
|
@ -19,10 +19,10 @@
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>debug</option> = <replaceable>False</replaceable></td>
|
||||
<td>(BoolOpt) Print debugging output (set logging level to DEBUG instead of default WARNING level).</td>
|
||||
<td>(BoolOpt) Print debugging output (set logging level to DEBUG instead of default INFO level).</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>default_log_levels</option> = <replaceable>amqp=WARN, amqplib=WARN, boto=WARN, qpid=WARN, sqlalchemy=WARN, suds=INFO, oslo.messaging=INFO, iso8601=WARN, requests.packages.urllib3.connectionpool=WARN, urllib3.connectionpool=WARN, websocket=WARN, requests.packages.urllib3.util.retry=WARN, urllib3.util.retry=WARN, keystonemiddleware=WARN, routes.middleware=WARN, stevedore=WARN</replaceable></td>
|
||||
<td><option>default_log_levels</option> = <replaceable>amqp=WARN, amqplib=WARN, boto=WARN, qpid=WARN, sqlalchemy=WARN, suds=INFO, oslo.messaging=INFO, iso8601=WARN, requests.packages.urllib3.connectionpool=WARN, urllib3.connectionpool=WARN, websocket=WARN, requests.packages.urllib3.util.retry=WARN, urllib3.util.retry=WARN, keystonemiddleware=WARN, routes.middleware=WARN, stevedore=WARN, taskflow=WARN</replaceable></td>
|
||||
<td>(ListOpt) List of logger=LEVEL pairs.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
@ -95,11 +95,11 @@
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>use_syslog_rfc_format</option> = <replaceable>True</replaceable></td>
|
||||
<td>(BoolOpt) (Optional) Enables or disables syslog rfc5424 format for logging. If enabled, prefixes the MSG part of the syslog message with APP-NAME (RFC5424). The format without the APP-NAME is deprecated in K, and will be removed in M, along with this option.</td>
|
||||
<td>(BoolOpt) (Optional) Enables or disables syslog rfc5424 format for logging. If enabled, prefixes the MSG part of the syslog message with APP-NAME (RFC5424). The format without the APP-NAME is deprecated in Kilo, and will be removed in Mitaka, along with this option.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>verbose</option> = <replaceable>False</replaceable></td>
|
||||
<td>(BoolOpt) Print more verbose output (set logging level to INFO instead of default WARNING level).</td>
|
||||
<td><option>verbose</option> = <replaceable>True</replaceable></td>
|
||||
<td>(BoolOpt) If set to false, will disable INFO logging level, making WARNING the default.</td>
|
||||
</tr>
|
||||
</tbody>
|
||||
</table>
|
||||
|
@ -257,6 +257,13 @@
|
||||
<td><option>vlan_start</option> = <replaceable>100</replaceable></td>
|
||||
<td>(IntOpt) First VLAN for private networks</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th colspan="2">[libvirt]</th>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>remote_filesystem_transport</option> = <replaceable>ssh</replaceable></td>
|
||||
<td>(StrOpt) Use ssh or rsync transport for creating, copying, removing files on the remote host.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th colspan="2">[vmware]</th>
|
||||
</tr>
|
||||
|
@ -19,7 +19,7 @@
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>pci_alias</option> = <replaceable>[]</replaceable></td>
|
||||
<td>(MultiStrOpt) An alias for a PCI passthrough device requirement. This allows users to specify the alias in the extra_spec for a flavor, without needing to repeat all the PCI property requirements. For example: pci_alias = { "name": "QuicAssist", "product_id": "0443", "vendor_id": "8086", "device_type": "ACCEL" } defines an alias for the Intel QuickAssist card. (multi valued)</td>
|
||||
<td>(MultiStrOpt) An alias for a PCI passthrough device requirement. This allows users to specify the alias in the extra_spec for a flavor, without needing to repeat all the PCI property requirements. For example: pci_alias = { "name": "QuickAssist", "product_id": "0443", "vendor_id": "8086", "device_type": "ACCEL" } defines an alias for the Intel QuickAssist card. (multi valued)</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>pci_passthrough_whitelist</option> = <replaceable>[]</replaceable></td>
|
||||
|
@ -25,10 +25,6 @@
|
||||
<td><option>periodic_fuzzy_delay</option> = <replaceable>60</replaceable></td>
|
||||
<td>(IntOpt) Range of seconds to randomly delay when starting the periodic task scheduler to reduce stampeding. (Disable by setting to 0)</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>run_external_periodic_tasks</option> = <replaceable>True</replaceable></td>
|
||||
<td>(BoolOpt) Some periodic tasks can be run in a separate process. Should we run them here?</td>
|
||||
</tr>
|
||||
</tbody>
|
||||
</table>
|
||||
</para>
|
||||
|
@ -71,7 +71,7 @@
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>send_single_reply</option> = <replaceable>False</replaceable></td>
|
||||
<td>(BoolOpt) Send a single AMQP reply to call message. The current behaviour since oslo-incubator is to send two AMQP replies - first one with the payload, a second one to ensure the other have finish to send the payload. We are going to remove it in the N release, but we must keep backward compatible at the same time. This option provides such compatibility - it defaults to False in Liberty and can be turned on for early adopters with a new installations or for testing. Please note, that this option will be removed in M release.</td>
|
||||
<td>(BoolOpt) Send a single AMQP reply to call message. The current behaviour since oslo-incubator is to send two AMQP replies - first one with the payload, a second one to ensure the other have finish to send the payload. We are going to remove it in the N release, but we must keep backward compatible at the same time. This option provides such compatibility - it defaults to False in Liberty and can be turned on for early adopters with a new installations or for testing. Please note, that this option will be removed in the Mitaka release.</td>
|
||||
</tr>
|
||||
</tbody>
|
||||
</table>
|
||||
|
@ -111,7 +111,7 @@
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>send_single_reply</option> = <replaceable>False</replaceable></td>
|
||||
<td>(BoolOpt) Send a single AMQP reply to call message. The current behaviour since oslo-incubator is to send two AMQP replies - first one with the payload, a second one to ensure the other have finish to send the payload. We are going to remove it in the N release, but we must keep backward compatible at the same time. This option provides such compatibility - it defaults to False in Liberty and can be turned on for early adopters with a new installations or for testing. Please note, that this option will be removed in M release.</td>
|
||||
<td>(BoolOpt) Send a single AMQP reply to call message. The current behaviour since oslo-incubator is to send two AMQP replies - first one with the payload, a second one to ensure the other have finish to send the payload. We are going to remove it in the N release, but we must keep backward compatible at the same time. This option provides such compatibility - it defaults to False in Liberty and can be turned on for early adopters with a new installations or for testing. Please note, that this option will be removed in the Mitaka release.</td>
|
||||
</tr>
|
||||
</tbody>
|
||||
</table>
|
||||
|
@ -41,10 +41,6 @@
|
||||
<td><option>rpc_response_timeout</option> = <replaceable>60</replaceable></td>
|
||||
<td>(IntOpt) Seconds to wait for a response from a call.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>rpc_thread_pool_size</option> = <replaceable>64</replaceable></td>
|
||||
<td>(IntOpt) Size of RPC thread pool.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th colspan="2">[cells]</th>
|
||||
</tr>
|
||||
|
@ -30,8 +30,8 @@
|
||||
<td>(ListOpt) Which filter class names to use for filtering baremetal hosts when not specified in the request.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>cpu_allocation_ratio</option> = <replaceable>16.0</replaceable></td>
|
||||
<td>(FloatOpt) Virtual CPU to physical CPU allocation ratio which affects all CPU filters. This configuration specifies a global ratio for CoreFilter. For AggregateCoreFilter, it will fall back to this configuration value if no per-aggregate setting found.</td>
|
||||
<td><option>cpu_allocation_ratio</option> = <replaceable>0.0</replaceable></td>
|
||||
<td>(FloatOpt) Virtual CPU to physical CPU allocation ratio which affects all CPU filters. This configuration specifies a global ratio for CoreFilter. For AggregateCoreFilter, it will fall back to this configuration value if no per-aggregate setting found. NOTE: This can be set per-compute, or if set to 0.0, the value set on the scheduler node(s) will be used and defaulted to 16.0</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>disk_allocation_ratio</option> = <replaceable>1.0</replaceable></td>
|
||||
@ -58,8 +58,8 @@
|
||||
<td>(IntOpt) Tells filters to ignore hosts that have this many or more instances currently in build, resize, snapshot, migrate, rescue or unshelve task states</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>ram_allocation_ratio</option> = <replaceable>1.5</replaceable></td>
|
||||
<td>(FloatOpt) Virtual ram to physical ram allocation ratio which affects all ram filters. This configuration specifies a global ratio for RamFilter. For AggregateRamFilter, it will fall back to this configuration value if no per-aggregate setting found.</td>
|
||||
<td><option>ram_allocation_ratio</option> = <replaceable>0.0</replaceable></td>
|
||||
<td>(FloatOpt) Virtual ram to physical ram allocation ratio which affects all ram filters. This configuration specifies a global ratio for RamFilter. For AggregateRamFilter, it will fall back to this configuration value if no per-aggregate setting found. NOTE: This can be set per-compute, or if set to 0.0, the value set on the scheduler node(s) will be used and defaulted to 1.5</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>ram_weight_multiplier</option> = <replaceable>1.0</replaceable></td>
|
||||
@ -82,7 +82,7 @@
|
||||
<td>(MultiStrOpt) Filter classes available to the scheduler which may be specified more than once. An entry of "nova.scheduler.filters.all_filters" maps to all filters included with nova.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>scheduler_default_filters</option> = <replaceable>RetryFilter, AvailabilityZoneFilter, RamFilter, ComputeFilter, ComputeCapabilitiesFilter, ImagePropertiesFilter, ServerGroupAntiAffinityFilter, ServerGroupAffinityFilter</replaceable></td>
|
||||
<td><option>scheduler_default_filters</option> = <replaceable>RetryFilter, AvailabilityZoneFilter, RamFilter, DiskFilter, ComputeFilter, ComputeCapabilitiesFilter, ImagePropertiesFilter, ServerGroupAntiAffinityFilter, ServerGroupAffinityFilter</replaceable></td>
|
||||
<td>(ListOpt) Which filter class names to use for filtering hosts when not specified in the request.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
|
@ -31,7 +31,7 @@
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>cluster_name</option> = <replaceable>None</replaceable></td>
|
||||
<td>(MultiStrOpt) Name of a VMware Cluster ComputeResource.</td>
|
||||
<td>(StrOpt) Name of a VMware Cluster ComputeResource.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>console_delay_seconds</option> = <replaceable>None</replaceable></td>
|
||||
@ -81,6 +81,14 @@
|
||||
<td><option>pbm_wsdl_location</option> = <replaceable>None</replaceable></td>
|
||||
<td>(StrOpt) PBM service WSDL file location URL. e.g. file:///opt/SDK/spbm/wsdl/pbmService.wsdl Not setting this will disable storage policy based placement of instances.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>serial_port_proxy_uri</option> = <replaceable>None</replaceable></td>
|
||||
<td>(StrOpt) Identifies a proxy service that provides network access to the serial_port_service_uri. This option is ignored if serial_port_service_uri is not specified.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>serial_port_service_uri</option> = <replaceable>None</replaceable></td>
|
||||
<td>(StrOpt) Identifies the remote system that serial port traffic will be sent to. If this is not set, no serial ports will be added to the created VMs.</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><option>task_poll_interval</option> = <replaceable>0.5</replaceable></td>
|
||||
<td>(FloatOpt) The interval used for polling of remote tasks.</td>
|
||||
|
@ -7,7 +7,6 @@ api_paste_config api
|
||||
api_rate_limit api
|
||||
auth_strategy authentication
|
||||
auto_assign_floating_ip network
|
||||
backdoor_port debug
|
||||
bandwidth_poll_interval quota
|
||||
baremetal_scheduler_default_filters scheduler
|
||||
bindir common
|
||||
@ -31,6 +30,7 @@ compute_stats_class compute
|
||||
compute_topic common
|
||||
config_drive_format configdrive
|
||||
config_drive_skip_versions configdrive
|
||||
console_allowed_origins console
|
||||
console_driver xen
|
||||
console_host compute
|
||||
console_manager compute
|
||||
@ -90,6 +90,7 @@ enable_network_quota quota
|
||||
enable_new_services api
|
||||
enabled_apis api
|
||||
enabled_ssl_apis api
|
||||
executor_thread_pool_size common
|
||||
fake_call testing
|
||||
fake_network testing
|
||||
fatal_deprecations logging
|
||||
@ -169,6 +170,7 @@ matchmaker_heartbeat_freq rpc
|
||||
matchmaker_heartbeat_ttl rpc
|
||||
max_age policy
|
||||
max_concurrent_builds compute
|
||||
max_concurrent_live_migrations livemigration
|
||||
max_header_line api
|
||||
max_instances_per_host scheduler
|
||||
max_io_ops_per_host scheduler
|
||||
@ -271,7 +273,6 @@ rpc_backend rpc
|
||||
rpc_cast_timeout rpc
|
||||
rpc_conn_pool_size rpc
|
||||
rpc_response_timeout rpc
|
||||
rpc_thread_pool_size rpc
|
||||
rpc_zmq_bind_address zeromq
|
||||
rpc_zmq_contexts zeromq
|
||||
rpc_zmq_host zeromq
|
||||
@ -279,7 +280,6 @@ rpc_zmq_ipc_dir zeromq
|
||||
rpc_zmq_matchmaker zeromq
|
||||
rpc_zmq_port zeromq
|
||||
rpc_zmq_topic_backlog zeromq
|
||||
run_external_periodic_tasks periodic
|
||||
running_deleted_instance_action compute
|
||||
running_deleted_instance_poll_interval compute
|
||||
running_deleted_instance_timeout compute
|
||||
@ -305,6 +305,7 @@ scheduler_topic scheduler
|
||||
scheduler_tracks_instance_changes scheduler
|
||||
scheduler_use_baremetal_filters scheduler
|
||||
scheduler_weight_classes scheduler
|
||||
secure_proxy_ssl_header api
|
||||
security_group_api network
|
||||
send_arp_for_ha network
|
||||
send_arp_for_ha_count network
|
||||
@ -338,6 +339,7 @@ use_ipv6 ipv6
|
||||
use_network_dns_servers network
|
||||
use_neutron_default_nets network
|
||||
use_project_ca ca
|
||||
use_rootwrap_daemon common
|
||||
use_single_default_gateway network
|
||||
use_stderr logging
|
||||
use_syslog logging
|
||||
@ -478,6 +480,8 @@ hyperv/instances_path_share hyperv
|
||||
hyperv/limit_cpu_features hyperv
|
||||
hyperv/mounted_disk_query_retry_count hyperv
|
||||
hyperv/mounted_disk_query_retry_interval hyperv
|
||||
hyperv/power_state_check_timeframe hyperv
|
||||
hyperv/power_state_event_polling_interval hyperv
|
||||
hyperv/qemu_img_cmd hyperv
|
||||
hyperv/volume_attach_retry_count volumes
|
||||
hyperv/volume_attach_retry_interval volumes
|
||||
@ -557,7 +561,12 @@ libvirt/iscsi_iface libvirt
|
||||
libvirt/iscsi_use_multipath libvirt
|
||||
libvirt/iser_use_multipath libvirt
|
||||
libvirt/live_migration_bandwidth livemigration
|
||||
libvirt/live_migration_completion_timeout livemigration
|
||||
libvirt/live_migration_downtime livemigration
|
||||
libvirt/live_migration_downtime_delay livemigration
|
||||
libvirt/live_migration_downtime_steps livemigration
|
||||
libvirt/live_migration_flag livemigration
|
||||
libvirt/live_migration_progress_timeout livemigration
|
||||
libvirt/live_migration_uri livemigration
|
||||
libvirt/mem_stats_period_seconds libvirt
|
||||
libvirt/nfs_mount_options volumes
|
||||
@ -570,6 +579,7 @@ libvirt/quobyte_client_cfg quobyte
|
||||
libvirt/quobyte_mount_point_base quobyte
|
||||
libvirt/rbd_secret_uuid volumes
|
||||
libvirt/rbd_user volumes
|
||||
libvirt/remote_filesystem_transport network
|
||||
libvirt/remove_unused_kernels libvirt
|
||||
libvirt/remove_unused_resized_minimum_age_seconds libvirt
|
||||
libvirt/rescue_image_id libvirt
|
||||
@ -601,6 +611,8 @@ metrics/required scheduler
|
||||
metrics/weight_multiplier scheduler
|
||||
metrics/weight_of_unavailable scheduler
|
||||
metrics/weight_setting scheduler
|
||||
mks/enabled console
|
||||
mks/mksproxy_base_url console
|
||||
neutron/admin_auth_url neutron
|
||||
neutron/admin_password neutron
|
||||
neutron/admin_tenant_id neutron
|
||||
@ -621,9 +633,9 @@ neutron/region_name neutron
|
||||
neutron/service_metadata_proxy neutron
|
||||
neutron/timeout neutron
|
||||
neutron/url neutron
|
||||
osapi_v3/enabled apiv3
|
||||
osapi_v3/extensions_blacklist apiv3
|
||||
osapi_v3/extensions_whitelist apiv3
|
||||
osapi_v21/enabled apiv21
|
||||
osapi_v21/extensions_blacklist apiv21
|
||||
osapi_v21/extensions_whitelist apiv21
|
||||
oslo_concurrency/disable_process_locking rpc
|
||||
oslo_concurrency/lock_path rpc
|
||||
oslo_messaging_amqp/allow_insecure_clients rpc
|
||||
@ -731,6 +743,8 @@ vmware/maximum_objects vmware
|
||||
vmware/pbm_default_policy vmware
|
||||
vmware/pbm_enabled vmware
|
||||
vmware/pbm_wsdl_location vmware
|
||||
vmware/serial_port_proxy_uri vmware
|
||||
vmware/serial_port_service_uri vmware
|
||||
vmware/task_poll_interval vmware
|
||||
vmware/use_linked_clone vmware
|
||||
vmware/vlan_interface network
|
||||
|
@ -1,4 +1,4 @@
|
||||
apiv3 API v3
|
||||
apiv21 API v2.1
|
||||
authentication authentication
|
||||
availabilityzones availability zones
|
||||
barbican Barbican
|
||||
|
Loading…
Reference in New Issue
Block a user