[codespell] Fixing Spelling Mistakes

This is the first in a series of commits to add support for codespell. This is continuning the process completed in ironic-python-agent.

Future Commits will add a Tox Target, CI support and potentially a git-blame-ignore-revs file if their are lots of spelling mistakes that could clutter git blame.

Change-Id: Id328ff64c352e85b58181e9d9e35973a8706ab7a
This commit is contained in:
Sharpz7 2024-02-09 23:48:37 +00:00
parent b0e443f77f
commit 949387bd80
154 changed files with 293 additions and 291 deletions

View File

@ -273,7 +273,7 @@ GET v1/lookup?node_uuid=$NID > lookup-node-response.json
# and the node's driver is "fake", to avoid potential races
# with internal processes that lock the Node
# this corrects an intentional ommission in some of the samples
# this corrects an intentional omission in some of the samples
PATCH v1/nodes/$NID node-update-driver-info-request.json > node-update-driver-info-response.json
GET v1/nodes/$NID/management/boot_device/supported > node-get-supported-boot-devices-response.json

View File

@ -52,7 +52,7 @@ parameters must be missing or match the provided node.
.. versionadded:: 1.79
A node with the same name as the allocation ``name`` is moved to the
start of the derived candidiate list.
start of the derived candidate list.
Normal response codes: 201

View File

@ -68,7 +68,7 @@ and method.
This endpoint passes the request directly to the hardware driver. The
HTTP BODY must be parseable JSON, which will be converted to parameters passed
to that function. Unparseable JSON, missing parameters, or excess parameters
to that function. Unparsable JSON, missing parameters, or excess parameters
will cause the request to be rejected with an HTTP 400 error.
Normal response code: 200 202

View File

@ -61,7 +61,7 @@ and method.
This endpoint passes the request directly to the Node's hardware driver. The
HTTP BODY must be parseable JSON, which will be converted to parameters passed
to that function. Unparseable JSON, missing parameters, or excess parameters
to that function. Unparsable JSON, missing parameters, or excess parameters
will cause the request to be rejected with an HTTP 400 error.
Normal response code: 200 202

View File

@ -292,7 +292,7 @@ r_description_contains:
Filter the list of returned nodes, and only return those containing
substring specified by ``description_contains``.
in: query
requred: false
required: false
type: string
r_driver:
description: |
@ -581,7 +581,7 @@ bios_setting_name:
type: string
bios_setting_read_only:
description: |
This Bios seting is read only and can't be changed.
This Bios setting is read only and can't be changed.
May be None.
in: body
required: true

View File

@ -1,7 +1,7 @@
# NOTE(TheJulia): This is a special bindep file which is independent of the
# project bindep file which is for general usage. This binde pfile is
# intended for execution from Devstack.
# The *primary* purpose being, devstack manages sql dependency mangement
# The *primary* purpose being, devstack manages sql dependency management
# and credential setup, so they can't be included here or it is installed
# prematurely.

View File

@ -1281,7 +1281,7 @@ function configure_ironic_dirs {
if [[ "$IRONIC_BOOT_MODE" == "uefi" ]]; then
if is_ubuntu; then
# NOTE(TheJulia): This is done separately here as this allows
# the script to have hirtuse/bionic compatability.
# the script to have hirtuse/bionic compatibility.
if [[ -f /usr/lib/ipxe/snponly.efi ]]; then
update_loader_copy_paths snponly.efi:/usr/lib/ipxe/snponly.efi
elif [[ -f /usr/lib/ipxe/ipxe.efi ]]; then
@ -1366,7 +1366,7 @@ function configure_ironic_provision_network {
if [[ "$IRONIC_IP_VERSION" == "6" ]]; then
# NOTE(TheJulia): Ideally we should let this happen
# with our global address, but iPXE seems to have in
# consistant behavior in this configuration with devstack.
# consistent behavior in this configuration with devstack.
# so we will setup a dummy interface and use that.
if ! ( ping -c1 -w1 $IRONIC_HOST_IPV6 ); then
# But before we create an interface, lets make sure it is
@ -1724,7 +1724,7 @@ function configure_ironic_conductor {
# - generate outside of this script
# - pass path in as $IRONIC_ANSIBLE_SSH_KEY
# - distribute private key to subnodes under the same path
# Similar to what we do for n-g-s, may be even re-use its key.
# Similar to what we do for n-g-s, may be even reuse its key.
ssh-keygen -t rsa -N '' -f $IRONIC_ANSIBLE_SSH_KEY
chmod 600 $IRONIC_ANSIBLE_SSH_KEY
fi
@ -3038,7 +3038,7 @@ function upload_baremetal_ironic_efiboot {
# request *or* we just do it as the project scoped admin using the admin
# project which in devstack's case is the demo project.
# In other words, we can't use devstack-system-admin to upload the image
# unless we set the project_id in the create reqeust.
# unless we set the project_id in the create request.
IRONIC_EFIBOOT_ID=$(openstack --os-cloud devstack-admin \
image create \
$efiboot_name \

View File

@ -78,7 +78,7 @@ def parse_field_list(content):
def create_bullet_list(input_dict, input_build_env):
"""Convert input_dict into a sphinx representaion of a bullet list."""
"""Convert input_dict into a sphinx representation of a bullet list."""
grp_field = GroupedField('grp_field', label='title')
bullet_list = nodes.paragraph()
@ -138,7 +138,7 @@ def split_list(input_list):
"""Split input_list into three sub-lists.
This function splits the input_list into three, one list containing the
inital non-empty items, one list containing items appearing after the
initial non-empty items, one list containing items appearing after the
string 'Success' in input_list; and the other list containing items
appearing after the string 'Failure' in input_list.
"""

View File

@ -126,7 +126,7 @@ glance:
openstack image create --file ./initrd.img --container-format ari \
--disk-format ari --shared anaconda-ramdisk-<version>
openstack image create --file ./squashfs.img --container-format ari \
--disk-format ari --shared anaconda-stage-<verison>
--disk-format ari --shared anaconda-stage-<version>
openstack image create --file ./os-image.tar.gz \
--container-format bare --disk-format raw --shared \
--property kernel_id=<glance_uuid_vmlinuz> \
@ -162,14 +162,14 @@ ironic node:
back to Ironic and indicate the state. This token is randomly generated
for every deploy, and is required. Specifically this is leveraged in the
template's ``pre``, ``onerror``, and ``post`` steps.
For more infomation on Agent Token, please see :doc:`/admin/agent-token`.
For more information on Agent Token, please see :doc:`/admin/agent-token`.
Standalone deployments
----------------------
While this deployment interface driver was developed around the use of other
OpenStack services, it is not explicitly required. For example HTTP(S) URLs
can be supplied by the API user to explictly set the expected baremetal node
can be supplied by the API user to explicitly set the expected baremetal node
``instance_info`` fields
.. code-block:: shell
@ -299,7 +299,7 @@ Limitations
* This deploy interface has only been tested with Red Hat based operating
systems that use anaconda. Other systems are not supported.
* Runtime TLS certifiate injection into ramdisks is not supported. Assets
* Runtime TLS certificate injection into ramdisks is not supported. Assets
such as ``ramdisk`` or a ``stage2`` ramdisk image need to have trusted
Certificate Authority certificates present within the images *or* the
Ironic API endpoint utilized should utilize a known trusted Certificate

View File

@ -445,7 +445,7 @@ power will be turned off via the management interface. Afterwards, the
While the deployment step framework also supports the
``execute_on_child_nodes`` and ``limit_child_node_execution`` parameters,
all of the step frameworks have a fundamental limitation in that child node
step execution is indended for syncronous actions which do not rely upon
step execution is intended for synchronous actions which do not rely upon
the ``ironic-python-agent`` running on any child nodes. This constraint may
be changed in the future.

View File

@ -115,7 +115,7 @@ Logging
Logging is implemented as custom Ansible callback module,
that makes use of ``oslo.log`` and ``oslo.config`` libraries
and can re-use logging configuration defined in the main ironic configuration
and can reuse logging configuration defined in the main ironic configuration
file to set logging for Ansible events, or use a separate file for this purpose.
It works best when ``journald`` support for logging is enabled.

View File

@ -720,7 +720,7 @@ Set BIOS Config
baremetal node passthru call <node> set_bios_config --arg "name=value"
Walkthrough of perfoming a BIOS configuration change:
Walkthrough of performing a BIOS configuration change:
The following section demonstrates how to change BIOS configuration settings,
detect that a commit and reboot are required, and act on them accordingly. The

View File

@ -361,7 +361,7 @@ Node configuration
before the Xena release.
* The following parameters are mandatory in ``driver_info``
if ``ilo-inspect`` inspect inteface is used and SNMPv3 inspection
if ``ilo-inspect`` inspect interface is used and SNMPv3 inspection
(`SNMPv3 Authentication` in `HPE iLO4 User Guide`_) is desired:
* ``snmp_auth_user`` : The SNMPv3 user.
@ -891,7 +891,7 @@ The hardware type ``ilo`` supports hardware inspection.
an error. This feature is available in proliantutils release
version >= 2.2.0.
* The iLO must be updated with SNMPv3 authentication details.
Pleae refer to the section `SNMPv3 Authentication` in `HPE iLO4 User Guide`_
Please refer to the section `SNMPv3 Authentication` in `HPE iLO4 User Guide`_
for setting up authentication details on iLO.
The following parameters are mandatory to be given in driver_info
for SNMPv3 inspection:
@ -1807,7 +1807,7 @@ refer to `HPE Integrated Lights-Out REST API Documentation <https://hewlettpacka
Allowed values are ``Enabled``, ``Disabled``.
- ``WorkloadProfile``:
Change the Workload Profile to accomodate your desired workload.
Change the Workload Profile to accommodate your desired workload.
Allowed values are ``GeneralPowerEfficientCompute``,
``GeneralPeakFrequencyCompute``, ``GeneralThroughputCompute``,
``Virtualization-PowerEfficient``, ``Virtualization-MaxPerformance``,

View File

@ -177,7 +177,7 @@ Configuration via ``driver_info``
- string representing filesystem path to directory which contains
certification file: In this case, iRMC driver uses certification file
stored at specified directory. Ironic conductor must be able to access
that directory. For iRMC to recongnize certification file, Ironic user
that directory. For iRMC to recognize certification file, Ironic user
must run ``openssl rehash <path_to_dir>``.
- string representing filesystem path to certification file: In this case,

View File

@ -97,7 +97,7 @@ field:
.. note::
The ``redfish_address``, ``redfish_username``, ``redfish_password``,
and ``redfish_verify_ca`` fields, if changed, will trigger a new session
to be establsihed and cached with the BMC. The ``redfish_auth_type`` field
to be established and cached with the BMC. The ``redfish_auth_type`` field
will only be used for the creation of a new cached session, or should
one be rejected by the BMC.
@ -447,7 +447,7 @@ of the remaining updates will pause. When the node is taken out of maintenance
mode, processing of the remaining updates will continue.
When updating the BMC firmware, the BMC may become unavailable for a period of
time as it resets. In this case, it may be desireable to have the cleaning step
time as it resets. In this case, it may be desirable to have the cleaning step
wait after the update has been applied before indicating that the
update was successful. This allows the BMC time to fully reset before further
operations are carried out against it. To cause the cleaning step to wait after
@ -627,7 +627,7 @@ Create Subscription
:widths: 25, 15, 15, 90
"Destination", "body", "string", "The URI of the destination Event Service"
"EventTypes (optional)", "body", "array", "List of ypes of events that shall be sent to the destination"
"EventTypes (optional)", "body", "array", "List of types of events that shall be sent to the destination"
"Context (optional)", "body", "string", "A client-supplied string that is stored with the event destination
subscription"
"Protocol (optional)", "body", "string", "The protocol type that the event will use for sending

View File

@ -213,7 +213,7 @@ to consume the output of the burn-in steps more easily, or even in real-time,
the nodes can be configured to store the output of the individual steps to
files in the ramdisk (from where they can be picked up by a logging pipeline).
The configuration of the outpout file is done via one of
The configuration of the output file is done via one of
``agent_burnin_cpu_outputfile``, ``agent_burnin_vm_outputfile``,
``agent_burnin_fio_disk_outputfile``, and
``agent_burnin_fio_network_outputfile`` parameters which need to be added

View File

@ -2,7 +2,7 @@ Inspection hooks
================
*Inspection hooks* are a type of the Bare Metal service plug-ins responsible
for processing data from in-band inspection. By confuguring these hooks, an
for processing data from in-band inspection. By configuring these hooks, an
operator can fully customize the inspection processing phase. How the data is
collected can be configured with `inspection collectors
<https://docs.openstack.org/ironic-python-agent/latest/admin/how_it_works.html#inspection-data>`_.

View File

@ -34,7 +34,7 @@ option:
[inspector]
callback_endpoint_override = https://example.com/baremetal-introspection/v1/continue
For the built-in inspection, the bare metal API endpoint can be overriden
For the built-in inspection, the bare metal API endpoint can be overridden
instead:
.. code-block:: ini

View File

@ -107,7 +107,7 @@ appended to the kernel command line::
``nomodeset`` as an option. This option is intended for troubleshooting,
and can greatly degrade performance with Matrox/Aspeed BMC Graphics
controllers which is very commonly used on physical servers. The
performance degredation can greatly reduce IO capacity upon every
performance degradation can greatly reduce IO capacity upon every
console graphics update being written to the screen.
Common options

View File

@ -55,11 +55,11 @@ is being worked in Neutron
specific issues, but the OVN DHCP server is an entirely different server
than the interfaces were tested upon.
Maxmium Transmission Units
Maximum Transmission Units
--------------------------
OVN's handling of MTUs has been identified by OVN as being incomplete.
The reality is that it assumes the MTU is not further constained beyond
The reality is that it assumes the MTU is not further constrained beyond
the gateway, which sort of works in some caess for virtual machines, but
might not be applicable with baremetal because your traffic may pass
through lower, or higher MTUs.
@ -106,7 +106,7 @@ a higher operational security posture for the deployment. Users of the
Rescue
------
Due to the aformentioned NAT issues, we know Rescue operations may not work.
Due to the aforementioned NAT issues, we know Rescue operations may not work.
This is being tracked as `bug 2033083 <https://bugs.launchpad.net/ironic/+bug/2033083>`_.

View File

@ -25,7 +25,7 @@ misbehaving BMCs which may delay or even block the synchronization otherwise.
logs). While Ironic automatically retries to get the power status
for the affected nodes, the failure rate may be reduced by increasing
the power sync cycle, e.g. to 300 seconds, and/or by reducing the number
of power sync workers, e.g. to 2. Pleae keep in mind, however, that
of power sync workers, e.g. to 2. Please keep in mind, however, that
depending on the concrete setup increasing the power sync interval may
have an impact on other components relying on up-to-date power states.

View File

@ -5,7 +5,7 @@ Ironic supports booting a user provided ramdisk or an ISO image (starting with
the Victoria release) instead of deploying a node.
Most commonly this is performed when an instance is booted via PXE, iPXE or
Virtual Media, with the only local storage contents being those in memory.
It is suported by ``pxe``, ``ipxe``, ``redfish-virtual-media`` and
It is supported by ``pxe``, ``ipxe``, ``redfish-virtual-media`` and
``ilo-virtual-media`` boot interfaces.
Configuration

View File

@ -72,7 +72,7 @@ will hence not be eligible for scheduling of new instances.
Equally, nodes with ``retired`` set to True cannot move from ``manageable``
to ``available``: the ``provide`` verb is blocked. This is to prevent
accidental re-use of nodes tagged for removal from the fleet. In order
accidental reuse of nodes tagged for removal from the fleet. In order
to move these nodes to ``available`` none the less, the ``retired`` field
needs to be removed first. This can be done via::

View File

@ -15,7 +15,7 @@ That being said, it would likely help anyone working to implement
customization of these policies to consult some reference material
in hopes of understanding the context.
* `Keystone Adminstrator Guide - Service API Protection <https://docs.openstack.org/keystone/latest/admin/service-api-protection.html>`_
* `Keystone Administrator Guide - Service API Protection <https://docs.openstack.org/keystone/latest/admin/service-api-protection.html>`_
* `Ironic Scoped Role Based Access Control Specification <https://specs.openstack.org/openstack/ironic-specs/specs/17.0/secure-rbac.html>`_
Historical Context - How we reached our access model
@ -64,7 +64,7 @@ indicate an associated ``project_id`` value.
The Secure RBAC model, since the introduction of the base capability has been
extended as a result of an OpenStack community goal to include a ``manager``
role in the project scope. By default, this access is equivelent to a Project
role in the project scope. By default, this access is equivalent to a Project
scoped ``admin`` user, however it may be delineated further as time moves
forward.

View File

@ -28,7 +28,7 @@ Similarities to Cleaning and Deployment
=======================================
Similar to Clean and Deploy steps, when invoked an operator can validate
the curent running steps by viewing the ``driver_internal_info`` field
the current running steps by viewing the ``driver_internal_info`` field
looking for a ``service_steps`` field. The *current* step being executed
can be viewed using the baremetal node ``service_step`` field, which is a
top level field.

View File

@ -191,7 +191,7 @@ location of the files will depend on the way you've built the ramdisk.
This will allow you to run commands within the filesystem, e.g. use package
manager. If the ramdisk is also systemd-based, and you have login
credentials set up, you can even boot a real ramdisk enviroment with
credentials set up, you can even boot a real ramdisk environment with
::
@ -548,7 +548,7 @@ of such wait states includes:
* ``clean wait`` for cleaning,
* ``inspect wait`` for introspection,
* ``rescue wait`` for rescueing, and
* ``rescue wait`` for rescuing, and
* ``wait call-back`` for deploying.
Communication issues between the conductor and the node
@ -641,7 +641,7 @@ stable state, and the node should not be provisioned. If the
``baremetal port delete`` command fails, this may indicate that
a known VIF is still attached. Generally if they are transitory from cleaning,
provisioning, rescuing, or even inspection, getting the node to the
``available`` state wil unblock your delete operation, that is unless there is
``available`` state will unblock your delete operation, that is unless there is
a tenant VIF attahment. In that case, the vif will need to be removed from
with-in the Bare Metal service using the
``baremetal node vif detach`` command.
@ -652,7 +652,7 @@ the port's ``internal_info`` field.
.. warning::
The ``maintenance`` flag can be used to force the node's port to be
deleted, however this will disable any check that would normally block
the user from issuing a delete and accidently orphaning the VIF attachment
the user from issuing a delete and accidentally orphaning the VIF attachment
record.
How do I resolve this?
@ -682,8 +682,8 @@ Using that, you can delete the port. Example:
My test VM image does not deploy -- mount point does not exist
==============================================================
What is likely occuring
-----------------------
What is likely occurring
------------------------
The image attempting to be deployed likely is a partition image where
the file system that the user wishes to boot from lacks the required
@ -845,7 +845,7 @@ How did I get here?
One of the major consumers of memory in a host running an ironic-conductor is
transformation of disk images using the ``qemu-img`` tool. This tool, because
the disk images it works with are both compressed and out of linear block
order, requires a considerable amount of memory to efficently re-assemble
order, requires a considerable amount of memory to efficiently re-assemble
and write-out a disk to a device, or to simply convert the format such as
to a ``raw`` image.
@ -986,7 +986,7 @@ baremetal operators.
These settings can be modified by using the
``[conductor]max_concurrent_deploy`` and ``[conductor]max_concurrent_clean``
settings from the ironic.conf file supporting the ``ironic-conductor``
service. Neither setting can be explicity disabled, however there is also no
service. Neither setting can be explicitly disabled, however there is also no
upper limit to the setting.
.. note::
@ -1112,7 +1112,7 @@ basic troubleshooting steps:
* If the version of ironic supports the feature, consult the node history
log, ``baremetal node history list`` and
``baremetal node history get <uuid>``.
* Consult the acutal console screen of the physical machine. *If* the ramdisk
* Consult the actual console screen of the physical machine. *If* the ramdisk
booted, you will generally want to investigate the controller logs and see
if an uploaded agent log is being stored on the conductor responsible for
the baremetal node. Consult `Retrieving logs from the deploy ramdisk`_.

View File

@ -160,7 +160,7 @@ user.
Often, depending on load, query patterns, periodic tasks, and so on and so
forth, additional indexes may be needed to help provide hints to the database
so it can most efficently attempt to reduce the number of rows which need to
so it can most efficiently attempt to reduce the number of rows which need to
be examined in order to return a result set.
Adding indexes

View File

@ -81,7 +81,7 @@ Feature Submission Process
#. As soon as a member of the team acknowledges the bug,
we will move it to the 'Review' state. As time goes on, Discussion
about the RFE, and whether to approve it will occur. If the RFE has not
been triaged and you'd like it to recieve immediate attention, add it to
been triaged and you'd like it to receive immediate attention, add it to
the Open Discussion section of our
`weekly meeting agenda <https://wiki.openstack.org/wiki/Meetings/Ironic>`,
and, timezone permitting, attend the meeting to advocate for your RFE.

View File

@ -13,7 +13,7 @@ standard pattern of entirely python based unit testing. In part this was done
for purposes of speed and to keep the declaration of the test context.
This also lended itself to be very useful due to the nature of A/B testing
which is requried to properly migrate the Ironic project from a project
which is required to properly migrate the Ironic project from a project
scoped universe where an ``admin project`` is utilized as the authenticating
factor coupled with two custom roles, ``baremetal_admin``, and
``baremetal_observer``.
@ -41,7 +41,7 @@ back would have been a heavier lift. As such, the tests largely look for
one of the following error codes.
* 200 - Got the item from the API - This is an database driven interaction.
* 201 - Created - This is databaes driven interaction. These are rare.
* 201 - Created - This is database driven interaction. These are rare.
* 204 - Accepted - This is a database driven interaction. These are rare.
* 403 - Forbidden - This tells us the policy worked as expected where
access was denied.
@ -78,7 +78,7 @@ Some general rules apply
Remember: owner and lessee admins are closer to System scoped Admin Members.
* Members can do some things, but not everything
* Readers can always read, but as we get into sensitive data later on
such as fields containing infrastucture internal addresses, these values
such as fields containing infrastructure internal addresses, these values
will become hidden and additional tests will examine this.
* Third party, or external/other Admins will find nothing but sadness
in empty lists, 403, 404, or even 500 errors.
@ -91,7 +91,7 @@ tests Role Based Access Control related capabilities will come in a
series of phases, styles vary a little.
The first phase is ``"legacy"``. In essence these are partially
programatically generated and then human reviewed and values populated
programmatically generated and then human reviewed and values populated
with expected values.
The second phase is remarkably similar to ``legacy``. It is the safety net
@ -108,7 +108,7 @@ to the ``legacy`` tests. The major difference being some header values,
and a user with a ``member`` role in the ``system`` scope now has some
rights.
The forth phase, is implementaiton of ``owner`` and ``lessee`` aware
The forth phase, is implementation of ``owner`` and ``lessee`` aware
project scoping. The testing approach is similar, however it is much more of
a "shotgun" approach. We test what we know should work, and what know should
not work, but we do not have redundant testing for each role as ``admin``

View File

@ -6,7 +6,7 @@ Contributor Vision
Background
==========
During the Rocky Project Teams Gathering (Feburary/March 2018),
During the Rocky Project Teams Gathering (February/March 2018),
The contributors in the room at that time took a few minutes to write out
each contributor's vision of where they see ironic in five years time.

View File

@ -26,7 +26,7 @@ IPMI is a relatively old protocol and may require additional set up on the
hardware side that the Bare Metal service cannot do automatically:
#. Make sure IPMI is enabled and the account you use have the permissions
to change power and boot devices. By default the adminstrator rights are
to change power and boot devices. By default the administrator rights are
expected, you can change it: see :ref:`ipmi-priv-level`.
#. Make sure the cipher suites are configured for maximum security. Suite 17 is

View File

@ -13,7 +13,7 @@ Configure ironic PXE driver for provisioning using IPv6 addressing
The PXE drivers operate in such a way that they are able to utilize
both IPv4 and IPv6 addresses based upon the deployment's operating state and
configuration. Internally, the drivers attempt to prepare configuration options for both formats, which allows ports which are IPv6 only to automatically
receieve boot parameters. As a result of this, it is critical that the
receive boot parameters. As a result of this, it is critical that the
``[DEFAULT]my_ipv6`` configuration parameter is set to the conductor's
IPv6 address. This option is unique per conductor, and due to the nature
of automatic address assignment, it cannot be "guessed" by the software.

View File

@ -12,12 +12,12 @@ as the underlying resources to provision the node must be available for
successful operations.
.. Warning:: This docuemntation is geared for use of OVS with Neutron along
.. Warning:: This documentation is geared for use of OVS with Neutron along
with the ``neutron-dhcp-agent``. It *is* possible to use OVN
with ``neutron-dhcp-agent``, and depending on version of OVN
and Neutron, OVN's own DHCP service for IPv4 clients, but that
is considered an advanced topic, and we encourage operators
interested in use of OVN to fully undestand it's capabilities
interested in use of OVN to fully understand it's capabilities
and state before attempting to utilize such a configuration.
Please see :doc:`/admin/ovn-networking` for more details.

View File

@ -300,7 +300,7 @@ on the Bare Metal service node(s) where ``ironic-conductor`` is running.
Setting the iPXE parameters noted in the code block above to no value,
in other words setting a line to something like ``ipxe_bootfile_name=``
will result in ironic falling back to the default values of the non-iPXE
PXE settings. This is for backwards compatability.
PXE settings. This is for backwards compatibility.
#. Ensure iPXE is the default PXE, if applicable.
@ -520,7 +520,7 @@ You may also populate other IPA options (e.g. ``ipa-debug=1`` for detailed
logging, ``ipa-inspection-collectors`` to customize the inspection process,
or ``ipa-api-url`` to enable :doc:`/admin/fast-track`).
Second, you need to configure DHCP for unknows hosts since the OpenStack
Second, you need to configure DHCP for unknown hosts since the OpenStack
Networking service won't be able to handle them. For instance, you can install
**dnsmasq** and use the following ``/etc/dnsmasq.conf``:

View File

@ -59,7 +59,7 @@ Bare Metal service comes with an example file for configuring the
/var/log/httpd).
#. Stop and disable the ironic-api service. If ironic-api service is
started, the port will be occupied. Apach will fail to start:
started, the port will be occupied. Apache will fail to start:
Fedora/RHEL8/CentOS8/SUSE::

View File

@ -66,7 +66,7 @@ You should make the following changes to ``/etc/ironic/ironic.conf``:
rpc_transport = json-rpc
JSON RPC also has its own authentication strategy. If it is not specified then
the stategy defaults to ``[DEFAULT]`` ``auth_strategy``. The following will
the strategy defaults to ``[DEFAULT]`` ``auth_strategy``. The following will
set JSON RPC to ``noauth``:
.. code-block:: ini

View File

@ -76,7 +76,7 @@ You need to specify image information in the node's ``instance_info``
* In case the image source requires HTTP(s) Basic Authentication ``RFC 7616``
then the relevant authentication strategy has to be configured as
``http_basic`` and supplied with credentials in the ironic global config
file. Further infromation about the authentication strategy selection
file. Further information about the authentication strategy selection
can be found in :doc:`/admin/user-image-basic-auth`.
* ``root_gb`` - size of the root partition, required for partition images.

View File

@ -271,22 +271,22 @@ class AllocationsController(pecan.rest.RestController):
:fields: fields
:owner: r_owner
"""
requestor = api_utils.check_list_policy('allocation', owner)
requester = api_utils.check_list_policy('allocation', owner)
self._check_allowed_allocation_fields(fields)
if owner is not None and not api_utils.allow_allocation_owner():
# Requestor has asked for an owner field/column match, but
# Requester has asked for an owner field/column match, but
# their client version does not support it.
raise exception.NotAcceptable()
if (owner is not None
and requestor is not None
and owner != requestor):
# The requestor is asking about other owner's records.
and requester is not None
and owner != requester):
# The requester is asking about other owner's records.
# Naughty!
raise exception.NotAuthorized()
if requestor is not None:
owner = requestor
if requester is not None:
owner = requester
return self._get_allocations_collection(node, resource_class, state,
owner, marker, limit,
@ -383,7 +383,7 @@ class AllocationsController(pecan.rest.RestController):
if req_alloc_owner != project_id:
msg = _("Cannot create allocation with an owner "
"Project ID value %(req_owner)s not matching "
"the requestor Project ID %(project)s. "
"the requester Project ID %(project)s. "
"Policy baremetal:allocation:create_restricted"
" is required for this capability."
) % {'req_owner': req_alloc_owner,
@ -427,7 +427,7 @@ class AllocationsController(pecan.rest.RestController):
if not api_utils.check_policy_true(
'baremetal:allocation:create_restricted'):
owner = cdict.get('project_id')
# Filter the candidate search by the requestor project ID
# Filter the candidate search by the requester project ID
# if any. The result is processes authenticating with system
# scope will not be impacted, where as project scoped requests
# will need additional authorization.

View File

@ -79,7 +79,7 @@ def get_next(collection, limit, url, key_field='uuid', **kwargs):
fields = kwargs.pop('fields', None)
# NOTE(saga): If fields argument is present in kwargs and not None. It
# is a list so convert it into a comma seperated string.
# is a list so convert it into a comma separated string.
if fields:
kwargs['fields'] = ','.join(fields)
q_args = ''.join(['%s=%s&' % (key, kwargs[key]) for key in kwargs])

View File

@ -313,7 +313,7 @@ class DeployTemplatesController(rest.RestController):
# This is due to the fact that the patch operation is always applied in
# the API. Ways to avoid this include passing the patch to the
# conductor to apply while holding a lock, or a collision detection
# & retry mechansim using e.g. the updated_at field.
# & retry mechanism using e.g. the updated_at field.
notify.emit_start_notification(context, rpc_template, 'update')
with notify.handle_error_notification(context, rpc_template, 'update'):
rpc_template.save()

View File

@ -923,7 +923,7 @@ class NodeStatesController(rest.RestController):
'modes': ', '.join(ALLOWED_TARGET_BOOT_MODES)})
raise exception.InvalidParameterValue(msg)
# NOTE(cenne): This currenly includes the ADOPTING state
# NOTE(cenne): This currently includes the ADOPTING state
if rpc_node.provision_state in ir_states.UNSTABLE_STATES:
msg = _("Node is in %(state)s state. Since node is transitioning, "
"the boot mode will not be set as this may interfere "
@ -971,7 +971,7 @@ class NodeStatesController(rest.RestController):
{'state': target})
raise exception.InvalidParameterValue(msg)
# NOTE(cenne): This currenly includes the ADOPTING state
# NOTE(cenne): This currently includes the ADOPTING state
if rpc_node.provision_state in ir_states.UNSTABLE_STATES:
msg = _("Node is in %(state)s state. Since node is transitioning, "
"the boot mode will not be set as this may interfere "
@ -1648,7 +1648,7 @@ def node_sanitize(node, fields, cdict=None,
:type fields: list of str
:param cdict: Context dictionary for policy values evaluation.
If not provided, it will be executed by the method,
however for enumerating node lists, it is more efficent
however for enumerating node lists, it is more efficient
to provide.
:param show_driver_secrets: A boolean value to allow external single
evaluation of policy instead of once per
@ -1781,7 +1781,7 @@ def _node_sanitize_extended(node, node_keys, target_dict, cdict):
and not policy.check("baremetal:node:get:last_error",
target_dict, cdict)):
# Guard the last error from being visible as it can contain
# hostnames revealing infrastucture internal details.
# hostnames revealing infrastructure internal details.
node['last_error'] = ('** Value Redacted - Requires '
'baremetal:node:get:last_error '
'permission. **')
@ -2822,7 +2822,7 @@ class NodesController(rest.RestController):
if requested_owner and requested_owner != project_id:
# Translation: If project scoped, and an owner has been
# requested, and that owner does not match the requestor's
# requested, and that owner does not match the requester's
# project ID value.
msg = _("Cannot create a node as a project scoped admin "
"with an owner other than your own project.")

View File

@ -366,7 +366,7 @@ def patched_validate_with_schema(patched_dict, schema, validator=None):
updates applied
:param schema: Any dict key not in the schema will be deleted from the
dict. If no validator is specified then the resulting ``patched_dict``
will be validated agains the schema
will be validated against the schema
:param validator: Optional validator to use if there is extra validation
required beyond the schema
:raises: exception.Invalid if validation fails

View File

@ -94,7 +94,7 @@ BASE_VERSION = 1
# v1.54: Add events support.
# v1.55: Add deploy templates API.
# v1.56: Add support for building configdrives.
# v1.57: Add support for updating an exisiting allocation.
# v1.57: Add support for updating an existing allocation.
# v1.58: Add support for backfilling allocations.
# v1.59: Add support vendor data in configdrives.
# v1.60: Add owner to the allocation object.

View File

@ -75,7 +75,7 @@ class FunctionDefinition(object):
#: The function arguments (list of :class:`FunctionArgument`)
self.arguments = []
#: If the body carry the datas of a single argument, its type
#: If the body carry the data of a single argument, its type
self.body_type = None
#: Status code
@ -86,7 +86,7 @@ class FunctionDefinition(object):
#: exceptions
self.ignore_extra_args = False
#: Dictionnary of protocol-specific options.
#: Dictionary of protocol-specific options.
self.extra_options = None
@staticmethod
@ -145,7 +145,7 @@ class signature(object):
:param body: If the function takes a final argument that is supposed to be
the request body by itself, its type.
:param status_code: HTTP return status code of the function.
:param ignore_extra_args: Allow extra/unknow arguments (default to False)
:param ignore_extra_args: Allow extra/unknown arguments (default to False)
Most of the time this decorator is not supposed to be used directly,
unless you are not using WSME on top of another framework.

View File

@ -112,7 +112,7 @@ def body(body_arg):
def format_exception(excinfo, debug=False):
"""Extract informations that can be sent to the client."""
"""Extract information that can be sent to the client."""
error = excinfo[1]
code = getattr(error, 'code', None)
if code and code in http_client.responses and (400 <= code < 500):

View File

@ -35,7 +35,7 @@ def _error_about_sqlite_usage():
# TODO(TheJulia): Make this a hard error in C*
LOG.error('We have detected the API is being launched with a SQLite '
'database backend. This is unsupported, and will be a hard '
'error in the future. This is becaues multi-process use of '
'error in the future. This is because multi-process use of '
'a single SQLite database is problematic in terms of '
'locking. A single process ironic model exists for use with '
'SQLite.')

View File

@ -327,7 +327,7 @@ def validate(*args, **kwargs):
"""Decorator which validates and transforms function arguments
"""
assert not args, 'Validators must be specifed by argument name'
assert not args, 'Validators must be specified by argument name'
assert kwargs, 'No validators specified'
validators = kwargs

View File

@ -87,7 +87,7 @@ def get_client(context, auth_from_config=False):
auth=user_auth or service_auth)
# Re-determine the endpoint so we can work with versions prior to
# Yoga, becuase the endpoint, based upon configuration, may require
# Yoga, because the endpoint, based upon configuration, may require
# project_id specific URLs.
if user_auth:
endpoint = keystone.get_endpoint('cinder', session=sess,

View File

@ -552,7 +552,7 @@ def create_boot_iso(context, output_filename, kernel_href,
elif CONF.esp_image:
esp_image_path = CONF.esp_image
# TODO(TheJulia): we should opportunisticly try to make bios
# TODO(TheJulia): we should opportunistically try to make bios
# bootable and UEFI. In other words, collapse a lot of this
# path since they are not mutually exclusive.
# UEFI boot mode, but Network iPXE -> ISO means bios bootable

View File

@ -139,7 +139,7 @@ def get_service_auth(context, endpoint, service_auth,
:param context: The RequestContext instance from which the user
auth_token is extracted.
:param endpoint: The requested endpoint to be utilized.
:param service_auth: The service authenticaiton credentals to be
:param service_auth: The service authentication credentals to be
used.
:param only_service_auth: Boolean, default False. When set to True,
the resulting Service token pair is generated

View File

@ -748,7 +748,7 @@ node_policies = [
deprecated_rule=deprecated_node_clear_maintenance
),
# NOTE(TheJulia): This should liekly be deprecated and be replaced with
# NOTE(TheJulia): This should likely be deprecated and be replaced with
# a cached object.
policy.DocumentedRuleDefault(
name='baremetal:node:get_boot_device',
@ -977,7 +977,7 @@ node_policies = [
name='baremetal:node:history:get',
check_str=SYSTEM_OR_OWNER_READER,
scope_types=['system', 'project'],
description='Filter to allow operators to retreive history records '
description='Filter to allow operators to retrieve history records '
'for a node.',
operations=[
{'path': '/nodes/{node_ident}/history', 'method': 'GET'},
@ -1929,7 +1929,7 @@ def init_enforcer(policy_file=None, rules=None,
rules=rules,
default_rule=default_rule,
use_conf=use_conf)
# NOTE(melwitt): Explictly disable the warnings for policies
# NOTE(melwitt): Explicitly disable the warnings for policies
# changing their default check_str. During policy-defaults-refresh
# work, all the policy defaults have been changed and warning for
# each policy started filling the logs limit for various tool.

View File

@ -780,7 +780,7 @@ def get_instance_image_info(task, ipxe_enabled=False):
# told anaconda how to execute.
if i_info.get('ks_template'):
# If the value is set, we always overwrite it, in the event
# a rebuild is occuring or something along those lines.
# a rebuild is occurring or something along those lines.
node.set_driver_internal_info('ks_template',
i_info['ks_template'])
else:

View File

@ -114,7 +114,7 @@ def is_valid_logical_name(hostname):
"""Determine if a logical name is valid.
The logical name may only consist of RFC3986 unreserved
characters, to wit:
characters:
ALPHA / DIGIT / "-" / "." / "_" / "~"
"""

View File

@ -212,7 +212,7 @@ def _allocate_node(context, allocation, nodes):
try:
# NOTE(dtantsur): retries are done for all nodes above, so disable
# per-node retry. Also disable loading the driver, since the
# current conductor may not have the requried hardware type or
# current conductor may not have the required hardware type or
# interfaces (it's picked at random).
with task_manager.acquire(context, node.uuid, shared=False,
retry=False, load_driver=False,

View File

@ -104,7 +104,7 @@ class BaseConductorManager(object):
"""Initialize the conductor host.
:param admin_context: the admin context to pass to periodic tasks.
:param start_consoles: If consoles should be started in intialization.
:param start_consoles: If consoles should be started in initialization.
:param start_allocations: If allocations should be started in
initialization.
:raises: RuntimeError when conductor is already running.

View File

@ -275,7 +275,7 @@ def execute_step_on_child_nodes(task, step):
:param task: The TaskManager object for the parent node.
:param step: The requested step to be executed.
:returns: None on Success, the resulting error message if a
failure has occured.
failure has occurred.
"""
# NOTE(TheJulia): We could just use nodeinfo list calls against
# dbapi.

View File

@ -520,7 +520,7 @@ def execute_step_on_child_nodes(task, step):
:param task: The TaskManager object for the parent node.
:param step: The requested step to be executed.
:returns: None on Success, the resulting error message if a
failure has occured.
failure has occurred.
"""
# NOTE(TheJulia): We could just use nodeinfo list calls against
# dbapi.

View File

@ -2678,7 +2678,7 @@ class ConductorManager(base_manager.BaseConductorManager):
except AttributeError:
# TODO(TheJulia): Remove this at some point, but right now
# don't inherently break on version mismatches when people
# disregard requriements.
# disregard requirements.
LOG.warning(
'get_sensors_data has been configured to collect '
'conductor metrics, however the installed ironic-lib '
@ -2686,7 +2686,7 @@ class ConductorManager(base_manager.BaseConductorManager):
'ironic-lib to a minimum of version 5.4.0.')
except Exception as e:
LOG.exception(
"An unknown error occured while attempting to collect "
"An unknown error occurred while attempting to collect "
"sensor data from within the conductor. Error: %(error)s",
{'error': e})
else:
@ -3588,7 +3588,7 @@ class ConductorManager(base_manager.BaseConductorManager):
{'node': node_id})
# Allow lookup to work by returning a value, it is just an
# unusable value that can't be verified against.
# This is important if the agent lookup has occured with
# This is important if the agent lookup has occurred with
# pre-generation of tokens with virtual media usage.
node.set_driver_internal_info('agent_secret_token', "******")
return node