Juju Charm - Keystone
Go to file
Arif Ali a85ae0e852 Update README.md and gitkeep consistancy
Remove 2 extra lines, and capitilsation of Juju, so the doc is
consitent. 2 empty directories either have .keep or .gitkeep
change .keep to .gitkeep to be consistent

Change-Id: Ib965a7e2a7afcd36423dadf548d33a8efbfaf6fe
2019-11-14 11:36:25 +00:00
actions Correct token hashing check 2019-03-19 15:39:35 +01:00
charmhelpers Policyd override implementation 2019-10-07 20:31:02 +01:00
files Sync charm/ceph helpers, tox, and requirements 2019-09-30 22:10:09 -05:00
hooks Policyd override implementation 2019-10-07 20:31:02 +01:00
lib Update README.md and gitkeep consistancy 2019-11-14 11:36:25 +00:00
scripts Convert the charm to Python 3 only 2018-09-21 09:09:47 +00:00
templates Replacing direct import for OAuth plugin 2019-09-24 14:14:23 +01:00
tests Enable functional tests for train 2019-10-16 15:31:54 +00:00
unit_tests Policyd override implementation 2019-10-07 20:31:02 +01:00
.gitignore Add disco-stein bundle to gate tests. 2019-08-26 11:15:53 -04:00
.gitreview OpenDev Migration Patch 2019-04-19 19:41:17 +00:00
.project Merge ssl-everywhere branch (may break stuff) 2014-03-27 10:54:38 +00:00
.pydevproject Fix issue with crontab enablement 2019-02-21 15:10:21 +00:00
.stestr.conf Configure stestr directly 2019-03-06 16:06:22 +01:00
.zuul.yaml Add Python 3 Train unit tests 2019-07-30 10:20:25 -04:00
actions.yaml Add security checklist to Keystone 2019-03-07 06:54:40 +00:00
charm-helpers-hooks.yaml Sync charm/ceph helpers, tox, and requirements 2019-09-30 22:10:09 -05:00
config.yaml Policyd override implementation 2019-10-07 20:31:02 +01:00
copyright Re-license charm as Apache-2.0 2016-07-03 16:39:47 +00:00
hardening.yaml Add hardening support 2016-03-24 13:17:48 +00:00
icon.svg Updated icon 2017-05-23 10:22:55 +01:00
LICENSE Re-license charm as Apache-2.0 2016-07-03 16:39:47 +00:00
Makefile Switch functional test framework from Amulet to Zaza 2018-08-21 04:48:44 +02:00
metadata.yaml Policyd override implementation 2019-10-07 20:31:02 +01:00
README.md Update README.md and gitkeep consistancy 2019-11-14 11:36:25 +00:00
requirements.txt Sync charm/ceph helpers, tox, and requirements 2019-09-30 22:10:09 -05:00
setup.cfg Add Python 3 Train unit tests 2019-07-30 10:20:25 -04:00
test-requirements.txt Sync charm/ceph helpers, tox, and requirements 2019-09-30 22:10:09 -05:00
tox.ini Sync charm/ceph helpers, tox, and requirements 2019-09-30 22:10:09 -05:00

Overview

This charm provides Keystone, the OpenStack identity service. Its target platform is (ideally) Ubuntu LTS + OpenStack.

Usage

The following interfaces are provided:

  • nrpe-external-master: Used to generate Nagios checks.

  • identity-service: OpenStack API endpoints request an entry in the Keystone service catalog + endpoint template catalog. When a relation is established, Keystone receives: service_name, region, public_url, admin_url and internal_url. It first checks that the requested service is listed as a supported service. This list should stay updated to support current OpenStack core services. If the service is supported, an entry in the service catalog is created, an endpoint template is created and an admin token is generated. The other end of the relation receives the token as well as info on which ports Keystone is listening on.

  • keystone-service: This is currently only used by Horizon/dashboard as its interaction with Keystone is different from other OpenStack API services. That is, Horizon requests a Keystone role and token exists. During a relation, Horizon requests its configured default role and Keystone responds with a token and the auth + admin ports on which Keystone is listening.

  • identity-admin: Charms use this relation to obtain the credentials for the admin user. This is intended for charms that automatically provision users, tenants, etc. or that otherwise automate using the OpenStack cluster deployment.

  • identity-notifications: Used to broadcast messages to any services listening on the interface.

  • identity-credentials: Charms use this relation to obtain keystone credentials without creating a service catalog entry. Set 'username' only on the relation and keystone will set defaults and return authentication details. Possible relation settings:

    • username Username to be created.
    • project Project (tenant) name to be created. Defaults to services project.
    • requested_roles Comma delimited list of roles to be created.
    • requested_grants Comma delimited list of roles to be granted. Defaults to Admin role.
    • domain Keystone v3 domain the user will be created in. Defaults to the Default domain.

Database

Keystone requires a database. The charm supports relation to a shared database server through the mysql-shared interface. When a new data store is configured, the charm ensures the minimum administrator credentials exist (as configured in charm configuration)

HA/Clustering

There are two mutually exclusive high availability options: using virtual IP(s) or DNS. In both cases, a relationship to hacluster is required which provides the corosync backend HA functionality.

To use virtual IP(s), the clustered nodes must be on the same subnet such that the VIP is a valid IP on the subnet for one of the node's interfaces and each node has an interface in said subnet. The VIP becomes a highly-available API endpoint.

At a minimum, the config option vip must be set in order to use virtual IP HA. If multiple networks are being used, a VIP should be provided for each network, separated by spaces. Optionally, vip_iface or vip_cidr may be specified.

To use DNS high availability, there are several prerequisites. However, DNS HA does not require the clustered nodes to be on the same subnet. Currently, the DNS HA feature is only available for MAAS 2.0 or greater environments. MAAS 2.0 requires Juju 2.0 or greater. The clustered nodes must have static or "reserved" IP addresses registered in MAAS. The DNS hostname(s) must be pre-registered in MAAS before use with DNS HA.

At a minimum, the configuration option dns-ha must be set to true and at least one of os-public-hostname, os-internal-hostname or os-admin-hostname must be set in order to use DNS HA. One or more of the above hostnames may be set.

The charm will throw an exception in the following circumstances:

  • If neither vip nor dns-ha is set and the charm is related to hacluster

  • If both vip and dns-ha are set as they are mutually exclusive

  • If dns-ha is set and none of the os-{admin,internal,public}-hostname configuration options are set

TLS/HTTPS

Support for TLS and HTTPS endpoints can be enabled through configuration options.

To enable TLS and HTTPS endpoints with a certificate signed by your own Certificate Authority, set the following configuration options:

  • ssl_ca

  • ssl_cert

  • ssl_key

Example bundle usage:

keystone:
  charm: cs:keystone
  num_units: 1
  options:
    ssl_ca:   include-base64://path-to-PEM-formatted-ca-data
    ssl_cert: include-base64://path-to-PEM-formatted-certificate-data
    ssl_key:  include-base64://path-to-PEM-formatted-key-data

Note

: If your certificate is signed by a Certificate Authority present in the CA Certificate Store in operating systems used in your deployment, you do not need to provide the ssl_ca configuration option.

Note

: The include-base64 bundle keyword tells Juju to source a file and Base64 encode it before storing it as a configuration option value. The path can be absolute or relative to the location of the bundle file.

Network Space Support

This charm supports the use of Juju Network Spaces, allowing the charm to be bound to network space configurations managed directly by Juju. This is only supported with Juju 2.0 and above.

API endpoints can be bound to distinct network spaces supporting the network separation of public, internal and admin endpoints.

Access to the underlying MySQL instance can also be bound to a specific space using the shared-db relation.

To use this feature, use the --bind option when deploying the charm:

juju deploy keystone --bind \
"public=public-space internal=internal-space admin=admin-space shared-db=internal-space"

Alternatively, these can also be provided as part of a Juju native bundle configuration:

keystone:
  charm: cs:xenial/keystone
  num_units: 1
  bindings:
    public: public-space
    admin: admin-space
    internal: internal-space
    shared-db: internal-space

NOTE: Spaces must be configured in the underlying provider prior to attempting to use them.

NOTE: Existing deployments using os\-\*-network configuration options will continue to function; these options are preferred over any network space binding provided if set.

Policy Overrides

This feature allows for policy overrides using the policy.d directory. This is an advanced feature and the policies that the OpenStack service supports should be clearly and unambiguously understood before trying to override, or add to, the default policies that the service uses. The charm also has some policy defaults. They should also be understood before being overridden.

Caution

: It is possible to break the system (for tenants and other services) if policies are incorrectly applied to the service.

Policy overrides are YAML files that contain rules that will add to, or override, existing policy rules in the service. The policy.d directory is a place to put the YAML override files. This charm owns the /etc/keystone/policy.d directory, and as such, any manual changes to it will be overwritten on charm upgrades.

Overrides are provided to the charm using a Juju resource called policyd-override. The resource is a ZIP file. This file, say overrides.zip, is attached to the charm by:

juju attach-resource keystone policyd-override=overrides.zip

The policy override is enabled in the charm using:

juju config keystone use-policyd-override=true

When use-policyd-override is True the status line of the charm will be prefixed with PO: indicating that policies have been overridden. If the installation of the policy override YAML files failed for any reason then the status line will be prefixed with PO (broken):. The log file for the charm will indicate the reason. No policy override files are installed if the PO (broken): is shown. The status line indicates that the overrides are broken, not that the policy for the service has failed. The policy will be the defaults for the charm and service.

Policy overrides on one service may affect the functionality of another service. Therefore, it may be necessary to provide policy overrides for multiple service charms to achieve a consistent set of policies across the OpenStack system. The charms for the other services that may need overrides should be checked to ensure that they support overrides before proceeding.

Token Support

As the keystone charm supports multiple releases of the OpenStack software, it also supports two keystone token systems: UUID and Fernet. The capabilities are:

  • pre 'ocata': UUID tokens only.
  • ocata and pike: UUID or Fernet tokens, configured via the 'token-provider' configuration parameter.
  • rocky and later: Fernet tokens only.

Fernet tokens were added to OpenStack to solve the problem of keystone being required to persist tokens to a common database (cluster) like UUID tokens, and solve the problem of size for PKI or PKIZ tokens.

For further information, please see Fernet - Frequently Asked Questions.

Theory of Operation

Fernet keys are used to generate tokens; they are generated by keystone and have an expiration date. The key repository is a directory, and each key is an integer number, with the highest number being the primary key. Key '0' is the staged key, that will be the next primary. Other keys are secondary keys.

New tokens are only ever generated from the primary key, whilst the secondary keys are used to validate existing tokens. The staging key is not used to generate tokens but can be used to validate tokens as the staging key might be the new primary key on the master due to a rotation and the keys have not yet been synchronised across all the units.

Fernet keys need to be rotated at periodic intervals, and the keys need to be synchronised to each of the other keystone units. Keys should only be rotated on the master keystone unit and must be synchronised before they are rotated again. Over rotation occurs if a unit rotates its keys such that there is no suitable decoding key on another unit that can decode a token that has been generated on the master. This happens if two key rotations are done on the master before a synchronisation has been successfully performed. This should be avoided. Over rotations can also cause validation keys to be removed before a token's expiration which would result in failed validations.

There are 3 parts to the Key Rotation Strategy:

  1. The rotation frequency
  2. The token lifespan
  3. The number of active keys

There needs to be at least 3 keys as a minimum. The actual number of keys is determined by the token lifespan and the rotation frequency. The max_active_keys must be one greater than the token lifespan / rotation frequency

To quote from the FAQ:

    The number of max_active_keys for a deployment can be determined by
    dividing the token lifetime, in hours, by the frequency of rotation in
    hours and adding two. Better illustrated as:

Configuring Key Lifetime

In the keystone charm, the rotation frequency is calculated automatically from the token-expiration and the fernet-max-active-keys configuration parameters. For example, with an expiration of 24 hours and 6 active keys, the rotation frequency is calculated as:

token_expiration = 24   # actually 3600, as it's in seconds
max_active_keys = 6
rotation_frequency = token_expiration / (max_active_keys - 2)

Thus, the fernet-max-active-keys can never be less than 3 (which is enforced in the charm), which would make the rotation frequency the same as the token expiration time.

NOTE: To increase the rotation frequency, either increase fernet-max-active-keys or reduce token-expiration, and, to decrease rotation frequency, do the opposite.

NOTE: If the configuration parameters are used to significantly reduce the key lifetime, then it is possible to over-rotate the verification keys such that services will hold tokens that cannot be verified but haven't yet expired. This should be avoided by only making small changes and verifying that current tokens will still be able to be verified. In particular, fernet-max-active-keys affects the rotation time.

Upgrades

When an older keystone charm is upgraded to this version, NO change will occur to the token system. That is, an ocata system will continue to use UUID tokens. In order to change the token system to Fernet, change the token-provider configuration item to fernet. This will switch the token system over. There may be a small outage in the control plane, but the running instances will be unaffected.