# Copyright 2018 Catalyst Cloud Ltd. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re import time from osc_lib import exceptions as osc_exc from oslo_config import cfg from oslo_log import log as logging from magnum.common import clients from magnum.common import context as magnum_context from magnum.common import exception from magnum.common import neutron LOG = logging.getLogger(__name__) CONF = cfg.CONF def wait_for_lb_deleted(octavia_client, deleted_lbs): """Wait for the loadbalancers to be deleted. Load balancer deletion API in Octavia is asynchronous so that the called needs to wait if it wants to guarantee the load balancer to be deleted. The timeout is necessary to avoid waiting infinitely. """ timeout = CONF.cluster.pre_delete_lb_timeout start_time = time.time() while True: lbs = octavia_client.load_balancer_list().get("loadbalancers", []) lbIDs = set( [lb["id"] for lb in lbs if lb["provisioning_status"] != "DELETED"] ) if not (deleted_lbs & lbIDs): break if (time.time() - timeout) > start_time: raise Exception("Timeout waiting for the load balancers " "%s to be deleted." % deleted_lbs) time.sleep(1) def _delete_loadbalancers(context, lbs, cluster, octavia_client, remove_fip=False, cascade=True): candidates = set() for lb in lbs: status = lb["provisioning_status"] if status not in ["PENDING_DELETE", "DELETED"]: LOG.info("Deleting load balancer %s for cluster %s", lb["id"], cluster.uuid) octavia_client.load_balancer_delete(lb["id"], cascade=cascade) candidates.add(lb["id"]) if remove_fip: neutron.delete_floatingip(context, lb["vip_port_id"], cluster) return candidates def delete_loadbalancers(context, cluster): """Delete loadbalancers for the cluster. The following load balancers are deleted: - The load balancers created for Kubernetes services and ingresses in the Kubernetes cluster. - The load balancers created for Kubernetes API and etcd for HA cluster. """ pattern = (r'Kubernetes .+ from cluster %s' % cluster.uuid) lb_resource_type = "Magnum::Optional::Neutron::LBaaS::LoadBalancer" adm_ctx = magnum_context.get_admin_context() adm_clients = clients.OpenStackClients(adm_ctx) user_clients = clients.OpenStackClients(context) candidates = set() try: octavia_client_adm = adm_clients.octavia() heat_client = user_clients.heat() octavia_client = user_clients.octavia() # Get load balancers created for service/ingress lbs = octavia_client.load_balancer_list().get("loadbalancers", []) lbs = [lb for lb in lbs if re.match(pattern, lb["description"])] deleted = _delete_loadbalancers(context, lbs, cluster, octavia_client_adm, remove_fip=True) candidates.update(deleted) # NOTE (brtknr): If stack has been deleted, cluster fails to delete # because stack_id resolves to None. Return if that is the case. if not cluster.stack_id: return # Get load balancers created for Kubernetes api/etcd lbs = [] lb_resources = heat_client.resources.list( cluster.stack_id, nested_depth=2, filters={"type": lb_resource_type}) for lb_res in lb_resources: lb_id = lb_res.physical_resource_id if not lb_id: continue try: lb = octavia_client.load_balancer_show(lb_id) lbs.append(lb) except osc_exc.NotFound: continue deleted = _delete_loadbalancers(context, lbs, cluster, octavia_client_adm, remove_fip=False) candidates.update(deleted) if not candidates: return wait_for_lb_deleted(octavia_client, candidates) except Exception as e: raise exception.PreDeletionFailed(cluster_uuid=cluster.uuid, msg=str(e))