04d4439606
This got decided at the PTG. The code is old, not maintained, not tested and most likely doesn't work anymore. Moreover it gave us a hard dependency on grpcio and protobuf, which is fairly problematic in Python and gave us all sorts of headaches. Change-Id: I0c8c91cdd3e1284e7a3c1e9fe04b4c0fbbde7e45
468 lines
19 KiB
Python
468 lines
19 KiB
Python
# Copyright (c) 2016 Mirantis, Inc.
|
|
# All Rights Reserved.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
import contextlib
|
|
import datetime
|
|
import functools
|
|
import itertools
|
|
import os
|
|
import ssl
|
|
import time
|
|
from urllib import parse
|
|
import urllib3
|
|
|
|
from oslo_log import log as logging
|
|
from oslo_serialization import jsonutils
|
|
import pytz
|
|
import requests
|
|
from requests import adapters
|
|
|
|
from kuryr.lib._i18n import _
|
|
from kuryr_kubernetes import config
|
|
from kuryr_kubernetes import constants
|
|
from kuryr_kubernetes import exceptions as exc
|
|
from kuryr_kubernetes import utils
|
|
|
|
CONF = config.CONF
|
|
LOG = logging.getLogger(__name__)
|
|
|
|
|
|
class K8sClient(object):
|
|
# REVISIT(ivc): replace with python-k8sclient if it could be extended
|
|
# with 'WATCH' support
|
|
|
|
def __init__(self, base_url):
|
|
self._base_url = base_url
|
|
cert_file = config.CONF.kubernetes.ssl_client_crt_file
|
|
key_file = config.CONF.kubernetes.ssl_client_key_file
|
|
ca_crt_file = config.CONF.kubernetes.ssl_ca_crt_file
|
|
self.verify_server = config.CONF.kubernetes.ssl_verify_server_crt
|
|
token_file = config.CONF.kubernetes.token_file
|
|
self.token = None
|
|
self.cert = (None, None)
|
|
self.are_events_enabled = config.CONF.kubernetes.use_events
|
|
|
|
# Setting higher numbers regarding connection pools as we're running
|
|
# with max of 1000 green threads.
|
|
self.session = requests.Session()
|
|
prefix = '%s://' % parse.urlparse(base_url).scheme
|
|
self.session.mount(prefix, adapters.HTTPAdapter(pool_maxsize=1000))
|
|
if token_file:
|
|
if os.path.exists(token_file):
|
|
with open(token_file, 'r') as f:
|
|
self.token = f.readline().rstrip('\n')
|
|
else:
|
|
raise RuntimeError(
|
|
_("Unable to find token_file : %s") % token_file)
|
|
else:
|
|
if cert_file and not os.path.exists(cert_file):
|
|
raise RuntimeError(
|
|
_("Unable to find ssl cert_file : %s") % cert_file)
|
|
if key_file and not os.path.exists(key_file):
|
|
raise RuntimeError(
|
|
_("Unable to find ssl key_file : %s") % key_file)
|
|
self.cert = (cert_file, key_file)
|
|
if self.verify_server:
|
|
if not ca_crt_file:
|
|
raise RuntimeError(
|
|
_("ssl_ca_crt_file cannot be None"))
|
|
elif not os.path.exists(ca_crt_file):
|
|
raise RuntimeError(
|
|
_("Unable to find ca cert_file : %s") % ca_crt_file)
|
|
else:
|
|
self.verify_server = ca_crt_file
|
|
|
|
# Let's setup defaults for our Session.
|
|
self.session.cert = self.cert
|
|
self.session.verify = self.verify_server
|
|
if self.token:
|
|
self.session.headers['Authorization'] = f'Bearer {self.token}'
|
|
# NOTE(dulek): Seems like this is the only way to set is globally.
|
|
self.session.request = functools.partial(
|
|
self.session.request, timeout=(
|
|
CONF.kubernetes.watch_connection_timeout,
|
|
CONF.kubernetes.watch_read_timeout))
|
|
|
|
def _raise_from_response(self, response):
|
|
if response.status_code == requests.codes.not_found:
|
|
raise exc.K8sResourceNotFound(response.text)
|
|
if response.status_code == requests.codes.conflict:
|
|
raise exc.K8sConflict(response.text)
|
|
if response.status_code == requests.codes.forbidden:
|
|
if 'because it is being terminated' in response.json()['message']:
|
|
raise exc.K8sNamespaceTerminating(response.text)
|
|
raise exc.K8sForbidden(response.text)
|
|
if response.status_code == requests.codes.unprocessable_entity:
|
|
# NOTE(gryf): on k8s API code 422 is also Forbidden, but specified
|
|
# to FieldValueForbidden. Perhaps there are other usages for
|
|
# throwing unprocessable entity errors in different cases.
|
|
if ('FieldValueForbidden' in response.text and
|
|
'Forbidden' in response.json()['message']):
|
|
raise exc.K8sFieldValueForbidden(response.text)
|
|
raise exc.K8sUnprocessableEntity(response.text)
|
|
if not response.ok:
|
|
raise exc.K8sClientException(response.text)
|
|
|
|
def get(self, path, json=True, headers=None):
|
|
LOG.debug("Get %(path)s", {'path': path})
|
|
url = self._base_url + path
|
|
response = self.session.get(url, headers=headers)
|
|
self._raise_from_response(response)
|
|
|
|
if json:
|
|
result = response.json()
|
|
kind = result['kind']
|
|
|
|
api_version = result.get('apiVersion')
|
|
if not api_version:
|
|
api_version = utils.get_api_ver(path)
|
|
|
|
# Strip List from e.g. PodList. For some reason `.items` of a list
|
|
# returned from API doesn't have `kind` set.
|
|
# NOTE(gryf): Also, for the sake of calculating selfLink
|
|
# equivalent, we need to have both: kind and apiVersion, while the
|
|
# latter is not present on items list for core resources, while
|
|
# for custom resources there are both kind and apiVersion..
|
|
if kind.endswith('List'):
|
|
kind = kind[:-4]
|
|
|
|
# NOTE(gryf): In case we get null/None for items from the API,
|
|
# we need to convert it to the empty list, otherwise it might
|
|
# be propagated to the consumers of this method and sent back
|
|
# to the Kubernetes as is, and fail as a result.
|
|
if result['items'] is None:
|
|
result['items'] = []
|
|
|
|
for item in result['items']:
|
|
if not item.get('kind'):
|
|
item['kind'] = kind
|
|
if not item.get('apiVersion'):
|
|
item['apiVersion'] = api_version
|
|
|
|
if not result.get('apiVersion'):
|
|
result['apiVersion'] = api_version
|
|
else:
|
|
result = response.text
|
|
|
|
return result
|
|
|
|
def _get_url_and_header(self, path, content_type):
|
|
url = self._base_url + path
|
|
header = {'Content-Type': content_type,
|
|
'Accept': 'application/json'}
|
|
|
|
return url, header
|
|
|
|
def patch(self, field, path, data):
|
|
LOG.debug("Patch %(path)s: %(data)s", {'path': path, 'data': data})
|
|
content_type = 'application/merge-patch+json'
|
|
url, header = self._get_url_and_header(path, content_type)
|
|
response = self.session.patch(url, json={field: data}, headers=header)
|
|
self._raise_from_response(response)
|
|
return response.json().get('status')
|
|
|
|
def patch_crd(self, field, path, data, action='replace'):
|
|
content_type = 'application/json-patch+json'
|
|
url, header = self._get_url_and_header(path, content_type)
|
|
|
|
if action == 'remove':
|
|
data = [{'op': action,
|
|
'path': f'/{field}/{data}'}]
|
|
else:
|
|
if data:
|
|
data = [{'op': action,
|
|
'path': f'/{field}/{crd_field}',
|
|
'value': value}
|
|
for crd_field, value in data.items()]
|
|
else:
|
|
data = [{'op': action,
|
|
'path': f'/{field}',
|
|
'value': data}]
|
|
|
|
LOG.debug("Patch %(path)s: %(data)s", {
|
|
'path': path, 'data': data})
|
|
|
|
response = self.session.patch(url, data=jsonutils.dumps(data),
|
|
headers=header)
|
|
self._raise_from_response(response)
|
|
return response.json().get('status')
|
|
|
|
def post(self, path, body):
|
|
LOG.debug("Post %(path)s: %(body)s", {'path': path, 'body': body})
|
|
url = self._base_url + path
|
|
header = {'Content-Type': 'application/json'}
|
|
|
|
response = self.session.post(url, json=body, headers=header)
|
|
self._raise_from_response(response)
|
|
return response.json()
|
|
|
|
def delete(self, path):
|
|
LOG.debug("Delete %(path)s", {'path': path})
|
|
url = self._base_url + path
|
|
header = {'Content-Type': 'application/json'}
|
|
|
|
response = self.session.delete(url, headers=header)
|
|
self._raise_from_response(response)
|
|
return response.json()
|
|
|
|
# TODO(dulek): add_finalizer() and remove_finalizer() have some code
|
|
# duplication, but I don't see a nice way to avoid it.
|
|
def add_finalizer(self, obj, finalizer):
|
|
if finalizer in obj['metadata'].get('finalizers', []):
|
|
return True
|
|
|
|
path = utils.get_res_link(obj)
|
|
LOG.debug(f"Add finalizer {finalizer} to {path}")
|
|
url, headers = self._get_url_and_header(
|
|
path, 'application/merge-patch+json')
|
|
|
|
for i in range(3): # Let's make sure it's not infinite loop
|
|
finalizers = obj['metadata'].get('finalizers', []).copy()
|
|
finalizers.append(finalizer)
|
|
|
|
data = {
|
|
'metadata': {
|
|
'finalizers': finalizers,
|
|
'resourceVersion': obj['metadata']['resourceVersion'],
|
|
},
|
|
}
|
|
|
|
response = self.session.patch(url, json=data, headers=headers)
|
|
|
|
if response.ok:
|
|
return True
|
|
|
|
try:
|
|
self._raise_from_response(response)
|
|
except (exc.K8sFieldValueForbidden, exc.K8sResourceNotFound):
|
|
# Object is being deleting or gone. Return.
|
|
return False
|
|
except exc.K8sConflict:
|
|
try:
|
|
obj = self.get(path)
|
|
except exc.K8sResourceNotFound:
|
|
# Object got removed before finalizer was set
|
|
return False
|
|
if finalizer in obj['metadata'].get('finalizers', []):
|
|
# Finalizer is there, return.
|
|
return True
|
|
|
|
# If after 3 iterations there's still conflict, just raise.
|
|
self._raise_from_response(response)
|
|
|
|
def remove_finalizer(self, obj, finalizer):
|
|
path = utils.get_res_link(obj)
|
|
LOG.debug(f"Remove finalizer {finalizer} from {path}")
|
|
url, headers = self._get_url_and_header(
|
|
path, 'application/merge-patch+json')
|
|
|
|
for i in range(3): # Let's make sure it's not infinite loop
|
|
finalizers = obj['metadata'].get('finalizers', []).copy()
|
|
try:
|
|
finalizers.remove(finalizer)
|
|
except ValueError:
|
|
# Finalizer is not there, return.
|
|
return True
|
|
|
|
data = {
|
|
'metadata': {
|
|
'finalizers': finalizers,
|
|
'resourceVersion': obj['metadata']['resourceVersion'],
|
|
},
|
|
}
|
|
|
|
response = self.session.patch(url, json=data, headers=headers)
|
|
|
|
if response.ok:
|
|
return True
|
|
|
|
try:
|
|
try:
|
|
self._raise_from_response(response)
|
|
except exc.K8sConflict:
|
|
obj = self.get(path)
|
|
except (exc.K8sFieldValueForbidden, exc.K8sResourceNotFound):
|
|
# Object is being deleted or gone already, stop.
|
|
return False
|
|
|
|
# If after 3 iterations there's still conflict, just raise.
|
|
self._raise_from_response(response)
|
|
|
|
def get_loadbalancer_crd(self, obj):
|
|
name = obj['metadata']['name']
|
|
namespace = obj['metadata']['namespace']
|
|
|
|
try:
|
|
crd = self.get('{}/{}/kuryrloadbalancers/{}'.format(
|
|
constants.K8S_API_CRD_NAMESPACES, namespace,
|
|
name))
|
|
except exc.K8sResourceNotFound:
|
|
return None
|
|
except exc.K8sClientException:
|
|
LOG.exception("Kubernetes Client Exception.")
|
|
raise
|
|
return crd
|
|
|
|
def annotate(self, path, annotations, resource_version=None):
|
|
"""Pushes a resource annotation to the K8s API resource
|
|
|
|
The annotate operation is made with a PATCH HTTP request of kind:
|
|
application/merge-patch+json as described in:
|
|
|
|
https://github.com/kubernetes/community/blob/master/contributors/devel/sig-architecture/api-conventions.md#patch-operations # noqa
|
|
"""
|
|
LOG.debug("Annotate %(path)s: %(names)s", {
|
|
'path': path, 'names': list(annotations)})
|
|
|
|
content_type = 'application/merge-patch+json'
|
|
url, header = self._get_url_and_header(path, content_type)
|
|
|
|
while itertools.count(1):
|
|
metadata = {"annotations": annotations}
|
|
if resource_version:
|
|
metadata['resourceVersion'] = resource_version
|
|
data = jsonutils.dumps({"metadata": metadata}, sort_keys=True)
|
|
response = self.session.patch(url, data=data, headers=header)
|
|
if response.ok:
|
|
return response.json()['metadata'].get('annotations', {})
|
|
if response.status_code == requests.codes.conflict:
|
|
resource = self.get(path)
|
|
new_version = resource['metadata']['resourceVersion']
|
|
retrieved_annotations = resource['metadata'].get(
|
|
'annotations', {})
|
|
|
|
for k, v in annotations.items():
|
|
if v != retrieved_annotations.get(k):
|
|
break
|
|
else:
|
|
LOG.debug("Annotations for %(path)s already present: "
|
|
"%(names)s", {'path': path,
|
|
'names': retrieved_annotations})
|
|
return retrieved_annotations
|
|
# Retry patching with updated resourceVersion
|
|
resource_version = new_version
|
|
continue
|
|
|
|
LOG.error("Exception response, headers: %(headers)s, "
|
|
"content: %(content)s, text: %(text)s"
|
|
% {'headers': response.headers,
|
|
'content': response.content, 'text': response.text})
|
|
|
|
self._raise_from_response(response)
|
|
|
|
def watch(self, path):
|
|
url = self._base_url + path
|
|
resource_version = None
|
|
|
|
attempt = 0
|
|
while True:
|
|
try:
|
|
params = {'watch': 'true'}
|
|
if resource_version:
|
|
params['resourceVersion'] = resource_version
|
|
with contextlib.closing(
|
|
self.session.get(
|
|
url, params=params, stream=True)) as response:
|
|
if not response.ok:
|
|
raise exc.K8sClientException(response.text)
|
|
attempt = 0
|
|
for line in response.iter_lines():
|
|
line = line.decode('utf-8').strip()
|
|
if line:
|
|
line_dict = jsonutils.loads(line)
|
|
yield line_dict
|
|
# Saving the resourceVersion in case of a restart.
|
|
# At this point it's safely passed to handler.
|
|
m = line_dict.get('object', {}).get('metadata', {})
|
|
resource_version = m.get('resourceVersion', None)
|
|
except (requests.ReadTimeout, requests.ConnectionError,
|
|
ssl.SSLError, requests.exceptions.ChunkedEncodingError,
|
|
urllib3.exceptions.SSLError):
|
|
t = utils.exponential_backoff(attempt)
|
|
log = LOG.debug
|
|
if attempt > 0:
|
|
# Only make it a warning if it's happening again, no need
|
|
# to inform about all the read timeouts.
|
|
log = LOG.warning
|
|
log('Connection error when watching %s. Retrying in %ds with '
|
|
'resourceVersion=%s', path, t,
|
|
params.get('resourceVersion'))
|
|
time.sleep(t)
|
|
attempt += 1
|
|
|
|
def add_event(self, resource, reason, message, type_='Normal',
|
|
component='kuryr-controller'):
|
|
"""Create an Event object for the provided resource."""
|
|
if not self.are_events_enabled:
|
|
return {}
|
|
|
|
if not resource:
|
|
return {}
|
|
|
|
involved_object = {'apiVersion': resource['apiVersion'],
|
|
'kind': resource['kind'],
|
|
'name': resource['metadata']['name'],
|
|
'namespace': resource['metadata']['namespace'],
|
|
'uid': resource['metadata']['uid']}
|
|
|
|
# This is needed for Event date, otherwise LAST SEEN/Age will be empty
|
|
# and misleading.
|
|
now = datetime.datetime.utcnow().replace(tzinfo=pytz.UTC)
|
|
date_time = now.strftime("%Y-%m-%dT%H:%M:%SZ")
|
|
|
|
name = ".".join((resource['metadata']['name'],
|
|
self._get_hex_timestamp(now)))
|
|
|
|
event = {'kind': 'Event',
|
|
'apiVersion': 'v1',
|
|
'firstTimestamp': date_time,
|
|
'metadata': {'name': name},
|
|
'reason': reason,
|
|
'message': message,
|
|
'type': type_,
|
|
'involvedObject': involved_object,
|
|
'source': {'component': component,
|
|
'host': utils.get_nodename()}}
|
|
|
|
try:
|
|
return self.post(f'{constants.K8S_API_BASE}/namespaces/'
|
|
f'{resource["metadata"]["namespace"]}/events',
|
|
event)
|
|
except exc.K8sNamespaceTerminating:
|
|
# We can't create events in a Namespace that is being terminated,
|
|
# there's no workaround, no need to log it, just ignore it.
|
|
return {}
|
|
except exc.K8sClientException:
|
|
LOG.warning(f'There was non critical error during creating an '
|
|
'Event for resource: "{resource}", with reason: '
|
|
f'"{reason}", message: "{message}" and type: '
|
|
f'"{type_}"')
|
|
return {}
|
|
|
|
def _get_hex_timestamp(self, datetimeobj):
|
|
"""Get hex representation for timestamp.
|
|
|
|
In Kuberenets, Event name is constructed name of the bounded object
|
|
and timestamp in hexadecimal representation.
|
|
Note, that Python timestamp is represented as floating figure:
|
|
1631622163.8534190654754638671875
|
|
while those which origin from K8s, after change to int:
|
|
1631622163915909162
|
|
so, to get similar integer, we need to multiply the float by
|
|
100000000 to get the same precision and cast to integer, to get rid
|
|
of the fractures, and finally convert it to hex representation.
|
|
"""
|
|
timestamp = datetime.datetime.timestamp(datetimeobj)
|
|
return format(int(timestamp * 100000000), 'x')
|