|
@@ -14,20 +14,35 @@ except ImportError:
|
|
|
# python3
|
|
|
import configparser as ConfigParser
|
|
|
|
|
|
+# pylint: disable=no-name-in-module, import-error, wrong-import-order
|
|
|
import copy
|
|
|
+import errno
|
|
|
+import json
|
|
|
+import re
|
|
|
import io
|
|
|
import os
|
|
|
import yaml
|
|
|
-from distutils.util import strtobool
|
|
|
-from distutils.version import LooseVersion
|
|
|
import struct
|
|
|
import socket
|
|
|
+from distutils.util import strtobool
|
|
|
+from distutils.version import LooseVersion
|
|
|
+
|
|
|
+# ignore pylint errors related to the module_utils import
|
|
|
+# pylint: disable=redefined-builtin, unused-wildcard-import, wildcard-import
|
|
|
+# import module snippets
|
|
|
+from ansible.module_utils.basic import * # noqa: F403
|
|
|
+from ansible.module_utils.facts import * # noqa: F403
|
|
|
+from ansible.module_utils.urls import * # noqa: F403
|
|
|
+from ansible.module_utils.six import iteritems, itervalues
|
|
|
+from ansible.module_utils.six.moves.urllib.parse import urlparse, urlunparse
|
|
|
+from ansible.module_utils._text import to_native
|
|
|
+
|
|
|
+HAVE_DBUS = False
|
|
|
|
|
|
-HAVE_DBUS=False
|
|
|
try:
|
|
|
from dbus import SystemBus, Interface
|
|
|
from dbus.exceptions import DBusException
|
|
|
- HAVE_DBUS=True
|
|
|
+ HAVE_DBUS = True
|
|
|
except ImportError:
|
|
|
pass
|
|
|
|
|
@@ -58,6 +73,7 @@ def migrate_docker_facts(facts):
|
|
|
}
|
|
|
if 'docker' not in facts:
|
|
|
facts['docker'] = {}
|
|
|
+ # pylint: disable=consider-iterating-dictionary
|
|
|
for role in params.keys():
|
|
|
if role in facts:
|
|
|
for param in params[role]:
|
|
@@ -76,6 +92,7 @@ def migrate_docker_facts(facts):
|
|
|
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
# TODO: We should add a generic migration function that takes source and destination
|
|
|
# paths and does the right thing rather than one function for common, one for node, etc.
|
|
|
def migrate_common_facts(facts):
|
|
@@ -86,6 +103,7 @@ def migrate_common_facts(facts):
|
|
|
}
|
|
|
if 'common' not in facts:
|
|
|
facts['common'] = {}
|
|
|
+ # pylint: disable=consider-iterating-dictionary
|
|
|
for role in params.keys():
|
|
|
if role in facts:
|
|
|
for param in params[role]:
|
|
@@ -93,6 +111,7 @@ def migrate_common_facts(facts):
|
|
|
facts['common'][param] = facts[role].pop(param)
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def migrate_node_facts(facts):
|
|
|
""" Migrate facts from various roles into node """
|
|
|
params = {
|
|
@@ -100,6 +119,7 @@ def migrate_node_facts(facts):
|
|
|
}
|
|
|
if 'node' not in facts:
|
|
|
facts['node'] = {}
|
|
|
+ # pylint: disable=consider-iterating-dictionary
|
|
|
for role in params.keys():
|
|
|
if role in facts:
|
|
|
for param in params[role]:
|
|
@@ -125,7 +145,9 @@ def migrate_hosted_facts(facts):
|
|
|
facts['hosted']['registry']['selector'] = facts['master'].pop('registry_selector')
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def migrate_admission_plugin_facts(facts):
|
|
|
+ """ Apply migrations for admission plugin facts """
|
|
|
if 'master' in facts:
|
|
|
if 'kube_admission_plugin_config' in facts['master']:
|
|
|
if 'admission_plugin_config' not in facts['master']:
|
|
@@ -139,6 +161,7 @@ def migrate_admission_plugin_facts(facts):
|
|
|
facts['master'].pop('kube_admission_plugin_config', None)
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def migrate_local_facts(facts):
|
|
|
""" Apply migrations of local facts """
|
|
|
migrated_facts = copy.deepcopy(facts)
|
|
@@ -149,6 +172,7 @@ def migrate_local_facts(facts):
|
|
|
migrated_facts = migrate_admission_plugin_facts(migrated_facts)
|
|
|
return migrated_facts
|
|
|
|
|
|
+
|
|
|
def first_ip(network):
|
|
|
""" Return the first IPv4 address in network
|
|
|
|
|
@@ -157,13 +181,14 @@ def first_ip(network):
|
|
|
Returns:
|
|
|
str: first IPv4 address
|
|
|
"""
|
|
|
- atoi = lambda addr: struct.unpack("!I", socket.inet_aton(addr))[0]
|
|
|
- itoa = lambda addr: socket.inet_ntoa(struct.pack("!I", addr))
|
|
|
+ atoi = lambda addr: struct.unpack("!I", socket.inet_aton(addr))[0] # noqa: E731
|
|
|
+ itoa = lambda addr: socket.inet_ntoa(struct.pack("!I", addr)) # noqa: E731
|
|
|
|
|
|
(address, netmask) = network.split('/')
|
|
|
netmask_i = (0xffffffff << (32 - atoi(netmask))) & 0xffffffff
|
|
|
return itoa((atoi(address) & netmask_i) + 1)
|
|
|
|
|
|
+
|
|
|
def hostname_valid(hostname):
|
|
|
""" Test if specified hostname should be considered valid
|
|
|
|
|
@@ -201,11 +226,8 @@ def choose_hostname(hostnames=None, fallback=''):
|
|
|
return hostname
|
|
|
|
|
|
ip_regex = r'\A\d{1,3}\.\d{1,3}\.\d{1,3}\.\d{1,3}\Z'
|
|
|
- ips = [i for i in hostnames
|
|
|
- if (i is not None and isinstance(i, basestring)
|
|
|
- and re.match(ip_regex, i))]
|
|
|
- hosts = [i for i in hostnames
|
|
|
- if i is not None and i != '' and i not in ips]
|
|
|
+ ips = [i for i in hostnames if i is not None and isinstance(i, basestring) and re.match(ip_regex, i)]
|
|
|
+ hosts = [i for i in hostnames if i is not None and i != '' and i not in ips]
|
|
|
|
|
|
for host_list in (hosts, ips):
|
|
|
for host in host_list:
|
|
@@ -225,11 +247,11 @@ def query_metadata(metadata_url, headers=None, expect_json=False):
|
|
|
Returns:
|
|
|
dict or list: metadata request result
|
|
|
"""
|
|
|
- result, info = fetch_url(module, metadata_url, headers=headers)
|
|
|
+ result, info = fetch_url(module, metadata_url, headers=headers) # noqa: F405
|
|
|
if info['status'] != 200:
|
|
|
raise OpenShiftFactsMetadataUnavailableError("Metadata unavailable")
|
|
|
if expect_json:
|
|
|
- return module.from_json(to_native(result.read()))
|
|
|
+ return module.from_json(to_native(result.read())) # noqa: F405
|
|
|
else:
|
|
|
return [to_native(line.strip()) for line in result.readlines()]
|
|
|
|
|
@@ -390,7 +412,7 @@ def normalize_openstack_facts(metadata, facts):
|
|
|
facts['network']['ip'] = local_ipv4
|
|
|
facts['network']['public_ip'] = metadata['ec2_compat']['public-ipv4']
|
|
|
|
|
|
- for f_var, h_var, ip_var in [('hostname', 'hostname', 'local-ipv4'),
|
|
|
+ for f_var, h_var, ip_var in [('hostname', 'hostname', 'local-ipv4'),
|
|
|
('public_hostname', 'public-hostname', 'public-ipv4')]:
|
|
|
try:
|
|
|
if socket.gethostbyname(metadata['ec2_compat'][h_var]) == metadata['ec2_compat'][ip_var]:
|
|
@@ -431,6 +453,7 @@ def normalize_provider_facts(provider, metadata):
|
|
|
facts = normalize_openstack_facts(metadata, facts)
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def set_flannel_facts_if_unset(facts):
|
|
|
""" Set flannel facts if not already present in facts dict
|
|
|
dict: the facts dict updated with the flannel facts if
|
|
@@ -448,6 +471,7 @@ def set_flannel_facts_if_unset(facts):
|
|
|
facts['common']['use_flannel'] = use_flannel
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def set_nuage_facts_if_unset(facts):
|
|
|
""" Set nuage facts if not already present in facts dict
|
|
|
dict: the facts dict updated with the nuage facts if
|
|
@@ -465,6 +489,7 @@ def set_nuage_facts_if_unset(facts):
|
|
|
facts['common']['use_nuage'] = use_nuage
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def set_node_schedulability(facts):
|
|
|
""" Set schedulable facts if not already present in facts dict
|
|
|
Args:
|
|
@@ -482,6 +507,7 @@ def set_node_schedulability(facts):
|
|
|
facts['node']['schedulable'] = True
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def set_selectors(facts):
|
|
|
""" Set selectors facts if not already present in facts dict
|
|
|
Args:
|
|
@@ -518,6 +544,7 @@ def set_selectors(facts):
|
|
|
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def set_dnsmasq_facts_if_unset(facts):
|
|
|
""" Set dnsmasq facts if not already present in facts
|
|
|
Args:
|
|
@@ -537,6 +564,7 @@ def set_dnsmasq_facts_if_unset(facts):
|
|
|
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def set_project_cfg_facts_if_unset(facts):
|
|
|
""" Set Project Configuration facts if not already present in facts dict
|
|
|
dict:
|
|
@@ -564,6 +592,7 @@ def set_project_cfg_facts_if_unset(facts):
|
|
|
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def set_identity_providers_if_unset(facts):
|
|
|
""" Set identity_providers fact if not already present in facts dict
|
|
|
|
|
@@ -590,6 +619,7 @@ def set_identity_providers_if_unset(facts):
|
|
|
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def set_url_facts_if_unset(facts):
|
|
|
""" Set url facts if not already present in facts dict
|
|
|
|
|
@@ -649,7 +679,6 @@ def set_url_facts_if_unset(facts):
|
|
|
host,
|
|
|
ports[prefix]))
|
|
|
|
|
|
-
|
|
|
r_lhn = "{0}:{1}".format(hostname, ports['api']).replace('.', '-')
|
|
|
r_lhu = "system:openshift-master/{0}:{1}".format(api_hostname, ports['api']).replace('.', '-')
|
|
|
facts['master'].setdefault('loopback_cluster_name', r_lhn)
|
|
@@ -665,6 +694,7 @@ def set_url_facts_if_unset(facts):
|
|
|
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def set_aggregate_facts(facts):
|
|
|
""" Set aggregate facts
|
|
|
|
|
@@ -760,6 +790,7 @@ def set_etcd_facts_if_unset(facts):
|
|
|
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def set_deployment_facts_if_unset(facts):
|
|
|
""" Set Facts that vary based on deployment_type. This currently
|
|
|
includes common.service_type, common.config_base, master.registry_url,
|
|
@@ -840,6 +871,7 @@ def set_deployment_facts_if_unset(facts):
|
|
|
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def set_version_facts_if_unset(facts):
|
|
|
""" Set version facts. This currently includes common.version and
|
|
|
common.version_gte_3_1_or_1_1.
|
|
@@ -878,7 +910,6 @@ def set_version_facts_if_unset(facts):
|
|
|
facts['common']['version_gte_3_3_or_1_3'] = version_gte_3_3_or_1_3
|
|
|
facts['common']['version_gte_3_4_or_1_4'] = version_gte_3_4_or_1_4
|
|
|
|
|
|
-
|
|
|
if version_gte_3_4_or_1_4:
|
|
|
examples_content_version = 'v1.4'
|
|
|
elif version_gte_3_3_or_1_3:
|
|
@@ -894,6 +925,7 @@ def set_version_facts_if_unset(facts):
|
|
|
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def set_manageiq_facts_if_unset(facts):
|
|
|
""" Set manageiq facts. This currently includes common.use_manageiq.
|
|
|
|
|
@@ -914,6 +946,7 @@ def set_manageiq_facts_if_unset(facts):
|
|
|
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def set_sdn_facts_if_unset(facts, system_facts):
|
|
|
""" Set sdn facts if not already present in facts dict
|
|
|
|
|
@@ -924,6 +957,7 @@ def set_sdn_facts_if_unset(facts, system_facts):
|
|
|
dict: the facts dict updated with the generated sdn facts if they
|
|
|
were not already present
|
|
|
"""
|
|
|
+ # pylint: disable=too-many-branches
|
|
|
if 'common' in facts:
|
|
|
use_sdn = facts['common']['use_openshift_sdn']
|
|
|
if not (use_sdn == '' or isinstance(use_sdn, bool)):
|
|
@@ -973,7 +1007,9 @@ def set_sdn_facts_if_unset(facts, system_facts):
|
|
|
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def set_nodename(facts):
|
|
|
+ """ set nodename """
|
|
|
if 'node' in facts and 'common' in facts:
|
|
|
if 'cloudprovider' in facts and facts['cloudprovider']['kind'] == 'openstack':
|
|
|
facts['node']['nodename'] = facts['provider']['metadata']['hostname'].replace('.novalocal', '')
|
|
@@ -981,6 +1017,7 @@ def set_nodename(facts):
|
|
|
facts['node']['nodename'] = facts['common']['hostname'].lower()
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def migrate_oauth_template_facts(facts):
|
|
|
"""
|
|
|
Migrate an old oauth template fact to a newer format if it's present.
|
|
@@ -1000,6 +1037,7 @@ def migrate_oauth_template_facts(facts):
|
|
|
facts['master']['oauth_templates']['login'] = facts['master']['oauth_template']
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def format_url(use_ssl, hostname, port, path=''):
|
|
|
""" Format url based on ssl flag, hostname, port and path
|
|
|
|
|
@@ -1022,6 +1060,7 @@ def format_url(use_ssl, hostname, port, path=''):
|
|
|
url = urlunparse((scheme, netloc, path, '', '', ''))
|
|
|
return url
|
|
|
|
|
|
+
|
|
|
def get_current_config(facts):
|
|
|
""" Get current openshift config
|
|
|
|
|
@@ -1051,10 +1090,9 @@ def get_current_config(facts):
|
|
|
)
|
|
|
|
|
|
kubeconfig_path = os.path.join(kubeconfig_dir, '.kubeconfig')
|
|
|
- if (os.path.isfile('/usr/bin/openshift')
|
|
|
- and os.path.isfile(kubeconfig_path)):
|
|
|
+ if os.path.isfile('/usr/bin/openshift') and os.path.isfile(kubeconfig_path):
|
|
|
try:
|
|
|
- _, output, _ = module.run_command(
|
|
|
+ _, output, _ = module.run_command( # noqa: F405
|
|
|
["/usr/bin/openshift", "ex", "config", "view", "-o",
|
|
|
"json", "--kubeconfig=%s" % kubeconfig_path],
|
|
|
check_rc=False
|
|
@@ -1085,6 +1123,7 @@ def get_current_config(facts):
|
|
|
|
|
|
return current_config
|
|
|
|
|
|
+
|
|
|
def build_kubelet_args(facts):
|
|
|
"""Build node kubelet_args
|
|
|
|
|
@@ -1129,6 +1168,7 @@ values provided as a list. Hence the gratuitous use of ['foo'] below.
|
|
|
#
|
|
|
# map() seems to be returning an itertools.imap object
|
|
|
# instead of a list. We cast it to a list ourselves.
|
|
|
+ # pylint: disable=unnecessary-lambda
|
|
|
labels_str = list(map(lambda x: '='.join(x), facts['node']['labels'].items()))
|
|
|
if labels_str != '':
|
|
|
kubelet_args['node-labels'] = labels_str
|
|
@@ -1139,6 +1179,7 @@ values provided as a list. Hence the gratuitous use of ['foo'] below.
|
|
|
facts = merge_facts({'node': {'kubelet_args': kubelet_args}}, facts, [], [])
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def build_controller_args(facts):
|
|
|
""" Build master controller_args """
|
|
|
cloud_cfg_path = os.path.join(facts['common']['config_base'],
|
|
@@ -1160,6 +1201,7 @@ def build_controller_args(facts):
|
|
|
facts = merge_facts({'master': {'controller_args': controller_args}}, facts, [], [])
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def build_api_server_args(facts):
|
|
|
""" Build master api_server_args """
|
|
|
cloud_cfg_path = os.path.join(facts['common']['config_base'],
|
|
@@ -1181,6 +1223,7 @@ def build_api_server_args(facts):
|
|
|
facts = merge_facts({'master': {'api_server_args': api_server_args}}, facts, [], [])
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def is_service_running(service):
|
|
|
""" Queries systemd through dbus to see if the service is running """
|
|
|
service_running = False
|
|
@@ -1200,6 +1243,7 @@ def is_service_running(service):
|
|
|
|
|
|
return service_running
|
|
|
|
|
|
+
|
|
|
def get_version_output(binary, version_cmd):
|
|
|
""" runs and returns the version output for a command """
|
|
|
cmd = []
|
|
@@ -1210,9 +1254,10 @@ def get_version_output(binary, version_cmd):
|
|
|
cmd.append(item)
|
|
|
|
|
|
if os.path.isfile(cmd[0]):
|
|
|
- _, output, _ = module.run_command(cmd)
|
|
|
+ _, output, _ = module.run_command(cmd) # noqa: F405
|
|
|
return output
|
|
|
|
|
|
+
|
|
|
def get_docker_version_info():
|
|
|
""" Parses and returns the docker version info """
|
|
|
result = None
|
|
@@ -1225,6 +1270,7 @@ def get_docker_version_info():
|
|
|
}
|
|
|
return result
|
|
|
|
|
|
+
|
|
|
def get_hosted_registry_insecure():
|
|
|
""" Parses OPTIONS from /etc/sysconfig/docker to determine if the
|
|
|
registry is currently insecure.
|
|
@@ -1239,10 +1285,12 @@ def get_hosted_registry_insecure():
|
|
|
options = config.get('root', 'OPTIONS')
|
|
|
if 'insecure-registry' in options:
|
|
|
hosted_registry_insecure = True
|
|
|
+ # pylint: disable=bare-except
|
|
|
except:
|
|
|
pass
|
|
|
return hosted_registry_insecure
|
|
|
|
|
|
+
|
|
|
def get_openshift_version(facts):
|
|
|
""" Get current version of openshift on the host.
|
|
|
|
|
@@ -1264,7 +1312,7 @@ def get_openshift_version(facts):
|
|
|
return chomp_commit_offset(facts['common']['version'])
|
|
|
|
|
|
if os.path.isfile('/usr/bin/openshift'):
|
|
|
- _, output, _ = module.run_command(['/usr/bin/openshift', 'version'])
|
|
|
+ _, output, _ = module.run_command(['/usr/bin/openshift', 'version']) # noqa: F405
|
|
|
version = parse_openshift_version(output)
|
|
|
elif 'common' in facts and 'is_containerized' in facts['common']:
|
|
|
version = get_container_openshift_version(facts)
|
|
@@ -1273,7 +1321,7 @@ def get_openshift_version(facts):
|
|
|
# This can be very slow and may get re-run multiple times, so we only use this
|
|
|
# if other methods failed to find a version.
|
|
|
if not version and os.path.isfile('/usr/local/bin/openshift'):
|
|
|
- _, output, _ = module.run_command(['/usr/local/bin/openshift', 'version'])
|
|
|
+ _, output, _ = module.run_command(['/usr/local/bin/openshift', 'version']) # noqa: F405
|
|
|
version = parse_openshift_version(output)
|
|
|
|
|
|
return chomp_commit_offset(version)
|
|
@@ -1363,9 +1411,10 @@ def apply_provider_facts(facts, provider_facts):
|
|
|
facts['provider'] = provider_facts
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
# Disabling pylint too many branches. This function needs refactored
|
|
|
# but is a very core part of openshift_facts.
|
|
|
-# pylint: disable=too-many-branches
|
|
|
+# pylint: disable=too-many-branches, too-many-nested-blocks
|
|
|
def merge_facts(orig, new, additive_facts_to_overwrite, protected_facts_to_overwrite):
|
|
|
""" Recursively merge facts dicts
|
|
|
|
|
@@ -1438,12 +1487,14 @@ def merge_facts(orig, new, additive_facts_to_overwrite, protected_facts_to_overw
|
|
|
if int(value) <= int(new[key]):
|
|
|
facts[key] = copy.deepcopy(new[key])
|
|
|
else:
|
|
|
- module.fail_json(msg='openshift_facts received a lower value for openshift.master.master_count')
|
|
|
+ # pylint: disable=line-too-long
|
|
|
+ module.fail_json(msg='openshift_facts received a lower value for openshift.master.master_count') # noqa: F405
|
|
|
# ha (bool) can not change unless it has been passed
|
|
|
# as a protected fact to overwrite.
|
|
|
if key == 'ha':
|
|
|
if safe_get_bool(value) != safe_get_bool(new[key]):
|
|
|
- module.fail_json(msg='openshift_facts received a different value for openshift.master.ha')
|
|
|
+ # pylint: disable=line-too-long
|
|
|
+ module.fail_json(msg='openshift_facts received a different value for openshift.master.ha') # noqa: F405
|
|
|
else:
|
|
|
facts[key] = value
|
|
|
# No other condition has been met. Overwrite the old fact
|
|
@@ -1463,6 +1514,7 @@ def merge_facts(orig, new, additive_facts_to_overwrite, protected_facts_to_overw
|
|
|
facts[key] = copy.deepcopy(new[key])
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def save_local_facts(filename, facts):
|
|
|
""" Save local facts
|
|
|
|
|
@@ -1478,7 +1530,7 @@ def save_local_facts(filename, facts):
|
|
|
if exception.errno != errno.EEXIST: # but it is okay if it is already there
|
|
|
raise # pass any other exceptions up the chain
|
|
|
with open(filename, 'w') as fact_file:
|
|
|
- fact_file.write(module.jsonify(facts))
|
|
|
+ fact_file.write(module.jsonify(facts)) # noqa: F405
|
|
|
os.chmod(filename, 0o600)
|
|
|
except (IOError, OSError) as ex:
|
|
|
raise OpenShiftFactsFileWriteError(
|
|
@@ -1514,6 +1566,7 @@ def get_local_facts_from_file(filename):
|
|
|
|
|
|
return local_facts
|
|
|
|
|
|
+
|
|
|
def sort_unique(alist):
|
|
|
""" Sorts and de-dupes a list
|
|
|
|
|
@@ -1531,6 +1584,7 @@ def sort_unique(alist):
|
|
|
|
|
|
return out
|
|
|
|
|
|
+
|
|
|
def safe_get_bool(fact):
|
|
|
""" Get a boolean fact safely.
|
|
|
|
|
@@ -1541,6 +1595,7 @@ def safe_get_bool(fact):
|
|
|
"""
|
|
|
return bool(strtobool(str(fact)))
|
|
|
|
|
|
+
|
|
|
def set_proxy_facts(facts):
|
|
|
""" Set global proxy facts and promote defaults from http_proxy, https_proxy,
|
|
|
no_proxy to the more specific builddefaults and builddefaults_git vars.
|
|
@@ -1556,13 +1611,11 @@ def set_proxy_facts(facts):
|
|
|
if 'common' in facts:
|
|
|
common = facts['common']
|
|
|
if 'http_proxy' in common or 'https_proxy' in common:
|
|
|
- if 'no_proxy' in common and \
|
|
|
- isinstance(common['no_proxy'], basestring):
|
|
|
+ if 'no_proxy' in common and isinstance(common['no_proxy'], basestring):
|
|
|
common['no_proxy'] = common['no_proxy'].split(",")
|
|
|
elif 'no_proxy' not in common:
|
|
|
common['no_proxy'] = []
|
|
|
- if 'generate_no_proxy_hosts' in common and \
|
|
|
- safe_get_bool(common['generate_no_proxy_hosts']):
|
|
|
+ if 'generate_no_proxy_hosts' in common and safe_get_bool(common['generate_no_proxy_hosts']):
|
|
|
if 'no_proxy_internal_hostnames' in common:
|
|
|
common['no_proxy'].extend(common['no_proxy_internal_hostnames'].split(','))
|
|
|
# We always add local dns domain and ourselves no matter what
|
|
@@ -1593,13 +1646,14 @@ def set_proxy_facts(facts):
|
|
|
# into admission_plugin_config
|
|
|
if 'admission_plugin_config' not in facts['master']:
|
|
|
facts['master']['admission_plugin_config'] = dict()
|
|
|
- if 'config' in builddefaults and ('http_proxy' in builddefaults or \
|
|
|
- 'https_proxy' in builddefaults):
|
|
|
+ if 'config' in builddefaults and ('http_proxy' in builddefaults or
|
|
|
+ 'https_proxy' in builddefaults):
|
|
|
facts['master']['admission_plugin_config'].update(builddefaults['config'])
|
|
|
facts['builddefaults'] = builddefaults
|
|
|
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
# pylint: disable=too-many-statements
|
|
|
def set_container_facts_if_unset(facts):
|
|
|
""" Set containerized facts.
|
|
@@ -1671,6 +1725,7 @@ def set_container_facts_if_unset(facts):
|
|
|
|
|
|
return facts
|
|
|
|
|
|
+
|
|
|
def set_installed_variant_rpm_facts(facts):
|
|
|
""" Set RPM facts of installed variant
|
|
|
Args:
|
|
@@ -1685,7 +1740,7 @@ def set_installed_variant_rpm_facts(facts):
|
|
|
['{0}-{1}'.format(base_rpm, r) for r in optional_rpms] + \
|
|
|
['tuned-profiles-%s-node' % base_rpm]
|
|
|
for rpm in variant_rpms:
|
|
|
- exit_code, _, _ = module.run_command(['rpm', '-q', rpm])
|
|
|
+ exit_code, _, _ = module.run_command(['rpm', '-q', rpm]) # noqa: F405
|
|
|
if exit_code == 0:
|
|
|
installed_rpms.append(rpm)
|
|
|
|
|
@@ -1693,7 +1748,6 @@ def set_installed_variant_rpm_facts(facts):
|
|
|
return facts
|
|
|
|
|
|
|
|
|
-
|
|
|
class OpenShiftFactsInternalError(Exception):
|
|
|
"""Origin Facts Error"""
|
|
|
pass
|
|
@@ -1761,12 +1815,12 @@ class OpenShiftFacts(object):
|
|
|
try:
|
|
|
# ansible-2.1
|
|
|
# pylint: disable=too-many-function-args,invalid-name
|
|
|
- self.system_facts = ansible_facts(module, ['hardware', 'network', 'virtual', 'facter'])
|
|
|
+ self.system_facts = ansible_facts(module, ['hardware', 'network', 'virtual', 'facter']) # noqa: F405
|
|
|
for (k, v) in self.system_facts.items():
|
|
|
self.system_facts["ansible_%s" % k.replace('-', '_')] = v
|
|
|
except UnboundLocalError:
|
|
|
# ansible-2.2
|
|
|
- self.system_facts = get_all_facts(module)['ansible_facts']
|
|
|
+ self.system_facts = get_all_facts(module)['ansible_facts'] # noqa: F405
|
|
|
|
|
|
self.facts = self.generate_facts(local_facts,
|
|
|
additive_facts_to_overwrite,
|
|
@@ -1799,7 +1853,6 @@ class OpenShiftFacts(object):
|
|
|
protected_facts_to_overwrite)
|
|
|
roles = local_facts.keys()
|
|
|
|
|
|
-
|
|
|
if 'common' in local_facts and 'deployment_type' in local_facts['common']:
|
|
|
deployment_type = local_facts['common']['deployment_type']
|
|
|
else:
|
|
@@ -1854,7 +1907,7 @@ class OpenShiftFacts(object):
|
|
|
"""
|
|
|
defaults = {}
|
|
|
ip_addr = self.system_facts['ansible_default_ipv4']['address']
|
|
|
- exit_code, output, _ = module.run_command(['hostname', '-f'])
|
|
|
+ exit_code, output, _ = module.run_command(['hostname', '-f']) # noqa: F405
|
|
|
hostname_f = output.strip() if exit_code == 0 else ''
|
|
|
hostname_values = [hostname_f, self.system_facts['ansible_nodename'],
|
|
|
self.system_facts['ansible_fqdn']]
|
|
@@ -1930,7 +1983,7 @@ class OpenShiftFacts(object):
|
|
|
defaults['docker'] = docker
|
|
|
|
|
|
if 'clock' in roles:
|
|
|
- exit_code, _, _ = module.run_command(['rpm', '-q', 'chrony'])
|
|
|
+ exit_code, _, _ = module.run_command(['rpm', '-q', 'chrony']) # noqa: F405
|
|
|
chrony_installed = bool(exit_code == 0)
|
|
|
defaults['clock'] = dict(
|
|
|
enabled=True,
|
|
@@ -2015,7 +2068,7 @@ class OpenShiftFacts(object):
|
|
|
|
|
|
# TODO: this is not exposed through module_utils/facts.py in ansible,
|
|
|
# need to create PR for ansible to expose it
|
|
|
- bios_vendor = get_file_content(
|
|
|
+ bios_vendor = get_file_content( # noqa: F405
|
|
|
'/sys/devices/virtual/dmi/id/bios_vendor'
|
|
|
)
|
|
|
if bios_vendor == 'Google':
|
|
@@ -2030,8 +2083,7 @@ class OpenShiftFacts(object):
|
|
|
if metadata:
|
|
|
metadata['project']['attributes'].pop('sshKeys', None)
|
|
|
metadata['instance'].pop('serviceAccounts', None)
|
|
|
- elif (virt_type == 'xen' and virt_role == 'guest'
|
|
|
- and re.match(r'.*\.amazon$', product_version)):
|
|
|
+ elif virt_type == 'xen' and virt_role == 'guest' and re.match(r'.*\.amazon$', product_version):
|
|
|
provider = 'aws'
|
|
|
metadata_url = 'http://169.254.169.254/latest/meta-data/'
|
|
|
metadata = get_provider_metadata(metadata_url)
|
|
@@ -2092,7 +2144,7 @@ class OpenShiftFacts(object):
|
|
|
|
|
|
# Determine if any of the provided variable structures match the fact.
|
|
|
matching_structure = None
|
|
|
- if openshift_env_structures != None:
|
|
|
+ if openshift_env_structures is not None:
|
|
|
for structure in openshift_env_structures:
|
|
|
if re.match(structure, openshift_env_fact):
|
|
|
matching_structure = structure
|
|
@@ -2116,7 +2168,7 @@ class OpenShiftFacts(object):
|
|
|
|
|
|
# Disabling too-many-branches and too-many-locals.
|
|
|
# This should be cleaned up as a TODO item.
|
|
|
- #pylint: disable=too-many-branches, too-many-locals
|
|
|
+ # pylint: disable=too-many-branches, too-many-locals
|
|
|
def init_local_facts(self, facts=None,
|
|
|
additive_facts_to_overwrite=None,
|
|
|
openshift_env=None,
|
|
@@ -2144,7 +2196,7 @@ class OpenShiftFacts(object):
|
|
|
if facts is not None:
|
|
|
facts_to_set[self.role] = facts
|
|
|
|
|
|
- if openshift_env != {} and openshift_env != None:
|
|
|
+ if openshift_env != {} and openshift_env is not None:
|
|
|
for fact, value in iteritems(openshift_env):
|
|
|
oo_env_facts = dict()
|
|
|
current_level = oo_env_facts
|
|
@@ -2173,7 +2225,7 @@ class OpenShiftFacts(object):
|
|
|
|
|
|
if 'docker' in new_local_facts:
|
|
|
# remove duplicate and empty strings from registry lists
|
|
|
- for cat in ['additional', 'blocked', 'insecure']:
|
|
|
+ for cat in ['additional', 'blocked', 'insecure']:
|
|
|
key = '{0}_registries'.format(cat)
|
|
|
if key in new_local_facts['docker']:
|
|
|
val = new_local_facts['docker'][key]
|
|
@@ -2190,7 +2242,7 @@ class OpenShiftFacts(object):
|
|
|
if new_local_facts != local_facts:
|
|
|
self.validate_local_facts(new_local_facts)
|
|
|
changed = True
|
|
|
- if not module.check_mode:
|
|
|
+ if not module.check_mode: # noqa: F405
|
|
|
save_local_facts(self.filename, new_local_facts)
|
|
|
|
|
|
self.changed = changed
|
|
@@ -2223,10 +2275,10 @@ class OpenShiftFacts(object):
|
|
|
invalid_facts = self.validate_master_facts(facts, invalid_facts)
|
|
|
if invalid_facts:
|
|
|
msg = 'Invalid facts detected:\n'
|
|
|
+ # pylint: disable=consider-iterating-dictionary
|
|
|
for key in invalid_facts.keys():
|
|
|
msg += '{0}: {1}\n'.format(key, invalid_facts[key])
|
|
|
- module.fail_json(msg=msg,
|
|
|
- changed=self.changed)
|
|
|
+ module.fail_json(msg=msg, changed=self.changed) # noqa: F405
|
|
|
|
|
|
# disabling pylint errors for line-too-long since we're dealing
|
|
|
# with best effort reduction of error messages here.
|
|
@@ -2278,13 +2330,14 @@ class OpenShiftFacts(object):
|
|
|
'Secrets must be 16, 24, or 32 characters in length.')
|
|
|
return invalid_facts
|
|
|
|
|
|
+
|
|
|
def main():
|
|
|
""" main """
|
|
|
# disabling pylint errors for global-variable-undefined and invalid-name
|
|
|
# for 'global module' usage, since it is required to use ansible_facts
|
|
|
# pylint: disable=global-variable-undefined, invalid-name
|
|
|
global module
|
|
|
- module = AnsibleModule(
|
|
|
+ module = AnsibleModule( # noqa: F405
|
|
|
argument_spec=dict(
|
|
|
role=dict(default='common', required=False,
|
|
|
choices=OpenShiftFacts.known_roles),
|
|
@@ -2299,18 +2352,18 @@ def main():
|
|
|
)
|
|
|
|
|
|
if not HAVE_DBUS:
|
|
|
- module.fail_json(msg="This module requires dbus python bindings")
|
|
|
+ module.fail_json(msg="This module requires dbus python bindings") # noqa: F405
|
|
|
|
|
|
- module.params['gather_subset'] = ['hardware', 'network', 'virtual', 'facter']
|
|
|
- module.params['gather_timeout'] = 10
|
|
|
- module.params['filter'] = '*'
|
|
|
+ module.params['gather_subset'] = ['hardware', 'network', 'virtual', 'facter'] # noqa: F405
|
|
|
+ module.params['gather_timeout'] = 10 # noqa: F405
|
|
|
+ module.params['filter'] = '*' # noqa: F405
|
|
|
|
|
|
- role = module.params['role']
|
|
|
- local_facts = module.params['local_facts']
|
|
|
- additive_facts_to_overwrite = module.params['additive_facts_to_overwrite']
|
|
|
- openshift_env = module.params['openshift_env']
|
|
|
- openshift_env_structures = module.params['openshift_env_structures']
|
|
|
- protected_facts_to_overwrite = module.params['protected_facts_to_overwrite']
|
|
|
+ role = module.params['role'] # noqa: F405
|
|
|
+ local_facts = module.params['local_facts'] # noqa: F405
|
|
|
+ additive_facts_to_overwrite = module.params['additive_facts_to_overwrite'] # noqa: F405
|
|
|
+ openshift_env = module.params['openshift_env'] # noqa: F405
|
|
|
+ openshift_env_structures = module.params['openshift_env_structures'] # noqa: F405
|
|
|
+ protected_facts_to_overwrite = module.params['protected_facts_to_overwrite'] # noqa: F405
|
|
|
|
|
|
fact_file = '/etc/ansible/facts.d/openshift.fact'
|
|
|
|
|
@@ -2322,24 +2375,15 @@ def main():
|
|
|
openshift_env_structures,
|
|
|
protected_facts_to_overwrite)
|
|
|
|
|
|
- file_params = module.params.copy()
|
|
|
+ file_params = module.params.copy() # noqa: F405
|
|
|
file_params['path'] = fact_file
|
|
|
- file_args = module.load_file_common_arguments(file_params)
|
|
|
- changed = module.set_fs_attributes_if_different(file_args,
|
|
|
+ file_args = module.load_file_common_arguments(file_params) # noqa: F405
|
|
|
+ changed = module.set_fs_attributes_if_different(file_args, # noqa: F405
|
|
|
openshift_facts.changed)
|
|
|
|
|
|
- return module.exit_json(changed=changed,
|
|
|
+ return module.exit_json(changed=changed, # noqa: F405
|
|
|
ansible_facts=openshift_facts.facts)
|
|
|
|
|
|
-# ignore pylint errors related to the module_utils import
|
|
|
-# pylint: disable=redefined-builtin, unused-wildcard-import, wildcard-import
|
|
|
-# import module snippets
|
|
|
-from ansible.module_utils.basic import *
|
|
|
-from ansible.module_utils.facts import *
|
|
|
-from ansible.module_utils.urls import *
|
|
|
-from ansible.module_utils.six import iteritems, itervalues
|
|
|
-from ansible.module_utils._text import to_native
|
|
|
-from ansible.module_utils.six import b
|
|
|
|
|
|
if __name__ == '__main__':
|
|
|
main()
|