logging.py 3.8 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697
  1. """
  2. Util functions for performing checks on an Elasticsearch, Fluentd, and Kibana stack
  3. """
  4. import json
  5. import os
  6. from openshift_checks import OpenShiftCheck, OpenShiftCheckException, get_var
  7. class LoggingCheck(OpenShiftCheck):
  8. """Base class for OpenShift aggregated logging component checks"""
  9. name = "logging"
  10. logging_namespace = "logging"
  11. @classmethod
  12. def is_active(cls, task_vars):
  13. logging_deployed = get_var(task_vars, "openshift_hosted_logging_deploy", default=False)
  14. return super(LoggingCheck, cls).is_active(task_vars) and cls.is_first_master(task_vars) and logging_deployed
  15. @staticmethod
  16. def is_first_master(task_vars):
  17. """Run only on first master. Returns: bool"""
  18. # Note: It would be nice to use membership in oo_first_master group, however for now it
  19. # seems best to avoid requiring that setup and just check this is the first master.
  20. hostname = get_var(task_vars, "ansible_ssh_host") or [None]
  21. masters = get_var(task_vars, "groups", "masters", default=None) or [None]
  22. return masters and masters[0] == hostname
  23. def run(self, tmp, task_vars):
  24. pass
  25. def get_pods_for_component(self, execute_module, namespace, logging_component, task_vars):
  26. """Get all pods for a given component. Returns: list of pods for component, error string"""
  27. pod_output = self.exec_oc(
  28. execute_module,
  29. namespace,
  30. "get pods -l component={} -o json".format(logging_component),
  31. [],
  32. task_vars
  33. )
  34. try:
  35. pods = json.loads(pod_output)
  36. if not pods or not pods.get('items'):
  37. raise ValueError()
  38. except ValueError:
  39. # successful run but non-parsing data generally means there were no pods in the namespace
  40. return None, 'No pods were found for the "{}" logging component.'.format(logging_component)
  41. return pods['items'], None
  42. @staticmethod
  43. def not_running_pods(pods):
  44. """Returns: list of pods not in a ready and running state"""
  45. return [
  46. pod for pod in pods
  47. if not pod.get("status", {}).get("containerStatuses") or any(
  48. container['ready'] is False
  49. for container in pod['status']['containerStatuses']
  50. ) or not any(
  51. condition['type'] == 'Ready' and condition['status'] == 'True'
  52. for condition in pod['status'].get('conditions', [])
  53. )
  54. ]
  55. @staticmethod
  56. def exec_oc(execute_module=None, namespace="logging", cmd_str="", extra_args=None, task_vars=None):
  57. """
  58. Execute an 'oc' command in the remote host.
  59. Returns: output of command and namespace,
  60. or raises OpenShiftCheckException on error
  61. """
  62. config_base = get_var(task_vars, "openshift", "common", "config_base")
  63. args = {
  64. "namespace": namespace,
  65. "config_file": os.path.join(config_base, "master", "admin.kubeconfig"),
  66. "cmd": cmd_str,
  67. "extra_args": list(extra_args) if extra_args else [],
  68. }
  69. result = execute_module("ocutil", args, None, task_vars)
  70. if result.get("failed"):
  71. msg = (
  72. 'Unexpected error using `oc` to validate the logging stack components.\n'
  73. 'Error executing `oc {cmd}`:\n'
  74. '{error}'
  75. ).format(cmd=args['cmd'], error=result['result'])
  76. if result['result'] == '[Errno 2] No such file or directory':
  77. msg = (
  78. "This host is supposed to be a master but does not have the `oc` command where expected.\n"
  79. "Has an installation been run on this host yet?"
  80. )
  81. raise OpenShiftCheckException(msg)
  82. return result.get("result", "")