Server IP : 192.64.112.168 / Your IP : 18.119.105.32 Web Server : Apache System : Linux nc-ph-2300-85.bluforrest.com 4.18.0-513.9.1.el8_9.x86_64 #1 SMP Sat Dec 2 05:23:44 EST 2023 x86_64 User : expressoneac ( 1128) PHP Version : 8.0.30 Disable Function : exec,passthru,shell_exec,system MySQL : OFF | cURL : ON | WGET : ON | Perl : ON | Python : ON | Sudo : ON | Pkexec : ON Directory : /lib/python3.6/site-packages/sos/report/plugins/ |
Upload File : |
# Copyright (C) 2016 Red Hat, Inc., Pep Turro Mauri <pep@redhat.com> # This file is part of the sos project: https://github.com/sosreport/sos # # This copyrighted material is made available to anyone wishing to use, # modify, copy, or redistribute it subject to the terms and conditions of # version 2 of the GNU General Public License. # # See the LICENSE file in the source distribution for further information. import os.path from sos.report.plugins import Plugin, RedHatPlugin # This plugin collects static configuration and runtime information # about OpenShift Origin based environments, like OpenShift Enterprise 3 # Some clarification on naming: # OpenShift Origin is the upstream project for OpenShift Enterprise, # OpenShift Container Platflorm, and Atomic Platform. # # However, the name "OpenShift Origin" refers to two different code bases: # * Origin M5 and later (https://github.com/openshift/origin) # which is upstream for OpenShift 3.x and later. # This is what this plugin handles # * Origin M4 and earlier (https://github.com/openshift/origin-server) # which is upstream for OpenShift 1.x and 2.x. # This is handled by the plugin in openshift.py # Note that this plugin should be used in conjunction with other plugins # in order to capture relevant data: the Kubernetes plugin for the # masters, the Docker plugin for the nodes, and also generic # plugins (e.g. for /etc/sysconfig entries, network setup etc) class OpenShiftOrigin(Plugin): short_desc = 'OpenShift Origin' plugin_name = "origin" files = None # file lists assigned after path setup below profiles = ('openshift',) master_base_dir = "/etc/origin/master" node_base_dir = "/etc/origin/node" master_cfg = os.path.join(master_base_dir, "master-config.yaml") master_env = os.path.join(master_base_dir, "master.env") node_cfg_file = "node-config.yaml" node_cfg = os.path.join(node_base_dir, node_cfg_file) node_kubeconfig = os.path.join(node_base_dir, "node.kubeconfig") static_pod_dir = os.path.join(node_base_dir, "pods") files = (master_cfg, node_cfg) # Master vs. node # # OpenShift Origin/3.x cluster members can be a master, a node, or both at # the same time: in most deployments masters are also nodes in order to get # access to the pod network, which some functionality (e.g. the API proxy) # requires. Therefore the following methods may all evaluate True on a # single instance (at least one must evaluate True if this is an OpenShift # installation) def is_master(self): """Determine if we are on a master""" return self.path_exists(self.master_cfg) def is_node(self): """Determine if we are on a node""" return self.path_exists(self.node_cfg) def is_static_etcd(self): """Determine if we are on a node running etcd""" return self.path_exists(self.path_join(self.static_pod_dir, "etcd.yaml")) def is_static_pod_compatible(self): """Determine if a node is running static pods""" return self.path_exists(self.static_pod_dir) def setup(self): bstrap_node_cfg = self.path_join(self.node_base_dir, "bootstrap-" + self.node_cfg_file) bstrap_kubeconfig = self.path_join(self.node_base_dir, "bootstrap.kubeconfig") node_certs = self.path_join(self.node_base_dir, "certs", "*") node_client_ca = self.path_join(self.node_base_dir, "client-ca.crt") admin_cfg = self.path_join(self.master_base_dir, "admin.kubeconfig") oc_cmd_admin = f"oc --config={admin_cfg}" static_pod_logs_cmd = "master-logs" # Note that a system can run both a master and a node. # See "Master vs. node" above. if self.is_master(): self.add_copy_spec([ self.master_cfg, self.master_env, self.path_join(self.master_base_dir, "*.crt"), ]) if self.is_static_pod_compatible(): self.add_copy_spec(self.path_join(self.static_pod_dir, "*.yaml")) self.add_cmd_output([ f"{static_pod_logs_cmd} api api", f"{static_pod_logs_cmd} controllers controllers", ]) if self.is_static_etcd(): self.add_cmd_output(f"{static_pod_logs_cmd} etcd etcd") # Possible enhancements: # some thoughts about information that might also be useful # to collect. However, these are maybe not needed in general # and/or present some challenges (scale, sensitive, ...) and need # some more thought. For now just leaving this comment here until # we decide if it's worth collecting: # # General project status: # oc status --all-namespaces (introduced in OSE 3.2) # -> deemed as not worthy in BZ#1394527 # Metrics deployment configurations # oc get -o json dc -n openshift-infra # Logging stack deployment configurations # oc get -o json dc -n logging # # Note: Information about nodes, events, pods, and services # is already collected by the Kubernetes plugin subcmds = [ "adm top images", "adm top imagestreams", "adm top nodes" ] self.add_cmd_output([ f'{oc_cmd_admin} {subcmd}' for subcmd in subcmds ]) jcmds = [ "hostsubnet", "clusternetwork", "netnamespaces" ] self.add_cmd_output([ f'{oc_cmd_admin} get -o json {jcmd}' for jcmd in jcmds ]) nmsps = [ 'default', 'openshift-web-console', 'openshift-ansible-service-broker', 'openshift-sdn', 'openshift-console' ] self.add_cmd_output([ f'{oc_cmd_admin} get -o json deploymentconfig,deployment,' f'daemonsets -n {n}' for n in nmsps ]) if not self.is_static_pod_compatible(): self.add_journal(units=["atomic-openshift-master", "atomic-openshift-master-api", "atomic-openshift-master-controllers"]) # get logs from the infrastructure pods running in the default ns pods = self.exec_cmd(f"{oc_cmd_admin} get pod -o name -n default") for pod in pods['output'].splitlines(): self.add_cmd_output(f"{oc_cmd_admin} logs -n default {pod}") # Note that a system can run both a master and a node. # See "Master vs. node" above. if self.is_node(): self.add_copy_spec([ self.node_cfg, self.node_kubeconfig, node_certs, node_client_ca, bstrap_node_cfg, bstrap_kubeconfig, self.path_join(self.node_base_dir, "*.crt"), self.path_join(self.node_base_dir, "resolv.conf"), self.path_join(self.node_base_dir, "node-dnsmasq.conf"), ]) self.add_journal(units="atomic-openshift-node") def postproc(self): # Clear env values from objects that can contain sensitive data # Sample JSON content: # { # "name": "MYSQL_PASSWORD", # "value": "mypassword" # }, # This will mask values when the "name" looks susceptible of # values worth obfuscating, i.e. if the name contains strings # like "pass", "pwd", "key" or "token". env_regexp = r'(?P<var>{\s*"name":\s*[^,]*' \ r'(pass|pwd|key|token|cred|secret' \ r'|PASS|PWD|KEY|TOKEN|CRED|SECRET)[^,]*,' \ r'\s*"value":)[^}]*' self.do_cmd_output_sub('oc*json', env_regexp, r'\g<var> "********"') # LDAP identity provider (bindPassword) # and github/google/OpenID identity providers (clientSecret) self.do_file_sub(self.master_cfg, r"(bindPassword|clientSecret):\s*(.*)", r'\1:"********"') class AtomicOpenShift(OpenShiftOrigin, RedHatPlugin): short_desc = 'OpenShift Enterprise / OpenShift Container Platform' packages = ('atomic-openshift',) # vim: set et ts=4 sw=4 :