But I find the Grafana is unable to add built-in Prometheus of openshift-monitoring project as data source. As an engineer responsible for maintaining a stack, metrics are one of the most important tools for understanding your infrastructure. KubeAPI has disappeared from Prometheus target discovery. The persistent volume claim size for each of the Prometheus instances. To configure core OpenShift Container Platform monitoring components, you must create the cluster-monitoring-config ConfigMap object in the openshift-monitoring project. The monitoring stack component for which you are setting a log level. The Grafana instance is not user-configurable. Currently supported authentication methods are bearer token (bearerToken) and client TLS (tlsConfig). 1 Answer. If you enabled the openshift_cluster_monitoring_operator_alertmanager_storage_enabled option, set a specific StorageClass to ensure that pods are configured to use the PVC with that storageclass. The following example sets the retention time to 24 hours for the Prometheus instance that monitors core OpenShift Container Platform components: To modify the retention time for the Prometheus instance that monitors user-defined projects: The following example sets the retention time to 24 hours for the Prometheus instance that monitors user-defined projects: Save the file to apply the changes. Creating additional ConfigMap objects, that cause the cluster monitoring Prometheus instance to include additional alerting and recording rules. The supported way of configuring OpenShift Container Platform Monitoring is by configuring it using the options described in this document. Ensure the user belongs to the cluster-monitoring-view role. Description: This is a DeadMansSwitch meant to ensure that the entire Alerting pipeline is functional. Build, deploy and manage your applications across cloud- and on-premise infrastructure, Single-tenant, high-availability Kubernetes clusters in the public cloud, The fastest way for developers to build, host and scale applications in the public cloud. The Alertmanager continuously sends notifications for the dead mans switch to the notification provider that supports this functionality. Log in. A value is required if this parameter is specified. While overriding CVO control for an Operator can be helpful during debugging, this is unsupported and the cluster administrator assumes full control of the individual component configurations and upgrades. Add Token of the SA as a bearer token . Defines a minimum resource request of 200 millicores for the Prometheus container. Prometheus has two replicas and Alertmanager has three replicas, which amounts to five PVs. This variable is set to false by default. See Dynamic Volume Provisioning for details. As the name indicates, someone should be paged for alerts that are critical. The monitoring stack imposes additional resource requirements. Etcd cluster "Job": instance Instance has seen X leader changes within the last hour. By using Prometheus and Grafana to collect and visualize the metrics of the cluster, and by using Portainer to simplify the deployment, you can effectively monitor your Swarm cluster and detect potential issues before they become critical. The new component placement configuration is applied automatically. Get started with Grafana and Prometheus. No translations currently exist. Check whether the cluster-monitoring-config ConfigMap object exists: Create the following YAML manifest. However, this built-in monitoring capability provides read-only cluster monitoring and does not allow monitoring any additional target. OpenShift Container Platform 3.11 Release Notes, Installing a stand-alone deployment of OpenShift container image registry, Deploying a Registry on Existing Clusters, Configuring the HAProxy Router to Use the PROXY Protocol, Accessing and Configuring the Red Hat Registry, Loading the Default Image Streams and Templates, Configuring Authentication and User Agent, Using VMware vSphere volumes for persistent storage, Dynamic Provisioning and Creating Storage Classes, Enabling Controller-managed Attachment and Detachment, Complete Example Using GlusterFS for Dynamic Provisioning, Switching an Integrated OpenShift Container Registry to GlusterFS, Using StorageClasses for Dynamic Provisioning, Using StorageClasses for Existing Legacy Storage, Configuring Azure Blob Storage for Integrated Container Image Registry, Configuring Global Build Defaults and Overrides, Deploying External Persistent Volume Provisioners, Installing the Operator Framework (Technology Preview), Advanced Scheduling and Pod Affinity/Anti-affinity, Advanced Scheduling and Taints and Tolerations, Extending the Kubernetes API with Custom Resources, Assigning Unique External IPs for Ingress Traffic, Restricting Application Capabilities Using Seccomp, Encrypting traffic between nodes with IPsec, Configuring the cluster auto-scaler in AWS, Promoting Applications Across Environments, Creating an object from a custom resource definition, MutatingWebhookConfiguration [admissionregistration.k8s.io/v1beta1], ValidatingWebhookConfiguration [admissionregistration.k8s.io/v1beta1], LocalSubjectAccessReview [authorization.k8s.io/v1], SelfSubjectAccessReview [authorization.k8s.io/v1], SelfSubjectRulesReview [authorization.k8s.io/v1], SubjectAccessReview [authorization.k8s.io/v1], ClusterRoleBinding [authorization.openshift.io/v1], ClusterRole [authorization.openshift.io/v1], LocalResourceAccessReview [authorization.openshift.io/v1], LocalSubjectAccessReview [authorization.openshift.io/v1], ResourceAccessReview [authorization.openshift.io/v1], RoleBindingRestriction [authorization.openshift.io/v1], RoleBinding [authorization.openshift.io/v1], SelfSubjectRulesReview [authorization.openshift.io/v1], SubjectAccessReview [authorization.openshift.io/v1], SubjectRulesReview [authorization.openshift.io/v1], CertificateSigningRequest [certificates.k8s.io/v1beta1], ImageStreamImport [image.openshift.io/v1], ImageStreamMapping [image.openshift.io/v1], EgressNetworkPolicy [network.openshift.io/v1], OAuthAuthorizeToken [oauth.openshift.io/v1], OAuthClientAuthorization [oauth.openshift.io/v1], AppliedClusterResourceQuota [quota.openshift.io/v1], ClusterResourceQuota [quota.openshift.io/v1], ClusterRoleBinding [rbac.authorization.k8s.io/v1], ClusterRole [rbac.authorization.k8s.io/v1], RoleBinding [rbac.authorization.k8s.io/v1], PriorityClass [scheduling.k8s.io/v1beta1], PodSecurityPolicyReview [security.openshift.io/v1], PodSecurityPolicySelfSubjectReview [security.openshift.io/v1], PodSecurityPolicySubjectReview [security.openshift.io/v1], RangeAllocation [security.openshift.io/v1], SecurityContextConstraints [security.openshift.io/v1], VolumeAttachment [storage.k8s.io/v1beta1], BrokerTemplateInstance [template.openshift.io/v1], TemplateInstance [template.openshift.io/v1], UserIdentityMapping [user.openshift.io/v1], Container-native Virtualization Installation, Container-native Virtualization Users Guide, Container-native Virtualization Release Notes, Configuring OpenShift Container Platform cluster monitoring, Accessing Prometheus, Alertmanager, and Grafana, Capacity Planning for Cluster Monitoring Operator, configure Dead Mans Snitch for PagerDuty. Store the machine learning models in S3 bucket as follows. Log warning and error messages only. You cannot access web UIs using unencrypted connections. Note that this behavior is known to cause a breaking behavior if applied, as Prometheus 2.0 will ship with a new rule file syntax. It collects information from the user, generates manifests, and uses terraform to provision and configure infrastructure that will compose a cluster. Defaults to node-role.kubernetes.io/infra=true. Description: The configuration of the instances of the Alertmanager cluster Service are out of sync. Add enabled: false for the alertmanagerMain component under data/config.yaml: Save the file to apply the changes. The following example configures the thanosRuler component to tolerate the example taint: Save the file to apply the changes. Overcommited CPU resource requests on Pods, cannot tolerate node failure. When you save changes to a monitoring config map, the pods and other resources in the related project might be redeployed. The pods affected by the new configuration are restarted automatically and the new storage configuration is applied. The OpenShift Container Platform monitoring stack includes a local Alertmanager instance that routes alerts from Prometheus. The following example configures the alertmanagerMain component to tolerate the example taint: To assign tolerations to a component that monitors user-defined projects: For example, oc adm taint nodes node1 key1=value1:NoSchedule adds a taint to node1 with the key key1 and the value value1. Description: Errors while sending alerts from Prometheus Namespace/Pod to Alertmanager Alertmanager, Summary: Prometheus is not connected to any Alertmanagers. You can configure remote write storage to enable Prometheus to send ingested metrics to remote systems for long-term storage. Using attributes that are bound to a limited set of possible values reduces the number of potential key-value pair combinations. The following example ConfigMap object configures a persistent volume claim (PVC) for Prometheus. Subscriber exclusive content. It is administrators responsibility to dedicate sufficient storage to ensure that the disk does not become full. Use Filesystem as the storage type value for the volumeMode parameter when you configure the persistent volume. ClusterMonitoringOperator has disappeared from Prometheus target discovery. $ ./deploy-monitoring.sh; Create a Route for the prometheus, alertmanager, and grafana Services. Log in to the OpenShift Container Platform (OCP) by using the OpenShift administrator credentials. Prometheus, Grafana and Alertmanager is the default platform monitoring stack in Redhat Openshift/IBM Cloud Pak for Data (CP4D). Backward compatibility for metrics, recording rules, or alerting rules is not guaranteed. For information on system requirements for persistent storage, see Prometheus database storage requirements. Install the Grafana operator in the same project or the namespace where you installed the Prometheus operator. Thanos stores metrics of all Prometheus instances (Prometheus instances installed on each OpenShift or Kubernetes cluster) which means you have a single place to see metrics from different clusters, delivering a global view of the fleet. . You can limit the number of samples that can be accepted per target scrape in user-defined projects. Summary: Targets are down. openshift_cluster_monitoring_operator_prometheus_storage_class_name. Create Service Account on a project. This might lead to previously collected metrics being lost if you have not yet followed the steps in the "Configuring persistent storage" section. Summary: Reloading Prometheus' configuration failed. The following sample shows how to forward a single metric called my_metric: See the Prometheus relabel_config documentation for information about write relabel configuration options. Running cluster monitoring with persistent storage means that your metrics are stored to a persistent volume (PV) and can survive a pod being restarted or recreated. See Alertmanager configuration for configuring alerting through different alert receivers. $ ./deploy-monitoring.sh. The OpenShift Container Platform Ansible openshift_cluster_monitoring_operator role configures and deploys the Cluster Monitoring Operator using the variables from the inventory file. The pods for the component restarts automatically when you apply the log-level change. If you want to resize a PV for a monitoring component such as Prometheus, Thanos Ruler, or Alertmanager, you can update the appropriate config map in which the component is configured. Dedicate sufficient local persistent storage to ensure that the disk does not become full. You can add external Alertmanager instances by configuring the cluster-monitoring-config config map in either the openshift-monitoring project or the user-workload-monitoring-config project. After you enable dynamic storage, you can also set the storageclass for the persistent volume claim for each component in the following parameters in the Ansible inventory file: openshift_cluster_monitoring_operator_prometheus_storage_class_name (default: ""), openshift_cluster_monitoring_operator_alertmanager_storage_class_name (default: ""). Apply the credentials file to the cluster: Now that you have configured authentication, visit the Targets page of the web interface again. You might do this if you do not need these dashboards and want to conserve resources in your cluster. Namespace/Pod (Container) is restarting times / second, Deployment Namespace/Deployment generation mismatch, Deployment Namespace/Deployment replica mismatch, StatefulSet Namespace/StatefulSet replica mismatch, StatefulSet Namespace/StatefulSet generation mismatch, Only X% of desired pods scheduled and ready for daemon set Namespace/DaemonSet. Choose prometheus-operator under A specific namespace on the clusterand subscribe. Etcd cluster "Job": member communication with To is taking X_s on etcd instance _Instance. See computing resources recommendations for details. The following example resizes the storage size for the Prometheus component in the openshift-monitoring namespace to 100Gi: Delete the underlying StatefulSet with the --cascade=orphan parameter: By default, the OpenShift Container Platform monitoring stack configures the retention time for Prometheus data to be 15 days. Kubernetes API server client 'Job/Instance' is experiencing X% errors. This change results in some components, including Prometheus and the Thanos Querier, being restarted. It also automatically generates monitoring target configurations based on familiar Kubernetes label queries. In addition to Prometheus and Alertmanager, OpenShift Container Platform Monitoring also includes a Grafana instance as well as pre-built dashboards for cluster monitoring troubleshooting. The following example configures a PVC that claims local persistent storage for the Prometheus instance that monitors core OpenShift Container Platform components: In the above example, the storage class created by the Local Storage Operator is called local-storage. Often, only a single key-value pair is used. This section explains what configuration is supported, shows how to configure the monitoring stack, and demonstrates several common configuration scenarios. The running monitoring processes in that project might also be restarted. Add the enforcedSampleLimit configuration to data/config.yaml to limit the number of samples that can be accepted per target scrape in user-defined projects: Save the file to apply the changes. Step 1: Create a project where GEL will be deployed: oc new-project $ {PROJECT_NAME} Step 2: Add the Grafana Helm charts repository: helm repo add grafana https://grafana.github.io/helm-charts helm repo update Step 3: Deploy the chart to your OpenShift cluster. The following sample config map configures an additional Alertmanager using Thanos Ruler with a bearer token and client TLS authentication: Save the file to apply the changes to the ConfigMap object. 1 Step 1: Setting up Openshift The simplest way to setup Openshift locally is using "CodeReady Containers". If you enabled the openshift_cluster_monitoring_operator_prometheus_storage_enabled option, set a specific StorageClass to ensure that pods are configured to use the PVC with that storageclass. For , substitute authentication and other configuration details for additional Alertmanager instances. Grafana and Prometheus: Download Prometheus and node_exporter. This script creates and configures the OpenShift resources needed to deploy Prometheus, Alertmanager, and Grafana in your OpenShift project. The following example checks the log level in the prometheus-operator deployment in the openshift-user-workload-monitoring project: Check that the pods for the component are running. In this example the file is called cluster-monitoring-config.yaml: Apply the configuration to create the ConfigMap object: To configure the components that monitor user-defined projects, you must create the user-workload-monitoring-config ConfigMap object in the openshift-user-workload-monitoring project. Currently X bytes are available. ', Kubernetes API server client 'Job/Instance' is experiencing X errors / sec.'. To get the addresses for accessing Prometheus, Alertmanager, and Grafana web UIs: Make sure to prepend https:// to these addresses. Add a remoteWrite: section under data/config.yaml/prometheusK8s. KubeScheduler has disappeared from Prometheus target discovery. The following log levels can be applied to the relevant component in the cluster-monitoring-config and user-workload-monitoring-config ConfigMap objects: debug. Add enabled: false for the grafana component under data/config.yaml: Save the file to apply the changes. You can assign tolerations to any of the monitoring stack components to enable moving them to tainted nodes. No service disruption occurs during this process. Job instance Instance will exhaust its file descriptors soon. '[Openshift] Prometheus + AlertManager ' . You can move any of the monitoring stack components to specific nodes. Instead of statically-provisioned storage, you can use dynamically-provisioned storage. If required, configure remote write for the Prometheus instance that monitors user-defined projects by changing the name and namespace metadata values as follows: Save the file to apply the changes to the ConfigMap object. You have access to the cluster as a user with the cluster-admin role or as a user with the user-workload-monitoring-config-edit role in the openshift-user-workload-monitoring project. Log informational, warning, and error messages. Often, only a single key-value pair is used. Now you can navigate to the web interface to see more information about the status of etcd monitoring. Prerequisites In this example the file is called user-workload-monitoring-config.yaml: Configurations applied to the user-workload-monitoring-config ConfigMap object are not activated unless a cluster administrator has enabled monitoring for user-defined projects. Application Monitoring on Red Hat OpenShift Container Platform (RHOCP) 4.2 with Prometheus and Grafana. This can cause collisions and load differences that cannot be accounted for, therefore the Prometheus setup can be unstable. Openshift 4.x comes with a built-in monitoring stack (Prometheus, Grafana & Alertmanager); The main Prometheus instance that is responsible for scraping infrastructure-related metrics is. Etcd cluster "Job": X% of requests for GRPC_Method failed on etcd instance Instance. Kubelet has disappeared from Prometheus target discovery. 100 * (count(up == 0) BY (job, namespace, service) / count(up) BY (job, Learn more about OpenShift Container Platform, OpenShift Container Platform 4.9 release notes, Selecting an installation method and preparing a cluster, Mirroring images for a disconnected installation, Installing a cluster on AWS with customizations, Installing a cluster on AWS with network customizations, Installing a cluster on AWS in a restricted network, Installing a cluster on AWS into an existing VPC, Installing a cluster on AWS into a government or secret region, Installing a cluster on AWS into a China region, Installing a cluster on AWS using CloudFormation templates, Installing a cluster on AWS in a restricted network with user-provisioned infrastructure, Installing a cluster on Azure with customizations, Installing a cluster on Azure with network customizations, Installing a cluster on Azure into an existing VNet, Installing a cluster on Azure into a government region, Installing a cluster on Azure using ARM templates, Manually creating IAM for Azure Stack Hub, Installing a cluster on Azure Stack Hub using ARM templates, Installing a cluster on GCP with customizations, Installing a cluster on GCP with network customizations, Installing a cluster on GCP in a restricted network, Installing a cluster on GCP into an existing VPC, Installing a cluster on GCP using Deployment Manager templates, Installing a cluster into a shared VPC on GCP using Deployment Manager templates, Installing a cluster on GCP in a restricted network with user-provisioned infrastructure, Installing a user-provisioned cluster on bare metal, Installing a user-provisioned bare metal cluster with network customizations, Installing a user-provisioned bare metal cluster on a restricted network, Preparing to install OpenShift on a single node, Setting up the environment for an OpenShift installation, Preparing to install with z/VM on IBM Z and LinuxONE, Installing a cluster with z/VM on IBM Z and LinuxONE, Restricted network IBM Z installation with z/VM, Preparing to install with RHEL KVM on IBM Z and LinuxONE, Installing a cluster with RHEL KVM on IBM Z and LinuxONE, Restricted network IBM Z installation with RHEL KVM, Restricted network IBM Power installation, Installing a cluster on OpenStack with customizations, Installing a cluster on OpenStack with Kuryr, Installing a cluster that supports SR-IOV compute machines on OpenStack, Installing a cluster on OpenStack on your own infrastructure, Installing a cluster on OpenStack with Kuryr on your own infrastructure, Installing a cluster on OpenStack on your own SR-IOV infrastructure, Installing a cluster on OpenStack in a restricted network, OpenStack cloud configuration reference guide, Uninstalling a cluster on OpenStack from your own infrastructure, Installing a cluster on RHV with customizations, Installing a cluster on RHV with user-provisioned infrastructure, Installing a cluster on RHV in a restricted network, Installing a cluster on vSphere with customizations, Installing a cluster on vSphere with network customizations, Installing a cluster on vSphere with user-provisioned infrastructure, Installing a cluster on vSphere with user-provisioned infrastructure and network customizations, Installing a cluster on vSphere in a restricted network, Installing a cluster on vSphere in a restricted network with user-provisioned infrastructure, Uninstalling a cluster on vSphere that uses installer-provisioned infrastructure, Using the vSphere Problem Detector Operator, Installing a cluster on VMC with customizations, Installing a cluster on VMC with network customizations, Installing a cluster on VMC in a restricted network, Installing a cluster on VMC with user-provisioned infrastructure, Installing a cluster on VMC with user-provisioned infrastructure and network customizations, Installing a cluster on VMC in a restricted network with user-provisioned infrastructure, Converting a connected cluster to a disconnected cluster, Configuring additional devices in an IBM Z or LinuxONE environment, Preparing to update to OpenShift Container Platform 4.9, Preparing to perform an EUS-to-EUS update, Performing update using canary rollout strategy, Updating a cluster that includes RHEL compute machines, About cluster updates in a disconnected environment, Mirroring the OpenShift Container Platform image repository, Updating a cluster in a disconnected environment using OSUS, Updating a cluster in a disconnected environment without OSUS, Updating hardware on nodes running on vSphere, Showing data collected by remote health monitoring, Using Insights to identify issues with your cluster, Using remote health reporting in a restricted network, Importing simple content access certificates with Insights Operator, Troubleshooting CRI-O container runtime issues, Troubleshooting the Source-to-Image process, Troubleshooting Windows container workload issues, OpenShift CLI developer command reference, OpenShift CLI administrator command reference, Knative CLI (kn) for use with OpenShift Serverless, Hardening Red Hat Enterprise Linux CoreOS, Replacing the default ingress certificate, Securing service traffic using service serving certificates, User-provided certificates for the API server, User-provided certificates for default ingress, Monitoring and cluster logging Operator component certificates, Retrieving Compliance Operator raw results, Performing advanced Compliance Operator tasks, Understanding the Custom Resource Definitions, Understanding the File Integrity Operator, Performing advanced File Integrity Operator tasks, Troubleshooting the File Integrity Operator, Allowing JavaScript-based access to the API server from additional hosts, Authentication and authorization overview, Understanding identity provider configuration, Configuring an htpasswd identity provider, Configuring a basic authentication identity provider, Configuring a request header identity provider, Configuring a GitHub or GitHub Enterprise identity provider, Configuring an OpenID Connect identity provider, Using RBAC to define and apply permissions, Understanding and creating service accounts, Using a service account as an OAuth client, Using manual mode with AWS Security Token Service, Understanding the Cluster Network Operator, Defining a default network policy for projects, Removing a pod from an additional network, About Single Root I/O Virtualization (SR-IOV) hardware networks, Configuring an SR-IOV Ethernet network attachment, Configuring an SR-IOV InfiniBand network attachment, About the OpenShift SDN default CNI network provider, Configuring an egress firewall for a project, Removing an egress firewall from a project, Considerations for the use of an egress router pod, Deploying an egress router pod in redirect mode, Deploying an egress router pod in HTTP proxy mode, Deploying an egress router pod in DNS proxy mode, Configuring an egress router pod destination list from a config map, About the OVN-Kubernetes network provider, Migrating from the OpenShift SDN cluster network provider, Rolling back to the OpenShift SDN cluster network provider, Converting to IPv4/IPv6 dual stack networking, Configuring ingress cluster traffic using an Ingress Controller, Configuring ingress cluster traffic using a load balancer, Configuring ingress cluster traffic on AWS using a Network Load Balancer, Configuring ingress cluster traffic using a service external IP, Configuring ingress cluster traffic using a NodePort, Troubleshooting node network configuration, Associating secondary interfaces metrics to network attachments, Persistent storage using AWS Elastic Block Store, Persistent storage using GCE Persistent Disk, Persistent storage using Red Hat OpenShift Container Storage, AWS Elastic Block Store CSI Driver Operator, AWS Elastic File Service CSI Driver Operator, Red Hat Virtualization CSI Driver Operator, Image Registry Operator in OpenShift Container Platform, Configuring the registry for AWS user-provisioned infrastructure, Configuring the registry for GCP user-provisioned infrastructure, Configuring the registry for OpenStack user-provisioned infrastructure, Configuring the registry for Azure user-provisioned infrastructure, Creating applications from installed Operators, Allowing non-cluster administrators to install Operators, Upgrading projects for newer Operator SDK versions, High-availability or single-node cluster detection and support, Configuring built-in monitoring with Prometheus, Migrating package manifest projects to bundle format, Setting up additional trusted certificate authorities for builds, Creating CI/CD solutions for applications using OpenShift Pipelines, Managing non-versioned and versioned cluster tasks, Using Tekton Hub with OpenShift Pipelines, Working with OpenShift Pipelines using the Developer perspective, Reducing resource consumption of OpenShift Pipelines, Setting compute resource quota for OpenShift Pipelines, Automatic pruning of task run and pipeline run, Using pods in a privileged security context, Authenticating pipelines using git secret, Using Tekton Chains for OpenShift Pipelines supply chain security, Viewing pipeline logs using the OpenShift Logging Operator, Configuring an OpenShift cluster by deploying an application with cluster configurations, Deploying a Spring Boot application with Argo CD, Configuring SSO for Argo CD using Keycloak, Running Control Plane Workloads on Infra nodes, Using the Cluster Samples Operator with an alternate registry, Using image streams with Kubernetes resources, Triggering updates on image stream changes, Creating applications using the Developer perspective, Viewing application composition using the Topology view, Getting started with service binding on IBM Power, IBM Z, and LinuxONE, Binding workloads using Service Binding Operator, Connecting an application to a service using the Developer perspective, Configuring custom Helm chart repositories, Understanding Deployments and DeploymentConfigs, Monitoring project and application metrics using the Developer perspective, Adding compute machines to clusters with user-provisioned infrastructure manually, Adding compute machines to AWS using CloudFormation templates, Adding compute machines to vSphere manually, Automatically scaling pods with the horizontal pod autoscaler, Automatically adjust pod resource levels with the vertical pod autoscaler, Using Device Manager to make devices available to nodes, Including pod priority in pod scheduling decisions, Configuring the default scheduler to control pod placement, Scheduling pods using a scheduler profile, Placing pods relative to other pods using pod affinity and anti-affinity rules, Controlling pod placement on nodes using node affinity rules, Controlling pod placement using node taints, Controlling pod placement using pod topology spread constraints, Running background tasks on nodes automatically with daemonsets, Viewing and listing the nodes in your cluster, Managing the maximum number of pods per node, Remediating nodes with the Poison Pill Operator, Deploying node health checks by using the Node Health Check Operator, Freeing node resources using garbage collection, Allocating specific CPUs for nodes in a cluster, Configuring the TLS security profile for the kubelet, Using Init Containers to perform tasks before a pod is deployed, Allowing containers to consume API objects, Using port forwarding to access applications in a container, Viewing system event information in a cluster, Configuring cluster memory to meet container memory and risk requirements, Configuring your cluster to place pods on overcommited nodes, Using remote worker node at the network edge, Red Hat OpenShift support for Windows Containers overview, Red Hat OpenShift support for Windows Containers release notes, Understanding Windows container workloads, Creating a Windows MachineSet object on AWS, Creating a Windows MachineSet object on Azure, Creating a Windows MachineSet object on vSphere, Using Bring-Your-Own-Host Windows instances as nodes, OpenShift sandboxed containers release notes, Understanding OpenShift sandboxed containers, Deploying OpenShift sandboxed containers workloads, Uninstalling OpenShift sandboxed containers workloads, Collecting OpenShift sandboxed containers data for Red Hat Support, About the Cluster Logging custom resource, Configuring CPU and memory limits for Logging components, Using tolerations to control Logging pod placement, Moving the Logging resources with node selectors, Collecting logging data for Red Hat Support, Recommended host practices for IBM Z & LinuxONE environments, Planning your environment according to object maximums, What huge pages do and how they are consumed by apps, Performance Addon Operator for low latency nodes, Performing latency tests for platform verification, Deploying distributed units manually on single-node OpenShift, Workload partitioning on single-node OpenShift, Deploying distributed units at scale in a disconnected environment, About specialized hardware and driver enablement, Overview of backup and restore operations, Installing and configuring OADP with Azure, Advanced OADP features and functionalities, Recovering from expired control plane certificates, About migrating from OpenShift Container Platform 3 to 4, Differences between OpenShift Container Platform 3 and 4, Installing MTC in a restricted network environment, Editing kubelet log level verbosity and gathering logs, LocalResourceAccessReview [authorization.openshift.io/v1], LocalSubjectAccessReview [authorization.openshift.io/v1], ResourceAccessReview [authorization.openshift.io/v1], SelfSubjectRulesReview [authorization.openshift.io/v1], SubjectAccessReview [authorization.openshift.io/v1], SubjectRulesReview [authorization.openshift.io/v1], LocalSubjectAccessReview [authorization.k8s.io/v1], SelfSubjectAccessReview [authorization.k8s.io/v1], SelfSubjectRulesReview [authorization.k8s.io/v1], SubjectAccessReview [authorization.k8s.io/v1], ClusterAutoscaler [autoscaling.openshift.io/v1], MachineAutoscaler [autoscaling.openshift.io/v1beta1], HelmChartRepository [helm.openshift.io/v1beta1], ConsoleCLIDownload [console.openshift.io/v1], ConsoleExternalLogLink [console.openshift.io/v1], ConsoleNotification [console.openshift.io/v1], ConsolePlugin [console.openshift.io/v1alpha1], ConsoleQuickStart [console.openshift.io/v1], ConsoleYAMLSample [console.openshift.io/v1], CustomResourceDefinition [apiextensions.k8s.io/v1], MutatingWebhookConfiguration [admissionregistration.k8s.io/v1], ValidatingWebhookConfiguration [admissionregistration.k8s.io/v1], ImageStreamImport [image.openshift.io/v1], ImageStreamMapping [image.openshift.io/v1], ContainerRuntimeConfig [machineconfiguration.openshift.io/v1], ControllerConfig [machineconfiguration.openshift.io/v1], KubeletConfig [machineconfiguration.openshift.io/v1], MachineConfigPool [machineconfiguration.openshift.io/v1], MachineConfig [machineconfiguration.openshift.io/v1], MachineHealthCheck [machine.openshift.io/v1beta1], MachineSet [machine.openshift.io/v1beta1], APIRequestCount [apiserver.openshift.io/v1], AlertmanagerConfig [monitoring.coreos.com/v1alpha1], PrometheusRule [monitoring.coreos.com/v1], ServiceMonitor [monitoring.coreos.com/v1], EgressNetworkPolicy [network.openshift.io/v1], EgressRouter [network.operator.openshift.io/v1], IPPool [whereabouts.cni.cncf.io/v1alpha1], NetworkAttachmentDefinition [k8s.cni.cncf.io/v1], PodNetworkConnectivityCheck [controlplane.operator.openshift.io/v1alpha1], OAuthAuthorizeToken [oauth.openshift.io/v1], OAuthClientAuthorization [oauth.openshift.io/v1], UserOAuthAccessToken [oauth.openshift.io/v1], Authentication [operator.openshift.io/v1], CloudCredential [operator.openshift.io/v1], ClusterCSIDriver [operator.openshift.io/v1], Config [imageregistry.operator.openshift.io/v1], Config [samples.operator.openshift.io/v1], CSISnapshotController [operator.openshift.io/v1], DNSRecord [ingress.operator.openshift.io/v1], ImageContentSourcePolicy [operator.openshift.io/v1alpha1], ImagePruner [imageregistry.operator.openshift.io/v1], IngressController [operator.openshift.io/v1], KubeControllerManager [operator.openshift.io/v1], KubeStorageVersionMigrator [operator.openshift.io/v1], OpenShiftAPIServer [operator.openshift.io/v1], OpenShiftControllerManager [operator.openshift.io/v1], OperatorPKI [network.operator.openshift.io/v1], CatalogSource [operators.coreos.com/v1alpha1], ClusterServiceVersion [operators.coreos.com/v1alpha1], InstallPlan [operators.coreos.com/v1alpha1], OperatorCondition [operators.coreos.com/v2], PackageManifest [packages.operators.coreos.com/v1], Subscription [operators.coreos.com/v1alpha1], ClusterRoleBinding [rbac.authorization.k8s.io/v1], ClusterRole [rbac.authorization.k8s.io/v1], RoleBinding [rbac.authorization.k8s.io/v1], ClusterRoleBinding [authorization.openshift.io/v1], ClusterRole [authorization.openshift.io/v1], RoleBindingRestriction [authorization.openshift.io/v1], RoleBinding [authorization.openshift.io/v1], AppliedClusterResourceQuota [quota.openshift.io/v1], ClusterResourceQuota [quota.openshift.io/v1], FlowSchema [flowcontrol.apiserver.k8s.io/v1beta1], PriorityLevelConfiguration [flowcontrol.apiserver.k8s.io/v1beta1], CertificateSigningRequest [certificates.k8s.io/v1], CredentialsRequest [cloudcredential.openshift.io/v1], PodSecurityPolicyReview [security.openshift.io/v1], PodSecurityPolicySelfSubjectReview [security.openshift.io/v1], PodSecurityPolicySubjectReview [security.openshift.io/v1], RangeAllocation [security.openshift.io/v1], SecurityContextConstraints [security.openshift.io/v1], CSIStorageCapacity [storage.k8s.io/v1beta1], StorageVersionMigration [migration.k8s.io/v1alpha1], VolumeSnapshot [snapshot.storage.k8s.io/v1], VolumeSnapshotClass [snapshot.storage.k8s.io/v1], VolumeSnapshotContent [snapshot.storage.k8s.io/v1], BrokerTemplateInstance [template.openshift.io/v1], TemplateInstance [template.openshift.io/v1], UserIdentityMapping [user.openshift.io/v1], Configuring the distributed tracing platform, Configuring distributed tracing data collection, Preparing your cluster for OpenShift Virtualization, Specifying nodes for OpenShift Virtualization components, Installing OpenShift Virtualization using the web console, Installing OpenShift Virtualization using the CLI, Uninstalling OpenShift Virtualization using the web console, Uninstalling OpenShift Virtualization using the CLI, Additional security privileges granted for kubevirt-controller and virt-launcher, Triggering virtual machine failover by resolving a failed node, Installing the QEMU guest agent on virtual machines, Viewing the QEMU guest agent information for virtual machines, Managing config maps, secrets, and service accounts in virtual machines, Installing VirtIO driver on an existing Windows virtual machine, Installing VirtIO driver on a new Windows virtual machine, Working with resource quotas for virtual machines, Configuring PXE booting for virtual machines, Enabling dedicated resources for a virtual machine, Importing virtual machine images with data volumes, Importing virtual machine images into block storage with data volumes, Enabling user permissions to clone data volumes across namespaces, Cloning a virtual machine disk into a new data volume, Cloning a virtual machine by using a data volume template, Cloning a virtual machine disk into a new block storage data volume, Configuring a virtual machine for the default pod network, Creating a service to expose a virtual machine, Connecting a virtual machine to a Linux bridge network, Connecting a virtual machine to an SR-IOV network, Configuring IP addresses for virtual machines, Viewing the IP address of NICs on a virtual machine, Using a MAC address pool for virtual machines, Configuring local storage for virtual machines, Configuring CDI to work with namespaces that have a compute resource quota, Uploading local disk images by using the web console, Uploading local disk images by using the virtctl tool, Uploading a local disk image to a block storage data volume, Moving a local virtual machine disk to a different node, Expanding virtual storage by adding blank disk images, Cloning a data volume using smart-cloning, Using container disks with virtual machines, Re-using statically provisioned persistent volumes, Enabling dedicated resources for a virtual machine template, Migrating a virtual machine instance to another node, Monitoring live migration of a virtual machine instance, Cancelling the live migration of a virtual machine instance, Configuring virtual machine eviction strategy, Managing node labeling for obsolete CPU models, Diagnosing data volumes using events and conditions, Viewing information about virtual machine workloads, Reviewing resource usage by virtual machines, OpenShift cluster monitoring, logging, and Telemetry, Preparing to install OpenShift Serverless, Overriding system deployment configurations, Reroute traffic using blue-green strategy, Configuring JSON Web Token authentication for Knative services, Using JSON Web Token authentication with Service Mesh 2.x, Using JSON Web Token authentication with Service Mesh 1.x, Domain mapping using the Developer perspective, Domain mapping using the Administrator perspective, Securing a mapped service using a TLS certificate, High availability for Knative services overview, Event source in the Administrator perspective, Connecting an event source to a sink using the Developer perspective, Configuring the default broker backing channel, Security configuration for Knative Kafka channels, Listing event sources and event source types, Listing event source types from the command line, Listing event source types from the Developer perspective, Listing event sources from the command line, Setting up OpenShift Serverless Functions, On-cluster function building and deploying, Function project configuration in func.yaml, Accessing secrets and config maps from functions, Configuration for scraping custom metrics, Finding logs for Knative Serving components, Finding logs for Knative Serving services, Using Red Hat OpenShift distributed tracing, Integrating Service Mesh with OpenShift Serverless, Integrating Serverless with the cost management service, Using NVIDIA GPU resources with serverless applications, Preparing to configure the monitoring stack, Moving monitoring components to different nodes, Assigning tolerations to monitoring components, Configuring a local persistent volume claim, Modifying the retention time for Prometheus metrics data, Controlling the impact of unbound metrics attributes in user-defined projects, Setting a scrape sample limit for user-defined projects, Configuring external alertmanager instances, Attaching additional labels to your time series and alerts, Setting log levels for monitoring components, OpenShift Container Platform documentation, Recommended configurable storage technology, Kubernetes documentation on PersistentVolumeClaims, Prometheus remote endpoints and storage documentation, Setting up remote write compatible endpoints, Determining why Prometheus is consuming a lot of disk space.