You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This project is currently under active development, and as such, all source code may not be included in any release. This means that the code is subject to change without notice, and that any information contained within the code should be considered as work in progress.
3
3
4
-
# Intel Technology Enabling for OpenShift
4
+
# Intel® Technology Enabling for OpenShift*
5
5
## General
6
6
7
7
Intel® Technology Enabling for OpenShift* project provides Intel data center and edge hardware features provisioning, related E2E solutions and the reference workloads for these Intel features on Red Hat OpenShift platform.
@@ -11,35 +11,30 @@ The goal of the project is to develop and release open, easy-to-use, integrated,
11
11
To achieve this goal, following OCP software stack development and management life cycle should be followed:
12
12
13
13
* The related software stacks must be designed and implemented properly for Cloud Native and OCP environment from Day 0
14
-
* The solutions and instructions must allow users to configure and provision their OCP cluster properly with technologies from intel and other vendors on Day 1
15
-
* The solutions and instructions to provision intel hardware features and manage the life cycle of containerized software stack must be present in Day 2
16
-
17
-
Multiple other opens source projects are used to provide the solutions. This GitHub project can also be used as an interface for users to access related projects and acquire support from open-source community. People are encouraged to use this project issue to report bugs and request features. Which can help them easily adopt Intel technologies with OCP and cloud-native environments.
18
-
19
-
As the downstream project, this project embodies upstream-first policy. None of the upstream projects will be forked, and no specific patches will be maintained in the project. The proper open-source feature request and bug triage process will be enforced before performing feature design, development, and bug fixes in the specific upstream projects. Everything must be down streamed to the OCP project before being used by the project.
14
+
* The solutions and instructions must allow users to configure and provision their OCP cluster properly with technologies from Intel and other vendors on Day 1
15
+
* The solutions and instructions to provision Intel hardware features and manage the life cycle of containerized software stack must be present in Day 2
20
16
21
17
## Intel Hardware features Provisioning for OpenShift
22
18
23
-
To provision Intel Hardware features on OCP, following open-source projects are used:
19
+
To provision Intel hardware features on OCP, following open-source projects are used:
24
20
***[Node Feature Discovery (NFD)](https://github.com/kubernetes-sigs/node-feature-discovery), [NFD Operator](https://github.com/openshift/cluster-nfd-operator)** are used to automatically label the nodes for Hardware provisioning operation.
25
21
***[Machine Config Operator (MCO)](https://github.com/openshift/machine-config-operator)** is used to configure the Red Hat Enterprise Linux Core OS (RHCOS) on the nodes.
26
22
***[Kernel Module Management (KMM)](https://github.com/kubernetes-sigs/kernel-module-management), [KMM Operator](https://github.com/rh-ecosystem-edge/kernel-module-management)** are used to manage deployment and lifecycle of Intel Data Center GPU Driver.
27
-
***[Intel Data Center GPU Driver For OpenShift](https://github.com/intel/intel-data-center-gpu-driver-for-openshift)** use **[Intel GPU Drivers](https://github.com/intel-gpu)** build, package certificate and release Intel dGPU driver container images for OCP.
23
+
***[Intel Data Center GPU Driver For OpenShift](https://github.com/intel/intel-data-center-gpu-driver-for-openshift)** use **[Intel GPU Drivers](https://github.com/intel-gpu)** build, package, certify and release Intel Data Center GPU driver container images for OCP.
28
24
***[Intel Device Plugins for Kubernetes project](https://github.com/intel/intel-device-plugins-for-kubernetes)** Provides Intel GPU/SGX/QAT device plugins images and the operator to deploy and manage the life cycle of these device plugins
## [Intel AI Inference E2E Solution for OpenShift](e2e/inference/README.md)
45
40
@@ -50,7 +45,7 @@ Below features are under consideration to be included in the future releases
50
45
* Third Generation Intel® Xeon® Scalable Processors (or later version) are used by the cluster.
51
46
* Contact your server or BIOS vendor for the BIOS setting to enable the feature.
52
47
53
-
### Intel® dGPU Card supported platform
48
+
### Intel® Data Center GPU Card supported platform
54
49
55
50
* The Intel® Data Center GPU Flex Series 140 or Intel® Data Center GPU Flex Series 170 Card is enabled on the nodes.
56
51
* Contact your server or BIOS vendor for the BIOS setting to enable the cards.
@@ -60,30 +55,30 @@ Below features are under consideration to be included in the future releases
60
55
* Contact your server or BIOS vendor for the BIOS setting to enable the feature.
61
56
62
57
## Get Started
63
-
To properly provision Intel Hardware features, deploy and manage the related e2e solutions as well as the reference workload, below OCP software stack development and life cycle management flow is followed by this project
58
+
To properly provision Intel hardware features, deploy and manage the related E2E solutions as well as the reference workload, below OCP software stack development and life cycle management flow is followed by this project
64
59
65
60
### Day 0 - Define the requirements of the OCP platform and design it.
66
61
Red Hat [OpenShift Operator](https://www.redhat.com/en/technologies/cloud-computing/openshift/what-are-openshift-operators) automates the creation, configuration, and management of instances of Kubernetes-native applications. It is based on [Kubernetes operator pattern](https://kubernetes.io/docs/concepts/extend-kubernetes/operator/). So the software stack running on OCP needs to be containerized and managed by a specific Operator. As an essential part of OCP, operators need to be well-designed from Day 0. Good examples are [Intel Device Plugins Operator](https://github.com/intel/intel-device-plugins-for-kubernetes) and [KMM Operator](https://github.com/rh-ecosystem-edge/kernel-module-management).
67
62
68
-
### Day 1 - provision the OCP platform and configure it to a working state.
69
-
This project mainly focuses on bare metal OCP cluster. [DCI](https://doc.distributed-ci.io/dci-openshift-agent/) is used to provision the bare metal OCP cluster. Users can also refer to [bare metal OCP cluster installation instructions](https://docs.openshift.com/container-platform/4.11/installing/installing_bare_metal_ipi/ipi-install-overview.html) to install the bare metal OCP cluster.
63
+
### Day 1 - Provision the OCP platform and configure it to a working state.
64
+
This project mainly focuses on bare metal OCP cluster. [Distributed CI](https://doc.distributed-ci.io/dci-openshift-agent/) is used to provision the bare metal OCP cluster. Users can also refer to [bare metal OCP cluster installation instructions](https://docs.openshift.com/container-platform/4.12/installing/installing_bare_metal_ipi/ipi-install-overview.html) to install the bare metal OCP cluster.
70
65
71
-
To avoid rebooting the nodes and some other issues on Day 2, Some Machine Configurations operations can be enforced on day 1 when provisioning the cluster. The related discussion is ongoing.
66
+
To avoid rebooting the nodes and some other issues on Day 2, Some machine configurations operations can be enforced on day 1 when provisioning the cluster. The related discussion is ongoing.
72
67
73
68
### Day 2.0 - Zero O’clock of Day 2
74
-
The Day 2.0 concept is introduced for users to provision intel hardware features right after provisioning an OCP cluster and before any user workloads are deployed. Refer to the steps below to provision Intel hardware features:
69
+
The Day 2.0 concept is introduced for users to provision Intel hardware features right after provisioning an OCP cluster and before any user workloads are deployed. Refer to the steps below to provision Intel hardware features:
75
70
76
-
***[Deploy Node Feature Descovery on OpenShift](nfd/README.md#steps-to-install-and-configure-nfd-operator-on-ocp-cluster)**
71
+
***[Deploy Node Feature Discovery on OpenShift](nfd/README.md#steps-to-install-and-configure-nfd-operator-on-ocp-cluster)**
77
72
***[Setup Machine Configuration on OpenShift](machine_configuration/README.md#general-configuration-for-provisioning-intel-hardware-features)**
78
73
79
74
Note: Running the above steps on Day 2.0 is recommended. However, if you want to provision the features above with the existing cluster on day 2, please be advised that some machine configuration operations might trigger the pods to drain and reboot the nodes. Some of the ongoing efforts in the MCO upstream are to set the machine configurations without rebooting.
80
75
81
76
### Day 2 - OCP platform is installed and ready to begin providing services.
82
-
Multiple operators are used to provision Intel hardware features and deploy, manage the e2e solutions, and reference workloads.
77
+
Multiple operators are used to provision Intel hardware features and deploy, manage the E2E solutions, and reference workloads.
83
78
84
79
**Provisioning Intel Hardware features on OpenShift**
85
80
***[Deploy Intel Data Center GPU Driver on OpenShift](kmmo/README.md#managing-intel-dgpu-driver-with-kmm-operator)**
86
81
***[Deploy Intel Device Plugins on OpenShift](device_plugins/README.md#deploy-intel-device-plugins-on-openshift)**
87
82
88
83
**Deploy E2E Solution**
89
-
***[Deploy Intel AI Reference E2E Solution](e2e/inference/README.md#deploy-intel-ai-inference-e2e-solution)**
84
+
***[Deploy Intel AI Reference E2E Solution](e2e/inference/README.md#deploy-intel-ai-inference-e2e-solution)**
0 commit comments