cnvrg.io is now Intel® Tiber™ AI Studio
Open Nav

Operationalizing AI/ML for the enterprise with Intel® Tiber™ AI Studio and Red Hat OpenShift MLOps solution

openshift-enterprise-mlops-solution

Machine Learning (ML) is rapidly becoming essential to all businesses and organizations around the world. However, this means that IT and DevOps teams are now facing the challenge of standardizing ML workloads and provisioning cloud, on-premise and hybrid compute resources that support the dynamic and intensive workflows that ML jobs and pipelines require. 

Not only that, but between all the many tools, scripts, plug-ins and disconnected stacks, developers and data scientists spend over 65% of their time on DevOps and managing infrastructure resources requests and hybrid cloud compute. This manual and labor intensive work pulls them from doing what they were hired to do – deliver high impact ML models. 

Intel® Tiber™ AI Studio on Red Hat OpenShift delivers an out of the box solution that empowers enterprise data science and DevOps teams to better manage infrastructure in the hybrid cloud and accelerate the ML workflow in one automated and unified platform. On top of Openshift, Intel® Tiber™ AI Studio has also collaborated with NVIDIA NGC’s registry of GPU-optimized AI software to provide a GPU optimized, distributed AI OS solution with NVIDIA GPUs and OpenShift foundations.

cnvrg-openshift-enterprise-mlops-solution

Intel® Tiber™ AI Studio on OpenShift: the ultimate MLOps solution

Red Hat OpenShift, the leading hybrid cloud, enterprise Kubernetes application platform allows data scientists to launch flexible, container-based jobs and pipelines, as well as enabling infrastructure teams to manage and monitor ML workloads in a single managed and cloud-native environment. Intel® Tiber™ AI Studio then enables data scientists to rapidly launch ML workloads on remote clusters without tinkering with infrastructure or complicated configuration. For infrastructure teams, Intel® Tiber™ AI Studio provides the ability to manage all ML compute resources in a unified and secure environment with advanced monitoring and administration capabilities built in.

Intel® Tiber™ AI Studio on OpenShift

How to operationalize Intel® Tiber™ AI Studio on OpenShift via OperatorHub

Intel® Tiber™ AI Studio is operator certified and resides in the RedHat Openshift OperatorHub. It provides everything a Data Science and DevOps team need to manage their ML workflow out-of-the-box:

  • Managed Kubernetes deployment on any cloud or on-premises environment
  • Fully automated installation and life cycle management
  • All tools data scientists need for ML/AI development: from research to deployment
  • Open & flexible, code-first data science platform, which integrates any open source tool

All you need to do is to click on the Intel® Tiber™ AI Studio tab, and the installation and configuration will be done out-of-the-box with OpenShift.

How to operationalize Intel® Tiber™ AI Studio on OpenShift via OperatorHub

Further, Intel® Tiber™ AI Studio will be monitored for anomalies and will be tracked to obtain healthy execution across its lifetime. Logs and metrics will be collected, allowing the operations team to monitor and take proactive actions if failure is expected. 

 

Once in the Intel® Tiber™ AI Studio platform you can begin by starting your own project, or browsing through example projects and start a new workspace. Here’s brief video tutorial to get you started

Utilizing all your compute resources with Intel® Tiber™ AI Studio ML pipelines

Utilizing all your compute resources with Intel® Tiber™ AI Studio ML pipelines

In Intel® Tiber™ AI Studio you can utilize all your compute resources whether on premise or cloud by controlling your jobs by task, and assigning each job to the most effective resource. You can do this quickly in our Flows feature – production-ready machine learning pipelines that allow you to build complex DAG (directed acyclic graph) pipelines and run your ML components (tasks) with just drag-n-drop. Here’s how to create an end to end ML pipeline with Intel® Tiber™ AI Studio.

 

Each task in a flow is an ML component that is fully customizable and can run on different compute resources with different docker images. You can assign each task to a different Kubernetes cluster, location and designating the scale of your resource. For example, you can have feature engineering running on a Spark cluster, followed by a training task running on a GPU instance on AWS. In addition, flows can be versioned, modified, shared, stored, revoked and customized. When a task completes, it frees the resource automatically so it can be utilized elsewhere.

openshift-flow

OpenShift is an excellent foundational technology for Intel® Tiber™ AI Studio, as it provides the ultimate automation for Kubernetes clusters and infrastructure life cycle management. Intel® Tiber™ AI Studio adds  advanced AI/ML capabilities such as model management, rapid experimentation and production ML on top of Openshift Kubernetes infrastructure with a strong and native integration. Together with Red Hat Openshift, Intel® Tiber™ AI Studio gives data scientists and DevOps engineers an out-of-the-box AI solution to accelerate machine learning life cycles, and provides teams agility, flexibility, portability, and scalability to train, deploy, and maintain ML models in production.

Intel® Tiber™ AI Studio is free to deploy, you can get started with Intel® Tiber™ AI Studio in one click via Red Hat OperartorHub.

You can learn more about the benefits of using Intel® Tiber™ AI Studio on Red Hat Openshift here

Top MLOps guides and news in your inbox every month