SUSE AI Deployment

Course SAI211v1

This course introduces the SUSE AI open infrastructure platform used to deploy and run GenAI workloads. It covers the deployment of the base operating system and cluster that will run the AI middleware, vector database and AI workloads. Once the base OS and Kubernetes cluster are deployed and imported into Rancher Prime, SUSE Storage is installed as the storage layer. At that point the AI related components such as the AI middleware (Ollama), vector database (Milvus) and AI chat application (Open WebUI) are installed. Finally SUSE Security is installed into the AI Cluster enabling the cluster to be secured and the SUSE Observability and OpenTelemetry agents are installed enabling the workloads to be observed.

Available in 6 languages, the source content is created in English. All translations are machine-generated and therefore might contain minor errors.

This course helps prepare students for the SUSE Certified Deployment Specialist (SCDS) in SUSE AI certification exam.

Downloadable Course Description

Course Information

Delivery Method: eLearning

Course Duration: Approximately 8 hours

Course Level: Intermediate

Key Objectives

Attendees will be taught the following concepts and skills:

  • Basic AI concepts and terminology
  • An understanding of the SUSE AI components
  • How to install and configure the base operating system and Kubernetes cluster for AI workloads
  • How to install install common AI middleware, vector databases and applications such as Ollama, Milvus and Open WebUI

Audience Summary

  • Consultants who regularly deploy the product for their customers.
  • Anyone who wants to learn how to deploy and begin to use SUSE AI.

Course Prerequisites

Attendees should have solid experience installing and managing SUSE Linux Enterprise Server or SUSE Linux Micro. They should also have experience using Kubernetes to deploy and manage containerized workloads.

Course Outline

  • Section 1: Course Overview
  • Section 2: Introduction to SUSE AI
    • Overview of AI Concepts and Terminology
    • AI Platform Challenges Addressed by SUSE AI
    • Overview of SUSE AI Components
  • Section 3: Operating System Configuration Requirements for an AI Cluster
    • Host OS Installation and Configuration
  • Section 4: Kubernetes Deployment and Configuration Requirements
    • RKE2 AI Cluster Requirements
    • RKE2 Deployment and Configuration
    • RKE2 Integration with Rancher Prime
  • Section 5: GPU Operator Deployment
    • NVIDIA GPU Operator Installation
  • Section 6: Storage Requirements for an AI Cluster
    • Storage Layer Installation and Configuration
  • Section 7: Introduction to the SUSE Application Collection
    • Overview of the Application Collection
  • Section 8: Ollama
    • Overview of Ollama and LLMs
  • Section 9: RAG and Vector Databases
    • Overview of RAG and Vector Databases
    • Install Milvus
  • Section 10: Open WebUI
    • Overview of Open WebUI
    • Install Open WebUI
  • Section 11: Secure AI Clusters
    • Overview of SUSE Security
    • Install SUSE Security into an AI Cluster
  • Section 12: Observe AI Workloads
    • How SUSE Observability Displays Information about AI Workloads
    • Configure GenAI Observability

Training Options