Product Documentation for Red Hat AI Inference Server 3.2
Version:
Welcome
-
Release notes
Highlights of what is new and what has changed with this Red Hat AI Inference Server release -
Getting started
Getting started with Red Hat AI Inference Server -
Deploying Red Hat AI Inference Server in a disconnected environment
Deploy Red Hat AI Inference Server in a disconnected environment using OpenShift Container Platform and a disconnected mirror image registry -
Deploying Red Hat AI Inference Server in OpenShift Container Platform
Deploy Red Hat AI Inference Server in OpenShift Container Platform clusters that have supported AI accelerators installed -
Supported product and hardware configurations
Supported hardware and software configurations for deploying Red Hat AI software -
Validated models
Red Hat AI validated models
Inference Operations
-
Inference serving language models in OCI-compliant model containers
Inferencing OCI-compliant models in Red Hat AI Inference Server -
vLLM server arguments
Server arguments for running Red Hat AI Inference Server -
Red Hat AI Model Optimization Toolkit
Compressing large language models with the LLM Compressor library -
Extending Red Hat AI Inference Server with tool calling capabilities
Configuring tool calling and chat templates for AI Inference Server
Related Products
-
Red Hat OpenShift AI
Find release notes and product documentation for using the OpenShift AI platform and its integrated MLOps capabilities to manage predictive and generative AI models at scale -
Red Hat Enterprise Linux AI
Switch to the Red Hat Enterprise Linux AI documentation
Additional Resources
-
Product life cycle
Understand the product life cycle to plan deployments and support applications using the product -
This content is not included.Red Hat AI learning hub
Explore a curated collection of learning resources designed to help you accomplish key tasks with Red Hat AI products and services