Skip to content
#

ai-stack

Here are 14 public repositories matching this topic...

Deploy a complete, self-hosted AI stack on your own server with one command. Includes Ollama (LLM), LiteLLM (AI gateway), Whisper (STT), Kokoro (TTS), Embeddings (RAG), and MCP Gateway. Most services run locally; LiteLLM optionally routes to external providers. Supports NVIDIA GPU (CUDA) acceleration.

  • Updated May 6, 2026

The **Podman AI Stack** is a secure, configurable, and systemd-native orchestration stack for deploying containerized AI environments (Open WebUI and Ollama). It leverages **Podman Quadlets** to provide a seamless integration with systemd, supporting both rootless and rootfull deployments on Fedora and other RPM-based distributions.

  • Updated Apr 21, 2026
  • Shell

Improve this page

Add a description, image, and links to the ai-stack topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the ai-stack topic, visit your repo's landing page and select "manage topics."

Learn more