
llama stack demos
Collection of demos for building Llama Stack based apps on OpenShift
Repository Info
About This Server
Collection of demos for building Llama Stack based apps on OpenShift
Model Context Protocol (MCP) - This server can be integrated with AI applications to provide additional context and capabilities, enabling enhanced AI interactions and functionality.
Documentation
Llama Stack Demos on OpenDataHub
This repository contains practical examples and demos designed to get you started quickly building AI apps with Llama Stack on Kubernetes or OpenShift. Whether you're a cluster admin looking to deploy the right GenAI infrastructure or a developer eager to innovate with AI Agents, the content in this repo should help you get started.
🛠️ Get Started
Requirements
To run these demos, ensure your environment meets the following:
- OpenShift Cluster 4.17+
- 2 GPUs with a minimum of 40GB VRAM each.
Deployment Instructions
Next, follow these simple steps to deploy the core components:
- Create a dedicated OpenShift project:
oc new-project llama-serve - Apply the Kubernetes manifests:
This will deploy the foundational Llama Stack services, vLLM model servers, and MCP tool servers.oc apply -k kubernetes/kustomize/overlay/all-models
Setting Up Your Development Environment
We use uv for managing Python dependencies, ensuring a consistent and efficient development experience. Here's how to get your environment ready:
- Install
uv:pip install uv - Synchronize your environment:
uv sync - Activate the virtual environment:
source .venv/bin/activate
Now you're all set to run any Python scripts or Jupyter notebooks within the demos/rag_agentic directory!
💡 Demo Architecture
The below diagram is an example architecture for a secure Llama Stack based application deployed on OpenShift (OCP) using both MCP tools and a Milvus vectorDB for its agentic and RAG based workflows. This is the same architecture that has been implemented in the RAG/Agentic demos.
!Architecture Diagram
We're excited to see what you build with Llama Stack! If you have any questions or feedback, please don't hesitate to open an issue. Happy building! 🎉
Quick Start
Clone the repository
git clone https://github.com/opendatahub-io/llama-stack-demosInstall dependencies
cd llama-stack-demos
npm installFollow the documentation
Check the repository's README.md file for specific installation and usage instructions.
Repository Details
Recommended MCP Servers
Discord MCP
Enable AI assistants to seamlessly interact with Discord servers, channels, and messages.
Knit MCP
Connect AI agents to 200+ SaaS applications and automate workflows.
Apify MCP Server
Deploy and interact with Apify actors for web scraping and data extraction.
BrowserStack MCP
BrowserStack MCP Server for automated testing across multiple browsers.
Zapier MCP
A Zapier server that provides automation capabilities for various apps.