ai/chat-demo
Full AI Chat Application stack, including the backend API, frontend interface, and model server.
345
This Compose setup lets you easily launch the full AI Chat Application stack, including the backend API, frontend interface, and model server. Ideal for real-time AI interactions, it runs fully containerized with Docker Compose.
latest
tag for CPU compatibility.gpu-latest
tag for GPU support.Note: This application requires at least 8 GB of available memory within the Docker VM and performs best on systems with at least 16 GB of available memory. Systems with higher memory, like 32 GB or more, will experience smoother performance, especially when handling intensive AI tasks.
Use the "View Compose" button on this page to download the file, saving it as compose.yaml
(recommended naming convention).
CPU Mode:
Copy the compose file from the latest
tag with the View Compose
button and save it as compose.yaml
.
Run the following command from the same directory as your compose.yaml
file:
docker compose up
GPU Mode:
To enable GPU support:
Ensure the nvidia
driver is installed.
Copy the compose file from the gpu-latest
tag with the View Compose
button and save it as compose.yaml
.
Run the following command from the same directory as your compose.yaml
file:
docker compose up
Specify a different model by setting the MODEL
variable:
MODEL=llama3.2:latest docker compose up
Note: Always include the model tag, even when using latest.
ai/chat-demo-backend:latest
8000:8000
MODEL_HOST=http://ollama:11434
ai/chat-demo-frontend:latest
3000:3000
PORT=3000
HOST=0.0.0.0
ai/chat-demo-model:latest
11434:11434
MODEL=${MODEL:-mistral:latest}
ollama_data
: Stores model data.GPU Requirement: Ensure your system has a compatible NVIDIA GPU if using GPU mode.
The Compose file includes a volume:
ollama_data