Files
gt-ai-os-community/docker-compose.x86-gpu.yml
HackWeasel 310491a557 GT AI OS Community v2.0.33 - Add NVIDIA NIM and Nemotron agents
- Updated python_coding_microproject.csv to use NVIDIA NIM Kimi K2
- Updated kali_linux_shell_simulator.csv to use NVIDIA NIM Kimi K2
  - Made more general-purpose (flexible targets, expanded tools)
- Added nemotron-mini-agent.csv for fast local inference via Ollama
- Added nemotron-agent.csv for advanced reasoning via Ollama
- Added wiki page: Projects for NVIDIA NIMs and Nemotron
2025-12-12 17:47:14 -05:00

25 lines
709 B
YAML

# Docker Compose x86_64 GPU Overlay
# Auto-applied by installer when NVIDIA GPU + Container Toolkit detected
# This overlay enables GPU passthrough for the vLLM embeddings container
services:
vllm-embeddings:
deploy:
resources:
# GPU mode: model loads into VRAM (~2.5GB), minimal system RAM needed
limits:
memory: 4G
reservations:
memory: 2G
devices:
- driver: nvidia
count: 1
capabilities: [gpu]
environment:
# GPU-specific settings
- CUDA_VISIBLE_DEVICES=0
- PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True
labels:
- "gt2.gpu=enabled"
- "gt2.gpu.vendor=nvidia"