GT AI OS Community v2.0.33 - Add NVIDIA NIM and Nemotron agents
- Updated python_coding_microproject.csv to use NVIDIA NIM Kimi K2 - Updated kali_linux_shell_simulator.csv to use NVIDIA NIM Kimi K2 - Made more general-purpose (flexible targets, expanded tools) - Added nemotron-mini-agent.csv for fast local inference via Ollama - Added nemotron-agent.csv for advanced reasoning via Ollama - Added wiki page: Projects for NVIDIA NIMs and Nemotron
This commit is contained in:
24
docker-compose.x86-gpu.yml
Normal file
24
docker-compose.x86-gpu.yml
Normal file
@@ -0,0 +1,24 @@
|
||||
# Docker Compose x86_64 GPU Overlay
|
||||
# Auto-applied by installer when NVIDIA GPU + Container Toolkit detected
|
||||
# This overlay enables GPU passthrough for the vLLM embeddings container
|
||||
|
||||
services:
|
||||
vllm-embeddings:
|
||||
deploy:
|
||||
resources:
|
||||
# GPU mode: model loads into VRAM (~2.5GB), minimal system RAM needed
|
||||
limits:
|
||||
memory: 4G
|
||||
reservations:
|
||||
memory: 2G
|
||||
devices:
|
||||
- driver: nvidia
|
||||
count: 1
|
||||
capabilities: [gpu]
|
||||
environment:
|
||||
# GPU-specific settings
|
||||
- CUDA_VISIBLE_DEVICES=0
|
||||
- PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True
|
||||
labels:
|
||||
- "gt2.gpu=enabled"
|
||||
- "gt2.gpu.vendor=nvidia"
|
||||
Reference in New Issue
Block a user