GT AI OS Community v2.0.33 - Add NVIDIA NIM and Nemotron agents

- Updated python_coding_microproject.csv to use NVIDIA NIM Kimi K2
- Updated kali_linux_shell_simulator.csv to use NVIDIA NIM Kimi K2
  - Made more general-purpose (flexible targets, expanded tools)
- Added nemotron-mini-agent.csv for fast local inference via Ollama
- Added nemotron-agent.csv for advanced reasoning via Ollama
- Added wiki page: Projects for NVIDIA NIMs and Nemotron
This commit is contained in:
HackWeasel
2025-12-12 17:47:14 -05:00
commit 310491a557
750 changed files with 232701 additions and 0 deletions

View File

@@ -0,0 +1,24 @@
# Docker Compose x86_64 GPU Overlay
# Auto-applied by installer when NVIDIA GPU + Container Toolkit detected
# This overlay enables GPU passthrough for the vLLM embeddings container
services:
vllm-embeddings:
deploy:
resources:
# GPU mode: model loads into VRAM (~2.5GB), minimal system RAM needed
limits:
memory: 4G
reservations:
memory: 2G
devices:
- driver: nvidia
count: 1
capabilities: [gpu]
environment:
# GPU-specific settings
- CUDA_VISIBLE_DEVICES=0
- PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True
labels:
- "gt2.gpu=enabled"
- "gt2.gpu.vendor=nvidia"