
NeMo Automodel MoE boosts large-scale PyTorch training
NVIDIA introduced NeMo Automodel MoE, an open-source library that accelerates large-scale Mixture-of-Experts training directly in PyTorch. The release targets teams building billion-parameter systems that need to scale across clusters without custom infrastructure. Moreover, The update lands alongside two notable open tooling moves. NVIDIA detailed a cuVS integration that speeds Faiss vector search on GPUs. The […]







