Ujval Kapasi – NVIDIA Technical Blog News and tutorials for developers, data scientists, and IT admins 2025-06-12T18:48:42Z http://www.open-lab.net/blog/feed/ Ujval Kapasi <![CDATA[How NVIDIA GB200 NVL72 and NVIDIA Dynamo Boost Inference Performance for MoE Models]]> http://www.open-lab.net/blog/?p=101457 2025-06-12T18:48:42Z 2025-06-06T19:00:00Z The latest wave of open source large language models (LLMs), like DeepSeek R1, Llama 4, and Qwen3, have embraced Mixture of Experts (MoE) architectures. Unlike...]]>

The latest wave of open source large language models (LLMs), like DeepSeek R1, Llama 4, and Qwen3, have embraced Mixture of Experts (MoE) architectures. Unlike traditional dense models, MoEs activate only a subset of specialized parameters—known as experts—during inference. This selective activation reduces computational overhead, leading to faster inference times and lower deployment costs.

Source

]]>
1
Ujval Kapasi <![CDATA[New Optimizations To Accelerate Deep Learning Training on NVIDIA GPUs]]> http://www.open-lab.net/blog/?p=12964 2023-02-13T17:46:37Z 2018-12-03T16:00:36Z The pace of AI adoption across diverse industries depends on maximizing data scientists�� productivity. NVIDIA releases optimized NGC containers every month...]]>

The pace of AI adoption across diverse industries depends on maximizing data scientists’ productivity. NVIDIA releases optimized NGC containers every month with improved performance for deep learning frameworks and libraries, helping scientists maximize their potential. NVIDIA continuously invests in the full data science stack, including GPU architecture, systems, and software stacks.

Source

]]>
0
���˳���97caoporen����