Today, AWS announced the general availability of the new Amazon EC2 G5 instances, powered by NVIDIA A10G Tensor Core GPUs. These instances are designed for the most demanding graphics-intensive applications, as well as machine learning inference and training simple to moderately complex machine learning models on the AWS cloud. The new EC2 G5 instances feature up to eight NVIDIA A10G Tensor��
]]>TIME magazine today named NVIDIA Omniverse one of the 100 Best Inventions of 2021, saying the project is ��making it easier to create ultra-realistic virtual spaces for��real-world purposes.�� Omniverse �� a scalable, multi-GPU, real-time reference development platform for 3D simulation and design collaboration �� is being evaluated by more than 700 companies and 70,000��
]]>The Nsight suite of Developer Tools provide insightful tracing, debugging, profiling, and other analyses to optimize your complex computational applications across NVIDIA GPUs, and CPUs including x86, Arm, and Power architectures. NVIDIA Nsight Systems is a performance analysis tool designed to visualize, analyze and optimize programming models, and tune to scale efficiently across any��
]]>Picture this: you��re having dinner at an upscale restaurant. You look at the menu and decide that you��re in the mood for a filet. You order the steak medium rare. The waiter brings it out, it��s plated beautifully, and the service is great. Yet you hear a little voice in the back of your head. ��I could have prepared this steak in my own kitchen exactly to my definition of medium rare!
]]>NVIDIA Nsight Deep Learning Designer is a new tool that helps ease the process of performant model design. DL Designer provides valuable insights into the structure of the model, and how well it performs on NVIDIA hardware. Models can be created with a user-friendly, drag-and-drop interface that features nodes for all of the commonly used operators available in the most popular deep learning��
]]>Today, NVIDIA announced Nsight Graphics 2021.5, the latest release, which supports Direct3D (11, 12, and DXR), Vulkan (1.2, NV Vulkan Ray Tracing Extension), OpenGL, OpenVR, and the Oculus SDK. Nsight Graphics is a standalone developer tool that enables you to debug, profile, and export frames built with high-fidelity 3D graphics applications. Developers now have full support for��
]]>Data scientists wrestle with many challenges that slow development. There are operational tasks, including software stack management, installation, and updates that impact productivity. Reproducing state-of-the-art assets can be difficult as modern workflows include many tedious and complex tasks. Access to the tools you need is not always fast or convenient. Also, the use of multiple tools and��
]]>The latest update to NVIDIA Nsight Systems��a performance analysis tool��is now available for download. Designed to help you tune and scale software across CPUs and GPUs, this release introduces several improvements aimed to enhance the profiling experience. Nsight Systems is part of the powerful debugging and profiling NVIDIA Nsight Tools Suite. You can start with Nsight Systems for an overall��
]]>The modern data center is becoming increasingly difficult to manage. There are billions of possible connection paths between applications and petabytes of log data. Static rules are insufficient to enforce security policies for dynamic microservices, and the sheer magnitude of log data is impossible for any human to analyze. AI provides the only path to the secure and self-managed data��
]]>We have been pursuing the creation of digital humans for years. Traditionally, digital humans have been widely used in media and entertainment, from video game characters to CGI characters in movies. But the process to create a digital human is extremely labor-intensive and manual. It requires hundreds of domain experts such as artists, programmers, and technical artists. Plus��
]]>At NVIDIA GTC this November, new software tools were announced that help developers build real-time speech applications, optimize inference for a variety of use-cases, optimize open-source interoperability for recommender systems, and more. Watch the keynote from CEO, Jensen Huang, to learn about the latest NVIDIA breakthroughs. Today, NVIDIA unveiled a new version of NVIDIA Riva with a��
]]>AI pioneer Andrew Ng is calling for a broad shift to a more data-centric approach to machine learning (ML). He recently held the first data-centric AI competition on data quality, which many claim represents 80% of the work in AI. ��I��m optimistic that the AI community before long will take as much interest in systematically improving data as architecting models,�� Ng wrote in his newsletter��
]]>NVIDIA recently unveiled new breakthroughs in NVIDIA Riva for speech AI, and NVIDIA NeMo for large-scale language modeling (LLM). Riva is a GPU-accelerated Speech AI SDK for enterprises to generate expressive human-like speech for their brand and virtual assistants. NeMo is an accelerated training framework for speech and NLU, that now has the capabilities to develop large-scale language models��
]]>Reallusion has released the iClone Connector for NVIDIA Omniverse, adding a complete character animation pipeline to Omniverse Create and Omniverse Machinima applications. When the Character Creator 3 Omniverse Connector launched in April 2021, creators of all skill levels could efficiently and easily create digital humans for any type of project. Reallusion now enhances the pipeline with the��
]]>Sign up for the latest Speech AI news from NVIDIA. Conversational AI is a set of technologies enabling human-like interactions between humans and devices based on the most natural interfaces for us: speech and natural language. Systems based on conversational AI can understand commands by recognizing speech and text, translating on-the-fly between different languages��
]]>In the past several months, many of us have grown accustomed to seeing our doctors over a video call. It��s certainly convenient, but after the call ends, those important pieces of advice from your doctor start to slip away. What was that new medication I needed to take? Were there any side effects to watch out for? Conversational AI can help in building an application to transcribe speech as��
]]>Sign up for the latest Speech AI news from NVIDIA. Virtual assistants have become part of our daily lives. We ask virtual assistants almost anything that we wonder about. In addition to providing convenience to our daily lives, virtual assistants are of tremendous help when it comes to enterprise applications. For example, we use online virtual agents to help navigate complex technical issues��
]]>Sign up for the latest Speech AI news from NVIDIA. There is a high chance that you have asked your smart speaker a question like, ��How tall is Mount Everest?�� If you did, it probably said, ��Mount Everest is 29,032 feet above sea level.�� Have you ever wondered how it found an answer for you? Question answering (QA) is loosely defined as a system consisting of information retrieval (IR)��
]]>Today NVIDIA announced the availability of NVIDIA PhysicsNeMo (previously known as SimNet), a platform to train neural networks using governing physics equations along with observed or simulated data. The robust and high-fidelity models produced by the PhysicsNeMo framework enable the acceleration of design exploration for multiphysics systems�Cideal for digital twin development.
]]>Quantum computing (QC) will drive a new wave of advances across a myriad of industries and domains. Simulating quantum circuits on the current class of computers is imperative to technology evolving to the next level of quantum computing in the future. NVIDIA cuQuantum is an SDK of optimized libraries and tools for accelerating quantum computing workflows. Developers can use cuQuantum while��
]]>The NVIDIA Jetson AGX Orin��the world��s smallest, most powerful, and energy-efficient AI supercomputer for robotics, autonomous machines, and edge computing applications��was announced today at NVIDIA GTC. The new Jetson single-chip computer is so small you can fit it in the palm of your hand, and yet it can deliver processing power similar to a GPU-enabled server. It maintains the same form��
]]>This post is part of a series about generating accurate speech transcription. For part 1, see Speech Recognition: Generating Accurate Domain-Specific Audio Transcriptions Using NVIDIA Riva. For part 2, see Speech Recognition: Customizing Models to Your Domain Using Transfer Learning. NVIDIA Riva is an AI speech SDK for developing real-time applications like transcription, virtual assistants��
]]>This post is part of a series about generating accurate speech transcription. For part 1, see Speech Recognition: Generating Accurate Transcriptions Using NVIDIA Riva. For part 3, see Speech Recognition: Deploying Models to Production. Creating a new AI deep learning model from scratch is an extremely time�C and resource-intensive process. A common solution to this problem is to employ��
]]>Today, developers can access the latest versions of RTX Global Illumination (RTXGI), RTX Direct Illumination (RTXDI), NVIDIA Real-Time Denoisers (NRD), and OptiX Ray Tracing Engine (OptiX). The updates are available through NVIDIA SDKs and select technologies through the NVIDIA Branch of Unreal Engine (NvRTX). Unreal Engine is the world��s most open and advanced real-time 3D creation platform.
]]>Join the NVIDIA Triton and NVIDIA TensorRT community to stay current on the latest product updates, bug fixes, content, best practices, and more. As of March 18, 2025, NVIDIA Triton Inference Server is now part of the NVIDIA Dynamo Platform and has been renamed to NVIDIA Dynamo Triton, accordingly. AI is a new way to write software and AI inference is running this software.
]]>With computation shifting from the CPU to faster GPUs for AI, ML and HPC applications, IO into and out of the GPU can become the primary bottleneck to the overall application performance. NVIDIA created Magnum IO GPUDirect Storage (GDS) to streamline data movement between storage and GPU memory and remove performance bottlenecks in the platform, like being forced to store and forward data��
]]>Watch NVIDIA founder and CEO Jensen Huang��s GTC keynote address streaming on Nov. 9 and in replay. Tune in to a healthcare special address by Kimberly Powell, NVIDIA VP of healthcare, on Nov. 9 at 10:30 a.m. Pacific. Subscribe to NVIDIA healthcare news. NVIDIA Clara Holoscan is the AI computing platform for medical devices that combines hardware systems for low-latency sensor and network��
]]>Deploying AI applications at the edge over a 5G connectivity fabric is an important step in driving digital transformation for enterprise operations. AI is reshaping how enterprises across every industry segment are doing their business. 5G offers superfast, wide-area, secure, low-latency and ultra-reliable connectivity, enabling enterprises to link their AI systems, whether on premises��
]]>Today, NVIDIA announced that it will help developers, researchers, and data scientists working with Graph Neural Networks (GNN) on large heterogeneous graphs with billions of edges by providing GPU-accelerated Deep Graph Library (DGL) containers. These containers will enable developers to work more efficiently in an integrated, GPU-accelerated environment that combines DGL and PyTorch.
]]>DeepStream SDK 6.0 is now available for download. A powerful AI streaming analytics toolkit, DeepStream helps developers build high-performance, low-latency, complex, video analytics applications, and services. This release introduces Graph Composer, a new low-code programming tool that enhances the DeepStream user experience. DeepStream and Graph Composer abstract the low-level��
]]>Traditional cybersecurity methods include creating barriers around your infrastructure to protect it from intruders with ill intentions. However, as enterprises continue along the path of digital transformation, faced with a proliferation of devices, more sophisticated cybersecurity attacks, and an incredibly vast network of data to protect, new cybersecurity methodologies must be explored.
]]>Supercomputers are significant investments. However they are extremely valuable tools for researchers and scientists. To effectively and securely share the computational might of these data centers, NVIDIA introduced the Cloud-Native Supercomputing architecture. It combines bare metal performance, multitenancy, and performance isolation for supercomputing. Magnum IO, the I/
]]>Today NVIDIA introduced the NVIDIA DOCA 1.2 software for NVIDIA BlueField DPUs, the world��s most advanced data processing unit (DPU). This latest release, scheduled for late November, builds on the momentum of the NVIDIA DOCA early access program to enable partners and customers to accelerate the development of applications and holistic zero trust solutions on the DPU. New authentication��
]]>NVIDIA Fleet Command is a cloud service that securely deploys, manages, and scales AI applications across distributed edge infrastructure. Since Fleet Command launched in July, several significant milestones have been achieved and are showcased at NVIDIA GTC. New features are constantly added to Fleet Command. In addition to making the platform more robust and secure��
]]>University of Pennsylvania researchers have used convolutional neural networks to catalog the morphology of 27 million galaxies, giving astronomers a massive dataset for studying the evolution of the universe. ��Galaxy morphology is one of the key aspects of galaxy evolution,�� said study author Helena Dom��nguez S��nchez, former postdoc at Penn. ��The shape and structure of galaxies has a lot of��
]]>CUDA is the software development platform for building GPU-accelerated applications, providing all the components you need to develop applications that use NVIDIA GPUs. CUDA is ideal for diverse workloads from high performance computing, data science analytics, and AI applications. The latest release, CUDA 11.3, and its features are focused on enhancing the programming model and performance of��
]]>The CUDA 11.3 release of the CUDA C++ compiler toolchain incorporates new features aimed at improving developer productivity and code performance. NVIDIA is introducing cu++flt, a standalone demangler tool that allows you to decode mangled function names to aid source code correlation. Starting with this release, the NVRTC shared library versioning scheme is relaxed to facilitate compatible��
]]>Tensor Cores, which are programmable matrix multiply and accumulate units, were first introduced in the V100 GPUs where they operated on half-precision (16-bit) multiplicands. Tensor Core functionality has been expanded in the following architectures, and in the Ampere A100 GPUs (compute capability 8.0) support for other data types was added, including double precision.
]]>We are pleased to announce that Ray Tracing Gems II, the follow up to 2019��s Ray Tracing Gems, will be available for digital download and print on August 4th, 2021. Today, as nearly every hardware vendor and 3D software platform embraces ray tracing, it is clear that real-time ray tracing is here to stay. Ray Tracing Gems II brings the community of rendering experts back together again to��
]]>Finding ways to improve performance and visual fidelity in your games and applications is challenging. To help during the game development process, NVIDIA has packaged and released a suite of SDKs through our branch of Unreal Engine for all developers, from independent to AAA, to harness the power of RTX. Today, NVIDIA released RTX Technology Showcase �C an interactive demo built from NVIDIA��s��
]]>Unity made real-time ray tracing available to all of their developers in 2019 with the release of 2019LTS. Before the end of 2021, NVIDIA DLSS (Deep Learning Super Sampling) will be natively supported for HDRP in Unity 2021.2. NVIDIA DLSS uses advanced AI rendering to produce image quality that��s comparable to native resolution�Cand sometimes even better�Cwhile only conventionally rendering a��
]]>GTC is a great opportunity to get hands-on with NVIDIA��s latest graphics technologies. Developers can apply now for access to RTX Direct Illumination (RTXDI), the latest advancement in real-time ray tracing. Nsight Perf, the next in a line of developer optimization tools, has just been made available to all members of the NVIDIA Developer Program. In addition, several exciting updates to aid game��
]]>Join the NVIDIA Triton and NVIDIA TensorRT community to stay current on the latest product updates, bug fixes, content, best practices, and more. When you are working on optimizing inference scenarios for the best performance, you may underestimate the effect of data preprocessing. These are the operations required before forwarding an input sample through the model. This post highlights the��
]]>Data science development faces many challenges in the areas of: Some estimates point to 70%-90% of the time is spent on experimentation �C much of which will run fast and efficiently on GPU-enabled mobile and desktop workstations. Running on a Linux mobile workstation, for example, presents another set of challenges �C including installing and configuring a data science stack��
]]>At GTC 2021, NVIDIA announced new software tools to help developers build optimized conversational AI, recommender, and video solutions. Watch the keynote from CEO, Jensen Huang, for insights on all of the latest GPU technologies. Today NVIDIA announced major conversational AI capabilities in NVIDIA Riva that will help enterprises build engaging and accurate applications for their��
]]>Conversational AI is opening new ways for enterprises to interact with customers in every industry using applications like real-time transcription, translation, chatbots, and virtual assistants. Building domain-specific interactive applications requires state-of-the-art models, optimizations for real-time performance, and tools to adapt those models with your data. This week at GTC��
]]>Join the NVIDIA Triton and NVIDIA TensorRT community to stay current on the latest product updates, bug fixes, content, best practices, and more. AI machine learning is unlocking breakthrough applications in fields such as online product recommendations, image classification, chatbots, forecasting, and manufacturing quality inspection. There are two parts to AI: training and inference.
]]>Data plays a crucial role in creating intelligent applications. To create an efficient AI/ ML app, you must train machine learning models with high-quality, labeled datasets. Generating and labeling such data from scratch has been a critical bottleneck for enterprises. Many companies prefer a one-stop solution to support their AI/ML workflow from data generation, data labeling, model training/
]]>NVIDIA Omniverse Audio2Face will be available later this week in open beta. With the Audio2Face app, Omniverse users can generate AI-driven facial animation from audio sources. The demand for digital humans is increasing across industries, from game development and visual effects to conversational AI and healthcare. But the animation process is tedious, manual and complex��
]]>3D deep learning researchers can enter NVIDIA Omniverse and simplify their workflows with the Omniverse Kaolin app, now available in open beta. The Omniverse platform provides researchers, developers, and engineers with the ability to virtually collaborate and work between different software applications. Omniverse Kaolin is an interactive application that acts as a companion to the NVIDIA��
]]>OptiX 7.3 brings temporal denoising and improvements to OptiX Curves primitives and new features to the OptiX Demand Loading library NVIDIA Optix Ray Tracing Engine is a scalable and seamless framework that offers optimal ray tracing performance on GPUs. In this spring��s update to the OptiX SDK, developers will be able to leverage temporal denoising, faster curve intersectors��
]]>Quantum computing has the potential to offer giant leaps in computational capabilities. Until it becomes a reality, scientists, developers, and researchers are simulating quantum circuits on classical computers. NVIDIA cuQuantum is an SDK of optimized libraries and tools for accelerating quantum computing workflows. Developers can use cuQuantum to speed up quantum circuit simulations based��
]]>NVIDIA announced the NVIDIA Arm HPC Developer Kit, an integrated hardware-software platform for creating, evaluating, and benchmarking HPC, AI and scientific computing applications. Developers can apply today for the early interest program. This DevKit targets heterogeneous GPU/CPU system development, and includes an Arm CPU, an NVIDIA A100 Tensor Core GPU, and the NVIDIA HPC SDK suite of��
]]>Today we announced the latest versions of Nsight Systems 2021.2 and Nsight Compute 2021.1 �C now available with new features for GPU profiling and performance optimization. We also announced the Nsight Visual Studio Code Edition, NVIDIA��s new addition to the series of world class developer tools for CUDA programming and debugging. Nsight Systems 2021.2 Nsight Systems 2021.2��
]]>Announcing NVIDIA Nsight Visual Studio Code Edition, an application development environment for heterogeneous platforms which brings CUDA development for GPUs into Microsoft Visual Studio Code. NVIDIA Nsight Visual Studio Code Edition allows you to build and debug GPU kernels and native CPU code, as well as inspect the state of the GPU and memory. It includes IntelliSense code��
]]>Today, NVIDIA is announcing the availability of the HPC SDK version 21.3. This software can be downloaded now free of charge. Download Now Pull Container from NGC What��s New See the HPC SDK Release Notes for more information. About the NVIDIA HPC SDK The NVIDIA HPC SDK is a comprehensive suite of compilers, libraries, and tools enabling developers to program the��
]]>DOCA is a software framework for developing applications on BlueField DPUs. By using DOCA, you can offload infrastructure workloads from the host CPU and accelerate them with the BlueField DPU. This enables an infrastructure that is software-defined yet hardware accelerated, maximizing both performance and flexibility in the data center. NVIDIA first introduced DOCA in October 2020.
]]>Today NVIDIA released the NVIDIA?DOCA SDK, a Data Center-on-a-Chip Architecture that provides developers with an easy way to program the BlueField DPU. Modern data centers are software-defined, agile, and built to serve highly distributed workloads from cloud to edge. But the traditional software defined data center approach imposes a significant performance tax that can easily consume 30%
]]>NVIDIA recently announced Morpheus, an AI application framework that provides cybersecurity developers with a highly optimized AI pipeline and pre-trained AI capabilities. Morpheus allows developers for the first time to instantaneously inspect all IP network communications through their data center fabric. Attacks are becoming more and more frequent and dangerous despite the advancements in��
]]>One of the main challenges and goals when creating an AI application is producing a robust model that is performant with high accuracy. Building such a deep learning model is time consuming. It can take weeks or months of retraining, fine-tuning, and optimizing until the model satisfies the necessary requirements. For many developers, building a deep learning AI pipeline from scratch is not a��
]]>Python plays a key role within the science, engineering, data analytics, and deep learning application ecosystem. NVIDIA has long been committed to helping the Python ecosystem leverage the accelerated massively parallel performance of GPUs to deliver standardized libraries, tools, and applications. Today, we��re introducing another step towards simplification of the developer experience with��
]]>In the field of medicine, advancements in artificial intelligence are constantly evolving. To keep up with the pace of innovation means adapting and providing the best experience to researchers, clinicians, and data scientists. NVIDIA Clara Train, an application framework for training medical imaging models, has undergone significant changes for its upcoming release at the beginning of May��
]]>The NVIDIA NGC catalog, a GPU-optimized hub for HPC, ML and AI applications, now has a new look and we��re really excited about it! Over the past few months we��ve been working with our community, design and research teams to bring you an enhanced user experience, engineered to deliver the most relevant content and features, faster than ever before. The new user interface��
]]>Reallusion released Character Creator 3 Connector for NVIDIA Omniverse, the open platform for 3D collaboration and simulation. The new Omniverse Connector adds a complete digital human creation pipeline to Omniverse Create and Omniverse Machinima apps. With Character Creator 3 and Omniverse, individuals or design teams can create and deploy digital characters as task performers��
]]>NVIDIA is raising the bar for XR streaming. Announced today, the NVIDIA CloudXR platform will be available on NVIDIA GPU-powered virtual machine instances on Azure. NVIDIA CloudXR is built on NVIDIA RTX GPUs to enable streaming of immersive AR, VR or mixed reality experiences from anywhere. By streaming from the cloud, enterprises can easily set up and scale immersive experiences from any��
]]>NVIDIA CUDA 11.3 NVIDIA has announced our newest release of the CUDA toolkit and development environment, consisting of GPU-accelerated libraries, debugging and optimization tools, a C/C++ compiler, and a runtime library to build and deploy your application on major architectures including NVIDIA Ampere, x86, Arm server processors, and POWER. CUDA 11.3 features are focused on enhancing��
]]>The upcoming NVIDIA CloudXR 2.1 release will deliver support for Apple iOS AR devices, including iPads and iPhones. Built on NVIDIA RTX technology, CloudXR is an advanced streaming technology that delivers VR and AR across 5G and Wi-Fi networks.XR users no longer need to be physically tethered to a high-performance computer to experience rich, immersive environments. The CloudXR SDK runs on��
]]>NVIDIA welcomes OpenCL 3.0��s focus on defining a baseline to enable developer-critical functionality to be widely adopted in future versions of the specification. With the recently released R465 display driver, NVIDIA is now officially OpenCL 3.0 conformant on both Windows and Linux. In September 2020, the Khronos Group released the OpenCL 3.0 final specification.
]]>NVIDIA is releasing the latest version of Variable Rate Supersampling (VRSS), which now includes gaze-tracked foveated rendering. VRSS 2 with dynamic foveated region at 8x shading rate. The blue foveated region has the highest image quality and location determined by eye gaze-tracking. Image of Boneworks courtesy of Stress Level Zero. Foveated rendering is a technique where a region of the��
]]>NVIDIA CloudXR is now publicly available through the AWS Marketplace as an Amazon Machine Image (AMI). Streaming from the cloud provides new opportunities for XR access and portability. NVIDIA CloudXR provides users with an advanced XR technology that brings deeply immersive graphics to any OpenVR application. With CloudXR on AWS, users can stream rich XR experiences wherever they are.
]]>Data engineering and data science workflows are often limited by the ability of platforms to process massively growing amounts of data. The integration of the Cloudera Data Platform (CDP), the RAPIDS Accelerator for Apache Spark 3.0, and NVIDIA computing, announced April 12, 2021, enables accelerated and scalable big data pre-processing, and workflows without code changes. With Cloudera CDP and��
]]>Natural Language Processing (NLP) has seen rapid progress in recent years as computation at scale has become more available and datasets have become larger. At the same time, recent work has shown large language models to be effective few-shot learners, with high accuracy on many NLP datasets without additional finetuning. As a result, state-of-the-art NLP models have grown at an exponential rate��
]]>Magnum IO is the collection of IO technologies from NVIDIA and Mellanox that make up the IO subsystem of the modern data center and enable applications at scale. If you are trying to scale up your application to multiple GPUs, or scaling it out across multiple nodes, you are probably using some of the libraries in Magnum IO. NVIDIA is now publishing the Magnum IO Developer Environment 21.04��
]]>Designing rich content and graphics for VR experiences means creating complex materials and high-resolution textures. But rendering all that content at VR resolutions and frame rates can be challenging, especially when rendering at the highest quality. You can address this challenge by using variable rate shading (VRS) to focus shader resources on certain parts of an image��specifically��
]]>