NVIDIA recently reached the 2 million registered developers milestone. To help commemorate the milestone, artist Pindar Van Arman this week at GTC painted a unique portrait of the NVIDIA Developer community. Using portraits submitted at the recent GTC AI Art Gallery, Pindar enlisted the help of an AI-painting robot to complete the job. The robot was trained for the final painting by��
]]>Nvidia is now offering precompiled drivers that are tested and packaged for targeted kernel versions. Precompiled drivers enable faster, seamless installation and driver updates, faster boot-up, and have reduced dependencies on external tools and repositories. Developers can now take advantage of NVIDIA precompiled drivers quite easily, using our kernel module (kmod) packages and modularity��
]]>NVIDIA GPUs have become mainstream for accelerating a variety of workloads from machine learning, high-performance computing (HPC), content creation workflows, and data center applications. For these enterprise use cases, NVIDIA provides a software stack powered by the CUDA platform: drivers, CUDA-X acceleration libraries, CUDA-optimized applications, and frameworks.
]]>This week at GTC 2020, synthetic biology startup Synvivia showcased protein switches being developed to control engineered organisms and aid in drug discovery for COVID-19. The full session is available in the GTC catalog to view on-demand. Synvivia uses GPU-accelerated molecular dynamics simulations to design protein molecule interactions and was able to observe potential additive effects of��
]]>We have selected the winners for our latest DXR Spotlight Contest. Thanks to all of the developers who participated. We received a lot of terrific entries, and it was very hard to narrow the list down to three. When combined with Microsoft DirectX? Raytracing and NVIDIA RTX GPUs, Unreal Engine 4��s ray tracing tools allow individual artists to deliver AAA results. Each of our finalists are��
]]>NVIDIA OptiX is an API for optimal ray tracing performance on the GPU that is used for film and video production as well as many other professional graphics applications. OptiX SDK 7.2 is the latest update to the new OptiX 7 API. This version introduces API changes to the OptiX denoiser to support layered AOV denoising, a new library for demand loading sparse textures��
]]>The nexus of 5G, IoT, and edge computing is turbocharging network performance. NVIDIA is working with the world��s leading telecommunications companies to build software-defined infrastructure that can meet the demand for real-time data processing at the edge for a variety of smart services. Today, we announced the first NVIDIA Aerial Developer Kit. Designed to jump start the execution of 5G��
]]>Machine learning workflows require iteration and experimentation. Improvements in performance, particularly speed, provides the opportunity to run additional experiments within a day, week, or month. The ability to increase iteration and experimentation potentially leads to improved models and impactful insights. Training benchmarks suites from industry consortiums, including MLPerf Training��
]]>Large language models such as Megatron and GPT-3 are transforming AI. We are excited about applications that can take advantage of these models to create better conversational AI. One main problem that generative language models have in conversational AI applications is their lack of controllability and consistency with real-world facts. In this work, we try to address this by making our large��
]]>For the past decade and a half, an increasing number of businesses have moved their traditional IT applications from on-premises to public clouds. This first phase of the revolution, which you could call ��enterprise cloud transformation,�� has allowed enterprises to reap the benefits of the scale, expertise, and flexibility of the cloud. The second phase of this revolution, the ��edge AI cloud��
]]>At NVIDIA, we are driving change in data science, machine learning, and artificial intelligence. Some of the key trends that drive us are as follows: At the intersection of these trends is Dask, an open-source library designed to provide parallelism to the existing Python stack. In this post, we talk about Dask, what it is, how we use it at NVIDIA, and why it has so much potential��
]]>At GTC Spring 2020, Adobe, Verizon Media, and Uber each discussed how they used Spark 3.0 with GPUs to accelerate and scale ML big data pre-processing, training, and tuning pipelines. There are multiple challenges when it comes to the performance of large-scale machine learning (ML) solutions: huge datasets, complex data preprocessing and feature engineering pipelines��
]]>NVIDIA CEO Jensen Huang Monday kicked off this week��s GPU Technology Conference. Huang made major announcements in data centers, edge AI, collaboration tools and healthcare in a talk simultaneously released in nine episodes, each under 10 minutes. The announcements touch on everything from healthcare to robotics to videoconferencing, Huang��s underlying story was simple: AI is changing��
]]>At GTC 2020, NVIDIA announced updates to 80 SDKs, including tools to help you build AI-powered video streaming solutions, conversational AI, recommendation systems, and more. Today, we announced NVIDIA Maxine, a cloud-native video streaming AI platform for services such as video conferencing. It includes state-of-the-art AI models and optimized pipelines that can run several��
]]>We first told you about their work in 2019 when they recreated a lost Picasso painting. Now as part of the GTC AI Art Gallery, artists George Cann and Anthony Bourached, and their company Oxia Palus, are debuting a recreated a masterpiece, a lost Leonardo da Vinci painting. The painting seen below, and shown publicly for the first time, is called Madonna.
]]>Today, NVIDIA announced the Jetson Nano 2GB Developer Kit, the ideal hands-on platform for teaching, learning, and developing AI and robotics applications. The NVIDIA Jetson platform introduced six years ago revolutionized embedded computing by delivering the power of artificial intelligence to edge computing devices. NVIDIA Jetson today is widely used in diverse fields such as robotics, retail��
]]>With the advent of new deep learning approaches based on transformer architecture, natural language processing (NLP) techniques have undergone a revolution in performance and capabilities. Cutting-edge NLP models are becoming the core of modern search engines, voice assistants, chatbots, and more. Modern NLP models can synthesize human-like text and answer questions posed in natural language.
]]>Cloud technologies are increasingly taking over the worldwide IT infrastructure market. With offerings that include elastic compute, storage, and networking, cloud service providers (CSPs) allow customers to rapidly scale their IT infrastructure up and down without having to build and manage it on their own. The increasing demand for differentiated and cost-effective cloud products and services is��
]]>Until today, artists had performance constraints that artificially limited lighting complexity; real-time renderers simply could not support more than a handful of dynamic lights. For years, NVIDIA sought methods to remove this barrier and enable real-time rendering of arbitrarily complex lighting. With NVIDIA��s Marbles at Night demo shown at the launch of the Ampere GeForce GPUs��
]]>Today, in his NVIDIA GTC Fall keynote, CEO Jensen Huang introduced a new kind of processor, the BlueField-2 data processing unit (DPU), a powerful new software development kit for the DPU, DOCA, along with a three year roadmap of DPU and AI innovation. The NVIDIA BlueField-2 DPU is the world��s first data center infrastructure on a chip architecture optimized for modern enterprise data centers.
]]>The world of live streaming has expanded rapidly this year: Twitch has seen an 89% increase in the number of streamers, while viewership is up 56%. Meanwhile more employees are working from home and collaborating by video conference. Background noise in the home and messy rooms can be an inconvenience for those hosting video meetings, and distracting to viewers tuning into a livestream.
]]>This is an updated version of Neural Modules for Fast Development of Speech and Language Models. This post upgrades the NeMo diagram with PyTorch and PyTorch Lightning support and updates the tutorial with the new code base. As a researcher building state-of-the-art speech and language models, you must be able to quickly experiment with novel network architectures.
]]>We previously announced Clara Guardian, an application framework and partner ecosystem that accelerates the development and deployment of smart sensors with multimodal AI anywhere in the hospital. Today four pre-trained models and NVIDIA Fleet Command are officially available for early access. Clara Guardian��s key components include healthcare pre-trained models for computer vision and��
]]>GPU Technology Conference (GTC) has just begun! The event started this morning and will run continuously for five days across seven time zones. Explore the latest announcements and releases for professional visualization and game development, including some of the newest tools and techniques in the graphics industry. Learn from leading experts on how they��re integrating GPU power to accelerate��
]]>This is the first post in the Accelerating IO series, which describes the architecture, components, storage, and benefits of Magnum IO, the IO subsystem of the modern data center. Previously the boundary of the unit of computing, sheet metal no longer constrains the resources that can be applied to a single problem or the data set that can be housed. The new unit is the data center.
]]>As the computing horsepower of GPUs increases, so does the demand for input/output (I/O) throughput and the need to strong scale with low-latency, high-bandwidth communication among GPUs. A GPU-accelerated supercomputer transforms a compute-bound problem into an I/O-bound problem. In multi-node GPU clusters, slow CPU single-thread performance is in the critical path of data access from local��
]]>NVIDIA Merlin is an open beta application framework and ecosystem that enables the end-to-end development of recommender systems, from data preprocessing to model training and inference, all accelerated on NVIDIA GPU. We announced Merlin in a previous post and have been continuously making updates to the open beta. In this post, we detail the new features added to the open beta NVIDIA Merlin��
]]>Recently, NVIDIA CEO Jensen Huang announced updates to the open beta of NVIDIA Merlin, an end-to-end framework that democratizes the development of large-scale deep learning recommenders. With NVIDIA Merlin, data scientists, machine learning engineers, and researchers can accelerate their entire workflow pipeline from ingesting and training to deploying GPU-accelerated recommenders (Figure 1).
]]>Hospitals today are seeking to overhaul their existing digital infrastructure to improve their internal processes, deliver better patient care, and reduce operational expenses. Such a transition is required if hospitals are to cope with the needs of a burgeoning human population, accumulation of medical patient data, and a pandemic. The goal is not only to digitize existing infrastructure but��
]]>AI, machine learning (ML), and deep learning (DL) are effective tools for solving diverse computing problems such as product recommendations, customer interactions, financial risk assessment, manufacturing defect detection, and more. Using an AI model in production, called inference serving, is the most complex part of incorporating AI in applications. Triton Inference Server takes care of all the��
]]>NVIDIA announced the Jetson Nano 2GB Developer Kit, the ideal hands-on platform for teaching, learning, and developing AI and robotics applications. The NVIDIA Jetson platform introduced six years ago revolutionized embedded computing by delivering the power of artificial intelligence to edge computing devices. NVIDIA Jetson today is widely used in diverse fields such as robotics, retail��
]]>At GTC, NVIDIA researchers introduced a robotics framework that combines model-based control and reinforcement learning to adaptively change contact sequences in real time. The system has the potential to help delivery robots and other autonomous machines function more effectively in environments and terrains the robot is not familiar with. The controller adapts to environmental changes��
]]>Building an AI application can be complex. It��s never a case of simply running and grabbing the latest framework container or simply downloading the model and code needed to build your AI app. It takes all of these components and more to build a solution for your use case, which is why we built the NGC catalog. The NGC catalog provides you with easy access to secure and optimized containers��
]]>The nexus of 5G, IoT, and MEC is expanding the horizons of digital transformation. NVIDIA Metropolis intelligent video analytics (IVA) platform makes sense of the flood of data created by trillions of cameras and sensors for traffic engineering in smart cities, frictionless retail, optical inspection on factory floors, and more. Mavenir��s 5G RAN & packet core platform transforms CSP��s��
]]>NVIDIA Omniverse, a computer graphics and simulation platform that enables artists to collaborate seamlessly in real time, is now available for early access customers in the architecture, engineering and construction (AEC) market. Using Pixar��s Universal Scene Description and NVIDIA RTX technology, Omniverse allows multiple people to easily work with popular content creation applications and��
]]>If you��re supporting the recent influx in remote work, you��ve probably noticed that business applications are more graphics-heavy than ever before. Applications such as Microsoft Office, Google Chrome, and PDF readers now offer graphics-rich features that require more power. In addition, 4K and multiple high-resolution monitors, as well as multimedia streaming, are becoming the new normal in the��
]]>Recommender systems drive every action that you take online, from the selection of this web page that you��re reading now to more obvious examples like online shopping. They play a critical role in driving user engagement on online platforms, selecting a few relevant goods or services from the exponentially growing number of available options. On some of the largest commercial platforms��
]]>NVIDIA founder and CEO Jensen Huang kicks off the company��s first ��kitchen keynote�� for NVIDIA GTC 2020. In part one, see NVIDIA��s work in response to the COVID-19 pandemic. Watch a new ��I AM AI�� video that vividly tells the story of AI through the work of our partners. Check out the future of data center-scale accelerated computing, enabled by NVIDIA and Mellanox technologies.
]]>NVIDIA today introduced the first GPU based on the NVIDIA Ampere architecture, the NVIDIA A100, is in full production and shipping to customers worldwide. The A100 draws on design breakthroughs in the NVIDIA Ampere architecture �� offering the company��s largest leap in performance to date within its eight generations of GPUs �� to unify AI training and inference and boost performance by up to��
]]>Organizations of all kinds are incorporating AI into their research, development, product, and business processes. This helps them meet and exceed their particular goals, and also helps them gain experience and knowledge to take on even bigger challenges. However, traditional compute infrastructures aren��t suitable for AI due to slow CPU architectures and varying system requirements for different��
]]>Recent work has demonstrated that larger language models dramatically advance the state of the art in natural language processing (NLP) applications such as question-answering, dialog systems, summarization, and article completion. However, during training, large models do not fit in the available memory of a single accelerator, requiring model parallelism to split the parameters across multiple��
]]>The NVIDIA mission is to accelerate the work of the da Vincis and Einsteins of our time. Scientists, researchers, and engineers are focused on solving some of the world��s most important scientific, industrial, and big data challenges using artificial intelligence (AI) and high performance computing (HPC). The NVIDIA HGX A100 with A100 Tensor Core GPUs delivers the next giant leap in our��
]]>Today, NVIDIA announced the NVIDIA Jetson Xavier NX Developer Kit , which is based on the Jetson Xavier NX module. Delivering up to 21 TOPS of compute in a compact form factor with under 15W of power, Jetson Xavier NX brings server-level performance and cloud-native workflows to edge AI devices and autonomous machines. With the Jetson Xavier NX Developer Kit, you can create amazing AI��
]]>BMW Group has selected the new NVIDIA Isaac robotics platform to enhance its automotive factories �� utilizing logistics robots built on advanced AI computing and visualization technologies, the companies announced today. The collaboration centers on implementing an end-to-end system based on NVIDIA technologies �� from training and testing through to deployment �� with robots developed using one��
]]>The NVIDIA Ampere GPU architecture has arrived! It��s time to make sure that your applications are getting the most out of the powerful compute resources in this new architecture. With the release of CUDA 11, we are adding several features to the Nsight family of Developer Tools to help you do just that. These additions improve usability, productivity, and make it easier for you to find bugs��
]]>NVIDIA today announced availability of the NVIDIA Jetson Xavier NX developer kit with cloud-native support �� and the extension of this support to the entire NVIDIA Jetson edge computing lineup for autonomous machines. The Jetson Xavier NX module is the ideal platform to accelerate AI applications, delivering greater than 10x higher performance compared to its widely adopted predecessor��
]]>Transfer learning is an important machine learning technique that uses a model��s knowledge of one task to make it perform better on another. Fine-tuning is one of the techniques to perform transfer learning. It is an essential part of the recipe for many state-of-the-art results where a base model is first pretrained on a task with abundant training data and then fine-tuned on different tasks of��
]]>NVIDIA today announced that it is collaborating with the open-source community to bring end-to-end GPU acceleration to Apache Spark 3.0, an analytics engine for big data processing used by more than 500,000 data scientists worldwide. With the anticipated late spring release of Spark 3.0, data scientists and machine learning engineers will for the first time be able to apply revolutionary GPU��
]]>Fifth-generation networks (5G) are ushering in a new era in wireless communications that delivers 1000X the bandwidth and 100X the speed at 1/10th the latency of 4G. 5G also allows for millions of connected devices per square km and is being deployed as an alternative to WiFi at edge locations like factories and retail stores. These applications demand a new network architecture that is fully��
]]>The world of healthcare is under a giant spotlight these days. In these unprecedented times, everyone is focusing on keeping loved ones safe and contributing to the community as much as possible. It is amazing to see how the community is uniting in the past two months, which has led to a global, rapid emergence of SARS-CoV-2 projects across the life sciences. At NVIDIA��
]]>CUDA is the most powerful software development platform for building GPU-accelerated applications, providing all the components needed to develop applications targeting every GPU platform. CUDA 11 introduces support for the new NVIDIA A100 based on the NVIDIA Ampere architecture, Arm server processors, performance-optimized libraries, and new developer tools and improvements for A100.
]]>To help unleash the performance advantages of the NVIDIA Ampere Architecture, the CUDA Toolkit 11 and Nsight Systems 2020.3 and Nsight Compute 2020.1 developer tools have been enhanced and scheduled for general availability at the end of May. The Nsight suite of developer tools provides insightful tracing, debugging, profiling, and other analyses to optimize your high-performance��
]]>The NVIDIA HPC SDK is a comprehensive suite of compilers and libraries enabling HPC developers to program the entire HPC platform from the GPU foundation to the CPU and through the interconnect. It is the only comprehensive, integrated SDK for programming accelerated computing systems. The NVIDIA HPC SDK C++ and Fortran compilers are the first and only compilers to support automatic GPU��
]]>NVIDIA announced Clara Guardian, an application framework and partner ecosystem that accelerates the development and deployment of smart sensors with multimodal AI anywhere in the hospital. In healthcare facilities, smart sensors such as AI-enabled cameras and microphones can help improve patient care and public safety while enhancing operational efficiency. NVIDIA Clara Guardian delivers AI��
]]>NVIDIA announced CloudXR 1.0 software development kit, which enables streaming augmented reality, mixed reality and virtual reality content over 5G, Wi-Fi and other high-performance networks. The NVIDIA CloudXR platform enhances immersive experiences by turning end devices such as head-mounted displays, smartphones, or tablets into a high-fidelity XR display, so users can power design reviews��
]]>At GTC 2020, NVIDIA announced and shipped a range of new AI SDKs, enabling developers to support the new Ampere architecture. For the first time, developers have the tools to build end-to-end deep learning-based pipelines for conversational AI and recommendation systems. Today, NVIDIA announced Riva, a fully accelerated application framework building multimodal conversational AI services.
]]>NVIDIA today announced new AI models to help the medical community better track, test and treat COVID-19. Available today, AI models developed jointly with the National Institutes of Health (NIH) can help researchers study the severity of COVID-19 from chest CT scans and develop new tools to better understand, measure and detect infections. The models are immediately available in the��
]]>NVIDIA announced a new technology embedded in its NVIDIA Mellanox ConnectX-6 Dx SmartNIC and BlueField-2 I/O Processing Unit to optimize 5G networks. Referred to as 5T-for-5G, or time-triggered transmission technology for telco, this new technology delivers superbly accurate time synchronization across front-haul and mid-haul networks, providing telecommunications providers with higher��
]]>Today, during the 2020 NVIDIA GTC keynote address, NVIDIA founder and CEO Jensen Huang introduced the new NVIDIA A100 GPU based on the new NVIDIA Ampere GPU architecture. This post gives you a look inside the new A100 GPU, and describes important new features of NVIDIA Ampere architecture GPUs. The diversity of compute-intensive applications running in modern cloud data centers has driven��
]]>Many of today��s speech synthesis models lack emotion and human-like expression. To help tackle this problem, a team of researchers from the NVIDIA Applied Deep Learning Research group developed a state-of-the-art model that generates more realistic expressions and provides better user control than previously published models. Named ��Flowtron��, the model debuted publicly for the first time as��
]]>The new NVIDIA A100 GPU based on the NVIDIA Ampere GPU architecture delivers the greatest generational leap in accelerated computing. The A100 GPU has revolutionary hardware capabilities and we��re excited to announce CUDA 11 in conjunction with A100. CUDA 11 enables you to leverage the new hardware capabilities to accelerate HPC, genomics, 5G, rendering, deep learning, data analytics��
]]>Today NVIDIA is releasing the ACM SIGGRAPH 2020 research paper showing how to render dynamic direct lighting and shadows from millions of area lights in real-time. This was previously impossible. Traditionally, games achieved complex lighting using baked solutions augmented by a few dynamic emitters. Recent RTX games have upgraded to dynamic area lights with physically correct shadows��
]]>