TensorFlow – NVIDIA Technical Blog News and tutorials for developers, data scientists, and IT admins 2025-07-03T22:20:47Z http://www.open-lab.net/blog/feed/ Nirmal Kumar Juluru <![CDATA[Transforming Industrial Defect Detection with NVIDIA TAO and Vision AI Models]]> http://www.open-lab.net/blog/?p=73760 2023-12-07T16:59:55Z 2023-11-20T17:00:00Z Efficiency is paramount in industrial manufacturing, where even minor gains can have significant financial implications. According to the American Society of...]]> Efficiency is paramount in industrial manufacturing, where even minor gains can have significant financial implications. According to the American Society of...An image of bottles in a manufacturing processing plant.

Efficiency is paramount in industrial manufacturing, where even minor gains can have significant financial implications. According to the American Society of Quality, ��Many organizations will have true quality-related costs as high as 15-20% of sales revenue, some going as high as 40% of total operations.�� These staggering statistics reveal a stark reality: defects in industrial applications not��

Source

]]>
1
Kamil Tokarski <![CDATA[Why Automatic Augmentation Matters]]> http://www.open-lab.net/blog/?p=64036 2023-06-06T23:22:25Z 2023-05-05T20:32:52Z Deep learning models require hundreds of gigabytes of data to generalize well on unseen samples. Data augmentation helps by increasing the variability of...]]> Deep learning models require hundreds of gigabytes of data to generalize well on unseen samples. Data augmentation helps by increasing the variability of...

Deep learning models require hundreds of gigabytes of data to generalize well on unseen samples. Data augmentation helps by increasing the variability of examples in datasets. The traditional approach to data augmentation dates to statistical learning when the choice of augmentation relied on the domain knowledge, skill, and intuition of the engineers that set up the model training.

Source

]]>
0
Gwena Cunha Sergio <![CDATA[Accelerating Quantized Networks with the NVIDIA QAT Toolkit for TensorFlow and NVIDIA TensorRT]]> http://www.open-lab.net/blog/?p=48838 2023-04-04T17:00:05Z 2022-06-16T17:28:18Z We��re excited to announce the NVIDIA Quantization-Aware Training (QAT) Toolkit for TensorFlow 2 with the goal of accelerating the quantized networks with...]]> We��re excited to announce the NVIDIA Quantization-Aware Training (QAT) Toolkit for TensorFlow 2 with the goal of accelerating the quantized networks with...

Join the NVIDIA Triton and NVIDIA TensorRT community to stay current on the latest product updates, bug fixes, content, best practices, and more. We��re excited to announce the NVIDIA Quantization-Aware Training (QAT) Toolkit for TensorFlow 2 with the goal of accelerating the quantized networks with NVIDIA TensorRT on NVIDIA GPUs. This toolkit provides you with an easy-to-use API to quantize��

Source

]]>
0
Michelle Horton <![CDATA[Just Released: TensorRT 8.4]]> http://www.open-lab.net/blog/?p=49102 2022-09-09T16:15:06Z 2022-06-16T17:00:00Z ]]> ]]> 0 Tomasz Grel <![CDATA[Training a Recommender System on DGX A100 with 100B+ Parameters in TensorFlow 2]]> http://www.open-lab.net/blog/?p=46168 2023-02-13T18:51:50Z 2022-04-05T22:36:21Z Deep learning recommender systems often use large embedding tables. It can be difficult to fit them in GPU memory. This post shows you how to use a combination...]]> Deep learning recommender systems often use large embedding tables. It can be difficult to fit them in GPU memory. This post shows you how to use a combination...

Deep learning recommender systems often use large embedding tables. It can be difficult to fit them in GPU memory. This post shows you how to use a combination of model parallel and data parallel training paradigms to solve this memory issue to train large deep learning recommender systems more quickly. I share the steps that my team took to efficiently train a 113 billion-parameter��

Source

]]>
0
Michelle Horton <![CDATA[Deep Learning Study Could Spark New Dinosaur Discoveries]]> http://www.open-lab.net/blog/?p=44700 2024-08-12T17:56:19Z 2022-02-28T22:28:20Z Applying new technology to studying ancient history, researchers are looking to expand their understanding of dinosaurs with a new AI algorithm. The study,...]]> Applying new technology to studying ancient history, researchers are looking to expand their understanding of dinosaurs with a new AI algorithm. The study,...

Applying new technology to studying ancient history, researchers are looking to expand their understanding of dinosaurs with a new AI algorithm. The study, published in Frontiers in Earth Science, uses high-resolution Computed Tomography (CT) imaging combined with deep learning models to scan and evaluate dinosaur fossils. The research is a step toward creating a new tool that would vastly change��

Source

]]>
0
Michelle Horton <![CDATA[Autonomous AI Outraces Gran Turismo World Champs]]> http://www.open-lab.net/blog/?p=44331 2023-05-22T19:41:49Z 2022-02-17T17:42:06Z Gran Turismo (GT) Sport competitors are facing a new, AI-supercharged contender thanks to the latest collaborative effort from Sony AI, Sony Interactive...]]> Gran Turismo (GT) Sport competitors are facing a new, AI-supercharged contender thanks to the latest collaborative effort from Sony AI, Sony Interactive...

Gran Turismo (GT) Sport competitors are facing a new, AI-supercharged contender thanks to the latest collaborative effort from Sony AI, Sony Interactive Entertainment (SIE), and Polyphony Digital Inc., the developers behind GT Sport. The autonomous AI racing agent, known as Gran Turismo Sophy (GT Sophy), recently beat the world��s best drivers in GT Sport. Published in Nature��

Source

]]>
0
Chintan Patel <![CDATA[New on NGC: Security Reports, Latest Containers for PyTorch, TensorFlow, HPC and More]]> http://www.open-lab.net/blog/?p=43583 2023-02-13T18:55:40Z 2022-01-26T22:54:42Z The NVIDIA NGC catalog is a hub for GPU-optimized deep learning, machine learning, and HPC applications. With highly performant software containers, pretrained...]]> The NVIDIA NGC catalog is a hub for GPU-optimized deep learning, machine learning, and HPC applications. With highly performant software containers, pretrained...

The NVIDIA NGC catalog is a hub for GPU-optimized deep learning, machine learning, and HPC applications. With highly performant software containers, pretrained models, industry-specific SDKs, and Jupyter Notebooks the content helps simplify and accelerate end-to-end workflows. New features, software, and updates to help you streamline your workflow and build your solutions faster on NGC��

Source

]]>
0
Michelle Horton <![CDATA[Researchers Create a Camera the Size of a Salt Grain Using Neural Nano-Optics]]> http://www.open-lab.net/blog/?p=42350 2023-05-22T19:44:39Z 2021-12-09T23:47:54Z A team of researchers from Princeton and the University of Washington created a new camera that captures stunning images and measures in at only a...]]> A team of researchers from Princeton and the University of Washington created a new camera that captures stunning images and measures in at only a...A picture of a finger with a half milimeter metasurface next to a blown-up version of the metasurface showing the optical cylinders spread out in a circular pattern.

A team of researchers from Princeton and the University of Washington created a new camera that captures stunning images and measures in at only a half-millimeter��the size of a coarse grain of salt. The new study, published in Nature Communications, outlines the use of optical metasurfaces with machine learning to produce high-quality color imagery, with a wide field of view.

Source

]]>
0
Vinh Nguyen <![CDATA[Accelerating Embedding with the HugeCTR TensorFlow Embedding Plugin]]> http://www.open-lab.net/blog/?p=37559 2022-08-21T23:52:42Z 2021-09-24T19:00:00Z Recommender systems are the economic engine of the Internet. It is hard to imagine any other type of applications with more direct impact in our daily digital...]]> Recommender systems are the economic engine of the Internet. It is hard to imagine any other type of applications with more direct impact in our daily digital...

Recommender systems are the economic engine of the Internet. It is hard to imagine any other type of applications with more direct impact in our daily digital lives: Trillions of items to be recommended to billions of people. Recommender systems filter products and services among an overwhelming number of options, easing the paradox of choice that most users face. As the amount of data��

Source

]]>
0
Vivek Kini <![CDATA[Using the NVIDIA CUDA Stream-Ordered Memory Allocator, Part 2]]> http://www.open-lab.net/blog/?p=35152 2022-08-21T23:52:21Z 2021-07-27T20:47:33Z In part 1 of this series, we introduced new API functions, cudaMallocAsync and cudaFreeAsync, that enable memory allocation and deallocation to be...]]> In part 1 of this series, we introduced new API functions, cudaMallocAsync and cudaFreeAsync, that enable memory allocation and deallocation to be...

In part 1 of this series, we introduced new API functions, and , that enable memory allocation and deallocation to be stream-ordered operations. In this post, we highlight the benefits of this new capability by sharing some big data benchmark results and provide a code migration guide for modifying your existing applications. We also cover advanced topics to take advantage of stream-ordered��

Source

]]>
12
Vivek Kini <![CDATA[Using the NVIDIA CUDA Stream-Ordered Memory Allocator, Part 1]]> http://www.open-lab.net/blog/?p=35109 2022-08-21T23:52:19Z 2021-07-27T20:46:43Z Most CUDA developers are familiar with the cudaMalloc and cudaFree API functions to allocate GPU accessible memory. However, there has long been an obstacle...]]> Most CUDA developers are familiar with the cudaMalloc and cudaFree API functions to allocate GPU accessible memory. However, there has long been an obstacle...

Most CUDA developers are familiar with the and API functions to allocate GPU accessible memory. However, there has long been an obstacle with these API functions: they aren��t stream ordered. In this post, we introduce new API functions, and , that enable memory allocation and deallocation to be stream-ordered operations. In part 2 of this series, we highlight the benefits of this new��

Source

]]>
1
Houman Abbasian <![CDATA[Speeding Up Deep Learning Inference Using TensorFlow, ONNX, and NVIDIA TensorRT]]> http://www.open-lab.net/blog/?p=16755 2022-08-21T23:39:52Z 2021-07-20T13:00:00Z This post was updated July 20, 2021 to reflect NVIDIA TensorRT 8.0 updates. In this post, you learn how to deploy TensorFlow trained deep learning models using...]]> This post was updated July 20, 2021 to reflect NVIDIA TensorRT 8.0 updates. In this post, you learn how to deploy TensorFlow trained deep learning models using...

Source

]]>
19
Michelle Horton <![CDATA[Transforming Brain Waves into Words with AI]]> http://www.open-lab.net/blog/?p=34822 2024-08-12T17:57:50Z 2021-07-19T18:38:58Z New research out of the University of California, San Francisco has given a paralyzed man the ability to communicate by translating his brain signals into...]]> New research out of the University of California, San Francisco has given a paralyzed man the ability to communicate by translating his brain signals into...Diagram of neuroprosthesis device

New research out of the University of California, San Francisco has given a paralyzed man the ability to communicate by translating his brain signals into computer generated writing. The study, published in The New England Journal of Medicine, marks a significant milestone toward restoring communication for people who have lost the ability to speak. ��To our knowledge, this is the first��

Source

]]>
0
Piotr Bigaj <![CDATA[Accelerating the Wide & Deep Model Workflow from 25 Hours to 10 Minutes Using NVIDIA GPUs]]> http://www.open-lab.net/blog/?p=29663 2024-10-28T19:02:41Z 2021-04-29T22:15:38Z Recommender systems drive engagement on many of the most popular online platforms. As data volume grows exponentially, data scientists increasingly turn from...]]> Recommender systems drive engagement on many of the most popular online platforms. As data volume grows exponentially, data scientists increasingly turn from...

Recommender systems drive engagement on many of the most popular online platforms. As data volume grows exponentially, data scientists increasingly turn from traditional machine learning methods to highly expressive, deep learning models to improve recommendation quality. Often, the recommendations are framed as modeling the completion of a user-item matrix, in which the user-item entry is the��

Source

]]>
1
Vinh Nguyen <![CDATA[Announcing the NVIDIA NVTabular Open Beta with Multi-GPU Support and New Data Loaders]]> http://www.open-lab.net/blog/?p=21200 2024-10-28T18:24:20Z 2020-10-05T13:00:00Z Recently, NVIDIA CEO Jensen Huang announced updates to the open beta of NVIDIA Merlin, an end-to-end framework that democratizes the development of large-scale...]]> Recently, NVIDIA CEO Jensen Huang announced updates to the open beta of NVIDIA Merlin, an end-to-end framework that democratizes the development of large-scale...

Recently, NVIDIA CEO Jensen Huang announced updates to the open beta of NVIDIA Merlin, an end-to-end framework that democratizes the development of large-scale deep learning recommenders. With NVIDIA Merlin, data scientists, machine learning engineers, and researchers can accelerate their entire workflow pipeline from ingesting and training to deploying GPU-accelerated recommenders (Figure 1).

Source

]]>
0
Ethem Can <![CDATA[Profiling and Optimizing Deep Neural Networks with DLProf and PyProf]]> http://www.open-lab.net/blog/?p=21005 2024-08-28T17:55:38Z 2020-09-28T18:33:08Z Software profiling is key for achieving the best performance on a system and that��s true for the data science and machine learning applications as well. In...]]> Software profiling is key for achieving the best performance on a system and that��s true for the data science and machine learning applications as well. In...

Software profiling is key for achieving the best performance on a system and that��s true for the data science and machine learning applications as well. In the era of GPU-accelerated deep learning, when profiling deep neural networks, it is important to understand CPU, GPU, and even memory bottlenecks, which could cause slowdowns in training or inference. In this post��

Source

]]>
13
Christian Hundt <![CDATA[Streaming Interactive Deep Learning Applications at Peak Performance]]> http://www.open-lab.net/blog/?p=20528 2022-08-21T23:40:37Z 2020-09-01T17:25:39Z Imagine that you have just finished implementing an awesome, interactive, deep learning pipeline on your NVIDIA-accelerated data science workstation using...]]> Imagine that you have just finished implementing an awesome, interactive, deep learning pipeline on your NVIDIA-accelerated data science workstation using...

Imagine that you have just finished implementing an awesome, interactive, deep learning pipeline on your NVIDIA-accelerated data science workstation using OpenCV for capturing your webcam stream and rendering the output. A colleague of yours mentions that exploiting the novel TF32 compute mode of the Ampere microarchitecture third-generation Tensor Cores might significantly accelerate your��

Source

]]>
2
Akhil Docca <![CDATA[Accelerating AI and ML Workflows with Amazon SageMaker and NVIDIA NGC]]> http://www.open-lab.net/blog/?p=19448 2022-10-20T21:49:02Z 2020-08-07T19:33:05Z AI is going mainstream and is quickly becoming pervasive in every industry��from autonomous vehicles to drug discovery. However, developing and deploying AI...]]> AI is going mainstream and is quickly becoming pervasive in every industry��from autonomous vehicles to drug discovery. However, developing and deploying AI...

AI is going mainstream and is quickly becoming pervasive in every industry��from autonomous vehicles to drug discovery. However, developing and deploying AI applications is a challenging endeavor. The process requires building a scalable infrastructure by combining hardware, software, and intricate workflows, which can be time-consuming as well as error-prone. To accelerate the end-to-end AI��

Source

]]>
0
Vinh Nguyen <![CDATA[Accelerating TensorFlow on NVIDIA A100 GPUs]]> http://www.open-lab.net/blog/?p=18957 2023-06-12T21:15:05Z 2020-07-24T22:22:06Z The NVIDIA A100, based on the NVIDIA Ampere GPU architecture, offers a suite of exciting new features: third-generation Tensor Cores, Multi-Instance GPU (MIG)...]]> The NVIDIA A100, based on the NVIDIA Ampere GPU architecture, offers a suite of exciting new features: third-generation Tensor Cores, Multi-Instance GPU (MIG)...

The NVIDIA A100, based on the NVIDIA Ampere GPU architecture, offers a suite of exciting new features: third-generation Tensor Cores, Multi-Instance GPU (MIG) and third-generation NVLink. Ampere Tensor Cores introduce a novel math mode dedicated for AI training: the TensorFloat-32 (TF32). TF32 is designed to accelerate the processing of FP32 data types, commonly used in DL workloads.

Source

]]>
0
Raphael Boissel <![CDATA[Announcing CUDA on Windows Subsystem for Linux 2]]> http://www.open-lab.net/blog/?p=18337 2022-08-21T23:40:15Z 2020-06-17T17:00:00Z [stextbox id="info"]WSL2 is available on Windows 11 outside the Windows Insider Preview. For more information about what is supported, see the CUDA on WSL User...]]> [stextbox id="info"]WSL2 is available on Windows 11 outside the Windows Insider Preview. For more information about what is supported, see the CUDA on WSL User...

WSL2 is available on Windows 11 outside the Windows Insider Preview. For more information about what is supported, see the CUDA on WSL User Guide. In response to popular demand, Microsoft announced a new feature of the Windows Subsystem for Linux 2 (WSL 2)��GPU acceleration��at the Build conference in May 2020. This feature opens the gate for many compute applications, professional tools��

Source

]]>
6
Micha? Marcinkiewicz <![CDATA[Accelerating Medical Image Segmentation with NVIDIA Tensor Cores and TensorFlow 2]]> http://www.open-lab.net/blog/?p=17253 2024-11-04T22:55:18Z 2020-05-09T20:20:15Z Figure 1. Example of a serial section Transmission Electron Microscopy image (ssTEM) and its corresponding segmentation. Medical image segmentation is a hot...]]> Figure 1. Example of a serial section Transmission Electron Microscopy image (ssTEM) and its corresponding segmentation. Medical image segmentation is a hot...

Medical image segmentation is a hot topic in the deep learning community. Proof of that is the number of challenges, competitions, and research projects being conducted in this area, which only rises year over year. Among all the different approaches to this problem, U-Net has become the backbone of many of the top-performing solutions for both 2D and 3D segmentation tasks. This is due to its��

Source

]]>
0
Jin Li <![CDATA[Jump-start AI Training with NGC Pretrained Models On-Premises and in the Cloud]]> http://www.open-lab.net/blog/?p=16731 2022-08-21T23:39:52Z 2020-03-26T20:51:00Z Figure 1. NGC software stack. The process of building an AI-powered solution from start to finish can be daunting. First, datasets must be curated and...]]> Figure 1. NGC software stack. The process of building an AI-powered solution from start to finish can be daunting. First, datasets must be curated and...

The process of building an AI-powered solution from start to finish can be daunting. First, datasets must be curated and pre-processed. Next, models need to be trained and tested for inference performance, and then finally deployed into a usable, customer-facing application. At each step along the way, developers are constantly face time-consuming challenges, such as building efficient��

Source

]]>
0
Yaki Tebeka <![CDATA[TensorFlow Performance Logging Plugin nvtx-plugins-tf Goes Public]]> http://www.open-lab.net/blog/?p=15258 2024-08-28T17:58:49Z 2019-07-16T13:00:05Z The new nvtx-plugins-tf library enables users to add performance logging nodes to TensorFlow graphs. (TensorFlow is an open source library widely used for training DNN��deep neural network��models). These nodes log performance data using the NVTX (NVIDIA��s Tools Extension) library. The logged performance data can then be viewed in tools such as NVIDIA Nsight Systems and NVIDIA Nsight Compute.

Source

]]>
1
Yan Cheng <![CDATA[Annotate, Build, and Adapt Models for Medical Imaging with the Clara Train SDK]]> http://www.open-lab.net/blog/?p=15017 2022-08-21T23:39:31Z 2019-06-26T14:00:12Z Deep Learning?in medical imaging has shown great potential for disease detection, localization, and classification within radiology. Deep Learning holds the...]]> Deep Learning?in medical imaging has shown great potential for disease detection, localization, and classification within radiology. Deep Learning holds the...CUDA AI Cube

Deep Learning in medical imaging has shown great potential for disease detection, localization, and classification within radiology. Deep Learning holds the potential to create solutions that can detect conditions that might have been overlooked and can improve the efficiency and effectiveness of the radiology team. However, for this to happen data scientists and radiologists need to collaborate��

Source

]]>
0
Amulya Vishwanath <![CDATA[Automatic Mixed Precision for NVIDIA Tensor Core Architecture in TensorFlow]]> http://www.open-lab.net/blog/?p=14054 2022-08-21T23:39:23Z 2019-03-18T22:19:17Z Whether to employ mixed precision to train your TensorFlow models is no longer a tough decision.?NVIDIA��s?Automatic Mixed Precision (AMP) feature for...]]> Whether to employ mixed precision to train your TensorFlow models is no longer a tough decision.?NVIDIA��s?Automatic Mixed Precision (AMP) feature for...Jetson Xavier Tensor Core Matrix

Whether to employ mixed precision to train your TensorFlow models is no longer a tough decision. NVIDIA��s Automatic Mixed Precision (AMP) feature for TensorFlow, recently announced at the 2019 GTC, features automatic mixed precision training by making all the required model and optimizer adjustments internally within TensorFlow with minimal programmer intervention.

Source

]]>
5
Michael Carilli <![CDATA[New Optimizations To Accelerate Deep Learning Training on NVIDIA GPUs]]> http://www.open-lab.net/blog/?p=12964 2023-02-13T17:46:37Z 2018-12-03T16:00:36Z The pace of AI adoption across diverse industries depends on maximizing data scientists�� productivity. NVIDIA releases optimized NGC containers every month...]]> The pace of AI adoption across diverse industries depends on maximizing data scientists�� productivity. NVIDIA releases optimized NGC containers every month...

The pace of AI adoption across diverse industries depends on maximizing data scientists�� productivity. NVIDIA releases optimized NGC containers every month with improved performance for deep learning frameworks and libraries, helping scientists maximize their potential. NVIDIA continuously invests in the full data science stack, including GPU architecture, systems, and software stacks.

Source

]]>
0
Chip Huyen <![CDATA[Mixed Precision Training for NLP and Speech Recognition with OpenSeq2Seq]]> http://www.open-lab.net/blog/?p=12300 2022-08-21T23:39:09Z 2018-10-09T13:00:45Z The success of neural networks thus far has been built on bigger datasets, better theoretical models, and reduced training time. Sequential models, in...]]> The success of neural networks thus far has been built on bigger datasets, better theoretical models, and reduced training time. Sequential models, in...

Source

]]>
1
Shiva Pentyala <![CDATA[Mixed-Precision ResNet-50 Using Tensor Cores with TensorFlow]]> http://www.open-lab.net/blog/?p=11599 2022-08-21T23:39:01Z 2018-08-28T16:55:47Z Mixed-Precision combines different numerical precisions in a computational method. Using precision lower than FP32 reduces memory usage, allowing deployment of...]]> Mixed-Precision combines different numerical precisions in a computational method. Using precision lower than FP32 reduces memory usage, allowing deployment of...Figure 1: The Tesla V100 Accelerator with Volta GV100 GPU. SXM2 Form Factor.

Mixed-Precision combines different numerical precisions in a computational method. Using precision lower than FP32 reduces memory usage, allowing deployment of larger neural networks. Data transfers take less time, and compute performance increases, especially on NVIDIA GPUs with Tensor Core support for that precision. Mixed-precision training of DNNs achieves two main objectives: This video��

Source

]]>
2
Siddharth Sharma <![CDATA[TensorRT 4 Accelerates Neural Machine Translation, Recommenders, and Speech]]> http://www.open-lab.net/blog/?p=10726 2023-03-14T19:00:22Z 2018-06-19T13:00:45Z NVIDIA has released TensorRT?4 at CVPR 2018. This new version of TensorRT, NVIDIA��s powerful inference optimizer and runtime engine provides: New Recurrent...]]> NVIDIA has released TensorRT?4 at CVPR 2018. This new version of TensorRT, NVIDIA��s powerful inference optimizer and runtime engine provides: New Recurrent...

NVIDIA has released TensorRT 4 at CVPR 2018. This new version of TensorRT, NVIDIA��s powerful inference optimizer and runtime engine provides: Additional features include the ability to execute custom neural network layers using FP16 precision and support for the Xavier SoC through NVIDIA DRIVE AI platforms. TensorRT 4 speeds up deep learning inference applications such as neural machine��

Source

]]>
0
Sami Kama <![CDATA[TensorRT Integration Speeds Up TensorFlow Inference]]> http://www.open-lab.net/blog/?p=9984 2022-08-21T23:38:49Z 2018-03-27T17:33:00Z Update, May 9, 2018: TensorFlow v1.7 and above integrates with TensorRT 3.0.4. NVIDIA is working on supporting the integration for a wider set of configurations...]]> Update, May 9, 2018: TensorFlow v1.7 and above integrates with TensorRT 3.0.4. NVIDIA is working on supporting the integration for a wider set of configurations...

Update, May 9, 2018: TensorFlow v1.7 and above integrates with TensorRT 3.0.4. NVIDIA is working on supporting the integration for a wider set of configurations and versions. We��ll publish updates when these become available. Meanwhile, if you��re using , simply download TensorRT files for Ubuntu 14.04 not16.04, no matter what version of Ubuntu you��re running.

Source

]]>
40
Shashank Prasanna <![CDATA[TensorRT 3: Faster TensorFlow Inference and Volta Support]]> http://www.open-lab.net/blog/parallelforall/?p=8664 2022-08-21T23:38:34Z 2017-12-04T17:00:59Z NVIDIA TensorRT™ is a high-performance deep learning inference optimizer and runtime that delivers low latency, high-throughput inference for deep...]]> NVIDIA TensorRT™ is a high-performance deep learning inference optimizer and runtime that delivers low latency, high-throughput inference for deep...CUDA AI hero image

NVIDIA TensorRT is a high-performance deep learning inference optimizer and runtime that delivers low latency, high-throughput inference for deep learning applications. NVIDIA released TensorRT last year with the goal of accelerating deep learning inference for production deployment. In this post we��ll introduce TensorRT 3, which improves performance versus previous versions and includes new��

Source

]]>
16
Greg Heinrich <![CDATA[Photo Editing with Generative Adversarial Networks (Part 2)]]> http://www.open-lab.net/blog/parallelforall/?p=7793 2023-01-13T17:32:41Z 2017-04-24T23:10:12Z In part 1 of this series I introduced Generative Adversarial Networks (GANs) and showed how to generate images of handwritten digits using a GAN. In this post I...]]> In part 1 of this series I introduced Generative Adversarial Networks (GANs) and showed how to generate images of handwritten digits using a GAN. In this post I...

In part 1 of this series I introduced Generative Adversarial Networks (GANs) and showed how to generate images of handwritten digits using a GAN. In this post I will do something much more exciting: use Generative Adversarial Networks to generate images of celebrity faces. I am going to use CelebA [1], a dataset of 200,000 aligned and cropped 178 x 218-pixel RGB images of celebrities.

Source

]]>
2
Greg Heinrich <![CDATA[Photo Editing with Generative Adversarial Networks (Part 1)]]> http://www.open-lab.net/blog/parallelforall/?p=7749 2023-01-13T17:32:49Z 2017-04-20T16:00:23Z Adversarial training (also called GAN for Generative Adversarial Networks), and the variations that are now being proposed, is the most interesting idea in the...]]> Adversarial training (also called GAN for Generative Adversarial Networks), and the variations that are now being proposed, is the most interesting idea in the...

You heard it from the Deep Learning guru: Generative Adversarial Networks [2] are a very hot topic in Machine Learning. In this post I will explore various ways of using a GAN to create previously unseen images. I provide source code in Tensorflow and a modified version of DIGITS that you are free to use if you wish to try it out yourself. Figure 1 gives a preview of what you will learn to do in��

Source

]]>
12
���˳���97caoporen����