Shashank Verma

Shashank Verma is a product research engineering manager at NVIDIA, where he leads the development and presentation of developer-focused content and proof-of-concept applications using the latest AI frameworks and platforms. He holds a master’s in Electrical Engineering from the University of Wisconsin-Madison, specializing in computer vision, security aspects in data science, and high-performance computing. Shashank is passionate about making advanced AI accessible by translating complex concepts into practical solutions for the developer community.
bio image for Shashank Verma NVIDIA

Posts by Shashank Verma

Generative AI

Run Hugging Face Models Instantly with Day-0 Support from NVIDIA NeMo Framework

As organizations strive to maximize the value of their generative AI investments, accessing the latest model developments is crucial to continued success. By... 6 MIN READ
Generative AI

Enhance Your AI Agent with Data Flywheels Using NVIDIA NeMo Microservices

Enterprise data is constantly changing. This presents significant challenges for maintaining AI system accuracy over time. As organizations increasingly rely on... 12 MIN READ
A larger and smaller cartoon llama on a sunny beach, wearing shirts that say 8B and 4B.
Generative AI

LLM Model Pruning and Knowledge Distillation with NVIDIA NeMo Framework

Model pruning and knowledge distillation are powerful cost-effective strategies for obtaining smaller language models from an initial larger sibling. ... 10 MIN READ
Generative AI

Leverage the Latest Open Models for Synthetic Data Generation with NVIDIA Nemotron-4-340B

This post was updated on August 16, 2024 to reflect the most recent Reward Bench results. Since the introduction and subsequent wide adoption of large language... 8 MIN READ
Generative AI

Customizing NVIDIA NIM for Domain-Specific Needs with NVIDIA NeMo

Large language models (LLMs) adopted for specific enterprise applications most often benefit from model customization. Enterprises need to tailor ?LLMs for... 11 MIN READ
Generative AI

Seamlessly Deploying a Swarm of LoRA Adapters with NVIDIA NIM

The latest state-of-the-art foundation large language models (LLMs) have billions of parameters and are pretrained on trillions of tokens of input text. They... 11 MIN READ