Fine-Tuning – NVIDIA Technical Blog News and tutorials for developers, data scientists, and IT admins 2025-07-10T13:00:00Z http://www.open-lab.net/blog/feed/ Adi Renduchintala <![CDATA[Introducing the Nemotron-H Reasoning Model Family: Throughput Gains Without Compromise]]> http://www.open-lab.net/blog/?p=101373 2025-06-12T18:48:43Z 2025-06-06T17:00:00Z As large language models increasingly take on reasoning-intensive tasks in areas like math and science, their output lengths are getting significantly...]]> As large language models increasingly take on reasoning-intensive tasks in areas like math and science, their output lengths are getting significantly...

As large language models increasingly take on reasoning-intensive tasks in areas like math and science, their output lengths are getting significantly longer��sometimes spanning tens of thousands of tokens. This shift makes efficient throughput a critical bottleneck, especially when deploying models in real-world, latency-sensitive environments. To address these challenges and enable the��

Source

]]>
0
Cheng-Han (Hank) Du <![CDATA[Improving Translation Quality with Domain-Specific Fine-Tuning and NVIDIA NIM]]> http://www.open-lab.net/blog/?p=95756 2025-06-25T17:52:57Z 2025-02-05T21:30:00Z Translation plays an essential role in enabling companies to expand across borders, with requirements varying significantly in terms of tone, accuracy, and...]]> Translation plays an essential role in enabling companies to expand across borders, with requirements varying significantly in terms of tone, accuracy, and...

Translation plays an essential role in enabling companies to expand across borders, with requirements varying significantly in terms of tone, accuracy, and technical terminology handling. The emergence of sovereign AI has highlighted critical challenges in large language models (LLMs), particularly their struggle to capture nuanced cultural and linguistic contexts beyond English-dominant��

Source

]]>
1
Japinder Singh <![CDATA[Fine-Tuning Small Language Models to Optimize Code Review Accuracy]]> http://www.open-lab.net/blog/?p=94078 2025-02-17T05:13:45Z 2024-12-17T17:58:31Z Generative AI is transforming enterprises by driving innovation and boosting efficiency across numerous applications. However, adopting large foundational...]]> Generative AI is transforming enterprises by driving innovation and boosting efficiency across numerous applications. However, adopting large foundational...

Source

]]>
0
Bethann Noble <![CDATA[Deploying Fine-Tuned AI Models with NVIDIA NIM]]> http://www.open-lab.net/blog/?p=91696 2024-12-17T00:07:21Z 2024-11-21T22:04:57Z For organizations adapting AI foundation models with domain-specific data, the ability to rapidly create and deploy fine-tuned models is key to efficiently...]]> For organizations adapting AI foundation models with domain-specific data, the ability to rapidly create and deploy fine-tuned models is key to efficiently...

For organizations adapting AI foundation models with domain-specific data, the ability to rapidly create and deploy fine-tuned models is key to efficiently delivering value with enterprise generative AI applications. NVIDIA NIM offers prebuilt, performance-optimized inference microservices for the latest AI foundation models, including seamless deployment of models customized using parameter��

Source

]]>
0
Nirmal Kumar Juluru <![CDATA[Deploy GPU-Optimized AI Software with One Click Using Brev.dev and NVIDIA NGC Catalog]]> http://www.open-lab.net/blog/?p=84373 2024-07-25T18:19:21Z 2024-07-01T16:00:00Z Brev.dev is making it easier to develop AI solutions by leveraging software libraries, frameworks, and Jupyter Notebooks on the NVIDIA NGC catalog. You can use...]]> Brev.dev is making it easier to develop AI solutions by leveraging software libraries, frameworks, and Jupyter Notebooks on the NVIDIA NGC catalog. You can use...

Brev.dev is making it easier to develop AI solutions by leveraging software libraries, frameworks, and Jupyter Notebooks on the NVIDIA NGC catalog. You can use Brev.dev to easily deploy software on an NVIDIA GPU by pairing a cloud orchestration tool with a simple UI. Get an on-demand GPU reliably from any cloud, access the notebook in-browser, or SSH directly into the machine with the Brev��

Source

]]>
0
Ali Taghibakhshi <![CDATA[Enhance Text-to-Image Fine-Tuning with DRaFT+, Now Part of NVIDIA NeMo]]> http://www.open-lab.net/blog/?p=81405 2024-05-02T19:07:04Z 2024-04-26T14:00:00Z Text-to-image diffusion models have been established as a powerful method for high-fidelity image generation based on given text. Nevertheless, diffusion models...]]> Text-to-image diffusion models have been established as a powerful method for high-fidelity image generation based on given text. Nevertheless, diffusion models...

Text-to-image diffusion models have been established as a powerful method for high-fidelity image generation based on given text. Nevertheless, diffusion models do not always grant the desired alignment between the given input text and the generated image, especially for complicated idiosyncratic prompts that are not encountered in real life. Hence, there is growing interest in efficiently fine��

Source

]]>
0
Amit Bleiweiss <![CDATA[Tune and Deploy LoRA LLMs with NVIDIA TensorRT-LLM]]> http://www.open-lab.net/blog/?p=80481 2024-09-04T22:42:54Z 2024-04-02T17:00:00Z Large language models (LLMs) have revolutionized natural language processing (NLP) with their ability to learn from massive amounts of text and generate fluent...]]> Large language models (LLMs) have revolutionized natural language processing (NLP) with their ability to learn from massive amounts of text and generate fluent...

Large language models (LLMs) have revolutionized natural language processing (NLP) with their ability to learn from massive amounts of text and generate fluent and coherent texts for various tasks and domains. However, customizing LLMs is a challenging task, often requiring a full training process that is time-consuming and computationally expensive. Moreover, training LLMs requires a diverse and��

Source

]]>
2
Mickael Ide <![CDATA[Accelerating Vector Search: Fine-Tuning GPU Index Algorithms]]> http://www.open-lab.net/blog/?p=69885 2024-11-18T21:16:14Z 2023-09-11T16:00:00Z In this post, we dive deeper into each of the GPU-accelerated indexes mentioned in part 1 and give a brief explanation of how the algorithms work, along with a...]]> In this post, we dive deeper into each of the GPU-accelerated indexes mentioned in part 1 and give a brief explanation of how the algorithms work, along with a...

In this post, we dive deeper into each of the GPU-accelerated indexes mentioned in part 1 and give a brief explanation of how the algorithms work, along with a summary of important parameters to fine-tune their behavior. We then go through a simple end-to-end example to demonstrate cuVS�� Python APIs on a question-and-answer problem with a pretrained large language model and provide a��

Source

]]>
0
Jiwei Liu <![CDATA[Fast Fine-Tuning of AI Transformers Using RAPIDS Machine Learning]]> http://www.open-lab.net/blog/?p=46373 2024-04-24T23:13:02Z 2022-04-14T03:05:21Z In recent years, transformers have emerged as a powerful deep neural network architecture that has been proven to beat the state of the art in many application...]]> In recent years, transformers have emerged as a powerful deep neural network architecture that has been proven to beat the state of the art in many application...

In recent years, transformers have emerged as a powerful deep neural network architecture that has been proven to beat the state of the art in many application domains, such as natural language processing (NLP) and computer vision. This post uncovers how you can achieve maximum accuracy with the fastest training time possible when fine-tuning transformers. We demonstrate how the cuML support��

Source

]]>
0
���˳���97caoporen����