Mathstral, an advanced AI model developed from the ground up, can deliver superior performance for enhanced learning of math, engineering, and science.
]]>Llama 3.1 Nemotron 70B Reward model helps generate high-quality training data that aligns with human preferences for finance, retail, healthcare, scientific research, telecommunications, and sovereign AI.
]]>Dracarys, fine-tuned from Llama 3.1 70B and available from NVIDIA NIM microservice, supports a variety of applications, including data analysis, text summarization, and multi-language support.
]]>The new model by Mistral excels at a variety of complex tasks including text summarization, multilingual translation and reasoning, programming, question and answering, and conversational AI.
]]>Deepseek Coder v2, available as an NVIDIA NIM microservice, enhances project-level coding and infilling tasks.
]]>Phi-3-Medium accelerates research with logic-rich features in both short (4K) and long (128K) context.
]]>Trained on 600+ programming languages, StarCoder2-15B is now packaged as a NIM inference microservice available for free from the NVIDIA API catalog.
]]>Gemma 2, the next generation of Google Gemma models, is now optimized with TensorRT-LLM and packaged as NVIDIA NIM inference microservice.
]]>Experience and test Llama3-ChatQA models at scale with performance optimized NVIDIA NIM inference microservice using the NVIDIA API catalog.
]]>Experience Codestral, packaged as an NVIDIA NIM inference microservice for code completion, writing tests, and debugging in over 80 languages using the NVIDIA API catalog.
]]>Enhance efficiency and performance in instruction-based NLP tasks with SOLAR-10.7B, especially in following instructions, reasoning, and mathematical tasks.
]]>The model demonstrates strong performance for tasks such as Q&A, multi-round chat, and summarization in both traditional Chinese and English.
]]>Experience the versatile embedding model designed for multilingual, multi-functional, and multi-granularity text retrieval tasks, excelling in dense, multi-vector, and sparse retrieval for inputs from short sentences to long documents.
]]>Experience the advanced LLM API for code generation, completion, mathematical reasoning, and instruction following with free cloud credits.
]]>With free NVIDIA cloud credits, you can start testing the model at scale on the API Catalog.
]]>This week’s model release features DBRX, a state-of-the-art large language model (LLM) developed by Databricks. With demonstrated strength in programming and coding tasks, DBRX is adept at handling specialized topics and writing specific algorithms in languages like Python. It can also be used for text completion tasks and few-turn interactions. DBRX long-context abilities can be used in RAG…
]]>