NVIDIA announced a new NVIDIA AI Foundry service and NVIDIA NIM inference microservices to boost generative AI for the world’s enterprises with the Llama 3.1 collection of openly available models also introduced.
NVIDIA AI Foundry is a service that allows enterprises to utilize data, accelerated computing, and software tools to develop and implement personalized models, which can significantly enhance their generative AI initiatives.
With NVIDIA AI Foundry, enterprises and nations can now create custom “supermodels” for their domain-specific industry use cases using Llama 3.1 and NVIDIA software, computing, and expertise. Enterprises can train these supermodels with their own data and synthetic data generated from Llama 3.1 405B and the NVIDIA Nemotron Reward model.
Many healthcare, energy, financial services, retail, transportation, and telecommunications companies are already using NVIDIA NIM microservices for Llama.
The new offerings come at a time when enterprises, as well as nations developing sovereign AI strategies, want to build custom large language models with domain-specific knowledge for generative AI applications that reflect their unique business or culture.
“Meta’s openly available Llama 3.1 models mark a pivotal moment for the adoption of generative AI within the world’s enterprises,” said Jensen Huang, founder and CEO of NVIDIA. “Llama 3.1 opens the floodgates for every enterprise and industry to build state-of-the-art generative AI applications. NVIDIA AI Foundry has integrated Llama 3.1 throughout and is ready to help enterprises build and deploy custom Llama supermodels.”
NVIDIA and Meta have also teamed to provide a distillation recipe for Llama 3.1 that developers can use to build smaller custom Llama 3.1 models for generative AI applications. This enables enterprises to run Llama-powered AI applications on a broader range of accelerated infrastructure, such as AI workstations and laptops.
“The new Llama 3.1 models are a super-important step for open-source AI,” said Mark Zuckerberg, founder and CEO of Meta. “With NVIDIA AI Foundry, companies can easily create and customize the state-of-the-art AI services people want and deploy them with NVIDIA NIM. I’m excited to get this in people’s hands.”
Enterprises can pair Llama 3.1 NIM microservices with new NVIDIA NeMo Retriever NIM microservices to create state-of-the-art retrieval pipelines for AI copilots, assistants, and digital human avatars.
Accenture is the first to adopt NVIDIA AI Foundry to build custom Llama 3.1 models, using the Accenture AI Refinery framework for its own use and for clients seeking to deploy generative AI applications.
“Accenture has been working with NVIDIA NIM inference microservices for our internal AI applications, and now, using NVIDIA AI Foundry, we can help clients quickly create and deploy custom Llama 3.1 models to power transformative AI applications for their own business priorities,” said Julie Sweet, chair and CEO of Accenture.
Hundreds of NVIDIA NIM partners providing enterprise, data, and infrastructure platforms can now integrate the new microservices in their AI solutions to boost generative AI for the NVIDIA community of more than 5 million developers and 19,000 startups.
Also Read: MediaTek Integrates NVIDIA TAO to Boost Edge AI Applications in IoT