NVIDIA AI Foundry

Unveiling NVIDIA’s New AI Foundry Service: Empowering Enterprises with Custom Generative AI Models


NVIDIA has revolutionized AI technology with the introduction of the AI Foundry service and NVIDIA NIM inference microservices. These innovations help enterprises and nations create custom generative AI models tailored to their specific industry needs. Let’s dive into the key highlights of these groundbreaking services.

What is NVIDIA AI Foundry?

NVIDIA AI Foundry empowers enterprises to build custom “supermodels” using the openly available Llama 3.1 collection of models. With NVIDIA’s software, computing resources, and expertise, businesses can train these models with proprietary and synthetic data generated from Llama 3.1 405B and the NVIDIA Nemotron Reward model. Moreover, the service runs on the NVIDIA DGX Cloud AI platform, providing scalable compute resources for any enterprise’s needs.

AI Refinery framework

Key Benefits of the NVIDIA Foundry

  1. Custom Model Creation: Enterprises can develop unique models that meet their specific industry requirements.
  2. Scalable Compute Resources: The service leverages the NVIDIA DGX Cloud AI platform to ensure businesses have the necessary computing power.
  3. Synthetic Data Training: Utilize both proprietary and synthetic data for comprehensive model training.

Accenture Partnership: A First Adopter

Accenture has become the first to adopt NVIDIA AI Foundry, leveraging its AI Refinery framework to build custom Llama 3.1 models. This partnership allows Accenture to deploy generative AI applications that reflect the culture, languages, and industries of its clients. Consequently, this partnership showcases the potential of NVIDIA AI Foundry in real-world applications.

AI Refinery framework

Model Customization and Deployment

NVIDIA AI Foundry offers a curated collection of enterprise-grade pretrained models as a starting point. Enterprises can then output their custom models as NVIDIA NIM inference microservices, enabling efficient deployment on accelerated infrastructure. Solutions like NVIDIA TensorRT-LLM optimize performance, reducing latency and cost. Thus, enterprises can ensure their AI models run efficiently and effectively.

Supporting the AI Ecosystem

NVIDIA AI Enterprise experts and global system integrator partners support customers throughout the development and deployment process. Together, they will enable the AI ecosystem to deploy Llama 3.1 endpoints and other open models on DGX Cloud using the NVIDIA GPU-accelerated inference stack. Therefore, businesses can rely on a robust support system to maximize their AI potential.

Advantages of the NVIDIA Foundry

  • Tailored AI Solutions: Enterprises can create AI models that address their specific needs.
  • Efficient Deployment: Custom models can be deployed efficiently as NVIDIA NIM inference microservices.
  • Optimized Performance: Solutions like NVIDIA TensorRT-LLM ensure high performance and low costs.

Summary

NVIDIA AI Foundry aims to supercharge generative AI adoption by providing enterprises with the tools, computing power, and expertise needed to create and deploy custom AI models. With partners like Accenture, NVIDIA is paving the way for a new era of AI-driven innovation tailored to the unique needs of various industries.


Discover more from My Profit Sense

Subscribe now to keep reading and get access to free newsletter.

Continue reading