Details
NIM microservices provide pre-optimized AI model containers including NVIDIA TensorRT, TensorRT-LLM, and other inference engines, along with industry-standard APIs, runtime dependencies, and enterprise-grade support. Developers can access NIM endpoints for over 40 AI models from NVIDIA and partners—including Meta Llama 3, Google Gemma, Microsoft Phi-3, and Mistral—through the NVIDIA Developer Program for free research and development, or in production via the NVIDIA AI Enterprise software platform. NIM reduces model deployment times from weeks to minutes and has been embedded into platforms from Amazon SageMaker, Microsoft Azure AI, and over 150 ecosystem partners.
Have evidence about NVIDIA's AI practices? Submit a report.
Report a Sighting →