NVIDIA NIM Connector Guide > Introduction to NVIDIA NIM Connector > NVIDIA NIM overview
  

NVIDIA NIM overview

NVIDIA NIM offers a cloud-native deployment model, allowing users to run models using NVIDIA API endpoints or in their own environment using NGC containers. It supports multi-model and multi-instance serving, efficiently utilizing resources by hosting multiple models concurrently with batching, and scaling across single or multiple GPUs. The flexible hosting options include on-premises deployment with NGC containers, NVIDIA cloud-hosted endpoints, and localhost inference endpoints commonly used for development.
NVIDIA NIM covers a broad range of use cases, supporting natural language processing tasks such as text generation, reranking, embeddings, summarization, classification, and translation. It also supports vision tasks including image classification, segmentation, and object detection.