
Merlin InferenceNVIDIA
Deploy NVTabular workflows and HugeCTR or TensorFlow models to Triton Inference server for production.
Vendor
NVIDIA
Company Website
Product details
The Merlin Inference container allows users to deploy NVTabular workflows and HugeCTR or TensorFlow models to Triton Inference server for production. This container is part of the NVIDIA Merlin framework, which accelerates the entire recommender systems pipeline on the GPU, from data ingestion and training to deployment. Merlin empowers data scientists, machine learning engineers, and researchers to build high-performing recommenders at scale. Each stage of the Merlin pipeline is optimized to support hundreds of terabytes of data, all accessible through easy-to-use APIs.
Features
- NVTabular: Performs data preprocessing and feature engineering for tabular data, scaling to manipulate terabyte-scale datasets.
- HugeCTR and TensorFlow: Used for training deep-learning based recommender system models.
- Triton Inference Server: Provides GPU-accelerated inference, simplifying the deployment of AI models at scale.
- Multi-Arch Support: Compatible with Linux/amd64 architecture.
- Security: Signed images and comprehensive security scanning.
Benefits
- High Performance: Accelerates the entire recommender systems pipeline on the GPU.
- Scalability: Supports large datasets, making it suitable for extensive data processing and model training.
- Ease of Deployment: Simplifies the deployment of trained models with Triton Inference Server.
- Versatility: Supports various AI frameworks and deployment environments.