Large Language Models (LLMs)
Deploying NVIDIA NIM Models
Description of your new file.
Adding nvcr.io
Docker Registry
-
Generate an API Key from https://org.ngc.nvidia.com/setup/api-keys
Make sure to give it access to
NGC Catalog
-
Add a Custom Docker Registry to the Platform
Registry URL:
nvcr.io
Username:$oauthtoken
Password: The API Key from the previous step
Adding NGC API Key to Secrets
-
Add the same API Key as a Secret on the Platform
Deploying a NIM Model
-
From
New Deployment
page, selectNVIDIA NIM
.- Select the workspace you want to deploy to
- Select the NVCR Model Registry Integration we created in the previous step
- Select the NGC API Key Secret we created in the previous step
- Select the model you want to deploy
-
Click Next and follow the rest of the steps to deploy the model.