The LLMs listed below are available on Databricks Marketplace as Model Serving Endpoints.
Deployment Instructions
-
Click Get access on the model listing page in Databricks Marketplace.
-
Once access is granted, click the Open button to view the model and available versions.
-
Select the model version with the production alias (recommended for stable deployments) and click Serve this model.
-
In the Create serving endpoint configuration:
-
Compute type: Select the appropriate compute type based on your workload requirements. Refer to the memory requirements table to determine the approximate memory needed for your selected model.
-
Environment variables: Add the following required environment variable:
SPARK_NLP_LICENSE="your-license-key"
-
-
Complete the serving endpoint configuration and click Create.
-
Wait for the endpoint to become active.
Note: Each model listing includes a detailed notebook with comprehensive instructions for creating serving endpoints using the Databricks API, along with inference examples and supported formats.