Snowflake announced its collaboration with NVIDIA during the Snowflake Summit 2024. This partnership aims to empower customers and partners to develop bespoke AI data applications within Snowflake, leveraging NVIDIA’s AI technology.
This collaboration sees Snowflake integrating NVIDIA AI Enterprise software, incorporating NeMo Retriever microservices into Snowflake Cortex AI, Snowflake’s managed LLM and vector search service. This integration allows organisations to link custom models to varied business data, delivering precise responses seamlessly.
Additionally, Snowflake Arctic, an open, enterprise-grade LLM, now supports NVIDIA TensorRT-LLM software, enhancing performance. Arctic is also accessible as an NVIDIA NIM inference microservice, broadening developers’ access to its capabilities.
As enterprises strive to maximise AI’s potential, the need for data-driven customisation grows. The Snowflake-NVIDIA collaboration facilitates rapid development of specific AI solutions, benefiting businesses across various sectors.
“Pairing NVIDIA’s full stack accelerated computing and software with Snowflake’s state-of-the-art AI capabilities in Cortex AI is game-changing,” stated Sridhar Ramaswamy, CEO of Snowflake. “Together, we are unlocking a new era of AI where customers from every industry and every skill level can build custom AI applications on their enterprise data with ease, efficiency, and trust.”
“Data is the essential raw material of the AI industrial revolution,” said Jensen Huang, founder and CEO of NVIDIA. “Together, NVIDIA and Snowflake will help enterprises refine their proprietary business data and transform it into valuable generative AI.”
Notable NVIDIA AI Enterprise software capabilities offered in Cortex AI include:
- NVIDIA NeMo Retriever: Provides accurate and high-performance information retrieval for enterprises.
- NVIDIA Triton Inference Server: Facilitates the deployment, running, and scaling of AI inference for various applications on any platform.
- NVIDIA NIM inference microservices, part of NVIDIA AI Enterprise, can be deployed within Snowflake as a native app using Snowpark Container Services. This setup allows organisations to deploy foundational models directly within Snowflake easily.
Quantiphi, an AI-first digital engineering firm and ‘Elite’ partner with both Snowflake and NVIDIA, exemplifies this innovation. Quantiphi’s Snowflake Native Apps, baioniq and Dociphi, are designed to enhance productivity and document processing within specific industries. These apps, developed using the NVIDIA NeMo framework, will be available on Snowflake Marketplace.
The Snowflake Arctic LLM, launched in April 2024 and trained on NVIDIA H100 Tensor Core GPUs, is now available as an NVIDIA NIM. This makes Arctic accessible in seconds, either via the NVIDIA API catalogue with free credits or as a downloadable NIM, offering flexible deployment options.
Earlier this year, Snowflake and NVIDIA expanded their collaboration to create a unified AI infrastructure and compute platform in the AI Data Cloud. Today’s announcements mark significant advancements in their joint mission to help customers excel in their AI initiatives.