
Snowflake has announced a new collaboration with NVIDIA that customers and partners can harness to build customised AI data applications in Snowflake, powered by NVIDIA AI.
With this latest collaboration, Snowflake has adopted NVIDIA AI Enterprise software to integrate NeMo Retriever microservices into Snowflake Cortex AI, Snowflake’s fully managed large language model (LLM) and vector search service. This will enable organisations to connect custom models to diverse business data and deliver highly accurate responses.
Snowflake Arctic is now fully supported with NVIDIA TensorRT-LLM software
This will provide users with highly optimised performance. Arctic is also now available as an NVIDIA NIM inference microservice, allowing more developers to access Arctic’s efficient intelligence.
As enterprises look for ways to further unlock the power of AI across their teams, there’s an increasing need to apply data to drive customisation. Through Snowflake’s collaboration with NVIDIA, organisations can rapidly create bespoke, use-case specific AI solutions, enabling businesses across industries to realise the potential of enterprise AI.
Sridhar Ramaswamy, CEO, Snowflake, said: “Pairing NVIDIA’s full stack accelerated computing and software with Snowflake’s state-of-the-art AI capabilities in Cortex AI is game-changing. Together, we are unlocking a new era of AI where customers from every industry and every skill level can build custom AI applications on their enterprise data with ease, efficiency, and trust.”
Jensen Huang, founder and CEO, NVIDIA, added: “Data is the essential raw material of the AI industrial revolution. Together, NVIDIA and Snowflake will help enterprises refine their proprietary

US Tariffs are shifting - will you react or anticipate?
Don’t let policy changes catch you off guard. Stay proactive with real-time data and expert analysis.
By GlobalDataNVIDIA AI Enterprise software capabilities to be offered in Cortex AI include: NVIDIA NeMo Retriever which provides information retrieval with high accuracy and powerful performance for enterprises building retrieval-augmented generation-based AI applications within Cortex AI. Plus. NVIDIA Triton Inference Server, which provides the ability to deploy, run, and scale AI inference for any application on any platform.