Subscribe to our daily and weekly newsletters to stay updated on the latest developments and exclusive content in the world of AI. Learn More
Nvidia and DataStax have unveiled groundbreaking technology aimed at significantly reducing storage requirements for companies implementing generative AI systems. This innovation also enables faster and more accurate information retrieval across multiple languages.
The newly introduced Nvidia NeMo Retriever microservices, in conjunction with DataStax’s AI platform, can reduce data storage volume by an impressive 35 times compared to traditional methods. This development is crucial as enterprise data is projected to exceed 20 zettabytes by 2027.
Kari Briski, VP of product management for AI at Nvidia, emphasized the significance of this breakthrough, stating, “Today’s enterprise unstructured data amounts to 11 zettabytes, with 83% being unstructured and 50% consisting of audio and video. By significantly reducing storage costs and enhancing information embedding and retrieval capabilities, companies can experience a game-changing transformation.”
This technology has already had a profound impact on Wikimedia Foundation, which utilized the integrated solution to reduce processing time for 10 million Wikipedia entries from 30 days to less than three days. The system efficiently manages real-time updates for hundreds of thousands of entries edited daily by 24,000 global volunteers.
Chet Kapoor, CEO of DataStax, highlighted the necessity of context from existing enterprise data in addition to large language models. He explained, “Our hybrid search capability combines semantic search and traditional text search, utilizing Nvidia’s re-ranker technology to deliver the most relevant results in real-time on a global scale.”
Enterprise data security meets AI accessibility
The collaboration addresses a crucial challenge for enterprises – how to make their extensive private data accessible to AI systems without compromising sensitive information to external language models.
Kapoor emphasized the importance of safeguarding data, citing the example of FedEx, where a significant portion of data is stored securely within their products, including detailed package delivery information spanning two decades. This data is not shared with external entities like Gemini or OpenAI.
Despite regulatory constraints, financial services firms are at the forefront of adopting this technology. Kapoor praised the advancements made by firms like Commonwealth Bank of Australia and Capital One.
The next frontier for AI: Multimodal document processing
Nvidia’s future plans include enhancing the technology to handle more intricate document formats, particularly in the realm of multimodal PDF processing. Briski expressed enthusiasm for addressing the complexities of understanding tables, graphs, charts, and images within documents and how they interrelate across pages.
This innovative solution offers enterprises drowning in unstructured data a pathway to prepare their information assets for AI implementation without compromising security or facing exorbitant storage costs. Interested parties can access the solution immediately through the Nvidia API catalog with a complimentary 90-day trial license.
This announcement underscores the increasing focus on enterprise AI infrastructure as companies transition from experimental phases to large-scale deployments. Effective data management and cost efficiency are becoming pivotal factors for success in this evolving landscape.