PR Archives:  LatestBy Company By Date


Press Release -- November 18th, 2024
Source: cloudian
Tags:

Cloudian Announces Support for NVIDIA GPUDirect Acceleration for Object Storage

Simplifies AI Data Storage with a Cost-Effective, Scalable, High-Performance Object Storage Solution

SYDNEY – 19 November 2024 – Cloudian, the global leader in secure S3-compatible AI data lake platforms, today announced its integration with NVIDIA Magnum IO™ GPUDirectÒ Storage technology, delivering breakthrough performance of over 200GB/s from a HyperStore system. Cloudian HyperStore® with GPUDirect access simplifies the management of AI training and inference datasets – at petabyte and exabyte scales – while reducing costs by eliminating the need for complex data migrations and legacy file storage layers.

Key Benefits of NVIDIA GPUDirect Storage for Object Storage in AI Training and Inference Workflows:

  • Limitless Scalability: Expands effortlessly to exabyte scale without disruption, supporting growing AI datasets without adding management complexity.
  • Reduced Costs and No Data Migrations: Removes legacy file layers and enables a single, unified data lake without the need for constant data movement between tiers.
  • High Performance: Delivers over 200GB/s from a single system with performance sustained over a 30-minute period without the use of data caching.
  • Maximised CPU for AI Workloads: Slashes CPU overhead by 45 per cent during data transfers, freeing computational resources for AI processing.
  • No Kernel Modifications: Eliminates the security exposure of vendor-specific kernel modifications, reducing security vulnerabilities
  • Integrated Metadata: Rich metadata facilitates rapid search without the need for external databases.

“Cloudian is proud to be at the forefront of transforming how enterprises and AI hyperscalers harness data to realise the power of AI,” said Michael Tso, CEO of Cloudian. “For too long, AI users have been saddled with the unnecessary complexity and performance bottlenecks of legacy storage solutions. With GPUDirect Storage integration, we are enabling AI workflows to directly leverage a simply scalable storage architecture so organisations can unleash the full potential of their data.”

“At Supermicro, we’re committed to delivering the most advanced and efficient solutions for AI and deep learning,” said Michael McNerney, SVP of Marketing and Network Security at Supermicro. “Cloudian’s integration of NVIDIA GPUDirect Storage with the HyperStore line of object storage appliances based on Supermicro systems – including the Hyper 2U and 1U servers, the high-density SuperStorage 90-bay storage servers, and the Simply Double 2U 24-bay storage servers – represents a significant innovation in the use of object storage for AI workloads. This will enable our mutual customers to deploy more powerful and cost-effective AI infrastructure at scale.”

“Fast, consistent, and scalable performance in object storage systems is crucial for AI workflows,” said Rob Davis, vice president of storage technology, NVIDIA. “It enables real-time processing and decision-making, which are essential for applications like fraud detection and personalised recommendations.”

Simplifies Data Management, Exabyte Scale Eliminates Data Migration
Legacy file-based storage systems in AI workflows often require frequent data movement between long-term and high-speed storage, adding management complexity. With Cloudian’s solution, AI training and inference happen directly on the data in-place, accelerating workflows and eliminating frequent migration. Cloudian HyperStore’s limitless scalability enables AI data lakes to grow to exabyte levels, while its centralised management ensures simple, unified control across multi-data centre and multi-tenant environments.

Fast Throughput for Higher GPU Utilisation
NVIDIA GPUDirect Storage with NVIDIA ConnectX and NVIDIA BlueField networking technologies optimise data transfer speeds by enabling direct communication between NVIDIA GPUs and multiple Cloudian storage nodes, bypassing the CPU. This direct parallel data transfer delivers consistent and scalable performance over 200GB/s from a HyperStore system—as measured on the industry-standard GOSBench benchmark over a sustained period without the use of data caching. As throughput can be easily and economically scaled, organisations can achieve better GPU utilisation and lower GPU communications latency.

Reduces Storage Costs
Managing the enormous datasets needed for AI workflows can be both costly and resource intensive. Cloudian’s software-defined platform helps address these challenges by eliminating the need for a separate file storage layer. With AI workflows occurring directly within the object-based data lake, organisations can streamline data management while significantly reducing operational and capital expenses, as well as overall complexity.

No Kernel Level Modifications

GPUDirect for Object Storage requires no vendor-driven kernel-level modifications. Unlike file solutions, this approach reduces potential vulnerabilities typically associated with kernel changes. By eliminating the need for such alterations, it simplifies system administration, decreases attack surfaces, and lowers the risk of security breaches.

Integrated Metadata for Simplicity and Accelerated Search
Metadata plays a crucial role in AI workflows by enabling rapid data discovery, retrieval, and access control. Cloudian accelerates AI data searches with integrated metadata support that allows for easy tagging, classification, and indexing of large datasets. Unlike legacy file-based systems, which depend on rigid directory structures and separate databases for metadata management, Cloudian natively handles metadata within the object storage platform, simplifying workflows and speeding up AI training and inference processes.

Enhanced Data Security
Data privacy and security are top priorities for enterprises adopting AI, as noted by Forrester analysts. Cloudian addresses these concerns with the industry’s most comprehensive range of security features. These include advanced access controls, encryption protocols, integrated key management, and S3 Object Lock for ransomware protection, helping ensure that sensitive AI data remains safe and secure throughout its lifecycle.

Reduced CPU Consumption

Cloudian’s integration with NVIDIA’s GPUDirect Storage technology enables direct data transfers between storage systems and GPU memory, bypassing the CPU. This direct path reduces CPU utilisation by 45 per cent during data transfers, allowing the CPU to focus on other tasks and improving overall system efficiency.

Cloudian HyperStore with NVIDIA Magnum IO GPUDirect Storage technology is available now.

“At Softsource vBridge, we’ve seen firsthand how data management challenges can hinder AI adoption,” said David Small, Group Technology Officer of Softsource vBridge. “Cloudian’s GPUDirect for Object Storage will simplify the entire AI data lifecycle, which could be the key to democratising AI across various business sectors, allowing companies of all sizes to harness the power of their data. We’re particularly excited about how this could accelerate AI projects for our mid-market clients who have previously found enterprise AI solutions out of reach.”

##

PR Archives: Latest, By Company, By Date