Latest news with #AlanBenjamin


Business Wire
02-05-2025
- Business
- Business Wire
GigaIO to Showcase Next-Generation AI Fabric Technology at ISC 2025
CARLSBAD, Calif.--(BUSINESS WIRE)--GigaIO, a pioneer in scalable edge-to-core AI platforms for all accelerators that are easy to deploy and manage, will showcase its latest innovations at ISC High Performance 2025, taking place June 10-13 in Hamburg, Germany. Visitors to stand H22 can see how GigaIO's revolutionary AI fabric technology, which seamlessly bridges from edge to core with a dynamic, open platform built for any accelerator, powers its two flagship products, SuperNODE and Gryf. "GigaIO's rail-optimized, PCIe-based AI fabric topologies offer up to 3.7x improved collective performance with an accelerator-agnostic design, ensuring adaptability across diverse AI workloads." SuperNODE is the world's most powerful and energy-efficient scale-up AI computing platform, and Gryf is the first suitcase-sized AI supercomputer that brings datacenter-class computing power directly to the edge. GigaIO's architecture, powered by its AI fabric, effortlessly integrates GPUs and inference accelerators from NVIDIA, AMD, d-Matrix, and more, enabling organizations to slash power and cooling requirements by up to 30% without compromising performance. GigaIO's AI fabric implements a native PCIe Gen5 architecture that enables direct memory-semantic communication between distributed computing resources, eliminating protocol translation overhead while maintaining sub-microsecond latencies for GPU-to-GPU transfers. This enables AI workloads to achieve near-linear scaling across pooled accelerators that appear as if locally attached to the host. GigaIO's groundbreaking paper, 'Rail Optimized PCIe Topologies for LLMs,' was selected for presentation at ISC 2025 on Thursday, 12 June 2025, from 9:00am to 9:25am in Hall F (2nd floor). This research explores optimized network architectures for large language model training and inference. Scaling LLMs efficiently requires innovative approaches to GPU interconnects, and GigaIO's rail-optimized, PCIe-based AI fabric topologies offer up to 3.7x improved collective performance with an accelerator-agnostic design, ensuring adaptability across diverse AI workloads. 'ISC 2025 arrives at a critical juncture, as AI workloads demand unprecedented hardware resources, making optimized infrastructure essential for organizations to achieve their performance targets,' said Alan Benjamin, CEO of GigaIO. 'Our expanded conference participation will demonstrate how our PCIe-based fabric technology delivers superior performance for LLM training and inference, while dramatically reducing power consumption and total cost of ownership.' Stop by stand H22 at ISC 2025 or schedule a meeting during the event with the GigaIO team. About GigaIO GigaIO redefines scalable AI infrastructure, seamlessly bridging from edge to core with a dynamic, open platform built for every accelerator. Reduce power draw with GigaIO's SuperNODE, the world's most powerful and energy-efficient scale-up AI computing platform. Run AI jobs anywhere with Gryf, the world's first suitcase-sized AI supercomputer that brings datacenter-class computing power directly to the edge. Both are easy to deploy and manage, utilizing GigaIO's patented AI fabric that provides ultra-low latency and direct memory-to-memory communication between GPUs for near-perfect scaling for AI workloads. Visit or follow on Twitter (X) and LinkedIn.


Business Wire
01-05-2025
- Business
- Business Wire
GigaIO and d-Matrix Advance Strategic Collaboration to Build World's Most Efficient Scalable Inference Solution for Enterprise AI Deployment
CARLSBAD, Calif.--(BUSINESS WIRE)--GigaIO, a pioneer in scalable edge-to-core AI platforms for all accelerators that are easy to deploy and manage, today announced the next phase of its strategic partnership with d-Matrix to deliver the world's most expansive inference solution for enterprises deploying AI at scale. Integrating d-Matrix's revolutionary Corsair inference platform into GigaIO's SuperNODE architecture creates an unparalleled solution that eliminates the complexity and performance bottlenecks traditionally associated with large-scale AI inference deployment. "Combining GigaIO's scale-up AI architecture with d-Matrix's purpose-built inference acceleration technology delivers unprecedented token generation speeds and memory bandwidth, while significantly reducing power consumption and total cost of ownership." Share This joint solution addresses the growing demand from enterprises for high-performance, energy-efficient AI inference capabilities that can scale seamlessly without the typical limitations of multi-node configurations. Combining GigaIO's industry-leading scale-up AI architecture with d-Matrix's purpose-built inference acceleration technology produces a solution that delivers unprecedented token generation speeds and memory bandwidth, while significantly reducing power consumption and total cost of ownership. Revolutionary Performance Through Technological Integration The new GigaIO SuperNODE platform, capable of supporting dozens of d-Matrix Corsair accelerators in a single node, is now the industry's most scalable AI inference platform. This integration enables enterprises to deploy ultra-low-latency batched inference workloads at scale without the complexity of traditional distributed computing approaches. 'By combining d-Matrix's Corsair PCIe cards with the industry-leading scale-up architecture of GigaIO's SuperNODE, we've created a transformative solution for enterprises deploying next-generation AI inference at scale,' said Alan Benjamin, CEO of GigaIO. 'Our single-node server eliminates complex multi-node configurations and simplifies deployment, enabling enterprises to quickly adapt to evolving AI workloads while significantly improving their TCO and operational efficiency.' The combined solution delivers exceptional performance metrics that redefine what's possible for enterprise AI inference: Processing capability of 30,000 tokens per second at just 2 milliseconds per token for models like Llama3 70B Up to 10x faster interactive speed compared with GPU-based solutions 3x better performance at a similar total cost of ownership 3x greater energy efficiency for more sustainable AI deployments 'When we started d-Matrix in 2019, we looked at the landscape of AI compute and made a bet that inference would be the largest computing opportunity of our lifetime,' said Sid Sheth, founder and CEO of d-Matrix. 'Our collaboration with GigaIO brings together our ultra-efficient in-memory compute architecture with the industry's most powerful scale-up platform, delivering a solution that makes enterprise-scale generative AI commercially viable and accessible.' This integration leverages GigaIO's cutting-edge PCIe Gen 5-based AI fabric, which delivers low-latency communication between multiple d-Matrix Corsair accelerators with near-zero latency. This architectural approach eliminates the traditional bottlenecks associated with distributed inference workloads while maximizing the efficiency of d-Matrix's Digital In-Memory Compute (DIMC) architecture, which delivers an industry-leading 150 TB/s memory bandwidth. Industry Recognition and Performance Validation This partnership builds on GigaIO's recent achievement of recording the highest tokens per second for a single node in the MLPerf Inference: Datacenter benchmark database, further validating the company's leadership in scale-up AI infrastructure. 'The market has been demanding more efficient, scalable solutions for AI inference workloads that don't compromise performance,' added Benjamin. 'Our partnership with d-Matrix brings together the tremendous engineering innovation of both companies, resulting in a solution that redefines what's possible for enterprise AI deployment.' Those interested in early access to SuperNODEs running Corsair accelerators can indicate interest here. About GigaIO GigaIO redefines scalable AI infrastructure, seamlessly bridging from edge to core with a dynamic, open platform built for every accelerator. Reduce power draw with GigaIO's SuperNODE, the world's most powerful and energy-efficient scale-up AI computing platform. Run AI jobs anywhere with Gryf, the world's first suitcase-sized AI supercomputer that brings datacenter-class computing power directly to the edge. Both are easy to deploy and manage, utilizing GigaIO's patented AI fabric that provides ultra-low latency and direct memory-to-memory communication between GPUs for near-perfect scaling for AI workloads. Visit or follow on Twitter (X) and LinkedIn. About d-Matrix d-Matrix is transforming the economics of large-scale inference with the world's most efficient AI computing platform for inference in data centers. The company's Corsair platform leverages innovative Digital In-Memory Compute (DIMC) architecture to accelerate AI inference workloads with industry-leading real-time performance, energy efficiency, and cost savings compared to GPUs and other alternatives. d-Matrix delivers ultra-low latency without compromising throughput, unlocking the next wave of Generative AI use cases while enabling commercially viable AI computing that scales with model size to empower companies of all sizes and budgets. For more information, visit


Business Wire
25-04-2025
- Business
- Business Wire
GigaIO Announces General Availability for Gryf, the World's First Portable AI Supercomputer
CARLSBAD, Calif.--(BUSINESS WIRE)--GigaIO, a pioneer in scalable edge-to-core AI platforms for all accelerators that are easy to deploy and manage, today announced the general availability of Gryf™, the world's first suitcase-sized AI supercomputer. Co-designed by GigaIO and SourceCode, Gryf delivers datacenter-class computing power directly to edge operations, enabling real-time intelligence and analytics in previously impossible field conditions. The platform has already secured significant orders from the U.S. Department of Defense and the intelligence community, validating its game-changing capabilities for mission-critical applications in challenging environments. "Gryf delivers datacenter-class computing power directly to edge operations, enabling real-time intelligence and analytics in previously impossible field conditions." Share Gryf redefines on-demand configurability in the field. Powered by GigaIO's AI memory fabric, Gryf enables users to dynamically deploy applications anywhere, at any time. The revolutionary computing platform allows organizations to process critical data on-site without latency issues from data transfers, providing unprecedented computing power in a ruggedized, field-ready design that can be deployed virtually anywhere. 'Gryf represents a fundamental shift in how organizations access and utilize high-performance computing at the edge,' said Alan Benjamin, CEO of GigaIO. 'By bringing supercomputing capabilities to field operations in a portable form factor, we're enabling real-time intelligence and analytics that were previously impossible without massive infrastructure. The strong interest from defense, intelligence, sports, media organizations, and the energy sector confirms the market need for this revolutionary approach to edge computing.' Scalable and Flexible Design One of Gryf's most innovative features is its scalability, allowing users to stack up to five units for increased performance while maintaining portability. The units interconnect across GigaIO's AI fabric, allowing any server to access any resource within the fabric as if it was on a single node. This configuration can be adjusted in real-time to meet changing application requirements. The system's customizable design, featuring GPU, compute, storage, and network sleds, enables organizations to optimize configurations for specific workload demands, maximizing return on investment while providing a smaller footprint and lower power draw than traditional solutions. Meeting Critical Defense and Intelligence Requirements Designed and built in the US, Gryf has quickly gained traction within defense and intelligence sectors, where its portable data center performance and AI-enhanced capabilities enable mission-critical workloads at the tactical edge. The platform is optimized for AI, Intelligence Surveillance Reconnaissance (ISR), cybersecurity, and tactical missions, providing commanders with immediate insights without requiring data transmission to remote processing centers. The seamless edge-to-core integration between Gryf and GigaIO's SuperNODE system enables field units to operate autonomously and then synchronize instantly with central computing resources upon connection. This revolutionary capability ensures continuity of operations across the disconnected, intermittent, and limited bandwidth environments common in defense operations. Transforming Sports Analytics and Performance Monitoring Gryf significantly elevates sports analytics at remote training locations, stadiums, race courses, and practice facilities by providing instant data analysis at the edge with its portable AI-powered data center capabilities. The platform's ability to analyze speed, workload, heart rate, and other performance metrics in real time enables data-driven decisions that optimize performance and prevent injuries, providing coaches and performance staff with immediate insights during training and competition. 'Sports teams competing at the highest levels need immediate access to performance insights,' noted Lauren Spurlin, Founder, Aurified Consulting. 'Waiting hours or days for data processing is no longer acceptable. Gryf eliminates those delays by bringing the data center directly to the field, changing how teams train, compete, and recover.' Accelerating Media Production and Broadcasting Early adopters of the Gryf in media and entertainment are already seeing dramatic improvements in production workflows. Deployed on set, Gryf decreases post-production time for sports broadcasts and entertainment productions by enabling on-site processing of high-resolution content. Gryf's high-performance computing and storage capabilities support on-site live broadcasting, rapid content processing and editing, real-time analytics, and enhanced graphics generation. Particularly effective at outdoor venues with challenging high-speed connection points, such as NASCAR race tracks and sports stadiums, Gryf can create 360-degree virtual camera environments and handle resource-intensive tasks like green screen video capture directly on location, completely transforming the economics of remote production. Edge AI for Energy / Oil & Gas Exploration Energy operations demand speed, precision, and resilience in remote and rugged locations such as offshore drilling rigs and overland exploration sites. Offshore rigs lack traditional IT infrastructure, relying on high-latency satellite communications (SATCOM) for data transmission, and exploratory teams often work in remote, power-limited areas for short durations, with limited visibility into whether data collection was successful until they return. Gryf enables teams to capture, process, and act on critical data right at the edge, helping to prevent dangerous events offshore or validate drilling targets in real time during land-based surveys, processing data locally so that teams can act faster, improve safety, and maintain control over sensitive information. Additional Applications and Availability Other applications for Gryf include: Healthcare/Medical Research, enabling diagnostic analysis and patient-critical decisions in clinical environments through its configurable, facility-ready design that processes data securely on-site. Scientific Research/Field Studies, empowering scientific breakthroughs directly at research sites through its configurable, field-ready design that processes complex data at the source. Industrial/Manufacturing, providing production analysis and process optimization directly at manufacturing operations through its configurable, factory-ready design that eliminates remote processing delays. Gryf is now generally available worldwide and shipping through GigaIO's global partner network. For more information about Gryf and its applications across different industries, visit About GigaIO GigaIO redefines scalable AI infrastructure, seamlessly bridging from edge to core with a dynamic, open platform built for every accelerator. Reduce power draw with GigaIO's SuperNODE, the world's most powerful and energy-efficient scale-up AI computing platform. Run AI jobs anywhere with Gryf, the world's first suitcase-sized AI supercomputer that brings datacenter-class computing power directly to the edge. Both are easy to deploy and manage, utilizing GigaIO's patented AI fabric that provides ultra-low latency and direct memory-to-memory communication between GPUs for near-perfect scaling for AI workloads. Visit or follow on Twitter (X) and LinkedIn.
Yahoo
25-04-2025
- Business
- Yahoo
GigaIO Announces General Availability for Gryf, the World's First Portable AI Supercomputer
Groundbreaking suitcase-sized AI supercomputer shipping to DoD and Intelligence Community, has immediate applications across defense, sports analytics, media production, energy, and more. CARLSBAD, Calif., April 25, 2025--(BUSINESS WIRE)--GigaIO, a pioneer in scalable edge-to-core AI platforms for all accelerators that are easy to deploy and manage, today announced the general availability of Gryf™, the world's first suitcase-sized AI supercomputer. Co-designed by GigaIO and SourceCode, Gryf delivers datacenter-class computing power directly to edge operations, enabling real-time intelligence and analytics in previously impossible field conditions. The platform has already secured significant orders from the U.S. Department of Defense and the intelligence community, validating its game-changing capabilities for mission-critical applications in challenging environments. Gryf redefines on-demand configurability in the field. Powered by GigaIO's AI memory fabric, Gryf enables users to dynamically deploy applications anywhere, at any time. The revolutionary computing platform allows organizations to process critical data on-site without latency issues from data transfers, providing unprecedented computing power in a ruggedized, field-ready design that can be deployed virtually anywhere. "Gryf represents a fundamental shift in how organizations access and utilize high-performance computing at the edge," said Alan Benjamin, CEO of GigaIO. "By bringing supercomputing capabilities to field operations in a portable form factor, we're enabling real-time intelligence and analytics that were previously impossible without massive infrastructure. The strong interest from defense, intelligence, sports, media organizations, and the energy sector confirms the market need for this revolutionary approach to edge computing." Scalable and Flexible Design One of Gryf's most innovative features is its scalability, allowing users to stack up to five units for increased performance while maintaining portability. The units interconnect across GigaIO's AI fabric, allowing any server to access any resource within the fabric as if it was on a single node. This configuration can be adjusted in real-time to meet changing application requirements. The system's customizable design, featuring GPU, compute, storage, and network sleds, enables organizations to optimize configurations for specific workload demands, maximizing return on investment while providing a smaller footprint and lower power draw than traditional solutions. Meeting Critical Defense and Intelligence Requirements Designed and built in the US, Gryf has quickly gained traction within defense and intelligence sectors, where its portable data center performance and AI-enhanced capabilities enable mission-critical workloads at the tactical edge. The platform is optimized for AI, Intelligence Surveillance Reconnaissance (ISR), cybersecurity, and tactical missions, providing commanders with immediate insights without requiring data transmission to remote processing centers. The seamless edge-to-core integration between Gryf and GigaIO's SuperNODE system enables field units to operate autonomously and then synchronize instantly with central computing resources upon connection. This revolutionary capability ensures continuity of operations across the disconnected, intermittent, and limited bandwidth environments common in defense operations. Transforming Sports Analytics and Performance Monitoring Gryf significantly elevates sports analytics at remote training locations, stadiums, race courses, and practice facilities by providing instant data analysis at the edge with its portable AI-powered data center capabilities. The platform's ability to analyze speed, workload, heart rate, and other performance metrics in real time enables data-driven decisions that optimize performance and prevent injuries, providing coaches and performance staff with immediate insights during training and competition. "Sports teams competing at the highest levels need immediate access to performance insights," noted Lauren Spurlin, Founder, Aurified Consulting. "Waiting hours or days for data processing is no longer acceptable. Gryf eliminates those delays by bringing the data center directly to the field, changing how teams train, compete, and recover." Accelerating Media Production and Broadcasting Early adopters of the Gryf in media and entertainment are already seeing dramatic improvements in production workflows. Deployed on set, Gryf decreases post-production time for sports broadcasts and entertainment productions by enabling on-site processing of high-resolution content. Gryf's high-performance computing and storage capabilities support on-site live broadcasting, rapid content processing and editing, real-time analytics, and enhanced graphics generation. Particularly effective at outdoor venues with challenging high-speed connection points, such as NASCAR race tracks and sports stadiums, Gryf can create 360-degree virtual camera environments and handle resource-intensive tasks like green screen video capture directly on location, completely transforming the economics of remote production. Edge AI for Energy / Oil & Gas Exploration Energy operations demand speed, precision, and resilience in remote and rugged locations such as offshore drilling rigs and overland exploration sites. Offshore rigs lack traditional IT infrastructure, relying on high-latency satellite communications (SATCOM) for data transmission, and exploratory teams often work in remote, power-limited areas for short durations, with limited visibility into whether data collection was successful until they return. Gryf enables teams to capture, process, and act on critical data right at the edge, helping to prevent dangerous events offshore or validate drilling targets in real time during land-based surveys, processing data locally so that teams can act faster, improve safety, and maintain control over sensitive information. Additional Applications and Availability Other applications for Gryf include: Healthcare/Medical Research, enabling diagnostic analysis and patient-critical decisions in clinical environments through its configurable, facility-ready design that processes data securely on-site. Scientific Research/Field Studies, empowering scientific breakthroughs directly at research sites through its configurable, field-ready design that processes complex data at the source. Industrial/Manufacturing, providing production analysis and process optimization directly at manufacturing operations through its configurable, factory-ready design that eliminates remote processing delays. Gryf is now generally available worldwide and shipping through GigaIO's global partner network. For more information about Gryf and its applications across different industries, visit About GigaIO GigaIO redefines scalable AI infrastructure, seamlessly bridging from edge to core with a dynamic, open platform built for every accelerator. Reduce power draw with GigaIO's SuperNODE, the world's most powerful and energy-efficient scale-up AI computing platform. Run AI jobs anywhere with Gryf, the world's first suitcase-sized AI supercomputer that brings datacenter-class computing power directly to the edge. Both are easy to deploy and manage, utilizing GigaIO's patented AI fabric that provides ultra-low latency and direct memory-to-memory communication between GPUs for near-perfect scaling for AI workloads. Visit or follow on Twitter (X) and LinkedIn. View source version on Contacts For media inquiries, please contact:Shannon Biggs760-487-8395shannon@ Sign in to access your portfolio