logo

Technology Innovation Institute Announces Falcon-H1 model availability as NVIDIA NIM to Deliver Sovereign AI at Scale

Mid East Info6 days ago

Flagship, top ranking, open-source AI model to be production-ready via new NVIDIA NIM microservices that deliver enterprise-ready inference for thousands of LLMs
Paris, France – Abu Dhabi, UAE – June 2025: Abu Dhabi's Technology Innovation Institute (TII), a leading global research center and the developer behind the globally ranked Falcon open-source AI models and privacy-preserving technologies, today announced that Falcon-H1, its next-generation, hybrid-architecture large language model, will be available as an NVIDIA NIM microservice.
The announcement, timed with NVIDIA's GTC Paris showcase, positions Falcon-H1 for seamless enterprise deployment across cloud, on-premise, or hybrid environments. Developers can soon access and scale Falcon-H1 with production-grade performance, without the engineering overhead typically required to adapt open-source models for real-world application.
Dr. Najwa Aaraj, CEO of TII, commented: 'Falcon-H1's availability on NVIDIA NIM reflects our ongoing leadership in shaping the future of open, sovereign, and cross-domain deployment ready AI. It demonstrates that breakthrough innovation from our region is not only competitive on the global stage – it's setting new benchmarks for scalable, secure, and enterprise-ready AI.'
At the heart of Falcon-H1 is a novel hybrid Transformer–Mamba architecture, combining the efficiency of state space models (SSMs) with the expressiveness of Transformer networks. Designed in-house by TII researchers, the architecture supports context windows of up to 256k tokens, an order-of-magnitude leap in long-context reasoning, while preserving high-speed inference and reduced memory demands. Multilingual by design, Falcon-H1 delivers robust performance ahead of models in its category, across both high- and low-resource languages, making it suited for global-scale applications.
Supported soon for deployment via the universal LLM NIM microservice, Falcon-H1 becomes a plug-and-play asset for enterprises building agentic systems, retrieval-augmented generation (RAG) workflows, or domain-specific assistants. Whether running with NVIDIA TensorRT-LLM, vLLM, or SGLang, NIM abstracts away the underlying inference stack, enabling developers to deploy Falcon-H1 in minutes using standard tools such as Docker and Hugging Face, with automated hardware optimization and enterprise-grade SLAs.
'Falcon-H1's availability on NVIDIA NIM bridges the gap between cutting-edge model design and real-world operability. It combines our hybrid architecture with the performance and reliability of NVIDIA microservices. Developers can integrate Falcon-H1 optimized for long-context reasoning, multilingual versatility, and real-world applications. What once required weeks of infrastructure tuning becomes achievable in minutes at scale, with multilingual depth, and production resilience', said Dr. Hakim Hacid, Chief AI Researcher at TII. The release also mark Falcon-H1's integration with NVIDIA NeMo microservices and NVIDIA AI Blueprints, giving developers access to full lifecycle tooling, from data curation and guardrailing to continuous evaluation and post-deployment tuning. Crucially, this makes Falcon-H1 viable in regulated, latency-sensitive and sovereign AI contexts, with full-stack NVIDIA support.
With over 55 million downloads to date, the Falcon series has become one of the most widely adopted open-source models from the Middle East region. Beyond its scale, Falcon-H1 smaller variants routinely outperform larger peers on reasoning and mathematical tasks, while the 34B model now leads several industry benchmarks.
TII's strategic alignment with NVIDIA's validated deployment framework reflects that open-source models are production-ready assets. Falcon-H1's availability on NIM cements its place among them as a sovereign, scalable, and secure alternative to closed-weight incumbents.
About the Technology Innovation Institute:
The Technology Innovation Institute (TII) is the dedicated applied research pillar of Abu Dhabi's Advanced Technology Research Council (ATRC). TII is a pioneering global research and development center that focuses on applied research and new-age technology capabilities. The Institute has 10 dedicated research centers in advanced materials, autonomous robotics, cryptography, AI and digital science, directed energy, quantum, secure systems, propulsion and space, biotechnology, and renewable and sustainable energy. By working with exceptional talent, universities, research institutions, and industry partners from all over the world, TII connects an intellectual community and contributes to building an R&D ecosystem that reinforces the status of Abu Dhabi and the UAE as a global hub for innovation.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Vertiv develops Energy-Efficient Cooling and Power Reference Architecture for the NVIDIA GB300 NVL72 Platform
Vertiv develops Energy-Efficient Cooling and Power Reference Architecture for the NVIDIA GB300 NVL72 Platform

Mid East Info

timea day ago

  • Mid East Info

Vertiv develops Energy-Efficient Cooling and Power Reference Architecture for the NVIDIA GB300 NVL72 Platform

Available as SimReady™ Assets in NVIDIA Omniverse Blueprint for AI Factory Design and Operations New design demonstrates end-to-end energy-efficient cooling and power to support densities up to 142kW per rack Vertiv (NYSE: VRT), a global leader in critical digital infrastructure, announced their energy-efficient 142kW cooling and power reference architecture for the NVIDIA GB300 NVL72 platform. In addition to fully integrated end-to-end cooling and power strategies for this next generation platform, the reference architecture unlocks a new era of infrastructure design by shifting from traditional buildouts. Vertiv solutions are available as SimReady™ 3D assets in the NVIDIA Omniverse Blueprint for AI factory design and operations. The reference architecture can be customized for bespoke data center designs to dramatically reduce planning times and reduce risk during the buildout process. Vertiv works with NVIDIA and its data center roadmaps, to develop AI infrastructure strategies and deployment-ready designs that anticipate increased rack power densities. For example, Vertiv is collaborating with NVIDIA to support 800 VDC data center power infrastructure for 1 MW IT racks and beyond, with Vertiv solutions available starting in 2026. 'Building upon our strong collaboration with NVIDIA and a shared vision to advance AI infrastructure, Vertiv is proud to release our comprehensive reference design and SimReady™ 3D assets for the NVIDIA GB300 NVL72 platform,' stated Scott Armul, executive vice president of global portfolio and business units at Vertiv. 'Vertiv solutions allow customers to not only scale faster, but to optimize their AI-focused data centers digitally before a single physical module is built.' 'By combining NVIDIA's advanced AI platforms with Vertiv's expertise in power and cooling infrastructure, we're enabling customers to deploy next-generation data centers that are more efficient, scalable, and ready for the most demanding AI workloads,' said Dion Harris, senior director of HPC and AI Infrastructure at NVIDIA. 'Together, we're helping organizations unlock new levels of performance and sustainability as they build the future of AI.' The reference architecture for NVIDIA GB300 NVL72 is part of the Vertiv™ 360AI infrastructure platform, designed to help customers navigate integrated solutions for powering and cooling AI workloads and other high-performance computing applications. Key benefits of Vertiv's reference architecture and SimReady™ 3D assets for the NVIDIA GB300 NVL72 design include: Simulation to Deployment in One Path: Leveraging NVIDIA Omniverse technologies in the architecture bridges physical and digital infrastructure, enabling real-time system-level collaboration across IT, facilities, and engineering teams – allowing customers to test, optimize and validate the design before a single component is built. Leveraging NVIDIA Omniverse technologies in the architecture bridges physical and digital infrastructure, enabling real-time system-level collaboration across IT, facilities, and engineering teams – allowing customers to test, optimize and validate the design before a single component is built. Built for the Age of AI Reasoning: Supporting rack densities up to 142kW, the reference architecture for NVIDIA GB300 NVL72 defines the new benchmark for accelerated infrastructure — engineered for AI factories at scale. Supporting rack densities up to 142kW, the reference architecture for NVIDIA GB300 NVL72 defines the new benchmark for accelerated infrastructure — engineered for AI factories at scale. Accelerated Performance, Scale, and Speed: Delivering 1.5x more AI performance, up to 50% faster on-site builds and in 30% less physical space, Vertiv's reference architecture for NVIDIA GB300 NVL72 enables customers to scale AI infrastructure with unprecedented agility and energy efficiency. Delivering 1.5x more AI performance, up to 50% faster on-site builds and in 30% less physical space, Vertiv's reference architecture for NVIDIA GB300 NVL72 enables customers to scale AI infrastructure with unprecedented agility and energy efficiency. Thermal Flexibility Meets Efficiency Gains: Liquid cooling-ready, supporting both air- and hybrid- cooled configurations, achieves up to 70% better annual energy efficiency allowed by higher water Liquid cooling-ready, supporting both air- and hybrid- cooled configurations, achieves up to 70% better annual energy efficiency allowed by higher water Design, Deploy, Operate – Globally: Backed by 4,000+ Vertiv field service engineers globally, Vertiv's reference architecture for GB300 NVL72 brings unmatched scalability, serviceability, and supply chain reliability to global AI rollouts. As global AI demand accelerates exponentially, Vertiv is constantly working to deliver critical energy-efficient cooling and power solutions, to enable AI-ready infrastructure to run the next generation of AI factories. To learn more, visit

Armenia to Launch $500M AI Factory with NVIDIA
Armenia to Launch $500M AI Factory with NVIDIA

See - Sada Elbalad

time4 days ago

  • See - Sada Elbalad

Armenia to Launch $500M AI Factory with NVIDIA

By Ahmad El-Assasy In a bold step toward becoming a global hub for artificial intelligence, Armenia has announced the construction of a $500 million AI factory in partnership with U.S. tech giant NVIDIA, investment group Firebird, and Team Group. The facility, set to be the largest AI initiative in the region, is scheduled to begin operations by 2026. Minister of High-Tech Industry Mkhitar Hayrapetyan described the initiative as a 'historic leap' for Armenia, positioning the country as a key innovator in the global AI revolution. 'This is not just progress — it's a transformative leap. Armenia is stepping into the global AI revolution as a creator, builder, and leader,' Hayrapetyan said. A Regional AI Powerhouse The AI facility will be equipped with thousands of NVIDIA Blackwell GPUs and backed by over 100 megawatts of infrastructure power, making it one of the most advanced in Eastern Europe and the Caucasus. More than just a data center, the project will serve as a comprehensive AI research, education, and innovation hub, empowering Armenian engineers, scientists, and students to drive digital transformation at home. NVIDIA CEO Jensen Huang stated: 'AI factories are the infrastructure of the 21st century. Our collaboration with Armenia will help build critical capabilities and unlock new economic and technological opportunities across the region.' Creating Jobs and Driving Innovation The project is expected to create thousands of high-skill jobs, attract foreign investment, and reverse the brain drain trend by providing local opportunities for Armenia's brightest minds. The Armenian government sees this as a strategic investment in long-term prosperity. Firebird co-founder and CEO Razmig Hovaghimian called it a 'launchpad for innovation': 'From Armenia to the world, we're building a platform for robotics, science, and AI development in partnership with top global universities.' Team Group, which runs Armenia's national telecom operator "Team Telecom Armenia" and is a stakeholder in Georgia's SkyTel, will provide the digital infrastructure and expertise needed to bring the project to life. 'This is a testament to Armenia's potential to become a regional leader in information technology and AI,' said Team Group co-founder Alexander Yesayan. Global Partnerships, Local Impact The Aviean Foundation, led by Noubar Afeyan, CEO of Flagship Pioneering, will also join as a founding investor and strategic advisor to Firebird. Afeyan emphasized the need for long-term global collaboration in making Armenia a visible player in the AI ecosystem. U.S. Ambassador to Armenia Kristina Kvien welcomed the initiative: 'We are excited about the potential for U.S. AI and tech exports to accelerate Armenia's innovation economy. Companies like NVIDIA are ideal partners for Armenia's transformation.' This landmark project reflects a strong synergy between government policy, private tech leadership, and global cooperation — all aimed at establishing Armenia as a vital AI and digital hub in the region. read more Gold prices rise, 21 Karat at EGP 3685 NATO's Role in Israeli-Palestinian Conflict US Expresses 'Strong Opposition' to New Turkish Military Operation in Syria Shoukry Meets Director-General of FAO Lavrov: confrontation bet. nuclear powers must be avoided News Iran Summons French Ambassador over Foreign Minister Remarks News Aboul Gheit Condemns Israeli Escalation in West Bank News Greek PM: Athens Plays Key Role in Improving Energy Security in Region News One Person Injured in Explosion at Ukrainian Embassy in Madrid News China Launches Largest Ever Aircraft Carrier Sports Former Al Zamalek Player Ibrahim Shika Passes away after Long Battle with Cancer Lifestyle Get to Know 2025 Eid Al Adha Prayer Times in Egypt Business Fear & Greed Index Plummets to Lowest Level Ever Recorded amid Global Trade War Arts & Culture Zahi Hawass: Claims of Columns Beneath the Pyramid of Khafre Are Lies News Flights suspended at Port Sudan Airport after Drone Attacks News Shell Unveils Cost-Cutting, LNG Growth Plan Videos & Features Video: Trending Lifestyle TikToker Valeria Márquez Shot Dead during Live Stream Technology 50-Year Soviet Spacecraft 'Kosmos 482' Crashes into Indian Ocean News 3 Killed in Shooting Attack in Thailand

Technology Innovation Institute Announces Falcon-H1 model availability as NVIDIA NIM to Deliver Sovereign AI at Scale
Technology Innovation Institute Announces Falcon-H1 model availability as NVIDIA NIM to Deliver Sovereign AI at Scale

Mid East Info

time6 days ago

  • Mid East Info

Technology Innovation Institute Announces Falcon-H1 model availability as NVIDIA NIM to Deliver Sovereign AI at Scale

Flagship, top ranking, open-source AI model to be production-ready via new NVIDIA NIM microservices that deliver enterprise-ready inference for thousands of LLMs Paris, France – Abu Dhabi, UAE – June 2025: Abu Dhabi's Technology Innovation Institute (TII), a leading global research center and the developer behind the globally ranked Falcon open-source AI models and privacy-preserving technologies, today announced that Falcon-H1, its next-generation, hybrid-architecture large language model, will be available as an NVIDIA NIM microservice. The announcement, timed with NVIDIA's GTC Paris showcase, positions Falcon-H1 for seamless enterprise deployment across cloud, on-premise, or hybrid environments. Developers can soon access and scale Falcon-H1 with production-grade performance, without the engineering overhead typically required to adapt open-source models for real-world application. Dr. Najwa Aaraj, CEO of TII, commented: 'Falcon-H1's availability on NVIDIA NIM reflects our ongoing leadership in shaping the future of open, sovereign, and cross-domain deployment ready AI. It demonstrates that breakthrough innovation from our region is not only competitive on the global stage – it's setting new benchmarks for scalable, secure, and enterprise-ready AI.' At the heart of Falcon-H1 is a novel hybrid Transformer–Mamba architecture, combining the efficiency of state space models (SSMs) with the expressiveness of Transformer networks. Designed in-house by TII researchers, the architecture supports context windows of up to 256k tokens, an order-of-magnitude leap in long-context reasoning, while preserving high-speed inference and reduced memory demands. Multilingual by design, Falcon-H1 delivers robust performance ahead of models in its category, across both high- and low-resource languages, making it suited for global-scale applications. Supported soon for deployment via the universal LLM NIM microservice, Falcon-H1 becomes a plug-and-play asset for enterprises building agentic systems, retrieval-augmented generation (RAG) workflows, or domain-specific assistants. Whether running with NVIDIA TensorRT-LLM, vLLM, or SGLang, NIM abstracts away the underlying inference stack, enabling developers to deploy Falcon-H1 in minutes using standard tools such as Docker and Hugging Face, with automated hardware optimization and enterprise-grade SLAs. 'Falcon-H1's availability on NVIDIA NIM bridges the gap between cutting-edge model design and real-world operability. It combines our hybrid architecture with the performance and reliability of NVIDIA microservices. Developers can integrate Falcon-H1 optimized for long-context reasoning, multilingual versatility, and real-world applications. What once required weeks of infrastructure tuning becomes achievable in minutes at scale, with multilingual depth, and production resilience', said Dr. Hakim Hacid, Chief AI Researcher at TII. The release also mark Falcon-H1's integration with NVIDIA NeMo microservices and NVIDIA AI Blueprints, giving developers access to full lifecycle tooling, from data curation and guardrailing to continuous evaluation and post-deployment tuning. Crucially, this makes Falcon-H1 viable in regulated, latency-sensitive and sovereign AI contexts, with full-stack NVIDIA support. With over 55 million downloads to date, the Falcon series has become one of the most widely adopted open-source models from the Middle East region. Beyond its scale, Falcon-H1 smaller variants routinely outperform larger peers on reasoning and mathematical tasks, while the 34B model now leads several industry benchmarks. TII's strategic alignment with NVIDIA's validated deployment framework reflects that open-source models are production-ready assets. Falcon-H1's availability on NIM cements its place among them as a sovereign, scalable, and secure alternative to closed-weight incumbents. About the Technology Innovation Institute: The Technology Innovation Institute (TII) is the dedicated applied research pillar of Abu Dhabi's Advanced Technology Research Council (ATRC). TII is a pioneering global research and development center that focuses on applied research and new-age technology capabilities. The Institute has 10 dedicated research centers in advanced materials, autonomous robotics, cryptography, AI and digital science, directed energy, quantum, secure systems, propulsion and space, biotechnology, and renewable and sustainable energy. By working with exceptional talent, universities, research institutions, and industry partners from all over the world, TII connects an intellectual community and contributes to building an R&D ecosystem that reinforces the status of Abu Dhabi and the UAE as a global hub for innovation.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store