Latest news with #Falcon-H1


Al Etihad
12-06-2025
- Business
- Al Etihad
Falcon-H1 language model to be offered as NVIDIA NIM Microservice
12 June 2025 19:17 ABU DHABI (ALETIHAD) The Technology Innovation Institute (TII), a leading global research centre headquartered in Abu Dhabi, has announced that its next-generation open-source language model, Falcon-H1, will be available as a NVIDIA Inference Microservice (NIM), a move that significantly enhances the model's enterprise deployment in parallel with NVIDIA's GTC showcase in Paris, the initiative positions Falcon-H1 for immediate integration across cloud, on-premise and hybrid environments, a statement from TII said. The availability of Falcon-H1 on NIM will enable developers to deploy it with production-grade performance, bypassing the need for complex infrastructure customisation.'Falcon-H1's availability on NVIDIA NIM reflects our ongoing leadership in shaping the future of open, sovereign, and cross-domain deployment-ready AI. It demonstrates that breakthrough innovation from our region is not only competitive on the global stage – it's setting new benchmarks for scalable, secure, and enterprise-ready AI,' said Dr. Najwa Aaraj, CEO of is built on an innovative hybrid Transformer–Mamba architecture, blending the efficiency of state space models (SSMs) with the advanced reasoning power of Transformer networks. This design supports context windows of up to 256,000 tokens, enabling long-context reasoning and high-speed inference with reduced memory overhead. Its multilingual design ensures competitive performance across both high- and low-resource to Dr. Hakim Hacid, Chief AI Researcher at TII, 'Falcon-H1's availability on NVIDIA NIM bridges the gap between cutting-edge model design and real-world operability. It combines our hybrid architecture with the performance and reliability of NVIDIA microservices. Developers can integrate Falcon-H1 optimised for long-context reasoning, multilingual versatility, and real-world applications. What once required weeks of infrastructure tuning becomes achievable in minutes at scale, with multilingual depth, and production resilience.'The integration with NVIDIA's NeMo microservices and AI Blueprints adds full lifecycle tooling for tasks such as data curation, safety guardrails, and post-deployment optimisation. This makes Falcon-H1 a fit for regulated and latency-sensitive AI deployments, further enhancing its credentials as a sovereign AI solution. With over 55 million downloads, the Falcon model family is among the most widely adopted open-source AI efforts from the Middle East. TII's alignment with NVIDIA's enterprise-grade deployment frameworks affirms Falcon-H1's status as a production-ready, sovereign AI alternative to proprietary systems.


Mid East Info
11-06-2025
- Business
- Mid East Info
Technology Innovation Institute Announces Falcon-H1 model availability as NVIDIA NIM to Deliver Sovereign AI at Scale
Flagship, top ranking, open-source AI model to be production-ready via new NVIDIA NIM microservices that deliver enterprise-ready inference for thousands of LLMs Paris, France – Abu Dhabi, UAE – June 2025: Abu Dhabi's Technology Innovation Institute (TII), a leading global research center and the developer behind the globally ranked Falcon open-source AI models and privacy-preserving technologies, today announced that Falcon-H1, its next-generation, hybrid-architecture large language model, will be available as an NVIDIA NIM microservice. The announcement, timed with NVIDIA's GTC Paris showcase, positions Falcon-H1 for seamless enterprise deployment across cloud, on-premise, or hybrid environments. Developers can soon access and scale Falcon-H1 with production-grade performance, without the engineering overhead typically required to adapt open-source models for real-world application. Dr. Najwa Aaraj, CEO of TII, commented: 'Falcon-H1's availability on NVIDIA NIM reflects our ongoing leadership in shaping the future of open, sovereign, and cross-domain deployment ready AI. It demonstrates that breakthrough innovation from our region is not only competitive on the global stage – it's setting new benchmarks for scalable, secure, and enterprise-ready AI.' At the heart of Falcon-H1 is a novel hybrid Transformer–Mamba architecture, combining the efficiency of state space models (SSMs) with the expressiveness of Transformer networks. Designed in-house by TII researchers, the architecture supports context windows of up to 256k tokens, an order-of-magnitude leap in long-context reasoning, while preserving high-speed inference and reduced memory demands. Multilingual by design, Falcon-H1 delivers robust performance ahead of models in its category, across both high- and low-resource languages, making it suited for global-scale applications. Supported soon for deployment via the universal LLM NIM microservice, Falcon-H1 becomes a plug-and-play asset for enterprises building agentic systems, retrieval-augmented generation (RAG) workflows, or domain-specific assistants. Whether running with NVIDIA TensorRT-LLM, vLLM, or SGLang, NIM abstracts away the underlying inference stack, enabling developers to deploy Falcon-H1 in minutes using standard tools such as Docker and Hugging Face, with automated hardware optimization and enterprise-grade SLAs. 'Falcon-H1's availability on NVIDIA NIM bridges the gap between cutting-edge model design and real-world operability. It combines our hybrid architecture with the performance and reliability of NVIDIA microservices. Developers can integrate Falcon-H1 optimized for long-context reasoning, multilingual versatility, and real-world applications. What once required weeks of infrastructure tuning becomes achievable in minutes at scale, with multilingual depth, and production resilience', said Dr. Hakim Hacid, Chief AI Researcher at TII. The release also mark Falcon-H1's integration with NVIDIA NeMo microservices and NVIDIA AI Blueprints, giving developers access to full lifecycle tooling, from data curation and guardrailing to continuous evaluation and post-deployment tuning. Crucially, this makes Falcon-H1 viable in regulated, latency-sensitive and sovereign AI contexts, with full-stack NVIDIA support. With over 55 million downloads to date, the Falcon series has become one of the most widely adopted open-source models from the Middle East region. Beyond its scale, Falcon-H1 smaller variants routinely outperform larger peers on reasoning and mathematical tasks, while the 34B model now leads several industry benchmarks. TII's strategic alignment with NVIDIA's validated deployment framework reflects that open-source models are production-ready assets. Falcon-H1's availability on NIM cements its place among them as a sovereign, scalable, and secure alternative to closed-weight incumbents. About the Technology Innovation Institute: The Technology Innovation Institute (TII) is the dedicated applied research pillar of Abu Dhabi's Advanced Technology Research Council (ATRC). TII is a pioneering global research and development center that focuses on applied research and new-age technology capabilities. The Institute has 10 dedicated research centers in advanced materials, autonomous robotics, cryptography, AI and digital science, directed energy, quantum, secure systems, propulsion and space, biotechnology, and renewable and sustainable energy. By working with exceptional talent, universities, research institutions, and industry partners from all over the world, TII connects an intellectual community and contributes to building an R&D ecosystem that reinforces the status of Abu Dhabi and the UAE as a global hub for innovation.


Web Release
22-05-2025
- Business
- Web Release
Middle East's Leading AI Powerhouse TII Launches Two New AI Models: Falcon Arabic – the First Arabic Model in the Falcon Series & Falcon-H1, a Best-in-Class High-Performance Model
The UAE's Technology Innovation Institute (TII), the applied research arm of Abu Dhabi's Advanced Technology Research Council (ATRC), today unveiled two major AI advancements: Falcon Arabic, the first-ever Arabic language model in the Falcon series – now the best-performing Arabic AI model in the region – and Falcon H1, a new model that redefines performance and portability through a new architectural design. In the small-to-medium size category of AI models (30 to 70 billion parameters), Falcon H1 outperforms comparable offerings from Meta's LLaMA and Alibaba's Qwen, enabling real-world AI on everyday devices and in resource-limited settings. The announcement was made during a keynote address by H.E. Faisal Al Bannai, Advisor to the UAE President and Secretary General of ATRC, at the Make it in the Emirates event. Built on top of Falcon 3-7B (7-billion-parameter), Falcon Arabic is one of the most advanced Arabic AI models developed to date. Trained on a high-quality native (non-translated) Arabic dataset spanning Modern Standard Arabic and regional dialects, it captures the full linguistic diversity of the Arab world. According to the Open Arabic LLM Leaderboard benchmarks, Falcon Arabic outperforms all other regionally available Arabic language models, reinforcing its leadership in sovereign, multilingual AI. It ranks as the best-performing Arabic model in its class, matching the performance of models up to 10 times its size, proving that smart architecture can outperform sheer scale. Separately, the newly launched Falcon H1 model is designed to dramatically expand access to high-performance AI by reducing the computing power and technical expertise traditionally required to run advanced systems. The announcement builds on the success of TII's Falcon 3 series, which ranked among the top global AI models capable of operating on a single graphics processing unit (GPU), a major breakthrough that enabled developers, startups, and institutions without high-end infrastructure to deploy cutting-edge AI, affordably. 'We're proud to finally bring Arabic to Falcon, and prouder still that the best-performing large language model in the Arab world was built in the UAE,' said H.E. Faisal Al Bannai at the Make it in the Emirates event in Abu Dhabi. Commenting on Falcon H1, he said: 'Today, AI leadership is not about scale for the sake of scale. It is about making powerful tools useful, usable, and universal. Falcon-H1 reflects our commitment to delivering AI that works for everyone – not just the few.' Falcon-H1 continues to support European-origin languages and for the first time has scalable capability to support over 100 languages, thanks to a multilingual tokenizer trained on diverse datasets. Smarter, Simpler, and More Inclusive Falcon-H1 was developed to meet the growing global demand for efficient, flexible, and easy-to-use AI systems. Falcon-H1, named 'H' for its hybrid architecture combining the strengths of Transformers and Mamba, enables significantly faster inference speeds and lower memory consumption, while maintaining high performance across a range of benchmarks. 'We approached Falcon-H1 not just as a research milestone but as an engineering challenge: how to deliver exceptional efficiency without compromise,' said Dr. Najwa Aaraj, CEO of TII. 'This model reflects our commitment to building technically rigorous systems with real-world utility. Falcon isn't just a model; it's a foundation that empowers researchers, developers, and innovators, especially in environments where resources are limited but ambitions are not.' The Falcon-H1 family includes models of various sizes: 34B, 7B, 3B, 1.5B, 1.5B-deep, and 500M. These models offer users a wide range of performance-to-efficiency ratios, allowing developers to choose the most appropriate model for their deployment scenarios. While the smaller models enable deployment on constrained edge devices, the flagship 34B model outperforms like-models from Meta's LlaMa and Alibaba's Qwen on complex tasks. 'The Falcon-H1 series demonstrates how new architectures can unlock new opportunities in AI training while showcasing the potential of ultra-compact models,' said Dr. Hakim Hacid, Chief Researcher at the AI and Digital Science Research Center at TII. 'This fundamentally shifts what's possible at the smallest scale, enabling powerful AI on edge devices where privacy, efficiency, and low latency are critical. Our focus has been on reducing complexity without compromising capability.' Each model in the Falcon-H1 family surpasses other models that are twice its size, setting a new standard for performance-to-efficiency ratios. The models additionally excel in mathematics, reasoning, coding, long-context understanding, and multilingual tasks. International Impact Falcon models are already powering real-world applications. In partnership with the Bill & Melinda Gates Foundation, Falcon has supported the development of AgriLLM, a solution that helps farmers make smarter decisions under challenging climate conditions. TII's Falcon ecosystem has been downloaded over 55 million times globally and is widely regarded as the most powerful and consistently high-performing family of open AI models to emerge from the Middle East region. While many AI models focus on narrow consumer use cases, TII has prioritized building foundational models that can be adapted to meet the demanding needs of industry, research, and public good, without compromising on accessibility. These models are designed to be applied across a variety of real-world scenarios, remaining accessible, resource-efficient, and adaptable to different environments. All Falcon models are open source and available on Hugging Face and under the TII Falcon License, an Apache 2.0-based license, which promotes responsible and ethical AI development.


Barnama
22-05-2025
- Business
- Barnama
Middle East's Leading AI Powerhouse TII Launches Two New AI Models: Falcon Arabic - the First Arabic Model in the Falcon Series & Falcon-H1, a Best-in-Class High-Performance Model
ABU DHABI, United Arab Emirates, May 22 (Bernama) -- The UAE's Technology Innovation Institute (TII), the applied research arm of Abu Dhabi's Advanced Technology Research Council (ATRC), today unveiled two major AI advancements: Falcon Arabic, the first-ever Arabic language model in the Falcon series - now the best-performing Arabic AI model in the region - and Falcon-H1, a new model that redefines performance and portability through a new architectural design. In the small-to-medium size category of AI models (30 to 70 billion parameters), Falcon-H1 outperforms comparable offerings from Meta's LlaMA and Alibaba's Qwen, enabling real-world AI on everyday devices and in resource-limited settings. The announcement was made during a keynote address by H.E. Faisal Al Bannai, Advisor to the UAE President and Secretary General of ATRC, at the Make it in the Emirates event. Built on top of Falcon 3-7B (7-billion-parameter), Falcon Arabic is one of the most advanced Arabic AI models developed to date. Trained on a high-quality native (non-translated) Arabic dataset spanning Modern Standard Arabic and regional dialects, it captures the full linguistic diversity of the Arab world. According to the Open Arabic LLM Leaderboard benchmarks, Falcon Arabic outperforms all other regionally available Arabic language models, reinforcing its leadership in sovereign, multilingual AI. It ranks as the best-performing Arabic model in its class, matching the performance of models up to 10 times its size, proving that smart architecture can outperform sheer scale.


Business Wire
21-05-2025
- Business
- Business Wire
Middle East's Leading AI Powerhouse TII Launches Two New AI Models: Falcon Arabic - the First Arabic Model in the Falcon Series & Falcon-H1, a Best-in-Class High-Performance Model
ABU DHABI, United Arab Emirates--(BUSINESS WIRE)--The UAE's Technology Innovation Institute (TII), the applied research arm of Abu Dhabi's Advanced Technology Research Council (ATRC), today unveiled two major AI advancements: Falcon Arabic, the first-ever Arabic language model in the Falcon series - now the best-performing Arabic AI model in the region - and Falcon-H1, a new model that redefines performance and portability through a new architectural design. In the small-to-medium size category of AI models (30 to 70 billion parameters), Falcon-H1 outperforms comparable offerings from Meta's LlaMA and Alibaba's Qwen, enabling real-world AI on everyday devices and in resource-limited settings. The announcement was made during a keynote address by H.E. Faisal Al Bannai, Advisor to the UAE President and Secretary General of ATRC, at the Make it in the Emirates event. Built on top of Falcon 3-7B (7-billion-parameter), Falcon Arabic is one of the most advanced Arabic AI models developed to date. Trained on a high-quality native (non-translated) Arabic dataset spanning Modern Standard Arabic and regional dialects, it captures the full linguistic diversity of the Arab world. According to the Open Arabic LLM Leaderboard benchmarks, Falcon Arabic outperforms all other regionally available Arabic language models, reinforcing its leadership in sovereign, multilingual AI. It ranks as the best-performing Arabic model in its class, matching the performance of models up to 10 times its size, proving that smart architecture can outperform sheer scale. Separately, the newly launched Falcon-H1 model is designed to dramatically expand access to high-performance AI by reducing the computing power and technical expertise traditionally required to run advanced systems. The announcement builds on the success of TII's Falcon 3 series, which ranked among the top global AI models capable of operating on a single graphics processing unit (GPU), a major breakthrough that enabled developers, startups, and institutions without high-end infrastructure to deploy cutting-edge AI, affordably. 'We're proud to finally bring Arabic to Falcon, and prouder still that the best-performing large language model in the Arab world was built in the UAE,' said H.E. Faisal Al Bannai at the Make it in the Emirates event in Abu Dhabi. Commenting on Falcon-H1, he said: 'Today, AI leadership is not about scale for the sake of scale. It is about making powerful tools useful, usable, and universal. Falcon-H1 reflects our commitment to delivering AI that works for everyone – not just the few.' Falcon-H1 continues to support European-origin languages and for the first time has scalable capability to support over 100 languages, thanks to a multilingual tokenizer trained on diverse datasets. Smarter, Simpler, and More Inclusive Falcon-H1 was developed to meet the growing global demand for efficient, flexible, and easy-to-use AI systems. Falcon-H1, named 'H' for its hybrid architecture combining the strengths of Transformers and Mamba, enables significantly faster inference speeds and lower memory consumption, while maintaining high performance across a range of benchmarks. 'We approached Falcon-H1 not just as a research milestone but as an engineering challenge: how to deliver exceptional efficiency without compromise,' said Dr. Najwa Aaraj, CEO of TII. 'This model reflects our commitment to building technically rigorous systems with real-world utility. Falcon isn't just a model; it's a foundation that empowers researchers, developers, and innovators, especially in environments where resources are limited but ambitions are not.' The Falcon-H1 family includes models of various sizes: 34B, 7B, 3B, 1.5B, 1.5B-deep, and 500M. These models offer users a wide range of performance-to-efficiency ratios, allowing developers to choose the most appropriate model for their deployment scenarios. While the smaller models enable deployment on constrained edge devices, the flagship 34B model outperforms like-models from Meta's LlaMA and Alibaba's Qwen on complex tasks. 'The Falcon-H1 series demonstrates how new architectures can unlock new opportunities in AI training while showcasing the potential of ultra-compact models,' said Dr. Hakim Hacid, Chief Researcher at the AI and Digital Science Research Center at TII. 'This fundamentally shifts what's possible at the smallest scale, enabling powerful AI on edge devices where privacy, efficiency, and low latency are critical. Our focus has been on reducing complexity without compromising capability.' Each model in the Falcon-H1 family surpasses other models that are twice its size, setting a new standard for performance-to-efficiency ratios. The models additionally excel in mathematics, reasoning, coding, long-context understanding, and multilingual tasks. International Impact Falcon models are already powering real-world applications. In partnership with the Bill & Melinda Gates Foundation, Falcon has supported the development of AgriLLM, a solution that helps farmers make smarter decisions under challenging climate conditions. TII's Falcon ecosystem has been downloaded over 55 million times globally and is widely regarded as the most powerful and consistently high-performing family of open AI models to emerge from the Middle East region. While many AI models focus on narrow consumer use cases, TII has prioritized building foundational models that can be adapted to meet the demanding needs of industry, research, and public good, without compromising on accessibility. These models are designed to be applied across a variety of real-world scenarios, remaining accessible, resource-efficient, and adaptable to different environments. All Falcon models are open source and available on Hugging Face and under the TII Falcon License, an Apache 2.0-based license, which promotes responsible and ethical AI development.