UAE’s Technology Innovation Institute Revolutionizes AI Language Models With New Architecture
The Falcon Mamba 7B is the no. 1 globally performing open source State Space Language Model (SSLM), as independently verified by Hugging Face
SSLMs have a low memory cost and don’t require additional memory to generate arbitrary long blocks of text
Falcon Mamba 7B also outperforms traditional transformer architecture models such Meta’s Llama 3.1 8B and Mistral’s 7B
New model reflects the innovation and pioneering approach of Abu Dhabi in AI research and development
H.E. Faisal Al Bannai, Secretary General of ATRC and Adviser to the UAE President for Strategic Research and Advanced Technology Affairs, said: “The Falcon Mamba 7B marks TII’s fourth consecutive top-ranked AI model, reinforcing Abu Dhabi as a global hub for AI research and development. This achievement highlights the UAE’s unwavering commitment to innovation.”
For transformer architecture models, Falcon Mamba 7B outperforms Meta’s Llama 3.1 8B, Llama 3 8B, and Mistral’s 7B on the newly introduced benchmarks from HuggingFace. Meanwhile for the other SSLMs, Falcon Mamba 7B beats all other open source models in the old benchmarks and it will be the be first model on HuggingFace’s new tougher benchmark leaderboard.
Dr. Najwa Aaraj, Chief Executive of TII, said: “The Technology Innovation Institute continues to push the boundaries of technology with its Falcon series of AI models. The Falcon Mamba 7B represents true pioneering work and paves the way for future AI innovations that will enhance human capabilities and improve lives.”
State Space models are extremely performant at understanding complex situations that evolve over time, such as a whole book. This is because SSLMs do not require additional memory to digest such large bits of information.
Transformer based models, on the other hand, are very efficient at remembering and using information they have processed earlier in a sequence. This makes them very good at tasks like content generation, however, because they compare every word with every other word, this requires significant computational power.
SSLMs can find applications in various fields such as estimation, forecasting, and control tasks. Similar to the transformer architecture models, they also excel in Natural Language Processing tasks and can be applied to machine translation, text summarization, computer vision, and audio processing.
Dr. Hakim Hacid, Acting Chief Researcher of the TII’s AI Cross-Center Unit, said: “As we introduce the Falcon Mamba 7B, I’m proud of the collaborative ecosystem of TII that nurtured its development. This release represents a significant stride forward, inspiring fresh perspectives and further fueling the quest for intelligent systems. At TII, we’re pushing the boundaries of both SSLM and transformer models to spark further innovation in generative AI.”
Falcon LLMs have been downloaded over 45 million times, proving the outstanding success of the models. Falcon Mamba 7B will be released under TII Falcon License 2.0, the permissive Apache 2.0-based software license which includes an acceptable use policy that promotes the responsible use of AI. More information on the new model can be found at FalconLLM.TII.ae.
*Source: AETOSWire
View source version on businesswire.com: https://www.businesswire.com/news/home/20240812019509/en/
Website: https://www.tii.ae/
Contact
The Technology Innovation Institute
Senior Director of Communications
Jennifer Dewan
Jennifer.dewan@tii.ae
This news is a press release provided by The Technology Innovation Institute. Korea Newswire follows these editorial guidelines.