Middle East AI News

Share this post

MBZUAI releases first LLM library

middleeastainews.substack.com

MBZUAI releases first LLM library

The Oryx Library contains three new domain-specific AI models

Carrington Malin
May 21, 2023
Share
Article voiceover
1ร—
0:00
-3:20
Audio playback is not supported on your browser. Please upgrade.
(Image credit: Carrington Malin via Midjourney)

MBZUAI (Mohamed bin Zayed University of Artificial Intelligence) has publicly released its first library of large-language models (LLMs) named the ๐—ข๐—ฟ๐˜†๐˜… ๐—Ÿ๐—ถ๐—ฏ๐—ฟ๐—ฎ๐—ฟ๐˜† via GitHub, to advance domain-specific applications. ClimateGPT, Video-ChatGPT and XrayGPT follow a number of LLM releases from Abu Dhabi-based R&D labs this year, including projects from Technology Innovation Institute (TII) and IIAI (Inception Institute of Artificial Intelligence)

The Oryx Library consists of projects and demos for large vision-language models developed at MBZUAI. The university's broad goal is to advance LLMs for multi-modal and domain-specific dialogues.

Three Oryx Library projects have just been released:

๐Ÿ’ป ๐—–๐—น๐—ถ๐—บ๐—ฎ๐˜๐—ฒ๐—š๐—ฃ๐—ง [English & Arabic] - a specialised LLM developed over the Vicuna framework for conversations related to Climate Change and Sustainability, in both English and Arabic languages. The goal is to develop an educational resource to support climate-change related conversations and help inform policymakers. ClimateGPT marks the first release of a large Arabic dataset (>500k samples) dedicated to climate change and sustainability. Codebase & demo available.

  • Watch the ClimateGPT results video (Eng)

  • Watch the ClimateGPT results video (Arabic)

๐Ÿ’ป ๐—ฉ๐—ถ๐—ฑ๐—ฒ๐—ผ-๐—–๐—ต๐—ฎ๐˜๐—š๐—ฃ๐—ง - a conversational engine for videos, based on a dedicated video-encoder and large language model (LLM) aligned using a simple linear projection, enabling video understanding and conversation about videos. The project includes the first set of high-quality 86k instruction data obtained specifically for video. Codebase, demo and a short results video available.

  • Watch the Video-ChatGPT results video.

๐Ÿ’ป ๐—ซ๐—ฟ๐—ฎ๐˜†๐—š๐—ฃ๐—ง - a model for automated analysis of chest radiographs based on the given X-ray images. Frozen medical vision encoder (MedCLIP) is aligned with a fine-tuned LLM (Vicuna) using a simple linear transformation.ย The LLM is fine-tuned on medical data (100k real conversations between patients and doctors) and further on ~30k radiology conversations to acquire domain-specific features. XrayGPT generates interactive and clean summaries (~217k) from free-text radiology reports. Codebase, demo and a short results video available.

  • Watch the XrayGPT results video.

The UAE has moved quickly to embrace the potential of LLMs. Earlier this month, the Artificial Intelligence, Digital Economy and Remote Work Applications Officeย published a guide for government workers using generative AI apps.

  • Read the full MBZUAI blogpost:
    https://lnkd.in/dknzJAnC

  • Check out the Oryx Library GitHub repositories:
    https://lnkd.in/daGnCgWQ

  • Read my article on TII's Falcon LLM:
    https://lnkd.in/e7kzZNrj

Share
Previous
Next
Comments
Top
New
Community

No posts

Ready for more?

ยฉ 2023 Carrington Malin
Privacy โˆ™ Terms โˆ™ Collection notice
Start WritingGet the app
Substackย is the home for great writing