First Multimodal AI Community Forum - An online AI UK Fringe event
Event details
Description
Event Objectives:
Multimodal AI, which integrates various data modalities such as text, image, sound, and others, is swiftly revolutionising our interaction with technology and data. In our recent Turing Interest Group event (22nd Nov), "The First Multimodal AI Research Sprint", we explored the diverse research states and methodologies in Multimodal AI across six areas and initiated the writing of a perspective paper on multimodal AI. Based on such past activities, this online forum aims to further bring together community members, from researchers to practitioners, to share their latest interdisciplinary perspectives and pioneering work in Multimodal AI. Our goal is to facilitate the exchange of fresh insights and foster connections and research progress within the Multimodal AI community.
Who Should Attend:
We welcome academics, industry professionals, and students engaged in or interested in Multimodal AI from anywhere in the world to join us online. And we encourage self-organised local community gathering.
This Online Forum Will Feature:
- A keynote presentation "LLaVA: A Vision-Language Approach to Computer Vision in the Wild" by Chunyuan Li, Research Lead at ByteDance/TikTok, (Co-)Lead Developer of LLaVA, Former Principal Researcher at Microsoft Research, Redmond
- An introduction to the six areas in our Multimodal AI perspective paper
- Pitches from the community (please express your interest below)
- Open discussions and conclusions
Organisers:
This event is brought to you by the Turing Interest Group on Meta-Learning for Multimodal Data (welcome to sign up) and the Multimodal AI Community (welcome to subscribe to our Google Group) supported by the Centre for Machine Intelligence at the University of Sheffield.
Please visit https://multimodalai.github.io/ for the latest information on this event.