Building Next-Gen Multimodal Foundation Models for General-Purpose Assistants

LLaVA is an open-source project, collaborating with research community to advance the state-of-the-art in AI. LLaVA represents the first end-to-end trained large multimodal model (LMM) that achieves impressive chat capabilities mimicking spirits of the multimodal GPT-4. The LLaVA family continues growing to support more modalities, capabilities, applications and beyond.

人员

Open research collaboration across universities in the research community and multiple Microsoft team, pushing the SoTA in new capabilities scale and applications etc.

Hao Cheng的肖像

Hao Cheng

Principal Researcher

Michel Galley的肖像

Michel Galley

Senior Principal Researcher

Jianfeng Gao的肖像

Jianfeng Gao

Distinguished Scientist & Vice President

Yong Jae Lee的肖像

Yong Jae Lee

Associate Professor

University of Wisconsin-Madison

Lars Liden的肖像

Lars Liden

Principal Research Software Engineer Manager

Haotian Liu的肖像

Haotian Liu

Ph.D. student

University of Wisconsin-Madison

Xiaodong Liu的肖像

Xiaodong Liu

Senior Principal Researcher

Yadong Lu的肖像

Yadong Lu

Researcher

Microsoft Azure AI

Matt Mazzola的肖像

Matt Mazzola

Senior Research Software Engineer

Tristan Naumann的肖像

Tristan Naumann

Principal Researcher

Hoifung Poon的肖像

Hoifung Poon

General Manager, Health Futures

Yelong Shen的肖像

Yelong Shen

Principal Researcher

Microsoft Azure AI

Swadheen Shukla的肖像

Swadheen Shukla

Principal Program Manager

Irina Spiridonova的肖像

Irina Spiridonova

Senior Software Engineer

Andrea Tupini的肖像

Andrea Tupini

Research Software Engineer

Naoto Usuyama的肖像

Naoto Usuyama

Principal Researcher

Cliff Wong的肖像

Cliff Wong

Principal Data Scientist

Jianwei Yang的肖像

Jianwei Yang

Principal Researcher

Sheng Zhang的肖像

Sheng Zhang

Principal Researcher