Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

【开源自荐】MoE-LLaVA:北大ChatLaw团队开源多模态稀疏MoE大模型! #405

Open
LinB203 opened this issue Feb 1, 2024 · 0 comments

Comments

@LinB203
Copy link

LinB203 commented Feb 1, 2024

  • 项目名称:

MoE-LLaVA:大型视觉语言模型的混合专家模型

  • 项目地址:

Github: https://github.com/PKU-YuanGroup/MoE-LLaVA
Paper: https://arxiv.org/abs/2401.15947
Demo: https://huggingface.co/spaces/LanguageBind/MoE-LLaVA

  • 项目简介 (100 字以内):

MoE-LLaVA只有3B个稀疏激活参数,表现与LLaVA-1.5-7B在各种视觉理解数据集上相当,并且在物体幻觉基准测试中甚至超越了LLaVA-1.5-13B。通过MoE-LLaVA,我们旨在建立稀疏LVLMs的基准,并为未来研究开发更高效和有效的多模态学习系统提供宝贵的见解。并且MoE-LLaVA团队已经开放了所有的数据、代码和模型。

  • 项目截图 (6张以内):

intro0

intro

framework

imagecli

20240126_205845.mp4

moe-llava

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant