Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Refine related work. #2

Open
LinB203 opened this issue Dec 14, 2024 · 0 comments
Open

Refine related work. #2

LinB203 opened this issue Dec 14, 2024 · 0 comments

Comments

@LinB203
Copy link

LinB203 commented Dec 14, 2024

The related work reviews existing research to outline the development and current state of the field, enabling readers to quickly grasp established methods, theories, and findings. Here are some suggestions, and I look forward to your feedback.

Including early works applying MoE layers to multimodal representation learning, such as EVE and LIMoE, which use soft routers for data fusion, would enrich the paper. Moreover, MoCLE clusters different instruction sets and distributes them to different experts. MoE-LLaVA also integrate the MoE. Although its performance appears inferior to DeepSeek-VL2, I believe it is an essential baseline.

Finally, thank you for your contributions to VLM. I believe applying MoE to VLM is a milestone.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant