You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The related work reviews existing research to outline the development and current state of the field, enabling readers to quickly grasp established methods, theories, and findings. Here are some suggestions, and I look forward to your feedback.
Including early works applying MoE layers to multimodal representation learning, such as EVE and LIMoE, which use soft routers for data fusion, would enrich the paper. Moreover, MoCLE clusters different instruction sets and distributes them to different experts. MoE-LLaVA also integrate the MoE. Although its performance appears inferior to DeepSeek-VL2, I believe it is an essential baseline.
Finally, thank you for your contributions to VLM. I believe applying MoE to VLM is a milestone.
The text was updated successfully, but these errors were encountered:
The related work reviews existing research to outline the development and current state of the field, enabling readers to quickly grasp established methods, theories, and findings. Here are some suggestions, and I look forward to your feedback.
Including early works applying MoE layers to multimodal representation learning, such as EVE and LIMoE, which use soft routers for data fusion, would enrich the paper. Moreover, MoCLE clusters different instruction sets and distributes them to different experts. MoE-LLaVA also integrate the MoE. Although its performance appears inferior to DeepSeek-VL2, I believe it is an essential baseline.
Finally, thank you for your contributions to VLM. I believe applying MoE to VLM is a milestone.
The text was updated successfully, but these errors were encountered: