- π¨βπ¦° Iβm currently a Master of Science candidate of Peking University (PKU).
- π¦ Before that, I received the Honours Bachelor, Huazhong University of Science and Technology (HUST).
- β€οΈβπ₯ Now, I am intersted in Multi-modal Learning especially MLLM.
- π₯ In 2023 summer, I take part in OSPP(Open Source Promotion Plan) Summer Camp , with the honor of contributing for MMPretrain to build prompt-based classifier.
- π₯ 2023.11 - 2024.5:
MPP-Qwen-Next
is released! All training is conducted on 3090/4090 GPUs. To prevent poverty (24GB of VRAM) from limiting imagination, I implemented an MLLM version based on deepspeed Pipeline Parallel. The Repo supports {video/image/multi-image} {single/multi-turn} conversations. Let's have a try! . - π₯ 2024.9: We release
ChartMoE
, a multimodal large language model with Mixture-of-Expert connector, for advanced chart 1)understanding, 2)replot, 3)editing, 4)highlighting and 5)transformation. - π₯π₯π₯ 2024.10: I am really fortunate to be involved in the development of
Aria
.Aria
is a Naive Multimodal MoE model, with best-in-class performance across multimodal, language, and coding tasks!