|
15 | 15 |
|
16 | 16 | <font size=7><div align='center' > [[📽 VITA-1.5 Demo Show! Here We Go! 🔥](https://youtu.be/tyi6SVFT5mM?si=fkMQCrwa5fVnmEe7)] </div></font>
|
17 | 17 | <font size=7><div align='center' > VITA-1.5 spports both **English** and **Chinese**.🌟 </div></font>
|
18 |
| -You can experience our [🤖 Basic Demo](https://modelscope.cn/studios/modelscope/VITA1.5_demo) on ModelScope directly. The Real-Time Interactive Demo needs to be configured according to the [instructions](#-real-time-interactive-demo). |
| 18 | +You can experience our [Basic Demo](https://modelscope.cn/studios/modelscope/VITA1.5_demo) on ModelScope directly. The Real-Time Interactive Demo needs to be configured according to the [instructions](#-real-time-interactive-demo). |
19 | 19 |
|
20 | 20 | ## 🔥 News
|
| 21 | +* **`2025.01.17`** 🌟 ModelScope has supported VITA-1.5! You could try our [Basic Demo](https://modelscope.cn/studios/modelscope/VITA1.5_demo) on it! |
21 | 22 | * **`2025.01.06`** 🌟 [VLMEvalKit](https://github.com/open-compass/VLMEvalKit) of OpenCompass has supported our both VITA-1.5 and VITA-1.0 models!
|
22 | 23 | * **`2025.01.06`** 🌟 The [technical report](https://huggingface.co/VITA-MLLM) of VITA-1.5 has been released!
|
23 | 24 | * **`2024.12.20`** 🌟 We are excited to introduce the **VITA-1.5**, a more powerful and more real-time version!
|
@@ -344,15 +345,15 @@ If you find our work helpful for your research, please consider citing our work.
|
344 | 345 |
|
345 | 346 | ```bibtex
|
346 | 347 | @article{fu2025vita,
|
347 |
| - title={VITA-1.5: Towards gpt-4o level real-time vision and speech interaction}, |
348 |
| - author={Fu, Chaoyou and Lin, Haojia and Wang, Xiong and Zhang, Yi-Fan and Shen, Yunhang and Liu, Xiaoyu and Li, Yangze and Long, Zuwei and Gao, Heting and others}, |
| 348 | + title={VITA-1.5: Towards GPT-4o Level Real-Time Vision and Speech Interaction}, |
| 349 | + author={Fu, Chaoyou and Lin, Haojia and Wang, Xiong and Zhang, Yi-Fan and Shen, Yunhang and Liu, Xiaoyu and Li, Yangze and Long, Zuwei and Gao, Heting and Li, Ke and others}, |
349 | 350 | journal={arXiv preprint arXiv:2501.01957},
|
350 | 351 | year={2025}
|
351 | 352 | }
|
352 | 353 |
|
353 | 354 | @article{fu2024vita,
|
354 | 355 | title={Vita: Towards open-source interactive omni multimodal llm},
|
355 |
| - author={Fu, Chaoyou and Lin, Haojia and Long, Zuwei and Shen, Yunhang and Zhao, Meng and Zhang, Yifan and Wang, Xiong and Yin, Di and Ma, Long and Zheng, Xiawu and others}, |
| 356 | + author={Fu, Chaoyou and Lin, Haojia and Long, Zuwei and Shen, Yunhang and Zhao, Meng and Zhang, Yifan and Dong, Shaoqi and Wang, Xiong and Yin, Di and Ma, Long and others}, |
356 | 357 | journal={arXiv preprint arXiv:2408.05211},
|
357 | 358 | year={2024}
|
358 | 359 | }
|
|
0 commit comments