We are contributing 🎉🎉🎉Video-LLaVA #846
LinB203
started this conversation in
Show and tell
Replies: 1 comment
-
Congrats to the great achievement. It is definitely an important direction to equip large multimodal models with video understanding capability. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello, esteemed LLaVA developer, thank you for contributing such robust code and data to the community.
We have extended LLaVA to Video-LLaVA to achieve advanced performance on MSRVTT,MSVD,TGIF,ACTIVITYNET.
Thank you again for your contributions to the large visual-language model!
Beta Was this translation helpful? Give feedback.
All reactions