Post

Conversation

Re Llama3V: Firstly, we want to apologize to the original authors of MiniCPM.and I posted Llama3V with @mustafaaljadery. Mustafa wrote the code for the project. Aksh and I were both excited about multimodal models and liked the architectural extensions on top of Idefics and Siglip that he described to us. Our role here was to help him promote the model on medium and twitter. We looked at recent papers to validate the novelty of the work but we were not informed of or made aware of any of the previous work by. After seeing the twitter posts about this topic yesterday, we asked Mustafa about proof of originality for Llama3V and asked for the training code but we haven’t seen any response so far. We were waiting for Mustafa to take the lead but instead we are releasing our own statement. We apologize to the authors of miniCBM for any inconvenience that we caused for not doing the full diligence to verify and peer review the novelty of this work. Going forward, we will be cautious and diligent, and we sincerely thank the community for bringing this to our attention. We've taken all references to Llama3V down and we apologize once again for the inconvenience we may have caused. - Siddharth and Aksh
Quote
PrimerYang
@yangzhizheng1
Shocked! Llama3-V project from a Stanford team plagiarized a lot from MiniCPM-Llama3-V 2.5! its code is a reformatting of MiniCPM-Llama3-V 2.5, and the model's behavior is highly similar to a noised version of MiniCPM-Llama3-V 2.5 checkpoint. Evidence: github.com/OpenBMB/MiniCP
Image
Image
Image

New to X?

Sign up now to get your own personalized timeline!
Create account
By signing up, you agree to the Terms of Service and Privacy Policy, including Cookie Use.
CS @ Stanford. Building. Prev,,
Something went wrong. Try reloading.