MiniCPM-V is a series of end-side multimodal LLMs (MLLMs) designed for vision-language understanding. The models take image and text as inputs and provide high-quality text outputs. Since February 2024, we have released 4 versions of the model, aiming to achieve strong performance and efficient deployment. The most notable models in this series currently include:
• MiniCPM-Llama3-V 2.5: 🔥🔥🔥 The latest and most capable model in the MiniCPM-V series. With a total of 8B parameters, the model surpasses proprietary models such as GPT-4V-1106, Gemini Pro, Qwen-VL-Max and Claude 3 in overall performance. Equipped with the enhanced OCR and instruction-following capability, the model can also support multimodal conversation for over 30 languages including English, Chinese, French, Spanish, German etc. With help of quantization, compilation optimizations, and several efficient inference techniques on CPUs and NPUs, MiniCPM-Llama3-V 2.5 can be efficiently deployed on end-side devices.
Relevant Links:
https://github.com/OpenBMB/MiniCPM-V
https://huggingface.co/openbmb/MiniCP...
https://huggingface.co/spaces/openbmb...
If you like to support me financially, It is totally optional and voluntary. Buy me a coffee here: https://www.buymeacoffee.com/rithesh
If you like such content please subscribe to the channel here:
https://www.youtube.com/c/RitheshSree...
Watch video A GPT-4V Level Multimodal LLM on Your Phone ??? MiniCPM-Llama3-V-2_5 online without registration, duration hours minute second in high quality. This video was added by user Rithesh Sreenivasan 06 June 2024, don't forget to share it with your friends and acquaintances, it has been viewed on our site 28 once and liked it people.