MiniGPT-4, is an open-source multimodal model similar to the version of GPT-4 that was shown during OpenAI’s presentation. It combines a Visual encoder with an LLM. They used Vicuna which is a fine-tuned version of LLaMA.
In the future, I hope more teams try to add new ideas to their models instead of creating more and more small language models.
Leave a Reply