LLaVA-1.5: Large Language and Vision Assistant
#17: LLMs in 2023, Llama impact grants, Visualization of transformers
LLaVA-1.5: Improved Baselines with Visual Instruction Tuning
LLaVA is a Large Language and Vision Assistant model, similar to GPT-4V(ision)
LLaVA-1.5 is able to achieve better baselines with simple modifications to LLaVA, namely, using CLIP-ViT-L-336px with an MLP projection and adding academic-task-oriented VQA data with simple response formatting prompts.
Video: Large Language Models in 2023
A very good video on “Large Language Models in 2023” by Hyung Won Chung of OpenAI.
He talks about the change in perspectives needed with the change in emergent abilities of LLMs over time, scaling of transformers based models etc.
A pity you did not keep this up.