Finetuning Large Visual Models on Visual Question Answering
-
Updated
May 23, 2024 - Jupyter Notebook
Finetuning Large Visual Models on Visual Question Answering
Open-source evaluation toolkit of large vision-language models (LVLMs), support GPT-4v, Gemini, QwenVLPlus, 40+ HF models, 20+ benchmarks
Davidsonian Scene Graph (DSG) for Text-to-Image Evaluation (ICLR 2024)
An extension of the Planner-Actor-Reporter framework applied to autonomous vehicles in Highway-Env and CARLA.
AIS: Vision, Graphics and AI for Streaming Workshop at CVPR 2024
Visual Question Answering Using CLIP + LSTM
OmniFusion — a multimodal model to communicate using text and images
Multimodal Instruction Tuning for Llama 3
visual question answering prompting recipes for large vision-language models
Chatbot Arena meets multi-modality! Multi-Modality Arena allows you to benchmark vision-language models side-by-side while providing images as inputs. Supports MiniGPT-4, LLaMA-Adapter V2, LLaVA, BLIP-2, and many more!
Official Implementation of WACV 2024 Paper "HIDRO-VQA : High Dynamic Range Oracle for Video Quality Assessment"
This package is a flexible python implementation of the Quantum Approximate Optimization Algorithm /Quantum Alternating Operator ansatz (QAOA) aimed at researchers to readily test the performance of a new ansatz, a new classical optimizers, etc.
Code for the MultipanelVQA benchmark "Muffin or Chihuahua? Challenging Large Vision-Language Models with Multipanel VQA"
How well do the GPT-4V, Gemini Pro Vision, and Claude 3 Opus models perform zero-shot vision tasks on data structures?
This repository contains the data and code of the paper titled "IllusionVQA: A Challenging Optical Illusion Dataset for Vision Language Models"
LLaVA inference with multiple images at once for cross-image analysis.
An implemention of CLIP-ViL Gradcam for VQA tasks
[ICLR'24] Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning
Add a description, image, and links to the vqa topic page so that developers can more easily learn about it.
To associate your repository with the vqa topic, visit your repo's landing page and select "manage topics."