[EMNLP 2023 Demo] Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding
#大语言模型#(AAAI 2024) BLIVA: A Simple Multimodal LLM for Better Handling of Text-rich Visual Questions
Automate Fashion Image Captioning using BLIP-2. Automatic generating descriptions of clothes on shopping websites, which can help customers without fashion knowledge to better understand the features ...
#计算机科学#Implementation of Qformer from BLIP2 in Zeta Lego blocks.
Official implementation and dataset for the NAACL 2024 paper "ComCLIP: Training-Free Compositional Image and Text Matching"
[ACM MM 2024] Improving Composed Image Retrieval via Contrastive Learning with Scaling Positives and Negatives
The Multimodal Model for Vietnamese Visual Question Answering (ViVQA)
CLIP Interrogator, fully in HuggingFace Transformers 🤗, with LongCLIP & CLIP's own words and / or *your* own words!
Modifying LAVIS' BLIP2 Q-former with models pretrained on Japanese datasets.
#大语言模型#This repository is for profiling, extracting, visualizing and reusing generative AI weights to hopefully build more accurate AI models and audit/scan weights at rest to identify knowledge domains for ...
Caption images across your datasets with state of the art models from Hugging Face and Replicate!
#数据仓库#Annotations on a Budget: Leveraging Geo-Data Similarity to Balance Model Performance and Annotation Cost
Finetuning Large Visual Models on Visual Question Answering
caption generator using lavis and argostranslate
In this we explore into visual Question Answering Using Gemini LLM and image was in URL or any other extension
#大语言模型#An offline AI-powered video analysis tool with object detection (YOLO), image captioning (BLIP), speech transcription (Whisper), audio event detection (PANNs), and AI-generated summaries (LLMs via Oll...
Too lazy to organize my desktop, make gpt + BLIP-2 do it /s