A collection of research on knowledge graphs
-
Updated
Oct 7, 2022 - JavaScript
A collection of research on knowledge graphs
🪩 Create Disco Diffusion artworks in one line
Represent, send, store and search multimodal data
PyTorch source code for "Stacked Cross Attention for Image-Text Matching" (ECCV 2018)
Analyze the unstructured data with Towhee, such as reverse image search, reverse video search, audio classification, question and answer systems, molecular search, etc.
A curated list of different papers and datasets in various areas of audio-visual processing
Effective prompting for Large Multimodal Models like GPT-4 Vision, LLaVA or CogVLM. 🔥
The implementation of AAAI-17 paper "Collective Deep Quantization of Efficient Cross-modal Retrieval"
Remote Sensing Sar-Optical Land-use Classfication Pytorch Pytorch高分辨率遥感语义分割/地物分割/地物分类
Weakly Supervised 3D Object Detection from Point Clouds (VS3D), ACM MM 2020
Unofficial Implementation of Google Deepmind's paper `Objects that Sound`
[CVPR 2023] Referring Image Matting
Code for journal paper "Learning Dual Semantic Relations with Graph Attention for Image-Text Matching", TCSVT, 2020.
[NAACL 2022]Mobile Text-to-Image search powered by multimodal semantic representation models(e.g., OpenAI's CLIP)
Official PyTorch implementation of our CVPR 2022 paper: Beyond a Pre-Trained Object Detector: Cross-Modal Textual and Visual Context for Image Captioning
Cross-modal convolutional neural networks
[Paper][AAAI 2023] DUET: Cross-modal Semantic Grounding for Contrastive Zero-shot Learning
Unleash the Potential of Image Branch for Cross-modal 3D Object Detection [NeurIPS2023]
Code for paper "direct speech-to-image translation"
Preserving Semantic Neighborhoods for Robust Cross-modal Retrieval [ECCV 2020]
Add a description, image, and links to the cross-modal topic page so that developers can more easily learn about it.
To associate your repository with the cross-modal topic, visit your repo's landing page and select "manage topics."