Code release of "Collective Deep Quantization of Efficient Cross-modal Retrieval" (AAAI 17)
-
Updated
Mar 15, 2017 - Python
Code release of "Collective Deep Quantization of Efficient Cross-modal Retrieval" (AAAI 17)
The implementation of AAAI-17 paper "Collective Deep Quantization of Efficient Cross-modal Retrieval"
Cross-modal convolutional neural networks
Implementation of Fast ml-CCA from the ICCV-2015 work "Multi-Label Cross-Modal Retrieval"
Unofficial Implementation of Google Deepmind's paper `Objects that Sound`
Generalized cross-modal NNs; new audiovisual benchmark (IEEE TNNLS 2019)
Code for paper "direct speech-to-image translation"
Preserving Semantic Neighborhoods for Robust Cross-modal Retrieval [ECCV 2020]
Implementation of `Objects that Sound` and `Look, Listen, and Learn` papers by Relja Arandjelovi´c and Andrew Zisserman
Search targeted pedestrians with the text.
An intentionally simple Image to Food cross-modal search. Created by Prithiviraj Damodaran.
Website for Cross Modal Learning and Application workshop - ACM ICMR 2019
Code, dataset and models for our CVPR 2022 publication "Text2Pos"
MMAct: A Large-Scale Dataset for Cross Modal Learning on Human Action Understanding
[ECCV2022] Contrastive Vision-Language Pre-training with Limited Resources
A collection of research on knowledge graphs
Official PyTorch implementation of our CVPR 2022 paper: Beyond a Pre-Trained Object Detector: Cross-Modal Textual and Visual Context for Image Captioning
Code for journal paper "Learning Dual Semantic Relations with Graph Attention for Image-Text Matching", TCSVT, 2020.
Python implementation of cross-modal hashing algorithms
Add a description, image, and links to the cross-modal topic page so that developers can more easily learn about it.
To associate your repository with the cross-modal topic, visit your repo's landing page and select "manage topics."