#
vqa
Here are 157 public repositories matching this topic...
Bottom-up attention model for image captioning and VQA, based on Faster R-CNN and Visual Genome
caffe
vqa
faster-rcnn
image-captioning
captioning-images
mscoco
mscoco-dataset
visual-question-answering
-
Updated
Feb 14, 2021 - Jupyter Notebook
An efficient PyTorch implementation of the winning entry of the 2017 VQA Challenge.
-
Updated
Sep 4, 2019 - Python
Visual Question Answering in Pytorch
-
Updated
Dec 11, 2019 - Python
Oscar and VinVL
-
Updated
May 24, 2021 - Python
Implementation for the paper "Compositional Attention Networks for Machine Reasoning" (Hudson and Manning, ICLR 2018)
tensorflow
vqa
question-answering
attention
clevr
machine-reasoning
compositional-attention-networks
-
Updated
Jul 10, 2021 - Python
[CVPR 2021 Best Student Paper Honorable Mention, Oral] Official PyTorch code for ClipBERT, an efficient framework for end-to-end learning on image-text and video-text tasks.
-
Updated
Jul 20, 2021 - Python
PyTorch implementation of "Transparency by Design: Closing the Gap Between Performance and Interpretability in Visual Reasoning"
-
Updated
Aug 21, 2018 - Jupyter Notebook
A curated list of Visual Question Answering(VQA)(Image/Video Question Answering),Visual Question Generation ,Visual Dialog ,Visual Commonsense Reasoning and related area.
-
Updated
Jul 26, 2021
PyTorch implementation for the Neuro-Symbolic Concept Learner (NS-CL).
-
Updated
Oct 24, 2020 - Python
A lightweight, scalable, and general framework for visual question answering research
-
Updated
Mar 10, 2021 - Python
Strong baseline for visual question answering
-
Updated
May 26, 2021 - Python
-
Updated
Jul 30, 2019 - Python
读过的CV方向的一些论文,图像生成文字、弱监督分割等
natural-language-processing
computer-vision
captions
vqa
cvpr
iccv
miccai
eccv
image2text
scene-text-detection-recognition
weakly-supervised-segmentation
-
Updated
May 16, 2020
[ICCV 2021- Oral] Official PyTorch implementation for Generic Attention-model Explainability for Interpreting Bi-Modal and Encoder-Decoder Transformers, a novel method to visualize any Transformer-based network. Including examples for DETR, VQA.
visualization
transformers
transformer
vqa
clip
interpretability
explainable-ai
explainability
detr
lxmert
visualbert
-
Updated
Jul 22, 2021 - Jupyter Notebook
Tensorflow Implementation of Deeper LSTM+ normalized CNN for Visual Question Answering
-
Updated
Apr 27, 2017 - Python
Improved Fusion of Visual and Language Representations by Dense Symmetric Co-Attention for Visual Question Answering
-
Updated
Oct 14, 2019 - Python
Implementation for the paper "Hierarchical Conditional Relation Networks for Video Question Answering" (Le et al., CVPR 2020, Oral)
-
Updated
May 6, 2020 - Python
This project is out of date, I don't remember the details inside...
-
Updated
Dec 2, 2017 - Python
[IEEE TIP'2021] "UGC-VQA: Benchmarking Blind Video Quality Assessment for User Generated Content", Zhengzhong Tu, Yilin Wang, Neil Birkbeck, Balu Adsumilli, Alan C. Bovik
evaluation
dataset
feature-extraction
vqa
user-generated-content
iqa
image-quality-assessment
video-quality-assessment
bvqa-model
-
Updated
Jun 7, 2021 - MATLAB
CloudCV Visual Question Answering Demo
-
Updated
Jun 10, 2021 - Lua
Hadamard Product for Low-rank Bilinear Pooling
-
Updated
Nov 6, 2017 - Lua
Bottom-up features extractor implemented in PyTorch.
-
Updated
Dec 5, 2019 - Python
Code for ICML 2019 paper "Probabilistic Neural-symbolic Models for Interpretable Visual Question Answering" [long-oral]
-
Updated
Mar 10, 2020 - Python
[CVPR 2021] Counterfactual VQA: A Cause-Effect Look at Language Bias
-
Updated
Jul 2, 2021 - Python
Counterfactual Samples Synthesizing for Robust VQA
-
Updated
Jul 8, 2020 - Python
RUBi : Reducing Unimodal Biases for Visual Question Answering
-
Updated
Mar 29, 2021 - Python
Improve this page
Add a description, image, and links to the vqa topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the vqa topic, visit your repo's landing page and select "manage topics."
File "/home/ubuntu/vqa/GMN/mmf/mmf/datasets/builders/visual_genome/dataset.py", line 44, in init
scene_graph_file = self._get_absolute_path(scene_graph_file)
AttributeError: 'VisualGenomeDataset' object has no attribute '_get_absolute_path'
Command that i run in shell
CUDA_VISIBLE_DEVICES="0" mmf_run config=projects/gmn/configs/visual_genome/defaults.yaml model=gm