#
vqa
Here are 104 public repositories matching this topic...
Bottom-up attention model for image captioning and VQA, based on Faster R-CNN and Visual Genome
caffe
vqa
faster-rcnn
image-captioning
captioning-images
mscoco
mscoco-dataset
visual-question-answering
-
Updated
Mar 3, 2020 - Jupyter Notebook
An efficient PyTorch implementation of the winning entry of the 2017 VQA Challenge.
-
Updated
Mar 4, 2020 - Python
Open
Batch first in LSTM
DenisDsh
commented
Aug 2, 2018
In the LSTM documentation is specified that we should feed to the RNN inputs of shape (seq_len, batch, input_size) however it seems to me that we are feeding inputs with shape (batch, seq_len, input_size).
Therefore I believe that the parameter batch_first should be set True.
Implementation for the paper "Compositional Attention Networks for Machine Reasoning" (Hudson and Manning, ICLR 2018)
tensorflow
vqa
question-answering
attention
clevr
machine-reasoning
compositional-attention-networks
-
Updated
Feb 26, 2020 - Python
PyTorch implementation of "Transparency by Design: Closing the Gap Between Performance and Interpretability in Visual Reasoning"
-
Updated
Feb 28, 2020 - Jupyter Notebook
PyTorch implementation for the Neuro-Symbolic Concept Learner (NS-CL).
-
Updated
Mar 2, 2020 - Python
A curated list of Visual Question Answering(VQA)(Image/Video Question Answering),Visual Question Generation ,Visual Dialog ,Visual Commonsense Reasoning and related area.
-
Updated
Mar 4, 2020
-
Updated
Feb 27, 2020 - Python
Strong baseline for visual question answering
-
Updated
Feb 27, 2020 - Python
A lightweight, scalable, and general framework for visual question answering (VQA) research
-
Updated
Mar 4, 2020 - Python
Tensorflow Implementation of Deeper LSTM+ normalized CNN for Visual Question Answering
-
Updated
Mar 2, 2020 - Python
Improved Fusion of Visual and Language Representations by Dense Symmetric Co-Attention for Visual Question Answering
-
Updated
Feb 2, 2020 - Python
读过的CV方向的一些论文,图像生成文字、弱监督分割等
natural-language-processing
computer-vision
captions
vqa
cvpr
iccv
miccai
eccv
image2text
scene-text-detection-recognition
weakly-supervised-segmentation
-
Updated
Feb 6, 2020
This project is out of date, I don't remember the details inside...
-
Updated
Feb 23, 2020 - Python
Hadamard Product for Low-rank Bilinear Pooling
-
Updated
Jan 15, 2020 - Lua
Code for ICML 2019 paper "Probabilistic Neural-symbolic Models for Interpretable Visual Question Answering" [long-oral]
-
Updated
Feb 22, 2020 - Python
CloudCV Visual Question Answering Demo
-
Updated
Feb 26, 2020 - Lua
CVPR'17 Spotlight: What’s in a Question: Using Visual Questions as a Form of Supervision
-
Updated
Aug 25, 2019 - Lua
Code for paper title "Learning Semantic Sentence Embeddings using Pair-wise Discriminator" COLING-2018
acl
torch
vqa
question-answering
rouge
emnlp
questions-and-answers
paraphrase
bleu
paraphrase-identification
visual-question-answering
paraphrase-generation
question-generation
coling2018
acl-2018
emnlp2018
emnlp2017
vqg
visual-questions-generation
question-parapharse
-
Updated
Mar 2, 2020 - Jupyter Notebook
Pytorch implementation of NIPS 2017 paper "Modulating early visual processing by language"
-
Updated
Feb 8, 2020 - Python
Bottom-up features extractor implemented in PyTorch.
-
Updated
Mar 1, 2020 - Python
RUBi : Reducing Unimodal Biases for Visual Question Answering
-
Updated
Feb 24, 2020 - Python
Co-attending Regions and Detections for VQA.
caffe
torch
vqa
faster-rcnn
object-detection
multi-modal
attention-mechanism
aaai
multi-gpu
visual-question-answering
-
Updated
Jan 24, 2020 - MATLAB
PyTorch VQA implementation that achieved top performances in the (ECCV18) VizWiz Grand Challenge: Answering Visual Questions from Blind People
-
Updated
Feb 26, 2020 - Jupyter Notebook
An VideoQA dataset based on the videos from ActivityNet
-
Updated
Jan 21, 2020 - Python
TensorFlow implementation of the CNN-LSTM, Relation Network and text-only baselines for the paper "FigureQA: An Annotated Figure Dataset for Visual Reasoning"
microsoft
deep-learning
tensorflow
python3
neural-networks
vqa
relation-network
relational-reasoning
visual-question-answering
figure-analysis
-
Updated
Nov 14, 2019 - Python
Code for NeurIPS 2019 paper ``Self-Critical Reasoning for Robust Visual Question Answering''
-
Updated
Feb 3, 2020 - Python
Improve this page
Add a description, image, and links to the vqa topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the vqa topic, visit your repo's landing page and select "manage topics."
This has been a long requested feature internally. With current architecture and SampleList, this should be really easy to tackle. ParlAI datasets all have build function. For #163, this should be looked into first.