V3ALab
V3ALab
Home
Projects
Publications
People
Job
Contact
Chunhua Shen
Latest
REVERIE: Remote Embodied Visual Referring Expression in Real Indoor Environments
Soft Expert Reward Learning for Vision-and-Language Navigation
FVQA: Fact-based visual question answering
Medical Data Inquiry Using a Question Answering Model
Heritage Image Annotation via Collective Knowledge
Multi-Label Image Classification with Regional Latent Semantic Dependencies
Mind Your Neighbours: Image Annotation with Metadata Neighbourhood Graph Co-Attention Networks
Image Captioning and Visual Question Answering Based on Attributes and Their Related External Knowledge
Visual Question Answering: A Survey of Models and Datasets
Are You Talking to Me? Reasoned Visual Dialog Generation through Adversarial Learning
Asking the Difficult Questions: Goal-Oriented Visual Question Generation via Intermediate Rewards
Parallel Attention: A Unified Framework for Visual Object Discovery through Dialogs and Queries
Visual Question Answering with Memory-Augmented Networks
HCVRD: a benchmark for large-scale Human-Centered Visual Relationship Detection
Kill Two Birds With One Stone: Weakly-Supervised Neural Network for Image Annotation and Tag Refinement
Explicit Knowledge-based Reasoning for Visual Question Answering
The VQA-Machine: Learning How to Use Existing Vision Algorithms to Answer New Questions
Ask Me Anything: Free-form Visual Question Answering Based on Knowledge from External Sources
What Value Do Explicit High Level Concepts Have in Vision to Language Problems?
Cite
×