


default search action
ViGIL@NeurIPS 2019: Vancouver, Canada
- Visually Grounded Interaction and Language (ViGIL), NeurIPS 2019 Workshop, Vancouver, Canada, December 13, 2019. 2019

- Khanh Nguyen, Hal Daumé III:

Help, Anna! Visual Navigation with Natural Multimodal Assistance via Retrospective Curiosity-Encouraging Imitation Learning. - Alexander Kuhnle, Ann A. Copestake:

What is needed for simple spatial language capabilities in VQA? - Tsu-Jui Fu, Yuta Tsuboi, Sosuke Kobayashi, Yuta Kikuchi:

Learning from Observation-Only Demonstration for Task-Oriented Language Grounding via Self-Examination. - Jiannan Xiang, Xin Wang, William Yang Wang:

Not All Actions Are Equal: Learning to Stop in Language-Grounded Urban Navigation. - Jialin Wu, Raymond J. Mooney:

Hidden State Guidance: Improving Image Captioning Using an Image Conditioned Autoencoder. - Nikhil Krishnaswamy, James Pustejovsky:

Situated Grounding Facilitates Multimodal Concept Learning for AI. - Catalina Cangea, Eugene Belilovsky, Pietro Liò, Aaron C. Courville:

VideoNavQA: Bridging the Gap between Visual and Embodied Question Answering. - Vasu Sharma, Ankita Kalra, Louis-Philippe Morency:

Induced Attention Invariance: Defending VQA Models against Adversarial Attacks. - Xin Wang, Vihan Jain, Eugene Ie, William Yang Wang, Zornitsa Kozareva, Sujith Ravi:

Natural Language Grounded Multitask Navigation. - Farley Lai, Ning Xie, Derek Doran, Asim Kadav:

Contextual Grounding of Natural Language Entities in Images. - Olga Kovaleva, Chaitanya Shivade, Satyananda Kashyap, Karina Kanjaria, Adam Coy, Deddeh Ballah, Yufan Guo, Joy T. Wu, Alexandros Karargyris, David Beymer, Anna Rumshisky, Vandana V. Mukherjee:

Visual Dialog for Radiology: Data Curation and FirstSteps. - Thomas M. Sutter, Imant Daunhawer, Julia E. Vogt:

Multimodal Generative Learning Utilizing Jensen-Shannon-Divergence. - Guan-Lin Chao, Abhinav Rastogi, Semih Yavuz, Dilek Hakkani-Tür, Jindong Chen, Ian R. Lane:

Learning Question-Guided Video Representation for Multi-Turn Video Question Answering. - Angeliki Lazaridou, Anna Potapenko, Olivier Tieleman:

Structural and functional learning for learning language use. - Olivier Tieleman, Angeliki Lazaridou, Shibl Mourad, Charles Blundell, Doina Precup:

Community size effect in artificial learning systems. - Harm de Vries, Dzmitry Bahdanau, Shikhar Murty, Aaron C. Courville, Philippe Beaudoin:

CLOSURE: Assessing Systematic Generalization of CLEVR Models. - Chihiro Fujiyama, Ichiro Kobayashi:

A Comprehensive Analysis of Semantic Compositionality in Text-to-Image Generation. - Alba Maria Hererra-Palacio, Carles Ventura, Carina Silberer, Ionut-Teodor Sorodoc, Gemma Boleda, Xavier Giró-i-Nieto:

Recurrent Instance Segmentation using Sequences of Referring Expressions. - T. S. Jayram, Vincent Albouy, Tomasz Kornuta, Emre Sevgen, Ahmet S. Ozcan:

Visually Grounded Video Reasoning in Selective Attention Memory. - Jean-Benoit Delbrouck:

Modulated Self-attention Convolutional Network for VQA. - Gabriel Ilharco, Vihan Jain, Alexander Ku, Eugene Ie, Jason Baldridge:

General Evaluation for Instruction Conditioned Navigation using Dynamic Time Warping. - Jingxiang Lin, Unnat Jain, Alexander G. Schwing:

A Simple Baseline for Visual Commonsense Reasoning. - Nicolas Lair, Cédric Colas, Rémy Portelas, Jean-Michel Dussoux, Peter F. Dominey, Pierre-Yves Oudeyer:

Language Grounding through Social Interactions and Curiosity-Driven Multi-Goal Learning. - Yen-Ling Kuo, Boris Katz, Andrei Barbu:

Deep compositional robotic planners that follow natural language commands. ViGIL@NeurIPS 2019 - Jean-Benoit Delbrouck:

Can adversarial training learn image captioning ? - Shachi H. Kumar, Eda Okur, Saurav Sahay, Jonathan Huang, Lama Nachman:

Leveraging Topics and Audio Features with Multimodal Attention for Audio Visual Scene-Aware Dialog. - Douwe Kiela, Suvrat Bhooshan, Hamed Firooz, Davide Testuggine:

Supervised Multimodal Bitransformers for Classifying Images and Text. - Jesse Mu, Percy Liang, Noah D. Goodman:

Shaping Visual Representations with Language for Few-shot Classification. - Geoffrey Cideron, Mathieu Seurin, Florian Strub, Olivier Pietquin:

Self-Educated Language Agent with Hindsight Experience Replay for Instruction Following. - Sanjay Subramanian, Sameer Singh, Matt Gardner:

Analyzing Compositionality in Visual Question Answering. - Yassine Mrabet, Dina Demner-Fushman:

On Agreements in Visual Understanding. - Homagni Saha, Vijay Venkataraman, Alberto Speranzon, Soumik Sarkar:

A perspective on multi-agent communication for information fusion. - Shabnam Daghaghi, Anshumali Shrivastava, Tharun Medini:

Cross-Modal Mapping for Generalized Zero-Shot Learning by Soft-Labeling. - Candace Ross, Cheahuychou Mao, Boris Katz, Andrei Barbu:

Learning Language from Vision. - Dian Yu, Chandra Khatri, Alexandros Papangelis, Andrea Madotto, Mahdi Namazifar, Joost Huizinga, Adrien Ecoffet, Huaixiu Zheng, Piero Molino, Jeff Clune, Zhou Yu, Kenji Sagae, Gökhan Tür:

Commonsense and Semantic-Guided Navigation through Language in Embodied Environment. ViGIL@NeurIPS 2019

manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.


Google
Google Scholar
Semantic Scholar
Internet Archive Scholar
CiteSeerX
ORCID














