default search action
ViGIL@NeurIPS 2019: Vancouver, Canada
- Visually Grounded Interaction and Language (ViGIL), NeurIPS 2019 Workshop, Vancouver, Canada, December 13, 2019. 2019
- Khanh Nguyen, Hal Daumé III:
Help, Anna! Visual Navigation with Natural Multimodal Assistance via Retrospective Curiosity-Encouraging Imitation Learning. - Alexander Kuhnle, Ann A. Copestake:
What is needed for simple spatial language capabilities in VQA? - Tsu-Jui Fu, Yuta Tsuboi, Sosuke Kobayashi, Yuta Kikuchi:
Learning from Observation-Only Demonstration for Task-Oriented Language Grounding via Self-Examination. - Jiannan Xiang, Xin Wang, William Yang Wang:
Not All Actions Are Equal: Learning to Stop in Language-Grounded Urban Navigation. - Jialin Wu, Raymond J. Mooney:
Hidden State Guidance: Improving Image Captioning Using an Image Conditioned Autoencoder. - Nikhil Krishnaswamy, James Pustejovsky:
Situated Grounding Facilitates Multimodal Concept Learning for AI. - Catalina Cangea, Eugene Belilovsky, Pietro Liò, Aaron C. Courville:
VideoNavQA: Bridging the Gap between Visual and Embodied Question Answering. - Vasu Sharma, Ankita Kalra, Louis-Philippe Morency:
Induced Attention Invariance: Defending VQA Models against Adversarial Attacks. - Xin Wang, Vihan Jain, Eugene Ie, William Yang Wang, Zornitsa Kozareva, Sujith Ravi:
Natural Language Grounded Multitask Navigation. - Farley Lai, Ning Xie, Derek Doran, Asim Kadav:
Contextual Grounding of Natural Language Entities in Images. - Olga Kovaleva, Chaitanya Shivade, Satyananda Kashyap, Karina Kanjaria, Adam Coy, Deddeh Ballah, Yufan Guo, Joy T. Wu, Alexandros Karargyris, David Beymer, Anna Rumshisky, Vandana V. Mukherjee:
Visual Dialog for Radiology: Data Curation and FirstSteps. - Thomas M. Sutter, Imant Daunhawer, Julia E. Vogt:
Multimodal Generative Learning Utilizing Jensen-Shannon-Divergence. - Guan-Lin Chao, Abhinav Rastogi, Semih Yavuz, Dilek Hakkani-Tür, Jindong Chen, Ian R. Lane:
Learning Question-Guided Video Representation for Multi-Turn Video Question Answering. - Angeliki Lazaridou, Anna Potapenko, Olivier Tieleman:
Structural and functional learning for learning language use. - Olivier Tieleman, Angeliki Lazaridou, Shibl Mourad, Charles Blundell, Doina Precup:
Community size effect in artificial learning systems. - Harm de Vries, Dzmitry Bahdanau, Shikhar Murty, Aaron C. Courville, Philippe Beaudoin:
CLOSURE: Assessing Systematic Generalization of CLEVR Models. - Chihiro Fujiyama, Ichiro Kobayashi:
A Comprehensive Analysis of Semantic Compositionality in Text-to-Image Generation. - Alba Maria Hererra-Palacio, Carles Ventura, Carina Silberer, Ionut-Teodor Sorodoc, Gemma Boleda, Xavier Giró-i-Nieto:
Recurrent Instance Segmentation using Sequences of Referring Expressions. - T. S. Jayram, Vincent Albouy, Tomasz Kornuta, Emre Sevgen, Ahmet S. Ozcan:
Visually Grounded Video Reasoning in Selective Attention Memory. - Jean-Benoit Delbrouck:
Modulated Self-attention Convolutional Network for VQA. - Gabriel Ilharco, Vihan Jain, Alexander Ku, Eugene Ie, Jason Baldridge:
General Evaluation for Instruction Conditioned Navigation using Dynamic Time Warping. - Jingxiang Lin, Unnat Jain, Alexander G. Schwing:
A Simple Baseline for Visual Commonsense Reasoning. - Nicolas Lair, Cédric Colas, Rémy Portelas, Jean-Michel Dussoux, Peter F. Dominey, Pierre-Yves Oudeyer:
Language Grounding through Social Interactions and Curiosity-Driven Multi-Goal Learning. - Yen-Ling Kuo, Boris Katz, Andrei Barbu:
Deep compositional robotic planners that follow natural language commands. ViGIL@NeurIPS 2019 - Jean-Benoit Delbrouck:
Can adversarial training learn image captioning ? - Shachi H. Kumar, Eda Okur, Saurav Sahay, Jonathan Huang, Lama Nachman:
Leveraging Topics and Audio Features with Multimodal Attention for Audio Visual Scene-Aware Dialog. - Douwe Kiela, Suvrat Bhooshan, Hamed Firooz, Davide Testuggine:
Supervised Multimodal Bitransformers for Classifying Images and Text. - Jesse Mu, Percy Liang, Noah D. Goodman:
Shaping Visual Representations with Language for Few-shot Classification. - Geoffrey Cideron, Mathieu Seurin, Florian Strub, Olivier Pietquin:
Self-Educated Language Agent with Hindsight Experience Replay for Instruction Following. - Sanjay Subramanian, Sameer Singh, Matt Gardner:
Analyzing Compositionality in Visual Question Answering. - Yassine Mrabet, Dina Demner-Fushman:
On Agreements in Visual Understanding. - Homagni Saha, Vijay Venkataraman, Alberto Speranzon, Soumik Sarkar:
A perspective on multi-agent communication for information fusion. - Shabnam Daghaghi, Anshumali Shrivastava, Tharun Medini:
Cross-Modal Mapping for Generalized Zero-Shot Learning by Soft-Labeling. - Candace Ross, Cheahuychou Mao, Boris Katz, Andrei Barbu:
Learning Language from Vision. - Dian Yu, Chandra Khatri, Alexandros Papangelis, Andrea Madotto, Mahdi Namazifar, Joost Huizinga, Adrien Ecoffet, Huaixiu Zheng, Piero Molino, Jeff Clune, Zhou Yu, Kenji Sagae, Gökhan Tür:
Commonsense and Semantic-Guided Navigation through Language in Embodied Environment. ViGIL@NeurIPS 2019
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.