Tianwei Chen

Follow

Generating author description...

Common Coauthors
Commonly Cited References
Action Title Year Authors # of times referenced
+ PDF Chat ImageNet Large Scale Visual Recognition Challenge 2015 Olga Russakovsky
Jia Deng
Hao Su
Jonathan Krause
Sanjeev Satheesh
Sean Ma
Zhiheng Huang
Andrej Karpathy
Aditya Khosla
Michael S. Bernstein
1
+ PDF Chat YFCC100M 2016 Bart Thomée
David A. Shamma
Gerald Friedland
Benjamin Elizalde
Karl Ni
Douglas N. Poland
Damian Borth
Li-Jia Li
1
+ PDF Chat GuessWhat?! Visual Object Discovery through Multi-modal Dialogue 2017 Harm de Vries
Florian Strub
Sarath Chandar
Olivier Pietquin
Hugo Larochelle
Aaron Courville
1
+ Overcoming catastrophic forgetting in neural networks 2017 James Kirkpatrick
Razvan Pascanu
Neil C. Rabinowitz
Joel Veness
Guillaume Desjardins
Andrei A. Rusu
Kieran Milan
John Quan
Tiago Ramalho
Agnieszka Grabska‐Barwińska
1
+ PDF Chat Making the V in VQA Matter: Elevating the Role of Image Understanding in Visual Question Answering 2017 Yash Goyal
Tejas Khot
Douglas Summers-Stay
Dhruv Batra
Devi Parikh
1
+ PDF Chat CLEVR: A Diagnostic Dataset for Compositional Language and Elementary Visual Reasoning 2017 Justin Johnson
Bharath Hariharan
Laurens van der Maaten
Li Fei-Fei
C. Lawrence Zitnick
Ross Girshick
1
+ PDF Chat VQA: Visual Question Answering 2016 Aishwarya Agrawal
Jiasen Lu
Stanislaw Antol
Margaret Mitchell
C. Lawrence Zitnick
Devi Parikh
Dhruv Batra
1
+ PDF Chat Flickr30k Entities: Collecting Region-to-Phrase Correspondences for Richer Image-to-Sentence Models 2016 Bryan A. Plummer
Liwei Wang
Chris M. Cervantes
Juan C. Caicedo
Julia Hockenmaier
Svetlana Lazebnik
1
+ PDF Chat Visual7W: Grounded Question Answering in Images 2016 Yuke Zhu
Oliver Groth
Michael S. Bernstein
Li Fei-Fei
1
+ PDF Chat Visual Question Answering as a Meta Learning Task 2018 Damien Teney
Anton van den Hengel
1
+ PDF Chat Generation and Comprehension of Unambiguous Object Descriptions 2016 Junhua Mao
Jonathan Huang
Alexander Toshev
Oana Camburu
Alan Yuille
Kevin Murphy
1
+ PDF Chat GQA: A New Dataset for Real-World Visual Reasoning and Compositional Question Answering 2019 Drew A. Hudson
Christopher D. Manning
1
+ A Corpus for Reasoning about Natural Language Grounded in Photographs 2019 Alane Suhr
Stephanie Zhou
Ally Zhang
Iris Zhang
Huajun Bai
Yoav Artzi
1
+ PDF Chat Multi-Task Learning of Hierarchical Vision-Language Representation 2019 Duy-Kien Nguyen
Takayuki Okatani
1
+ PDF Chat VizWiz Grand Challenge: Answering Visual Questions from Blind People 2018 Danna Gurari
Qing Li
Abigale Stangl
Anhong Guo
Chi Lin
Kristen Grauman
Jiebo Luo
Jeffrey P. Bigham
1
+ PDF Chat A Survey of Current Datasets for Vision and Language Research 2015 Francis Ferraro
Nasrin Mostafazadeh
Ting-Hao Huang
Lucy Vanderwende
Jacob Devlin
Michel Galley
Margaret Mitchell
1
+ PDF Chat How to Read Paintings: Semantic Art Understanding with Multi-modal Retrieval 2019 Noa García
George Vogiatzis
1
+ PDF Chat Taskonomy: Disentangling Task Transfer Learning 2018 Amir Zamir
Alexander F. Sax
William Shen
Leonidas Guibas
Jitendra Malik
Silvio Savarese
1
+ LXMERT: Learning Cross-Modality Encoder Representations from Transformers 2019 Hao Tan
Mohit Bansal
1
+ PDF Chat Unicoder-VL: A Universal Encoder for Vision and Language by Cross-Modal Pre-Training 2020 Gen Li
Nan Duan
Yuejian Fang
Ming Gong
Daxin Jiang
1
+ PDF Chat 12-in-1: Multi-Task Vision and Language Representation Learning 2020 Jiasen Lu
Vedanuj Goswami
Marcus Rohrbach
Devi Parikh
Stefan Lee
1
+ PDF Chat UNITER: UNiversal Image-TExt Representation Learning 2020 Yen-Chun Chen
Linjie Li
Licheng Yu
Ahmed El Kholy
Faisal Ahmed
Zhe Gan
Yu Cheng
Jingjing Liu
1
+ PDF Chat Oscar: Object-Semantics Aligned Pre-training for Vision-Language Tasks 2020 Xiujun Li
Xi Yin
Chunyuan Li
Pengchuan Zhang
Xiaowei Hu
Lei Zhang
Lijuan Wang
Houdong Hu
Dong Li
Furu Wei
1
+ PDF Chat Big Transfer (BiT): General Visual Representation Learning 2020 Alexander Kolesnikov
Lucas Beyer
Xiaohua Zhai
Joan Puigcerver
Jessica Yung
Sylvain Gelly
Neil Houlsby
1
+ PDF Chat UniT: Multimodal Multitask Learning with a Unified Transformer 2021 Ronghang Hu
Amanpreet Singh
1
+ PDF Chat Diagnosing Vision-and-Language Navigation: What Really Matters 2022 Wanrong Zhu
Yuankai Qi
Pradyumna Narayana
Kazoo Sone
Sugato Basu
Xin Wang
Qi Wu
Miguel P. Eckstein
William Yang Wang
1
+ PDF Chat VinVL: Revisiting Visual Representations in Vision-Language Models 2021 Pengchuan Zhang
Xiujun Li
Xiaowei Hu
Jianwei Yang
Lei Zhang
Lijuan Wang
Yejin Choi
Jianfeng Gao
1
+ PDF Chat Trends in Integration of Vision and Language Research: A Survey of Tasks, Datasets, and Methods 2021 Aditya Mogadala
Marimuthu Kalimuthu
Dietrich Klakow
1
+ Factors of Influence for Transfer Learning Across Diverse Appearance Domains and Task Types 2021 Thomas Mensink
Jasper Uijlings
Alina Kuznetsova
Michael Gygli
Vittorio Ferrari
1
+ PDF Chat Hierarchical Modular Network for Video Captioning 2022 Hanhua Ye
Guorong Li
Yuankai Qi
Shuhui Wang
Qingming Huang
Ming–Hsuan Yang
1
+ PDF Chat CLIP Models are Few-Shot Learners: Empirical Studies on VQA and Visual Entailment 2022 Haoyu Song
Dong Li
Weinan Zhang
Ting Liu
Furu Wei
1
+ PDF Chat Multi-Attention Network for Compressed Video Referring Object Segmentation 2022 Weidong Chen
Dexiang Hong
Yuankai Qi
Zhenjun Han
Shuhui Wang
Laiyun Qing
Qingming Huang
Guorong Li
1
+ PDF Chat VL-ADAPTER: Parameter-Efficient Transfer Learning for Vision-and-Language Tasks 2022 Yi-Lin Sung
Jaemin Cho
Mohit Bansal
1
+ PDF Chat High-Resolution Image Synthesis with Latent Diffusion Models 2022 Robin Rombach
Andreas Blattmann
Dominik Lorenz
Patrick Esser
Björn Ommer
1
+ PDF Chat Vision and Structured-Language Pretraining for Cross-Modal Food Retrieval 2023 Mustafa Shukor
Nicolas Thome
Matthieu Cord
1
+ PDF Chat All in One: Exploring Unified Video-Language Pre-Training 2023 Jinpeng Wang
Yixiao Ge
Rui Yan
Yuying Ge
Kevin Qinghong Lin
Satoshi Tsutsui
Xudong Lin
Guanyu Cai
Jianping Wu
Ying Shan
1