Te-Lin Wu (吳德霖)

PhD Student
University of California, Los Angeles
Email: telinwu [at] cs (dot) ucla (dot) edu

Google Scholar / LinkedIn / Github

Photo

I am open to 2024 industry job opportunities!
If you think I might be a good fit for your team/group, I'm delighted to chat!

I am a final year PhD student at UCLA PlusLab advised by Nanyun (Violet) Peng, where my research focuses on multimodal models across NLP and computer vision.
I have also worked with Joseph J. Lim on reinforcement learning and vision for robotics topics.
Prior to PhD, I obtained my M.S. at Stanford University where I was advised by Silvio Savarese, and I did my undergrad at National Tsing-Hua University (國立清華大學).

Over the summers, I've been lucky to work as research interns in several wonderful groups, including Google Research, Meta Reality Labs, Amazon AI, and Adobe Research.

Selected Publications

For a full list of my publications, please see here. (* denotes equal contributions.)
 
emnlp23_jarvis
Localizing Active Objects from Egocentric Vision with Symbolic World Knowledge
Te-Lin Wu*, Yu Zhou*, Nanyun Peng
EMNLP 2023 / Paper / Video

A novel technique to ground active objects in egocentric vision with LLM-enhanced knowledge.
 
emnlp23_acquired
ACQUIRED: A Dataset for Answering Counterfactual Questions In Real-Life Videos
Te-Lin Wu*, Zi-Yi Dou*, Qingyuan Hu*, Yu Hou, Nischal Chandra, Marjorie Freedman, Ralph Weischedel, Nanyun Peng
EMNLP 2023 / Paper / Video

A novel dataset for understanding counterfactual commonsense reasoning in videos.
 
acl23_simmcvr
SIMMC-VR: A Task-oriented Multimodal Dialog Dataset with Situated and Immersive VR Streams
Te-Lin Wu, Satwik Kottur, Andrea Madotto, Mahmoud Azab, Pedro Rodriguez, Babak Damavandi, Nanyun Peng, Seungwhan Moon
ACL 2023 / Paper / Video

A dataset for situated conversational agent with applications in AR/VR shopping domains.
acl23-condition
Learning Action Conditions from Instructional Manuals for Instruction Understanding
Te-Lin Wu, Caiqi Zhang, Qingyuan Hu, Alex Spangher, Nanyun Peng
ACL 2023 / Paper / Video

We learn a model to perform pre- and post-conditon inference to actionables in instructional manuals via a weakly-supervised method.
emnlp22-char
Character-Centric Story Visualization via Visual Planning and Token Alignment
Hong Chen, Rujun Han, Te-Lin Wu, Hideki Nakayama, Nanyun Peng
EMNLP 2022 / Paper / Code

A method that utilizes grad-cam to propose plausible character plans for story generation (visualization) task.
acl22_sort
Understanding Multimodal Procedural Knowledge by Sequencing Multimodal Instructional Manuals
Te-Lin Wu, Alex Spangher, Pegah Alipoormolabashi, Marjorie Freedman, Ralph Weischedel, Nanyun Peng
ACL 2022 / Paper / Video

We propose several sequence-aware pre-training objectives to equip multimodal models with task-order knowledge.
emnlp21_hyperexpan
HyperExpan: Taxonomy Expansion with Hyperbolic Representation Learning
Mingyu Ma, Muhao Chen*, Te-Lin Wu*, Nanyun Peng,
Findings of EMNLP 2021

Using a hyperbolic representation learning scheme is more effective more KG taxonomy expansion.
acl21_com2sense
COM2SENSE: A Commonsense Reasoning Benchmark with Complementary Sentences
Shikhar Singh∗, Nuan Wen∗, Yu Hou, Pegah Alipoormolabashi, Te-Lin Wu, Xuezhe Ma, Nanyun Peng
Findings of ACL 2021 / Paper / Dataset & Codes /

A dataset for complementary commonsense reasoning collected via a model-in-the-loop gamified session.
aaai21_melinda
MELINDA: A Multimodal Dataset for Biomedical Experiment Method Classification
Te-Lin Wu, Shikhar Singh, Sayan Paul, Gully Burns, Nanyun Peng
AAAI 2021 / Paper / Dataset & Codes

A dataset for multimodal biomedical method classification.
naacl21_lampret
LAMPRET: Layout-Aware Multimodal PreTraining for Document Understanding
Te-Lin Wu, Cheng Li, Mingyang Zhang, Tao Chen, Spurthi Amba Hombaiah, Michael Bendersky
ViGIL Workshop, NAACL 2021 / Paper

A pre-training paradigm to exploit document layout to learn a document representation.
iclr20_pga
Program Guided Agent
Shao-Hua Sun, Te-Lin Wu, Joseph J. Lim
ICLR 2020

A framework to programmatically control an RL-trained agent.
cvpr18_demo2vec
Demo2Vec: Reasoning Object Affordances from Online Videos
Te-Lin Wu*, Kuan Fang*, Daniel Yang, Silvio Savarese, Joseph J. Lim
CVPR 2018

Learning to infer object affordance with a video demonstration of how to interact with objects.
cvpr17_feedback_net
Feedback Networks
Te-Lin Wu*, Amir R. Zamir*, Lin Sun, William B. Shen, Bertram E. Shi, Jitendra Malik, Silvio Savarese
CVPR 2017

A study of feedback mechanism of convolutional neural networks.