
Zhiwei Yu
Exploring the World
Senior Researcher at Multimodal Interaction Research Center
BAAI
BAAI
About Me
I design smart systems—digital or embodied— to turn real-world chaos into innovation—where AI,
product, and business collide. My recent focus is long-horizon loco-manipulation challenges in huimanoid legged robots.
Research Interests
- Multimodal & Embodied AI
- Knowledge Computing
- Human-Robot Interaction
- Computational Humor
Selected Publications
TIARA: Multi-grained Retrieval for Robust Question Answering over Large
Knowledge Bases
Pre-trained language models (PLMs) have shown their effectiveness in multiple scenarios.
However, KBQA remains challenging, especially regarding coverage and generalization settings.
This is due to two main factors: i) understanding the semantics of both questions and relevant
knowledge from the KB; ii) generating executable logical forms with both semantic and syntactic
correctness. In this paper, we present a new KBQA model, TIARA, which addresses those issues by
applying multi-grained retrieval to help the PLM focus on the most relevant KB contexts, viz.,
entities, exemplary logical forms, and schema items. Moreover, constrained decoding is used to
control the output space and reduce generation errors. Experiments over important benchmarks
demonstrate the effectiveness of our approach. TIARA outperforms previous SOTA, including those
using PLMs or oracle entity annotations, by at least 4.1 and 1.1 F1 points on GrailQA and
WebQuestionsSP, respectively.
EMNLP, 2022
ReTraCk: A Flexible and Efficient Framework for Knowledge Base Question
Answering
We present Retriever-Transducer-Checker (ReTraCk), a neural semantic parsing framework for large
scale knowledge base question answering (KBQA). ReTraCk is designed as a modular framework to
maintain high flexibility. It includes a retriever to retrieve relevant KB items efficiently, a
transducer to generate logical form with syntax correctness guarantees and a checker to improve
transduction procedure. ReTraCk is ranked at top1 overall performance on the GrailQA leaderboard
and obtains highly competitive performance on the typical WebQuestionsSP benchmark. Our system
can interact with users timely, demonstrating the efficiency of the proposed framework.
ACL, 2021
A Neural Approach to Pun Generation
Automatic pun generation is an interesting and challenging text generation task. Previous
efforts rely on templates or laboriously manually annotated pun datasets, which heavily
constrains the quality and diversity of generated puns. Since sequence-to-sequence models
provide an effective technique for text generation, it is promising to investigate these models
on the pun generation task. In this paper, we propose neural network models for homographic pun
generation, and they can generate puns without requiring any pun data for training. We first
train a conditional neural language model from a general text corpus, and then generate puns
from the language model with an elaborately designed decoding algorithm. Automatic and human
evaluations show that our models are able to generate homographic puns of good readability and
quality.
ACL, 2018
For Prospective Students
I am always looking for motivated graduate students and postdocs to join my research group. If you are interested in machine learning research, please send me an email with your CV, research interests, and a brief description of your background.