Zhiwei Yu
Exploring the World
Senior Researcher at Multimodal Interaction Research Center
Beijing Academy of Artificial Intelligence
Beijing Academy of Artificial Intelligence
About Me
I design smart systems—digital or embodied— to turn real-world chaos into innovation—where AI,
product, and business collide. My research interests span World Modeling, Embodied AI, and Knowledge Computing. My recent focus is long-horizon loco-manipulation challenges in huimanoid robots.
Previously, I graduated from AAIS and WICT at Peking University, advised by Xiaojun Wan. Then I joined MSRA as a researacher. During my stay in MS, I directed research on knowledge-base question answering (KBQA) and steered AI integration across flagship products such as Office, Bing and Windows.
Previously, I graduated from AAIS and WICT at Peking University, advised by Xiaojun Wan. Then I joined MSRA as a researacher. During my stay in MS, I directed research on knowledge-base question answering (KBQA) and steered AI integration across flagship products such as Office, Bing and Windows.
Selected Publications
TIARA: Multi-grained Retrieval for Robust Question Answering over Large
Knowledge Bases
Pre-trained language models (PLMs) have shown their effectiveness in multiple scenarios.
However, KBQA remains challenging, especially regarding coverage and generalization settings.
This is due to two main factors: i) understanding the semantics of both questions and relevant
knowledge from the KB; ii) generating executable logical forms with both semantic and syntactic
correctness. In this paper, we present a new KBQA model, TIARA, which addresses those issues by
applying multi-grained retrieval to help the PLM focus on the most relevant KB contexts, viz.,
entities, exemplary logical forms, and schema items. Moreover, constrained decoding is used to
control the output space and reduce generation errors. Experiments over important benchmarks
demonstrate the effectiveness of our approach. TIARA outperforms previous SOTA, including those
using PLMs or oracle entity annotations, by at least 4.1 and 1.1 F1 points on GrailQA and
WebQuestionsSP, respectively.
EMNLP, 2022
ReTraCk: A Flexible and Efficient Framework for Knowledge Base Question
Answering
We present Retriever-Transducer-Checker (ReTraCk), a neural semantic parsing framework for large
scale knowledge base question answering (KBQA). ReTraCk is designed as a modular framework to
maintain high flexibility. It includes a retriever to retrieve relevant KB items efficiently, a
transducer to generate logical form with syntax correctness guarantees and a checker to improve
transduction procedure. ReTraCk is ranked at top1 overall performance on the GrailQA leaderboard
and obtains highly competitive performance on the typical WebQuestionsSP benchmark. Our system
can interact with users timely, demonstrating the efficiency of the proposed framework.
ACL, 2021
A Neural Approach to Pun Generation
Automatic pun generation is an interesting and challenging text generation task. Previous
efforts rely on templates or laboriously manually annotated pun datasets, which heavily
constrains the quality and diversity of generated puns. Since sequence-to-sequence models
provide an effective technique for text generation, it is promising to investigate these models
on the pun generation task. In this paper, we propose neural network models for homographic pun
generation, and they can generate puns without requiring any pun data for training. We first
train a conditional neural language model from a general text corpus, and then generate puns
from the language model with an elaborately designed decoding algorithm. Automatic and human
evaluations show that our models are able to generate homographic puns of good readability and
quality.
ACL, 2018
For Prospective Students
I am always looking for motivated graduate students and postdocs to join my research group. If you are interested in machine learning research, please send me an email with your CV, research interests, and a brief description of your background.