News
- Oct 19, 2023 - Honored to be awarded the
Baosteel Outstanding Student Award 2023
as the ONLY undergrad student among science and technology departments in RUC! Special thanks to
NLPIR lab
!
- Sep 20, 2023 - Glad to announce that I am currently working on a brand-new topic of
fine-grained control on interactive language models with
Dr. Weiyan Shi from Stanford University! Paper will come out soon!
- Jun 7, 2023 - Excited to share that I'll be joining
UIUC Blender Lab
this summer as a student researcher!
- Mar 15, 2023 - My talk on LARGE language models for
Capital of Statistics
will take place at 7:00 PM Mar 17, 2023 BJT! Click
here
for more details. (Update: slides, video)
- Jan 12, 2023 - I will give a talk on pre-trained models and their applications
at 2:00 PM Jan 13, 2023 BJT for
Mingli College! For more information, click
here.
(Update: slides)
- Dec 12, 2022 - I posted an article introducing ChatGPT on
Capital of Statistics.
Do not miss it if you want to know more about ChatGPT!
(Only Chinese version available currently)
(link)
|
Research Interest
Research Experiences
|
MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback
Xingyao Wang*,
Zihan Wang*,
Jiateng Liu,
Yangyi Chen,
Lifan Yuan,
Hao Peng,
Heng Ji
International Conference on Learning Representations (ICLR) 2024
[paper]
[code]
[website]
Current LLM evaluations focus on single-turn and overlook multi-turn real-world scenarios.
We introduce the MINT benchmark to assess LLMs in interaction with tools and language feedback.
Our study of 20 LLMs shows they benefit from multi-turn interactions, but current RLHF and SIFT methods might hinder this.
MINT aims to encourage research on LLM multi-turn capabilities, especially in open-source models.
|
|
NOVO: Learnable and Interpretable Document Identifiers for Model Based IR
Zihan Wang,
Yujia Zhou,
Yiteng Tu,
Zhicheng Dou.
Conference on Information and Knowledge Management (CIKM) 2023, Oral Presentation
[paper]
[code]
We propose Neural Optimized Vocabularial doc-ids for model-based information retrieval.
NOVO doc-ids consist of non-overlapping n-gram sets to identify documents,
optimized through query denoising modeling and retrieval tasks.
We further demonstrate how N-grams lead to interpretability in understanding documents and queries.
|
|
RetaLLM: A Retrieval-Augmented Large Language Model Toolkit
Jiongnan Liu,
Jiajie Jin,
Zihan Wang,
Jiehan Cheng,
Zhicheng Dou,
Ji-Rong Wen
Arxiv Preprint
[paper]
[code]
We develop a RETreival-Augmented LLM toolkit.
It provides more plug-and-play modules to support better interaction between IR systems and LLMs,
including request rewriting, document retrieval, passage extraction, answer generation, and fact checking modules.
|
|
Learning on Structured Documents for Conditional Question Answering
Zihan Wang,
Hongjin Qian,
Zhicheng Dou
China National Conference on Computational Linguistics (CCL) 2023
[paper][code][poster]
We propose a self-supervised learning method on structured documents for conditional question answering,
consisting of a conditional question generation approach and a contrastive learning objective.
We propose a pipeline able to generate multiple answers with detailed conditions.
|
Misc
- I like to work and chat with people from diverse backgrounds (🌈), which I believe is the key to true innovation. Feel free to contact me.
- I love Sandbox games like Minecraft and Danmaku games like Touhou Project.
I also loved designing RPG games when I was in primary school (with
RMXP on WindowsXP), although
they cannot be launched anymore on Win10.
- My dream was to be a vlogger and I posted
videos
on bilibili, including vlogs, game playing records and some parody videos.
- Besides Chinese and English, I can speak a little Japanese due to my passion in Anime
in my childhood. My favorite Anime was ワンピース and Fate/stay night.
|
Awards
- Baosteel Outstanding Student Award, 7/30000+, Renmin Univ. of China, 2023
- First Class Academic Excellence Award (top 3% GPA), Renmin Univ. of China, 2021
- Provincal First Prize, Contemporary Undergraduate Mathematical Contest in
Modeling, 2021
- Honorable Mention, Mathematical Contest in Modeling and Interdisciplinary
Contest in Modeling, 2021
|
|