Zihan Wang

I am an undergraduate student from GSAI, Renmin University of China. I am fortunate to be advised by Zhicheng Dou, working on Natural Language Processing and Information Retrieval. I also work closely with Heng Ji from UIUC and Weiyan Shi from Stanford University. I had a fantastic experience as an exchange student at UC Berkeley.

I'm looking for a Ph.D. position starting from Fall 2024. Feel free to contact me if you are interested!

Resume  /  Research Statement  
  Email  /  Twitter  /  Github  /  Semantic Scholar  /  Zhihu

profile photo
News
  • Oct 19, 2023 - Honored to be awarded the Baosteel Outstanding Student Award 2023 as the ONLY undergrad student among science and technology departments in RUC! Special thanks to NLPIR lab !
  • Sep 20, 2023 - Glad to announce that I am currently working on a brand-new topic of fine-grained control on interactive language models with Dr. Weiyan Shi from Stanford University! Paper will come out soon!
  • Jun 7, 2023 - Excited to share that I'll be joining UIUC Blender Lab this summer as a student researcher!
  • Mar 15, 2023 - My talk on LARGE language models for Capital of Statistics will take place at 7:00 PM Mar 17, 2023 BJT! Click here for more details. (Update: slides, video)
  • Jan 12, 2023 - I will give a talk on pre-trained models and their applications at 2:00 PM Jan 13, 2023 BJT for Mingli College! For more information, click here. (Update: slides)
  • Dec 12, 2022 - I posted an article introducing ChatGPT on Capital of Statistics. Do not miss it if you want to know more about ChatGPT! (Only Chinese version available currently) (link)
Research Interest

I work on Augmented Language Models. I have done several works regarding (1) general interaction, e.g., interaction benchmark, (2) the cross application of language models and information retrieval (IR) systems, e.g., retrieval augmented LMs and LM-based IR. I'm currently working on a boarder range of LM Augmentation: controlling interactive language models. I'm also interested in AI alignment topics like scalable oversight.

Research Experiences
MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback
Xingyao Wang*, Zihan Wang*, Jiateng Liu, Yangyi Chen, Lifan Yuan, Hao Peng, Heng Ji
International Conference on Learning Representations (ICLR) 2024
[paper] [code] [website]

Current LLM evaluations focus on single-turn and overlook multi-turn real-world scenarios. We introduce the MINT benchmark to assess LLMs in interaction with tools and language feedback. Our study of 20 LLMs shows they benefit from multi-turn interactions, but current RLHF and SIFT methods might hinder this. MINT aims to encourage research on LLM multi-turn capabilities, especially in open-source models.


NOVO: Learnable and Interpretable Document Identifiers for Model Based IR
Zihan Wang, Yujia Zhou, Yiteng Tu, Zhicheng Dou.
Conference on Information and Knowledge Management (CIKM) 2023, Oral Presentation
[paper] [code]

We propose Neural Optimized Vocabularial doc-ids for model-based information retrieval. NOVO doc-ids consist of non-overlapping n-gram sets to identify documents, optimized through query denoising modeling and retrieval tasks. We further demonstrate how N-grams lead to interpretability in understanding documents and queries.


RetaLLM: A Retrieval-Augmented Large Language Model Toolkit
Jiongnan Liu, Jiajie Jin, Zihan Wang, Jiehan Cheng, Zhicheng Dou, Ji-Rong Wen
Arxiv Preprint
[paper] [code]

We develop a RETreival-Augmented LLM toolkit. It provides more plug-and-play modules to support better interaction between IR systems and LLMs, including request rewriting, document retrieval, passage extraction, answer generation, and fact checking modules.

Learning on Structured Documents for Conditional Question Answering
Zihan Wang, Hongjin Qian, Zhicheng Dou
China National Conference on Computational Linguistics (CCL) 2023
[paper][code][poster]

We propose a self-supervised learning method on structured documents for conditional question answering, consisting of a conditional question generation approach and a contrastive learning objective. We propose a pipeline able to generate multiple answers with detailed conditions.

Misc
  • I like to work and chat with people from diverse backgrounds (🌈), which I believe is the key to true innovation. Feel free to contact me.
  • I love Sandbox games like Minecraft and Danmaku games like Touhou Project. I also loved designing RPG games when I was in primary school (with RMXP on WindowsXP), although they cannot be launched anymore on Win10.
  • My dream was to be a vlogger and I posted videos on bilibili, including vlogs, game playing records and some parody videos.
  • Besides Chinese and English, I can speak a little Japanese due to my passion in Anime in my childhood. My favorite Anime was ワンピース and Fate/stay night.
Awards
  • Baosteel Outstanding Student Award, 7/30000+, Renmin Univ. of China, 2023
  • First Class Academic Excellence Award (top 3% GPA), Renmin Univ. of China, 2021
  • Provincal First Prize, Contemporary Undergraduate Mathematical Contest in Modeling, 2021
  • Honorable Mention, Mathematical Contest in Modeling and Interdisciplinary Contest in Modeling, 2021

Website design from Jon Barron