Research Overview
I use AI to explain programs, models, and datasets. My goal is to empower humans to achieve what they cannot achieve on their own. See presentation slides here and my talk here to get a sense of my research interests.
|
Blogs
Representative Work
Explaining Datasets in Words: Statistical Models with Natural Language Parameters
Ruiqi Zhong,
Heng Wang,
Dan Klein,
Jacob Steinhardt
NeurIPS 2024
[paper]
[code]
[blog]
|
Learning Task Decomposition to Assist Humans in Competitive Programming
Jiaxin Wen, Ruiqi Zhong, Pei Ke, Zhihong Shao, Hongning Wang, Minlie Huang
ACL 2024
[paper] [tweet]
|
Do Models Explain Themselves? Counterfactual Simulatability of Natural Language Explanations
Yanda Chen, Ruiqi Zhong, Narutatsu Ri, Chen Zhao, He He, Jacob Steinhardt, Zhou Yu, Kathleen McKeown
ICML 2024
[paper] [tweet]
|
Others
Language Models Learn to Mislead Humans via RLHF
Jiaxin Wen, Ruiqi Zhong, Akbir Khan, Ethan Perez, Jacob Steinhardt, Minlie Huang, Samuel R. Boman, He He, Shi Feng
arXiv 2024
[paper]
[tweet]
|
Describing Differences in Image Sets with Natural Language
Lisa Dunlap*, Yuhui Zhang*, Xiaohan Wang, Ruiqi Zhong, Trevor Darrell*, Jacob Steinhardt*, Joseph E Gonzalez*, Serena Yeung-Levy*
CVPR 2024
[paper][website] [tweet]
|
Goal Driven Discovery of Distributional Differences via Language Descriptions
Ruiqi Zhong,
Peter Zhang,
Steve Li,
Jinwoo Ahn,
Dan Klein,
Jacob Steinhardt
NeurIPS 2023
[paper]
[code]
|
Goal-Driven Explainable Clustering via Language Descriptions
Zihan Wang, Jingbo Shang, Ruiqi Zhong
EMNLP 2023
[paper][code]
|
Non-Programmers Can Label Programs Indirectly via Active Examples: A Case Study with Text-to-SQL
Ruiqi Zhong*,
Charlie Snell*,
Dan Klein,
Jason Eisner
EMNLP 2023
[paper] [code] [demo] [tweet]
|
DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation
Yuhang Lai*, Chengxi Li*, Yiming Wang*, Tianyi Zhang*, Ruiqi Zhong*, Luke Zettlemoyer, Scott Wen-tau Yih, Daniel Fried, Sida Wang, Tao Yu
ICML 2023
[paper][data]
|
InCoder: A Generative Model for Code Infilling and Synthesis
Daniel Fried, Armen Aghajanyan, Jessy Lin, Sida Wang, Eric Wallace, Freda Shi, Ruiqi Zhong, Wen-tau Yih, Luke Zettlemoyer, Mike Lewis
ICLR 2023
[paper]
|
Learning by Distilling Context
Charlie Snell, Dan Klein, Ruiqi Zhong
arXiv 2022
[paper]
|
Describing Differences between Text Distributions with Natural Language
Ruiqi Zhong,
Charlie Snell,
Dan Klein,
Jacob Steinhardt
ICML 2022
[paper]
[code]
|
UnifiedSKG: Unifying and Multi-Tasking Structured Knowledge Grounding with Text-to-Text Language Models
Tianbao Xie, Chen Henry Wu, Peng Shi, Ruiqi Zhong,
Torsten Scholak, Michihiro Yasunaga, Chien-Sheng Wu, Ming Zhong, Pengcheng Yin, Sida I. Wang, Victor Zhong, Bailin Wang, Chengzu Li, Connor Boyle, Ansong Ni, Ziyu Yao, Dragomir Radev, Caiming Xiong, Lingpeng Kong, Rui Zhang, Noah A. Smith, Luke Zettlemoyer, Tao Yu
EMNLP 2022
[paper]
|
Meta-learning via Language Model In-context
Tuning
Yanda Chen, Ruiqi Zhong, Sheng Zha, George
Karypis, He He
ACL 2022
[paper]
|
The Effect of Model Size on Worst-Group Generalization
Alan Pham*, Eunice Chan*, Vikranth Srivatsa*, Dhruba Ghosh*,
Yaoqing Yang, Yaodong Yu, Ruiqi Zhong, Joseph E.
Gonzalez, Jacob Steinhardt
NeurIPS 2021 Workshop on Distribution Shifts
[paper]
|
Approximating How Single Head Attention
Learns
Charlie Snell*, Ruiqi Zhong*, Dan Klein, Jacob
Steinhardt
arXiv 2021
[paper] [slides] [code] [blog]
|
Adapting Language Models for Zero-shot Learning by Meta-tuning on Dataset and Prompt Collections
Ruiqi Zhong,
Kristy Lee*, Zheng Zhang*,
Dan Klein
EMNLP 2021, Findings
[paper][slides][code]
|
Are Larger Pretrained Language Models Uniformly Better? Comparing Performance at the Instance Level
Ruiqi Zhong,
Dhruba Ghosh,
Dan Klein,
Jacob Steinhardt
ACL 2021, Findings
[paper] [slides] [code]
|
Semantic Evaluation for Text-to-SQL with Distilled Test Suites
Ruiqi Zhong,
Tao Yu,
Dan Klein
EMNLP 2020
[paper] [slides] [code]
|
Semantic Scaffolds for Pseudocode-to-Code
Generation
Ruiqi Zhong, Mitchell Stern, Dan Klein
ACL 2020
[paper] [slides] [code] [video]
|
Detecting and Reducing Bias in a High Stakes Domain
Ruiqi Zhong, Yanda Chen, Desmond Patton, Charlotte Selous, Kathy
McKeown
EMNLP 2019
[paper] [poster] [code]
|
Fine-grained Sentiment Analysis with Faithful
Attention
Ruiqi Zhong, Steven Shao, Kathy McKeown
arXiv 2019
[paper]`
|
Detecting Gang-involved Escalation on Social
Media Using Context
Serina Chang, Ruiqi Zhong, Ethan Adams, Fei-Tzin Lee, Siddharth
Varia, Desmond Patton, William Frey, Chris Kedzie, Kathy McKeown
EMNLP 2018
[paper] [code]
|
Subspace Embedding and Linear Regression with
Orlicz Norm
Alexandr Andoni, Chengyu Lin, Ying Sheng, Peilin Zhong,
Ruiqi Zhong
ICML 2018
[paper] [video] [slides]
|
GAIA - A Multi-media Multi-lingual Knowledge Extraction and Hypothesis Generation System
Tongtao Zhang, Ananya Subburathinam, Ge Shi, Lifu Huang, Di Lu, Xiaoman Pan,
Manling Li, Boliang Zhang, Qingyun Wang, Spencer Whitehead,
Heng Ji, Alireza Zareian, Hassan Akbari, Brian Chen,
Ruiqi Zhong, Steven Shao, Emily Allaway, Shih-Fu
Chang, Kathleen R. McKeown, Dongyu Li, Xin Huang, Kexuan Sun, Xujun Peng, Ryan Gabbard, Marjorie Freedman, Mayank Kejriwal, Ram Nevatia, Pedro A. Szekely, T. K. Satish Kumar, Ali Sadeghian, Giacomo Bergami, Sourav Dutta, Miguel E. Rodríguez, Daisy Zhe Wang
TAC 2018
[paper]
|
External Presentations
- [05/02/2024] At Sasha Rush's group meeting, on explaining datasets with LLMs
- [04/16/2024] At HKU NLP class, on Scalable Oversight
- [05/18/2023] 由安远和智源社区共同举办,on Scalable Oversight (中文的幻灯片)
- [04/20/2023] At JHU CS 601, on Scalable Oversight
- [04/05/2023] At Anthropic, on Scalable Oversight
- [03/30/2023] At USC NLP Seminar, on Scalable Oversight
- [10/27/2022] At NYU, on Scalable Oversight
- [10/25/2022] At Cornell, Sasha Rush's group meeting, on Scalable Oversight
- [10/24/2022] At Columbia NLP Seminar, on Scalable Oversight
- [09/07/2022] At Redwood Research, on Scalable Oversight
- [08/23/2022] At University of Toronto, Roger Grosse's group meeting, on Scalable Oversight
- [07/28/2022] At Codex Community reading group, on Active Programming by Example with a Natural Language Prior
- [06/23/2022] At Microsoft Semantic Machines, on Scalable Oversight
|
Miscellaneous
- I represented Columbia University in ACM-ICPC and
Putnam Math Competition during my Sophormore year
(though it seems I was the bottleneck of our teams).
- I sleep at 11 p.m. and do not respond to later messages. Sometimes, however, I am actually awake; but I pretend not to see them anyways.
- My favorite animation character and role model is
Wenli Yang in Legend of Galactical Heroes.
|
Awards
- Berkeley Graduate Student Fellowship
- Theodore R. Bashkow Award (research), Academic Excellence Award (GPA)
- CRA Outstanding Undergraduate Research Award
Honorable Mention * 2 (2018, 2019)
- William Lowell Putnam Math Competition top 5% * 3
(2015, 2016, 2018)
|
Undergrad Advising
Undergrads/grads that I have mentored:
- Xinyi Han (now Ph.D. at MIT)
- Yanda Chen (now Ph.D. at Columbia)
- Charlie Snell (now Ph.D. at UC Berkeley)
- Dhruba Ghosh (now Ph.D. at University of Washington)
- Sicheng Tang
- Kristy Lee (now 5th year Master at UC Berkeley)
- Zheng Zhang (now 5th year Master at UC Berkeley)
- Harry Zhao (now 5th year Master at UC Berkeley)
- Pulkit Bhasin
- Dong Yang
- Peter Zhang
- Oscar Xu (now Ph.D. at University of Pennsylvania)
- Steve Li
- JinWoo Ahn
- Vedant Kumud
- Heng Wang
- Anu Soneye
- Jiaxin Wen
- Dominic Sobhani
|
|