Hi! I am a second-year PhD student at Language Technologies Institute (LTI), Carnegie Mellon University (CMU), advised by Prof. Chenyan Xiong. My primary research interests are exploring novel ways to efficiently train and apply large language models in sample-limited and computation-limited scenarios. I am currently working on data valuation, hopefully derived from model preference, for a more efficient pre-training procedure.
Previously, I graduated from Tsinghua University in 2023 with a major in Computer Science and Technology. I was honored to be a member of THUNLP, advised by Prof. Zhiyuan Liu, working closely with Dr. Tianyu Gao and Dr. Zhengyan Zhang in prompting and few-shot learning. I was a research intern at UWNLP, advised by Prof. Sheng Wang, and an intern at Baidu NLP group.
When I am not doing research, I like to work out, play guitar, and watch movies.
Updates:
- June 2024: Check out our pretraining data curation paper: MATESπ§βπ€βπ§: Model-Aware Data Selection for Efficient Pretraining with Data Influence Models at NeurIPS 2024 (poster) β¨
- December 2023: Check out our benchmarking LLMs paper: An In-depth Look at Gemini's Language Abilities β¨
- August 2023: Begin my PhD at CMU πͺ
- May 2023: Check out our generic retrieval augmentation paper: Augmentation-Adapted Retriever Improves Generalization of Language Models as Generic Plug-In at ACL 2023 (oral presentation) β¨
- August 2022: Check out our automatic prompting paper: Automatic Label Sequence Generation for Prompting Sequence-to-sequence Models at COLING 2022 (oral presentation) β¨