Yujia Qin (秦禹嘉)
yujiaqin16 <at> gmail.com

I am a Ph.D. student at Tsinghua, advised by Prof. Zhiyuan Liu . I have broad interests in natural language processing and machine learning, with a focus on pre-trained language models and their applications in information extraction and question answering .

Google Scholar  /  Twitter


B.E. Electronic Information Science and Technology, Tsinghua University, 2016-2020

Ph.D. Computer Science, Tsinghua University, 2020-2025 (expected)

Selected Publication
(*indicates equal contribution)
Exploring Mode Connectivity for Pre-trained Language Models
Yujia Qin*, Cheng Qian*, Jing Yi*, Weize Chen, Yankai Lin, Xu Han, Zhiyuan Liu, Maosong Sun and Jie Zhou
EMNLP 2022
paper  /  code
Different Tunes Played with Equal Skill: Exploring a Unified Optimization Subspace for Parameter-Efficient Tuning
Jing Yi*, Weize Chen*, Yujia Qin*, Yankai Lin, Ning Ding, Xu Han, Zhiyuan Liu, Maosong Sun and Jie Zhou
Findings of EMNLP 2022
paper  /  code
FPT: Improving Prompt Tuning Efficiency via Progressive Training
Yufei Huang*, Yujia Qin*, Huadong Wang, Yichun Yin, Maosong Sun, Zhiyuan Liu, Qun Liu
Findings of EMNLP 2022, Short Paper
paper  /  code
Moderate-fitting as a Natural Backdoor Defender for Pre-trained Language Models
Biru Zhu*, Yujia Qin*, Ganqu Cui, Yangyi Chen, Weilin Zhao, Chong Fu, Yangdong Deng, Zhiyuan Liu, Jingang Wang, Wei Wu, Maosong Sun, Ming Gu
NeurIPS 2022
paper  /  code
Delta Tuning: A Comprehensive Study of Parameter Efficient Methods for Pre-trained Language Models
Ning Ding*, Yujia Qin*, Guang Yang, Fuchao Wei, Zonghan Yang, Yusheng Su, Shengding Hu, Yulin Chen, Chi-Min Chan, Weize Chen, Jing Yi, Weilin Zhao, Xiaozhi Wang, Zhiyuan Liu, Hai-Tao Zheng, Jianfei Chen, Yang Liu, Jie Tang, Juanzi Li, Maosong Sun
paper  /  code
Exploring Universal Intrinsic Task Subspace via Prompt Tuning
Yujia Qin*, Xiaozhi Wang*, Yusheng Su, Yankai Lin, Ning Ding, Zhiyuan Liu, Juanzi Li, Lei Hou, Peng Li, Maosong Sun, Jie Zhou
previously accepted by findings of ACL 2022, and findings of EMNLP 2022
paper  /  code
Pass off Fish Eyes for Pearls: Attacking Model Selection of Pre-trained Models
Biru Zhu*, Yujia Qin*, Fanchao Qi, Yangdong Deng, Zhiyuan Liu, Maosong Sun, Ming Gu
ACL 2022, main conf
paper  /  code
ELLE: Efficient Lifelong Pre-training for Emerging Data
Yujia Qin*, Jiajie Zhang*, Yankai Lin, Zhiyuan Liu, Peng Li, Maosong Sun, Jie Zhou
ACL 2022, findings
paper  /  code
Knowledge Inheritance for Pre-trained Language Models
Yujia Qin, Yankai Lin, Jing Yi, Jiajie Zhang, Xu Han, Zhengyan Zhang, Yusheng Su, Zhiyuan Liu, Peng Li, Maosong Sun, Jie Zhou
NAACL 2022 (oral)
paper  /  code
ERICA: Improving Entity and Relation Understanding for Pre-trained Language Models via Contrastive Learning
Yujia Qin, Yankai Lin, Ryuichi Takanobu, Zhiyuan Liu, Peng Li, Heng Ji, Minlie Huang, Maosong Sun, Jie Zhou
ACL-IJCNLP 2021 (oral)
paper  /  code
Learning from Explanations with Neural Execution Tree
Ziqi Wang*, Yujia Qin*, Wenxuan Zhou, Jun Yan, Qinyuan Ye, Leonardo Neves, Zhiyuan Liu, Xiang Ren
ICLR 2020
paper  /  code  /  Project NExT
Improving Sequence Modeling Ability of Recurrent Neural Networks via Sememes
Yujia Qin*, Fanchao Qi*, Sicong Ouyang, Zhiyuan Liu, Cheng Yang, Yasheng Wang, Qun Liu, Maosong Sun
TASLP 2020
paper  /  code

Pattern recognition group, @Tencent.Wechat, Beijing, Summer Research Intern hosted by Dr. Peng Li and Dr. Yankai Lin, 2020.5 - 2020.9

Laboratory of Multimedia and Information Processing, Tsinghua University (Thesis), Research Intern advised by Prof. Ji Wu, 2019.9 - 2020.6

Intelligence and Knowledge Discovery Research Lab, USC, Research Intern hosted by Prof. Xiang Ren , 2019.6 - 2019.9

Laboratory of Natural Language Processing, Tsinghua University, advised by Prof. Zhiyuan Liu , 2018.5 - 2019.6

Pacific Century CyberWorks, Hong Kong, Summer 2018


First Prize Scholarship of 2021 Tencent Rhino-Bird Research Elite Program



The design of this website is borrowed from Jon Barron, last updated: 2022.7