


Остановите войну!
for scientists:


default search action
Weizhu Chen
Person information

Refine list

refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
showing all ?? records
2020 – today
- 2023
- [c91]Chenxiao Liu, Shuai Lu, Weizhu Chen, Daxin Jiang, Alexey Svyatkovskiy, Shengyu Fu, Neel Sundaresan, Nan Duan:
Code Execution with Pre-trained Language Models. ACL (Findings) 2023: 4984-4999 - [c90]Yifei Li, Zeqi Lin, Shizhuo Zhang, Qiang Fu, Bei Chen, Jian-Guang Lou, Weizhu Chen:
Making Language Models Better Reasoners with Step-Aware Verifier. ACL (1) 2023: 5315-5333 - [c89]Weizhou Shen, Yeyun Gong, Yelong Shen, Song Wang, Xiaojun Quan, Nan Duan, Weizhu Chen:
Joint Generator-Ranker Learning for Natural Language Generation. ACL (Findings) 2023: 7681-7699 - [c88]Xuxi Chen, Tianlong Chen, Weizhu Chen, Ahmed Hassan Awadallah, Zhangyang Wang, Yu Cheng:
DSEE: Dually Sparsity-embedded Efficient Tuning of Pre-trained Language Models. ACL (1) 2023: 8208-8222 - [c87]Bei Chen, Fengji Zhang, Anh Nguyen, Daoguang Zan, Zeqi Lin, Jian-Guang Lou, Weizhu Chen:
CodeT: Code Generation with Generated Tests. ICLR 2023 - [c86]Pengcheng He, Jianfeng Gao, Weizhu Chen:
DeBERTaV3: Improving DeBERTa using ELECTRA-Style Pre-Training with Gradient-Disentangled Embedding Sharing. ICLR 2023 - [c85]Zhendong Wang, Huangjie Zheng, Pengcheng He, Weizhu Chen, Mingyuan Zhou:
Diffusion-GAN: Training GANs with Diffusion. ICLR 2023 - [c84]Qingru Zhang, Minshuo Chen, Alexander Bukharin, Pengcheng He, Yu Cheng, Weizhu Chen, Tuo Zhao:
Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning. ICLR 2023 - [c83]Huangjie Zheng, Pengcheng He, Weizhu Chen, Mingyuan Zhou:
Truncated Diffusion Probabilistic Models and Diffusion-based Adversarial Auto-Encoders. ICLR 2023 - [c82]Yixiao Li, Yifan Yu, Qingru Zhang, Chen Liang, Pengcheng He, Weizhu Chen, Tuo Zhao:
LoSparse: Structured Compression of Large Language Models based on Low-Rank and Sparse Approximation. ICML 2023: 20336-20350 - [c81]Chen Liang, Simiao Zuo, Qingru Zhang, Pengcheng He, Weizhu Chen, Tuo Zhao:
Less is More: Task-aware Layer-wise Distillation for Language Model Compression. ICML 2023: 20852-20867 - [c80]Zhenghao Lin, Yeyun Gong, Yelong Shen, Tong Wu, Zhihao Fan, Chen Lin, Nan Duan, Weizhu Chen:
Text Generation with Diffusion Language Models: A Pre-training Approach with Continuous Paragraph Denoise. ICML 2023: 21051-21064 - [c79]Jason Phang, Yi Mao, Pengcheng He, Weizhu Chen:
HyperTuning: Toward Adapting Large Language Models without Back-propagation. ICML 2023: 27854-27875 - [c78]Zhihong Shao, Yeyun Gong, Yelong Shen, Minlie Huang, Nan Duan, Weizhu Chen:
Synthetic Prompting: Generating Chain-of-Thought Demonstrations for Large Language Models. ICML 2023: 30706-30775 - [i91]Zhihong Shao, Yeyun Gong, Yelong Shen, Minlie Huang, Nan Duan
, Weizhu Chen:
Synthetic Prompting: Generating Chain-of-Thought Demonstrations for Large Language Models. CoRR abs/2302.00618 (2023) - [i90]Baolin Peng, Michel Galley, Pengcheng He, Hao Cheng, Yujia Xie, Yu Hu, Qiuyuan Huang, Lars Liden, Zhou Yu, Weizhu Chen, Jianfeng Gao:
Check Your Facts and Try Again: Improving Large Language Models with External Knowledge and Automated Feedback. CoRR abs/2302.12813 (2023) - [i89]Anh Nguyen, Nikos Karampatziakis, Weizhu Chen:
Meet in the Middle: A New Pre-training Paradigm. CoRR abs/2303.07295 (2023) - [i88]Qingru Zhang, Minshuo Chen, Alexander Bukharin, Pengcheng He, Yu Cheng, Weizhu Chen, Tuo Zhao:
Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning. CoRR abs/2303.10512 (2023) - [i87]Fengji Zhang, Bei Chen, Yue Zhang, Jin Liu, Daoguang Zan, Yi Mao, Jian-Guang Lou, Weizhu Chen:
RepoCoder: Repository-Level Code Completion Through Iterative Retrieval and Generation. CoRR abs/2303.12570 (2023) - [i86]Xingwei He, Zhenghao Lin, Yeyun Gong, A-Long Jin, Hang Zhang, Chen Lin, Jian Jiao, Siu Ming Yiu, Nan Duan, Weizhu Chen:
AnnoLLM: Making Large Language Models to Be Better Crowdsourced Annotators. CoRR abs/2303.16854 (2023) - [i85]Wanjun Zhong, Ruixiang Cui, Yiduo Guo, Yaobo Liang, Shuai Lu, Yanlin Wang, Amin Saied, Weizhu Chen, Nan Duan:
AGIEval: A Human-Centric Benchmark for Evaluating Foundation Models. CoRR abs/2304.06364 (2023) - [i84]Zhendong Wang, Yifan Jiang, Huangjie Zheng, Peihao Wang, Pengcheng He, Zhangyang Wang, Weizhu Chen, Mingyuan Zhou:
Patch Diffusion: Faster and More Data-Efficient Training of Diffusion Models. CoRR abs/2304.12526 (2023) - [i83]Zhendong Wang, Yifan Jiang, Yadong Lu, Yelong Shen, Pengcheng He, Weizhu Chen, Zhangyang Wang, Mingyuan Zhou:
In-Context Learning Unlocked for Diffusion Models. CoRR abs/2305.01115 (2023) - [i82]Chenxiao Liu, Shuai Lu, Weizhu Chen, Daxin Jiang, Alexey Svyatkovskiy, Shengyu Fu, Neel Sundaresan, Nan Duan:
Code Execution with Pre-trained Language Models. CoRR abs/2305.05383 (2023) - [i81]Tong Wu, Zhihao Fan, Xiao Liu, Yeyun Gong, Yelong Shen, Jian Jiao, Hai-Tao Zheng, Juntao Li, Zhongyu Wei, Jian Guo, Nan Duan, Weizhu Chen:
AR-Diffusion: Auto-Regressive Diffusion Model for Text Generation. CoRR abs/2305.09515 (2023) - [i80]Zhibin Gou, Zhihong Shao, Yeyun Gong, Yelong Shen, Yujiu Yang, Nan Duan, Weizhu Chen:
CRITIC: Large Language Models Can Self-Correct with Tool-Interactive Critiquing. CoRR abs/2305.11738 (2023) - [i79]Shengnan An, Bo Zhou, Zeqi Lin, Qiang Fu, Bei Chen, Nanning Zheng, Weizhu Chen, Jian-Guang Lou:
Skill-Based Few-Shot Selection for In-Context Learning. CoRR abs/2305.14210 (2023) - [i78]Woojeong Jin, Subhabrata Mukherjee, Yu Cheng, Yelong Shen, Weizhu Chen, Ahmed Hassan Awadallah, Damien Jose, Xiang Ren:
GRILL: Grounded Vision-language Pre-training via Aligning Text and Image Regions. CoRR abs/2305.14676 (2023) - [i77]Zhihong Shao, Yeyun Gong, Yelong Shen, Minlie Huang, Nan Duan, Weizhu Chen:
Enhancing Retrieval-Augmented Large Language Models with Iterative Retrieval-Generation Synergy. CoRR abs/2305.15294 (2023) - [i76]Yixiao Li, Yifan Yu, Qingru Zhang, Chen Liang, Pengcheng He, Weizhu Chen, Tuo Zhao:
LoSparse: Structured Compression of Large Language Models based on Low-Rank and Sparse Approximation. CoRR abs/2306.11222 (2023) - [i75]Alexander Bukharin, Yixiao Li, Pengcheng He, Weizhu Chen, Tuo Zhao:
Deep Reinforcement Learning from Hierarchical Weak Preference Feedback. CoRR abs/2309.02632 (2023) - 2022
- [j2]Caihong Mu, Weizhu Chen, Yi Liu
, Dongchang Lei, Ruochen Liu:
Virtual information core optimization for collaborative filtering recommendation based on clustering and evolutionary algorithms. Appl. Soft Comput. 116: 108355 (2022) - [c77]Xiaoze Jiang, Yaobo Liang, Weizhu Chen, Nan Duan:
XLM-K: Improving Cross-Lingual Language Model Pre-training with Multilingual Knowledge. AAAI 2022: 10840-10848 - [c76]Zhuocheng Gong, Di He, Yelong Shen, Tie-Yan Liu, Weizhu Chen, Dongyan Zhao, Ji-Rong Wen, Rui Yan:
Finding the Dominant Winning Ticket in Pre-Trained Language Models. ACL (Findings) 2022: 1459-1472 - [c75]Woojeong Jin, Yu Cheng, Yelong Shen, Weizhu Chen, Xiang Ren:
A Good Prompt Is Worth Millions of Parameters: Low-resource Prompt-based Learning for Vision-Language Models. ACL (1) 2022: 2763-2775 - [c74]Jing Qian, Li Dong, Yelong Shen, Furu Wei, Weizhu Chen:
Controllable Natural Language Generation with Contrastive Prefixes. ACL (Findings) 2022: 2912-2924 - [c73]Wei Chen, Yeyun Gong, Song Wang, Bolun Yao, Weizhen Qi, Zhongyu Wei, Xiaowu Hu, Bartuer Zhou, Yi Mao, Weizhu Chen, Biao Cheng, Nan Duan:
DialogVED: A Pre-trained Latent Variable Encoder-Decoder Model for Dialog Response Generation. ACL (1) 2022: 4852-4864 - [c72]Tianyu Liu, Yizhe Zhang, Chris Brockett, Yi Mao, Zhifang Sui, Weizhu Chen, Bill Dolan:
A Token-level Reference-free Hallucination Detection Benchmark for Free-form Text Generation. ACL (1) 2022: 6723-6737 - [c71]Chen Liang, Pengcheng He, Yelong Shen, Weizhu Chen, Tuo Zhao:
CAMERO: Consistency Regularized Ensemble of Perturbed Language Models with Weight Sharing. ACL (1) 2022: 7162-7175 - [c70]Jiachang Liu, Dinghan Shen, Yizhe Zhang, Bill Dolan, Lawrence Carin, Weizhu Chen:
What Makes Good In-Context Examples for GPT-3? DeeLIO@ACL 2022: 100-114 - [c69]Xuxi Chen, Tianlong Chen, Yu Cheng, Weizhu Chen, Ahmed Hassan Awadallah, Zhangyang Wang:
Scalable Learning to Optimize: A Learned Optimizer Can Train Big Models. ECCV (23) 2022: 389-405 - [c68]Xiaonan Li, Daya Guo, Yeyun Gong, Yun Lin, Yelong Shen, Xipeng Qiu, Daxin Jiang, Weizhu Chen, Nan Duan:
Soft-Labeled Contrastive Pre-Training for Function-Level Code Representation. EMNLP (Findings) 2022: 118-129 - [c67]Xinyu Pi, Qian Liu, Bei Chen, Morteza Ziyadi, Zeqi Lin, Qiang Fu, Yan Gao, Jian-Guang Lou, Weizhu Chen:
Reasoning Like Program Executors. EMNLP 2022: 761-779 - [c66]Xiaonan Li, Yeyun Gong, Yelong Shen, Xipeng Qiu, Hang Zhang, Bolun Yao, Weizhen Qi, Daxin Jiang, Weizhu Chen, Nan Duan:
CodeRetriever: A Large Scale Contrastive Pre-Training Method for Code Search. EMNLP 2022: 2898-2910 - [c65]Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen:
LoRA: Low-Rank Adaptation of Large Language Models. ICLR 2022 - [c64]Chen Liang, Haoming Jiang, Simiao Zuo, Pengcheng He, Xiaodong Liu, Jianfeng Gao, Weizhu Chen, Tuo Zhao:
No Parameters Left Behind: Sensitivity Guided Adaptive Learning Rate for Training Large Transformer Models. ICLR 2022 - [c63]Qian Liu, Bei Chen, Jiaqi Guo, Morteza Ziyadi, Zeqi Lin, Weizhu Chen, Jian-Guang Lou:
TAPEX: Table Pre-training via Learning a Neural SQL Executor. ICLR 2022 - [c62]Hang Zhang, Yeyun Gong, Yelong Shen, Jiancheng Lv, Nan Duan, Weizhu Chen:
Adversarial Retriever-Ranker for Dense Text Retrieval. ICLR 2022 - [c61]Qingru Zhang, Simiao Zuo, Chen Liang, Alexander Bukharin, Pengcheng He, Weizhu Chen, Tuo Zhao:
PLATON: Pruning Large Transformer Models with Upper Confidence Bound of Weight Importance. ICML 2022: 26809-26823 - [c60]Daoguang Zan, Bei Chen, Dejian Yang, Zeqi Lin, Minsu Kim, Bei Guan, Yongji Wang, Weizhu Chen, Jian-Guang Lou:
CERT: Continual Pre-training on Sketches for Library-oriented Code Generation. IJCAI 2022: 2369-2375 - [c59]Zhengbao Jiang, Yi Mao, Pengcheng He, Graham Neubig, Weizhu Chen:
OmniTab: Pretraining with Natural and Synthetic Data for Few-shot Table-based Question Answering. NAACL-HLT 2022: 932-942 - [c58]Shujian Zhang, Chengyue Gong, Xingchao Liu, Pengcheng He, Weizhu Chen, Mingyuan Zhou:
ALLSH: Active Learning Guided by Local Sensitivity and Hardness. NAACL-HLT (Findings) 2022: 1328-1342 - [c57]Simiao Zuo, Qingru Zhang, Chen Liang, Pengcheng He, Tuo Zhao, Weizhu Chen:
MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided Adaptation. NAACL-HLT 2022: 1610-1623 - [i74]Xiaonan Li, Yeyun Gong, Yelong Shen, Xipeng Qiu, Hang Zhang, Bolun Yao, Weizhen Qi, Daxin Jiang, Weizhu Chen, Nan Duan:
CodeRetriever: Unimodal and Bimodal Contrastive Learning. CoRR abs/2201.10866 (2022) - [i73]Xinyu Pi, Qian Liu, Bei Chen, Morteza Ziyadi, Zeqi Lin, Yan Gao, Qiang Fu, Jian-Guang Lou, Weizhu Chen:
Reasoning Like Program Executors. CoRR abs/2201.11473 (2022) - [i72]Chen Liang, Haoming Jiang, Simiao Zuo, Pengcheng He, Xiaodong Liu, Jianfeng Gao, Weizhu Chen, Tuo Zhao:
No Parameters Left Behind: Sensitivity Guided Adaptive Learning Rate for Training Large Transformer Models. CoRR abs/2202.02664 (2022) - [i71]Huangjie Zheng, Pengcheng He, Weizhu Chen, Mingyuan Zhou:
Mixing and Shifting: Exploiting Global and Local Dependencies in Vision MLPs. CoRR abs/2202.06510 (2022) - [i70]Huangjie Zheng, Pengcheng He, Weizhu Chen, Mingyuan Zhou:
Truncated Diffusion Probabilistic Models. CoRR abs/2202.09671 (2022) - [i69]Jing Qian, Li Dong, Yelong Shen, Furu Wei, Weizhu Chen:
Controllable Natural Language Generation with Contrastive Prefixes. CoRR abs/2202.13257 (2022) - [i68]Shengnan An, Yifei Li, Zeqi Lin, Qian Liu, Bei Chen, Qiang Fu, Weizhu Chen, Nanning Zheng, Jian-Guang Lou:
Input-Tuning: Adapting Unfamiliar Inputs to Frozen Pretrained Models. CoRR abs/2203.03131 (2022) - [i67]Greg Yang, Edward J. Hu, Igor Babuschkin, Szymon Sidor, Xiaodong Liu, David Farhi, Nick Ryder, Jakub Pachocki, Weizhu Chen, Jianfeng Gao:
Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer. CoRR abs/2203.03466 (2022) - [i66]Chen Liang, Pengcheng He, Yelong Shen, Weizhu Chen, Tuo Zhao:
CAMERO: Consistency Regularized Ensemble of Perturbed Language Models with Weight Sharing. CoRR abs/2204.06625 (2022) - [i65]Simiao Zuo, Qingru Zhang, Chen Liang, Pengcheng He, Tuo Zhao, Weizhu Chen:
MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided Adaptation. CoRR abs/2204.07675 (2022) - [i64]Wei Chen, Yeyun Gong, Song Wang, Bolun Yao, Weizhen Qi, Zhongyu Wei, Xiaowu Hu, Bartuer Zhou, Yi Mao, Weizhu Chen, Biao Cheng, Nan Duan
:
DialogVED: A Pre-trained Latent Variable Encoder-Decoder Model for Dialog Response Generation. CoRR abs/2204.13031 (2022) - [i63]Shujian Zhang, Chengyue Gong, Xingchao Liu, Pengcheng He, Weizhu Chen, Mingyuan Zhou:
ALLSH: Active Learning Guided by Local Sensitivity and Hardness. CoRR abs/2205.04980 (2022) - [i62]Weizhen Qi, Yeyun Gong, Yelong Shen, Jian Jiao, Yu Yan, Houqiang Li, Ruofei Zhang, Weizhu Chen, Nan Duan
:
A Self-Paced Mixed Distillation Method for Non-Autoregressive Generation. CoRR abs/2205.11162 (2022) - [i61]Zhendong Wang, Huangjie Zheng, Pengcheng He, Weizhu Chen, Mingyuan Zhou:
Diffusion-GAN: Training GANs with Diffusion. CoRR abs/2206.02262 (2022) - [i60]Yifei Li, Zeqi Lin, Shizhuo Zhang, Qiang Fu, Bei Chen, Jian-Guang Lou, Weizhu Chen:
On the Advance of Making Language Models Better Reasoners. CoRR abs/2206.02336 (2022) - [i59]Daoguang Zan, Bei Chen, Dejian Yang, Zeqi Lin, Minsu Kim, Bei Guan, Yongji Wang, Weizhu Chen, Jian-Guang Lou:
CERT: Continual Pre-Training on Sketches for Library-Oriented Code Generation. CoRR abs/2206.06888 (2022) - [i58]Qingru Zhang, Simiao Zuo, Chen Liang, Alexander Bukharin, Pengcheng He, Weizhu Chen, Tuo Zhao:
PLATON: Pruning Large Transformer Models with Upper Confidence Bound of Weight Importance. CoRR abs/2206.12562 (2022) - [i57]Weizhou Shen, Yeyun Gong, Yelong Shen, Song Wang, Xiaojun Quan, Nan Duan
, Weizhu Chen:
Joint Generator-Ranker Learning for Natural Language Generation. CoRR abs/2206.13974 (2022) - [i56]Zhengbao Jiang, Yi Mao, Pengcheng He, Graham Neubig, Weizhu Chen:
OmniTab: Pretraining with Natural and Synthetic Data for Few-shot Table-based Question Answering. CoRR abs/2207.03637 (2022) - [i55]Bei Chen, Fengji Zhang, Anh Nguyen, Daoguang Zan, Zeqi Lin, Jian-Guang Lou, Weizhu Chen:
CodeT: Code Generation with Generated Tests. CoRR abs/2207.10397 (2022) - [i54]Chen Liang, Simiao Zuo, Qingru Zhang, Pengcheng He, Weizhu Chen, Tuo Zhao:
Less is More: Task-aware Layer-wise Distillation for Language Model Compression. CoRR abs/2210.01351 (2022) - [i53]Xiaonan Li, Daya Guo, Yeyun Gong, Yun Lin, Yelong Shen, Xipeng Qiu, Daxin Jiang
, Weizhu Chen, Nan Duan
:
Soft-Labeled Contrastive Pre-training for Function-level Code Representation. CoRR abs/2210.09597 (2022) - [i52]Kun Zhou, Yeyun Gong, Xiao Liu, Wayne Xin Zhao, Yelong Shen, Anlei Dong, Jingwen Lu, Rangan Majumder, Ji-Rong Wen, Nan Duan
, Weizhu Chen:
SimANS: Simple Ambiguous Negatives Sampling for Dense Text Retrieval. CoRR abs/2210.11773 (2022) - [i51]Biyang Guo, Yeyun Gong, Yelong Shen, Songqiao Han, Hailiang Huang, Nan Duan
, Weizhu Chen:
GENIUS: Sketch-based Language Model Pre-training via Extreme and Selective Masking for Text Generation and Augmentation. CoRR abs/2211.10330 (2022) - [i50]Jason Phang, Yi Mao, Pengcheng He, Weizhu Chen:
HyperTuning: Toward Adapting Large Language Models without Back-propagation. CoRR abs/2211.12485 (2022) - [i49]Dong Li, Yelong Shen, Ruoming Jin, Yi Mao, Kuan Wang, Weizhu Chen:
Generation-Augmented Query Expansion For Code Retrieval. CoRR abs/2212.10692 (2022) - [i48]Zhenghao Lin
, Yeyun Gong, Yelong Shen, Tong Wu, Zhihao Fan, Chen Lin, Weizhu Chen, Nan Duan
:
GENIE: Large Scale Pre-training for Text Generation with Diffusion Model. CoRR abs/2212.11685 (2022) - 2021
- [c56]Yuning Mao, Pengcheng He, Xiaodong Liu, Yelong Shen, Jianfeng Gao, Jiawei Han, Weizhu Chen:
Reader-Guided Passage Reranking for Open-Domain Question Answering. ACL/IJCNLP (Findings) 2021: 344-350 - [c55]Dayiheng Liu, Yu Yan, Yeyun Gong, Weizhen Qi, Hang Zhang, Jian Jiao, Weizhu Chen, Jie Fu, Linjun Shou, Ming Gong, Pengcheng Wang, Jiusheng Chen, Daxin Jiang, Jiancheng Lv, Ruofei Zhang, Winnie Wu, Ming Zhou, Nan Duan:
GLGE: A New General Language Generation Evaluation Benchmark. ACL/IJCNLP (Findings) 2021: 408-420 - [c54]Hao Cheng, Yelong Shen, Xiaodong Liu, Pengcheng He, Weizhu Chen, Jianfeng Gao:
UnitedQA: A Hybrid Approach for Open Domain Question Answering. ACL/IJCNLP (1) 2021: 3080-3090 - [c53]Yuning Mao, Pengcheng He, Xiaodong Liu, Yelong Shen, Jianfeng Gao, Jiawei Han, Weizhu Chen:
Generation-Augmented Retrieval for Open-Domain Question Answering. ACL/IJCNLP (1) 2021: 4089-4100 - [c52]Yuekai Zhao, Li Dong, Yelong Shen, Zhihua Zhang, Furu Wei, Weizhu Chen:
Memory-Efficient Differentiable Transformer Architecture Search. ACL/IJCNLP (Findings) 2021: 4254-4264 - [c51]Jiaao Chen, Dinghan Shen, Weizhu Chen, Diyi Yang:
HiddenCut: Simple Data Augmentation for Natural Language Understanding with Better Generalizability. ACL/IJCNLP (1) 2021: 4380-4390 - [c50]Chen Liang, Simiao Zuo, Minshuo Chen, Haoming Jiang, Xiaodong Liu, Pengcheng He, Tuo Zhao, Weizhu Chen:
Super Tickets in Pre-Trained Language Models: From Model Compression to Improving Generalization. ACL/IJCNLP (1) 2021: 6524-6538 - [c49]Chen Liang, Haoming Jiang, Xiaodong Liu, Pengcheng He, Weizhu Chen, Jianfeng Gao, Tuo Zhao:
Token-wise Curriculum Learning for Neural Machine Translation. EMNLP (Findings) 2021: 3658-3670 - [c48]Simiao Zuo, Chen Liang, Haoming Jiang, Pengcheng He, Xiaodong Liu, Jianfeng Gao, Weizhu Chen, Tuo Zhao:
ARCH: Efficient Adversarial Regularized Training with Caching. EMNLP (Findings) 2021: 4118-4131 - [c47]Simiao Zuo, Chen Liang, Haoming Jiang, Xiaodong Liu, Pengcheng He, Jianfeng Gao, Weizhu Chen, Tuo Zhao:
Adversarial Regularization as Stackelberg Game: An Unrolled Optimization Approach. EMNLP (1) 2021: 6562-6577 - [c46]Jiaxin Huang, Chunyuan Li, Krishan Subudhi, Damien Jose, Shobana Balakrishnan, Weizhu Chen, Baolin Peng, Jianfeng Gao, Jiawei Han:
Few-Shot Named Entity Recognition: An Empirical Baseline Study. EMNLP (1) 2021: 10408-10423 - [c45]Jungo Kasai, Hao Peng, Yizhe Zhang, Dani Yogatama, Gabriel Ilharco, Nikolaos Pappas, Yi Mao, Weizhu Chen, Noah A. Smith:
Finetuning Pretrained Transformers into RNNs. EMNLP (1) 2021: 10630-10643 - [c44]Pengcheng He, Xiaodong Liu, Jianfeng Gao, Weizhu Chen:
Deberta: decoding-Enhanced Bert with Disentangled Attention. ICLR 2021 - [c43]Kevin J. Liang, Weituo Hao, Dinghan Shen, Yufan Zhou, Weizhu Chen, Changyou Chen, Lawrence Carin:
MixKD: Towards Efficient Distillation of Large-scale Language Models. ICLR 2021 - [c42]Yanru Qu, Dinghan Shen, Yelong Shen, Sandra Sajeev, Weizhu Chen, Jiawei Han:
CoDA: Contrast-enhanced and Diversity-promoting Data Augmentation for Natural Language Understanding. ICLR 2021 - [c41]Weizhen Qi, Yeyun Gong, Jian Jiao, Yu Yan, Weizhu Chen, Dayiheng Liu, Kewen Tang, Houqiang Li, Jiusheng Chen, Ruofei Zhang, Ming Zhou, Nan Duan:
BANG: Bridging Autoregressive and Non-autoregressive Generation with Large Scale Pretraining. ICML 2021: 8630-8639 - [c40]Hang Zhang, Yeyun Gong, Yelong Shen, Weisheng Li, Jiancheng Lv, Nan Duan, Weizhu Chen:
Poolingformer: Long Document Modeling with Pooling Attention. ICML 2021: 12437-12446 - [c39]Sandra Sajeev, Jade Huang, Nikos Karampatziakis, Matthew Hall, Sebastian Kochman, Weizhu Chen:
Contextual Bandit Applications in a Customer Support Bot. KDD 2021: 3522-3530 - [c38]Ge Yang, Edward J. Hu, Igor Babuschkin, Szymon Sidor, Xiaodong Liu, David Farhi, Nick Ryder, Jakub Pachocki, Weizhu Chen, Jianfeng Gao:
Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer. NeurIPS 2021: 17084-17097 - [i47]Sewon Min, Jordan L. Boyd-Graber, Chris Alberti, Danqi Chen, Eunsol Choi, Michael Collins, Kelvin Guu, Hannaneh Hajishirzi, Kenton Lee, Jennimaria Palomaki, Colin Raffel, Adam Roberts, Tom Kwiatkowski, Patrick S. H. Lewis
, Yuxiang Wu, Heinrich Küttler, Linqing Liu, Pasquale Minervini, Pontus Stenetorp, Sebastian Riedel, Sohee Yang, Minjoon Seo, Gautier Izacard, Fabio Petroni, Lucas Hosseini, Nicola De Cao, Edouard Grave, Ikuya Yamada, Sonse Shimaoka, Masatoshi Suzuki, Shumpei Miyawaki, Shun Sato, Ryo Takahashi, Jun Suzuki, Martin Fajcik, Martin Docekal, Karel Ondrej, Pavel Smrz, Hao Cheng, Yelong Shen, Xiaodong Liu, Pengcheng He, Weizhu Chen, Jianfeng Gao, Barlas Oguz, Xilun Chen, Vladimir Karpukhin, Stan Peshterliev, Dmytro Okhonko, Michael Sejr Schlichtkrull, Sonal Gupta, Yashar Mehdad, Wen-tau Yih:
NeurIPS 2020 EfficientQA Competition: Systems, Analyses and Lessons Learned. CoRR abs/2101.00133 (2021) - [i46]Hao Cheng, Yelong Shen, Xiaodong Liu, Pengcheng He, Weizhu Chen, Jianfeng Gao:
UnitedQA: A Hybrid Approach for Open Domain Question Answering. CoRR abs/2101.00178 (2021) - [i45]Yuning Mao, Pengcheng He, Xiaodong Liu, Yelong Shen, Jianfeng Gao, Jiawei Han, Weizhu Chen:
Reader-Guided Passage Reranking for Open-Domain Question Answering. CoRR abs/2101.00294 (2021) - [i44]Jiachang Liu, Dinghan Shen, Yizhe Zhang, Bill Dolan, Lawrence Carin, Weizhu Chen:
What Makes Good In-Context Examples for GPT-3? CoRR abs/2101.06804 (2021) - [i43]Chen Liang, Haoming Jiang, Xiaodong Liu, Pengcheng He, Weizhu Chen, Jianfeng Gao, Tuo Zhao:
Token-wise Curriculum Learning for Neural Machine Translation. CoRR abs/2103.11088 (2021) - [i42]Jungo Kasai, Hao Peng, Yizhe Zhang, Dani Yogatama, Gabriel Ilharco, Nikolaos Pappas, Yi Mao, Weizhu Chen, Noah A. Smith:
Finetuning Pretrained Transformers into RNNs. CoRR abs/2103.13076 (2021) - [i41]Simiao Zuo, Chen Liang, Haoming Jiang, Xiaodong Liu, Pengcheng He, Jianfeng Gao, Weizhu Chen, Tuo Zhao:
Adversarial Training as Stackelberg Game: An Unrolled Optimization Approach. CoRR abs/2104.04886 (2021) - [i40]Tianyu Liu, Yizhe Zhang, Chris Brockett, Yi Mao, Zhifang Sui, Weizhu Chen, Bill Dolan:
A Token-level Reference-free Hallucination Detection Benchmark for Free-form Text Generation. CoRR abs/2104.08704 (2021) - [i39]Hang Zhang, Yeyun Gong, Yelong Shen, Weisheng Li, Jiancheng Lv, Nan Duan, Weizhu Chen:
Poolingformer: Long Document Modeling with Pooling Attention. CoRR abs/2105.04371 (2021) - [i38]Chen Liang, Simiao Zuo, Minshuo Chen, Haoming Jiang, Xiaodong Liu, Pengcheng He, Tuo Zhao, Weizhu Chen:
Super Tickets in Pre-Trained Language Models: From Model Compression to Improving Generalization. CoRR abs/2105.12002 (2021) - [i37]Yuekai Zhao, Li Dong, Yelong Shen, Zhihua Zhang, Furu Wei, Weizhu Chen:
Memory-Efficient Differentiable Transformer Architecture Search. CoRR abs/2105.14669 (2021) - [i36]Jiaao Chen, Dinghan Shen, Weizhu Chen, Diyi Yang:
HiddenCut: Simple Data Augmentation for Natural Language Understanding with Better Generalization. CoRR abs/2106.00149 (2021) - [i35]Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Weizhu Chen:
LoRA: Low-Rank Adaptation of Large Language Models. CoRR abs/2106.09685 (2021) - [i34]Simiao Zuo, Chen Liang, Haoming Jiang, Pengcheng He, Xiaodong Liu, Jianfeng Gao, Weizhu Chen, Tuo Zhao:
ARCH: Efficient Adversarial Regularized Training with Caching. CoRR abs/2109.07048 (2021) - [i33]Xiaoze Jiang, Yaobo Liang, Weizhu Chen, Nan Duan:
XLM-K: Improving Cross-Lingual Language Model Pre-Training with Multilingual Knowledge. CoRR abs/2109.12573 (2021) - [i32]Hang Zhang, Yeyun Gong, Yelong Shen, Jiancheng Lv, Nan Duan, Weizhu Chen:
Adversarial Retriever-Ranker for dense text retrieval. CoRR abs/2110.03611 (2021) - [i31]