| CARVIEW |
Select Language
HTTP/2 200
server: GitHub.com
content-type: text/html; charset=utf-8
last-modified: Tue, 11 Jun 2024 05:30:14 GMT
access-control-allow-origin: *
etag: W/"6667e0e6-16539"
expires: Mon, 29 Dec 2025 14:56:56 GMT
cache-control: max-age=600
content-encoding: gzip
x-proxy-cache: MISS
x-github-request-id: DC03:2D8B9D:8F2BE2:A0AB39:69529460
accept-ranges: bytes
age: 0
date: Mon, 29 Dec 2025 14:46:56 GMT
via: 1.1 varnish
x-served-by: cache-bom-vanm7210075-BOM
x-cache: MISS
x-cache-hits: 0
x-timer: S1767019616.256004,VS0,VE203
vary: Accept-Encoding
x-fastly-request-id: 2e5ea514832b90a4bdacfb706bbe2230a6efaf51
content-length: 13683
Zhilin Yang
Zhilin Yang
About
I am working on a startup.
In 2019, I obtained my PhD degree from the School of Computer Science, Carnegie Mellon University, advised by Ruslan Salakhutdinov and William W. Cohen. Prior to that, in 2015, I received my bachelor's degree from Tsinghua University, advised by Jie Tang.
I worked at Meta AI with Jason Weston, and Google Brain with Quoc V. Le.
Reach out to me at A@B, where A=kimi.zhilin.yang and B=gmail.com.
[Google Scholar] [GitHub]
我的名字是杨植麟。
Papers
NLP From Scratch Without Large-Scale Pretraining: A Simple and Efficient Framework
Xingcheng Yao, Yanan Zheng, Xiaocong Yang, Zhilin Yang
ICML 2022
GLM: General Language Model Pretraining with Autoregressive Blank Infilling
Zhengxiao Du, Yujie Qian, Xiao Liu, Ming Ding, Jiezhong Qiu, Zhilin Yang, Jie Tang
ACL 2022
P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally Across Scales and Tasks
Xiao Liu, Kaixuan Ji, Yicheng Fu, Weng Lam Tam, Zhengxiao Du, Zhilin Yang, Jie Tang
ACL 2022
FewNLU: Benchmarking State-of-the-Art Methods for Few-Shot Natural Language Understanding
Yanan Zheng, Jing Zhou, Yujie Qian, Ming Ding, Chonghua Liao, Jian Li, Ruslan Salakhutdinov, Jie Tang, Sebastian Ruder, Zhilin Yang
ACL 2022
Flipda: Effective and Robust Data Augmentation for Few-Shot Learning
Jing Zhou, Yanan Zheng, Jie Tang, Jian Li, Zhilin Yang
ACL 2022
Controllable Generation from Pre-trained Language Models via Inverse Prompting
Xu Zou, Da Yin, Qingyang Zhong, Ming Ding, Hongxia Yang, Zhilin Yang, Jie Tang
KDD 2021
Distribution Matching for Rationalization
Yongfeng Huang, Yujun Chen, Yulun Du, Zhilin Yang
AAAI 2021
XLNet: Generalized Autoregressive Pretraining for Language Understanding
Zhilin Yang*, Zihang Dai*, Yiming Yang, Jaime Carbonell, Ruslan Salakhutdinov, Quoc V. Le
NeurIPS 2019 (*: equal contribution)
Oral, acceptance rate 0.5%
Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context
Zihang Dai*, Zhilin Yang*, Yiming Yang, Jaime Carbonell, Quoc V. Le, Ruslan Salakhutdinov
ACL 2019 (*: equal contribution)
Mixtape: Breaking the Softmax Bottleneck Efficiently
Zhilin Yang, Thang Luong, Ruslan Salakhutdinov, Quoc V. Le
NeurIPS 2019
[PDF]
HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering
Zhilin Yang*, Peng Qi*, Saizheng Zhang*, Yoshua Bengio, William W. Cohen, Ruslan Salakhutdinov, Christopher D. Manning
EMNLP 2018 (*: equal contribution)
GLoMo: Unsupervised Learning of Transferable Relational Graphs
Zhilin Yang*, Jake Zhao*, Bhuwan Dhingra, Kaiming He, William W. Cohen, Ruslan Salakhutdinov, Yann LeCun
NIPS 2018 (*: equal contribution)
Neural Cross-Lingual Named Entity Recognition with Minimal Resources
Jiateng Xie, Zhilin Yang, Graham Neubig, Noah A. Smith, Jaime Carbonell
EMNLP 2018
[PDF]
Neural Models for Reasoning over Multiple Mentions using Coreference
Bhuwan Dhingra, Qiao Jin, Zhilin Yang, William W. Cohen, Ruslan Salakhutdinov
NAACL 2018, short paper
[PDF] [Code and Data]
Breaking the Softmax Bottleneck: A High-Rank RNN Language Model
Zhilin Yang*, Zihang Dai*, Ruslan Salakhutdinov, William W. Cohen
ICLR 2018 (*: equal contribution)
Oral, acceptance rate 2%
Mastering the Dungeon: Grounded Language Learning by Mechanical Turker Descent
Zhilin Yang, Saizheng Zhang, Jack Urbanek, Will Feng, Alexander H. Miller, Arthur Szlam, Douwe Kiela, Jason Weston
ICLR 2018
[PDF]
Good Semi-supervised Learning that Requires a Bad GAN
Zihang Dai*, Zhilin Yang*, Fan Yang, William W. Cohen, Ruslan Salakhutdinov
NIPS 2017 (*: equal contribution)
Differentiable Learning of Logical Rules for Knowledge Base Reasoning
Fan Yang, Zhilin Yang, William W. Cohen
NIPS 2017
Linguistic Knowledge as Memory for Recurrent Neural Networks
Bhuwan Dhingra, Zhilin Yang, William W. Cohen, Ruslan Salakhutdinov
Preprint 2017
[PDF]
Semi-Supervised QA with Generative Domain-Adaptive Nets
Zhilin Yang, Junjie Hu, Ruslan Salakhutdinov, William W. Cohen
ACL 2017
Gated-Attention Readers for Text Comprehension
Bhuwan Dhingra, Hanxiao Liu, Zhilin Yang, William W. Cohen, Ruslan Salakhutdinov
ACL 2017
Transfer Learning for Sequence Tagging with Hierarchical Recurrent Networks
Zhilin Yang, Ruslan Salakhutdinov, William W. Cohen
ICLR 2017
Words or Characters? Fine-grained Gating for Reading Comprehension
Zhilin Yang, Bhuwan Dhingra, Ye Yuan, Junjie Hu, William W. Cohen, Ruslan Salakhutdinov
ICLR 2017
Review Networks for Caption Generation
Zhilin Yang, Ye Yuan, Yuexin Wu, Ruslan Salakhutdinov, William W. Cohen
NIPS 2016
[PDF] [Code and data]
Multi-Task Cross-Lingual Sequence Tagging from Scratch
Zhilin Yang, Ruslan Salakhutdinov, William Cohen
Preprint 2016
[PDF]
Revisiting Semi-Supervised Learning with Graph Embeddings
Zhilin Yang, William Cohen, Ruslan Salakhutdinov
ICML 2016
[PDF] [Code and data]
Multi-Modal Bayesian Embeddings for Learning Social Knowledge Graphs
Zhilin Yang, Jie Tang, William Cohen
IJCAI 2016
Our model is deployed on AMiner for extracting research interests.
Collaborative Embedding Features and Diversified Ensemble for E-Commerce Repeat Buyer Prediction
Zhanpeng Fang*, Zhilin Yang*, Yutao Zhang (* indicates equal contribution)
IJCAI Workshop 2015
Invited paper, competition winners.
COSNET: Connecting Heterogeneous Social Networks with Local and Global Consistency
Yutao Zhang, Jie Tang, Zhilin Yang, Jian Pei, Philip Yu
KDD 2015
Oral presentation, acceptance rate 19%.
Active Learning for Streaming Networked Data
Zhilin Yang, Jie Tang, Yutao Zhang
CIKM 2014
Full paper, acceptance rate 21%.
[PDF]
Active Learning for Networked Data Based on Non-Progressive Diffusion Model
Zhilin Yang, Jie Tang, Bin Xu, Chunxiao Xing
WSDM 2014
Full-length oral presentation (5%), acceptance rate 18%.
[PDF]
SAE: Social Analytic Engine for Large Networks
Yang Yang, Jianfei Wang, Yutao Zhang, Wei Chen, Jing Zhang, Honglei Zhuang, Zhilin Yang, Bo Ma, Zhanpeng Fang, Sen Wu, Xiaoxiao Li, Debing Liu, Jie Tang
KDD Demo 2013
[PDF]