Takip et
Zhengyan Zhang
Zhengyan Zhang
mails.tsinghua.edu.cn üzerinde doğrulanmış e-posta adresine sahip - Ana Sayfa
Başlık
Alıntı yapanlar
Alıntı yapanlar
Yıl
Graph neural networks: A review of methods and applications
J Zhou, G Cui, S Hu, Z Zhang, C Yang, Z Liu, L Wang, C Li, M Sun
AI Open 1, 57-81, 2020
21232020
ERNIE: Enhanced Language Representation with Informative Entities
Z Zhang, X Han, Z Liu, X Jiang, M Sun, Q Liu
arXiv preprint arXiv:1905.07129, 2019
7022019
KEPLER: A unified model for knowledge embedding and pre-trained language representation
X Wang, T Gao, Z Zhu, Z Zhang, Z Liu, J Li, J Tang
arXiv preprint arXiv:1911.06136, 2019
1652019
TransNet: Translation-Based Network Representation Learning for Social Relation Extraction.
C Tu, Z Zhang, Z Liu, M Sun
IJCAI, 2864-2870, 2017
802017
A unified framework for community detection and network representation learning
C Tu, X Zeng, H Wang, Z Zhang, Z Liu, M Sun, B Zhang, L Lin
IEEE Transactions on Knowledge and Data Engineering 31 (6), 1051-1065, 2018
742018
Pre-trained models: Past, present and future
X Han, Z Zhang, N Ding, Y Gu, X Liu, Y Huo, J Qiu, Y Yao, A Zhang, ...
AI Open 2, 225-250, 2021
622021
CPM: A large-scale generative Chinese pre-trained language model
Z Zhang, X Han, H Zhou, P Ke, Y Gu, D Ye, Y Qin, Y Su, H Ji, J Guan, F Qi, ...
AI Open 2, 93-99, 2021
382021
Cpm-2: Large-scale cost-effective pre-trained language models
Z Zhang, Y Gu, X Han, S Chen, C Xiao, Z Sun, Y Yao, F Qi, J Guan, P Ke, ...
AI Open 2, 216-224, 2021
252021
Train No Evil: Selective Masking for Task-guided Pre-training
Y Gu, Z Zhang, X Wang, Z Liu, M Sun
arXiv preprint arXiv:2004.09733, 2020
252020
Cpt: Colorful prompt tuning for pre-trained vision-language models
Y Yao, A Zhang, Z Zhang, Z Liu, TS Chua, M Sun
arXiv preprint arXiv:2109.11797, 2021
232021
Hidden Killer: Invisible Textual Backdoor Attacks with Syntactic Trigger
F Qi, M Li, Y Chen, Z Zhang, Z Liu, Y Wang, M Sun
arXiv preprint arXiv:2105.12400, 2021
232021
Cokebert: Contextual knowledge selection and embedding towards enhanced pre-trained language models
Y Su, X Han, Z Zhang, Y Lin, P Li, Z Liu, J Zhou, M Sun
AI Open 2, 127-134, 2021
142021
Know what you don't need: Single-Shot Meta-Pruning for attention heads
Z Zhang, F Qi, Z Liu, Q Liu, M Sun
AI Open 2, 36-42, 2021
132021
Better Robustness by More Coverage: Adversarial Training with Mixup Augmentation for Robust Fine-tuning
C Si, Z Zhang, F Qi, Z Liu, Y Wang, Q Liu, M Sun
arXiv preprint arXiv:2012.15699, 2020
112020
Knowledge Inheritance for Pre-trained Language Models
Y Qin, Y Lin, J Yi, J Zhang, X Han, Z Zhang, Y Su, Z Liu, P Li, M Sun, ...
arXiv preprint arXiv:2105.13880, 2021
72021
COSINE: Compressive network embedding on large-scale information networks
Z Zhang, C Yang, Z Liu, M Sun, Z Fang, B Zhang, L Lin
IEEE Transactions on Knowledge and Data Engineering, 2020
62020
Red Alarm for Pre-trained Models: Universal Vulnerability to Neuron-Level Backdoor Attacks
Z Zhang, G Xiao, Y Li, T Lv, F Qi, Z Liu, Y Wang, X Jiang, M Sun
arXiv preprint arXiv:2101.06969, 2021
52021
MoEfication: Conditional Computation of Transformer Models for Efficient Inference
Z Zhang, Y Lin, Z Liu, P Li, M Sun, J Zhou
arXiv preprint arXiv:2110.01786, 2021
32021
SHUOWEN-JIEZI: Linguistically Informed Tokenizers For Chinese Language Model Pretraining
C Si, Z Zhang, Y Chen, F Qi, X Wang, Z Liu, M Sun
arXiv preprint arXiv:2106.00400, 2021
32021
CUGE: A Chinese Language Understanding and Generation Evaluation Benchmark
Y Yao, Q Dong, J Guan, B Cao, Z Zhang, C Xiao, X Wang, F Qi, J Bao, ...
arXiv preprint arXiv:2112.13610, 2021
22021
Sistem, işlemi şu anda gerçekleştiremiyor. Daha sonra yeniden deneyin.
Makaleler 1–20