+86-15927551363
上海交通大学电院3号楼225
speaker recognition

Biography

Zhengyang Chen received the B.Eng. degree from the Department of Electronic and Information Engineering, Huazhong University of Science and Technology, Wuhan, China, in 2019. He is currently working toward the Ph.D. degree in Shanghai Jiao Tong University, Shanghai, China, under the supervision of Yanmin Qian. His current research interests include speaker recognition, speaker diarization and deep learning.

Education

2019-Present, Shanghai Jiao Tong University, Department of Computer Science and Engineering: Ph.D. Candidate in Computer Science Supervised by Prof. Yanmin Qian 2015-2019, Huazhong University of Science & Technology, Department of Electronic and Information Engineering: B.E in Electronic and Information Engineering

Research Interests

  • Speaker Recognition
  • Speaker Diarization
  • Speech Pre-training

Selected Honors

  • 2022, Winner of CN-Celeb Speaker Recognition Challenge 2022 (track 1)
  • 2022, ICASSP 2022 student travel grant
  • 2021, SdSv Challenge 2021 (3rd place in track1)

Activities

Open-source toolkit

Publications

  • Zhengyang Chen, Sanyuan Chen, Yu Wu, Yao Qian, Chengyi Wang, Shujie Liu, Yanmin Qian, and Michael Zeng, “Large-scale self-supervised speech representation learning for automatic speaker verification,” in ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2022, pp. 6147–6151.
  • Zhengyang Chen, Shuai Wang, and Yanmin Qian, “Self-supervised learning based domain adaptation for robust speaker verification,” in ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2021, pp. 5834–5838.
  • Zhengyang Chen, Shuai Wang, and Yanmin Qian, “Multi-modality matters: A performance leap on voxceleb,” Proc. Interspeech 2020, pp. 2252–2256, 2020.
  • Zhengyang Chen, Shuai Wang, Yanmin Qian, and Kai Yu, “Channel invariant speaker embedding learning with joint multi-task and adversarial training,” in ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2020, pp. 6574–6578.
  • Zhengyang Chen, Shuai Wang, and Yanmin Qian, “Adversarial domain adaptation for speaker verification using partially shared network,” Proc. Interspeech, 2020, pp. 3017–3021, 2020.
  • Sanyuan Chen (Equal Contribution), Chengyi Wang (Equal Contribution), Zhengyang Chen (Equal Contribution), Yu Wu (Equal Contribution), Shujie Liu, Zhuo Chen, Jinyu Li, Naoyuki Kanda, Takuya Yoshioka, Xiong Xiao, Jian Wu, Long Zhou, Shuo Ren, Yanmin Qian, Yao Qian, Jian Wu, Michael Zeng, and Furu Wei, “Wavlm: Large-scale self-supervised pre-training for full stack speech processing,” accepted by IEEE Journal of Selected Topics in Signal Processing.
  • Yanmin Qian, Zhengyang Chen, and Shuai Wang, “Audio-visual deep neural network for robust person verification,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 29, pp. 1079–1092, 2021.
  • Bing Han, Zhengyang Chen, and Yanmin Qian, “Local information modeling with self-attention for speaker verification,” in ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2022, pp. 6727–6731.
  • Bing Han, Zhengyang Chen, Bei Liu, and Yanmin Qian, “Mlp-svnet: A multi-layer perceptrons based network for speaker verification,” in ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2022, pp. 7522–7526.
  • Bei Liu, Haoyu Wang, Zhengyang Chen, Shuai Wang, and Yanmin Qian, “Self-knowledge distillation via feature enhancement for speaker verification,” in ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2022, pp. 7542–7546.
  • Sanyuan Chen, Yu Wu, Chengyi Wang, Zhengyang Chen, Zhuo Chen, Shujie Liu, Jian Wu, Yao Qian, Furu Wei, Jinyu Li, and Xiangzhan Yu, “Unispeech-sat: Universal speech representation learning with speaker aware pre-training,” in ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2022, pp. 6152–6156.
  • Leying Zhang, Zhengyang Chen, and Yanmin Qian, “Knowledge Distillation from Multi-Modality to SingleModality for Person Verification,” in Proc. Interspeech 2021, 2021, pp. 1897–1901.
  • Bing Han, Zhengyang Chen, Zhikai Zhou, and Yanmin Qian, “The SJTU System for Short-Duration Speaker Verification Challenge 2021,” in Proc. Interspeech 2021, 2021, pp. 2332–2336.
  • Xun Gong, Zhengyang Chen, Yexin Yang, Shuai Wang, Lan Wang, and Yanmin Qian, “Speaker embedding augmentation with noise distribution matching,” in 2021 12th International Symposium on Chinese Spoken Language Processing (ISCSLP). IEEE, 2021, pp. 1–5.
  • Yexin Yang, Hongji Wang, Heinrich Dinkel, Zhengyang Chen, Shuai Wang, Yanmin Qian, and Kai Yu, “The sjtu robust anti-spoofing system for the asvspoof 2019 challenge.,” in Interspeech, 2019, pp. 1038–1042.