Shaolei Zhang (张绍磊) is currently working toward his Ph.D. degree in the Key Laboratory of Intelligent Information Processing, Institute of Computing Technology, Chinese Academy of Sciences (中国科学院计算技术研究所), advised by Yang Feng (冯洋). He received his bachelor’s degree from Beijing University of Posts and Telecommunications in 2020, majoring in computer science and technology (北京邮电大学计算机科学与技术实验班).

His research interests include nature language processing, machine translation, simultaneous translation and large language model. He has published over 10 papers at the top international AI/NLP conferences such as ACL, EMNLP, NeurIPS, ICLR, AAAI. He won the first place in the streaming transcription track of AutoSimTrans 2021.

🔥 News

  • 2023.12:  🎉 One paper is accepted by ICASSP 2024!
  • 2023.10:  🎉 Two papers are accepted by EMNLP 2023!
  • 2023.09:  👏 Serve as Area Chair of ACL/EACL/NAACL ARR 2023!
  • 2023.09:  🎉 One paper is accepted by NeurIPS 2023!
  • 2023.06:  🎉 Our cross-lingual aligned LLM BayLing is released.
  • 2023.05:  🎉 Two papers are accepted by ACL 2023.
  • 2023.01:  🎉 One paper is accepted by ICLR 2023 (spotlight)!
  • 2022.10:  🎉 Three papers are accepted by EMNLP 2022!
  • 2022.02:  🎉 Three papers are accepted by ACL 2022!

📝 Publications

Preprint 2023
sym
Demo img

BayLing: Bridging Cross-lingual Alignment and Instruction Following through Interactive Translation for Large Language Models
Shaolei Zhang, Qingkai Fang, Zhuocheng Zhang, Zhengrui Ma, Yan Zhou, Langlin Huang, Mengyu Bu, Shangtong Gui, Yunji Chen, Xilin Chen, Yang Feng

paper homepage demo huggingface

  • BayLing (百聆) is a LLM equipped with advanced language alignment.
  • BayLing is the first research to use interactive translation tasks to complete the cross-language transfer of LLM’s generation capabilities.
  • BayLing is selected for inclusion in the 2022-2023 Top 100 Opensource achievements: Open100 (2022-2023), launched by the International Open Benchmark Council (BenchCouncil).
Preprint 2024
sym
Demo img

TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful Space
Shaolei Zhang *, Tian Yu *, Yang Feng

paper huggingface

  • TruthX is an inference-time method to elicit the truthfulness of LLMs by editing their internal representations in truthful space, thereby mitigating the hallucinations of LLMs.
  • TruthX can control LLMs to generate truthful or hallucinatory responses by editing only a vector in truthful space.
  • On TruthfulQA benchmark, TruthX yields an average enhancement of 20% in truthfulness across 13 advanced LLMs.
GitHub Repo
sym

Awesome Simultaneous Translation
Shaolei Zhang

  • A repository that collects the tookits, common datasets and paper list related to the research on Simultaneous Translation, including text-to-text machine translation and speech-to-text translation.

2024

2023

2022

2021

2019

🏆 Honors and Awards

  • [2022] ICT’s Special Scholarship (Xia Peisu Award) (计算所所长特别奖(夏培肃奖), highest award in ICT/CAS)
  • [2022] National Scholarship (国家奖学金)
  • [2021] First place in the streaming transcription track of AutoSimTrans 2021
  • [2020] Beijing Outstanding Graduates Award (北京市优秀毕业生)
  • [2018] Beijing Merit Student (北京市三好学生)
  • [2017] National Scholarship (国家奖学金)

👏 Services

  • Area Chair of ACL/EACL/NAACL ARR 2023
  • Reviewer of ACL/EMNLP/COLING/NAACL/EACL
  • Session Chair of Student Seminar in CCL 2024
  • Session Chair of Student Seminar in YSSNLP 2024
  • 中国中文信息学会青年工作委员会 学生执委会主任
  • Programming Chair of CSSNLP 2020/2021/2023

📖 Educations

💬 Invited Talks

  • “大模型时代的科研选题和实践分享” on MLNLP Academic Seminar [Slides]
  • “跨语言对齐增强大模型——百聆” on AI TIME 大模型嘉年华 [Slides] [Video]
  • “如何在大模型时代找到科研切入点?” on CCMT 2023 [Slides] [Video]
  • “从机器翻译到同声传译:挑战与进展” on MLNLP Academic Seminar [Slides] [Video]
  • AI Time Youth Talk for ICLR 2023 [Video]

💻 Internships

  • 2019.12 - 2021.12, Huawei Noah’s Ark Lab, industry-university-research collaboration project, China.