Zhen Yang    杨珍

I am a PhD student at Knowledge Engineering Group (KEG), Department of Computer Science and Technology of Tsinghua University, fortunately working with Prof. Jie Tang. I obtained my master and bachelor degree from Tsinghua University and Xidian University, respectively.

My current research interests lie in specialized mathematical models, large language models, multi-modal large language models, and graph representation learning. If these areas align with your interests, I welcome you to reach out via email. I am always open to discussing potential collaborations and exploring new ideas together.

Email: yangz21 [at] mails.tsinghua.edu.cn  /  CV_ZH  /  CV_EN  /  Google Scholar  /  Github

profile photo
News
  • August 2024: Checkout our specialized mathematical multi-modal large language model MathGLM-Vision!
  • August 2024: Checkout our multi-modal scientific benchmark VisScience!
  • June 2024: Checkout our GLM Team paper ChatGLM!
  • February 2024: Our survey paper about Negative Sampling is accepted by TPAMI 2024!
  • December 2023: Our paper TriSampler is accepted by AAAI 2024!
  • September 2023: Checkout our specialized mathematical model MathGLM!
  • September 2023: I am currently doing an internship at the ChatGLM group of ZhipuAI!
  • August 2023: Our paper ViLTA gets accepted to ICCV 2023!
  • May 2023: Our paper BatchSampler gets accepted to KDD 2023!
  • February 2022: Our paper RecNS gets accepted to TKDE 2022!
  • January 2022: Our paper STAM gets accepted to WWW 2022!
  • May 2021: One paper MixGCF gets accepted to KDD 2021!
  • May 2020: One paper MCNS gets accepted to KDD 2020!
Publications
fqn MathGLM-Vision: Solving Mathematical Problems with Multi-Modal Large Language Model
Zhen Yang*, Jinhao Chen*, Zhengxiao Du, Wenmeng Yu, Weihan Wang, Wenyi Hong, Zhihuan Jiang, Bin Xu, Yuxiao Dong, Jie Tang
Manuscript, 2024
paper, code

We aim to construct a fine-tuning dataset MathVL, and develop a series of specialized mathematical MLLMs MathGLM-Vision with various parameter-scale backbones.

fqn VisScience: An Extensive Benchmark for Evaluating K12 Educational Multi-modal Scientific Reasoning
Zhihuan Jiang*, Zhen Yang*, Jinhao Chen, Zhengxiao Du, Weihan Wang, Bin Xu, Yuxiao Dong, Jie Tang
Manuscript, 2024
paper, code

We meticulously construct a comprehensive benchmark, named VisScience, which is utilized to assess the multi-modal scientific reasoning across the three disciplines of mathematics, physics, and chemistry.

streamv2v Does Negative Sampling Matter? A Review with Insights into its Theory and Applications
Zhen Yang, Ming Ding, Tinglin Huang, Yukuo Cen, Junshuai Song, Bin Xu, Yuxiao Dong, Jie Tang
TPAMI, 2024
paper

We explore the history of negative sampling, categorize the strategies used to select negative samples, and examine their practical applications.

ovseg TriSampler: A Better Negative Sampling Principle for Dense Retrieval
Zhen Yang, Zhou Shao, Yuxiao Dong, Jie Tang,
AAAI, 2024
paper

We design the quasi-triangular principle and introduce TriSampler to selectively sample more informative negatives within a prescribed constrained region.

ovseg GPT Can Solve Mathematical Problems Without a Calculator
Zhen Yang*, Ming Ding*, Qingsong Lv, Zhihuan Jiang, Zehai He, Yuyi Guo, Jinfei Bai, Jie Tang,
Manuscript, 2023
arxiv, code

We propose a 2 billion-parameter language model MathGLM that can accurately perform multi-digit arithmetic operations with almost 100% accuracy without data leakage.

supmae Batchsampler: Sampling Mini-batches for Contrastive Learning in Vision, Language, and Graphs
Zhen Yang*, Tinglin Huang*, Ming Ding*, Yuxiao Dong, Rex Ying, Yukuo Cen, Yangliao Geng, Jie Tang
KDD, 2023
paper, code

We present BatchSampler to sample mini-batches of hard-to-distinguish (i.e., hard and true negatives to each other) instances by leveraging the proximity graph and a random walk with restart.

declip ViLTA: Enhancing Vision-Language Pre-training through Textual Augmentation
Weihan Wang*, Zhen Yang*, Bin Xu, Juanzi Li, Yankui Sun
ICCV, 2023
paper

We propose a novel method ViLTA that utilize a cross-distillation method to generate soft labels for enhancing the robustness of model.

ant STAM: A Spatiotemporal Aggregation Method for Graph Neural Network-based Recommendation
Zhen Yang, Ming Ding, Bin Xu, Hongxia Yang, Jie Tang
WWW, 2022
paper, code

We propose a spatiotemporal aggregation method STAM to efficiently incorporate temporal information into neighbor embedding learning.

repre Region or Global? A Principle for Negative Sampling in Graph-based Recommendation
Zhen Yang, Ming Ding, Xu Zou, Jie Tang, Bin Xu, Chang Zhou, Hongxia Yang
TKDE, 2022, Long oral
paper, code

We design three region principle to select negative candidate and propose RecNS method to sythesize hard negatives.

crnas MixGCF: An Improved Training Method for Graph Neural Network-based Recommender Systems
Tinglin Huang, Yuxiao Dong, Ming Ding, Zhen Yang, Wenzheng Feng, Xinyu Wang, Jie Tang
KDD, 2021
paper, code

We present MixGCF that can study negative sampling by leveraging both the user-item graph structure and GNNs’ aggregation process to design the hop mixing technique to synthesize hard negatives.

oqa Understanding Negative Sampling in Graph Representation Learning
Zhen Yang*, Ming Ding*, Chang Zhou, Hongxia Yang, Jingren Zhou, Jie Tang,
KDD, 2020
paper, code

We develop a theory and quantify that a nice negative sampling distribution is \( p_n(u|v) \propto p_d(u|v)^\alpha \), \( 0 < \alpha < 1 \). Additionly, we propose Markov chain Monte Carlo Negative Sampling (MCNS), an effective and scalable negative sampling strategy for various tasks in graph representation learning.

Academic Services

  • Reviewer of Journals: TKDE
  • Reviewer of Conferences: KDD 2021/2022, WWW 2023/2024/2025
  • Honors and Awards
    • Huawei Scholarship, 2023.
    • Excellent Graduate of Beijing, 2019.
    • Outstanding Academic Paper Award, Tsinghua University, 2019.
    • 129 Scholarship of Tsinghua University, 2018.
    • National Scholarship by Ministry of Education of China, 2018.
    • National Scholarship by Ministry of Education of China, 2014.

    Thanks to Jon Barron