Jingjing Zheng (郑晢晢)

"Theory without practice is empty, but equally, practice without theory is blind." — I. Kant

Hello! I'm Jingjing Zheng (she/her), a Ph.D. student in Mathematics at the University of British Columbia, supervised by Prof. Yankai Cao. My research interests include efficient training/inference of large models grounded in theory, low-rank/tensor methods, and sparse representation learning. My academic background spans art and design (B.A.), mathematics (M.S. and current Ph.D.), and computer science (completed Ph.D. degree).

I received the Borealis AI Fellowship (awarded to ten AI researchers across Canada) and the Government Award for Outstanding Self-financed Students Abroad (awarded to 650 outstanding young talents worldwide).

🌈 I am committed to supporting LGBTQ+ visibility, inclusion, and diversity within academia and STEM communities.

Email  /  CV  /  Scholar  /  GitHub  /  LinkedIn

profile photo
News Research Experience Awards Community Service Presentations Teaching

Recent News

[2026] This summer, I will have a short-term visit to Prof. Qibin Zhao's group at RIKEN.
[2026] One paper accepted to CVPR 2026 (see you in Denver!).
[2025] Joined the Organizing Committee of Women and Gender-diverse Mathematicians at UBC (WGM).
[2025–2026] Appointed to the UBC Green College Academic Committee.
[2025] Our startup GradientX was selected for the Lab2Market Validate Program (Funded).
[2025] Two papers accepted to NeurIPS 2025 (see you in San Diego!).

Selected Research

My research focuses on efficient training/inference of large models, low-rank/sparse representation learning, and tensor optimization. For the full list, please visit Google Scholar.

* corresponding author    † supervisor

ReFTA ReFTA: Breaking the Weight Reconstruction Bottleneck in Tensorized Parameter-Efficient Fine-Tuning
Jingjing Zheng, Anda Tang, Qiangqiang Mao, Zhouchen Lin*, Yankai Cao*,†
CVPR, 2026

Breaks the weight reconstruction bottleneck in tensorized parameter-efficient fine-tuning for large models.

Tensorized PEFT methods represent weight updates in compact tensor formats for high parameter efficiency, but incur significant overhead at inference due to weight reconstruction. ReFTA proposes a reconstruction-free tensorized adaptation framework that eliminates this bottleneck, achieving both parameter efficiency during fine-tuning and computational efficiency at inference, with strong performance on vision and language benchmarks.

arXiv 2026
On Catastrophic Forgetting in Low-Rank Decomposition-Based Parameter-Efficient Fine-Tuning
Muhammad Ahmad, Jingjing Zheng, Yankai Cao*,†
arXiv:2603.09684, 2026
paper

Investigates catastrophic forgetting in low-rank decomposition-based PEFT methods and proposes mitigation strategies for continual adaptation of large models.

Low-rank decomposition-based PEFT methods such as LoRA are widely adopted for efficient adaptation, but exhibit catastrophic forgetting in continual learning settings. This paper systematically studies the forgetting phenomenon across representative low-rank PEFT methods, identifies its root causes in the low-rank constraint, and proposes mitigation strategies that preserve prior knowledge while enabling effective sequential adaptation of large pre-trained models.

AdaMSS AdaMSS: Adaptive Multi-Subspace Approach for Parameter-Efficient Fine-Tuning
Jingjing Zheng, Wanglong Lu, Yiming Dong, Chaojie Ji, Yankai Cao*,†, Zhouchen Lin*
NeurIPS, 2025
paper

Leverages subspace segmentation to adaptively reduce trainable parameters, achieving better generalization than LoRA/PiSSA while using far fewer parameters.

AdaMSS proposes an adaptive multi-subspace approach for parameter-efficient fine-tuning of large models. Unlike traditional PEFT methods that operate in a single large subspace, AdaMSS applies subspace segmentation to obtain multiple smaller subspaces and adaptively reduces trainable parameters during training, ultimately updating only those associated with the subspaces most relevant to the target task. Theoretical analyses show better generalization guarantees than LoRA and PiSSA. On ViT-Large, AdaMSS achieves 4.7% higher average accuracy than LoRA across seven tasks using only 15.4% of LoRA’s trainable parameters. On RoBERTa-Large, it outperforms PiSSA by 7% in average accuracy across six tasks while reducing trainable parameters by ~94.4%.

ReLU+Argmin Differentiable Decision Tree via “ReLU+Argmin” Reformulation
Qiangqiang Mao, Jiayang Ren, Yixiu Wang, Chenxuanyin Zou, Jingjing Zheng, Yankai Cao*,†
NeurIPS, 2025  (Spotlight)
paper

Proposes a differentiable decision tree reformulation using ReLU+Argmin for end-to-end training.

Decision trees are inherently non-differentiable due to hard branching decisions at each node, preventing end-to-end gradient-based training. This paper introduces a “ReLU+Argmin” reformulation that makes the entire tree fully differentiable, enabling joint optimization with neural network components via standard backpropagation. The approach achieves strong performance on tabular data benchmarks while preserving the interpretability of tree-structured models.

arXiv 2024
Adaptive Principal Components Allocation with the ℓ2,g-regularized Gaussian Graphical Model for Efficient Fine-Tuning Large Models
Jingjing Zheng, Yankai Cao*,†
arXiv:2412.08592, 2024
paper

Proposes adaptive principal component allocation guided by an ℓ2,g-regularized Gaussian graphical model to improve efficiency in fine-tuning large language models.

Standard PEFT methods assign equal rank to all layers, wasting capacity on already well-adapted layers and under-fitting others. This work models inter-layer dependencies of weight matrices using an ℓ2,g-regularized Gaussian graphical model and uses the inferred structure to guide adaptive principal component allocation across layers, achieving better efficiency-accuracy trade-offs in fine-tuning large language models.

Multi-Objective Tensor Recovery Handling The Non-Smooth Challenge in Tensor SVD: A Multi-Objective Tensor Recovery Framework
Jingjing Zheng, Wanglong Lu, Wenzhe Wang, Yankai Cao*,†, Xiaoqin Zhang†, Xianta Jiang†
ECCV, 2024
paper

Addresses non-smooth tensor data recovery via a learnable tensor nuclear norm and the proposed APMM optimization algorithm with KKT convergence guarantees.

Tensor SVD (t-SVD)-based recovery methods show strong results on visual data such as color images and videos, but suffer severe performance degradation when tensor data exhibits non-smooth changes—a common real-world phenomenon largely ignored by prior work. This paper introduces a novel tensor recovery model with a learnable tensor nuclear norm to address this challenge, and develops the Alternating Proximal Multiplier Method (APMM) to iteratively solve the proposed model. Theoretical analysis proves convergence of APMM to the KKT point of the optimization problem. A multi-objective tensor recovery framework is further proposed based on APMM to efficiently exploit correlations across tensor dimensions, extending t-SVD-based methods to higher-order tensor cases. Numerical experiments on tensor completion demonstrate the effectiveness of the approach.

Bayesian Tensor Norm Bayesian-Driven Learning of A New Weighted Tensor Norm for Tensor Recovery
Jingjing Zheng, Yankai Cao*,†
ICLR, 2024 (Tiny Papers Track)
paper / code

Learns a data-dependent weighted tensor norm via Bayesian Optimization within a bilevel tensor completion framework, addressing non-smooth changes and imbalanced low-rankness.

t-SVD-based tensor recovery methods suffer from performance limitations caused by non-smooth changes and imbalanced low-rankness in tensor data. This work introduces a novel bilevel tensor completion model that integrates the learning of a data-dependent weighted tensor norm as an upper-level problem within the tensor completion framework. The bilevel optimization is treated as a black-box problem, and Bayesian Optimization (BO) is employed for efficient learning of the proposed tensor norm. Numerical experiments demonstrate superior performance compared to state-of-the-art tensor completion methods.

Structured Sparsity Structured Sparsity Optimization with Non-Convex Surrogates of ℓ2,0-Norm: A Unified Algorithmic Framework
Xiaoqin Zhang*,†, Jingjing Zheng, Di Wang, Guiying Tang, Zhengyuan Zhou, Zhouchen Lin
IEEE TPAMI, 2023 (IF: 20.8)
paper / code

A unified algorithmic framework for structured sparsity optimization with non-convex surrogates.

Enforcing structured group sparsity—crucial for network pruning and feature selection—requires minimizing the ℓ2,0-norm, which is NP-hard. This paper proposes a unified framework using non-convex surrogate functions, deriving a family of proximal algorithms with global convergence guarantees. The framework subsumes several prior methods as special cases and achieves better sparsity-accuracy trade-offs on diverse tasks including neural network compression and multi-task learning.

IEEE TNNLS 2022
Tensor Recovery With Weighted Tensor Average Rank
Xiaoqin Zhang*,†, Jingjing Zheng, Li Zhao, Zhengyuan Zhou, Zhouchen Lin
IEEE Transactions on Neural Networks and Learning Systems, 2022 (IF: 11.1)
paper / code

Proposes Weighted Tensor Average Rank (WTAR) to capture relationships between differently-transposed tensors, with convex/non-convex surrogates and GTSVT solver for robust tensor recovery.

This paper investigates a curious phenomenon in tensor recovery: do different transpose operations on observation tensors yield identical recovered results? If not, information within the data may be lost under certain transpose operators. To address this, a new tensor rank called Weighted Tensor Average Rank (WTAR) is proposed to learn relationships between tensors obtained via a series of transpose operators. WTAR is applied to three-order tensor robust PCA (TRPCA) to validate its effectiveness. To balance effectiveness and solvability, both convex and non-convex surrogates of the model are studied, with corresponding worst-case error bounds derived. A generalized tensor singular value thresholding (GTSVT) method and an associated optimization algorithm are proposed to efficiently solve the generalized model.

AAAI 2022
Handling Slice Permutations Variability in Tensor Recovery
Jingjing Zheng, Xiaoqin Zhang*,†, Wenzhe Wang, Xianta Jiang†
AAAI, 2022
paper / supp / video / poster

Addresses slice permutation variability in tensor recovery with a novel optimization approach.

Tensor SVD (t-SVD) based recovery methods are sensitive to the ordering of frontal slices, yet this variability is rarely addressed in practice. This paper analyzes how slice permutations affect the t-SVD rank structure and recovery quality, and proposes an optimization framework that jointly estimates the latent tensor and its optimal slice ordering. Experiments on hyperspectral image recovery and video completion demonstrate consistent improvements by accounting for slice permutation variability.

Experience

  • RIKEN, Japan — Visiting Student (Summer 2026, upcoming)
    Upcoming short-term visit to Prof. Qibin Zhao's group at RIKEN AIP, Japan, focused on tensor decomposition and efficient learning.
  • ZERO Lab, Peking University — Visiting Student (05/2024 – 09/2024)
    Advised by Prof. Zhouchen Lin. Developed AdaMSS (NeurIPS 2025): adaptive multi-subspace PEFT that improves average accuracy by +4.7% on ViT-Large across 7 image classification datasets and +6.9% on GSM8K with LLaMA 2-7B, requiring only 15% and 1.25% of LoRA’s trainable parameters, respectively.
  • Nasdaq (Verafin) — Research Intern (05/2022 – 09/2022)
    Advised by John Hawkin. Developed an unsupervised financial fraud detection system using low-rank recovery (Outlier Pursuit and non-convex variants). Published at Canadian AI 2023; supported by Mitacs Accelerate Award (CAD $15,000).
  • GradientX Technologies Inc. — Co-Founder (2025 – Present)
    Co-founded a Vancouver-based startup building the next generation of personalized financial intelligence. Selected for the Lab2Market Validate Program (2025, funded at $10,000).

Selected Awards & Grants

Awards & Honors

  • BPOC Graduate Excellence Award, 2024
  • The Borealis AI Fellowship (awarded to ten AI researchers from across Canada), 2023
  • Government Award for Outstanding Self-financed Students Abroad (globally awarded to 650 young talents every year), 2023
  • Fellow of the School of Graduate Studies, 2023
  • MUN Outstanding Research Award, 2022
  • National Scholarship, China, 2019
  • Outstanding Graduates of Zhejiang Province, China, 2019
  • CISC Outstanding Paper Award, China, 2018
  • National Post-Graduate Mathematical Contest in Modeling, China (Second Prize, Team Leader), 2017

Grants

  • Mitacs Business Strategy Internship (BSI), 2025, fund: CAD $10,000
  • Mitacs Accelerate Award with Verafin, Unsupervised Financial Fraud Detection Using Low-rank Recovery, CAD $15,000, 2022.5–2022.9
  • Science and Technology Innovation Program for College Students in Zhejiang Province, Image Classification Based on New Norm and Its Generalization, Jingjing Zheng (Principal Investigator), Xiaoju Lu, Guiying Tang, 2018–2020, fund: RMB ¥10,000

Community Service

  • Women and Gender-diverse Mathematicians at UBC (WGM), Organizing Committee, 2025–2026
  • UBC Green College Academic Committee, 2025–2026
  • UBC Applied Mathematics Meeting, Organizer, 2025
  • Reviewer: AAAI, WACV, ICLR, CVPR, ICCV, NeurIPS, Canadian AI; IEEE Transactions on Industrial Informatics
  • Mentoring: Suleman Ahmad (Engineering, UBC, 08/2025–Present); Wenzhe Wang, Zhiwei Huang, Xixiang Chen (Wenzhou University, 2019–2023); Mengqing Sun (Wenzhou University, 2018–2021)

Presentations

  1. UBC Math Graduate and Postdoc Seminar, Vancouver, Nov. 20, 2025
  2. CAN-CWiC West 2025 (poster), Vancouver, Nov. 7, 2025
  3. UBC-hosted event collocated with ICML 2025 (poster), Vancouver, Jul. 15, 2025
  4. The 18th European Conference on Computer Vision (ECCV) (poster), MiCo Milano, 2024
  5. Canadian Conference on Artificial Intelligence (oral + poster), Montreal, 2023
  6. AAAI Conference on Artificial Intelligence (poster), Vancouver (remote), 2022
  7. The First Annual SEA Conference (poster), Newfoundland, 2022
  8. AARMS CRG Workshop (oral), Newfoundland, Jun. 2, 2022

Teaching

Teaching Assistant

  1. MATH 340: Introduction to Linear Programming, UBC, 2025 Winter Term 1
  2. MATH 340: Introduction to Linear Programming, UBC, 2024 Winter Term 2
  3. MATH 340: Introduction to Linear Programming, UBC, 2024 Winter Term 1
  4. Abstract Linear Algebra, UBC, 2023 Winter Term 2
  5. Matrix Algebra, UBC, 2023 Winter Term 1
  6. Math Learning Center, UBC, 2023–2025
  7. Computer Science 2002: Data Structures and Algorithms, Memorial University of Newfoundland, Winter 2022

Design and source code from Jon Barron's website.