Picture for Jingzhao Zhang

Jingzhao Zhang

NeuralDB: Scaling Knowledge Editing in LLMs to 100,000 Facts with Neural KV Database

Add code
Jul 24, 2025
Viaarxiv icon

QuestA: Expanding Reasoning Capacity in LLMs via Question Augmentation

Add code
Jul 17, 2025
Viaarxiv icon

Solving Convex-Concave Problems with $\tilde{\mathcal{O}}(ε^{-4/7})$ Second-Order Oracle Complexity

Add code
Jun 10, 2025
Viaarxiv icon

Data Mixing Can Induce Phase Transitions in Knowledge Acquisition

Add code
May 23, 2025
Viaarxiv icon

Understanding Nonlinear Implicit Bias via Region Counts in Input Space

Add code
May 16, 2025
Viaarxiv icon

Scalable Model Merging with Progressive Layer-wise Distillation

Add code
Feb 18, 2025
Viaarxiv icon

Task Generalization With AutoRegressive Compositional Structure: Can Learning From $\d$ Tasks Generalize to $\d^{T}$ Tasks?

Add code
Feb 13, 2025
Viaarxiv icon

Second-Order Min-Max Optimization with Lazy Hessians

Add code
Oct 12, 2024
Figure 1 for Second-Order Min-Max Optimization with Lazy Hessians
Figure 2 for Second-Order Min-Max Optimization with Lazy Hessians
Figure 3 for Second-Order Min-Max Optimization with Lazy Hessians
Viaarxiv icon

From Sparse Dependence to Sparse Attention: Unveiling How Chain-of-Thought Enhances Transformer Sample Efficiency

Add code
Oct 07, 2024
Figure 1 for From Sparse Dependence to Sparse Attention: Unveiling How Chain-of-Thought Enhances Transformer Sample Efficiency
Figure 2 for From Sparse Dependence to Sparse Attention: Unveiling How Chain-of-Thought Enhances Transformer Sample Efficiency
Figure 3 for From Sparse Dependence to Sparse Attention: Unveiling How Chain-of-Thought Enhances Transformer Sample Efficiency
Figure 4 for From Sparse Dependence to Sparse Attention: Unveiling How Chain-of-Thought Enhances Transformer Sample Efficiency
Viaarxiv icon

Functionally Constrained Algorithm Solves Convex Simple Bilevel Problems

Add code
Sep 10, 2024
Viaarxiv icon