Picture for Jingfeng Wu

Jingfeng Wu

Large Stepsize Gradient Descent for Logistic Loss: Non-Monotonicity of the Loss Improves Optimization Efficiency

Add code
Feb 24, 2024
Figure 1 for Large Stepsize Gradient Descent for Logistic Loss: Non-Monotonicity of the Loss Improves Optimization Efficiency
Figure 2 for Large Stepsize Gradient Descent for Logistic Loss: Non-Monotonicity of the Loss Improves Optimization Efficiency
Figure 3 for Large Stepsize Gradient Descent for Logistic Loss: Non-Monotonicity of the Loss Improves Optimization Efficiency
Viaarxiv icon

In-Context Learning of a Linear Transformer Block: Benefits of the MLP Component and One-Step GD Initialization

Add code
Feb 22, 2024
Figure 1 for In-Context Learning of a Linear Transformer Block: Benefits of the MLP Component and One-Step GD Initialization
Viaarxiv icon

Risk Bounds of Accelerated SGD for Overparameterized Linear Regression

Add code
Nov 23, 2023
Viaarxiv icon

How Many Pretraining Tasks Are Needed for In-Context Learning of Linear Regression?

Add code
Oct 12, 2023
Viaarxiv icon

Private Federated Frequency Estimation: Adapting to the Hardness of the Instance

Add code
Jun 15, 2023
Figure 1 for Private Federated Frequency Estimation: Adapting to the Hardness of the Instance
Figure 2 for Private Federated Frequency Estimation: Adapting to the Hardness of the Instance
Figure 3 for Private Federated Frequency Estimation: Adapting to the Hardness of the Instance
Viaarxiv icon

Implicit Bias of Gradient Descent for Logistic Regression at the Edge of Stability

Add code
May 19, 2023
Viaarxiv icon

Fixed Design Analysis of Regularization-Based Continual Learning

Add code
Mar 17, 2023
Viaarxiv icon

Learning High-Dimensional Single-Neuron ReLU Networks with Finite Samples

Add code
Mar 03, 2023
Viaarxiv icon

The Power and Limitation of Pretraining-Finetuning for Linear Regression under Covariate Shift

Add code
Aug 03, 2022
Figure 1 for The Power and Limitation of Pretraining-Finetuning for Linear Regression under Covariate Shift
Figure 2 for The Power and Limitation of Pretraining-Finetuning for Linear Regression under Covariate Shift
Viaarxiv icon

Risk Bounds of Multi-Pass SGD for Least Squares in the Interpolation Regime

Add code
Mar 07, 2022
Figure 1 for Risk Bounds of Multi-Pass SGD for Least Squares in the Interpolation Regime
Figure 2 for Risk Bounds of Multi-Pass SGD for Least Squares in the Interpolation Regime
Viaarxiv icon