Picture for Yang You

Yang You

Helen: Optimizing CTR Prediction Models with Frequency-wise Hessian Eigenvalue Regularization

Add code
Feb 23, 2024
Figure 1 for Helen: Optimizing CTR Prediction Models with Frequency-wise Hessian Eigenvalue Regularization
Figure 2 for Helen: Optimizing CTR Prediction Models with Frequency-wise Hessian Eigenvalue Regularization
Figure 3 for Helen: Optimizing CTR Prediction Models with Frequency-wise Hessian Eigenvalue Regularization
Figure 4 for Helen: Optimizing CTR Prediction Models with Frequency-wise Hessian Eigenvalue Regularization
Viaarxiv icon

Neural Network Diffusion

Add code
Feb 20, 2024
Viaarxiv icon

Two Trades is not Baffled: Condensing Graph via Crafting Rational Gradient Matching

Add code
Feb 08, 2024
Viaarxiv icon

Navigating Complexity: Toward Lossless Graph Condensation via Expanding Window Matching

Add code
Feb 07, 2024
Viaarxiv icon

RAP: Retrieval-Augmented Planning with Contextual Memory for Multimodal LLM Agents

Add code
Feb 06, 2024
Figure 1 for RAP: Retrieval-Augmented Planning with Contextual Memory for Multimodal LLM Agents
Figure 2 for RAP: Retrieval-Augmented Planning with Contextual Memory for Multimodal LLM Agents
Figure 3 for RAP: Retrieval-Augmented Planning with Contextual Memory for Multimodal LLM Agents
Figure 4 for RAP: Retrieval-Augmented Planning with Contextual Memory for Multimodal LLM Agents
Viaarxiv icon

GliDe with a CaPE: A Low-Hassle Method to Accelerate Speculative Decoding

Add code
Feb 03, 2024
Figure 1 for GliDe with a CaPE: A Low-Hassle Method to Accelerate Speculative Decoding
Figure 2 for GliDe with a CaPE: A Low-Hassle Method to Accelerate Speculative Decoding
Figure 3 for GliDe with a CaPE: A Low-Hassle Method to Accelerate Speculative Decoding
Figure 4 for GliDe with a CaPE: A Low-Hassle Method to Accelerate Speculative Decoding
Viaarxiv icon

OpenMoE: An Early Effort on Open Mixture-of-Experts Language Models

Add code
Jan 29, 2024
Figure 1 for OpenMoE: An Early Effort on Open Mixture-of-Experts Language Models
Figure 2 for OpenMoE: An Early Effort on Open Mixture-of-Experts Language Models
Figure 3 for OpenMoE: An Early Effort on Open Mixture-of-Experts Language Models
Figure 4 for OpenMoE: An Early Effort on Open Mixture-of-Experts Language Models
Viaarxiv icon

AutoChunk: Automated Activation Chunk for Memory-Efficient Long Sequence Inference

Add code
Jan 19, 2024
Figure 1 for AutoChunk: Automated Activation Chunk for Memory-Efficient Long Sequence Inference
Figure 2 for AutoChunk: Automated Activation Chunk for Memory-Efficient Long Sequence Inference
Figure 3 for AutoChunk: Automated Activation Chunk for Memory-Efficient Long Sequence Inference
Figure 4 for AutoChunk: Automated Activation Chunk for Memory-Efficient Long Sequence Inference
Viaarxiv icon

ProvNeRF: Modeling per Point Provenance in NeRFs as a Stochastic Process

Add code
Jan 18, 2024
Viaarxiv icon

Must: Maximizing Latent Capacity of Spatial Transcriptomics Data

Add code
Jan 15, 2024
Viaarxiv icon