Picture for Caglar Gulcehre

Caglar Gulcehre

The Effect of Scheduling and Preemption on the Efficiency of LLM Inference Serving

Add code
Nov 12, 2024
Figure 1 for The Effect of Scheduling and Preemption on the Efficiency of LLM Inference Serving
Figure 2 for The Effect of Scheduling and Preemption on the Efficiency of LLM Inference Serving
Figure 3 for The Effect of Scheduling and Preemption on the Efficiency of LLM Inference Serving
Figure 4 for The Effect of Scheduling and Preemption on the Efficiency of LLM Inference Serving
Viaarxiv icon

Beyond Autoregression: Fast LLMs via Self-Distillation Through Time

Add code
Oct 28, 2024
Figure 1 for Beyond Autoregression: Fast LLMs via Self-Distillation Through Time
Figure 2 for Beyond Autoregression: Fast LLMs via Self-Distillation Through Time
Figure 3 for Beyond Autoregression: Fast LLMs via Self-Distillation Through Time
Figure 4 for Beyond Autoregression: Fast LLMs via Self-Distillation Through Time
Viaarxiv icon

Unpacking SDXL Turbo: Interpreting Text-to-Image Models with Sparse Autoencoders

Add code
Oct 28, 2024
Figure 1 for Unpacking SDXL Turbo: Interpreting Text-to-Image Models with Sparse Autoencoders
Figure 2 for Unpacking SDXL Turbo: Interpreting Text-to-Image Models with Sparse Autoencoders
Figure 3 for Unpacking SDXL Turbo: Interpreting Text-to-Image Models with Sparse Autoencoders
Figure 4 for Unpacking SDXL Turbo: Interpreting Text-to-Image Models with Sparse Autoencoders
Viaarxiv icon

SIKeD: Self-guided Iterative Knowledge Distillation for mathematical reasoning

Add code
Oct 24, 2024
Figure 1 for SIKeD: Self-guided Iterative Knowledge Distillation for mathematical reasoning
Figure 2 for SIKeD: Self-guided Iterative Knowledge Distillation for mathematical reasoning
Figure 3 for SIKeD: Self-guided Iterative Knowledge Distillation for mathematical reasoning
Figure 4 for SIKeD: Self-guided Iterative Knowledge Distillation for mathematical reasoning
Viaarxiv icon

The Role of Deep Learning Regularizations on Actors in Offline RL

Add code
Sep 11, 2024
Figure 1 for The Role of Deep Learning Regularizations on Actors in Offline RL
Figure 2 for The Role of Deep Learning Regularizations on Actors in Offline RL
Figure 3 for The Role of Deep Learning Regularizations on Actors in Offline RL
Figure 4 for The Role of Deep Learning Regularizations on Actors in Offline RL
Viaarxiv icon

In Search for Architectures and Loss Functions in Multi-Objective Reinforcement Learning

Add code
Jul 23, 2024
Figure 1 for In Search for Architectures and Loss Functions in Multi-Objective Reinforcement Learning
Figure 2 for In Search for Architectures and Loss Functions in Multi-Objective Reinforcement Learning
Figure 3 for In Search for Architectures and Loss Functions in Multi-Objective Reinforcement Learning
Figure 4 for In Search for Architectures and Loss Functions in Multi-Objective Reinforcement Learning
Viaarxiv icon

Investigating Low-Rank Training in Transformer Language Models: Efficiency and Scaling Analysis

Add code
Jul 13, 2024
Figure 1 for Investigating Low-Rank Training in Transformer Language Models: Efficiency and Scaling Analysis
Figure 2 for Investigating Low-Rank Training in Transformer Language Models: Efficiency and Scaling Analysis
Figure 3 for Investigating Low-Rank Training in Transformer Language Models: Efficiency and Scaling Analysis
Figure 4 for Investigating Low-Rank Training in Transformer Language Models: Efficiency and Scaling Analysis
Viaarxiv icon

HiPPO-Prophecy: State-Space Models can Provably Learn Dynamical Systems in Context

Add code
Jul 12, 2024
Figure 1 for HiPPO-Prophecy: State-Space Models can Provably Learn Dynamical Systems in Context
Figure 2 for HiPPO-Prophecy: State-Space Models can Provably Learn Dynamical Systems in Context
Figure 3 for HiPPO-Prophecy: State-Space Models can Provably Learn Dynamical Systems in Context
Figure 4 for HiPPO-Prophecy: State-Space Models can Provably Learn Dynamical Systems in Context
Viaarxiv icon

Self-Recognition in Language Models

Add code
Jul 09, 2024
Figure 1 for Self-Recognition in Language Models
Figure 2 for Self-Recognition in Language Models
Figure 3 for Self-Recognition in Language Models
Figure 4 for Self-Recognition in Language Models
Viaarxiv icon

Building on Efficient Foundations: Effectively Training LLMs with Structured Feedforward Layers

Add code
Jun 24, 2024
Figure 1 for Building on Efficient Foundations: Effectively Training LLMs with Structured Feedforward Layers
Figure 2 for Building on Efficient Foundations: Effectively Training LLMs with Structured Feedforward Layers
Figure 3 for Building on Efficient Foundations: Effectively Training LLMs with Structured Feedforward Layers
Figure 4 for Building on Efficient Foundations: Effectively Training LLMs with Structured Feedforward Layers
Viaarxiv icon