Picture for Yoon Kim

Yoon Kim

Richard

Log-Linear Attention

Add code
Jun 05, 2025
Figure 1 for Log-Linear Attention
Figure 2 for Log-Linear Attention
Figure 3 for Log-Linear Attention
Figure 4 for Log-Linear Attention
Viaarxiv icon

FlashFormer: Whole-Model Kernels for Efficient Low-Batch Inference

Add code
May 28, 2025
Viaarxiv icon

PaTH Attention: Position Encoding via Accumulating Householder Transformations

Add code
May 22, 2025
Viaarxiv icon

Multimodal LLM Augmented Reasoning for Interpretable Visual Perception Analysis

Add code
Apr 16, 2025
Viaarxiv icon

reWordBench: Benchmarking and Improving the Robustness of Reward Models with Transformed Inputs

Add code
Mar 14, 2025
Figure 1 for reWordBench: Benchmarking and Improving the Robustness of Reward Models with Transformed Inputs
Figure 2 for reWordBench: Benchmarking and Improving the Robustness of Reward Models with Transformed Inputs
Figure 3 for reWordBench: Benchmarking and Improving the Robustness of Reward Models with Transformed Inputs
Figure 4 for reWordBench: Benchmarking and Improving the Robustness of Reward Models with Transformed Inputs
Viaarxiv icon

On the Duality between Gradient Transformations and Adapters

Add code
Feb 19, 2025
Viaarxiv icon

Vision-Language Models Do Not Understand Negation

Add code
Jan 16, 2025
Figure 1 for Vision-Language Models Do Not Understand Negation
Figure 2 for Vision-Language Models Do Not Understand Negation
Figure 3 for Vision-Language Models Do Not Understand Negation
Figure 4 for Vision-Language Models Do Not Understand Negation
Viaarxiv icon

Ladder-residual: parallelism-aware architecture for accelerating large model inference with communication overlapping

Add code
Jan 11, 2025
Figure 1 for Ladder-residual: parallelism-aware architecture for accelerating large model inference with communication overlapping
Figure 2 for Ladder-residual: parallelism-aware architecture for accelerating large model inference with communication overlapping
Figure 3 for Ladder-residual: parallelism-aware architecture for accelerating large model inference with communication overlapping
Figure 4 for Ladder-residual: parallelism-aware architecture for accelerating large model inference with communication overlapping
Viaarxiv icon

The Surprising Effectiveness of Test-Time Training for Abstract Reasoning

Add code
Nov 11, 2024
Figure 1 for The Surprising Effectiveness of Test-Time Training for Abstract Reasoning
Figure 2 for The Surprising Effectiveness of Test-Time Training for Abstract Reasoning
Figure 3 for The Surprising Effectiveness of Test-Time Training for Abstract Reasoning
Figure 4 for The Surprising Effectiveness of Test-Time Training for Abstract Reasoning
Viaarxiv icon

The Semantic Hub Hypothesis: Language Models Share Semantic Representations Across Languages and Modalities

Add code
Nov 07, 2024
Figure 1 for The Semantic Hub Hypothesis: Language Models Share Semantic Representations Across Languages and Modalities
Figure 2 for The Semantic Hub Hypothesis: Language Models Share Semantic Representations Across Languages and Modalities
Figure 3 for The Semantic Hub Hypothesis: Language Models Share Semantic Representations Across Languages and Modalities
Figure 4 for The Semantic Hub Hypothesis: Language Models Share Semantic Representations Across Languages and Modalities
Viaarxiv icon