Picture for Ron Banner

Ron Banner

Workspace Optimization: How to Train Your Agent

Add code
May 10, 2026
Viaarxiv icon

Retrieval from Within: An Intrinsic Capability of Attention-Based Models

Add code
May 07, 2026
Viaarxiv icon

Normalized Architectures are Natively 4-Bit

Add code
May 07, 2026
Viaarxiv icon

FP4 All the Way: Fully Quantized Training of LLMs

Add code
May 25, 2025
Figure 1 for FP4 All the Way: Fully Quantized Training of LLMs
Figure 2 for FP4 All the Way: Fully Quantized Training of LLMs
Figure 3 for FP4 All the Way: Fully Quantized Training of LLMs
Figure 4 for FP4 All the Way: Fully Quantized Training of LLMs
Viaarxiv icon

EXAQ: Exponent Aware Quantization For LLMs Acceleration

Add code
Oct 04, 2024
Figure 1 for EXAQ: Exponent Aware Quantization For LLMs Acceleration
Figure 2 for EXAQ: Exponent Aware Quantization For LLMs Acceleration
Figure 3 for EXAQ: Exponent Aware Quantization For LLMs Acceleration
Figure 4 for EXAQ: Exponent Aware Quantization For LLMs Acceleration
Viaarxiv icon

DropCompute: simple and more robust distributed synchronous training via compute variance reduction

Add code
Jun 18, 2023
Figure 1 for DropCompute: simple and more robust distributed synchronous training via compute variance reduction
Figure 2 for DropCompute: simple and more robust distributed synchronous training via compute variance reduction
Figure 3 for DropCompute: simple and more robust distributed synchronous training via compute variance reduction
Figure 4 for DropCompute: simple and more robust distributed synchronous training via compute variance reduction
Viaarxiv icon

Optimal Fine-Grained N:M sparsity for Activations and Neural Gradients

Add code
Mar 21, 2022
Figure 1 for Optimal Fine-Grained N:M sparsity for Activations and Neural Gradients
Figure 2 for Optimal Fine-Grained N:M sparsity for Activations and Neural Gradients
Figure 3 for Optimal Fine-Grained N:M sparsity for Activations and Neural Gradients
Figure 4 for Optimal Fine-Grained N:M sparsity for Activations and Neural Gradients
Viaarxiv icon

Energy awareness in low precision neural networks

Add code
Feb 06, 2022
Figure 1 for Energy awareness in low precision neural networks
Figure 2 for Energy awareness in low precision neural networks
Figure 3 for Energy awareness in low precision neural networks
Figure 4 for Energy awareness in low precision neural networks
Viaarxiv icon

On Recoverability of Graph Neural Network Representations

Add code
Jan 30, 2022
Figure 1 for On Recoverability of Graph Neural Network Representations
Figure 2 for On Recoverability of Graph Neural Network Representations
Figure 3 for On Recoverability of Graph Neural Network Representations
Figure 4 for On Recoverability of Graph Neural Network Representations
Viaarxiv icon

Logarithmic Unbiased Quantization: Practical 4-bit Training in Deep Learning

Add code
Dec 19, 2021
Figure 1 for Logarithmic Unbiased Quantization: Practical 4-bit Training in Deep Learning
Figure 2 for Logarithmic Unbiased Quantization: Practical 4-bit Training in Deep Learning
Figure 3 for Logarithmic Unbiased Quantization: Practical 4-bit Training in Deep Learning
Figure 4 for Logarithmic Unbiased Quantization: Practical 4-bit Training in Deep Learning
Viaarxiv icon