Picture for Hemanth Saratchandran

Hemanth Saratchandran

Transformers Pretrained on Procedural Data Contain Modular Structures for Algorithmic Reasoning

Add code
May 28, 2025
Viaarxiv icon

Compressing Sine-Activated Low-Rank Adapters through Post-Training Quantization

Add code
May 28, 2025
Viaarxiv icon

Leaner Transformers: More Heads, Less Depth

Add code
May 27, 2025
Viaarxiv icon

Structured Initialization for Vision Transformers

Add code
May 26, 2025
Viaarxiv icon

Enhancing Transformers Through Conditioned Embedded Tokens

Add code
May 19, 2025
Viaarxiv icon

Always Skip Attention

Add code
May 04, 2025
Viaarxiv icon

RandLoRA: Full-rank parameter-efficient fine-tuning of large models

Add code
Feb 03, 2025
Viaarxiv icon

Rethinking Softmax: Self-Attention with Polynomial Activations

Add code
Oct 24, 2024
Viaarxiv icon

Weight Conditioning for Smooth Optimization of Neural Networks

Add code
Sep 05, 2024
Viaarxiv icon

Invertible Neural Warp for NeRF

Add code
Jul 17, 2024
Figure 1 for Invertible Neural Warp for NeRF
Figure 2 for Invertible Neural Warp for NeRF
Figure 3 for Invertible Neural Warp for NeRF
Figure 4 for Invertible Neural Warp for NeRF
Viaarxiv icon