Picture for Clara Na

Clara Na

Energy Considerations of Large Language Model Inference and Efficiency Optimizations

Add code
Apr 24, 2025
Viaarxiv icon

Scalable Data Ablation Approximations for Language Models through Modular Training and Merging

Add code
Oct 21, 2024
Figure 1 for Scalable Data Ablation Approximations for Language Models through Modular Training and Merging
Figure 2 for Scalable Data Ablation Approximations for Language Models through Modular Training and Merging
Figure 3 for Scalable Data Ablation Approximations for Language Models through Modular Training and Merging
Figure 4 for Scalable Data Ablation Approximations for Language Models through Modular Training and Merging
Viaarxiv icon

Energy and Carbon Considerations of Fine-Tuning BERT

Add code
Nov 17, 2023
Viaarxiv icon

To Build Our Future, We Must Know Our Past: Contextualizing Paradigm Shifts in Natural Language Processing

Add code
Oct 11, 2023
Viaarxiv icon

The Framework Tax: Disparities Between Inference Efficiency in Research and Deployment

Add code
Feb 13, 2023
Viaarxiv icon

Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible Models

Add code
May 25, 2022
Figure 1 for Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible Models
Figure 2 for Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible Models
Figure 3 for Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible Models
Figure 4 for Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible Models
Viaarxiv icon