Picture for Newsha Ardalani

Newsha Ardalani

Demystifying Synthetic Data in LLM Pre-training: A Systematic Study of Scaling Laws, Benefits, and Pitfalls

Add code
Oct 02, 2025
Viaarxiv icon

Quagmires in SFT-RL Post-Training: When High SFT Scores Mislead and What to Use Instead

Add code
Oct 02, 2025
Viaarxiv icon

Carbon Aware Transformers Through Joint Model-Hardware Optimization

Add code
May 02, 2025
Viaarxiv icon

Text Quality-Based Pruning for Efficient Training of Language Models

Add code
Apr 26, 2024
Figure 1 for Text Quality-Based Pruning for Efficient Training of Language Models
Figure 2 for Text Quality-Based Pruning for Efficient Training of Language Models
Figure 3 for Text Quality-Based Pruning for Efficient Training of Language Models
Figure 4 for Text Quality-Based Pruning for Efficient Training of Language Models
Viaarxiv icon

Decoding Data Quality via Synthetic Corruptions: Embedding-guided Pruning of Code Data

Add code
Dec 05, 2023
Viaarxiv icon

Data Acquisition: A New Frontier in Data-centric AI

Add code
Nov 22, 2023
Figure 1 for Data Acquisition: A New Frontier in Data-centric AI
Figure 2 for Data Acquisition: A New Frontier in Data-centric AI
Figure 3 for Data Acquisition: A New Frontier in Data-centric AI
Figure 4 for Data Acquisition: A New Frontier in Data-centric AI
Viaarxiv icon

MAD Max Beyond Single-Node: Enabling Large Machine Learning Model Acceleration on Distributed Systems

Add code
Oct 18, 2023
Viaarxiv icon

SIEVE: Multimodal Dataset Pruning Using Image Captioning Models

Add code
Oct 03, 2023
Viaarxiv icon

Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference

Add code
Mar 10, 2023
Figure 1 for Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference
Figure 2 for Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference
Figure 3 for Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference
Figure 4 for Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference
Viaarxiv icon

MP-Rec: Hardware-Software Co-Design to Enable Multi-Path Recommendation

Add code
Feb 21, 2023
Viaarxiv icon