Picture for Armen Aghajanyan

Armen Aghajanyan

SkillRater: Untangling Capabilities in Multimodal Data

Add code
Feb 12, 2026
Viaarxiv icon

Improving MoE Compute Efficiency by Composing Weight and Data Sparsity

Add code
Jan 21, 2026
Viaarxiv icon

When Worse is Better: Navigating the compression-generation tradeoff in visual tokenization

Add code
Dec 20, 2024
Figure 1 for When Worse is Better: Navigating the compression-generation tradeoff in visual tokenization
Figure 2 for When Worse is Better: Navigating the compression-generation tradeoff in visual tokenization
Figure 3 for When Worse is Better: Navigating the compression-generation tradeoff in visual tokenization
Figure 4 for When Worse is Better: Navigating the compression-generation tradeoff in visual tokenization
Viaarxiv icon

MoMa: Efficient Early-Fusion Pre-training with Mixture of Modality-Aware Experts

Add code
Jul 31, 2024
Figure 1 for MoMa: Efficient Early-Fusion Pre-training with Mixture of Modality-Aware Experts
Figure 2 for MoMa: Efficient Early-Fusion Pre-training with Mixture of Modality-Aware Experts
Figure 3 for MoMa: Efficient Early-Fusion Pre-training with Mixture of Modality-Aware Experts
Figure 4 for MoMa: Efficient Early-Fusion Pre-training with Mixture of Modality-Aware Experts
Viaarxiv icon

Small Molecule Optimization with Large Language Models

Add code
Jul 26, 2024
Viaarxiv icon

Text Quality-Based Pruning for Efficient Training of Language Models

Add code
Apr 26, 2024
Figure 1 for Text Quality-Based Pruning for Efficient Training of Language Models
Figure 2 for Text Quality-Based Pruning for Efficient Training of Language Models
Figure 3 for Text Quality-Based Pruning for Efficient Training of Language Models
Figure 4 for Text Quality-Based Pruning for Efficient Training of Language Models
Viaarxiv icon

DOMINO: A Dual-System for Multi-step Visual Language Reasoning

Add code
Oct 04, 2023
Viaarxiv icon

Jointly Training Large Autoregressive Multimodal Models

Add code
Sep 28, 2023
Figure 1 for Jointly Training Large Autoregressive Multimodal Models
Figure 2 for Jointly Training Large Autoregressive Multimodal Models
Figure 3 for Jointly Training Large Autoregressive Multimodal Models
Figure 4 for Jointly Training Large Autoregressive Multimodal Models
Viaarxiv icon

Scaling Autoregressive Multi-Modal Models: Pretraining and Instruction Tuning

Add code
Sep 05, 2023
Figure 1 for Scaling Autoregressive Multi-Modal Models: Pretraining and Instruction Tuning
Figure 2 for Scaling Autoregressive Multi-Modal Models: Pretraining and Instruction Tuning
Figure 3 for Scaling Autoregressive Multi-Modal Models: Pretraining and Instruction Tuning
Figure 4 for Scaling Autoregressive Multi-Modal Models: Pretraining and Instruction Tuning
Viaarxiv icon

D4: Improving LLM Pretraining via Document De-Duplication and Diversification

Add code
Aug 23, 2023
Figure 1 for D4: Improving LLM Pretraining via Document De-Duplication and Diversification
Figure 2 for D4: Improving LLM Pretraining via Document De-Duplication and Diversification
Figure 3 for D4: Improving LLM Pretraining via Document De-Duplication and Diversification
Figure 4 for D4: Improving LLM Pretraining via Document De-Duplication and Diversification
Viaarxiv icon