Alert button
Picture for Armen Aghajanyan

Armen Aghajanyan

Alert button

DOMINO: A Dual-System for Multi-step Visual Language Reasoning

Add code
Bookmark button
Alert button
Oct 04, 2023
Peifang Wang, Olga Golovneva, Armen Aghajanyan, Xiang Ren, Muhao Chen, Asli Celikyilmaz, Maryam Fazel-Zarandi

Figure 1 for DOMINO: A Dual-System for Multi-step Visual Language Reasoning
Figure 2 for DOMINO: A Dual-System for Multi-step Visual Language Reasoning
Figure 3 for DOMINO: A Dual-System for Multi-step Visual Language Reasoning
Figure 4 for DOMINO: A Dual-System for Multi-step Visual Language Reasoning
Viaarxiv icon

Jointly Training Large Autoregressive Multimodal Models

Add code
Bookmark button
Alert button
Sep 28, 2023
Emanuele Aiello, Lili Yu, Yixin Nie, Armen Aghajanyan, Barlas Oguz

Figure 1 for Jointly Training Large Autoregressive Multimodal Models
Figure 2 for Jointly Training Large Autoregressive Multimodal Models
Figure 3 for Jointly Training Large Autoregressive Multimodal Models
Figure 4 for Jointly Training Large Autoregressive Multimodal Models
Viaarxiv icon

Scaling Autoregressive Multi-Modal Models: Pretraining and Instruction Tuning

Add code
Bookmark button
Alert button
Sep 05, 2023
Lili Yu, Bowen Shi, Ramakanth Pasunuru, Benjamin Muller, Olga Golovneva, Tianlu Wang, Arun Babu, Binh Tang, Brian Karrer, Shelly Sheynin, Candace Ross, Adam Polyak, Russell Howes, Vasu Sharma, Puxin Xu, Hovhannes Tamoyan, Oron Ashual, Uriel Singer, Shang-Wen Li, Susan Zhang, Richard James, Gargi Ghosh, Yaniv Taigman, Maryam Fazel-Zarandi, Asli Celikyilmaz, Luke Zettlemoyer, Armen Aghajanyan

Figure 1 for Scaling Autoregressive Multi-Modal Models: Pretraining and Instruction Tuning
Figure 2 for Scaling Autoregressive Multi-Modal Models: Pretraining and Instruction Tuning
Figure 3 for Scaling Autoregressive Multi-Modal Models: Pretraining and Instruction Tuning
Figure 4 for Scaling Autoregressive Multi-Modal Models: Pretraining and Instruction Tuning
Viaarxiv icon

D4: Improving LLM Pretraining via Document De-Duplication and Diversification

Add code
Bookmark button
Alert button
Aug 23, 2023
Kushal Tirumala, Daniel Simig, Armen Aghajanyan, Ari S. Morcos

Figure 1 for D4: Improving LLM Pretraining via Document De-Duplication and Diversification
Figure 2 for D4: Improving LLM Pretraining via Document De-Duplication and Diversification
Figure 3 for D4: Improving LLM Pretraining via Document De-Duplication and Diversification
Figure 4 for D4: Improving LLM Pretraining via Document De-Duplication and Diversification
Viaarxiv icon

MEGABYTE: Predicting Million-byte Sequences with Multiscale Transformers

Add code
Bookmark button
Alert button
May 19, 2023
Lili Yu, Dániel Simig, Colin Flaherty, Armen Aghajanyan, Luke Zettlemoyer, Mike Lewis

Figure 1 for MEGABYTE: Predicting Million-byte Sequences with Multiscale Transformers
Figure 2 for MEGABYTE: Predicting Million-byte Sequences with Multiscale Transformers
Figure 3 for MEGABYTE: Predicting Million-byte Sequences with Multiscale Transformers
Figure 4 for MEGABYTE: Predicting Million-byte Sequences with Multiscale Transformers
Viaarxiv icon

Scaling Laws for Generative Mixed-Modal Language Models

Add code
Bookmark button
Alert button
Jan 10, 2023
Armen Aghajanyan, Lili Yu, Alexis Conneau, Wei-Ning Hsu, Karen Hambardzumyan, Susan Zhang, Stephen Roller, Naman Goyal, Omer Levy, Luke Zettlemoyer

Figure 1 for Scaling Laws for Generative Mixed-Modal Language Models
Figure 2 for Scaling Laws for Generative Mixed-Modal Language Models
Figure 3 for Scaling Laws for Generative Mixed-Modal Language Models
Figure 4 for Scaling Laws for Generative Mixed-Modal Language Models
Viaarxiv icon

BARTSmiles: Generative Masked Language Models for Molecular Representations

Add code
Bookmark button
Alert button
Nov 29, 2022
Gayane Chilingaryan, Hovhannes Tamoyan, Ani Tevosyan, Nelly Babayan, Lusine Khondkaryan, Karen Hambardzumyan, Zaven Navoyan, Hrant Khachatrian, Armen Aghajanyan

Figure 1 for BARTSmiles: Generative Masked Language Models for Molecular Representations
Figure 2 for BARTSmiles: Generative Masked Language Models for Molecular Representations
Figure 3 for BARTSmiles: Generative Masked Language Models for Molecular Representations
Figure 4 for BARTSmiles: Generative Masked Language Models for Molecular Representations
Viaarxiv icon

Retrieval-Augmented Multimodal Language Modeling

Add code
Bookmark button
Alert button
Nov 22, 2022
Michihiro Yasunaga, Armen Aghajanyan, Weijia Shi, Rich James, Jure Leskovec, Percy Liang, Mike Lewis, Luke Zettlemoyer, Wen-tau Yih

Figure 1 for Retrieval-Augmented Multimodal Language Modeling
Figure 2 for Retrieval-Augmented Multimodal Language Modeling
Figure 3 for Retrieval-Augmented Multimodal Language Modeling
Figure 4 for Retrieval-Augmented Multimodal Language Modeling
Viaarxiv icon

Memorization Without Overfitting: Analyzing the Training Dynamics of Large Language Models

Add code
Bookmark button
Alert button
May 22, 2022
Kushal Tirumala, Aram H. Markosyan, Luke Zettlemoyer, Armen Aghajanyan

Figure 1 for Memorization Without Overfitting: Analyzing the Training Dynamics of Large Language Models
Figure 2 for Memorization Without Overfitting: Analyzing the Training Dynamics of Large Language Models
Figure 3 for Memorization Without Overfitting: Analyzing the Training Dynamics of Large Language Models
Figure 4 for Memorization Without Overfitting: Analyzing the Training Dynamics of Large Language Models
Viaarxiv icon