Picture for Alexander Toshev

Alexander Toshev

Apple

DataComp-LM: In search of the next generation of training sets for language models

Add code
Jun 18, 2024
Viaarxiv icon

Grounding Multimodal Large Language Models in Actions

Add code
Jun 12, 2024
Viaarxiv icon

MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training

Add code
Mar 22, 2024
Figure 1 for MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training
Figure 2 for MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training
Figure 3 for MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training
Figure 4 for MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training
Viaarxiv icon

Scalable Pre-training of Large Autoregressive Image Models

Add code
Jan 16, 2024
Figure 1 for Scalable Pre-training of Large Autoregressive Image Models
Figure 2 for Scalable Pre-training of Large Autoregressive Image Models
Figure 3 for Scalable Pre-training of Large Autoregressive Image Models
Figure 4 for Scalable Pre-training of Large Autoregressive Image Models
Viaarxiv icon

Pre-trained Language Models Do Not Help Auto-regressive Text-to-Image Generation

Add code
Nov 27, 2023
Figure 1 for Pre-trained Language Models Do Not Help Auto-regressive Text-to-Image Generation
Figure 2 for Pre-trained Language Models Do Not Help Auto-regressive Text-to-Image Generation
Figure 3 for Pre-trained Language Models Do Not Help Auto-regressive Text-to-Image Generation
Figure 4 for Pre-trained Language Models Do Not Help Auto-regressive Text-to-Image Generation
Viaarxiv icon

Large Language Models as Generalizable Policies for Embodied Tasks

Add code
Oct 26, 2023
Viaarxiv icon

Data Filtering Networks

Add code
Oct 02, 2023
Figure 1 for Data Filtering Networks
Figure 2 for Data Filtering Networks
Figure 3 for Data Filtering Networks
Figure 4 for Data Filtering Networks
Viaarxiv icon

Mobile V-MoEs: Scaling Down Vision Transformers via Sparse Mixture-of-Experts

Add code
Sep 08, 2023
Figure 1 for Mobile V-MoEs: Scaling Down Vision Transformers via Sparse Mixture-of-Experts
Figure 2 for Mobile V-MoEs: Scaling Down Vision Transformers via Sparse Mixture-of-Experts
Figure 3 for Mobile V-MoEs: Scaling Down Vision Transformers via Sparse Mixture-of-Experts
Figure 4 for Mobile V-MoEs: Scaling Down Vision Transformers via Sparse Mixture-of-Experts
Viaarxiv icon

Principles and Guidelines for Evaluating Social Robot Navigation Algorithms

Add code
Jun 29, 2023
Figure 1 for Principles and Guidelines for Evaluating Social Robot Navigation Algorithms
Figure 2 for Principles and Guidelines for Evaluating Social Robot Navigation Algorithms
Figure 3 for Principles and Guidelines for Evaluating Social Robot Navigation Algorithms
Figure 4 for Principles and Guidelines for Evaluating Social Robot Navigation Algorithms
Viaarxiv icon

Value function estimation using conditional diffusion models for control

Add code
Jun 09, 2023
Figure 1 for Value function estimation using conditional diffusion models for control
Figure 2 for Value function estimation using conditional diffusion models for control
Figure 3 for Value function estimation using conditional diffusion models for control
Figure 4 for Value function estimation using conditional diffusion models for control
Viaarxiv icon