Picture for Binhang Yuan

Binhang Yuan

A Survey of Multimodal Large Language Model from A Data-centric Perspective

Add code
May 26, 2024
Figure 1 for A Survey of Multimodal Large Language Model from A Data-centric Perspective
Figure 2 for A Survey of Multimodal Large Language Model from A Data-centric Perspective
Figure 3 for A Survey of Multimodal Large Language Model from A Data-centric Perspective
Figure 4 for A Survey of Multimodal Large Language Model from A Data-centric Perspective
Viaarxiv icon

Chinese Tiny LLM: Pretraining a Chinese-Centric Large Language Model

Add code
Apr 09, 2024
Figure 1 for Chinese Tiny LLM: Pretraining a Chinese-Centric Large Language Model
Figure 2 for Chinese Tiny LLM: Pretraining a Chinese-Centric Large Language Model
Figure 3 for Chinese Tiny LLM: Pretraining a Chinese-Centric Large Language Model
Figure 4 for Chinese Tiny LLM: Pretraining a Chinese-Centric Large Language Model
Viaarxiv icon

DeFT: Flash Tree-attention with IO-Awareness for Efficient Tree-search-based LLM Inference

Add code
Mar 30, 2024
Figure 1 for DeFT: Flash Tree-attention with IO-Awareness for Efficient Tree-search-based LLM Inference
Figure 2 for DeFT: Flash Tree-attention with IO-Awareness for Efficient Tree-search-based LLM Inference
Figure 3 for DeFT: Flash Tree-attention with IO-Awareness for Efficient Tree-search-based LLM Inference
Figure 4 for DeFT: Flash Tree-attention with IO-Awareness for Efficient Tree-search-based LLM Inference
Viaarxiv icon

Exploring the Robustness of Decentralized Training for Large Language Models

Add code
Dec 01, 2023
Viaarxiv icon

Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time

Add code
Oct 26, 2023
Figure 1 for Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Figure 2 for Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Figure 3 for Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Figure 4 for Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Viaarxiv icon

Serving Deep Learning Model in Relational Databases

Add code
Oct 10, 2023
Figure 1 for Serving Deep Learning Model in Relational Databases
Figure 2 for Serving Deep Learning Model in Relational Databases
Figure 3 for Serving Deep Learning Model in Relational Databases
Figure 4 for Serving Deep Learning Model in Relational Databases
Viaarxiv icon

Auto-Differentiation of Relational Computations for Very Large Scale Machine Learning

Add code
Jun 07, 2023
Figure 1 for Auto-Differentiation of Relational Computations for Very Large Scale Machine Learning
Figure 2 for Auto-Differentiation of Relational Computations for Very Large Scale Machine Learning
Figure 3 for Auto-Differentiation of Relational Computations for Very Large Scale Machine Learning
Figure 4 for Auto-Differentiation of Relational Computations for Very Large Scale Machine Learning
Viaarxiv icon

High-throughput Generative Inference of Large Language Models with a Single GPU

Add code
Mar 13, 2023
Figure 1 for High-throughput Generative Inference of Large Language Models with a Single GPU
Figure 2 for High-throughput Generative Inference of Large Language Models with a Single GPU
Figure 3 for High-throughput Generative Inference of Large Language Models with a Single GPU
Figure 4 for High-throughput Generative Inference of Large Language Models with a Single GPU
Viaarxiv icon

Holistic Evaluation of Language Models

Add code
Nov 16, 2022
Figure 1 for Holistic Evaluation of Language Models
Figure 2 for Holistic Evaluation of Language Models
Figure 3 for Holistic Evaluation of Language Models
Figure 4 for Holistic Evaluation of Language Models
Viaarxiv icon

Stochastic Gradient Descent without Full Data Shuffle

Add code
Jun 12, 2022
Figure 1 for Stochastic Gradient Descent without Full Data Shuffle
Figure 2 for Stochastic Gradient Descent without Full Data Shuffle
Figure 3 for Stochastic Gradient Descent without Full Data Shuffle
Figure 4 for Stochastic Gradient Descent without Full Data Shuffle
Viaarxiv icon