Picture for Mohit Bansal

Mohit Bansal

Shammie

Vision Transformers are Parameter-Efficient Audio-Visual Learners

Add code
Dec 15, 2022
Figure 1 for Vision Transformers are Parameter-Efficient Audio-Visual Learners
Figure 2 for Vision Transformers are Parameter-Efficient Audio-Visual Learners
Figure 3 for Vision Transformers are Parameter-Efficient Audio-Visual Learners
Figure 4 for Vision Transformers are Parameter-Efficient Audio-Visual Learners
Viaarxiv icon

VindLU: A Recipe for Effective Video-and-Language Pretraining

Add code
Dec 09, 2022
Figure 1 for VindLU: A Recipe for Effective Video-and-Language Pretraining
Figure 2 for VindLU: A Recipe for Effective Video-and-Language Pretraining
Figure 3 for VindLU: A Recipe for Effective Video-and-Language Pretraining
Figure 4 for VindLU: A Recipe for Effective Video-and-Language Pretraining
Viaarxiv icon

Mutual Exclusivity Training and Primitive Augmentation to Induce Compositionality

Add code
Nov 28, 2022
Viaarxiv icon

Perceiver-VL: Efficient Vision-and-Language Modeling with Iterative Latent Attention

Add code
Nov 21, 2022
Viaarxiv icon

Evaluating the Factual Consistency of Large Language Models Through Summarization

Add code
Nov 15, 2022
Figure 1 for Evaluating the Factual Consistency of Large Language Models Through Summarization
Figure 2 for Evaluating the Factual Consistency of Large Language Models Through Summarization
Figure 3 for Evaluating the Factual Consistency of Large Language Models Through Summarization
Figure 4 for Evaluating the Factual Consistency of Large Language Models Through Summarization
Viaarxiv icon

Are Hard Examples also Harder to Explain? A Study with Human and Model-Generated Explanations

Add code
Nov 14, 2022
Viaarxiv icon

Evaluating and Improving Factuality in Multimodal Abstractive Summarization

Add code
Nov 04, 2022
Viaarxiv icon

Exclusive Supermask Subnetwork Training for Continual Learning

Add code
Oct 18, 2022
Figure 1 for Exclusive Supermask Subnetwork Training for Continual Learning
Figure 2 for Exclusive Supermask Subnetwork Training for Continual Learning
Figure 3 for Exclusive Supermask Subnetwork Training for Continual Learning
Figure 4 for Exclusive Supermask Subnetwork Training for Continual Learning
Viaarxiv icon

TVLT: Textless Vision-Language Transformer

Add code
Sep 28, 2022
Figure 1 for TVLT: Textless Vision-Language Transformer
Figure 2 for TVLT: Textless Vision-Language Transformer
Figure 3 for TVLT: Textless Vision-Language Transformer
Figure 4 for TVLT: Textless Vision-Language Transformer
Viaarxiv icon

Summarization Programs: Interpretable Abstractive Summarization with Neural Modular Trees

Add code
Sep 21, 2022
Figure 1 for Summarization Programs: Interpretable Abstractive Summarization with Neural Modular Trees
Figure 2 for Summarization Programs: Interpretable Abstractive Summarization with Neural Modular Trees
Figure 3 for Summarization Programs: Interpretable Abstractive Summarization with Neural Modular Trees
Figure 4 for Summarization Programs: Interpretable Abstractive Summarization with Neural Modular Trees
Viaarxiv icon