Picture for Barun Patra

Barun Patra

sPhinX: Sample Efficient Multilingual Instruction Fine-Tuning Through N-shot Guided Prompting

Add code
Jul 16, 2024
Figure 1 for sPhinX: Sample Efficient Multilingual Instruction Fine-Tuning Through N-shot Guided Prompting
Figure 2 for sPhinX: Sample Efficient Multilingual Instruction Fine-Tuning Through N-shot Guided Prompting
Figure 3 for sPhinX: Sample Efficient Multilingual Instruction Fine-Tuning Through N-shot Guided Prompting
Figure 4 for sPhinX: Sample Efficient Multilingual Instruction Fine-Tuning Through N-shot Guided Prompting
Viaarxiv icon

Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone

Add code
Apr 23, 2024
Figure 1 for Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone
Figure 2 for Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone
Figure 3 for Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone
Figure 4 for Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone
Viaarxiv icon

A Glitch in the Matrix? Locating and Detecting Language Model Grounding with Fakepedia

Add code
Dec 04, 2023
Viaarxiv icon

Language Is Not All You Need: Aligning Perception with Language Models

Add code
Mar 01, 2023
Figure 1 for Language Is Not All You Need: Aligning Perception with Language Models
Figure 2 for Language Is Not All You Need: Aligning Perception with Language Models
Figure 3 for Language Is Not All You Need: Aligning Perception with Language Models
Figure 4 for Language Is Not All You Need: Aligning Perception with Language Models
Viaarxiv icon

A Length-Extrapolatable Transformer

Add code
Dec 20, 2022
Figure 1 for A Length-Extrapolatable Transformer
Figure 2 for A Length-Extrapolatable Transformer
Figure 3 for A Length-Extrapolatable Transformer
Figure 4 for A Length-Extrapolatable Transformer
Viaarxiv icon

TorchScale: Transformers at Scale

Add code
Nov 23, 2022
Figure 1 for TorchScale: Transformers at Scale
Figure 2 for TorchScale: Transformers at Scale
Figure 3 for TorchScale: Transformers at Scale
Figure 4 for TorchScale: Transformers at Scale
Viaarxiv icon

Beyond English-Centric Bitexts for Better Multilingual Language Representation Learning

Add code
Oct 26, 2022
Figure 1 for Beyond English-Centric Bitexts for Better Multilingual Language Representation Learning
Figure 2 for Beyond English-Centric Bitexts for Better Multilingual Language Representation Learning
Figure 3 for Beyond English-Centric Bitexts for Better Multilingual Language Representation Learning
Figure 4 for Beyond English-Centric Bitexts for Better Multilingual Language Representation Learning
Viaarxiv icon

Foundation Transformers

Add code
Oct 19, 2022
Figure 1 for Foundation Transformers
Figure 2 for Foundation Transformers
Figure 3 for Foundation Transformers
Figure 4 for Foundation Transformers
Viaarxiv icon

Language Model Decoding as Likelihood-Utility Alignment

Add code
Oct 13, 2022
Figure 1 for Language Model Decoding as Likelihood-Utility Alignment
Figure 2 for Language Model Decoding as Likelihood-Utility Alignment
Figure 3 for Language Model Decoding as Likelihood-Utility Alignment
Figure 4 for Language Model Decoding as Likelihood-Utility Alignment
Viaarxiv icon

On the Representation Collapse of Sparse Mixture of Experts

Add code
Apr 20, 2022
Figure 1 for On the Representation Collapse of Sparse Mixture of Experts
Figure 2 for On the Representation Collapse of Sparse Mixture of Experts
Figure 3 for On the Representation Collapse of Sparse Mixture of Experts
Figure 4 for On the Representation Collapse of Sparse Mixture of Experts
Viaarxiv icon