Picture for Leonid Karlinsky

Leonid Karlinsky

Comparison Visual Instruction Tuning

Add code
Jun 13, 2024
Figure 1 for Comparison Visual Instruction Tuning
Figure 2 for Comparison Visual Instruction Tuning
Figure 3 for Comparison Visual Instruction Tuning
Figure 4 for Comparison Visual Instruction Tuning
Viaarxiv icon

ConMe: Rethinking Evaluation of Compositional Reasoning for Modern VLMs

Add code
Jun 12, 2024
Figure 1 for ConMe: Rethinking Evaluation of Compositional Reasoning for Modern VLMs
Figure 2 for ConMe: Rethinking Evaluation of Compositional Reasoning for Modern VLMs
Figure 3 for ConMe: Rethinking Evaluation of Compositional Reasoning for Modern VLMs
Figure 4 for ConMe: Rethinking Evaluation of Compositional Reasoning for Modern VLMs
Viaarxiv icon

$\textit{Trans-LoRA}$: towards data-free Transferable Parameter Efficient Finetuning

Add code
May 27, 2024
Figure 1 for $\textit{Trans-LoRA}$: towards data-free Transferable Parameter Efficient Finetuning
Figure 2 for $\textit{Trans-LoRA}$: towards data-free Transferable Parameter Efficient Finetuning
Figure 3 for $\textit{Trans-LoRA}$: towards data-free Transferable Parameter Efficient Finetuning
Figure 4 for $\textit{Trans-LoRA}$: towards data-free Transferable Parameter Efficient Finetuning
Viaarxiv icon

Adaptive Memory Replay for Continual Learning

Add code
Apr 18, 2024
Viaarxiv icon

NumeroLogic: Number Encoding for Enhanced LLMs' Numerical Reasoning

Add code
Mar 30, 2024
Figure 1 for NumeroLogic: Number Encoding for Enhanced LLMs' Numerical Reasoning
Figure 2 for NumeroLogic: Number Encoding for Enhanced LLMs' Numerical Reasoning
Figure 3 for NumeroLogic: Number Encoding for Enhanced LLMs' Numerical Reasoning
Figure 4 for NumeroLogic: Number Encoding for Enhanced LLMs' Numerical Reasoning
Viaarxiv icon

Meta-Prompting for Automating Zero-shot Visual Recognition with LLMs

Add code
Mar 19, 2024
Viaarxiv icon

Towards Multimodal In-Context Learning for Vision & Language Models

Add code
Mar 19, 2024
Figure 1 for Towards Multimodal In-Context Learning for Vision & Language Models
Figure 2 for Towards Multimodal In-Context Learning for Vision & Language Models
Figure 3 for Towards Multimodal In-Context Learning for Vision & Language Models
Figure 4 for Towards Multimodal In-Context Learning for Vision & Language Models
Viaarxiv icon

CAMELoT: Towards Large Language Models with Training-Free Consolidated Associative Memory

Add code
Feb 21, 2024
Figure 1 for CAMELoT: Towards Large Language Models with Training-Free Consolidated Associative Memory
Figure 2 for CAMELoT: Towards Large Language Models with Training-Free Consolidated Associative Memory
Figure 3 for CAMELoT: Towards Large Language Models with Training-Free Consolidated Associative Memory
Figure 4 for CAMELoT: Towards Large Language Models with Training-Free Consolidated Associative Memory
Viaarxiv icon

3VL: using Trees to teach Vision & Language models compositional concepts

Add code
Dec 28, 2023
Viaarxiv icon

Learning Human Action Recognition Representations Without Real Humans

Add code
Nov 10, 2023
Figure 1 for Learning Human Action Recognition Representations Without Real Humans
Figure 2 for Learning Human Action Recognition Representations Without Real Humans
Figure 3 for Learning Human Action Recognition Representations Without Real Humans
Figure 4 for Learning Human Action Recognition Representations Without Real Humans
Viaarxiv icon