Alert button
Picture for Jack Hessel

Jack Hessel

Alert button

Personalized Soups: Personalized Large Language Model Alignment via Post-hoc Parameter Merging

Add code
Bookmark button
Alert button
Oct 17, 2023
Joel Jang, Seungone Kim, Bill Yuchen Lin, Yizhong Wang, Jack Hessel, Luke Zettlemoyer, Hannaneh Hajishirzi, Yejin Choi, Prithviraj Ammanabrolu

Viaarxiv icon

Reading Books is Great, But Not if You Are Driving! Visually Grounded Reasoning about Defeasible Commonsense Norms

Add code
Bookmark button
Alert button
Oct 16, 2023
Seungju Han, Junhyeok Kim, Jack Hessel, Liwei Jiang, Jiwan Chung, Yejin Son, Yejin Choi, Youngjae Yu

Viaarxiv icon

VisIT-Bench: A Benchmark for Vision-Language Instruction Following Inspired by Real-World Use

Add code
Bookmark button
Alert button
Aug 12, 2023
Yonatan Bitton, Hritik Bansal, Jack Hessel, Rulin Shao, Wanrong Zhu, Anas Awadalla, Josh Gardner, Rohan Taori, Ludwig Schimdt

Figure 1 for VisIT-Bench: A Benchmark for Vision-Language Instruction Following Inspired by Real-World Use
Figure 2 for VisIT-Bench: A Benchmark for Vision-Language Instruction Following Inspired by Real-World Use
Figure 3 for VisIT-Bench: A Benchmark for Vision-Language Instruction Following Inspired by Real-World Use
Figure 4 for VisIT-Bench: A Benchmark for Vision-Language Instruction Following Inspired by Real-World Use
Viaarxiv icon

OpenFlamingo: An Open-Source Framework for Training Large Autoregressive Vision-Language Models

Add code
Bookmark button
Alert button
Aug 07, 2023
Anas Awadalla, Irena Gao, Josh Gardner, Jack Hessel, Yusuf Hanafy, Wanrong Zhu, Kalyani Marathe, Yonatan Bitton, Samir Gadre, Shiori Sagawa, Jenia Jitsev, Simon Kornblith, Pang Wei Koh, Gabriel Ilharco, Mitchell Wortsman, Ludwig Schmidt

Figure 1 for OpenFlamingo: An Open-Source Framework for Training Large Autoregressive Vision-Language Models
Figure 2 for OpenFlamingo: An Open-Source Framework for Training Large Autoregressive Vision-Language Models
Figure 3 for OpenFlamingo: An Open-Source Framework for Training Large Autoregressive Vision-Language Models
Figure 4 for OpenFlamingo: An Open-Source Framework for Training Large Autoregressive Vision-Language Models
Viaarxiv icon

FunQA: Towards Surprising Video Comprehension

Add code
Bookmark button
Alert button
Jun 26, 2023
Binzhu Xie, Sicheng Zhang, Zitang Zhou, Bo Li, Yuanhan Zhang, Jack Hessel, Jingkang Yang, Ziwei Liu

Figure 1 for FunQA: Towards Surprising Video Comprehension
Figure 2 for FunQA: Towards Surprising Video Comprehension
Viaarxiv icon

Symbolic Chain-of-Thought Distillation: Small Models Can Also "Think" Step-by-Step

Add code
Bookmark button
Alert button
Jun 24, 2023
Liunian Harold Li, Jack Hessel, Youngjae Yu, Xiang Ren, Kai-Wei Chang, Yejin Choi

Figure 1 for Symbolic Chain-of-Thought Distillation: Small Models Can Also "Think" Step-by-Step
Figure 2 for Symbolic Chain-of-Thought Distillation: Small Models Can Also "Think" Step-by-Step
Figure 3 for Symbolic Chain-of-Thought Distillation: Small Models Can Also "Think" Step-by-Step
Figure 4 for Symbolic Chain-of-Thought Distillation: Small Models Can Also "Think" Step-by-Step
Viaarxiv icon

How Far Can Camels Go? Exploring the State of Instruction Tuning on Open Resources

Add code
Bookmark button
Alert button
Jun 07, 2023
Yizhong Wang, Hamish Ivison, Pradeep Dasigi, Jack Hessel, Tushar Khot, Khyathi Raghavi Chandu, David Wadden, Kelsey MacMillan, Noah A. Smith, Iz Beltagy, Hannaneh Hajishirzi

Figure 1 for How Far Can Camels Go? Exploring the State of Instruction Tuning on Open Resources
Figure 2 for How Far Can Camels Go? Exploring the State of Instruction Tuning on Open Resources
Figure 3 for How Far Can Camels Go? Exploring the State of Instruction Tuning on Open Resources
Figure 4 for How Far Can Camels Go? Exploring the State of Instruction Tuning on Open Resources
Viaarxiv icon

Text encoders are performance bottlenecks in contrastive vision-language models

Add code
Bookmark button
Alert button
May 24, 2023
Amita Kamath, Jack Hessel, Kai-Wei Chang

Figure 1 for Text encoders are performance bottlenecks in contrastive vision-language models
Figure 2 for Text encoders are performance bottlenecks in contrastive vision-language models
Figure 3 for Text encoders are performance bottlenecks in contrastive vision-language models
Figure 4 for Text encoders are performance bottlenecks in contrastive vision-language models
Viaarxiv icon

Multimodal C4: An Open, Billion-scale Corpus of Images Interleaved With Text

Add code
Bookmark button
Alert button
Apr 14, 2023
Wanrong Zhu, Jack Hessel, Anas Awadalla, Samir Yitzhak Gadre, Jesse Dodge, Alex Fang, Youngjae Yu, Ludwig Schmidt, William Yang Wang, Yejin Choi

Figure 1 for Multimodal C4: An Open, Billion-scale Corpus of Images Interleaved With Text
Figure 2 for Multimodal C4: An Open, Billion-scale Corpus of Images Interleaved With Text
Figure 3 for Multimodal C4: An Open, Billion-scale Corpus of Images Interleaved With Text
Figure 4 for Multimodal C4: An Open, Billion-scale Corpus of Images Interleaved With Text
Viaarxiv icon