Alert button
Picture for Haokun Liu

Haokun Liu

Alert button

Dense Training, Sparse Inference: Rethinking Training of Mixture-of-Experts Language Models

Add code
Bookmark button
Alert button
Apr 08, 2024
Bowen Pan, Yikang Shen, Haokun Liu, Mayank Mishra, Gaoyuan Zhang, Aude Oliva, Colin Raffel, Rameswar Panda

Viaarxiv icon

Hypothesis Generation with Large Language Models

Add code
Bookmark button
Alert button
Apr 05, 2024
Yangqiaoyu Zhou, Haokun Liu, Tejes Srivastava, Hongyuan Mei, Chenhao Tan

Viaarxiv icon

Learning to Route Among Specialized Experts for Zero-Shot Generalization

Add code
Bookmark button
Alert button
Feb 08, 2024
Mohammed Muqeeth, Haokun Liu, Yufan Liu, Colin Raffel

Viaarxiv icon

LLM-Based Human-Robot Collaboration Framework for Manipulation Tasks

Add code
Bookmark button
Alert button
Aug 29, 2023
Haokun Liu, Yaonan Zhu, Kenji Kato, Izumi Kondo, Tadayoshi Aoyama, Yasuhisa Hasegawa

Figure 1 for LLM-Based Human-Robot Collaboration Framework for Manipulation Tasks
Figure 2 for LLM-Based Human-Robot Collaboration Framework for Manipulation Tasks
Viaarxiv icon

Git-Theta: A Git Extension for Collaborative Development of Machine Learning Models

Add code
Bookmark button
Alert button
Jun 07, 2023
Nikhil Kandpal, Brian Lester, Mohammed Muqeeth, Anisha Mascarenhas, Monty Evans, Vishal Baskaran, Tenghao Huang, Haokun Liu, Colin Raffel

Figure 1 for Git-Theta: A Git Extension for Collaborative Development of Machine Learning Models
Figure 2 for Git-Theta: A Git Extension for Collaborative Development of Machine Learning Models
Figure 3 for Git-Theta: A Git Extension for Collaborative Development of Machine Learning Models
Figure 4 for Git-Theta: A Git Extension for Collaborative Development of Machine Learning Models
Viaarxiv icon

Soft Merging of Experts with Adaptive Routing

Add code
Bookmark button
Alert button
Jun 06, 2023
Mohammed Muqeeth, Haokun Liu, Colin Raffel

Figure 1 for Soft Merging of Experts with Adaptive Routing
Figure 2 for Soft Merging of Experts with Adaptive Routing
Figure 3 for Soft Merging of Experts with Adaptive Routing
Figure 4 for Soft Merging of Experts with Adaptive Routing
Viaarxiv icon

The Semantic Scholar Open Data Platform

Add code
Bookmark button
Alert button
Jan 24, 2023
Rodney Kinney, Chloe Anastasiades, Russell Authur, Iz Beltagy, Jonathan Bragg, Alexandra Buraczynski, Isabel Cachola, Stefan Candra, Yoganand Chandrasekhar, Arman Cohan, Miles Crawford, Doug Downey, Jason Dunkelberger, Oren Etzioni, Rob Evans, Sergey Feldman, Joseph Gorney, David Graham, Fangzhou Hu, Regan Huff, Daniel King, Sebastian Kohlmeier, Bailey Kuehl, Michael Langan, Daniel Lin, Haokun Liu, Kyle Lo, Jaron Lochner, Kelsey MacMillan, Tyler Murray, Chris Newell, Smita Rao, Shaurya Rohatgi, Paul Sayre, Zejiang Shen, Amanpreet Singh, Luca Soldaini, Shivashankar Subramanian, Amber Tanaka, Alex D. Wade, Linda Wagner, Lucy Lu Wang, Chris Wilhelm, Caroline Wu, Jiangjiang Yang, Angele Zamarron, Madeleine Van Zuylen, Daniel S. Weld

Figure 1 for The Semantic Scholar Open Data Platform
Figure 2 for The Semantic Scholar Open Data Platform
Figure 3 for The Semantic Scholar Open Data Platform
Figure 4 for The Semantic Scholar Open Data Platform
Viaarxiv icon

Beyond Ensemble Averages: Leveraging Climate Model Ensembles for Subseasonal Forecasting

Add code
Bookmark button
Alert button
Nov 29, 2022
Elena Orlova, Haokun Liu, Raphael Rossellini, Benjamin Cash, Rebecca Willett

Figure 1 for Beyond Ensemble Averages: Leveraging Climate Model Ensembles for Subseasonal Forecasting
Figure 2 for Beyond Ensemble Averages: Leveraging Climate Model Ensembles for Subseasonal Forecasting
Figure 3 for Beyond Ensemble Averages: Leveraging Climate Model Ensembles for Subseasonal Forecasting
Figure 4 for Beyond Ensemble Averages: Leveraging Climate Model Ensembles for Subseasonal Forecasting
Viaarxiv icon

Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning

Add code
Bookmark button
Alert button
May 11, 2022
Haokun Liu, Derek Tam, Mohammed Muqeeth, Jay Mohta, Tenghao Huang, Mohit Bansal, Colin Raffel

Figure 1 for Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning
Figure 2 for Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning
Figure 3 for Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning
Figure 4 for Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning
Viaarxiv icon

Fine-Tuned Transformers Show Clusters of Similar Representations Across Layers

Add code
Bookmark button
Alert button
Sep 20, 2021
Jason Phang, Haokun Liu, Samuel R. Bowman

Figure 1 for Fine-Tuned Transformers Show Clusters of Similar Representations Across Layers
Figure 2 for Fine-Tuned Transformers Show Clusters of Similar Representations Across Layers
Figure 3 for Fine-Tuned Transformers Show Clusters of Similar Representations Across Layers
Figure 4 for Fine-Tuned Transformers Show Clusters of Similar Representations Across Layers
Viaarxiv icon