Alert button
Picture for Felix Yu

Felix Yu

Alert button

Metric-aware LLM inference

Add code
Bookmark button
Alert button
Mar 07, 2024
Michal Lukasik, Harikrishna Narasimhan, Aditya Krishna Menon, Felix Yu, Sanjiv Kumar

Figure 1 for Metric-aware LLM inference
Figure 2 for Metric-aware LLM inference
Figure 3 for Metric-aware LLM inference
Figure 4 for Metric-aware LLM inference
Viaarxiv icon

ReST meets ReAct: Self-Improvement for Multi-Step Reasoning LLM Agent

Add code
Bookmark button
Alert button
Dec 15, 2023
Renat Aksitov, Sobhan Miryoosefi, Zonglin Li, Daliang Li, Sheila Babayan, Kavya Kopparapu, Zachary Fisher, Ruiqi Guo, Sushant Prakash, Pranesh Srinivasan, Manzil Zaheer, Felix Yu, Sanjiv Kumar

Viaarxiv icon

SpecTr: Fast Speculative Decoding via Optimal Transport

Add code
Bookmark button
Alert button
Oct 23, 2023
Ziteng Sun, Ananda Theertha Suresh, Jae Hun Ro, Ahmad Beirami, Himanshu Jain, Felix Yu

Viaarxiv icon

Large Language Models with Controllable Working Memory

Add code
Bookmark button
Alert button
Nov 09, 2022
Daliang Li, Ankit Singh Rawat, Manzil Zaheer, Xin Wang, Michal Lukasik, Andreas Veit, Felix Yu, Sanjiv Kumar

Figure 1 for Large Language Models with Controllable Working Memory
Figure 2 for Large Language Models with Controllable Working Memory
Figure 3 for Large Language Models with Controllable Working Memory
Figure 4 for Large Language Models with Controllable Working Memory
Viaarxiv icon

Preserving In-Context Learning ability in Large Language Model Fine-tuning

Add code
Bookmark button
Alert button
Nov 01, 2022
Yihan Wang, Si Si, Daliang Li, Michal Lukasik, Felix Yu, Cho-Jui Hsieh, Inderjit S Dhillon, Sanjiv Kumar

Figure 1 for Preserving In-Context Learning ability in Large Language Model Fine-tuning
Figure 2 for Preserving In-Context Learning ability in Large Language Model Fine-tuning
Figure 3 for Preserving In-Context Learning ability in Large Language Model Fine-tuning
Figure 4 for Preserving In-Context Learning ability in Large Language Model Fine-tuning
Viaarxiv icon

Large Models are Parsimonious Learners: Activation Sparsity in Trained Transformers

Add code
Bookmark button
Alert button
Oct 12, 2022
Zonglin Li, Chong You, Srinadh Bhojanapalli, Daliang Li, Ankit Singh Rawat, Sashank J. Reddi, Ke Ye, Felix Chern, Felix Yu, Ruiqi Guo, Sanjiv Kumar

Figure 1 for Large Models are Parsimonious Learners: Activation Sparsity in Trained Transformers
Figure 2 for Large Models are Parsimonious Learners: Activation Sparsity in Trained Transformers
Figure 3 for Large Models are Parsimonious Learners: Activation Sparsity in Trained Transformers
Figure 4 for Large Models are Parsimonious Learners: Activation Sparsity in Trained Transformers
Viaarxiv icon

FedDM: Iterative Distribution Matching for Communication-Efficient Federated Learning

Add code
Bookmark button
Alert button
Jul 20, 2022
Yuanhao Xiong, Ruochen Wang, Minhao Cheng, Felix Yu, Cho-Jui Hsieh

Figure 1 for FedDM: Iterative Distribution Matching for Communication-Efficient Federated Learning
Figure 2 for FedDM: Iterative Distribution Matching for Communication-Efficient Federated Learning
Figure 3 for FedDM: Iterative Distribution Matching for Communication-Efficient Federated Learning
Figure 4 for FedDM: Iterative Distribution Matching for Communication-Efficient Federated Learning
Viaarxiv icon

Correlated quantization for distributed mean estimation and optimization

Add code
Bookmark button
Alert button
Mar 09, 2022
Ananda Theertha Suresh, Ziteng Sun, Jae Hun Ro, Felix Yu

Figure 1 for Correlated quantization for distributed mean estimation and optimization
Figure 2 for Correlated quantization for distributed mean estimation and optimization
Figure 3 for Correlated quantization for distributed mean estimation and optimization
Figure 4 for Correlated quantization for distributed mean estimation and optimization
Viaarxiv icon

HD-cos Networks: Efficient Neural Architectures for Secure Multi-Party Computation

Add code
Bookmark button
Alert button
Oct 28, 2021
Wittawat Jitkrittum, Michal Lukasik, Ananda Theertha Suresh, Felix Yu, Gang Wang

Figure 1 for HD-cos Networks: Efficient Neural Architectures for Secure Multi-Party Computation
Figure 2 for HD-cos Networks: Efficient Neural Architectures for Secure Multi-Party Computation
Figure 3 for HD-cos Networks: Efficient Neural Architectures for Secure Multi-Party Computation
Figure 4 for HD-cos Networks: Efficient Neural Architectures for Secure Multi-Party Computation
Viaarxiv icon

InFillmore: Neural Frame Lexicalization for Narrative Text Infilling

Add code
Bookmark button
Alert button
Mar 08, 2021
Jiefu Ou, Nathaniel Weir, Anton Belyy, Felix Yu, Benjamin Van Durme

Figure 1 for InFillmore: Neural Frame Lexicalization for Narrative Text Infilling
Figure 2 for InFillmore: Neural Frame Lexicalization for Narrative Text Infilling
Figure 3 for InFillmore: Neural Frame Lexicalization for Narrative Text Infilling
Figure 4 for InFillmore: Neural Frame Lexicalization for Narrative Text Infilling
Viaarxiv icon