Alert button
Picture for Tushar Krishna

Tushar Krishna

Alert button

H3DFact: Heterogeneous 3D Integrated CIM for Factorization with Holographic Perceptual Representations

Add code
Bookmark button
Alert button
Apr 05, 2024
Zishen Wan, Che-Kai Liu, Mohamed Ibrahim, Hanchen Yang, Samuel Spetalnick, Tushar Krishna, Arijit Raychowdhury

Viaarxiv icon

Abstracting Sparse DNN Acceleration via Structured Sparse Tensor Decomposition

Add code
Bookmark button
Alert button
Mar 12, 2024
Geonhwa Jeong, Po-An Tsai, Abhimanyu R. Bambhaniya, Stephen W. Keckler, Tushar Krishna

Figure 1 for Abstracting Sparse DNN Acceleration via Structured Sparse Tensor Decomposition
Figure 2 for Abstracting Sparse DNN Acceleration via Structured Sparse Tensor Decomposition
Figure 3 for Abstracting Sparse DNN Acceleration via Structured Sparse Tensor Decomposition
Figure 4 for Abstracting Sparse DNN Acceleration via Structured Sparse Tensor Decomposition
Viaarxiv icon

GEAR: An Efficient KV Cache Compression Recipe for Near-Lossless Generative Inference of LLM

Add code
Bookmark button
Alert button
Mar 11, 2024
Hao Kang, Qingru Zhang, Souvik Kundu, Geonhwa Jeong, Zaoxing Liu, Tushar Krishna, Tuo Zhao

Figure 1 for GEAR: An Efficient KV Cache Compression Recipe for Near-Lossless Generative Inference of LLM
Figure 2 for GEAR: An Efficient KV Cache Compression Recipe for Near-Lossless Generative Inference of LLM
Figure 3 for GEAR: An Efficient KV Cache Compression Recipe for Near-Lossless Generative Inference of LLM
Figure 4 for GEAR: An Efficient KV Cache Compression Recipe for Near-Lossless Generative Inference of LLM
Viaarxiv icon

GEAR: An Efficient KV Cache Compression Recipefor Near-Lossless Generative Inference of LLM

Add code
Bookmark button
Alert button
Mar 08, 2024
Hao Kang, Qingru Zhang, Souvik Kundu, Geonhwa Jeong, Zaoxing Liu, Tushar Krishna, Tuo Zhao

Figure 1 for GEAR: An Efficient KV Cache Compression Recipefor Near-Lossless Generative Inference of LLM
Figure 2 for GEAR: An Efficient KV Cache Compression Recipefor Near-Lossless Generative Inference of LLM
Figure 3 for GEAR: An Efficient KV Cache Compression Recipefor Near-Lossless Generative Inference of LLM
Figure 4 for GEAR: An Efficient KV Cache Compression Recipefor Near-Lossless Generative Inference of LLM
Viaarxiv icon

Algorithm-Hardware Co-Design of Distribution-Aware Logarithmic-Posit Encodings for Efficient DNN Inference

Add code
Bookmark button
Alert button
Mar 08, 2024
Akshat Ramachandran, Zishen Wan, Geonhwa Jeong, John Gustafson, Tushar Krishna

Figure 1 for Algorithm-Hardware Co-Design of Distribution-Aware Logarithmic-Posit Encodings for Efficient DNN Inference
Figure 2 for Algorithm-Hardware Co-Design of Distribution-Aware Logarithmic-Posit Encodings for Efficient DNN Inference
Figure 3 for Algorithm-Hardware Co-Design of Distribution-Aware Logarithmic-Posit Encodings for Efficient DNN Inference
Figure 4 for Algorithm-Hardware Co-Design of Distribution-Aware Logarithmic-Posit Encodings for Efficient DNN Inference
Viaarxiv icon

Progressive Gradient Flow for Robust N:M Sparsity Training in Transformers

Add code
Bookmark button
Alert button
Feb 07, 2024
Abhimanyu Rajeshkumar Bambhaniya, Amir Yazdanbakhsh, Suvinay Subramanian, Sheng-Chun Kao, Shivani Agrawal, Utku Evci, Tushar Krishna

Viaarxiv icon

Towards Cognitive AI Systems: a Survey and Prospective on Neuro-Symbolic AI

Add code
Bookmark button
Alert button
Jan 02, 2024
Zishen Wan, Che-Kai Liu, Hanchen Yang, Chaojian Li, Haoran You, Yonggan Fu, Cheng Wan, Tushar Krishna, Yingyan Lin, Arijit Raychowdhury

Viaarxiv icon

Subgraph Stationary Hardware-Software Inference Co-Design

Add code
Bookmark button
Alert button
Jun 21, 2023
Payman Behnam, Jianming Tong, Alind Khare, Yangyu Chen, Yue Pan, Pranav Gadikar, Abhimanyu Rajeshkumar Bambhaniya, Tushar Krishna, Alexey Tumanov

Figure 1 for Subgraph Stationary Hardware-Software Inference Co-Design
Figure 2 for Subgraph Stationary Hardware-Software Inference Co-Design
Figure 3 for Subgraph Stationary Hardware-Software Inference Co-Design
Figure 4 for Subgraph Stationary Hardware-Software Inference Co-Design
Viaarxiv icon

Chakra: Advancing Performance Benchmarking and Co-design using Standardized Execution Traces

Add code
Bookmark button
Alert button
May 26, 2023
Srinivas Sridharan, Taekyung Heo, Louis Feng, Zhaodong Wang, Matt Bergeron, Wenyin Fu, Shengbao Zheng, Brian Coutinho, Saeed Rashidi, Changhai Man, Tushar Krishna

Figure 1 for Chakra: Advancing Performance Benchmarking and Co-design using Standardized Execution Traces
Figure 2 for Chakra: Advancing Performance Benchmarking and Co-design using Standardized Execution Traces
Figure 3 for Chakra: Advancing Performance Benchmarking and Co-design using Standardized Execution Traces
Figure 4 for Chakra: Advancing Performance Benchmarking and Co-design using Standardized Execution Traces
Viaarxiv icon