Alert button
Picture for Tianda Li

Tianda Li

Alert button

SPDF: Sparse Pre-training and Dense Fine-tuning for Large Language Models

Add code
Bookmark button
Alert button
Mar 18, 2023
Vithursan Thangarasa, Abhay Gupta, William Marshall, Tianda Li, Kevin Leong, Dennis DeCoste, Sean Lie, Shreyas Saxena

Figure 1 for SPDF: Sparse Pre-training and Dense Fine-tuning for Large Language Models
Figure 2 for SPDF: Sparse Pre-training and Dense Fine-tuning for Large Language Models
Figure 3 for SPDF: Sparse Pre-training and Dense Fine-tuning for Large Language Models
Figure 4 for SPDF: Sparse Pre-training and Dense Fine-tuning for Large Language Models
Viaarxiv icon

Towards Understanding Label Regularization for Fine-tuning Pre-trained Language Models

Add code
Bookmark button
Alert button
May 25, 2022
Ivan Kobyzev, Aref Jafari, Mehdi Rezagholizadeh, Tianda Li, Alan Do-Omri, Peng Lu, Ali Ghodsi, Pascal Poupart

Figure 1 for Towards Understanding Label Regularization for Fine-tuning Pre-trained Language Models
Figure 2 for Towards Understanding Label Regularization for Fine-tuning Pre-trained Language Models
Figure 3 for Towards Understanding Label Regularization for Fine-tuning Pre-trained Language Models
Figure 4 for Towards Understanding Label Regularization for Fine-tuning Pre-trained Language Models
Viaarxiv icon

A Short Study on Compressing Decoder-Based Language Models

Add code
Bookmark button
Alert button
Oct 16, 2021
Tianda Li, Yassir El Mesbahi, Ivan Kobyzev, Ahmad Rashid, Atif Mahmud, Nithin Anchuri, Habib Hajimolahoseini, Yang Liu, Mehdi Rezagholizadeh

Figure 1 for A Short Study on Compressing Decoder-Based Language Models
Figure 2 for A Short Study on Compressing Decoder-Based Language Models
Figure 3 for A Short Study on Compressing Decoder-Based Language Models
Figure 4 for A Short Study on Compressing Decoder-Based Language Models
Viaarxiv icon

How to Select One Among All? An Extensive Empirical Study Towards the Robustness of Knowledge Distillation in Natural Language Understanding

Add code
Bookmark button
Alert button
Sep 20, 2021
Tianda Li, Ahmad Rashid, Aref Jafari, Pranav Sharma, Ali Ghodsi, Mehdi Rezagholizadeh

Figure 1 for How to Select One Among All? An Extensive Empirical Study Towards the Robustness of Knowledge Distillation in Natural Language Understanding
Figure 2 for How to Select One Among All? An Extensive Empirical Study Towards the Robustness of Knowledge Distillation in Natural Language Understanding
Figure 3 for How to Select One Among All? An Extensive Empirical Study Towards the Robustness of Knowledge Distillation in Natural Language Understanding
Figure 4 for How to Select One Among All? An Extensive Empirical Study Towards the Robustness of Knowledge Distillation in Natural Language Understanding
Viaarxiv icon

Unsupervised Pre-training with Structured Knowledge for Improving Natural Language Inference

Add code
Bookmark button
Alert button
Sep 08, 2021
Xiaoyu Yang, Xiaodan Zhu, Zhan Shi, Tianda Li

Figure 1 for Unsupervised Pre-training with Structured Knowledge for Improving Natural Language Inference
Figure 2 for Unsupervised Pre-training with Structured Knowledge for Improving Natural Language Inference
Figure 3 for Unsupervised Pre-training with Structured Knowledge for Improving Natural Language Inference
Figure 4 for Unsupervised Pre-training with Structured Knowledge for Improving Natural Language Inference
Viaarxiv icon

Learning to Retrieve Entity-Aware Knowledge and Generate Responses with Copy Mechanism for Task-Oriented Dialogue Systems

Add code
Bookmark button
Alert button
Dec 22, 2020
Chao-Hong Tan, Xiaoyu Yang, Zi'ou Zheng, Tianda Li, Yufei Feng, Jia-Chen Gu, Quan Liu, Dan Liu, Zhen-Hua Ling, Xiaodan Zhu

Figure 1 for Learning to Retrieve Entity-Aware Knowledge and Generate Responses with Copy Mechanism for Task-Oriented Dialogue Systems
Figure 2 for Learning to Retrieve Entity-Aware Knowledge and Generate Responses with Copy Mechanism for Task-Oriented Dialogue Systems
Figure 3 for Learning to Retrieve Entity-Aware Knowledge and Generate Responses with Copy Mechanism for Task-Oriented Dialogue Systems
Figure 4 for Learning to Retrieve Entity-Aware Knowledge and Generate Responses with Copy Mechanism for Task-Oriented Dialogue Systems
Viaarxiv icon

DialBERT: A Hierarchical Pre-Trained Model for Conversation Disentanglement

Add code
Bookmark button
Alert button
Apr 08, 2020
Tianda Li, Jia-Chen Gu, Xiaodan Zhu, Quan Liu, Zhen-Hua Ling, Zhiming Su, Si Wei

Figure 1 for DialBERT: A Hierarchical Pre-Trained Model for Conversation Disentanglement
Figure 2 for DialBERT: A Hierarchical Pre-Trained Model for Conversation Disentanglement
Figure 3 for DialBERT: A Hierarchical Pre-Trained Model for Conversation Disentanglement
Figure 4 for DialBERT: A Hierarchical Pre-Trained Model for Conversation Disentanglement
Viaarxiv icon

Speaker-Aware BERT for Multi-Turn Response Selection in Retrieval-Based Chatbots

Add code
Bookmark button
Alert button
Apr 07, 2020
Jia-Chen Gu, Tianda Li, Quan Liu, Xiaodan Zhu, Zhen-Hua Ling, Zhiming Su, Si Wei

Figure 1 for Speaker-Aware BERT for Multi-Turn Response Selection in Retrieval-Based Chatbots
Figure 2 for Speaker-Aware BERT for Multi-Turn Response Selection in Retrieval-Based Chatbots
Figure 3 for Speaker-Aware BERT for Multi-Turn Response Selection in Retrieval-Based Chatbots
Figure 4 for Speaker-Aware BERT for Multi-Turn Response Selection in Retrieval-Based Chatbots
Viaarxiv icon

Pre-Trained and Attention-Based Neural Networks for Building Noetic Task-Oriented Dialogue Systems

Add code
Bookmark button
Alert button
Apr 04, 2020
Jia-Chen Gu, Tianda Li, Quan Liu, Xiaodan Zhu, Zhen-Hua Ling, Yu-Ping Ruan

Figure 1 for Pre-Trained and Attention-Based Neural Networks for Building Noetic Task-Oriented Dialogue Systems
Figure 2 for Pre-Trained and Attention-Based Neural Networks for Building Noetic Task-Oriented Dialogue Systems
Figure 3 for Pre-Trained and Attention-Based Neural Networks for Building Noetic Task-Oriented Dialogue Systems
Figure 4 for Pre-Trained and Attention-Based Neural Networks for Building Noetic Task-Oriented Dialogue Systems
Viaarxiv icon

Several Experiments on Investigating Pretraining and Knowledge-Enhanced Models for Natural Language Inference

Add code
Bookmark button
Alert button
Apr 27, 2019
Tianda Li, Xiaodan Zhu, Quan Liu, Qian Chen, Zhigang Chen, Si Wei

Figure 1 for Several Experiments on Investigating Pretraining and Knowledge-Enhanced Models for Natural Language Inference
Figure 2 for Several Experiments on Investigating Pretraining and Knowledge-Enhanced Models for Natural Language Inference
Figure 3 for Several Experiments on Investigating Pretraining and Knowledge-Enhanced Models for Natural Language Inference
Figure 4 for Several Experiments on Investigating Pretraining and Knowledge-Enhanced Models for Natural Language Inference
Viaarxiv icon