Picture for Minlie Huang

Minlie Huang

EJ

Enhancing Long-form Text Generation Efficacy with Task-adaptive Tokenization

Add code
Oct 23, 2023
Figure 1 for Enhancing Long-form Text Generation Efficacy with Task-adaptive Tokenization
Figure 2 for Enhancing Long-form Text Generation Efficacy with Task-adaptive Tokenization
Figure 3 for Enhancing Long-form Text Generation Efficacy with Task-adaptive Tokenization
Figure 4 for Enhancing Long-form Text Generation Efficacy with Task-adaptive Tokenization
Viaarxiv icon

ToRA: A Tool-Integrated Reasoning Agent for Mathematical Problem Solving

Add code
Oct 04, 2023
Figure 1 for ToRA: A Tool-Integrated Reasoning Agent for Mathematical Problem Solving
Figure 2 for ToRA: A Tool-Integrated Reasoning Agent for Mathematical Problem Solving
Figure 3 for ToRA: A Tool-Integrated Reasoning Agent for Mathematical Problem Solving
Figure 4 for ToRA: A Tool-Integrated Reasoning Agent for Mathematical Problem Solving
Viaarxiv icon

Language Model Decoding as Direct Metrics Optimization

Add code
Oct 02, 2023
Figure 1 for Language Model Decoding as Direct Metrics Optimization
Figure 2 for Language Model Decoding as Direct Metrics Optimization
Figure 3 for Language Model Decoding as Direct Metrics Optimization
Figure 4 for Language Model Decoding as Direct Metrics Optimization
Viaarxiv icon

SafetyBench: Evaluating the Safety of Large Language Models with Multiple Choice Questions

Add code
Sep 13, 2023
Figure 1 for SafetyBench: Evaluating the Safety of Large Language Models with Multiple Choice Questions
Figure 2 for SafetyBench: Evaluating the Safety of Large Language Models with Multiple Choice Questions
Figure 3 for SafetyBench: Evaluating the Safety of Large Language Models with Multiple Choice Questions
Figure 4 for SafetyBench: Evaluating the Safety of Large Language Models with Multiple Choice Questions
Viaarxiv icon

On Large Language Models' Selection Bias in Multi-Choice Questions

Add code
Sep 08, 2023
Figure 1 for On Large Language Models' Selection Bias in Multi-Choice Questions
Figure 2 for On Large Language Models' Selection Bias in Multi-Choice Questions
Figure 3 for On Large Language Models' Selection Bias in Multi-Choice Questions
Figure 4 for On Large Language Models' Selection Bias in Multi-Choice Questions
Viaarxiv icon

AgentBench: Evaluating LLMs as Agents

Add code
Aug 07, 2023
Figure 1 for AgentBench: Evaluating LLMs as Agents
Figure 2 for AgentBench: Evaluating LLMs as Agents
Figure 3 for AgentBench: Evaluating LLMs as Agents
Figure 4 for AgentBench: Evaluating LLMs as Agents
Viaarxiv icon

Facilitating Multi-turn Emotional Support Conversation with Positive Emotion Elicitation: A Reinforcement Learning Approach

Add code
Jul 16, 2023
Figure 1 for Facilitating Multi-turn Emotional Support Conversation with Positive Emotion Elicitation: A Reinforcement Learning Approach
Figure 2 for Facilitating Multi-turn Emotional Support Conversation with Positive Emotion Elicitation: A Reinforcement Learning Approach
Figure 3 for Facilitating Multi-turn Emotional Support Conversation with Positive Emotion Elicitation: A Reinforcement Learning Approach
Figure 4 for Facilitating Multi-turn Emotional Support Conversation with Positive Emotion Elicitation: A Reinforcement Learning Approach
Viaarxiv icon

DecompEval: Evaluating Generated Texts as Unsupervised Decomposed Question Answering

Add code
Jul 13, 2023
Figure 1 for DecompEval: Evaluating Generated Texts as Unsupervised Decomposed Question Answering
Figure 2 for DecompEval: Evaluating Generated Texts as Unsupervised Decomposed Question Answering
Figure 3 for DecompEval: Evaluating Generated Texts as Unsupervised Decomposed Question Answering
Figure 4 for DecompEval: Evaluating Generated Texts as Unsupervised Decomposed Question Answering
Viaarxiv icon

Ethicist: Targeted Training Data Extraction Through Loss Smoothed Soft Prompting and Calibrated Confidence Estimation

Add code
Jul 10, 2023
Figure 1 for Ethicist: Targeted Training Data Extraction Through Loss Smoothed Soft Prompting and Calibrated Confidence Estimation
Figure 2 for Ethicist: Targeted Training Data Extraction Through Loss Smoothed Soft Prompting and Calibrated Confidence Estimation
Figure 3 for Ethicist: Targeted Training Data Extraction Through Loss Smoothed Soft Prompting and Calibrated Confidence Estimation
Figure 4 for Ethicist: Targeted Training Data Extraction Through Loss Smoothed Soft Prompting and Calibrated Confidence Estimation
Viaarxiv icon

Mitigating the Learning Bias towards Repetition by Self-Contrastive Training for Open-Ended Generation

Add code
Jul 04, 2023
Viaarxiv icon