Picture for Deyi Xiong

Deyi Xiong

Efficiently Exploring Large Language Models for Document-Level Machine Translation with In-context Learning

Add code
Jun 11, 2024
Viaarxiv icon

CRiskEval: A Chinese Multi-Level Risk Evaluation Benchmark Dataset for Large Language Models

Add code
Jun 07, 2024
Viaarxiv icon

Benchmark Underestimates the Readiness of Multi-lingual Dialogue Agents

Add code
May 28, 2024
Figure 1 for Benchmark Underestimates the Readiness of Multi-lingual Dialogue Agents
Figure 2 for Benchmark Underestimates the Readiness of Multi-lingual Dialogue Agents
Figure 3 for Benchmark Underestimates the Readiness of Multi-lingual Dialogue Agents
Figure 4 for Benchmark Underestimates the Readiness of Multi-lingual Dialogue Agents
Viaarxiv icon

Decoding at the Speed of Thought: Harnessing Parallel Decoding of Lexical Units for LLMs

Add code
May 24, 2024
Figure 1 for Decoding at the Speed of Thought: Harnessing Parallel Decoding of Lexical Units for LLMs
Figure 2 for Decoding at the Speed of Thought: Harnessing Parallel Decoding of Lexical Units for LLMs
Figure 3 for Decoding at the Speed of Thought: Harnessing Parallel Decoding of Lexical Units for LLMs
Figure 4 for Decoding at the Speed of Thought: Harnessing Parallel Decoding of Lexical Units for LLMs
Viaarxiv icon

ConTrans: Weak-to-Strong Alignment Engineering via Concept Transplantation

Add code
May 22, 2024
Viaarxiv icon

LFED: A Literary Fiction Evaluation Dataset for Large Language Models

Add code
May 16, 2024
Figure 1 for LFED: A Literary Fiction Evaluation Dataset for Large Language Models
Figure 2 for LFED: A Literary Fiction Evaluation Dataset for Large Language Models
Figure 3 for LFED: A Literary Fiction Evaluation Dataset for Large Language Models
Figure 4 for LFED: A Literary Fiction Evaluation Dataset for Large Language Models
Viaarxiv icon

LHMKE: A Large-scale Holistic Multi-subject Knowledge Evaluation Benchmark for Chinese Large Language Models

Add code
Mar 19, 2024
Figure 1 for LHMKE: A Large-scale Holistic Multi-subject Knowledge Evaluation Benchmark for Chinese Large Language Models
Figure 2 for LHMKE: A Large-scale Holistic Multi-subject Knowledge Evaluation Benchmark for Chinese Large Language Models
Figure 3 for LHMKE: A Large-scale Holistic Multi-subject Knowledge Evaluation Benchmark for Chinese Large Language Models
Figure 4 for LHMKE: A Large-scale Holistic Multi-subject Knowledge Evaluation Benchmark for Chinese Large Language Models
Viaarxiv icon

OpenEval: Benchmarking Chinese LLMs across Capability, Alignment and Safety

Add code
Mar 18, 2024
Figure 1 for OpenEval: Benchmarking Chinese LLMs across Capability, Alignment and Safety
Figure 2 for OpenEval: Benchmarking Chinese LLMs across Capability, Alignment and Safety
Figure 3 for OpenEval: Benchmarking Chinese LLMs across Capability, Alignment and Safety
Figure 4 for OpenEval: Benchmarking Chinese LLMs across Capability, Alignment and Safety
Viaarxiv icon

FineMath: A Fine-Grained Mathematical Evaluation Benchmark for Chinese Large Language Models

Add code
Mar 12, 2024
Figure 1 for FineMath: A Fine-Grained Mathematical Evaluation Benchmark for Chinese Large Language Models
Figure 2 for FineMath: A Fine-Grained Mathematical Evaluation Benchmark for Chinese Large Language Models
Figure 3 for FineMath: A Fine-Grained Mathematical Evaluation Benchmark for Chinese Large Language Models
Figure 4 for FineMath: A Fine-Grained Mathematical Evaluation Benchmark for Chinese Large Language Models
Viaarxiv icon

Exploring Multilingual Human Value Concepts in Large Language Models: Is Value Alignment Consistent, Transferable and Controllable across Languages?

Add code
Feb 28, 2024
Viaarxiv icon