Picture for Preslav Nakov

Preslav Nakov

Mohamed bin Zayed University of Artificial Intelligence

Unconditional Truthfulness: Learning Conditional Dependency for Uncertainty Quantification of Large Language Models

Add code
Aug 20, 2024
Figure 1 for Unconditional Truthfulness: Learning Conditional Dependency for Uncertainty Quantification of Large Language Models
Figure 2 for Unconditional Truthfulness: Learning Conditional Dependency for Uncertainty Quantification of Large Language Models
Figure 3 for Unconditional Truthfulness: Learning Conditional Dependency for Uncertainty Quantification of Large Language Models
Figure 4 for Unconditional Truthfulness: Learning Conditional Dependency for Uncertainty Quantification of Large Language Models
Viaarxiv icon

LLM-DetectAIve: a Tool for Fine-Grained Machine-Generated Text Detection

Add code
Aug 08, 2024
Figure 1 for LLM-DetectAIve: a Tool for Fine-Grained Machine-Generated Text Detection
Figure 2 for LLM-DetectAIve: a Tool for Fine-Grained Machine-Generated Text Detection
Figure 3 for LLM-DetectAIve: a Tool for Fine-Grained Machine-Generated Text Detection
Figure 4 for LLM-DetectAIve: a Tool for Fine-Grained Machine-Generated Text Detection
Viaarxiv icon

Web2Code: A Large-scale Webpage-to-Code Dataset and Evaluation Framework for Multimodal LLMs

Add code
Jun 28, 2024
Figure 1 for Web2Code: A Large-scale Webpage-to-Code Dataset and Evaluation Framework for Multimodal LLMs
Figure 2 for Web2Code: A Large-scale Webpage-to-Code Dataset and Evaluation Framework for Multimodal LLMs
Figure 3 for Web2Code: A Large-scale Webpage-to-Code Dataset and Evaluation Framework for Multimodal LLMs
Figure 4 for Web2Code: A Large-scale Webpage-to-Code Dataset and Evaluation Framework for Multimodal LLMs
Viaarxiv icon

Benchmarking Uncertainty Quantification Methods for Large Language Models with LM-Polygraph

Add code
Jun 21, 2024
Figure 1 for Benchmarking Uncertainty Quantification Methods for Large Language Models with LM-Polygraph
Figure 2 for Benchmarking Uncertainty Quantification Methods for Large Language Models with LM-Polygraph
Figure 3 for Benchmarking Uncertainty Quantification Methods for Large Language Models with LM-Polygraph
Figure 4 for Benchmarking Uncertainty Quantification Methods for Large Language Models with LM-Polygraph
Viaarxiv icon

Can Machines Resonate with Humans? Evaluating the Emotional and Empathic Comprehension of LMs

Add code
Jun 17, 2024
Figure 1 for Can Machines Resonate with Humans? Evaluating the Emotional and Empathic Comprehension of LMs
Figure 2 for Can Machines Resonate with Humans? Evaluating the Emotional and Empathic Comprehension of LMs
Figure 3 for Can Machines Resonate with Humans? Evaluating the Emotional and Empathic Comprehension of LMs
Figure 4 for Can Machines Resonate with Humans? Evaluating the Emotional and Empathic Comprehension of LMs
Viaarxiv icon

Exploring the Limitations of Detecting Machine-Generated Text

Add code
Jun 16, 2024
Figure 1 for Exploring the Limitations of Detecting Machine-Generated Text
Figure 2 for Exploring the Limitations of Detecting Machine-Generated Text
Figure 3 for Exploring the Limitations of Detecting Machine-Generated Text
Figure 4 for Exploring the Limitations of Detecting Machine-Generated Text
Viaarxiv icon

Corpus Poisoning via Approximate Greedy Gradient Descent

Add code
Jun 07, 2024
Figure 1 for Corpus Poisoning via Approximate Greedy Gradient Descent
Figure 2 for Corpus Poisoning via Approximate Greedy Gradient Descent
Figure 3 for Corpus Poisoning via Approximate Greedy Gradient Descent
Figure 4 for Corpus Poisoning via Approximate Greedy Gradient Descent
Viaarxiv icon

Missci: Reconstructing Fallacies in Misrepresented Science

Add code
Jun 05, 2024
Viaarxiv icon

MemeMQA: Multimodal Question Answering for Memes via Rationale-Based Inferencing

Add code
May 18, 2024
Figure 1 for MemeMQA: Multimodal Question Answering for Memes via Rationale-Based Inferencing
Figure 2 for MemeMQA: Multimodal Question Answering for Memes via Rationale-Based Inferencing
Figure 3 for MemeMQA: Multimodal Question Answering for Memes via Rationale-Based Inferencing
Figure 4 for MemeMQA: Multimodal Question Answering for Memes via Rationale-Based Inferencing
Viaarxiv icon

OpenFactCheck: A Unified Framework for Factuality Evaluation of LLMs

Add code
May 09, 2024
Figure 1 for OpenFactCheck: A Unified Framework for Factuality Evaluation of LLMs
Figure 2 for OpenFactCheck: A Unified Framework for Factuality Evaluation of LLMs
Figure 3 for OpenFactCheck: A Unified Framework for Factuality Evaluation of LLMs
Figure 4 for OpenFactCheck: A Unified Framework for Factuality Evaluation of LLMs
Viaarxiv icon