Picture for Bertie Vidgen

Bertie Vidgen

APEX-SWE

Add code
Jan 13, 2026
Viaarxiv icon

Classification is a RAG problem: A case study on hate speech detection

Add code
Aug 08, 2025
Viaarxiv icon

Why human-AI relationships need socioaffective alignment

Add code
Feb 04, 2025
Viaarxiv icon

MSTS: A Multimodal Safety Test Suite for Vision-Language Models

Add code
Jan 17, 2025
Viaarxiv icon

LMUnit: Fine-grained Evaluation with Natural Language Unit Tests

Add code
Dec 17, 2024
Viaarxiv icon

The Responsible Foundation Model Development Cheatsheet: A Review of Tools & Resources

Add code
Jun 26, 2024
Figure 1 for The Responsible Foundation Model Development Cheatsheet: A Review of Tools & Resources
Figure 2 for The Responsible Foundation Model Development Cheatsheet: A Review of Tools & Resources
Figure 3 for The Responsible Foundation Model Development Cheatsheet: A Review of Tools & Resources
Viaarxiv icon

Risks and Opportunities of Open-Source Generative AI

Add code
May 14, 2024
Figure 1 for Risks and Opportunities of Open-Source Generative AI
Figure 2 for Risks and Opportunities of Open-Source Generative AI
Figure 3 for Risks and Opportunities of Open-Source Generative AI
Figure 4 for Risks and Opportunities of Open-Source Generative AI
Viaarxiv icon

WorkBench: a Benchmark Dataset for Agents in a Realistic Workplace Setting

Add code
May 01, 2024
Viaarxiv icon

Near to Mid-term Risks and Opportunities of Open Source Generative AI

Add code
Apr 25, 2024
Figure 1 for Near to Mid-term Risks and Opportunities of Open Source Generative AI
Figure 2 for Near to Mid-term Risks and Opportunities of Open Source Generative AI
Figure 3 for Near to Mid-term Risks and Opportunities of Open Source Generative AI
Figure 4 for Near to Mid-term Risks and Opportunities of Open Source Generative AI
Viaarxiv icon

The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models

Add code
Apr 24, 2024
Figure 1 for The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models
Figure 2 for The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models
Figure 3 for The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models
Figure 4 for The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models
Viaarxiv icon