Picture for Alexander Wan

Alexander Wan

The 2025 Foundation Model Transparency Index

Add code
Dec 11, 2025
Figure 1 for The 2025 Foundation Model Transparency Index
Figure 2 for The 2025 Foundation Model Transparency Index
Figure 3 for The 2025 Foundation Model Transparency Index
Figure 4 for The 2025 Foundation Model Transparency Index
Viaarxiv icon

What Evidence Do Language Models Find Convincing?

Add code
Feb 19, 2024
Figure 1 for What Evidence Do Language Models Find Convincing?
Figure 2 for What Evidence Do Language Models Find Convincing?
Figure 3 for What Evidence Do Language Models Find Convincing?
Figure 4 for What Evidence Do Language Models Find Convincing?
Viaarxiv icon

Poisoning Language Models During Instruction Tuning

Add code
May 01, 2023
Figure 1 for Poisoning Language Models During Instruction Tuning
Figure 2 for Poisoning Language Models During Instruction Tuning
Figure 3 for Poisoning Language Models During Instruction Tuning
Figure 4 for Poisoning Language Models During Instruction Tuning
Viaarxiv icon

GLUECons: A Generic Benchmark for Learning Under Constraints

Add code
Feb 16, 2023
Viaarxiv icon