Picture for Philip Torr

Philip Torr

Understanding and Improving In-Context Learning on Vision-language Models

Add code
Nov 29, 2023
Viaarxiv icon

Self-Discovering Interpretable Diffusion Latent Directions for Responsible Text-to-Image Generation

Add code
Nov 28, 2023
Viaarxiv icon

A Survey on Transferability of Adversarial Examples across Deep Neural Networks

Add code
Oct 26, 2023
Figure 1 for A Survey on Transferability of Adversarial Examples across Deep Neural Networks
Figure 2 for A Survey on Transferability of Adversarial Examples across Deep Neural Networks
Viaarxiv icon

Managing AI Risks in an Era of Rapid Progress

Add code
Oct 26, 2023
Viaarxiv icon

Real-Fake: Effective Training Data Synthesis Through Distribution Matching

Add code
Oct 16, 2023
Figure 1 for Real-Fake: Effective Training Data Synthesis Through Distribution Matching
Figure 2 for Real-Fake: Effective Training Data Synthesis Through Distribution Matching
Figure 3 for Real-Fake: Effective Training Data Synthesis Through Distribution Matching
Figure 4 for Real-Fake: Effective Training Data Synthesis Through Distribution Matching
Viaarxiv icon

PoRF: Pose Residual Field for Accurate Neural Surface Reconstruction

Add code
Oct 12, 2023
Figure 1 for PoRF: Pose Residual Field for Accurate Neural Surface Reconstruction
Figure 2 for PoRF: Pose Residual Field for Accurate Neural Surface Reconstruction
Figure 3 for PoRF: Pose Residual Field for Accurate Neural Surface Reconstruction
Figure 4 for PoRF: Pose Residual Field for Accurate Neural Surface Reconstruction
Viaarxiv icon

Interpreting Reward Models in RLHF-Tuned Language Models Using Sparse Autoencoders

Add code
Oct 12, 2023
Viaarxiv icon

AttributionLab: Faithfulness of Feature Attribution Under Controllable Environments

Add code
Oct 10, 2023
Figure 1 for AttributionLab: Faithfulness of Feature Attribution Under Controllable Environments
Figure 2 for AttributionLab: Faithfulness of Feature Attribution Under Controllable Environments
Figure 3 for AttributionLab: Faithfulness of Feature Attribution Under Controllable Environments
Figure 4 for AttributionLab: Faithfulness of Feature Attribution Under Controllable Environments
Viaarxiv icon

Exploring Non-additive Randomness on ViT against Query-Based Black-Box Attacks

Add code
Sep 12, 2023
Viaarxiv icon

Neural Collapse Terminus: A Unified Solution for Class Incremental Learning and Its Variants

Add code
Aug 03, 2023
Figure 1 for Neural Collapse Terminus: A Unified Solution for Class Incremental Learning and Its Variants
Figure 2 for Neural Collapse Terminus: A Unified Solution for Class Incremental Learning and Its Variants
Figure 3 for Neural Collapse Terminus: A Unified Solution for Class Incremental Learning and Its Variants
Figure 4 for Neural Collapse Terminus: A Unified Solution for Class Incremental Learning and Its Variants
Viaarxiv icon