Picture for Yuuki Yamanaka

Yuuki Yamanaka

Sparse-Autoencoder-Guided Internal Representation Unlearning for Large Language Models

Add code
Sep 19, 2025
Viaarxiv icon

Concept Unlearning in Large Language Models via Self-Constructed Knowledge Triplets

Add code
Sep 19, 2025
Viaarxiv icon

Positive-Unlabeled Diffusion Models for Preventing Sensitive Data Generation

Add code
Mar 05, 2025
Figure 1 for Positive-Unlabeled Diffusion Models for Preventing Sensitive Data Generation
Figure 2 for Positive-Unlabeled Diffusion Models for Preventing Sensitive Data Generation
Figure 3 for Positive-Unlabeled Diffusion Models for Preventing Sensitive Data Generation
Figure 4 for Positive-Unlabeled Diffusion Models for Preventing Sensitive Data Generation
Viaarxiv icon

Deep Positive-Unlabeled Anomaly Detection for Contaminated Unlabeled Data

Add code
May 29, 2024
Figure 1 for Deep Positive-Unlabeled Anomaly Detection for Contaminated Unlabeled Data
Figure 2 for Deep Positive-Unlabeled Anomaly Detection for Contaminated Unlabeled Data
Figure 3 for Deep Positive-Unlabeled Anomaly Detection for Contaminated Unlabeled Data
Figure 4 for Deep Positive-Unlabeled Anomaly Detection for Contaminated Unlabeled Data
Viaarxiv icon

LogELECTRA: Self-supervised Anomaly Detection for Unstructured Logs

Add code
Feb 16, 2024
Figure 1 for LogELECTRA: Self-supervised Anomaly Detection for Unstructured Logs
Figure 2 for LogELECTRA: Self-supervised Anomaly Detection for Unstructured Logs
Figure 3 for LogELECTRA: Self-supervised Anomaly Detection for Unstructured Logs
Figure 4 for LogELECTRA: Self-supervised Anomaly Detection for Unstructured Logs
Viaarxiv icon

ARDIR: Improving Robustness using Knowledge Distillation of Internal Representation

Add code
Nov 01, 2022
Figure 1 for ARDIR: Improving Robustness using Knowledge Distillation of Internal Representation
Figure 2 for ARDIR: Improving Robustness using Knowledge Distillation of Internal Representation
Figure 3 for ARDIR: Improving Robustness using Knowledge Distillation of Internal Representation
Figure 4 for ARDIR: Improving Robustness using Knowledge Distillation of Internal Representation
Viaarxiv icon