Picture for Marc Dymetman

Marc Dymetman

Xerox Research Centre Europe, Grenoble

FaST: Feature-aware Sampling and Tuning for Personalized Preference Alignment with Limited Data

Add code
Aug 06, 2025
Viaarxiv icon

Guaranteed Generation from Large Language Models

Add code
Oct 09, 2024
Figure 1 for Guaranteed Generation from Large Language Models
Figure 2 for Guaranteed Generation from Large Language Models
Figure 3 for Guaranteed Generation from Large Language Models
Figure 4 for Guaranteed Generation from Large Language Models
Viaarxiv icon

Compositional preference models for aligning LMs

Add code
Oct 17, 2023
Figure 1 for Compositional preference models for aligning LMs
Figure 2 for Compositional preference models for aligning LMs
Figure 3 for Compositional preference models for aligning LMs
Figure 4 for Compositional preference models for aligning LMs
Viaarxiv icon

Should you marginalize over possible tokenizations?

Add code
Jun 30, 2023
Viaarxiv icon

disco: a toolkit for Distributional Control of Generative Models

Add code
Mar 08, 2023
Figure 1 for disco: a toolkit for Distributional Control of Generative Models
Figure 2 for disco: a toolkit for Distributional Control of Generative Models
Figure 3 for disco: a toolkit for Distributional Control of Generative Models
Figure 4 for disco: a toolkit for Distributional Control of Generative Models
Viaarxiv icon

Aligning Language Models with Preferences through f-divergence Minimization

Add code
Feb 16, 2023
Figure 1 for Aligning Language Models with Preferences through f-divergence Minimization
Figure 2 for Aligning Language Models with Preferences through f-divergence Minimization
Figure 3 for Aligning Language Models with Preferences through f-divergence Minimization
Figure 4 for Aligning Language Models with Preferences through f-divergence Minimization
Viaarxiv icon

On Reinforcement Learning and Distribution Matching for Fine-Tuning Language Models with no Catastrophic Forgetting

Add code
Jun 01, 2022
Figure 1 for On Reinforcement Learning and Distribution Matching for Fine-Tuning Language Models with no Catastrophic Forgetting
Figure 2 for On Reinforcement Learning and Distribution Matching for Fine-Tuning Language Models with no Catastrophic Forgetting
Figure 3 for On Reinforcement Learning and Distribution Matching for Fine-Tuning Language Models with no Catastrophic Forgetting
Figure 4 for On Reinforcement Learning and Distribution Matching for Fine-Tuning Language Models with no Catastrophic Forgetting
Viaarxiv icon

Sampling from Discrete Energy-Based Models with Quality/Efficiency Trade-offs

Add code
Dec 10, 2021
Figure 1 for Sampling from Discrete Energy-Based Models with Quality/Efficiency Trade-offs
Figure 2 for Sampling from Discrete Energy-Based Models with Quality/Efficiency Trade-offs
Figure 3 for Sampling from Discrete Energy-Based Models with Quality/Efficiency Trade-offs
Figure 4 for Sampling from Discrete Energy-Based Models with Quality/Efficiency Trade-offs
Viaarxiv icon

Controlling Conditional Language Models with Distributional Policy Gradients

Add code
Dec 01, 2021
Figure 1 for Controlling Conditional Language Models with Distributional Policy Gradients
Figure 2 for Controlling Conditional Language Models with Distributional Policy Gradients
Figure 3 for Controlling Conditional Language Models with Distributional Policy Gradients
Figure 4 for Controlling Conditional Language Models with Distributional Policy Gradients
Viaarxiv icon

Energy-Based Models for Code Generation under Compilability Constraints

Add code
Jun 09, 2021
Figure 1 for Energy-Based Models for Code Generation under Compilability Constraints
Figure 2 for Energy-Based Models for Code Generation under Compilability Constraints
Figure 3 for Energy-Based Models for Code Generation under Compilability Constraints
Figure 4 for Energy-Based Models for Code Generation under Compilability Constraints
Viaarxiv icon