Picture for Wei Shen

Wei Shen

Amazon

FecTek: Enhancing Term Weight in Lexicon-Based Retrieval with Feature Context and Term-level Knowledge

Add code
Apr 18, 2024
Figure 1 for FecTek: Enhancing Term Weight in Lexicon-Based Retrieval with Feature Context and Term-level Knowledge
Figure 2 for FecTek: Enhancing Term Weight in Lexicon-Based Retrieval with Feature Context and Term-level Knowledge
Figure 3 for FecTek: Enhancing Term Weight in Lexicon-Based Retrieval with Feature Context and Term-level Knowledge
Figure 4 for FecTek: Enhancing Term Weight in Lexicon-Based Retrieval with Feature Context and Term-level Knowledge
Viaarxiv icon

EndoGSLAM: Real-Time Dense Reconstruction and Tracking in Endoscopic Surgeries using Gaussian Splatting

Add code
Mar 22, 2024
Viaarxiv icon

Improving Reinforcement Learning from Human Feedback Using Contrastive Rewards

Add code
Mar 14, 2024
Figure 1 for Improving Reinforcement Learning from Human Feedback Using Contrastive Rewards
Figure 2 for Improving Reinforcement Learning from Human Feedback Using Contrastive Rewards
Figure 3 for Improving Reinforcement Learning from Human Feedback Using Contrastive Rewards
Figure 4 for Improving Reinforcement Learning from Human Feedback Using Contrastive Rewards
Viaarxiv icon

Overcoming Reward Overoptimization via Adversarial Policy Optimization with Lightweight Uncertainty Estimation

Add code
Mar 08, 2024
Figure 1 for Overcoming Reward Overoptimization via Adversarial Policy Optimization with Lightweight Uncertainty Estimation
Figure 2 for Overcoming Reward Overoptimization via Adversarial Policy Optimization with Lightweight Uncertainty Estimation
Figure 3 for Overcoming Reward Overoptimization via Adversarial Policy Optimization with Lightweight Uncertainty Estimation
Viaarxiv icon

GaussianObject: Just Taking Four Images to Get A High-Quality 3D Object with Gaussian Splatting

Add code
Feb 20, 2024
Figure 1 for GaussianObject: Just Taking Four Images to Get A High-Quality 3D Object with Gaussian Splatting
Figure 2 for GaussianObject: Just Taking Four Images to Get A High-Quality 3D Object with Gaussian Splatting
Figure 3 for GaussianObject: Just Taking Four Images to Get A High-Quality 3D Object with Gaussian Splatting
Figure 4 for GaussianObject: Just Taking Four Images to Get A High-Quality 3D Object with Gaussian Splatting
Viaarxiv icon

Training Large Language Models for Reasoning through Reverse Curriculum Reinforcement Learning

Add code
Feb 08, 2024
Figure 1 for Training Large Language Models for Reasoning through Reverse Curriculum Reinforcement Learning
Figure 2 for Training Large Language Models for Reasoning through Reverse Curriculum Reinforcement Learning
Figure 3 for Training Large Language Models for Reasoning through Reverse Curriculum Reinforcement Learning
Figure 4 for Training Large Language Models for Reasoning through Reverse Curriculum Reinforcement Learning
Viaarxiv icon

StepCoder: Improve Code Generation with Reinforcement Learning from Compiler Feedback

Add code
Feb 05, 2024
Figure 1 for StepCoder: Improve Code Generation with Reinforcement Learning from Compiler Feedback
Figure 2 for StepCoder: Improve Code Generation with Reinforcement Learning from Compiler Feedback
Figure 3 for StepCoder: Improve Code Generation with Reinforcement Learning from Compiler Feedback
Figure 4 for StepCoder: Improve Code Generation with Reinforcement Learning from Compiler Feedback
Viaarxiv icon

ViTree: Single-path Neural Tree for Step-wise Interpretable Fine-grained Visual Categorization

Add code
Jan 30, 2024
Figure 1 for ViTree: Single-path Neural Tree for Step-wise Interpretable Fine-grained Visual Categorization
Figure 2 for ViTree: Single-path Neural Tree for Step-wise Interpretable Fine-grained Visual Categorization
Figure 3 for ViTree: Single-path Neural Tree for Step-wise Interpretable Fine-grained Visual Categorization
Figure 4 for ViTree: Single-path Neural Tree for Step-wise Interpretable Fine-grained Visual Categorization
Viaarxiv icon

Linear Alignment: A Closed-form Solution for Aligning Human Preferences without Tuning and Feedback

Add code
Jan 21, 2024
Figure 1 for Linear Alignment: A Closed-form Solution for Aligning Human Preferences without Tuning and Feedback
Figure 2 for Linear Alignment: A Closed-form Solution for Aligning Human Preferences without Tuning and Feedback
Figure 3 for Linear Alignment: A Closed-form Solution for Aligning Human Preferences without Tuning and Feedback
Figure 4 for Linear Alignment: A Closed-form Solution for Aligning Human Preferences without Tuning and Feedback
Viaarxiv icon

Secrets of RLHF in Large Language Models Part II: Reward Modeling

Add code
Jan 12, 2024
Figure 1 for Secrets of RLHF in Large Language Models Part II: Reward Modeling
Figure 2 for Secrets of RLHF in Large Language Models Part II: Reward Modeling
Figure 3 for Secrets of RLHF in Large Language Models Part II: Reward Modeling
Figure 4 for Secrets of RLHF in Large Language Models Part II: Reward Modeling
Viaarxiv icon