Picture for Yitian Li

Yitian Li

PhysInOne: Visual Physics Learning and Reasoning in One Suite

Add code
Apr 10, 2026
Viaarxiv icon

Hypothesis Testing Prompting Improves Deductive Reasoning in Large Language Models

Add code
May 09, 2024
Figure 1 for Hypothesis Testing Prompting Improves Deductive Reasoning in Large Language Models
Figure 2 for Hypothesis Testing Prompting Improves Deductive Reasoning in Large Language Models
Figure 3 for Hypothesis Testing Prompting Improves Deductive Reasoning in Large Language Models
Figure 4 for Hypothesis Testing Prompting Improves Deductive Reasoning in Large Language Models
Viaarxiv icon

Logical Negation Augmenting and Debiasing for Prompt-based Methods

Add code
May 08, 2024
Figure 1 for Logical Negation Augmenting and Debiasing for Prompt-based Methods
Figure 2 for Logical Negation Augmenting and Debiasing for Prompt-based Methods
Figure 3 for Logical Negation Augmenting and Debiasing for Prompt-based Methods
Figure 4 for Logical Negation Augmenting and Debiasing for Prompt-based Methods
Viaarxiv icon

Comparable Demonstrations are Important in In-Context Learning: A Novel Perspective on Demonstration Selection

Add code
Dec 12, 2023
Figure 1 for Comparable Demonstrations are Important in In-Context Learning: A Novel Perspective on Demonstration Selection
Figure 2 for Comparable Demonstrations are Important in In-Context Learning: A Novel Perspective on Demonstration Selection
Figure 3 for Comparable Demonstrations are Important in In-Context Learning: A Novel Perspective on Demonstration Selection
Figure 4 for Comparable Demonstrations are Important in In-Context Learning: A Novel Perspective on Demonstration Selection
Viaarxiv icon

Chain-of-Thought Tuning: Masked Language Models can also Think Step By Step in Natural Language Understanding

Add code
Oct 18, 2023
Figure 1 for Chain-of-Thought Tuning: Masked Language Models can also Think Step By Step in Natural Language Understanding
Figure 2 for Chain-of-Thought Tuning: Masked Language Models can also Think Step By Step in Natural Language Understanding
Figure 3 for Chain-of-Thought Tuning: Masked Language Models can also Think Step By Step in Natural Language Understanding
Figure 4 for Chain-of-Thought Tuning: Masked Language Models can also Think Step By Step in Natural Language Understanding
Viaarxiv icon

Accurate Use of Label Dependency in Multi-Label Text Classification Through the Lens of Causality

Add code
Oct 11, 2023
Viaarxiv icon

Unlock the Potential of Counterfactually-Augmented Data in Out-Of-Distribution Generalization

Add code
Oct 10, 2023
Figure 1 for Unlock the Potential of Counterfactually-Augmented Data in Out-Of-Distribution Generalization
Figure 2 for Unlock the Potential of Counterfactually-Augmented Data in Out-Of-Distribution Generalization
Figure 3 for Unlock the Potential of Counterfactually-Augmented Data in Out-Of-Distribution Generalization
Figure 4 for Unlock the Potential of Counterfactually-Augmented Data in Out-Of-Distribution Generalization
Viaarxiv icon

Improving the Out-Of-Distribution Generalization Capability of Language Models: Counterfactually-Augmented Data is not Enough

Add code
Feb 18, 2023
Viaarxiv icon

MaxGNR: A Dynamic Weight Strategy via Maximizing Gradient-to-Noise Ratio for Multi-Task Learning

Add code
Feb 18, 2023
Figure 1 for MaxGNR: A Dynamic Weight Strategy via Maximizing Gradient-to-Noise Ratio for Multi-Task Learning
Figure 2 for MaxGNR: A Dynamic Weight Strategy via Maximizing Gradient-to-Noise Ratio for Multi-Task Learning
Figure 3 for MaxGNR: A Dynamic Weight Strategy via Maximizing Gradient-to-Noise Ratio for Multi-Task Learning
Figure 4 for MaxGNR: A Dynamic Weight Strategy via Maximizing Gradient-to-Noise Ratio for Multi-Task Learning
Viaarxiv icon

ReadNet:Towards Accurate ReID with Limited and Noisy Samples

Add code
May 12, 2020
Figure 1 for ReadNet:Towards Accurate ReID with Limited and Noisy Samples
Figure 2 for ReadNet:Towards Accurate ReID with Limited and Noisy Samples
Figure 3 for ReadNet:Towards Accurate ReID with Limited and Noisy Samples
Figure 4 for ReadNet:Towards Accurate ReID with Limited and Noisy Samples
Viaarxiv icon