Picture for Tianyu Gao

Tianyu Gao

Long-Context Language Modeling with Parallel Context Encoding

Add code
Feb 26, 2024
Viaarxiv icon

Improving Language Understanding from Screenshots

Add code
Feb 21, 2024
Figure 1 for Improving Language Understanding from Screenshots
Figure 2 for Improving Language Understanding from Screenshots
Figure 3 for Improving Language Understanding from Screenshots
Figure 4 for Improving Language Understanding from Screenshots
Viaarxiv icon

Harmonizing Covariance and Expressiveness for Deep Hamiltonian Regression in Crystalline Material Research: a Hybrid Cascaded Regression Framework

Add code
Jan 15, 2024
Figure 1 for Harmonizing Covariance and Expressiveness for Deep Hamiltonian Regression in Crystalline Material Research: a Hybrid Cascaded Regression Framework
Figure 2 for Harmonizing Covariance and Expressiveness for Deep Hamiltonian Regression in Crystalline Material Research: a Hybrid Cascaded Regression Framework
Figure 3 for Harmonizing Covariance and Expressiveness for Deep Hamiltonian Regression in Crystalline Material Research: a Hybrid Cascaded Regression Framework
Figure 4 for Harmonizing Covariance and Expressiveness for Deep Hamiltonian Regression in Crystalline Material Research: a Hybrid Cascaded Regression Framework
Viaarxiv icon

Evaluating Large Language Models at Evaluating Instruction Following

Add code
Oct 11, 2023
Figure 1 for Evaluating Large Language Models at Evaluating Instruction Following
Figure 2 for Evaluating Large Language Models at Evaluating Instruction Following
Figure 3 for Evaluating Large Language Models at Evaluating Instruction Following
Figure 4 for Evaluating Large Language Models at Evaluating Instruction Following
Viaarxiv icon

Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning

Add code
Oct 10, 2023
Viaarxiv icon

Fine-Tuning Language Models with Just Forward Passes

Add code
May 27, 2023
Figure 1 for Fine-Tuning Language Models with Just Forward Passes
Figure 2 for Fine-Tuning Language Models with Just Forward Passes
Figure 3 for Fine-Tuning Language Models with Just Forward Passes
Figure 4 for Fine-Tuning Language Models with Just Forward Passes
Viaarxiv icon

Enabling Large Language Models to Generate Text with Citations

Add code
May 24, 2023
Figure 1 for Enabling Large Language Models to Generate Text with Citations
Figure 2 for Enabling Large Language Models to Generate Text with Citations
Figure 3 for Enabling Large Language Models to Generate Text with Citations
Figure 4 for Enabling Large Language Models to Generate Text with Citations
Viaarxiv icon

What In-Context Learning "Learns" In-Context: Disentangling Task Recognition and Task Learning

Add code
May 16, 2023
Figure 1 for What In-Context Learning "Learns" In-Context: Disentangling Task Recognition and Task Learning
Figure 2 for What In-Context Learning "Learns" In-Context: Disentangling Task Recognition and Task Learning
Figure 3 for What In-Context Learning "Learns" In-Context: Disentangling Task Recognition and Task Learning
Figure 4 for What In-Context Learning "Learns" In-Context: Disentangling Task Recognition and Task Learning
Viaarxiv icon

STM-UNet: An Efficient U-shaped Architecture Based on Swin Transformer and Multi-scale MLP for Medical Image Segmentation

Add code
Apr 25, 2023
Figure 1 for STM-UNet: An Efficient U-shaped Architecture Based on Swin Transformer and Multi-scale MLP for Medical Image Segmentation
Figure 2 for STM-UNet: An Efficient U-shaped Architecture Based on Swin Transformer and Multi-scale MLP for Medical Image Segmentation
Figure 3 for STM-UNet: An Efficient U-shaped Architecture Based on Swin Transformer and Multi-scale MLP for Medical Image Segmentation
Figure 4 for STM-UNet: An Efficient U-shaped Architecture Based on Swin Transformer and Multi-scale MLP for Medical Image Segmentation
Viaarxiv icon

The CRINGE Loss: Learning what language not to model

Add code
Nov 10, 2022
Figure 1 for The CRINGE Loss: Learning what language not to model
Figure 2 for The CRINGE Loss: Learning what language not to model
Figure 3 for The CRINGE Loss: Learning what language not to model
Figure 4 for The CRINGE Loss: Learning what language not to model
Viaarxiv icon