Picture for Xiaolong Huang

Xiaolong Huang

Synthetic Data (Almost) from Scratch: Generalized Instruction Tuning for Language Models

Add code
Feb 20, 2024
Figure 1 for Synthetic Data (Almost) from Scratch: Generalized Instruction Tuning for Language Models
Figure 2 for Synthetic Data (Almost) from Scratch: Generalized Instruction Tuning for Language Models
Figure 3 for Synthetic Data (Almost) from Scratch: Generalized Instruction Tuning for Language Models
Figure 4 for Synthetic Data (Almost) from Scratch: Generalized Instruction Tuning for Language Models
Viaarxiv icon

Multilingual E5 Text Embeddings: A Technical Report

Add code
Feb 08, 2024
Viaarxiv icon

One Step Learning, One Step Review

Add code
Jan 19, 2024
Figure 1 for One Step Learning, One Step Review
Figure 2 for One Step Learning, One Step Review
Figure 3 for One Step Learning, One Step Review
Figure 4 for One Step Learning, One Step Review
Viaarxiv icon

Improving Text Embeddings with Large Language Models

Add code
Dec 31, 2023
Figure 1 for Improving Text Embeddings with Large Language Models
Figure 2 for Improving Text Embeddings with Large Language Models
Figure 3 for Improving Text Embeddings with Large Language Models
Figure 4 for Improving Text Embeddings with Large Language Models
Viaarxiv icon

Large Search Model: Redefining Search Stack in the Era of LLMs

Add code
Oct 23, 2023
Figure 1 for Large Search Model: Redefining Search Stack in the Era of LLMs
Figure 2 for Large Search Model: Redefining Search Stack in the Era of LLMs
Figure 3 for Large Search Model: Redefining Search Stack in the Era of LLMs
Figure 4 for Large Search Model: Redefining Search Stack in the Era of LLMs
Viaarxiv icon

Text Embeddings by Weakly-Supervised Contrastive Pre-training

Add code
Dec 07, 2022
Figure 1 for Text Embeddings by Weakly-Supervised Contrastive Pre-training
Figure 2 for Text Embeddings by Weakly-Supervised Contrastive Pre-training
Figure 3 for Text Embeddings by Weakly-Supervised Contrastive Pre-training
Figure 4 for Text Embeddings by Weakly-Supervised Contrastive Pre-training
Viaarxiv icon

2nd Place Solution to Google Universal Image Embedding

Add code
Oct 19, 2022
Figure 1 for 2nd Place Solution to Google Universal Image Embedding
Figure 2 for 2nd Place Solution to Google Universal Image Embedding
Figure 3 for 2nd Place Solution to Google Universal Image Embedding
Viaarxiv icon

Effective and Efficient Query-aware Snippet Extraction for Web Search

Add code
Oct 17, 2022
Figure 1 for Effective and Efficient Query-aware Snippet Extraction for Web Search
Figure 2 for Effective and Efficient Query-aware Snippet Extraction for Web Search
Figure 3 for Effective and Efficient Query-aware Snippet Extraction for Web Search
Figure 4 for Effective and Efficient Query-aware Snippet Extraction for Web Search
Viaarxiv icon

LexMAE: Lexicon-Bottlenecked Pretraining for Large-Scale Retrieval

Add code
Aug 31, 2022
Figure 1 for LexMAE: Lexicon-Bottlenecked Pretraining for Large-Scale Retrieval
Figure 2 for LexMAE: Lexicon-Bottlenecked Pretraining for Large-Scale Retrieval
Figure 3 for LexMAE: Lexicon-Bottlenecked Pretraining for Large-Scale Retrieval
Figure 4 for LexMAE: Lexicon-Bottlenecked Pretraining for Large-Scale Retrieval
Viaarxiv icon

SimLM: Pre-training with Representation Bottleneck for Dense Passage Retrieval

Add code
Jul 06, 2022
Figure 1 for SimLM: Pre-training with Representation Bottleneck for Dense Passage Retrieval
Figure 2 for SimLM: Pre-training with Representation Bottleneck for Dense Passage Retrieval
Figure 3 for SimLM: Pre-training with Representation Bottleneck for Dense Passage Retrieval
Figure 4 for SimLM: Pre-training with Representation Bottleneck for Dense Passage Retrieval
Viaarxiv icon