Picture for Linbo Qiao

Linbo Qiao

Let the Agent Search: Autonomous Exploration Beats Rigid Workflows in Temporal Question Answering

Add code
Mar 02, 2026
Viaarxiv icon

ParaDySe: A Parallel-Strategy Switching Framework for Dynamic Sequence Lengths in Transformer

Add code
Nov 17, 2025
Viaarxiv icon

A Survey on Memory-Efficient Large-Scale Model Training in AI for Science

Add code
Jan 21, 2025
Figure 1 for A Survey on Memory-Efficient Large-Scale Model Training in AI for Science
Figure 2 for A Survey on Memory-Efficient Large-Scale Model Training in AI for Science
Figure 3 for A Survey on Memory-Efficient Large-Scale Model Training in AI for Science
Figure 4 for A Survey on Memory-Efficient Large-Scale Model Training in AI for Science
Viaarxiv icon

Two-stage Generative Question Answering on Temporal Knowledge Graph Using Large Language Models

Add code
Feb 26, 2024
Figure 1 for Two-stage Generative Question Answering on Temporal Knowledge Graph Using Large Language Models
Figure 2 for Two-stage Generative Question Answering on Temporal Knowledge Graph Using Large Language Models
Figure 3 for Two-stage Generative Question Answering on Temporal Knowledge Graph Using Large Language Models
Figure 4 for Two-stage Generative Question Answering on Temporal Knowledge Graph Using Large Language Models
Viaarxiv icon

TFDMNet: A Novel Network Structure Combines the Time Domain and Frequency Domain Features

Add code
Jan 29, 2024
Viaarxiv icon

A Unified Generative Framework based on Prompt Learning for Various Information Extraction Tasks

Add code
Sep 23, 2022
Figure 1 for A Unified Generative Framework based on Prompt Learning for Various Information Extraction Tasks
Figure 2 for A Unified Generative Framework based on Prompt Learning for Various Information Extraction Tasks
Figure 3 for A Unified Generative Framework based on Prompt Learning for Various Information Extraction Tasks
Figure 4 for A Unified Generative Framework based on Prompt Learning for Various Information Extraction Tasks
Viaarxiv icon

Merak: An Efficient Distributed DNN Training Framework with Automated 3D Parallelism for Giant Foundation Models

Add code
Jun 21, 2022
Figure 1 for Merak: An Efficient Distributed DNN Training Framework with Automated 3D Parallelism for Giant Foundation Models
Figure 2 for Merak: An Efficient Distributed DNN Training Framework with Automated 3D Parallelism for Giant Foundation Models
Figure 3 for Merak: An Efficient Distributed DNN Training Framework with Automated 3D Parallelism for Giant Foundation Models
Figure 4 for Merak: An Efficient Distributed DNN Training Framework with Automated 3D Parallelism for Giant Foundation Models
Viaarxiv icon

DELTA: Dynamically Optimizing GPU Memory beyond Tensor Recomputation

Add code
Mar 30, 2022
Figure 1 for DELTA: Dynamically Optimizing GPU Memory beyond Tensor Recomputation
Figure 2 for DELTA: Dynamically Optimizing GPU Memory beyond Tensor Recomputation
Figure 3 for DELTA: Dynamically Optimizing GPU Memory beyond Tensor Recomputation
Figure 4 for DELTA: Dynamically Optimizing GPU Memory beyond Tensor Recomputation
Viaarxiv icon

Inertial Proximal Deep Learning Alternating Minimization for Efficient Neutral Network Training

Add code
Jan 30, 2021
Figure 1 for Inertial Proximal Deep Learning Alternating Minimization for Efficient Neutral Network Training
Figure 2 for Inertial Proximal Deep Learning Alternating Minimization for Efficient Neutral Network Training
Figure 3 for Inertial Proximal Deep Learning Alternating Minimization for Efficient Neutral Network Training
Figure 4 for Inertial Proximal Deep Learning Alternating Minimization for Efficient Neutral Network Training
Viaarxiv icon

ADMMiRNN: Training RNN with Stable Convergence via An Efficient ADMM Approach

Add code
Jun 17, 2020
Figure 1 for ADMMiRNN: Training RNN with Stable Convergence via An Efficient ADMM Approach
Figure 2 for ADMMiRNN: Training RNN with Stable Convergence via An Efficient ADMM Approach
Figure 3 for ADMMiRNN: Training RNN with Stable Convergence via An Efficient ADMM Approach
Figure 4 for ADMMiRNN: Training RNN with Stable Convergence via An Efficient ADMM Approach
Viaarxiv icon