Alert button
Picture for Irina Rish

Irina Rish

Alert button

Deep Generative Sampling in the Dual Divergence Space: A Data-efficient & Interpretative Approach for Generative AI

Add code
Bookmark button
Alert button
Apr 10, 2024
Sahil Garg, Anderson Schneider, Anant Raj, Kashif Rasul, Yuriy Nevmyvaka, Sneihil Gopal, Amit Dhurandhar, Guillermo Cecchi, Irina Rish

Viaarxiv icon

Simple and Scalable Strategies to Continually Pre-train Large Language Models

Add code
Bookmark button
Alert button
Mar 26, 2024
Adam Ibrahim, Benjamin Thérien, Kshitij Gupta, Mats L. Richter, Quentin Anthony, Timothée Lesort, Eugene Belilovsky, Irina Rish

Figure 1 for Simple and Scalable Strategies to Continually Pre-train Large Language Models
Figure 2 for Simple and Scalable Strategies to Continually Pre-train Large Language Models
Figure 3 for Simple and Scalable Strategies to Continually Pre-train Large Language Models
Figure 4 for Simple and Scalable Strategies to Continually Pre-train Large Language Models
Viaarxiv icon

Unsupervised Concept Discovery Mitigates Spurious Correlations

Add code
Bookmark button
Alert button
Feb 20, 2024
Md Rifat Arefin, Yan Zhang, Aristide Baratin, Francesco Locatello, Irina Rish, Dianbo Liu, Kenji Kawaguchi

Viaarxiv icon

Towards Machines that Trust: AI Agents Learn to Trust in the Trust Game

Add code
Bookmark button
Alert button
Dec 20, 2023
Ardavan S. Nobandegani, Irina Rish, Thomas R. Shultz

Viaarxiv icon

Lag-Llama: Towards Foundation Models for Time Series Forecasting

Add code
Bookmark button
Alert button
Oct 12, 2023
Kashif Rasul, Arjun Ashok, Andrew Robert Williams, Arian Khorasani, George Adamopoulos, Rishika Bhagwatkar, Marin Biloš, Hena Ghonia, Nadhir Vincent Hassen, Anderson Schneider, Sahil Garg, Alexandre Drouin, Nicolas Chapados, Yuriy Nevmyvaka, Irina Rish

Figure 1 for Lag-Llama: Towards Foundation Models for Time Series Forecasting
Figure 2 for Lag-Llama: Towards Foundation Models for Time Series Forecasting
Figure 3 for Lag-Llama: Towards Foundation Models for Time Series Forecasting
Figure 4 for Lag-Llama: Towards Foundation Models for Time Series Forecasting
Viaarxiv icon

LORD: Low Rank Decomposition Of Monolingual Code LLMs For One-Shot Compression

Add code
Bookmark button
Alert button
Sep 25, 2023
Ayush Kaushal, Tejas Vaidhya, Irina Rish

Viaarxiv icon

Amplifying Pathological Detection in EEG Signaling Pathways through Cross-Dataset Transfer Learning

Add code
Bookmark button
Alert button
Sep 19, 2023
Mohammad-Javad Darvishi-Bayazi, Mohammad Sajjad Ghaemi, Timothee Lesort, Md Rifat Arefin, Jocelyn Faubert, Irina Rish

Figure 1 for Amplifying Pathological Detection in EEG Signaling Pathways through Cross-Dataset Transfer Learning
Figure 2 for Amplifying Pathological Detection in EEG Signaling Pathways through Cross-Dataset Transfer Learning
Figure 3 for Amplifying Pathological Detection in EEG Signaling Pathways through Cross-Dataset Transfer Learning
Figure 4 for Amplifying Pathological Detection in EEG Signaling Pathways through Cross-Dataset Transfer Learning
Viaarxiv icon

Continual Pre-Training of Large Language Models: How to (re)warm your model?

Add code
Bookmark button
Alert button
Aug 08, 2023
Kshitij Gupta, Benjamin Thérien, Adam Ibrahim, Mats L. Richter, Quentin Anthony, Eugene Belilovsky, Irina Rish, Timothée Lesort

Figure 1 for Continual Pre-Training of Large Language Models: How to (re)warm your model?
Figure 2 for Continual Pre-Training of Large Language Models: How to (re)warm your model?
Figure 3 for Continual Pre-Training of Large Language Models: How to (re)warm your model?
Figure 4 for Continual Pre-Training of Large Language Models: How to (re)warm your model?
Viaarxiv icon

GOKU-UI: Ubiquitous Inference through Attention and Multiple Shooting for Continuous-time Generative Models

Add code
Bookmark button
Alert button
Jul 11, 2023
Germán Abrevaya, Mahta Ramezanian-Panahi, Jean-Christophe Gagnon-Audet, Irina Rish, Pablo Polosecki, Silvina Ponce Dawson, Guillermo Cecchi, Guillaume Dumas

Figure 1 for GOKU-UI: Ubiquitous Inference through Attention and Multiple Shooting for Continuous-time Generative Models
Figure 2 for GOKU-UI: Ubiquitous Inference through Attention and Multiple Shooting for Continuous-time Generative Models
Figure 3 for GOKU-UI: Ubiquitous Inference through Attention and Multiple Shooting for Continuous-time Generative Models
Figure 4 for GOKU-UI: Ubiquitous Inference through Attention and Multiple Shooting for Continuous-time Generative Models
Viaarxiv icon