Alert button
Picture for Aleksandr Drozd

Aleksandr Drozd

Alert button

Aurora-M: The First Open Source Multilingual Language Model Red-teamed according to the U.S. Executive Order

Add code
Bookmark button
Alert button
Mar 30, 2024
Taishi Nakamura, Mayank Mishra, Simone Tedeschi, Yekun Chai, Jason T Stillerman, Felix Friedrich, Prateek Yadav, Tanmay Laud, Vu Minh Chien, Terry Yue Zhuo, Diganta Misra, Ben Bogin, Xuan-Son Vu, Marzena Karpinska, Arnav Varma Dantuluri, Wojciech Kusa, Tommaso Furlanello, Rio Yokota, Niklas Muennighoff, Suhas Pai, Tosin Adewumi, Veronika Laippala, Xiaozhe Yao, Adalberto Junior, Alpay Ariyak, Aleksandr Drozd, Jordan Clive, Kshitij Gupta, Liangyu Chen, Qi Sun, Ken Tsui, Noah Persaud, Nour Fahmy, Tianlong Chen, Mohit Bansal, Nicolo Monti, Tai Dang, Ziyang Luo, Tien-Tung Bui, Roberto Navigli, Virendra Mehta, Matthew Blumberg, Victor May, Huu Nguyen, Sampo Pyysalo

Viaarxiv icon

Myths and Legends in High-Performance Computing

Add code
Bookmark button
Alert button
Jan 06, 2023
Satoshi Matsuoka, Jens Domke, Mohamed Wahib, Aleksandr Drozd, Torsten Hoefler

Figure 1 for Myths and Legends in High-Performance Computing
Figure 2 for Myths and Legends in High-Performance Computing
Figure 3 for Myths and Legends in High-Performance Computing
Viaarxiv icon

Outliers Dimensions that Disrupt Transformers Are Driven by Frequency

Add code
Bookmark button
Alert button
May 23, 2022
Giovanni Puccetti, Anna Rogers, Aleksandr Drozd, Felice Dell'Orletta

Figure 1 for Outliers Dimensions that Disrupt Transformers Are Driven by Frequency
Figure 2 for Outliers Dimensions that Disrupt Transformers Are Driven by Frequency
Figure 3 for Outliers Dimensions that Disrupt Transformers Are Driven by Frequency
Figure 4 for Outliers Dimensions that Disrupt Transformers Are Driven by Frequency
Viaarxiv icon

MLPerf HPC: A Holistic Benchmark Suite for Scientific Machine Learning on HPC Systems

Add code
Bookmark button
Alert button
Oct 26, 2021
Steven Farrell, Murali Emani, Jacob Balma, Lukas Drescher, Aleksandr Drozd, Andreas Fink, Geoffrey Fox, David Kanter, Thorsten Kurth, Peter Mattson, Dawei Mu, Amit Ruhela, Kento Sato, Koichi Shirahata, Tsuguchika Tabaru, Aristeidis Tsaris, Jan Balewski, Ben Cumming, Takumi Danjo, Jens Domke, Takaaki Fukai, Naoto Fukumoto, Tatsuya Fukushi, Balazs Gerofi, Takumi Honda, Toshiyuki Imamura, Akihiko Kasagi, Kentaro Kawakami, Shuhei Kudo, Akiyoshi Kuroda, Maxime Martinasso, Satoshi Matsuoka, Henrique Mendonça, Kazuki Minami, Prabhat Ram, Takashi Sawada, Mallikarjun Shankar, Tom St. John, Akihiro Tabuchi, Venkatram Vishwanath, Mohamed Wahib, Masafumi Yamazaki, Junqi Yin

Figure 1 for MLPerf HPC: A Holistic Benchmark Suite for Scientific Machine Learning on HPC Systems
Figure 2 for MLPerf HPC: A Holistic Benchmark Suite for Scientific Machine Learning on HPC Systems
Figure 3 for MLPerf HPC: A Holistic Benchmark Suite for Scientific Machine Learning on HPC Systems
Figure 4 for MLPerf HPC: A Holistic Benchmark Suite for Scientific Machine Learning on HPC Systems
Viaarxiv icon

Generalization in NLI: Ways (Not) To Go Beyond Simple Heuristics

Add code
Bookmark button
Alert button
Oct 04, 2021
Prajjwal Bhargava, Aleksandr Drozd, Anna Rogers

Figure 1 for Generalization in NLI: Ways (Not) To Go Beyond Simple Heuristics
Figure 2 for Generalization in NLI: Ways (Not) To Go Beyond Simple Heuristics
Figure 3 for Generalization in NLI: Ways (Not) To Go Beyond Simple Heuristics
Figure 4 for Generalization in NLI: Ways (Not) To Go Beyond Simple Heuristics
Viaarxiv icon

Scaling Distributed Deep Learning Workloads beyond the Memory Capacity with KARMA

Add code
Bookmark button
Alert button
Aug 26, 2020
Mohamed Wahib, Haoyu Zhang, Truong Thao Nguyen, Aleksandr Drozd, Jens Domke, Lingqi Zhang, Ryousei Takano, Satoshi Matsuoka

Figure 1 for Scaling Distributed Deep Learning Workloads beyond the Memory Capacity with KARMA
Figure 2 for Scaling Distributed Deep Learning Workloads beyond the Memory Capacity with KARMA
Figure 3 for Scaling Distributed Deep Learning Workloads beyond the Memory Capacity with KARMA
Figure 4 for Scaling Distributed Deep Learning Workloads beyond the Memory Capacity with KARMA
Viaarxiv icon