Alert button
Picture for Hailu Xu

Hailu Xu

Alert button

Improving CNN-base Stock Trading By Considering Data Heterogeneity and Burst

Mar 14, 2023
Keer Yang, Guanqun Zhang, Chuan Bi, Qiang Guan, Hailu Xu, Shuai Xu

Figure 1 for Improving CNN-base Stock Trading By Considering Data Heterogeneity and Burst
Figure 2 for Improving CNN-base Stock Trading By Considering Data Heterogeneity and Burst
Figure 3 for Improving CNN-base Stock Trading By Considering Data Heterogeneity and Burst
Figure 4 for Improving CNN-base Stock Trading By Considering Data Heterogeneity and Burst

In recent years, there have been quite a few attempts to apply intelligent techniques to financial trading, i.e., constructing automatic and intelligent trading framework based on historical stock price. Due to the unpredictable, uncertainty and volatile nature of financial market, researchers have also resorted to deep learning to construct the intelligent trading framework. In this paper, we propose to use CNN as the core functionality of such framework, because it is able to learn the spatial dependency (i.e., between rows and columns) of the input data. However, different with existing deep learning-based trading frameworks, we develop novel normalization process to prepare the stock data. In particular, we first empirically observe that the stock data is intrinsically heterogeneous and bursty, and then validate the heterogeneity and burst nature of stock data from a statistical perspective. Next, we design the data normalization method in a way such that the data heterogeneity is preserved and bursty events are suppressed. We verify out developed CNN-based trading framework plus our new normalization method on 29 stocks. Experiment results show that our approach can outperform other comparing approaches.

Viaarxiv icon

Making Machine Learning Datasets and Models FAIR for HPC: A Methodology and Case Study

Nov 03, 2022
Pei-Hung Lin, Chunhua Liao, Winson Chen, Tristan Vanderbruggen, Murali Emani, Hailu Xu

Figure 1 for Making Machine Learning Datasets and Models FAIR for HPC: A Methodology and Case Study
Figure 2 for Making Machine Learning Datasets and Models FAIR for HPC: A Methodology and Case Study
Figure 3 for Making Machine Learning Datasets and Models FAIR for HPC: A Methodology and Case Study
Figure 4 for Making Machine Learning Datasets and Models FAIR for HPC: A Methodology and Case Study

The FAIR Guiding Principles aim to improve the findability, accessibility, interoperability, and reusability of digital content by making them both human and machine actionable. However, these principles have not yet been broadly adopted in the domain of machine learning-based program analyses and optimizations for High-Performance Computing (HPC). In this paper, we design a methodology to make HPC datasets and machine learning models FAIR after investigating existing FAIRness assessment and improvement techniques. Our methodology includes a comprehensive, quantitative assessment for elected data, followed by concrete, actionable suggestions to improve FAIRness with respect to common issues related to persistent identifiers, rich metadata descriptions, license and provenance information. Moreover, we select a representative training dataset to evaluate our methodology. The experiment shows the methodology can effectively improve the dataset and model's FAIRness from an initial score of 19.1% to the final score of 83.0%.

Viaarxiv icon