@inproceedings{huang2023spatio, abstract = {RUL prediction plays a crucial role in the health management of industrial systems. Given the increasing complexity of systems, data-driven predictive models have attracted significant research interest. Upon reviewing the existing literature, it appears that many studies either do not fully integrate both spatial and temporal features or employ only a single attention mechanism. Furthermore, there seems to be inconsistency in the choice of data normalization methods, particularly concerning operating conditions, which might influence predictive performance. To bridge these observations, this study presents the Spatio-Temporal Attention Graph Neural Network. Our model combines graph neural networks and temporal convolutional neural networks for spatial and temporal feature extraction, respectively. The cascade of these extractors, combined with multihead attention mechanisms for both spatio-temporal dimensions, aims to improve predictive precision and refine model explainability. Comprehensive experiments were conducted on the CMAPSS dataset to evaluate the impact of unified versus clustering normalization. The findings suggest that our model performs state-of-the-art results using only the unified normalization. Additionally, when dealing with datasets with multiple operating conditions, cluster normalization enhances the performance of our proposed model by up to 27%.}, author = {Huang, Zhixin and He, Yujiang and Sick, Bernhard}, booktitle = {Computational Science and Computational Intelligence (CSCI)}, interhash = {d2c044e372780d668eb0140a724454e0}, intrahash = {ad927931297c4668fb4cda3fa215b717}, note = {(accepted)}, publisher = {IEEE}, title = {Spatio-Temporal Attention Graph Neural Network for Remaining Useful Life Prediction}, year = 2023 } @article{he2023proud, abstract = {Anomaly detection methods applied to time series are mostly viewed as a black box, which solely provides a deterministic answer to the detected target. Without a convincing explanation, domain experts can hardly trust the detection results and conduct a further time series diagnosis in real-world applications. To overcome the challenge, we mathematically analyzed the sources of anomalies and novelties in multivariate time series as well as their relationships from the perspective of Gaussian-distributed non-stationary noise. Furthermore, we proposed mathematical methods to generate artificial time series and synthetic anomalies, with the goal of solving the problem that it is difficult to train and evaluate models in real-world applications due to the lack of sufficient data. In addition, we designed extbf{Pr}obabilistic extbf{Ou}tlier extbf{D}etection (PrOuD), which is a general solution to provide interpretable detection results to assist domain experts in time series analysis. PrOuD can convert a predictive uncertainty of a trained model about a time series value into an estimated uncertainty of the detected outlier through Monte Carlo Estimation. The experimental results obtained on both artificial time series and real-world photovoltaic inverter data demonstrated that the proposed solution could detect emerging anomalies accurately and quickly. The implemented PrOuD demo case shows the potential to makes the detection results of the existing detection methods more convincing so that domain experts can more efficiently complete their tasks, such as time series diagnosis and anomalous pattern clustering.}, author = {He, Yujiang and Huang, Zhixin and Vogt, Stephan and Sick, Bernhard}, doi = {10.3390/en17010064}, interhash = {5f47e8f6ca1b59328f48f6f775540ed2}, intrahash = {f3ffc0a2138c25f36764613251b966ff}, journal = {Energies (MDPI)}, number = 1, pages = 64, publisher = {MDPI}, title = {PrOuD: Probabilistic Outlier Detection Solution for Time Series Analysis on Real-world Photovoltaic Inverters}, url = {https://www.mdpi.com/1996-1073/17/1/64}, volume = 17, year = 2024 } @inproceedings{huang2023active, abstract = {Active learning strategies aim to minimize the number of queried samples for model training. However, two challenges in pool-based deep active learning on imbalanced datasets are observed in experiments: (1) the declining performance of active learning strategies due to imbalanced class distribution; (2)~the lack of sample diversity in acquisition batches due to the absence of timely model updates. This paper proposes the AL-FaMoUS, a general solution combining fast model updates and class-balanced minibatch selection to the active learning process. Furthermore, a simplification of the AL-FaMoUS, which selects one single sample in each acquisition minibatch, is experimentally evaluated on four image and three time-series imbalanced datasets. The results demonstrate that the implemented AL-FaMoUS outperforms the other adopted AL strategies, including uncertainty sampling and BALD solely combined with either the fast model update or the class balance selection strategy, in terms of accuracy and Macro F1 score.}, author = {Huang, Zhixin and He, Yujiang and Herde, Marek and Huseljic, Denis and Sick, Bernhard}, booktitle = {Workshop on Interactive Adapative Learning (IAL), ECML PKDD}, interhash = {628e2871115e97a27bb1eee0484e1209}, intrahash = {53a9d44816a8f463ce66e8efa287fb0b}, pages = {28--45}, title = {Active Learning with Fast Model Updates and Class-Balanced Selection for Imbalanced Datasets}, url = {https://ceur-ws.org/Vol-3470/paper5.pdf}, year = 2023 } @inproceedings{he2022design, abstract = {Continual learning can enable neural networks to evolve by learning new tasks sequentially in task-changing scenarios. However, two general challenges should be overcome in further research before we apply this technique to real-world applications. Firstly, newly collected novelties from the data stream in applications could contain anomalies that are meaningless for continual learning. Instead of viewing them as a new task for updating, we have to filter out such anomalies to reduce the disturbance of extremely high-entropy data for the progression of convergence. Secondly, fewer efforts have been put into research regarding the explainability of continual learning, which leads to a lack of transparency and credibility of the updated neural networks. Elaborated explanations about the process and result of continual learning can help experts in judgment and making decisions. Therefore, we propose the conceptual design of an explainability module with experts in the loop based on techniques, such as dimension reduction, visualization, and evaluation strategies. This work aims to overcome the mentioned challenges by sufficiently explaining and visualizing the identified anomalies and the updated neural network. With the help of this module, experts can be more confident in decision-making regarding anomaly filtering, dynamic adjustment of hyperparameters, data backup, etc.}, author = {He, Yujiang and Huang, Zhixin and Sick, Bernhard}, booktitle = {Workshop on Interactive Machine Learning Workshop (IMLW), AAAI}, interhash = {e1229de8e285fd3b266fd73c3f5287c1}, intrahash = {040c3c0e2cae3bcf668f6c6a67aed6be}, pages = {1--6}, title = {Design of Explainability Module with Experts in the Loop for Visualization and Dynamic Adjustment of Continual Learning}, url = {https://arxiv.org/abs/2202.06781}, year = 2022 } @incollection{he2022adaptive, abstract = {Compared with traditional deep learning techniques, continual learning enables deep neural networks to learn continually and adaptively. Deep neural networks have to learn unseen tasks and overcome forgetting the knowledge obtained from previously learned tasks as the amount of data keeps increasing in applications. This article proposes two continual learning application scenarios, i.e., the target-domain incremental scenario and the data-domain incremental scenario, to describe the potential challenges in this context. Based on our previous work regarding the CLeaR framework, which is short for continual learning for regression tasks, models will be enabled to extend themselves and to learn data successively. Research topics are related, but not limited, to developing continual deep learning algorithms, strategies for non-stationarity detection in data streams, explainable and visualizable artificial intelligence, etc. Moreover, the framework- and algorithm-related hyperparameters should be dynamically updated in applications. Forecasting experiments will be conducted based on power generation and consumption data collected from real-world applications. A series of comprehensive evaluation metrics and visualization tools are applied to access the experimental results. The proposed framework is expected to be generally applied to other constantly changing scenarios.}, author = {He, Yujiang}, booktitle = {Organic Computing -- Doctoral Dissertation Colloquium 2021}, editor = {Tomforde, Sven and Krupitzer, Christian}, interhash = {2d8b5c76bf2af7ee1c60eba5d29ffe3b}, intrahash = {8c277a6ad33b732d6a7dd6c87f36d924}, pages = {125--140}, publisher = {kassel university press}, title = {Adaptive Explainable Continual Learning Framework for Regression Problems with Focus on Power Forecasts}, year = 2022 } @inproceedings{HHVS21, author = {Huang, Zhixin and He, Yujiang and Vogt, Stephan and Sick, Bernhard}, booktitle = {Workshop on Interactive Adaptive Learning (IAL), ECML PKDD}, interhash = {8c8035d8b4e56913c0aff00a75baceb2}, intrahash = {ded4abeb9d7f10a2d290b2826d3364f8}, note = {(accepted)}, title = {Uncertainty and Utility Sampling with Pre-Clustering}, year = 2021 } @article{HS21, abstract = {Catastrophic forgetting means that a trained neural network model gradually forgets the previously learned tasks when being retrained on new tasks. Overcoming the forgetting problem is a major problem in machine learning. Numerous continual learning algorithms are very successful in incremental learning of classification tasks, where new samples with their labels appear frequently. However, there is currently no research that addresses the catastrophic forgetting problem in regression tasks as far as we know. This problem has emerged as one of the primary constraints in some applications, such as renewable energy forecasts. This article clarifies problem-related definitions and proposes a new methodological framework that can forecast targets and update itself by means of continual learning. The framework consists of forecasting neural networks and buffers, which store newly collected data from a non-stationary data stream in an application. The changed probability distribution of the data stream, which the framework has identified, will be learned sequentially. The framework is called CLeaR (Continual Learning for Regression Tasks), where components can be flexibly customized for a specific application scenario. We design two sets of experiments to evaluate the CLeaR framework concerning fitting error (training), prediction error (test), and forgetting ratio. The first one is based on an artificial time series to explore how hyperparameters affect the CLeaR framework. The second one is designed with data collected from European wind farms to evaluate the CLeaR framework's performance in a real-world application. The experimental results demonstrate that the CLeaR framework can continually acquire knowledge in the data stream and improve the prediction accuracy. The article concludes with further research issues arising from requirements to extend the framework.}, author = {He, Yujiang and Sick, Bernhard}, day = 16, doi = {10.1186/s42467-021-00009-8}, interhash = {422b571040aa11037d8cfbcfe784f824}, intrahash = {df315115321132c0bb801b89437f1d31}, issn = {2523-398X}, journal = {AI Perspectives}, month = jul, number = 1, pages = 2, title = {CLeaR: An adaptive continual learning framework for regression tasks}, url = {https://doi.org/10.1186/s42467-021-00009-8}, volume = 3, year = 2021 }