Next Article in Journal
Mapping Tropical Forested Wetlands Biomass with LiDAR: A Machine Learning Comparison
Previous Article in Journal
A 1.8 m Class Pathfinder Raman LIDAR for the Northern Site of the Cherenkov Telescope Array Observatory—Technical Design
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Solar Flare Prediction Using Multivariate Time Series of Photospheric Magnetic Field Parameters: A Comparative Analysis of Vector, Time Series, and Graph Data Representations

by
Onur Vural
*,
Shah Muhammad Hamdi
and
Soukaina Filali Boubrahimi
Department of Computer Science, Utah State University, Logan, UT 84322, USA
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(6), 1075; https://doi.org/10.3390/rs17061075
Submission received: 30 January 2025 / Revised: 9 March 2025 / Accepted: 13 March 2025 / Published: 18 March 2025

Abstract

:
The purpose of this study is to provide a comprehensive resource for the selection of data representations for machine learning-oriented models and components in solar flare prediction tasks. Major solar flares occurring in the solar corona and heliosphere can bring potential destructive consequences, posing significant risks to astronauts, space stations, electronics, communication systems, and numerous technological infrastructures. For this reason, the accurate detection of major flares is essential for mitigating these hazards and ensuring the safety of our technology-dependent society. In response, leveraging machine learning techniques for predicting solar flares has emerged as a significant application within the realm of data science, relying on sensor data collected from solar active region photospheric magnetic fields by space- and ground-based observatories. In this research, three distinct solar flare prediction strategies utilizing the photospheric magnetic field parameter-based multivariate time series dataset are evaluated, with a focus on data representation techniques. Specifically, we examine vector-based, time series-based, and graph-based approaches to identify the most effective data representation for capturing key characteristics of the dataset. The vector-based approach condenses multivariate time series into a compressed vector form, the time series representation leverages temporal patterns, and the graph-based method models interdependencies between magnetic field parameters. The results demonstrate that the vector representation approach exhibits exceptional robustness in predicting solar flares, consistently yielding strong and reliable classification outcomes by effectively encapsulating the intricate relationships within photospheric magnetic field data when coupled with appropriate downstream machine learning classifiers.

1. Introduction

A solar flare is a sudden burst in magnetic flux emanating from the Sun’s surface within the solar corona and heliosphere. In accordance with peak soft X-Ray flux in the 1–8 Å wavelength range, flares can be logarithmically categorized as A, B, C, M, and X, with increasing intensities where M and X category flares correspond to the most intense flare categories [1]. These major solar events, as in Figure 1, can produce emissions such as Gamma Ray, X-Ray, and Extreme Ultraviolet (EUV). As a result, astronauts operating in space stations can face radiation-based health risks, electronic devices such as GPS and radio may fail to function, and serious infrastructure problems may occur [2,3]. Furthermore, the infamous 1859 Carrington Event shows that a similar massive solar superstorm hitting our contemporary technology-reliant society could lead to a prolonged nationwide blackout in the United States, lasting for months. The aftermath is projected to inflict economic havoc, with anticipated damages reaching USD trillions [4]. Currently, there is no well-established theoretical connection between the influx of magnetic fields and the occurrence of extreme events in solar active regions. Accordingly, in the heliophysics community, this draws emphasis on detailed data analysis and investigation of different methodologies to obtain robust flare prediction performance from magnetic field state data of solar active regions, with sensor observations from space-based and ground-based instruments playing a crucial role in data collection.
Recently (2020), Space Weather Analytics for Solar Flares (SWAN-SF) [5] has been proposed as an openly accessible, comprehensive dataset encapsulating the multivariate time series (MVTS) representation which is obtained from solar photospheric vector magnetograms in the Space-weather HMI Active Region Patch (SHARP) series. Including data from solar active regions that are collected from observations between May 2010 and December 2018, SWAN-SF contains features of 24 photospheric field parameters and integrates over 10,000 flare reports. In SWAN-SF, each MVTS instance captures magnetic field data over a predetermined observation period, with labeled flare classes occurring subsequent to a specified prediction time. In this respect, SWAN-SF allows the task of solar flare prediction to be formulated as an MVTS classification task. SWAN-SF enhances the accuracy of solar flare predictions and enables comprehensive investigations into elusive predictors and early indicators of flares. These improvements yield advantages in both practical operational applications and fundamental research [6].
To date, predicting solar flares through photospheric magnetic field data has been explored through various machine learning methodologies, naturally incorporating different data modalities, including the use of SWAN-SF for MVTS classification. However, in these solar flare prediction tasks, there are serious challenges that degrade flare prediction quality, which come about for three reasons: the extreme class imbalance between minor and major flare events, the intricate temporal dynamics, and the interdependencies of magnetic field parameters. To address these challenges and improve predictive performance, obtaining meaningful data representations is a crucial part of developing effective classification models [7]. By leveraging different representations, we can assess how the structure of the input data impacts model performance in downstream tasks, particularly under conditions of extreme class imbalance and noisy data. Accordingly, in this study, we compare the effectiveness of data representation methods for representing the MVTS data instances within the SWAN-SF dataset under conditions of extreme class imbalance, undersampling, and optimized preprocessing. Three distinct methodologies merit examination: (1) a vector-based approach where MVTS instances of SWAN-SF are reduced to summary vectors, encapsulating key features of an MVTS instance; (2) the maintenance of the native MVTS structure of SWAN-SF, with each univariate time series capturing distinct magnetic field parameters over time; and (3) a graph-based representation, wherein each univariate time series, corresponding to a specific magnetic field parameter, is modeled as a node, and the interdependencies between these parameters are expressed through edges. A thorough comparison of these representations could yield deeper insights into the optimal approach for accurately predicting solar flare occurrences and guide future research through the selection of proper representations while designing state-of-the-art frameworks.
To obtain vector representations, we apply two distinct techniques. First, we extract the last timestamps of the MVTS instances, assuming that the most recent data point from each series encapsulates critical information. Second, we generate summary vectors by aggregating key statistical features across the entire time series to capture the overall temporal dynamics in a compact form. Both techniques reduce the dimensionality of the data while preserving essential characteristics for predictive modeling. For the time series representation, we maintain the full temporal resolution of the data. Each magnetic field parameter is treated as an individual univariate time series, preserving the original time dependencies between parameters. This approach allows us to leverage time series-specific classification models to exploit temporal patterns and capture long-range dependencies critical to solar flare prediction. For the graph representation, we transform the MVTS instances into graphs using two measures defining relationships between variables: correlation and distance similarity. After constructing the graphs, we employ three different techniques to obtain graph embeddings: (1) summarizing each graph using node degree information; (2) utilizing graph embedding algorithms; and (3) using graph neural networks. This allows us to capture both local and global dependencies between magnetic field parameters across the graph structure, offering a new dimension to model interactions effectively. The contributions of this paper are as follows:
  • An in-depth exploration of the utilization of three distinct data representations of the SWAN-SF dataset, where each proposed method leverages its unique attributes to enhance predictive performance.
  • The full introduction of graph-based machine learning into solar flare prediction tasks by obtaining graph representations, applying multiple graph embedding techniques, and creating a graph convolutional network module.
  • The experimental evaluation of the strengths and shortcomings of each data representation method under different conditions of sampling and preprocessing in light of the appropriate selection of performance metrics that align with the nuanced characteristics of the benchmark dataset.
The rest of this paper is organized as follows. In Section 2, we cover related work. In Section 3, we discuss the SWAN-SF dataset and our proposed data representation methods. In Section 4, we present our experimental findings. Finally, we provide concluding remarks and discuss future work in Section 5.

2. Related Work

2.1. Solar Flare Prediction

The earliest attempt at solar flare prediction was THEO, an expert system that relied on human inputs. It was officially adopted for use by the Space Environment Center (SEC) of the National Oceanic and Atmospheric Administration (NOAA) in 1987 [8]. In the following years, with the help of an abundance of magnetic field data collected by different space-based and ground-based observatories, the task of solar prediction was often treated as a data science task. Focusing on how the data are represented, we review three machine learning-driven techniques for solar flare prediction, categorizing them into vector-based, time series-based, and graph-based approaches.
Vector-based approaches have gained prominence in solar flare prediction, with line-of-sight magnetogram-based models characterizing solar active region photospheric magnetic field parameters by the line-of-sight component, and vector magnetogram-based models capturing solar active region parameters using the full-disk photospheric vector magnetic field [9]. Since 2010, the launch of NASA’s Solar Dynamics Observatory (SDO), the Helioseismic and Magnetic Imager (HMI) instrument has continuously mapped the full-disk vector magnetic field every 12 min [10]. As a result, most of the current models utilize this stream of vector magnetogram data from SDO. Nonlinear statistical models, mainly machine learning classifiers, have been widely utilized for solar flare prediction, including logistic regression [11], C4.5 decision tree [12], fully connected neural network [13], support vector machine [3], and relevance vector machine [14].
In the time series-based approach, a significant advancement was made by extending single timestamp-based models to incorporate temporal window-based flare prediction, allowing for more accurate predictions over extended time periods [6]. This resulted in the creation of SWAN-SF [5], an MVTS dataset where each instance records magnetic field data over a preset observation time, providing labeled flare classes occurring after a specific prediction time window. From that point, various MVTS classification approaches were used, such as statistical summarization for k-nearest neighbors training [2], MVTS decision trees with clustering as a preprocessing step [15], deep sequence modeling for end-to-end flare classification with automated feature learning [16], and contrastive representation learning for addressing challenges of temporal dependencies and extreme class imbalance [7,17]. Additionally, there have been recent attempts to provide comparative analysis of deep learning models that utilize full-disk line-of-sight magnetograms and active region parameters, highlighting the strengths and weaknesses of some vector-based and time series-based models in capturing the dynamics of flare-imminent active regions [18]. In other studies, the focus was extended to the diversity of sequence models for time series of active region parameters, including long short-term memory (LSTM), LSTM with attention (LSTM-A), bidirectional LSTM (BLSTM), and BLSTM with attention (BLSTM-A), to predict flare classes [19]. However, there is still limited exploration of graph models in this area of research.
While graph-based representation learning has not been explored as extensively as vector-based and time series-based approaches, there have been some efforts to apply it to solar flare prediction. One such example is the use of functional network embedding together with sequence modeling to capture both temporal and spatial relationships of MVTS instances. The graph approach is evident in the functional network embedding stage where the inter-parameter dependencies are modeled as the edges in graphs [20]. Section 2.2 provides an overview of recent advancements in graph-based machine learning techniques to obtain fixed-dimension embeddings from graph data. Overall, these diverse methodologies showcase the evolution from traditional linear models to sophisticated machine learning techniques in predicting solar flare activities.

2.2. Graph Representation Learning

Graph representation learning is a fundamental part of graph-based machine learning research, a process aimed at embedding graph structures into a lower-dimensional space while preserving the underlying relationships and dependencies between nodes and edges that make up the graph [21]. By learning these embeddings, the intricate structures of graphs (e.g., node proximity, connectivity patterns, and community structures) can be effectively captured and utilized in machine learning models [22,23]. This method allows complex graph data to be transformed into meaningful representations that can be used for a variety of downstream tasks such as node classification, link prediction, and graph classification. Among these representation learning strategies, graph embedding algorithms provide alternative ways to represent components of graph structures. One example is the matrix decomposition-based Laplacian Eigenmaps algorithm, which uses eigenvectors derived from the graph Laplacian matrix to cover the one-hop neighborhood of nodes [24]. GraRep learns one d-dimensional embedding for each k-th order proximity, integrating global structural information of the graph into the learning process [25]. Another example is tensor decomposition-based node embedding, which uses higher-order transition probability matrices to construct one or more third-order tensors [26]. DeepWalk and Node2Vec employ flexible, biased random walks that can balance between a local and global view of the network [27,28]. Moreover, neural networks adapted and modified specifically for graph tasks have increasingly gained popularity in graph representation learning. Graph neural networks (GNNs) extend traditional neural networks, enabling them to operate on graph-structured data to model complex relationships and dependencies. Following a local aggregation mechanism, the GNN processes the node embedding vector of a node by recursively aggregating the embedding vectors of local neighborhood nodes [29,30]. This novel deep learning-based approach has led to many variations. One example is the graph convolutional network (GCN), which provides a localized first-order approximation of spectral graph convolutions [31]. Another variation is Graph-SAGE, an inductive framework that leverages node feature information to efficiently generate embeddings for previously unseen data [32]. The graph attention network (GAT) introduces masked self-attentional layers to overcome the limitations of earlier methods based on graph convolutions or their approximations [33]. These advancements highlight significant achievements in graph representation learning.

3. Materials and Methods

3.1. Data Collection

Various studies aimed to utilize machine learning for the task of solar flare prediction; however, they were using point-in-time measurements [3]. As a result, it becomes difficult to ascertain whether variations in accuracy or skill score values across these studies stem from the inherent stochasticity in flare occurrence, the preprocessing and sampling approaches employed, or the execution of machine learning models [1]. The SWAN-SF dataset [5] was created to address the issues, being composed of MVTS, to enable impartial flare forecasting and contribute to significant advancements in forecasting that go beyond incremental improvements. SWAN-SF primarily relies on sensor data obtained from NASA’s SDO, specifically the SHARPs from the Joint Science Operations Center (JSOC). These patches are derived from solar vector magnetograms captured by the HMI instrument that functions as a sensor onboard SDO, continuously observing the Sun and providing detailed magnetic field data from the solar photosphere. However, the SHARPs do not contain flare information, which is instead sourced from the Geostationary Operational Environmental Satellites (GOES) operated by NOAA. GOES, equipped with X-ray and particle detectors, has been detecting solar flares since 1975 and provides a catalog of flare events, including key details like flare time, classification, and spatial location on the solar disk, thereby providing the labeling information. The dataset creation process involves three key steps. First, solar flare reports from GOES are cleaned to resolve conflicting information. Second, these flare reports are matched with solar magnetic data from the HMI, either through matching NOAA AR numbers with HARP numbers in SHARPs or using a spatiotemporal overlap procedure to align flare occurrences with HMI active region patches. Lastly, sampling biases are eliminated to ensure a balanced and accurate dataset for training machine learning models [6]. Figure 2 illustrates the data pipeline.
Data points within SWAN-SF belong to one of five distinct categories—FQ, B, C, M, and X—each representing events with varying intensity. FQ contains both flare-quiet and category A flare events. For the task of solar flare prediction, we categorize SWAN-SF instances into two main classes, nonflare (NF) and flare (F) superclasses, encompassing all mentioned categories. The NF superclass includes FQ, B, and C category instances that are either quiet or minor flaring events. Conversely, the F superclass contains M and X flare categories, which are major flare types that can lead to severe consequences, including radiation-related health risks for astronauts, damage to infrastructure, and failures of electronic equipment and signals [2,3,4].
Each instance in SWAN-SF corresponds to an MVTS slice of multiple magnetic field parameters extracted from solar active regions with a sliding window. Accordingly, for a flare having a unique ID, multiple equal-length MVTS data instances are obtained. This time frame is referred to as the observation window ( T obs ). If s i denotes the starting point of the i-th time series segment, then the subsequent ( i + 1 ) -th segment starts at s i + τ , where τ represents the step size in the sliding process. Each extracted MVTS is labeled based on the class of the most intense flare observed within a predefined temporal window following T obs . This window is known as the prediction window ( T pred ) and commences precisely at the end of T obs . In the SWAN-SF context, T obs and T pred extend over 12 and 24 h, respectively, with τ set at 1 h [1,6]. Each MVTS data instance in SWAN-SF can be represented as M ( k ) R τ × N , where 1 k K (i.e., K is the total number of MVTS instances), a collection of univariate time series corresponding to N magnetic field parameters, each having a step size length of τ . The N magnetic field parameters as univariate time series match the first 24 parameters derived in Bobra et al.’s work [3]. Table 1 provides a full list of these magnetic field parameters, including their corresponding mathematical formulas.
SWAN-SF is divided into five partitions (i.e., P1, P2, P3, P4, and P5), where each partition covers a different observation period. Due to the nature of the dataset, where major flare events occur infrequently, SWAN-SF experiences a significant class imbalance issue between NF and F examples. In our task of binary solar flare prediction, F-class examples represent positive classes, whereas NF-class examples represent negative classes. Because of the negative-class MVTS instances dominating the dataset, classification results tend to favor the majority class, resulting in low flare prediction performance. In analyzing the data for partitions 1 through 5, we observe notably low flare-to-nonflare (F:NF) ratios of 1:58, 1:62, 1:29, 1:43, and 1:75, respectively. Figure 3 demonstrates the class distributions in detail. This extreme class imbalance presents a significant challenge in predicting solar flares in flare-forecasting research. Accordingly, for a robust and goal-based evaluation of the classification results, proper evaluation metrics must be selected. These evaluation metrics are discussed under Section 4.1.

3.2. Overview: Data Representation Methods

For clarity and organization, we categorize data representations into three distinct methods and present multiple strategies within each, capturing different structural and temporal aspects of MVTS data. The vector representation method transforms MVTS instances into a fixed-length feature vector through two strategies: vector of last timestamp (VLT), which retains the final recorded state of each time series, and vector of statistical summary (VSTAT), which encapsulates key statistical properties. The time series representation method preserves temporal dependencies, treating the task as a time series classification (TSC) problem and utilizing specialized classifiers to learn from time-ordered information, capturing meaningful time-evolving patterns. The graph representation method models each univariate time series as a node within a graph, with edges encoding relationships between magnetic field parameters. This approach comprises three strategies: graph node degree (GND), which captures structural importance; graph node embedding (GNE), which learns latent representations; and graph neural networks (GNNs), which exploit complex graph-based interdependencies between features. An overview of these methods is provided in Table 2.

3.3. Method 1: Vector Representation of SWAN-SF

In this paper, we use two strategies for vector-based representation. The aim is to summarize and simplify the features of MVTS instances such that they can be effectively utilized by the downstream classifiers to produce reliable solar flare prediction results.

3.3.1. Vector of Last Timestamp Representation

Inspired by the work presented in [3], we represent vector magnetogram data by utilizing only the last timestamp of the MVTS instances for all magnetic field parameters. The reason for this choice is that the last timestamp is temporally closest to the flaring event. Parallel to this, for each MVTS instance M ( k ) R τ × N , we extract a one-dimensional feature vector V V L T ( k ) R N . After obtaining this representation vector for each MVTS instance, denoted as V V L T R K × N where K is the total number of MVTS instances, we use V V L T as input to train the downstream module in a supervised manner to provide the class prediction. Figure 4 gives an insight into the process of solar flare prediction using VLT representation.

3.3.2. Vector of Statistical Summary Representation

Inspired by the work presented in [1], we extract a limited set of descriptive statistics from each univariate time series within the MVTS instances, specifically the median, mean, standard deviation, skewness, and kurtosis. Unlike the original work, we also incorporate the point-in-time last-value feature in a manner similar to the VLT approach to have greater representation capabilities. For each MVTS instance M ( k ) R τ × N , we extract a feature vector V V S T A T ( k ) R 6 N by taking the mentioned six descriptive parameters. After obtaining this representation vector for each MVTS instance, denoted as V V S T A T R K × 6 N with K representing the total number of MVTS instances, we use V V S T A T as input to train the downstream module in a supervised setting. Figure 5 gives an insight into the process of solar flare prediction using VSTAT representation.

3.4. Method 2: Time Series Representation of SWAN-SF

Considering the MVTS data instances in their inherent modality, we leverage TSC strategy for solar flare prediction in this method. Figure 6 gives an insight into the process of solar flare prediction using time series representation. In this respect, we employ five distinct models utilizing the MVTS representation, including time series classifiers and sequence models:
  • Shapelet Transform (ST): ST aims to effectively capture multivariate features by focusing on shapelets. Shapelets are small, distinctive subsequences of a time series that are particularly effective for classification tasks. They are characterized by being phase-independent, meaning they can identify patterns regardless of their position in the time series. Shapelets are designed to capture unique and discriminative patterns that distinguish between different classes, making them a powerful tool for time series classification [34,35].
  • Time Series Forest (TSF): TSF utilizes a blend of the entropy gain and a distance measure, termed the Entrance gain, to assess splits in the tree nodes. The algorithm randomly selects features at each tree node, exhibiting a computational complexity linear to the length of a time series [36].
  • Random Convolutional Kernel Transform (ROCKET): ROCKET expands upon the recent achievements of convolutional neural networks in time series classification, employing randomly generated convolutional kernels to achieve state-of-the-art accuracy while being efficient in the use of computational resources [37].
  • Recurrent Neural Network (RNN): As a neural network designed for sequential data processing, the RNN maintains hidden states to retain information from previous inputs, making it suitable for sequence modeling tasks, including time series analysis and natural language processing [38].
  • Long Short-Term Memory (LSTM): As a specialized type of RNN that addresses the challenges of learning long-range dependencies of the input sequences, LSTM introduces memory cells and gating mechanisms to selectively store and retrieve information [39].

3.5. Method 3: Graph Representation of SWAN-SF

3.5.1. Functional Network-Based Graph Creation

Most of the existing MVTS classification methods concentrate on the temporal dependencies of the time series. Yet, in tasks of this nature, there might exist an intricate interdependence among MVTS variables. Capturing this relationship not only constitutes a noteworthy contribution to enhancing classification performance but also poses a significant challenge. Using graph-based representations is a strong candidate since complex pairwise dependencies among multivariate variables can be better described using advanced graph methods, where each variable is considered as a node in the graph, and their dependencies (e.g., positive correlations) are regarded as edges [21]. In this paper, we use two strategies to convert each MVTS instance, M ( k ) R τ × N , into graphs G = ( V , E ) , where V denotes the nodes representing the individual magnetic field parameters and E represents the edges capturing the relationships between these parameters. We aim to utilize the graph-based approach to uncover spatial properties and discover interaction patterns between magnetic field parameters within MVTS instances.
  • Correlation-Based Graph (COR) Creation: For each MVTS instance, M ( k ) R τ × N , we first calculate the Pearson correlation matrix C ( k ) R N × N such that C i , j ( k ) corresponds to the Pearson correlation coefficient value c (where 1 c 1 ) between univariate time series i ( k ) and j ( k ) corresponding to the i-th and j-th column of M ( k ) , respectively. We utilize this symmetric matrix C ( k ) as the adjacency matrix in the graph creation stage where each univariate time series is considered as a node in the graph. After applying a zero threshold to enforce the condition of only creating edges with positive weight, we create the correlation graph instances, G C ( k ) = ( V , E ) , from the correlation matrices of each MVTS instance.
  • Distance Similarity-Based Graph (DS) Creation: For each MVTS instance, M ( k ) R τ × N , we calculate the Euclidean distance matrix E ( k ) R N × N such that E i , j ( k ) corresponds to the Euclidean distance similarity measure e between univariate time series i ( k ) and j ( k ) corresponding to the i-th and j-th column of M ( k ) , respectively. In the following step, we utilize the symmetric matrix E ( k ) as the adjacency matrix in the graph creation stage where each univariate time series is considered as a node in the graph. After applying a threshold t in creating edges, we create the Euclidian graph instances, G E ( k ) = ( V , E ) , from the Euclidian distance similarity matrices of each MVTS instance.

3.5.2. Graph Node Degree Representation

After obtaining graphs from each MVTS instance M ( k ) as G C ( k ) and G E ( k ) , we vectorize the graph instances by extracting the degree information of nodes in the graph. The degree of a node can be described as the number of connections that the particular node forms with other nodes in the graph. Each graph instance yields a degree vector of D ( k ) R N containing the degrees of each node in the graph to compactly describe the graph structure. In this respect, for all graph instances in the training set, we obtain the degree matrix, D R K × N , with K being the number of graphs. We use this information as input in the training stage for the downstream module. Figure 7 gives an insight into the process of solar flare prediction using degree representation.

3.5.3. Graph Node Embedding Representation

To investigate the effect of using node embeddings on solar flare prediction, we process the created graphs G C and G E by obtaining their node embeddings. Accordingly, we select two node embedding algorithms for this task. Figure 8 gives an insight into the process of solar flare prediction using the embeddings obtained from MVTS graphs.
  • Laplacian Eigenmaps (LAP) [24]: In this method, for each created graph G ( k ) , we obtain the corresponding Laplacian matrix as the first step. In the following step, we extract the eigenvectors and eigenvalues with eigendecomposition. After sorting the eigenvalues in descending order, we select the top-d eigenvectors representing each node in the graph by d dimensions. This results in an embedding matrix, L ( k ) R N × d , for each graph. After flattening this matrix, a feature vector of size N d is used as the representation vector of the entire graph. In this respect, for all graph instances in the training set, we obtain the matrix L R K × N d , with K being the number of graphs. This information is fed into the downstream module for training purposes.
  • Node2Vec (N2V) [28]: In this method, we utilize the Node2Vec algorithm to generate d-dimensional node embeddings for each node in the graph. Node2Vec is a graph embedding algorithm that leverages random walks to explore the structure of the graph. It performs biased random walks to capture both local and global structural information. Specifically, Node2Vec generates embeddings by performing random walks starting from each node, with the walk length w fixed. During these walks, the algorithm uses two parameters, p and q, to control the trade-off between breadth-first search (BFS) and depth-first search (DFS), allowing it to capture different types of structural relationships in the graph. The resulting node sequences from the random walks are then used to train a skip-gram model to learn node representations based on their co-occurrence in the walks. As a result, for each graph G ( k ) we obtain an embedding vector N V ( k ) R N × d . After flattening the matrix and obtaining the N d -sized representation vector for each graph, N V R K × N d where K represents the number of graphs, N V is used to train the downstream module.

3.5.4. Graph Neural Network Representation

To harness the power of GNNs on graph instances, we construct a two-layer GCN module. For each training instance, the GCN requires the graph as well as the node attributes. In this module architecture, for each layer l, the GCN convolutional layer considers the l-hop neighborhood of each node. The first layer has an input channel dimension size corresponding to the node attribute size. The GCN update process for a particular node v’s representation in graph G = ( V , E ) , where V is the set of nodes and E is the set of edges in the graph, can be expressed as
h v [ 0 ] = x v
h v [ l + 1 ] = ReLU W g [ l ] u N ( v ) w u v h u [ l ] | N ( v ) | + B g [ l ] h v [ l ] , l { 0 , 1 , , L 1 }
z v = h v l
z G = 1 | V | v V z v
Let L be the total number of GCN layers, x v R τ be node v’s representation, h v [ l ] R d g be node v’s representation in layer l, W g [ l ] R d g × d g be layer l’s weight matrix, B g [ l ] R d g be the bias vector of the l t h layer, N ( v ) be the set of neighbor nodes of a particular node v, w u v be the weight for the edge between node v and neighbor node u, z v be the final node representation of v at the end of L iterations of neighborhood aggregation, and z G be the graph representation found by node-level representation averaging [20,31].
In our case, the input channel size is τ , since each node, each univariate time series, has τ values corresponding to their length. The first layer outputs n hidden channels, followed by the rectified linear unit (ReLU) activation function and dropout. ReLU introduces nonlinearity by zeroing out negative values, helping the model learn complex patterns. Dropout, on the other hand, randomly sets some of the activations to zero during training, reducing overfitting and improving generalization. The second layer has a matching input channel size of n and an output channel size of m. Followed by a node-wise average pooling, the GCN is followed by a fully connected layer in an end-to-end fashion to generate raw output values for the class labels. Then, a sigmoid function is applied to these raw outputs, converting them into probabilities suitable for a binary classification task, tailored to our solar flare prediction needs. Figure 9 gives an insight into the process of solar flare prediction using our GCN model.

4. Results

In this section, we demonstrate our experimental findings. We conducted our experiments on a Windows machine having an AMD Ryzen 7 5800H processor, NVIDIA GeForce RTX 3060 GPU, and 16 GB memory, and we used Python 3.9.13, PyTorch 2.1.0, numpy 1.25.2, scikit-learn 1.2.0, and networkx 2.8.8 with CUDA 11.8 for implementing our models. The source code, including all methods, experimentation phases, and experimentation dataset, is available in our GitHub repository (GitHub repository: https://github.com/OnurVural/Data_Representation_Analysis_SWAN-SF, accessed on 16 March 2025).

4.1. Evaluation Metrics

To evaluate the performance of the mentioned strategies in terms of their effectiveness for the binary prediction task, we selected six performance evaluation metrics: accuracy, F1 score, receiver operating characteristic area under the curve (ROC AUC), Heidke skill score (HSS2), Gilbert skill score (GS), and true skill statistic (TSS). As SWAN-SF has a significant class imbalance between the positive class of F examples and the negative class of NF examples, accuracy alone, which simply reports the number of correctly predicted classes among all examples, is not sufficient for proper evaluation of models. Because of that, metrics such as ROC AUC, HSS2, GS, and TSS are used as effective evaluation metrics in flare prediction [2,3,7,10,17,19,40,41]. ROC AUC measures a binary classifier’s ability to distinguish between positive and negative classes by evaluating its performance across all possible thresholds, reflecting how well it ranks positive samples higher than negative ones, and offering a more reliable metric than accuracy, especially in imbalanced datasets. HSS2 assesses the improvement over a random prediction scenario, and GS considers the likelihood of obtaining true positives by chance. Among all metrics, TSS shows the most effectiveness against extreme class imbalance, as it shows the difference between the true-positive rate and false-positive rate by taking values in the range of −1 and 1, with a score of 1 indicating all correct predictions, a score of −1 indicating all wrong predictions, and a score of 0 indicating random predictions. The work of Bobra et al. [3] suggested that TSS is the most effective measure for solar flare prediction. For a classification result of true positives (TP) as correctly classified F examples, true negatives (TN) as correctly classified NF examples, false positives (FP) as incorrectly classified NF examples, and false negatives (FN) as incorrectly classified F examples, the performance evaluation can be expressed as
Accuracy = T P + T N T P + T N + F P + F N
F 1 = T P T P + 1 2 ( F P + F N )
HSS 2 = 2 × [ ( T P × T N ) ( F N × F P ) ] [ ( T P + F N ) × ( F N + T N ) ] + [ ( T N + F P ) × ( T P + F P ) ]
GS = T P C T P + F P + F N C
where C = ( T P + F P ) × ( T P + F N ) T P + T N + F P + F N
TSS = T P T P + F N F P T N + F P

4.2. Preprocessing

We employed two types of preprocessing techniques in our study. First, we applied a traditional approach commonly utilized in other flare prediction research, which we implemented in both imbalanced and oversampled settings. Additionally, we incorporated an optimized preprocessing approach to address the challenges associated with the dataset for evaluating the differences in performance and effectiveness.

4.2.1. Traditional Preprocessing Approach

For the traditional approach, while preprocessing the partitions that will be used in the training stage, among NF examples, we only selected the examples belonging to the FQ category for the train sets. The reason behind this decision is the two other categories among NF examples, namely, B and C categories, showing notable magnetic field similarities to M and X major flare types. These similarities emerge because of shared traits in the fundamental patterns and behaviors of solar activity [1]. As a result, in the training phase of classifiers to distinguish between F and NF examples, the presence of categories B and C among NF examples introduces a challenge. The classifier may face difficulties in terms of discerning between genuine flare instances and those that show a resemblance to major flare types within the NF class. As a subsequent step, to avoid skewing the data representations, we eliminated each MVTS instance containing null values in any of their entries corresponding to selected magnetic field parameters. Moving forward in the process, we performed Z-score normalization on each MVTS instance M ( k ) , ensuring that the magnetic field parameter values were standardized to a common scale. The normalization can be expressed as
m n < t > = m n < t > μ n σ n ,
In Equation (11), m n < t > is the t-th timestamp of the n-th magnetic field parameter time series of the k-th MVTS instance M ( k ) R τ × N , where 1 n N , N being the number of magnetic field parameters, and 1 t τ , τ being the time series length. μ n is the mean value of time series n, and σ n is the standard deviation of time series n of instance k.

4.2.2. Optimized Preprocessing Approach

For the optimized approach, we followed multiple steps to address the chronic challenges present in SWAN-SF such as class imbalance, class overlap, a high percentage of missing values, varying scales, and varying skewness. The preprocessing pipeline began with fast Pearson correlation-based k-nearest neighbors (FPCKNN) imputation [42], which handled missing data by estimating values based on the similarity of instances. Following this, a combination of log normalization, square root normalization, Box–Cox normalization, Z-score normalization, and min–max normalization (LSBZM) were applied to standardize the dataset, ensuring that all features have a zero mean and unit variance, thereby addressing varying scales and skewness [43]. To mitigate class overlap issues, the elimination of minor flare category samples allowed for emphasis on more distinct class characteristics between F and NF classes. Next, TimeGAN, a generative adversarial network, generated synthetic time series data that retain the temporal dynamics of the original dataset [44]. The Tomek links method was then applied to clean the dataset by removing ambiguous instances that lie between classes, thereby clarifying class boundaries. Finally, random undersampling was employed to address class imbalance by randomly removing instances from majority classes, resulting in a more equitable distribution of classes [45].

4.3. Analysis of Created Graphs

Here, we analyze the Pearson correlation and Euclidean distance graphs to explore the relationships among univariate time series instances, each representing solar magnetic field parameters. For the created Pearson correlation graphs, Figure 10 displays the inter-parameter connections to provide a focused view into the structures of graph classes. Our first observation is that the F and NF graphs express different connections for magnetic field parameters as expected. For example, in the selected F-class graph, the TOTZ parameter forms significantly more connections, resulting in a higher node degree compared to the selected NF-class graph, where the TOTZ parameter has relatively fewer connections with other magnetic field parameters, leading to a lower node degree. Figure 11 shows various randomly selected graphs to provide an overview.
For Euclidean distance graphs, in the process of edge creation between nodes, we performed experiments to enforce that the created graph had the lowest possible edge density while still being connected (i.e., without having isolated nodes). The reason behind this choice was to enforce a high similarity condition between univariate time series instances as nodes get connected with edges. As seen in Figure 12, with the increase in threshold, the graph structure becomes highly connected where the average degree and graph density become high. This creates a challenge for the classifier models since with an increasing number of edges, the models will have difficulty distinguishing the most important interaction patterns. In contrast, with the decrease in threshold, the graph starts losing the fully connected property. Therefore, selecting the threshold enforces keeping the graph structure fully connected and keeping the edge density low to filter the most valuable structural information. Accordingly, the graph instances G E ( k ) were created from the Euclidean distance matrices of each MVTS instance with a threshold t of 10 for imbalanced and undersampled settings and with a threshold t of 2 for optimized settings.
Figure 13 displays the difference in inter-parameter connections between F and NF graphs. The selected F-class graph exhibits a structure that contains two densely connected subgraphs bridged by a sparse connection. These two components represent clusters of parameters with high intra-cluster connectivity, reflecting strong correlations within each group. The presence of two such clusters suggests that the parameters within each group are strongly interdependent, while the interaction between the two clusters is relatively limited, leading to this topology. In contrast, the NF-class graph features a more uniformly connected structure, resembling a single connected component with fewer interactions overall. This suggests a more diffuse correlation pattern, where the parameters do not form distinct, highly connected groups but instead exhibit weaker, more distributed interactions across the entire graph. The reduced density of connections indicates that parameter interactions in the NF graph are more sparsely distributed, with fewer instances of strong correlations between specific parameters. Figure 14 shows various randomly selected graphs to provide an overview, which further confirms the presence of the mentioned structural patterns consistently observed in multiple F and NF graphs.

4.4. Training and Hyperparameter Settings

In this paper, we used MVTS instances that belong to consecutive partitions for training and testing purposes, respectively (e.g., training the models with MVTS instances of P1 and testing the models with MVTS instances of P2). We included all five partitions and thus had a total of four temporally consecutive train–test pairs (i.e., P1–P2, P2–P3, P3–P4, and P4–P5). We used five selected downstream classifiers to train and test our vector-based models. The support vector machine (SVM) classifier was trained with a radial basis function (RBF) kernel, a regularization parameter C of 1.0, and a tolerance of 0.001 for the stopping criteria. The k-nearest neighbors (KNN) classifier was trained with k of five nearest neighbors, using uniform weights so that each neighbor contributed equally to the prediction and applying the Minkowski distance with p of 2, equivalent to Euclidean distance. The multilayer perceptron (MLP) classifier was trained with a hidden layer size of (100, 50) and a maximum iteration of 1000. This model used a ReLU activation function for the hidden layers and the Adam optimizer with a learning rate of 0.001 set to a constant schedule. The logistic regression (LR) classifier was trained with a penalty of l2 norm and maximum iterations of 100,000. Lastly, the decision tree (DT) classifier was trained using the Gini impurity criterion for splits with no maximum depth restriction, allowing it to grow until all leaves were pure or the minimum sample split of two was reached. For the VSTAT models, we repeated our experiments both without normalization (VSTAT) and with Z-score normalization (VSTATN) separately to generate more diverse results and assess the impact of normalization on model performance.
To train our time series-based models, we applied the following hyperparameter settings. The ST model was trained using a Rotation Forest [46] with three estimators, 100 shapelet samples, a maximum of 10 shapelets per class, and a batch size of 20. The TSF model was trained using 100 estimators and a minimum interval length of three. The ROCKET model was trained using 500 kernels. The RNN and LSTM models had an input size of 24 (corresponding to the number of time series features) and a hidden dimension of 64, followed by a fully connected layer with an output size of one for binary classification. They were trained for 10 epochs with binary cross-entropy loss, using the Adam optimizer with a learning rate of 10 3 (selected from a search space of [ 10 1 , 10 2 , 10 3 , 10 4 ]).
To train our graph-based models, we used the following settings. For the GND and GNE models, the same downstream classifiers—SVM, KNN, MLP, LR, and DT—were used with configurations identical to those in the vector-based method as alternative downstream module selections. In obtaining the embedding stage, the Laplacian embedding extraction process was configured with a dimension d of 14, which was chosen after our experiments for its ability to provide over 75% representation power for the graphs. Node2Vec embedding process was configured with a walk length of 10, p of 1, q of 1, and d of 32. For our GNN approach, the first layer had an input channel size of 60 (corresponding to MVTS timestamp length) and a hidden channel size of 64, the second layer had an input channel size of 64 and an output channel size of 32 followed by the fully connected layer. The module was trained for 20 epochs with binary cross-entropy loss, using the Adam optimizer with a learning rate of 10 2 (selected from a search space of [ 10 1 , 10 2 , 10 3 , 10 4 ]).

4.5. Solar Flare Prediction Performance

In the experimentation phase, we compare the solar flare prediction performance of our proposed strategies belonging to the three data representation methods discussed in Section 3 on the benchmark dataset under three different settings: under extreme class imbalance (i.e., as present in the dataset), under undersampling, and finally, under optimized preprocessing. The vector representation strategies include VLT and VSTAT, while the time series representation strategy is denoted as TSC, and graph representation consists of GND, GNE, and GNN. All models that belong to those strategies are trained and tested with data instances under consecutive partitions of the SWAN-SF dataset (i.e., P1–P2, P2–P3, P3–P4, and P4–P5), and their performance are evaluated with the performance evaluation metrics in Section 4.1. Each classifier model is designated by a codename structured as <ClassifierName>-<DataRepresentation>. The first component of the codename corresponds to the classifier employed (e.g., SVM, KNN, TSF, LSTM, or GCN). The second component refers to the specific data representation utilized (e.g., VSTAT for statistical summary vector, CORLAP for correlation-based graph node embeddings via Laplacian Eigenmaps, or DSN2V for distance similarity-based graph node embeddings via Node2Vec). For convenient reference, a comprehensive list of classifiers and data representations is provided in Table 3.
Evaluation is presented in three views: (1) a summary of overall model performance, which provides a clear comparison across all metrics; (2) a visualization of performance distribution, showing which data representation method shows the most stable performance and less variability as a group; and (3) a detailed breakdown of results by partition, highlighting individual performance differences that are not captured in the overall summary.

4.5.1. Solar Flare Prediction Performance Under Extreme Class Imbalance

After the experiments under extreme class imbalance, the test results for all train and test partition pairs are averaged and presented in Table 4. Results documented in Table 4 show that, as expected, most methods tend to show high-accuracy results since class imbalance produces a tendency towards always predicting the negative class. The highest average accuracy performance is shared among various classifiers that are trained with different data representation methods, namely, SVM-CORLAP, LR-CORLAP, LR-DSLAP, LR-CORN2V, LR-DSN2V, LR-VLT, and LR-VSTAT. The high-accuracy results are not necessarily reflective of good performance. Therefore, in our experiments, we seek insight into other evaluation metrics that help to recognize the real performance of the classifiers trained with different data representations. Accordingly, the results demonstrate that under crude conditions of extreme class imbalance and no application of sampling techniques, vector-based data representation is the most suitable method for solar flare prediction tasks. The vector-based method achieves superior performance in four metrics, as SVM-VLT achieves an average TSS of 0.6635 and an average ROC AUC score of 0.8318. SVM-VSTAT also shows a similarly strong performance in the same metrics by achieving an average TSS of 0.6621 and an average ROC AUC score of 0.831. LR-VSTATN is the best performer in the other two metrics, where it shows an average HSS2 of 0.3871 and an average F1 score of 0.3999. The highest average GS performance is produced by the sequence model time series representation method, namely, RNN, achieving a GS of 0.602.
After identifying the top performers across all selected performance metrics, we use Figure 15 to group the three methods and visualize the distribution of results to highlight the average performance across the methods. Upon examination of Figure 15, it is evident that the vector representation method consistently outperforms other methods overall, exhibiting higher median performance across TSS, HSS2, F1, GS, and ROC AUC metrics excluding accuracy. The time series representation method comes next in TSS, HSS2, F1, GS, and ROC AUC metrics and shows the highest median performance in accuracy, making the graph representation the least effective method in all five performance metrics, only being able to show the second highest median performance in accuracy. Observing the distribution, the graph representation method has the lowest overall variability, followed by time series and vector representation methods. These findings underscore the importance of considering both the central tendency and variability of performance metrics when evaluating the effectiveness of different data representation methods in solar flare prediction tasks.
After investigating the distributional characteristics of the results and showcasing the superior performance of the vector-based method, our analysis proceeds with a direct focus on the TSS, the most resilient metric against class imbalance. This examination aims to showcase the TSS performances of each model in each train–test partition pair. Accordingly, in Figure 16, when we concentrate on the best performers (Table 5), for the train–test pair of P1–P2, the top five TSS results are achieved by SVM-VLT, SVM-VSTAT, DT-VSTAT, DT-VLT, and LR-VSTATN, with TSS results of 0.6027, 0.5998, 0.5657, 0.5431, and 0.5199, respectively. For the train–test pair of P2–P3, the top five TSS results are achieved by SVM-VSTAT, SVM-VLT, DT-VSTAT, DT-VLT, and TSF-TS with values of 0.6661, 0.6488, 0.5671, 0.5534, and 0.4578, respectively. For the train–test pair of P3–P4, the top five TSS results are achieved by SVM-VLT, SVM-VSTAT, LR-VSTATN, DT-VSTAT, and KNN-VSTATN with values of 0.7957, 0.7461, 0.6037, 0.5688 and 0.5564, respectively. For the train–test pair of P4–P5, the top five TSS results are achieved by SVM-VSTAT, DT-VLT, LR-VSTATN, SVM-VLT, and DT-VSTAT with values of 0.6365, 0.6246, 0.6197, 0.6069, and 0.5268, respectively. Based on the findings, it can be inferred that SVM-VSTAT and SVM-VLT consistently compete to achieve top performance in attaining high TSS scores.
In the culmination of our experiments, the vector-based representation method consistently emerges as the optimal choice for addressing the challenges posed by the extreme class imbalance inherent in the SWAN-SF dataset. Within this category, both VLT and VSTAT representations exhibit noteworthy performance. The SVM-VLT and SVM-VSTAT representations consistently demonstrate robust representation capabilities, positioning them prominently within the dataset analysis. These findings affirm not only the efficacy of the vector-based representation methods but also the prowess of SVM as the preferred downstream classifier in our experimental context. This collective evidence emphasizes the resilience and generalizability of the selected representation methods across diverse train–test partition pairs. The comparison between SVM-VLT and SVM-VSTAT, spanning various partition pairs, underscores their versatility and effectiveness in capturing meaningful patterns within the dataset.

4.5.2. Solar Flare Prediction Performance Under Undersampling

In the next phase of our experimental evaluation, we utilize the balanced SWAN-SF dataset, where class imbalance issues are already addressed by undersampling the majority class instances to a 1:1 ratio. For each train partition, we randomly select MVTS instances within the NF group corresponding to the same number of instances as the F group. The sizes of the undersampled training sets are presented in Table 6. Table 7 presents the averaged test results of all train and test partition pairs after conducting the experiments under undersampling. The first observation from the table indicates that, overall, the implementation of undersampling techniques results in improved performance across the evaluation metrics. Moreover, TSC models demonstrated a significant leap in performance, with all of them performing markedly better than their counterparts. This suggests that by reducing the size of the majority class, we enhance the model’s ability to focus on the underrepresented minority class. This finding underscores the importance of addressing class imbalance in solar flare tasks, as undersampling can help mitigate the risks of bias and overfitting, ultimately contributing to more robust and reliable model predictions. Regarding the results, the vector method attains the top performance in three metrics, the time series method attains the top performance in three metrics, and the graph method attains the top performance in three metrics. The highest average accuracy of 0.979 is achieved by LR-VLT and LR-VSTAT, clearly shown to be a result of overfitting. The highest average TSS score of 0.739 is achieved by TSF-TS, closely followed by SVM-VSTAT and SVM-VLT, which attained average TSS scores of 0.7286 and 0.7248, respectively. In both HSS2 and F1, DT-VSTAT stands out with the highest average performance, achieving scores of 0.2652 and 0.2899, respectively. In GS, SVM-CORN2V achieves the highest average performance with a score of 0.0221. Finally, in terms of ROC AUC, TSF-TS leads with a score of 0.8695, while SVM-VLT and SVM-VSTAT follow with scores of 0.8644 and 0.8624, respectively.
Figure 17 shows the result distribution after identifying the top performers to emphasize the average performance across the methods. Similar to the extreme class imbalance setting, the vector representation method continues to be the top performer, demonstrating higher median performance across all six metrics, followed by the time series representation method, while the graph representation method is identified as the least effective. Considering the variability in performance distribution, the time series representation method shows a broader performance range, with the best models making significant progress compared to previous settings. However, vector-based methods remain superior overall.
Figure 18 displays the results after comparing the TSS performances of each model in each train–test partition pair. Focusing on the best performers (Table 5), for the train–test pair of P1–P2, the top five TSS results are achieved by TSF-TS, ROCKET-TS, SVM-VLT, SVM-VSTAT, and DT-VSTAT, with TSS values of 0.6965, 0.6154, 0.6027, 0.5998, and 0.5657, respectively. For the train–test pair of P2–P3, the top five TSS results are achieved by DT-VSTAT, SVM-VSTAT, SVM-VLT, DT-VLT, and KNN-VSTATN with TSS values of 0.7526, 0.7404, 0.7375, 0.7348, and 0.7082, respectively. For the train–test pair of P3–P4, the top five TSS results are achieved by SVM-VLT, SVM-VSTAT, TSF-TS, DT-VSTAT, and ROCKET-TS with TSS values of 0.7885, 0.7826, 0.7564, 0.7465, and 0.6979, respectively. For the train–test pair of P4–P5, the top five TSS results are achieved by TSF-TS, SVM-VLT, SVM-VSTAT, DT-VSTAT, and LR-VSTATN with TSS values of 0.7891, 0.7857, 0.7762, 0.7174, and 0.6974, respectively.
In the culmination of our experiments under undersampled conditions, while the vector representation method demonstrates the most consistent performance across models, both the vector-based and time series-based approaches exhibit noteworthy results. In particular, the best performers in previous imbalanced settings, the SVM-based vector models—SVM-VLT and SVM-VSTAT—continue to show promising solar flare prediction performance, with improvements across all metrics. Additionally, the vector-based DT-VSTAT model achieves the highest F1 and GS scores. On the other hand, the time series-based TSF-TS model outperforms in terms of TSS and ROC AUC scores. Ultimately, the choice of the best model depends on the specific metric of interest and the task’s requirements, as each method offers distinct strengths. The results consistently show that the graph-based method is the least effective across all measures.

4.6. Solar Flare Prediction Performance Under Optimized Preprocessing

In the final stage of our experimental evaluation, we assess the solar flare prediction performance of our models utilizing the optimized SWAN-SF dataset, which has undergone a distinct set of preprocessing techniques, including advanced normalization and a combination of undersampling and oversampling methods. The sizes of the new training sets are presented in Table 8.
The averaged test results of all train and test partition pairs are presented in Table 9. Out of the six metrics evaluated, vector-based models demonstrate the highest performance in four metrics, while time series-based models excel in the remaining metrics. The SVM-VSTAT model achieves the highest accuracy at 0.9525. Regarding the GS and ROC AUC metrics, the RNN-TS model performs best, recording scores of 0.023 and 0.8764, respectively. The LR-VSTAT model excels in the TSS, HSS2, and F1 metrics, with scores of 0.7415, 0.2407, and 0.2718, thus demonstrating the most prominent performance in optimized preprocessing settings. We can infer that among time series models, the sequence models LSTM-TS and RNN-TS exhibit a significant leap in performance, likely resulting from the optimized preprocessing that leverages the temporal characteristics of SWAN-SF. Conversely, there is a noticeable decrease in previous top performers belonging to the vector representation, such as SVM and DT models, indicating a shift in effectiveness under these conditions. Another noteworthy observation is that the GNN-based method demonstrates commendable performance for the first time in this setting, with GCN-DS using distance similarity-based graph inputs. This finding will be a focal point for future research aimed at understanding the underlying reasons for this improvement and exploring strategies to enhance the resilience and robustness of graph-based methods.
Figure 19 shows the median performance of three data representation methods after we identify the top performers. In the optimized setting, the time series representation method demonstrates the best median performance with low variability, surpassing the vector-based methods in all six metrics, while the graph-based representation method consistently ranks the lowest.
We compare the TSS performances of each model in each train–test partition pair, and Figure 20 displays the results. Regarding the top performers (Table 5), for the train–test pair of P1–P2, the top five TSS results are achieved by RNN-TS, LSTM-TS, LR-VSTAT, KNN-VLT, and LR-VLT, with TSS values of 0.8036, 0.8015, 0.7819, 0.7585, and 0.7061, respectively. For the train–test pair of P2–P3, the top five TSS results are achieved by TSF-TS, LR-VSTAT, RNN-TS, KNN-VLT, and ST-TS with TSS values of 0.768, 0.7445, 0.6899, 0.6628, and 0.6551, respectively. For the train–test pair of P3–P4, the top five TSS results are achieved by LR-VLT, RNN-TS, LR-VSTAT, TSF-TS, and LSTM-TS with TSS values of 0.8094, 0.7648, 0.7117, 0.6875, and 0.6671, respectively. For the train–test pair of P4–P5, the top five TSS results are achieved by TSF-TS, LR-VSTAT, RNN-TS, ST-TS, and KNN-VLT with TSS values of 0.7645, 0.7278, 0.6873, 0.6464, and 0.6434, respectively.
In the final analysis of our experiments under optimized preprocessing conditions, the time series representation method exhibits the most consistent performance, particularly with sequence models. While the vector representation models that previously performed best tend to underperform in this new setting, the vector representation still manages to secure the top results with the LR-VSTAT model. The findings indicate that the graph-based approach remains the least effective across all evaluation criteria in this context as well, rendering the graph representation a suboptimal choice for solar flare prediction and analysis tasks.

5. Discussion

In this work, we evaluated three solar flare prediction approaches regarding data representations of the SWAN-SF dataset: vector-based, time series-based, and graph-based methods. The objective of this study was to provide a comprehensive guide for selecting machine learning-oriented models and components in the field of solar flare prediction research, focusing on various conditions under which different models and data representations are effectively chosen. In this context, we aimed to identify the data representation method that most effectively represents the characteristics of the solar flare dataset based on photospheric magnetic field parameters under three settings: imbalanced, undersampled, and custom-preprocessed. The vector representation intended to encapsulate the summarization of the multivariate time series, the time series representation focused on capturing temporal patterns, and the graph representation aimed to elucidate inter-parameter dependencies. The key findings of this study, as demonstrated by experimental validation, are as follows:
(1)
Under extreme class imbalance conditions, superior performance was achieved by the vector-based approach, with SVM-VLT and SVM-VSTAT being the best performers in this method.
(2)
Under undersampling conditions (i.e., when the training partition is balanced through undersampling), the best-performing vector and time series models demonstrated competitive performance. Specifically, the vector-based SVM-VLT and SVM-VSTAT models continued to show high performance, the DT-VSTAT model achieved the highest F1 and GS scores, and the time series-based TSF model excelled in TSS and ROC AUC scores. Ultimately, the utilization of these models depends on the specific metric of interest.
(3)
Under optimized preprocessing conditions, the vector representation still secured leading outcomes with the LR-VSTAT model, while the time series representation method exhibited high consistency in performance, particularly with sequence models.
(4)
In all settings, vector-based methods proved to be highly resilient in solar flare prediction tasks when coupled with an appropriate downstream classifier, demonstrating robust prediction results by effectively capturing complex relationships and summarizing multivariate data.
In future studies, we aim to expand the scope of data representations and investigate their effects under extended conditions, as this approach is deemed highly valuable and is expected to significantly influence model selections in subsequent research. Furthermore, we will experiment with different graph features, such as the clustering coefficient [47] and graphlet kernel [48,49], to yield better results. In addition, we have yet to explore the application of other normalization and preprocessing techniques or the utilization of node clustering methods to further investigate the relation between photospheric magnetic field parameters. Finally, we plan a specific study to improve the representation power of graph-based approaches in the SWAN-SF dataset by combining three types of features to capture temporal, spatial, and both higher-order and lower-order details. In this respect, MVTS data, the Pearson correlation matrix, and frequent subgraph similarity results will compose the feature space where we will select the most statistically significant features to train downstream classifiers.

Author Contributions

Conceptualization, O.V. and S.M.H.; methodology, O.V.; software, O.V.; validation, O.V.; formal analysis, O.V.; investigation, O.V.; resources, O.V., S.M.H., and S.F.B.; data curation, O.V.; writing—original draft preparation, O.V.; writing—review and editing, O.V. and S.M.H.; visualization, O.V.; supervision, S.M.H. and S.F.B.; project administration, O.V. and S.M.H.; funding acquisition, S.M.H. and S.F.B. All authors have read and agreed to the published version of the manuscript.

Funding

This project has been supported in part by funding from the Division of Atmospheric and Geospace Sciences within the Directorate for Geosciences, under NSF awards #2301397, #2204363, and #2240022, and by funding from the Office of Advanced Cyberinfrastructure within the Directorate for Computer and Information Science and Engineering, under NSF award #2305781.

Data Availability Statement

The SWAN-SF dataset, utilized in this study’s experiments, is publicly accessible via https://dataverse.harvard.edu/dataset.xhtml?persistentId=doi:10.7910/DVN/EBCFKM (accessed on 30 January 2025).

Acknowledgments

The authors acknowledge all those involved with the GOES missions as well as the SDO mission.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Ahmadzadeh, A.; Aydin, B.; Georgoulis, M.K.; Kempton, D.J.; Mahajan, S.S.; Angryk, R.A. How to train your flare prediction model: Revisiting robust sampling of rare events. Astrophys. J. Suppl. Ser. 2021, 254, 23. [Google Scholar] [CrossRef]
  2. Hamdi, S.M.; Kempton, D.; Ma, R.; Boubrahimi, S.F.; Angryk, R.A. A time series classification-based approach for solar flare prediction. In Proceedings of the 2017 IEEE International Conference on Big Data (Big Data), Boston, MA, USA, 11–14 December 2017; pp. 2543–2551. [Google Scholar]
  3. Bobra, M.G.; Couvidat, S. Solar flare prediction using SDO/HMI vector magnetic field data with a machine-learning algorithm. Astrophys. J. 2015, 798, 135. [Google Scholar] [CrossRef]
  4. Eastwood, J.; Biffis, E.; Hapgood, M.; Green, L.; Bisi, M.; Bentley, R.; Wicks, R.; McKinnell, L.A.; Gibbs, M.; Burnett, C. The economic impact of space weather: Where do we stand? Risk Anal. 2017, 37, 206–218. [Google Scholar] [CrossRef] [PubMed]
  5. Angryk, R.; Martens, P.; Aydin, B.; Kempton, D.; Mahajan, S.; Basodi, S.; Ahmadzadeh, A.; Cai, X.; Filali Boubrahimi, S.; Hamdi, S.M.; et al. SWAN-SF; Harvard Dataverse: Cambridge, MA, USA, 2020. [Google Scholar] [CrossRef]
  6. Angryk, R.A.; Martens, P.C.; Aydin, B.; Kempton, D.; Mahajan, S.S.; Basodi, S.; Ahmadzadeh, A.; Cai, X.; Filali Boubrahimi, S.; Hamdi, S.M.; et al. Multivariate time series dataset for space weather data analytics. Sci. Data 2020, 7, 227. [Google Scholar] [CrossRef]
  7. Vural, O.; Hamdi, S.M.; Filali Boubrahimi, S. EXCON: Extreme Instance-based Contrastive Representation Learning of Severely Imbalanced Multivariate Time Series for Solar Flare Prediction. In Proceedings of the 2024 IEEE International Conference on Big Data (BigData), Washington, DC, USA, 15–18 December 2024; pp. 1476–1483. [Google Scholar] [CrossRef]
  8. McIntosh, P.S. The classification of sunspot groups. Sol. Phys. 1990, 125, 251–267. [Google Scholar] [CrossRef]
  9. Boubrahimi, S.F.; Aydin, B.; Kempton, D.; Angryk, R. Spatio-temporal interpolation methods for solar events metadata. In Proceedings of the 2016 IEEE International Conference on Big Data (Big Data), Washington, DC, USA, 5–8 December 2016; pp. 3149–3157. [Google Scholar]
  10. Mason, J.P.; Hoeksema, J. Testing automated solar flare forecasting with 13 years of Michelson Doppler Imager magnetograms. Astrophys. J. 2010, 723, 634. [Google Scholar] [CrossRef]
  11. Song, H.; Tan, C.; Jing, J.; Wang, H.; Yurchyshyn, V.; Abramenko, V. Statistical assessment of photospheric magnetic features in imminent solar flare predictions. Sol. Phys. 2009, 254, 101–125. [Google Scholar] [CrossRef]
  12. Yu, D.; Huang, X.; Wang, H.; Cui, Y. Short-term solar flare prediction using a sequential supervised learning method. Sol. Phys. 2009, 255, 91–105. [Google Scholar] [CrossRef]
  13. Ahmed, O.W.; Qahwaji, R.; Colak, T.; Higgins, P.A.; Gallagher, P.T.; Bloomfield, D.S. Solar flare prediction using advanced feature extraction, machine learning, and feature selection. Sol. Phys. 2013, 283, 157–175. [Google Scholar] [CrossRef]
  14. Al-Ghraibah, A.; Boucheron, L.; McAteer, R. An automated classification approach to ranking photospheric proxies of magnetic energy build-up. Astron. Astrophys. 2015, 579, A64. [Google Scholar] [CrossRef]
  15. Ma, R.; Boubrahimi, S.F.; Hamdi, S.M.; Angryk, R.A. Solar flare prediction using multivariate time series decision trees. In Proceedings of the 2017 IEEE International Conference on Big Data (Big Data), Boston, MA, USA, 11–14 December 2017; pp. 2569–2578. [Google Scholar]
  16. Muzaheed, A.A.M.; Hamdi, S.M.; Boubrahimi, S.F. Sequence model-based end-to-end solar flare classification from multivariate time series data. In Proceedings of the 2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA), Boston, MA, USA, 11–14 December 2021; pp. 435–440. [Google Scholar]
  17. Vural, O.; Hamdi, S.M.; Boubrahimi, S.F. Contrastive Representation Learning for Predicting Solar Flares from Extremely Imbalanced Multivariate Time Series Data. In Proceedings of the 2024 International Conference on Machine Learning and Applications (ICMLA), Miami, FL, USA, 18–20 December 2024; pp. 1077–1082. [Google Scholar] [CrossRef]
  18. Sun, Z.; Bobra, M.G.; Wang, X.; Wang, Y.; Sun, H.; Gombosi, T.; Chen, Y.; Hero, A. Predicting solar flares using CNN and LSTM on two solar cycles of active region data. Astrophys. J. 2022, 931, 163. [Google Scholar] [CrossRef]
  19. Zheng, Y.; Qin, W.; Li, X.; Ling, Y.; Huang, X.; Li, X.; Yan, P.; Yan, S.; Lou, H. Comparative analysis of machine learning models for solar flare prediction. Astrophys. Space Sci. 2023, 368, 53. [Google Scholar] [CrossRef]
  20. Hamdi, S.M.; Ahmad, A.F.; Filali Boubrahimi, S. Multivariate time series-based solar flare prediction by functional network embedding and sequence modeling. In Proceedings of the CIKM workshop for Applied Machine Learning Methods for Time Series Forecasting (AMLTS 2022), Atlanta, GA, USA, 21 October 2022. [Google Scholar]
  21. Duan, Z.; Xu, H.; Wang, Y.; Huang, Y.; Ren, A.; Xu, Z.; Sun, Y.; Wang, W. Multivariate time-series classification with hierarchical variational graph pooling. Neural Netw. 2022, 154, 481–490. [Google Scholar] [CrossRef] [PubMed]
  22. Cai, H.; Zheng, V.W.; Chang, K.C.C. A comprehensive survey of graph embedding: Problems, techniques, and applications. IEEE Trans. Knowl. Data Eng. 2018, 30, 1616–1637. [Google Scholar] [CrossRef]
  23. Goyal, P.; Ferrara, E. Graph embedding techniques, applications, and performance: A survey. Knowl.-Based Syst. 2018, 151, 78–94. [Google Scholar] [CrossRef]
  24. Ng, A.; Jordan, M.; Weiss, Y. On spectral clustering: Analysis and an algorithm. Adv. Neural Inf. Process. Syst. 2001, 14, 1–8. [Google Scholar]
  25. Cao, S.; Lu, W.; Xu, Q. Grarep: Learning graph representations with global structural information. In Proceedings of the 24th ACM International on Conference on Information and Knowledge Management, Melbourne, Australia, 19–23 October 2015; pp. 891–900. [Google Scholar]
  26. Hamdi, S.M.; Filali Boubrahimi, S.; Angryk, R. Tensor decomposition-based node embedding. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, Beijing, China, 3–7 November 2019; pp. 2105–2108. [Google Scholar]
  27. Perozzi, B.; Al-Rfou, R.; Skiena, S. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 701–710. [Google Scholar]
  28. Grover, A.; Leskovec, J. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 855–864. [Google Scholar]
  29. Scarselli, F.; Gori, M.; Tsoi, A.C.; Hagenbuchner, M.; Monfardini, G. The graph neural network model. IEEE Trans. Neural Netw. 2008, 20, 61–80. [Google Scholar] [CrossRef]
  30. Zhou, J.; Cui, G.; Hu, S.; Zhang, Z.; Yang, C.; Liu, Z.; Wang, L.; Li, C.; Sun, M. Graph neural networks: A review of methods and applications. AI Open 2020, 1, 57–81. [Google Scholar] [CrossRef]
  31. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  32. Hamilton, W.; Ying, Z.; Leskovec, J. Inductive representation learning on large graphs. Adv. Neural Inf. Process. Syst. 2017, 30, 1–11. [Google Scholar]
  33. Velickovic, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. Stat 2017, 1050, 10-48550. [Google Scholar]
  34. Lines, J.; Davis, L.M.; Hills, J.; Bagnall, A. A shapelet transform for time series classification. In Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Beijing, China, 12–16 August 2012; pp. 289–297. [Google Scholar]
  35. Bostrom, A.; Bagnall, A. A shapelet transform for multivariate time series classification. arXiv 2017, arXiv:1712.06428. [Google Scholar]
  36. Deng, H.; Runger, G.; Tuv, E.; Vladimir, M. A time series forest for classification and feature extraction. Inf. Sci. 2013, 239, 142–153. [Google Scholar] [CrossRef]
  37. Dempster, A.; Petitjean, F.; Webb, G.I. ROCKET: Exceptionally fast and accurate time series classification using random convolutional kernels. Data Min. Knowl. Discov. 2020, 34, 1454–1495. [Google Scholar] [CrossRef]
  38. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning Internal Representations by Error Propagation; Stanford University: Stanford, CA, USA, 1985. [Google Scholar]
  39. Hochreiter, S.; Schmidhuber, J. Long Short-term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  40. Barnes, G.; Leka, K. Evaluating the performance of solar flare forecasting methods. Astrophys. J. 2008, 688, L107. [Google Scholar] [CrossRef]
  41. Ji, A.; Aydin, B.; Georgoulis, M.K.; Angryk, R. All-clear flare prediction using interval-based time series classifiers. In Proceedings of the 2020 IEEE International Conference on Big Data (Big Data), Atlanta, GA, USA, 10–13 December 2020; pp. 4218–4225. [Google Scholar]
  42. Batista, G.E.; Monard, M.C. A study of K-nearest neighbour as an imputation method. His 2002, 87, 48. [Google Scholar]
  43. EskandariNasab, M.; Hamdi, S.M.; Boubrahimi, S.F. Impacts of data preprocessing and sampling techniques on solar flare prediction from multivariate time series data of photospheric magnetic field parameters. Astrophys. J. Suppl. Ser. 2024, 275, 6. [Google Scholar] [CrossRef]
  44. Yoon, J.; Jarrett, D.; van der Schaar, M. Time-series Generative Adversarial Networks. In Proceedings of the Advances in Neural Information Processing Systems; Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2019; Volume 32. [Google Scholar]
  45. Elhassan, T.; Aljurf, M. Classification of imbalance data using tomek link (t-link) combined with random under-sampling (rus) as a data reduction method. Glob. J. Technol. Optim. 2016, 1, 2016. [Google Scholar]
  46. Rodriguez, J.; Kuncheva, L.; Alonso, C. Rotation Forest: A New Classifier Ensemble Method. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 1619–1630. [Google Scholar] [CrossRef]
  47. Masuda, N.; Sakaki, M.; Ezaki, T.; Watanabe, T. Clustering coefficients for correlation networks. Front. Neuroinform. 2018, 12, 7. [Google Scholar] [CrossRef] [PubMed]
  48. Shervashidze, N.; Vishwanathan, S.V.N.; Petri, T.; Mehlhorn, K.; Borgwardt, K.M. Efficient graphlet kernels for large graph comparison. In Proceedings of the International Conference on Artificial Intelligence and Statistics, Clearwater Beach, FL, USA, 16–18 April 2009. [Google Scholar]
  49. Shervashidze, N.; Schweitzer, P.; van Leeuwen, E.J.; Mehlhorn, K.; Borgwardt, K.M. Weisfeiler-Lehman Graph Kernels. J. Mach. Learn. Res. 2011, 12, 2539–2561. [Google Scholar]
Figure 1. NASA’s Solar Dynamics Observatory captured this image of an X5.8-category solar flare peaking at 9:23 p.m. EDT on 10 May 2024.
Figure 1. NASA’s Solar Dynamics Observatory captured this image of an X5.8-category solar flare peaking at 9:23 p.m. EDT on 10 May 2024.
Remotesensing 17 01075 g001
Figure 2. Photospheric magnetic field parameter-based MVTS data pipeline for flare prediction using GOES and SDO sensor observations.
Figure 2. Photospheric magnetic field parameter-based MVTS data pipeline for flare prediction using GOES and SDO sensor observations.
Remotesensing 17 01075 g002
Figure 3. Class distributions for each partition are presented in a stacked bar plot format. The plot displays five flare classes for each partition, along with their corresponding values.
Figure 3. Class distributions for each partition are presented in a stacked bar plot format. The plot displays five flare classes for each partition, along with their corresponding values.
Remotesensing 17 01075 g003
Figure 4. An overview of the vector-based solar flare prediction process by extracting the last timestamp from the MVTS instance.
Figure 4. An overview of the vector-based solar flare prediction process by extracting the last timestamp from the MVTS instance.
Remotesensing 17 01075 g004
Figure 5. An overview of the vector-based solar flare prediction process by extracting the statistical features and last timestamp features from the MVTS instance.
Figure 5. An overview of the vector-based solar flare prediction process by extracting the statistical features and last timestamp features from the MVTS instance.
Remotesensing 17 01075 g005
Figure 6. An overview of the time series-based solar flare prediction process by feeding each MVTS instance into the time series classification module as shown.
Figure 6. An overview of the time series-based solar flare prediction process by feeding each MVTS instance into the time series classification module as shown.
Remotesensing 17 01075 g006
Figure 7. An overview of the graph-based solar flare prediction process by converting each MVTS instance into a graph and modeling inter-parameter relations as edges. Node degrees are extracted in the following step for training purposes.
Figure 7. An overview of the graph-based solar flare prediction process by converting each MVTS instance into a graph and modeling inter-parameter relations as edges. Node degrees are extracted in the following step for training purposes.
Remotesensing 17 01075 g007
Figure 8. An overview of the graph-based solar flare prediction process where, after converting each MVTS instance into a graph, embeddings are learned to capture underlying patterns for prediction purposes.
Figure 8. An overview of the graph-based solar flare prediction process where, after converting each MVTS instance into a graph, embeddings are learned to capture underlying patterns for prediction purposes.
Remotesensing 17 01075 g008
Figure 9. An overview of the graph-based solar flare prediction process is presented, wherein each MVTS instance is transformed into a graph consisting of nodes that represent magnetic field parameters, with their corresponding univariate time series serving as node features. Subsequently, each graph is processed through our graph convolutional module.
Figure 9. An overview of the graph-based solar flare prediction process is presented, wherein each MVTS instance is transformed into a graph consisting of nodes that represent magnetic field parameters, with their corresponding univariate time series serving as node features. Subsequently, each graph is processed through our graph convolutional module.
Remotesensing 17 01075 g009
Figure 10. A view of the structure of randomly selected opposite-class correlation graphs created from Pearson correlation matrices, which illustrate the inter-parameter connections. The nodes have a one-to-one correspondence with the magnetic field parameters from Table 1. The F graph represents an X2.2 flare observed in NOAA active region 377, spanning 14 February 2011 from 02:00 to 13:48 UTC. The NF graph corresponds to an FQ event in NOAA active region 1038, with a timestamp from 9 November 2011 at 20:12 UTC to 10 November 2011 at 08:00 UTC.
Figure 10. A view of the structure of randomly selected opposite-class correlation graphs created from Pearson correlation matrices, which illustrate the inter-parameter connections. The nodes have a one-to-one correspondence with the magnetic field parameters from Table 1. The F graph represents an X2.2 flare observed in NOAA active region 377, spanning 14 February 2011 from 02:00 to 13:48 UTC. The NF graph corresponds to an FQ event in NOAA active region 1038, with a timestamp from 9 November 2011 at 20:12 UTC to 10 November 2011 at 08:00 UTC.
Remotesensing 17 01075 g010
Figure 11. An overview of correlation-based F- and NF-class graphs randomly selected from P1 train set. The P1 train set covers events between 1 May 2010 and 13 March 2012. For F and NF graphs, different connection patterns are observed, highlighting varied inter-parameter relationships between events.
Figure 11. An overview of correlation-based F- and NF-class graphs randomly selected from P1 train set. The P1 train set covers events between 1 May 2010 and 13 March 2012. For F and NF graphs, different connection patterns are observed, highlighting varied inter-parameter relationships between events.
Remotesensing 17 01075 g011
Figure 12. Experimentation of graph creation regarding different threshold values for Euclidian distance matrix randomly selected from P4 train set. Subfigures show the effect of changing the threshold on average degree, graph density, and graph structure. The figure includes the frequency distribution of the graph with respect to Euclidian distance values, providing a breakdown of the count of distance values that fall between certain intervals.
Figure 12. Experimentation of graph creation regarding different threshold values for Euclidian distance matrix randomly selected from P4 train set. Subfigures show the effect of changing the threshold on average degree, graph density, and graph structure. The figure includes the frequency distribution of the graph with respect to Euclidian distance values, providing a breakdown of the count of distance values that fall between certain intervals.
Remotesensing 17 01075 g012
Figure 13. A view of the structure of randomly selected opposite-class distance similarity-based graphs created from thresholded Euclidian distance matrices, which illustrate the inter-parameter connections. The nodes have a one-to-one correspondence with the magnetic field parameters from Table 1. The F graph represents an X1.8 flare observed in NOAA active region 4920, spanning 19 December 2014 at 03:12 UTC to 19 December 2014 at 15:00 UTC. The NF graph corresponds to an FQ event in NOAA active region 4618, with a timestamp from 28 September 2014 at 00:48 UTC to 28 September 2014 at 12:36 UTC.
Figure 13. A view of the structure of randomly selected opposite-class distance similarity-based graphs created from thresholded Euclidian distance matrices, which illustrate the inter-parameter connections. The nodes have a one-to-one correspondence with the magnetic field parameters from Table 1. The F graph represents an X1.8 flare observed in NOAA active region 4920, spanning 19 December 2014 at 03:12 UTC to 19 December 2014 at 15:00 UTC. The NF graph corresponds to an FQ event in NOAA active region 4618, with a timestamp from 28 September 2014 at 00:48 UTC to 28 September 2014 at 12:36 UTC.
Remotesensing 17 01075 g013
Figure 14. An overview of thresholded distance similarity-based F- and NF-class graphs randomly selected from P4 train set. The P4 train set covers events between 2 June 2014 and 18 March 2015. For F and NF graphs, different inter-parameter relationships are visible. The F-class graphs mostly show two tightly connected subgraphs with sparse connections between them, while the NF-class graphs have a more uniform, distributed connection pattern.
Figure 14. An overview of thresholded distance similarity-based F- and NF-class graphs randomly selected from P4 train set. The P4 train set covers events between 2 June 2014 and 18 March 2015. For F and NF graphs, different inter-parameter relationships are visible. The F-class graphs mostly show two tightly connected subgraphs with sparse connections between them, while the NF-class graphs have a more uniform, distributed connection pattern.
Remotesensing 17 01075 g014
Figure 15. Demonstration of accuracy, TSS, HSS2, F1, GS, and ROC AUC performance results of vector (blue), time series (green), and graph (red) data representation methods in boxplot format under the extreme class imbalance setting. For each method, the classification results of all submethods are aggregated to illustrate the overall performance of the group.
Figure 15. Demonstration of accuracy, TSS, HSS2, F1, GS, and ROC AUC performance results of vector (blue), time series (green), and graph (red) data representation methods in boxplot format under the extreme class imbalance setting. For each method, the classification results of all submethods are aggregated to illustrate the overall performance of the group.
Remotesensing 17 01075 g015
Figure 16. The TSS results for each model, based on the given data representation methods, are demonstrated using all train–test partition pairs under the imbalanced settings.
Figure 16. The TSS results for each model, based on the given data representation methods, are demonstrated using all train–test partition pairs under the imbalanced settings.
Remotesensing 17 01075 g016
Figure 17. Demonstration of accuracy, TSS, HSS2, F1, GS, and ROC AUC performance results of vector (blue), time series (green), and graph (red) data representation methods in boxplot format under undersampling setting. For each method, the classification results of all submethods are aggregated to illustrate the overall performance of the group.
Figure 17. Demonstration of accuracy, TSS, HSS2, F1, GS, and ROC AUC performance results of vector (blue), time series (green), and graph (red) data representation methods in boxplot format under undersampling setting. For each method, the classification results of all submethods are aggregated to illustrate the overall performance of the group.
Remotesensing 17 01075 g017
Figure 18. The TSS results for each model, based on the given data representation methods, are demonstrated using all train–test partition pairs under the undersampled settings.
Figure 18. The TSS results for each model, based on the given data representation methods, are demonstrated using all train–test partition pairs under the undersampled settings.
Remotesensing 17 01075 g018
Figure 19. Demonstration of accuracy, TSS, HSS2, F1, GS, and ROC AUC performance results of vector (blue), time series (green), and graph (red) data representation methods in boxplot format under optimized preprocessing setting. For each method, the classification results of all submethods are aggregated to illustrate the overall performance of the group.
Figure 19. Demonstration of accuracy, TSS, HSS2, F1, GS, and ROC AUC performance results of vector (blue), time series (green), and graph (red) data representation methods in boxplot format under optimized preprocessing setting. For each method, the classification results of all submethods are aggregated to illustrate the overall performance of the group.
Remotesensing 17 01075 g019
Figure 20. The TSS results for each model, based on the given data representation methods, are demonstrated using all train–test partition pairs under the optimized settings.
Figure 20. The TSS results for each model, based on the given data representation methods, are demonstrated using all train–test partition pairs under the optimized settings.
Remotesensing 17 01075 g020
Table 1. Solar active region photospheric magnetic field parameters.
Table 1. Solar active region photospheric magnetic field parameters.
AbbreviationDescriptionFormula
ABSNJZHAbsolute value of the net current helicity H c a b s B z · J z
EPSXSum of x-component of normalized Lorentz force δ F x B x B z B 2
EPSYSum of y-component of normalized Lorentz force δ F y B y B z B 2
EPSZSum of z-component of normalized Lorentz force δ F z B x 2 + B y 2 B z 2 B 2
MEANALPMean characteristic twist parameter, α α total J z · B z B z 2
MEANGAMMean angle of field from radial γ ¯ = 1 N arctan B h B z
MEANGBHMean gradient of horizontal field | B h | ¯ = 1 N B h x 2 + B h y 2
MEANGBTMean gradient of total field | B t o t | ¯ = 1 N B x 2 + B y 2
MEANGBZMean gradient of vertical field | B z | ¯ = 1 N B z x 2 + B z y 2
MEANJZDMean vertical current density J z ¯ 1 N B y x B x y
MEANJZHMean current helicity ( B z contribution) H c ¯ 1 N B z · J z
MEANPOTMean photospheric magnetic free energy ρ ¯ 1 N B O b s B P o t 2
MEANSHRMean shear angle Γ ¯ = 1 N arccos B O b s · B Pot B O b s B P o t
R_VALUESum of flux near polarity inversion line Φ = Σ B LoS d A ( within R mask )
SAVNCPPSum of the modulus of the net current per polarity J z s u m | B z + J z d A | + | B z J z d A |
SHRGT45Fraction of area with shear > 45 ° Area with shear > 45 ° /total area
TOTBSQTotal magnitude of Lorentz force F B 2
TOTFXSum of x-component of Lorentz force F x B x B z d A
TOTFYSum of y-component of Lorentz force F y B y B z d A
TOTFZSum of z-component of Lorentz force F z B x 2 + B y 2 B z 2 d A
TOTPOTTotal photospheric magnetic free energy density ρ t o t ( B O b s B P o t ) 2 d A
TOTUSJHTotal unsigned current helicity H c total B z · J z
TOTUSJZTotal unsigned vertical current J z t o t a l = J z d A
USFLUXTotal unsigned flux Φ = B z d A
Table 2. Overview of data representation methods and the strategies within each.
Table 2. Overview of data representation methods and the strategies within each.
AbbreviationName
Vector representation
VLTVector of last timestamp
VSTATVector of statistical summary
Time series representation
TSCTime series classification
Graph representation
GNDGraph node degree
GNEGraph node embedding
GNNGraph neural network
Table 3. Overview of experimented classifiers and data representations.
Table 3. Overview of experimented classifiers and data representations.
AbbreviationFull Name
Classifiers
DTDecision tree
GCNGraph convolutional network
KNNK-nearest neighbors
LRLogistic regression
LSTMLong short-term memory
MLPMultilayer perceptron
RNNRecurrent neural network
ROCKETRandom convolutional kernel transform
STShapelet transform
SVMSupport vector machine
TSFTime series forest
Data Representations
CORCorrelation-based graph
CORGNDCorrelation-based graph node degrees
CORLAPCorrelation-based graph node embeddings via Laplacian Eigenmaps
CORN2VCorrelation-based graph node embeddings via Node2Vec
DSDistance similarity-based graph
DSGNDDistance similarity-based graph node degrees
DSLAPDistance similarity-based graph node embeddings via Laplacian Eigenmaps
DSN2VDistance similarity-based graph node embeddings via Node2Vec
TSTime series
VLTVector of last timestamp
VSTATVector of statistical summary
VSTATNVector of statistical summary normalized
Table 4. Performance results of classifier models for data representations of SWAN-SF across averaged train–test pairs with extreme class imbalance, where red fonts represent maximum values.
Table 4. Performance results of classifier models for data representations of SWAN-SF across averaged train–test pairs with extreme class imbalance, where red fonts represent maximum values.
ModelsAccuracyTSSHSS2F1GSROC AUC
VLT
SVM-VLT0.8776 ± 0.0320.6635 ± 0.09050.1808 ± 0.05190.2098 ± 0.05910.01576 ± 0.00720.8318 ± 0.0453
KNN-VLT0.9577 ± 0.01250.3347 ± 0.10410.2371 ± 0.0620.2569 ± 0.06810.0076 ± 0.00450.6674 ± 0.052
MLP-VLT0.9713 ± 0.01530.0818 ± 0.13680.0666 ± 0.08560.0745 ± 0.09490.0017 ± 0.00290.5409 ± 0.0684
LR-VLT0.9791 ± 0.0090.0 ± 0.00.0 ± 0.00.0 ± 0.00.0 ± 0.00.5 ± 0.0
DT-VLT0.9529 ± 0.01540.5134 ± 0.12590.3012 ± 0.06070.3216 ± 0.06150.011 ± 0.00570.7567 ± 0.063
VSTAT
SVM-VSTAT0.891 ± 0.02690.6621 ± 0.06220.1977 ± 0.05610.2255 ± 0.06270.0155 ± 0.00710.831 ± 0.0311
KNN-VSTAT0.9581 ± 0.01310.3258 ± 0.08440.2361 ± 0.05640.2558 ± 0.06260.0074 ± 0.00450.6629 ± 0.0422
MLP-VSTAT0.9356 ± 0.03810.2779 ± 0.21380.1315 ± 0.04820.1536 ± 0.0570.0061 ± 0.00490.639 ± 0.1069
LR-VSTAT0.9791 ± 0.0090.0 ± 0.00.0 ± 0.00.0 ± 0.00.0 ± 0.00.5 ± 0.0
DT-VSTAT0.9537 ± 0.01390.5575 ± 0.01950.3241 ± 0.0280.3438 ± 0.0350.0121 ± 0.00550.7788 ± 0.0097
SVM-VSTATN0.9654 ± 0.0130.3102 ± 0.1070.2603 ± 0.03770.276 ± 0.03730.0063 ± 0.00260.6551 ± 0.0535
KNN-VSTATN0.9704 ± 0.00710.3942 ± 0.14950.3327 ± 0.04120.3465 ± 0.0420.0077 ± 0.00280.697 ± 0.0748
MLP-VSTATN0.9681 ± 0.00860.3903 ± 0.1650.3127 ± 0.03350.327 ± 0.03650.0077 ± 0.0030.6952 ± 0.0825
LR-VSTATN0.9714 ± 0.00510.4927 ± 0.18210.3871 ± 0.07530.3999 ± 0.07460.0093 ± 0.00210.7464 ± 0.0911
DT-VSTATN0.9536 ± 0.01440.4171 ± 0.11430.2603 ± 0.05970.2797 ± 0.05840.0083 ± 0.00170.7085 ± 0.0571
TSC
ST-TS0.9693 ± 0.01640.0246 ± 0.02260.028 ± 0.02370.0398 ± 0.03420.0006 ± 0.00070.5124 ± 0.0111
TSF-TS0.9678 ± 0.01210.3718 ± 0.12150.3133 ± 0.07650.3293 ± 0.08140.0085 ± 0.00570.6859 ± 0.0608
ROCKET-TS0.9782 ± 0.00910.0379 ± 0.02950.0654 ± 0.04910.0698 ± 0.05010.0009 ± 0.00070.5189 ± 0.0147
LSTM-TS0.9772 ± 0.00910.0533 ± 0.06730.0794 ± 0.09210.0853 ± 0.09440.0011 ± 0.00140.5267 ± 0.0337
RNN-TS0.9788 ± 0.00920.008 ± 0.01160.0147 ± 0.02090.016 ± 0.02280.602 ± 0.80330.504 ± 0.0058
GND
SVM-CORGND0.8429 ± 0.02630.2362 ± 0.05310.057 ± 0.02140.0909 ± 0.03280.0058 ± 0.00270.6181 ± 0.0266
KNN-CORGND0.9726 ± 0.00860.0293 ± 0.00880.041 ± 0.01130.0515 ± 0.01330.0006 ± 0.00030.5147 ± 0.0044
MLP-CORGND0.9737 ± 0.00770.0251 ± 0.01930.0328 ± 0.02290.0409 ± 0.02820.5741 ± 1.14680.5126 ± 0.0097
LR-CORGND0.9688 ± 0.02120.0359 ± 0.07010.0227 ± 0.04210.0298 ± 0.05480.0008 ± 0.00150.518 ± 0.035
DT-CORGND0.9459 ± 0.01210.0562 ± 0.01420.0392 ± 0.0090.0636 ± 0.01480.0012 ± 0.00050.5281 ± 0.0071
SVM-DSGND0.8286 ± 0.0240.2959 ± 0.0270.0646 ± 0.02040.0989 ± 0.03240.0074 ± 0.00310.648 ± 0.0135
KNN-DSGND0.9708 ± 0.00890.0287 ± 0.00440.0385 ± 0.01050.0507 ± 0.00920.0006 ± 0.00030.5144 ± 0.0022
MLP-DSGND0.9585 ± 0.01010.0621 ± 0.02310.0525 ± 0.0060.0711 ± 0.01020.0012 ± 0.00040.5311 ± 0.0116
LR-DSGND0.9781 ± 0.00890.0128 ± 0.01230.0229 ± 0.02140.0256 ± 0.02230.0002 ± 0.00030.5064 ± 0.0062
DT-DSGND0.9464 ± 0.00830.0794 ± 0.02550.0538 ± 0.01380.0782 ± 0.01850.0017 ± 0.00070.5397 ± 0.0128
GNE
SVM-CORLAP0.9791 ± 0.0090.0003 ± 0.00060.0006 ± 0.00110.0006 ± 0.00110.0 ± 0.00.5002 ± 0.0003
KNN-CORLAP0.9781 ± 0.00890.0027 ± 0.00530.0046 ± 0.00890.0066 ± 0.01130.0001 ± 0.00010.5014 ± 0.0026
MLP-CORLAP0.9473 ± 0.05150.034 ± 0.05680.0134 ± 0.01160.0225 ± 0.02550.0045 ± 0.00750.517 ± 0.0284
LR-CORLAP0.9791 ± 0.0090.0 ± 0.00.0 ± 0.00.0 ± 0.00.0 ± 0.00.5 ± 0.0
DT-CORLAP0.9549 ± 0.0210.0458 ± 0.03320.0331 ± 0.02290.0518 ± 0.03670.001 ± 0.00080.5229 ± 0.0167
SVM-DSLAP0.4206 ± 0.02090.4078 ± 0.02230.0278 ± 0.01130.0667 ± 0.02690.0204 ± 0.00860.7039 ± 0.0111
KNN-DSLAP0.9783 ± 0.00910.0006 ± 0.00110.001 ± 0.0020.0026 ± 0.0038−1.2425 ± 2.4850.5003 ± 0.0005
MLP-DSLAP0.9516 ± 0.04120.0191 ± 0.02810.0101 ± 0.00560.0236 ± 0.01460.0003 ± 0.00050.5095 ± 0.0141
LR-DSLAP0.9791 ± 0.00910.0 ± 0.00.0 ± 0.00.0 ± 0.00.0 ± 0.00.5 ± 0.0
DT-DSLAP0.9425 ± 0.01420.0242 ± 0.00880.0154 ± 0.0020.0409 ± 0.0050.0005 ± 0.00010.5121 ± 0.0044
SVM-CORN2V0.5856 ± 0.04150.0955 ± 0.01670.0105 ± 0.00330.0549 ± 0.01830.0037 ± 0.00070.5478 ± 0.0083
KNN-CORN2V0.9789 ± 0.0089−0.9586 ± 1.9167−0.0004 ± 0.00050.0 ± 0.0−0.0 ± 0.00.4999 ± 0.0002
MLP-CORN2V0.9516 ± 0.04120.0191 ± 0.02810.0101 ± 0.00560.0236 ± 0.01460.0003 ± 0.00050.5095 ± 0.0141
LR-CORN2V0.9791 ± 0.0090.0 ± 0.00.0 ± 0.00.0 ± 0.00.0 ± 0.00.5 ± 0.0
DT-CORN2V0.9554 ± 0.02140.0347 ± 0.02710.0246 ± 0.0170.0434 ± 0.0295−1.1866 ± 2.37520.5174 ± 0.0135
SVM-DSN2V0.545 ± 0.01230.2909 ± 0.02740.0253 ± 0.01010.0639 ± 0.02540.0112 ± 0.00510.6454 ± 0.0137
KNN-DSN2V0.9788 ± 0.0090.0014 ± 0.00250.0026 ± 0.00460.0034 ± 0.0054−0.4885 ± 0.97720.5007 ± 0.0012
MLP-DSN2V0.9706 ± 0.01410.0012 ± 0.00240.0007 ± 0.00130.0055 ± 0.0110.0 ± 0.00.5006 ± 0.0012
LR-DSN2V0.9791 ± 0.0090.0 ± 0.00.0 ± 0.00.0 ± 0.00.0 ± 0.00.5 ± 0.0
DT-DSN2V0.9438 ± 0.01350.0053 ± 0.006−2.4172 ± 4.84320.0284 ± 0.0083−0.5559 ± 1.1120.5027 ± 0.003
GNN
GCN-COR0.9749 ± 0.01540.0061 ± 0.0140.0077 ± 0.01880.0144 ± 0.02870.0002 ± 0.00050.5032 ± 0.007
GCN-DS0.9602 ± 0.0337−0.0143 ± 0.0269−0.0062 ± 0.00940.0018 ± 0.0035−0.0003 ± 0.00050.4928 ± 0.0135
Table 5. Top 5 models and TSS comparison for different conditions and partitions, where red fonts represent maximum values.
Table 5. Top 5 models and TSS comparison for different conditions and partitions, where red fonts represent maximum values.
ConditionP1–P2P2–P3P3–P4P4–P5
Model TSS Model TSS Model TSS Model TSS
ImbalancedSVM-VLT0.6027SVM-VSTAT0.6661SVM-VLT0.7957SVM-VSTAT0.6365
SVM-VSTAT0.5998SVM-VLT0.6488SVM-VSTAT0.7461DT-VLT0.6246
DT-VSTAT0.5657DT-VSTAT0.5671LR-VSTATN0.6037LR-VSTATN0.6197
DT-VLT0.5431DT-VLT0.5534DT-VSTAT0.5688SVM-VLT0.6069
LR-VSTATN0.5199TSF-TS0.4578KNN-VSTATN0.5564DT-VSTAT0.5268
UndersampledTSF-TS0.6965DT-VSTAT0.7526SVM-VLT0.7885TSF-TS0.7891
ROCKET-TS0.6154SVM-VSTAT0.7404SVM-VSTAT0.7826SVM-VLT0.7857
SVM-VLT0.6027SVM-VLT0.7375TSF-TS0.7564SVM-VSTAT0.7762
SVM-VSTAT0.5998DT-VLT0.7348DT-VSTAT0.7465DT-VSTAT0.7174
DT-VSTAT0.5657KNN-VSTATN0.7082ROCKET-TS0.6979LR-VSTATN0.6974
OptimizedRNN-TS0.8036TSF-TS0.768LR-VLT0.8094TSF-TS0.7645
LSTM-TS0.8015LR-VSTAT0.7445RNN-TS0.7648LR-VSTAT0.7278
LR-VSTAT0.7819RNN-TS0.6899LR-VSTAT0.7117RNN-TS0.6873
KNN-VLT0.7585KNN-VLT0.6628TSF-TS0.6875ST-TS0.6464
LR-VLT0.7061ST-TS0.6551LSTM-TS0.6671KNN-VLT0.6434
Table 6. Class sizes before and after undersampling for each training partition.
Table 6. Class sizes before and after undersampling for each training partition.
PartitionBefore UndersamplingAfter Undersampling
F NF F NF
P1118056,31911801180
P2128565,36412851285
P3127730,76612771277
P489036,667890890
Table 7. Performance results of classifier models for data representations of SWAN-SF across averaged train–test pairs with undersampling, where red fonts represent maximum values.
Table 7. Performance results of classifier models for data representations of SWAN-SF across averaged train–test pairs with undersampling, where red fonts represent maximum values.
ModelsAccuracyTSSHSS2F1GSROC AUC
VLT
SVM-VLT0.8408 ± 0.02990.7286 ± 0.08710.1531 ± 0.03380.1843 ± 0.04440.0181 ± 0.00880.8644 ± 0.0436
KNN-VLT0.8989 ± 0.05260.5712 ± 0.17590.1936 ± 0.02820.221 ± 0.0340.0138 ± 0.00870.7856 ± 0.0879
MLP-VLT0.7268 ± 0.20790.2906 ± 0.15290.065 ± 0.03790.0981 ± 0.04310.0099 ± 0.00830.6453 ± 0.0764
LR-VLT0.979 ± 0.0090.0 ± 0.00.0 ± 0.00.0 ± 0.00.0 ± 0.00.5 ± 0.0
DT-VLT0.9172 ± 0.03850.6162 ± 0.08440.2406 ± 0.05280.2662 ± 0.05530.0146 ± 0.0090.8081 ± 0.0422
VSTAT
SVM-VSTAT0.8452 ± 0.03890.7248 ± 0.08530.1566 ± 0.0280.1877 ± 0.03920.018 ± 0.0090.8624 ± 0.0427
KNN-VSTAT0.8995 ± 0.05290.5652 ± 0.1850.1924 ± 0.02730.2198 ± 0.03250.0136 ± 0.00840.7826 ± 0.0925
MLP-VSTAT0.7214 ± 0.18430.3985 ± 0.11310.0635 ± 0.01810.0994 ± 0.02170.0132 ± 0.00890.6993 ± 0.0565
LR-VSTAT0.979 ± 0.0090.0 ± 0.00.0 ± 0.00.0 ± 0.00.0 ± 0.00.5 ± 0.0
DT-VSTAT0.9185 ± 0.03890.6956 ± 0.08790.2652 ± 0.03530.2899 ± 0.03690.0161 ± 0.00880.8478 ± 0.044
SVM-VSTATN0.7527 ± 0.15120.3067 ± 0.16150.111 ± 0.13210.1431 ± 0.12290.0103 ± 0.01040.6534 ± 0.0808
KNN-VSTATN0.761 ± 0.14690.5866 ± 0.18330.1489 ± 0.12660.1797 ± 0.11750.0177 ± 0.01160.7932 ± 0.0916
MLP-VSTATN0.7666 ± 0.14270.5872 ± 0.20230.1463 ± 0.11770.177 ± 0.10830.0176 ± 0.01180.7936 ± 0.1011
LR-VSTATN0.7491 ± 0.15390.6295 ± 0.07460.1675 ± 0.17690.199 ± 0.16530.0186 ± 0.01050.8148 ± 0.0373
DT-VSTATN0.787 ± 0.11280.4175 ± 0.24390.0949 ± 0.07420.129 ± 0.0750.0135 ± 0.01280.7087 ± 0.1219
TSC
ST-TS0.8585 ± 0.02440.3033 ± 0.06010.0786 ± 0.0250.1114 ± 0.03560.0072 ± 0.00290.6517 ± 0.0301
TSF-TS0.8912 ± 0.02690.739 ± 0.04180.2146 ± 0.04520.2423 ± 0.05090.017 ± 0.00710.8695 ± 0.0209
ROCKET-TS0.7978 ± 0.04140.6182 ± 0.05660.1096 ± 0.03060.1432 ± 0.04130.0161 ± 0.00710.8091 ± 0.0283
LSTM-TS0.6979 ± 0.03770.3204 ± 0.04980.0417 ± 0.01290.0786 ± 0.02650.0095 ± 0.0040.6602 ± 0.0249
RNN-TS0.726 ± 0.09740.3126 ± 0.10790.0647 ± 0.04240.1008 ± 0.03880.0091 ± 0.00450.6692 ± 0.0673
GND
SVM-CORGND0.708 ± 0.01940.3564 ± 0.04670.0385 ± 0.01020.0685 ± 0.01440.0084 ± 0.00260.6782 ± 0.0234
KNN-CORGND0.695 ± 0.01890.2486 ± 0.04010.0261 ± 0.00760.0566 ± 0.01140.0059 ± 0.00150.6243 ± 0.0201
MLP-CORGND0.7496 ± 0.05090.2618 ± 0.04250.035 ± 0.01580.0648 ± 0.01850.0057 ± 0.00130.6309 ± 0.0213
LR-CORGND0.6507 ± 0.0360.2804 ± 0.06180.0259 ± 0.00930.0566 ± 0.01360.0073 ± 0.00310.6402 ± 0.0309
DT-CORGND0.6794 ± 0.01420.1845 ± 0.04790.0186 ± 0.00720.0495 ± 0.0120.0046 ± 0.00220.5922 ± 0.024
SVM-DSGND0.682 ± 0.01960.4282 ± 0.02750.042 ± 0.00810.0721 ± 0.01270.0104 ± 0.00270.7141 ± 0.0138
KNN-DSGND0.6543 ± 0.02220.359 ± 0.02120.0326 ± 0.0050.0631 ± 0.00930.009 ± 0.00140.6795 ± 0.0106
MLP-DSGND0.7597 ± 0.01860.3324 ± 0.01460.0437 ± 0.01190.0732 ± 0.01590.0071 ± 0.00120.6662 ± 0.0073
LR-DSGND0.6681 ± 0.03410.3865 ± 0.05870.0369 ± 0.01050.0672 ± 0.01480.0097 ± 0.00340.6932 ± 0.0294
DT-DSGND0.7133 ± 0.01620.3357 ± 0.02640.0371 ± 0.00970.0672 ± 0.01410.0078 ± 0.00190.6678 ± 0.0132
GNE
SVM-CORLAP0.5379 ± 0.01570.0453 ± 0.01810.0043 ± 0.00290.0438 ± 0.01880.0018 ± 0.00120.5227 ± 0.009
KNN-CORLAP0.436 ± 0.01040.1315 ± 0.02110.0097 ± 0.00520.0494 ± 0.02140.0067 ± 0.0040.5658 ± 0.0106
MLP-CORLAP0.536 ± 0.01090.0635 ± 0.01350.0059 ± 0.00380.0454 ± 0.01980.0026 ± 0.00170.5317 ± 0.0068
LR-CORLAP0.4953 ± 0.0065−0.0068 ± 0.0148−0.0007 ± 0.0010.0393 ± 0.0159−0.0004 ± 0.00050.4966 ± 0.0074
DT-CORLAP0.6701 ± 0.02710.3053 ± 0.03310.0358 ± 0.01270.0733 ± 0.02680.0093 ± 0.00370.6526 ± 0.0165
SVM-DSLAP0.554 ± 0.02110.2668 ± 0.02870.024 ± 0.01060.0627 ± 0.02570.0101 ± 0.00440.6334 ± 0.0143
KNN-DSLAP0.7028 ± 0.01720.264 ± 0.01610.0347 ± 0.01330.072 ± 0.02750.0078 ± 0.00360.632 ± 0.0081
MLP-DSLAP0.6021 ± 0.01420.2306 ± 0.03970.0234 ± 0.01050.0618 ± 0.02510.0079 ± 0.00330.6153 ± 0.0199
LR-DSLAP0.5014 ± 0.00540.0111 ± 0.0020.0009 ± 0.00040.0404 ± 0.0159−1.1157 ± 2.23210.5056 ± 0.001
DT-DSLAP0.6532 ± 0.00860.299 ± 0.01560.0327 ± 0.01060.0709 ± 0.02610.0092 ± 0.00340.6495 ± 0.0078
SVM-CORN2V0.4156 ± 0.01560.4018 ± 0.01430.0296 ± 0.01030.0717 ± 0.02340.0221 ± 0.00730.7009 ± 0.0071
KNN-CORN2V0.5522 ± 0.01150.327 ± 0.01850.0309 ± 0.00850.0724 ± 0.02140.0131 ± 0.00350.6635 ± 0.0092
MLP-CORN2V0.5989 ± 0.05770.3135 ± 0.02320.0301 ± 0.00970.0712 ± 0.02170.013 ± 0.00620.6567 ± 0.0116
LR-CORN2V0.4238 ± 0.00640.3996 ± 0.01970.0298 ± 0.01020.0719 ± 0.02340.0216 ± 0.00760.6998 ± 0.0099
DT-CORN2V0.6237 ± 0.00880.2602 ± 0.00230.0297 ± 0.010.0708 ± 0.02260.0093 ± 0.00290.6301 ± 0.0012
SVM-DSN2V0.5174 ± 0.00170.2804 ± 0.01760.0249 ± 0.00780.0669 ± 0.02080.0122 ± 0.0040.6402 ± 0.0088
KNN-DSN2V0.5488 ± 0.00570.1326 ± 0.01310.0128 ± 0.00460.0551 ± 0.01770.0055 ± 0.00210.5663 ± 0.0066
MLP-DSN2V0.4824 ± 0.02590.266 ± 0.01780.0221 ± 0.00720.0643 ± 0.02040.0127 ± 0.00520.633 ± 0.0089
LR-DSN2V0.5835 ± 0.00420.27 ± 0.00290.0277 ± 0.00850.0692 ± 0.02130.0104 ± 0.00340.635 ± 0.0014
DT-DSN2V0.5739 ± 0.03380.1624 ± 0.06770.0167 ± 0.00850.0548 ± 0.01850.0051 ± 0.00220.5624 ± 0.0073
GNN
GCN-COR0.6115 ± 0.25170.0878 ± 0.12670.0092 ± 0.01120.0445 ± 0.02040.0097 ± 0.00610.5197 ± 0.0226
GCN-DS0.5347 ± 0.38280.0852 ± 0.12820.0219 ± 0.01840.0314 ± 0.00930.0052 ± 0.00730.5094 ± 0.0095
Table 8. Class sizes before and after optimized preprocessing for each training partition.
Table 8. Class sizes before and after optimized preprocessing for each training partition.
PartitionBefore Optimized PreprocessingAfter Optimized Preprocessing
F NF F NF
P1118056,31987789995
P2128565,364980710,000
P3127730,76699689997
P489036,667932010,000
Table 9. Performance results of classifier models for data representations of SWAN-SF across averaged train–test pairs with optimized preprocessing, where red fonts represent maximum values.
Table 9. Performance results of classifier models for data representations of SWAN-SF across averaged train–test pairs with optimized preprocessing, where red fonts represent maximum values.
ModelsAccuracyTSSHSS2F1GSROC AUC
VLT
SVM-VLT0.6098 ± 0.47130.2208 ± 0.26090.074 ± 0.07440.1072 ± 0.04690.015 ± 0.01560.6104 ± 0.1304
KNN-VLT0.8399 ± 0.10730.5562 ± 0.27170.1624 ± 0.06560.1942 ± 0.06240.0163 ± 0.01130.7781 ± 0.1359
MLP-VLT0.6799 ± 0.39850.3392 ± 0.28780.1251 ± 0.11660.1563 ± 0.08840.0161 ± 0.01490.6679 ± 0.141
LR-VLT0.6589 ± 0.21170.6378 ± 0.2140.0952 ± 0.06340.1361 ± 0.0550.0229 ± 0.00870.8189 ± 0.107
DT-VLT0.4971 ± 0.24190.0027 ± 0.62650.0291 ± 0.08510.0605 ± 0.09830.0019 ± 0.05230.5014 ± 0.3133
VSTAT
SVM-VSTAT0.9525 ± 0.01880.1667 ± 0.25930.0727 ± 0.08870.0888 ± 0.09370.0029 ± 0.00430.5833 ± 0.1296
KNN-VSTAT0.9013 ± 0.0040.2856 ± 0.00880.1188 ± 0.04150.1514 ± 0.04930.0074 ± 0.00280.6428 ± 0.0044
MLP-VSTAT0.8616 ± 0.15130.474 ± 0.20620.1889 ± 0.07310.219 ± 0.05430.015 ± 0.01330.7369 ± 0.1031
LR-VSTAT0.8577 ± 0.09540.7415 ± 0.03510.2407 ± 0.14610.2718 ± 0.13760.0209 ± 0.00960.873 ± 0.0175
DT-VSTAT0.5508 ± 0.0833−0.1962 ± 0.3067−0.0294 ± 0.03550.0172 ± 0.0199−0.0105 ± 0.01320.4019 ± 0.1534
TSC
ST-TS0.8814 ± 0.10450.3915 ± 0.34280.1001 ± 0.0810.126 ± 0.10190.0123 ± 0.01420.6958 ± 0.1714
TSF-TS0.8855 ± 0.09350.4913 ± 0.41150.1452 ± 0.10780.1751 ± 0.1220.0164 ± 0.01580.7457 ± 0.2058
ROCKET-TS0.7279 ± 0.30830.2639 ± 0.18130.0586 ± 0.03430.0973 ± 0.03660.0144 ± 0.01620.6319 ± 0.0907
LSTM-TS0.7363 ± 0.31720.6067 ± 0.23090.2089 ± 0.19790.2433 ± 0.17860.0207 ± 0.01010.8033 ± 0.1154
RNN-TS0.7751 ± 0.07150.7364 ± 0.05780.1343 ± 0.00690.1722 ± 0.01090.023 ± 0.00870.8764 ± 0.0289
GND
SVM-CORGND0.8911 ± 0.02370.141 ± 0.00750.0608 ± 0.03360.0947 ± 0.03890.0037 ± 0.00110.5705 ± 0.0038
KNN-CORGND0.8424 ± 0.02310.1143 ± 0.01060.0345 ± 0.01790.0731 ± 0.02810.0032 ± 0.00120.5572 ± 0.0053
MLP-CORGND0.8741 ± 0.03040.144 ± 0.01380.053 ± 0.02330.0892 ± 0.03280.0039 ± 0.00150.572 ± 0.0069
LR-CORGND0.7514 ± 0.01010.1611 ± 0.01770.0302 ± 0.01450.0721 ± 0.02790.0052 ± 0.00240.5805 ± 0.0088
DT-CORGND0.8427 ± 0.01340.1052 ± 0.01850.0316 ± 0.0170.0708 ± 0.02830.0031 ± 0.00160.5526 ± 0.0092
SVM-DSGND0.8911 ± 0.02370.141 ± 0.00750.0608 ± 0.03360.0947 ± 0.03890.0037 ± 0.00110.5705 ± 0.0038
KNN-DSGND0.8424 ± 0.02310.1143 ± 0.01060.0345 ± 0.01790.0731 ± 0.02810.0032 ± 0.00120.5572 ± 0.0053
MLP-DSGND0.8759 ± 0.01670.1614 ± 0.0230.0577 ± 0.02070.0938 ± 0.02930.0042 ± 0.00110.5807 ± 0.0115
LR-DSGND0.7514 ± 0.01010.1611 ± 0.01770.0302 ± 0.01450.0721 ± 0.02790.0052 ± 0.00240.5805 ± 0.0088
DT-DSGND0.8427 ± 0.01340.1052 ± 0.01850.0316 ± 0.0170.0708 ± 0.02830.0031 ± 0.00160.5526 ± 0.0092
GNE
SVM-CORLAP0.7718 ± 0.05960.2244 ± 0.020.0438 ± 0.01090.085 ± 0.02360.0071 ± 0.00360.6122 ± 0.01
KNN-CORLAP0.7491 ± 0.02570.1703 ± 0.02790.0317 ± 0.01510.0732 ± 0.02810.0054 ± 0.00220.5852 ± 0.014
MLP-CORLAP0.7634 ± 0.0130.1981 ± 0.01220.0375 ± 0.01360.0789 ± 0.0270.0062 ± 0.00240.599 ± 0.0061
LR-CORLAP0.5774 ± 0.1054−0.0056 ± 0.0115−0.0004 ± 0.00110.0447 ± 0.0168−0.0002 ± 0.00060.4972 ± 0.0057
DT-CORLAP0.7767 ± 0.01850.1655 ± 0.00610.034 ± 0.01510.0751 ± 0.02770.005 ± 0.00190.5827 ± 0.003
SVM-DSLAP0.8405 ± 0.05250.5152 ± 0.15250.1375 ± 0.08280.1716 ± 0.08690.0138 ± 0.00660.7576 ± 0.0763
KNN-DSLAP0.6636 ± 0.10590.4167 ± 0.15220.0628 ± 0.05270.1026 ± 0.06230.0142 ± 0.00720.7083 ± 0.0761
MLP-DSLAP0.759 ± 0.03990.4047 ± 0.10470.069 ± 0.02270.1084 ± 0.03430.0125 ± 0.00710.7023 ± 0.0524
LR-DSLAP0.5594 ± 0.0768−0.0139 ± 0.0247−0.001 ± 0.00230.0432 ± 0.0191−0.0003 ± 0.00070.493 ± 0.0123
DT-DSLAP0.7247 ± 0.08740.2625 ± 0.07530.0413 ± 0.01390.0824 ± 0.02940.0093 ± 0.00630.6312 ± 0.0377
SVM-CORN2V0.5856 ± 0.04150.0955 ± 0.01670.0105 ± 0.00330.0549 ± 0.01830.0037 ± 0.00070.5478 ± 0.0083
KNN-CORN2V0.4816 ± 0.02220.0384 ± 0.0090.0036 ± 0.0020.049 ± 0.0180.002 ± 0.00110.5192 ± 0.0045
MLP-CORN2V0.572 ± 0.16060.0605 ± 0.00970.008 ± 0.00580.0517 ± 0.01940.0025 ± 0.00050.5303 ± 0.0049
LR-CORN2V0.6374 ± 0.05050.0905 ± 0.01890.0109 ± 0.00050.0551 ± 0.01620.0032 ± 0.00080.5453 ± 0.0094
DT-CORN2V0.5305 ± 0.01290.016 ± 0.00840.0016 ± 0.00090.0469 ± 0.01660.0007 ± 0.00040.508 ± 0.0042
SVM-DSN2V0.4544 ± 0.10740.3046 ± 0.04770.0299 ± 0.01950.0765 ± 0.03030.017 ± 0.00320.6523 ± 0.0239
KNN-DSN2V0.5044 ± 0.07220.1931 ± 0.03820.0207 ± 0.01320.0677 ± 0.02440.0097 ± 0.00320.5965 ± 0.0191
MLP-DSN2V0.4014 ± 0.05560.2804 ± 0.05460.0241 ± 0.01370.0714 ± 0.02540.018 ± 0.0060.6402 ± 0.0273
LR-DSN2V0.4687 ± 0.09880.3106 ± 0.04770.031 ± 0.01970.0776 ± 0.03040.0168 ± 0.00350.6553 ± 0.0238
DT-DSN2V0.4927 ± 0.03990.1435 ± 0.03910.0149 ± 0.00950.0623 ± 0.02130.0076 ± 0.00360.5717 ± 0.0196
GNN
GCN-COR0.658 ± 0.5498−0.0016 ± 0.0035−0.0003 ± 0.00060.0143 ± 0.02570.0072 ± 0.01280.5001 ± 0.0002
GCN-DS0.5652 ± 0.01990.4258 ± 0.07930.0466 ± 0.03180.09 ± 0.05260.0189 ± 0.01240.7128 ± 0.0397
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Vural, O.; Hamdi, S.M.; Boubrahimi, S.F. Solar Flare Prediction Using Multivariate Time Series of Photospheric Magnetic Field Parameters: A Comparative Analysis of Vector, Time Series, and Graph Data Representations. Remote Sens. 2025, 17, 1075. https://doi.org/10.3390/rs17061075

AMA Style

Vural O, Hamdi SM, Boubrahimi SF. Solar Flare Prediction Using Multivariate Time Series of Photospheric Magnetic Field Parameters: A Comparative Analysis of Vector, Time Series, and Graph Data Representations. Remote Sensing. 2025; 17(6):1075. https://doi.org/10.3390/rs17061075

Chicago/Turabian Style

Vural, Onur, Shah Muhammad Hamdi, and Soukaina Filali Boubrahimi. 2025. "Solar Flare Prediction Using Multivariate Time Series of Photospheric Magnetic Field Parameters: A Comparative Analysis of Vector, Time Series, and Graph Data Representations" Remote Sensing 17, no. 6: 1075. https://doi.org/10.3390/rs17061075

APA Style

Vural, O., Hamdi, S. M., & Boubrahimi, S. F. (2025). Solar Flare Prediction Using Multivariate Time Series of Photospheric Magnetic Field Parameters: A Comparative Analysis of Vector, Time Series, and Graph Data Representations. Remote Sensing, 17(6), 1075. https://doi.org/10.3390/rs17061075

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop