Next Article in Journal
Automated Segmentation of Optical Coherence Tomography Images of the Human Tympanic Membrane Using Deep Learning
Next Article in Special Issue
Data-Driven Analysis of Student Engagement in Time-Limited Computer Laboratories
Previous Article in Journal
Tensor-Based Approaches for Nonlinear and Multilinear Systems Modeling and Identification
Previous Article in Special Issue
Transfer Learning and Analogical Inference: A Critical Comparison of Algorithms, Methods, and Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Nonsmooth Optimization-Based Hyperparameter-Free Neural Networks for Large-Scale Regression

1
Department of Computing, University of Turku, FI-20014 Turku, Finland
2
School of Science, RMIT University, Melbourne 3000, Australia
3
Department of Mathematics and Statistics, University of Turku, FI-20014 Turku, Finland
4
Siili Solutions Oyj, FI-60100 Seinäjoki, Finland
5
Centre for Smart Analytics, Federation University Australia, Ballarat 3350, Australia
*
Author to whom correspondence should be addressed.
Algorithms 2023, 16(9), 444; https://doi.org/10.3390/a16090444
Submission received: 18 August 2023 / Revised: 11 September 2023 / Accepted: 12 September 2023 / Published: 14 September 2023
(This article belongs to the Special Issue Machine Learning Algorithms for Big Data Analysis)

Abstract

:
In this paper, a new nonsmooth optimization-based algorithm for solving large-scale regression problems is introduced. The regression problem is modeled as fully-connected feedforward neural networks with one hidden layer, piecewise linear activation, and the L 1 -loss functions. A modified version of the limited memory bundle method is applied to minimize this nonsmooth objective. In addition, a novel constructive approach for automated determination of the proper number of hidden nodes is developed. Finally, large real-world data sets are used to evaluate the proposed algorithm and to compare it with some state-of-the-art neural network algorithms for regression. The results demonstrate the superiority of the proposed algorithm as a predictive tool in most data sets used in numerical experiments.

1. Introduction

Regression models and methods are extensively utilized for prediction and approximation in various real-world scenarios. When data involve complex relationships between the response and explanatory variables, regression methods designed using neural networks (NNs) have emerged as powerful alternatives to traditional regression methods [1]. Hence, we focus on NNs for regression (NNR): we introduce a new approach for modeling and solving regression problems using fully-connected feedforward NNs with the rectified linear unit (ReLU) activation function, the L 1 -loss function, and the L 1 -regularization. We call this problem the ReLU-NNR problem, and the method for solving it is the limited memory bundle NNR (LMBNNR) algorithm.
The ReLU-NNR problem is nonconvex and nonsmooth. Note that ReLU itself is nonsmooth. Thus, even if we used a smooth loss function and regularization, the underlying optimization problem would still be nonsmooth. Conventional global optimization methods, including those based on global random search, become time-consuming as this problem contains a large number of variables. In addition, the authors in [2] showed that it is impossible to give any guarantee that the global minimizer is found by a general global (quasi-)random search algorithm with reasonable accuracy when the dimension is large. On the other hand, local search methods are sensitive to the choice of starting points and, in general, end up at the closest local solutions, which may be significantly different from the global ones. To address the nonconvexity of the ReLU-NNR problem, we propose a constructive approach. More specifically, we construct initial weights by using the solution from the previous iteration. Such an approach allows us to find either global or deep local minimizers for the ReLU-NNR problem. To solve the underlying nonsmooth optimization problems, we apply a slightly modified version of the limited memory bundle method (LMBM) developed by Karmitsa (née Haarala) [3,4]. We use this method since it is one of the few algorithms capable of handling large dimensions, nonconvexity, and nonsmoothness all at once. In addition, the LMBM has already proven itself in solving machine learning problems such as clustering [5], cardinality and clusterwise linear regression [6,7], and missing value imputation  [8].
We consider NNs with only one hidden layer but the number of hidden nodes is determined automatically using a novel constructive approach and an automated stopping procedure (ASP). More precisely, the number of nodes is determined incrementally by starting from one node. ASP bases the intelligent selection of initial weights on the iteratively self-updated regularization parameter. It is applied at each iteration of the constructive algorithm to stop training if there is no further improvement in the model. Thus, there are no tuneable hyperparameters in LMBNNR.
To summarize, the LMBNNR algorithm has some remarkable features, including:
  • It takes full advantage of nonsmooth models and nonsmooth optimization in solving NNR problems (no need for smoothing etc.);
  • It is hyperparameter-free due to the automated determination of the proper number of nodes;
  • It is applicable to large-scale regression problems;
  • It is an efficient and accurate predictive tool.
The structure of the paper is as follows. An overview of the related work is given in Section 2, while Section 3 provides the theoretical background on nonsmooth optimization and NNR. The problem statement—the ReLU-NNR problem—is given in Section 4, and in Section 5, the LMBNNR algorithm together with ASP is introduced. In Section 6, we present the performance of the LMBNNR algorithm and compare it with some state-of-the-art NNR algorithms. Section 7 concludes the paper.

2. Related Work

Optimization is at the core of machine learning. Although it is a well-known fact that many machine learning problems lead to solving nonsmooth optimization problems (e.g., hinge-loss, lasso, and ReLU), nonsmooth optimization methods are scarcely used in the machine learning society. The common practice is to minimize nonsmooth functions by ignoring the nonsmoothness and employing a popular and simple smooth solver (like the Newton method, e.g., [9]) or by applying some smoothing techniques (see, e.g., [10,11]). However, there are some successful exceptions. For example, abs-linear forms of prediction tasks are solved using a successive piecewise linearization method in [12,13], a primal-dual prox method for problems in which both the loss function and the regularizer are nonsmooth is developed in [14], various nonsmooth optimization methods are applied to solve clustering, classification, and regression problems in [5,6,15,16,17], and finally, nonsmooth optimization approaches are combined with support vector machines in [18,19,20].
NNs are among the most popular and powerful machine learning techniques. There are two main properties in any NNs: an activation function and an error (loss) function. The simplest activation function is linear, and NNs with this function can be easily trained. Nevertheless, they cannot learn complex mapping functions, resulting in a poor outcome in the testing phase. On the other hand, smooth nonlinear activation functions, like the sigmoid and the hyperbolic tangent activation functions, may lead to highly complex nonconvex loss functions and require numerous training iterations as well as many hidden nodes [21].
In [22], it is theoretically discussed that NNs with nonsmooth activation functions demonstrate high performance. Among these functions, the piecewise linear ReLU has lately become the default activation function for many types of NNs. This function is nonsmooth, but it has a simple mathematical form of f ( x ) = max { 0 , x } .  ReLU provides more sensitivity to the summed activation of the node compared with traditional smooth sigmoid and hyperbolic tangent functions, and it avoids easy saturation. Due to the computational advantages of its simple structure and, thus, strong training ability, ReLU is preferable for training complex relationships in NNs. In practice, instead of minimizing the resultant nonsmooth loss function directly, it is often replaced with a smoothed surrogate loss function (see, e.g., [11]). However, algorithms based on smoothing techniques involve the choice of smoothing parameters. The number of such parameters becomes large in large data sets, and their choice becomes problematic, which affects the accuracy of algorithms.
The stochastic (sub)gradient descent method (SGD) is another commonly used method to minimize the loss function in NNs algorithms. Although this method has been used intuitively to solve machine learning problems with nonsmooth activation and loss functions for years, its convergence for such functions has been proved only very recently [23]. SGD is efficient since it does not depend on the size of the data, but it is not accurate as an optimization method and may require a large number of function and subgradient evaluations. Moreover, SGD may easily diverge if the learning rate (step size) is too large. These drawbacks are due to the fact that SGD is based on the subgradient method for convex problems, while NNs problems are highly nonconvex. For more discussions on NNs, we refer to [24].
The most commonly used NNR algorithms are the feedforward backpropagation network (FFBPN) [25,26,27,28] and the radial-basis network (RBN) [29,30,31]. FFBPN and most of its variants converge to only locally optimal solutions [26,27], and they only work under the precondition that all the functions involved in NNR are differentiable. RBN and its modifications can be trained without local minima issues [32], but they involve a heuristic procedure to select parameters and hyperparameters.
The choice of the hyperparameters in NNs is a challenging problem. These parameters define the structure of NNs, and their optimal choice leads to an accurate model. Generally, the hyperparameters can be determined either by an algorithm or manually by the user. Tuning the hyperparameters manually is a tedious and time-consuming process, and thus, several algorithms with varying levels of automaticity have been proposed for this purpose (see, e.g., [33,34,35,36,37,38]). Nevertheless, there is no guarantee that in these algorithms the selected number of hidden units (usually the number of nodes) is optimal, and the question of a good hyperparameter tuning procedure remains open. Therefore, it is imperative to develop algorithms that can minimize nonsmooth loss functions as they are (i.e., without smoothing) and automate the calibration of the hyperparameters.

3. Theoretical Background and Notations

In this section, we provide some theoretical background and notations that are used throughout the paper.

3.1. Nonsmooth Optimization

Nonsmooth optimization refers to the problem of minimizing (or maximizing) functions that are not continuously differentiable [39]. We denote the d-dimensional Euclidean space by R d and the inner product by x y = i = 1 d x i y i , where x , y R d . The associated L 2 and L 1 -norms are x 2 = ( x x ) 1 / 2 and x 1 = i = 1 d | x i | , respectively.
A function f : R d R is called locally Lipschitz continuous on R d if for any bounded subset X R d there exists L > 0 such that
| f ( x ) f ( y ) | L x y 2 for all x , y X .
The Clarke subdifferential  f ( x ) of a locally Lipschitz continuous function f : R d R at a point x R d is given by [39,40]
f ( x ) = conv lim i f ( x i ) x i x and f ( x i ) exists ,
where “conv” denotes the convex hull of a set. A vector ξ f ( x ) is called a subgradient. The point x * R d is stationary, if 0 f ( x * ) . Note that stationarity is a necessary condition for local optimality [39].

3.2. Neural Networks for Regression

Let A be a given data set with n samples: A = { ( x i , y i ) R m × R i = 1 , , n } , where x i R m are the values of m input features and y i R are their outputs. In regression analysis, the aim is to find a function φ : R m R such that φ ( x i ) is a “good approximation” of y i . In other words, the following regression error is minimized:
i = 1 n | φ ( x i ) y i | p , p > 0 .
Applying NNs to the regression problem can lead to significantly higher predictive power compared with traditional regression. In addition, it can model more complex scenarios just by increasing the number of nodes. This concept is proved in the universal approximation theorem, which states that a single hidden layer feedforward network of sufficient complexity is able to approximate any given regression function on a compact set to an arbitrary degree [41].
In practice, NNs take several input features and, as a part of the learning process, multiply them by their weights and run them through an activation function and a loss function, which closely resemble the regression error (1). The loss function is used to estimate the error in the current model so that the weights can be updated to reduce this error on the next evaluation. The most commonly used loss function for NNR problems is the mean squared error (MSE or L 2 -norm). However, there are at least two valid reasons to choose the mean absolute error (MAE or L 1 -norm) over the L 2 -norm: first, the L 1 -norm makes the loss function with the ReLU activation function simpler, and second, the regression models with the L 1 -norm are more robust to outliers (see, e.g., [42]). Once the is trained, the optimal weights for the model (regression coefficients) are found to fit the data.

4. Nonsmooth Optimization Model of ReLU-NNR

In this section, we formulate the nonsmooth optimization model for the RELU-NNR problem, but first, we give some notations. Let us denote the number of hidden nodes in NNs by H and let w be the weight connection vector and w s be its element in the index place s, where s is the index triplet s = a b c (see Figure 1). Then w a b c states the weight that connects a-th layer’s b-th node to ( a + 1 ) -th layer’s c-th node. In addition, denote by P an output index value for the weights connecting the hidden layer’s nodes to the output node of the NN. In contrast to the hidden nodes, no activation function is applied to the output node. Finally, let B denote the middle index value for the weights connecting a layer’s bias term to a node in the next layer. The bias terms with this index can be thought of as the last node in each layer. The weight connection vector w for m input features and H hidden nodes has H ( m + 2 ) + 1 components. For the sake of clarity, we organize w = w H in an “increasing order” as
w H = ( w 011 , w 012 , , w 01 H , w 021 , w 022 , , w 02 H , , w 0 m 1 , w 0 m 2 , , w 0 m H , w 0 B 1 , w 0 B 2 , , w 0 B H , w 11 P , w 12 P , , w 1 H P , w 1 B P ) .
We define the following
s i j = w 0 B j + k = 1 m w 0 k j x k i , i = 1 , , n , j = 1 , , H , t i = w 1 B P + j = 1 H w 1 j P max { 0 , s i j } , i = 1 , , n ,
where x k i denotes the k-th coordinate of the i-th sample. The loss function using the L 1 -norm (cf. regression error (1) with p = 1 ) is
F H ( w ) = i = 1 n | t i y i | .
To avoid overfitting in the learning process, we add an extra element to the loss function F H . In most cases, L 1 -regularization is preferable as it reduces the weight values of less important input features. In addition, it is robust and insensitive to outliers. Thus, we rewrite (2) as
f H ( w ) = F H ( w ) + ρ H w 1 ,
where ρ H > 0 is an iteratively self-updated regularization parameter (to be described later). Furthermore, the nonsmooth, nonconvex optimization formulation for ReLU-NNR can be expressed as
minimize f H ( w ) subject to w R H ( m + 2 ) + 1 .

5. LMBNNR Algorithm

To solve the ReLU-NNR problem (3), we now introduce the LMBNNR algorithm. In addition, we recall the LMBM in the form used here and briefly discuss its convergence. A more detailed description of the algorithm can be found in the technical report [43].
The LMBNNR algorithm computes hidden nodes incrementally and uses ASP to detect the proper number of these nodes. More precisely, it starts with one node in the hidden layer and adds a new node to that layer at each iteration of the algorithm. It is worth noting that each time a node is added, m + 2 new connection weights appear. Starting from the initial weight w 1 R m + 3 , the LMBNNR algorithm (Algorithm 1) applies the LMBM (Algorithm 2) to solve the underlying ReLU-NNR problem. The solution is employed to generate initial weights for the next iteration. This procedure is repeated until ASP is activated or the maximum number of hidden nodes H max is reached. ASP is designed based on the value of the objective function; if the value of the objective is not improved in three subsequent iterations, then the LMBNNR algorithm is stopped. Note that the initialization of weights and the regularization parameter (described below) are determined in such a way that we have f H f H + 1 for all H = 1 , , H max and f H is an optimal value of (3) with H nodes.
We select the initial weights w 1 as
w 1 = 1 m , 1 m , , 1 m , 0 , 1 , 0 R m + 3 .
Here, the first m components are the weights from the nodes of the input layer to the node of the hidden layer. The weights from the bias terms w 1 B P and w 0 B 1 are set to zero, and the weight w 11 P from the hidden layer to the output is set to one. Figure 2 illustrates the weights in different iterations and the progress of the LMBNNR algorithm.
At subsequent iterations H = 2 , , H final (i.e., when we add a new node to the hidden layer, here H final is the final number of hidden nodes), we initialize the weights w H as follows: let w ¯ H 1 be the solution to the ( H 1 ) -th ReLU-NNR problem. First, we set weights for all but the last two hidden nodes in the output layer as
w 1 j P H = w ¯ 1 j P H 1 for all j = 1 , , H 2 , w 1 B P H = w ¯ 1 B P H 1
and the weights from the input layer to all but the last two hidden nodes as
w 0 i j H = w ¯ 0 i j H 1 for all i = 1 , , m , j = 1 , , H 2 , w 0 B j H = w ¯ 0 B j H 1 for all j = 1 , , H 2 .
Note that in the case of H = 2 , we set w 1 B P 2 = w ¯ 1 B P 1 . Otherwise, we skip this step as there are only two nodes in the hidden layer. Then we take the most recent weights obtained at the previous iteration and split these weights to get the initial weights connecting the input layer and the last two hidden nodes of the current iteration, that is,
w 0 i H H = w 0 i ( H 1 ) H = 1 2 w ¯ 0 i ( H 1 ) H 1 for all i = 1 , , m , w 0 B H H = w 0 B ( H 1 ) H = 1 2 w ¯ 0 B ( H 1 ) H 1 .
Weights from the last two hidden nodes to the output are
w 1 H P H = w 1 ( H 1 ) P H = w ¯ 1 ( H 1 ) P H 1 .
To update the regularization parameter ρ H , we use the value of the objective at the previous iteration and the weight connecting the newest node of the hidden layer to the output as follows:
ρ H = ρ H 1 · f H 1 f H 1 + | w 1 H P H | , with ρ 1 = 1.0 .
Remark 1.
With the used initialization of weights w H and the regularization parameter ρ H we always have f H f H + 1 for H = 1 , , H final . ASP is activated if the value of the objective is no longer improving. In other words, when adding more nodes to the hidden layer does not give us a better model after optimization of the weights.
Now we are ready to give the LMBNNR algorithm.
Algorithm 1: LMBNNR
Algorithms 16 00444 i001
Remark 2.
There are no tuneable hyperparameters in LMBNNR. Hence, there is no need for a validation set. It is sufficient to choose the maximum number of hidden nodes H max big enough (see Section 6 for suitable values).
Next, we describe the LMBM with slight modifications to its original version for solving the underlying ReLU-NNR problems in the LMBNNR algorithm. This method is called at every iteration of LMBNNR. For more details of the LMBM, we refer to [3,4].
Remark 3.
We use a nonmonotone line search [4,44] to find step sizes t L k and t R k when Algorithm 2 is combined with Algorithm 1. In addition, as in [9], we use a relatively low maximum number of iterations, k max , to avoid overfitting.
Remark 4.
The search direction in Algorithm 2 is computed using the L-BFGS update after a serious step and the L-SR1 update after a null step. The updating formulae are similar to those in the classical limited memory variable metric methods for smooth optimization [45]. Nevertheless, the correction vectors u k and s k are obtained using subgradients instead of gradients and the auxiliary point instead of the new iteration point.
Remark 5.
The classical linearization error may be negative in the case of a nonconvex objective function. Therefore, a subgradient locality measure β k , which is a generalization of the linearization error for nonconvex functions (see, e.g., [46]), is used in Algorithm 2.
Algorithm 2: LMBM for ReLU-NNR problems
Algorithms 16 00444 i002
We now recall the convergence properties of the LMBM in the case of ReLU-NNR problems. Since the objective function f H : R H ( m + 2 ) + 1 R is locally Lipschitz continuous and upper semi-smooth (see e.g., [47]), and the level set { w H R H ( m + 2 ) + 1 f H ( w H ) f H ( w 1 H ) } is bounded for every starting weight w 1 H R H ( m + 2 ) + 1 , all the assumptions needed for the global convergence of the original LMBM are satisfied. Therefore, the theorems on the convergence of the LMBM proved in [3,4] can be modified for the ReLU-NNR problems as follows.
Theorem 1.
If the LMBM terminates after a finite number of iterations, say at iteration k, then the weight w k H is a stationary point of the  ReLU-NNR problem (3).
Theorem 2.
Every accumulation point w ¯ H of the sequence { w k H } generated by the  LMBM is a stationary point of the  ReLU-NNR problem (3).

6. Numerical Experiments

Using some real-world data sets and performance measures, we evaluate the performance of the proposed LMBNNR algorithm. In addition, we compare it with three different widely used NNR algorithms whose implementations are freely available. That is, the backpropagation NNR (BPN) algorithm utilizing TensorFlow (https://www.tensorflow.org/(accessed on 10 September 2023)), Extreme Learning Machine (ELM) [48], and Monotone Multi-Layer Perceptron Neural Networks (MONMLP) [49].

6.1. Data Sets and Performance Measures

A brief description of data sets is given in Table 1 and the references therein. The data sets are divided randomly into training (80%) and test (20%) sets. To get comparable results, we use the same training and test sets for all the methods. We apply the following performance measures: root mean square error (RMSE), mean absolute error (MAE), coefficient of determination ( R 2 ), and Pearson’s correlation coefficient (r) (see the Supplementary Material for more details).

6.2. Implementation of Algorithms

The proposed LMBNNR algorithm was implemented in Fortran 2003, and the computational experiments were carried out on an iMac (macOS Big Sur 11.6) with a 4.0 GHz Intel(R) Core(TM) i7 machine and 16 GB of RAM. The source code is available at http://napsu.karmitsa.fi/lmbnnr (accessed on 10 September 2023). There are no tuneable parameters in the LMBNNR algorithm.
The BPN algorithm is implemented using TensorFlow in Google Colab. We use the ReLU activation for the hidden layer, the linear activation for the output layer, and the MSE loss function. We use MSE since it usually worked better than MAE in our preliminary experiments with the BPN algorithm. This is probably due to the smoothness of MSE. Naturally, with the proposed LMBNNR algorithm, we do not have difficulties with the nonsmoothness as it applies the nonsmooth optimization solver LMBM. In the BPN algorithm, the Keras optimizer SGD with the default parameters and the following three different combinations of batch size (the number of samples that will be propagated through the network) and number of epochs (the number of complete passes through the training data) are used:
  • Mini-Batch Gradient Descent (MBGD): batch size = 32 and number of epochs = 1 . These are the default values for the BPN algorithm;
  • Batch Gradient Descent (BGD): batch size = size of the training data and number of epochs = 1 . These choices mimic the proposed LMBNNR algorithm;
  • Stochastic Gradient Descent (SGD): batch size = 1 and number of epochs = 100 for data sets with less than 100,000 samples and number of epochs = 10 for larger data. We reduce the number of epochs in the latter case due to very long computational times and the fact that the larger number of epochs often leads to NaN loss function values. SGD aims to be as a stochastic version of the BPN algorithm as possible.
The algorithms ELM and MONMLP are implemented in R using the packages “elmNNRcpp” (ELM is available at https://CRAN.R-project.org/package=elmNNRcpp (accessed on 10 September 2023)) and “monmlp” (MONMLP is available at https://CRAN.R-project.org/package=monmlp (accessed on 10 September 2023)). The parameters are the default values provided in the given references, but the number of hidden layers with MONMLP was set to one. The tests with ELM and MONMLP were run on Windows, 2.6 GHz Intel(R) Core(TM) i7-9750H. We run ELM, MONMLP, and the BPN algorithms with the number of hidden nodes set to 2, 5, 10, 50, 100, 200, 500, and 5000. Moreover, since SGD is used as an optimizer in the BPN algorithms as a stochastic method, we run the BPN algorithms ten times for all problems and report the average. Note that LMBNNR needs to be applied to solve the ReLU-NNR problem only once for each data set.
Remark 6.
At every iteration of the LMBNNR algorithm, we solve an optimization problem with the dimension H ( m + 2 ) + 1 , where H is the number of hidden nodes and m is the number of features in training samples. The convergence rate of the LMBM, employed as an underlying solver in the proposed LMBNNR algorithm, has not been studied extensively, but, as a bundle method, it is at most linear (see, e.g., [62]), while the search direction can be computed within O ( m ) operations [3]. For comparison purposes, we mention that the time complexity of a BPN algorithm is typically O ( n ) , where n is the number of training samples. A more specific analysis of the time complexity of LMBNNR remains a subject for future research.

6.3. Results and Discussion

The results of our experiments are given in Table 2 and Table 3, where we provide the RMSE values for the test set and the used CPU time in seconds. The results using the other three performance measures (MAE, R 2 , and r) are given as Supplementary Material for this paper. Since the algorithms are implemented in different programming languages and run on different platforms, the CPU times reported here are not directly comparable. Nevertheless, we can still compare the magnitudes of the computational times used. For the BPN algorithms, the results (including CPU times) are the average of results over ten runs, unless some of the runs lead to NaN loss function values. In that case, the results are the average of the successful runs. Note that obtaining a NaN loss function value is a natural property of SGD and related methods if the learning rate is too large. If there is “NaN” in the tables, then all ten runs lead to the NaN solution. In all separate runs, the maximum time limit is set to 1 h when n × m < 10 6 (n and m are the number of samples and features, respectively) and 2 h otherwise. In tables, “t-lim” means that an algorithm gives no solution within the time limit.
To evaluate the reliability of ASP, we force LMBNNR to solve ReLU-NNR problems up to 200 nodes. With LMBNNR, the results with smaller numbers of nodes are obtained as a side product. In tables, we report the results obtained with different numbers of nodes (the same numbers that we used for testing the other algorithms), the best solution with respect to the RMSE of the test set, and the results when applying ASP. The last two are denoted by “Best” and “ASP”, respectively.
As we report all the results with different model parameter combinations (i.e., results with different numbers of nodes or different combinations of batch sizes and epochs), we do not use separate validation sets in our experiments. Instead, we use the best results, namely the smallest RSME of the test set, obtained with any other NNR algorithm than LMBNNR, in our comparison. Naturally, in real-world predictions, we would not know which result is the best, and a separate validation set should be employed to tune the hyperparameters with NNR algorithms but LMBNNR. Note that this kind of experimental setting favors the other methods over the proposed one. In tables, the best results obtained with any other NNR algorithm than LMBNNR are presented using boldface font. To make the comparison even more challenging for the LMBNNR algorithm, we compare the result that it gives with ASP to the best result obtained with any other tested algorithm. We point out that the results using the other three performance measures support the conclusions drawn from RMSE. Nevertheless, MAE often indicates slightly better performance of the LMBNNR algorithm than RMSE (see the Supplementary Material). This is, in particular, due to the used loss function or the regularization term.
The predictions with the LMBNNR algorithm in termination are better than those of any other tested NNR algorithm in 5 data sets out of 12. In addition, MONMLP with the selected best number of hidden nodes is the most accurate algorithm in five data sets, and both SGD and ELM are in one data set. MBGD and BGD never produce the most accurate results in our experiments. Therefore, in terms of accuracy, the only noteworthy challenger for LMBNNR is MONMLP. However, the pairwise comparison of these two algorithms shows that the required computational times are clearly in favor of LMBNNR. Indeed, MONMLP is the most time-consuming of the algorithms tested (sometimes together with SGD): in the largest data sets, CIFAR-10 and Greenhouse, MONMLP finds no solution within the time limit, and even in the smallest data tested, it fails to give a solution with larger numbers of hidden nodes ( H = 500 and 5000). Moreover, MONMLP has a big deviation in the prediction accuracy obtained with different numbers of hidden nodes. For example, in residential building data it gives the most accurate prediction of all tested algorithms (RMSE = 60.932) with two hidden nodes and almost the worst prediction (RMSE = 1200.323) with 50 hidden nodes. In practice, this means that finding good hyperparameters for MONMLP may become an issue.
Although SGD is the most accurate version of the BPN algorithm when the number of features is relatively low, it usually requires a lot more computational time than LMBNNR. Moreover, SGD fails almost always when the number of features is large.
If we only consider computational times, then MBGD, BGD, and ELM would be our choices. From these, BGD is out of the question, as the predictions it produces are not at all accurate. Obviously, the parameter choices used in BGD are not suitable for the BPN algorithm, and the main reason to keep this version here is purely theoretical as these parameters mimic the LMBNNR algorithm. In addition, MBGD, a version using the default parameters of the BPN algorithm, often fails to produce accurate predictions, although it is clearly better than BGD. The pairwise comparison with LMBNNR shows that MBGD produces more accurate prediction results only in one data set: online news popularity. Moreover, similar to SGD, MBGD fails when the number of features is large. Therefore, in large data sets, the only real challenger for LMBNNR is ELM. Indeed, ELM is a very efficient method, usually using only a few seconds to solve an individual ReLU-NNR problem. Nevertheless, LMBNNR produces more accurate predictions than ELM in 9 data sets out of 12, and it is fairly efficient as well.
ASP seems to work pretty well: it usually triggers within a few iterations after the best solution is obtained, and the accuracy of the prediction is very close to the best one. The average number of hidden nodes used before ASP is 40. This is considerably less than what is needed to obtain an accurate prediction with any of the BPN algorithms.
It is worth noting that finding good hyperparameters for most NNR algorithms is time-consuming and may require the utilization of a separate validation. For instance, if we just run a NNR algorithm with two different numbers of nodes, the computational time is doubled—not to mention the time needed to prepare the separate runs and validate the results. With the LMBNNR algorithm, the intermediate results are obtained within the time required by the largest number of nodes without the necessity of separate runs. In addition, there is no need to use a validation set to fit hyperparameters in LMBNNR, as there is none.

7. Conclusions

In this paper, we introduce a novel neural networks (NNs) algorithm, LMBNNR, to solve regression problems in large data sets. The regression problem is modeled using NNs with one hidden layer, the piecewise linear activation function known as ReLU, and the L 1 -based loss function. We utilize a modified version of the limited memory bundle method to minimize the objective, as this method is able to handle large dimensions, nonconvexity, and nonsmoothness very efficiently.
The proposed algorithm requires no hyperparameter tuning. It starts with one hidden node and gradually adds more nodes with each iteration. The solution of the previous iteration is used as a starting point for the next one to obtain either global or deep local minimizers for the ReLU-NNR problem. The algorithm terminates when the optimal value of the loss function cannot be improved in several successive iterations or when the maximum number of hidden nodes is reached.
The LMBNNR algorithm is tested using 12 large real-world data sets and compared with the backpropagation NNR (BPN) algorithm using TensorFlow, Extreme Learning Machine (ELM), and Monotone Multi-Layer Perceptron Neural Networks (MONMLP). The results show that the proposed algorithm outperforms the BPN algorithm, even if we tested the latter with different hyperparameter settings and used the best results in comparison. In addition, the LMBNNR algorithm is far more accurate than ELM and significantly faster than MONMLP. Thus, we conclude that the proposed algorithm is accurate and efficient for solving regression problems with large numbers of samples and large numbers of input features.
The results presented in this paper demonstrate the importance of nonsmooth optimization for NNs: the use of simple but nonsmooth activation functions together with powerful nonsmooth optimization methods can lead to the development of accurate and efficient hyperparameter-free NNs algorithms. The approach proposed in this paper can be extended to model NNs for classification and to build NNs with more than one hidden layer in order to develop robust and effective deep learning algorithms.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/a16090444/s1, Description of used performance measures with Tables S1–S12 including obtained MAE, R 2 and r results in various data sets.

Author Contributions

Investigation, N.K. and S.T.; Methodology, N.K., P.P. and A.M.B.; Software, N.K., K.J. and P.P.; Writing—original draft, N.K., S.T., K.J., A.M.B. and M.M.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was financially supported by Research Council of Finland grants #289500, #319274, #345804, and #345805, and by the Australian Government through the Australian Research Council’s Discovery Projects funding scheme (Project no. DP190100580).

Data Availability Statement

The proposed LMBNNR algorithm is available at http://napsu.karmitsa.fi/lmbnnr (accessed on 10 September 2023). The algorithms ELM and MONMLP are implemented in R using the packages “elmNNRcpp” and “monmlp” available in CRAN. More precisely, ELM is available at https://CRAN.R-project.org/package=elmNNRcpp (accessed on 10 September 2023) and MONMLP is available at https://CRAN.R-project.org/package=monmlp (accessed on 10 September 2023). The used data sets are from UC Irvine Machine Learning Repository (UCI, https://archive.ics.uci.edu/ (accessed on 10 September 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Malte, J. Artificial neural network regression models in a panel setting: Predicting economic growth. Econ. Model. 2020, 91, 148–154. [Google Scholar]
  2. Pepelyshev, A.; Zhigljavsky, A.; Žilinskas, A. Performance of global random search algorithms for large dimensions. J. Glob. Optim. 2018, 71, 57–71. [Google Scholar] [CrossRef]
  3. Haarala, N.; Miettinen, K.; Mäkelä, M.M. Globally Convergent Limited Memory Bundle Method for Large-Scale Nonsmooth Optimization. Math. Program. 2007, 109, 181–205. [Google Scholar] [CrossRef]
  4. Karmitsa, N. Limited Memory Bundle Method and Its Variations for Large-Scale Nonsmooth Optimization. In Numerical Nonsmooth Optimization: State of the Art Algorithms; Bagirov, A.M., Gaudioso, M., Karmitsa, N., Mäkelä, M.M., Taheri, S., Eds.; Springer: Cham, Switzerland, 2020; pp. 167–200. [Google Scholar]
  5. Bagirov, A.M.; Karmitsa, N.; Taheri, S. Partitional Clustering via Nonsmooth Optimization: Clustering via Optimization; Springer: Cham, Switzerland, 2020. [Google Scholar]
  6. Halkola, A.; Joki, K.; Mirtti, T.; Mäkelä, M.M.; Aittokallio, T.; Laajala, T. OSCAR: Optimal subset cardinality regression using the L0-pseudonorm with applications to prognostic modelling of prostate cancer. PLoS Comput. Biol. 2023, 19, e1010333. [Google Scholar] [CrossRef] [PubMed]
  7. Karmitsa, N.; Bagirov, A.M.; Taheri, S.; Joki, K. Limited Memory Bundle Method for Clusterwise Linear Regression. In Computational Sciences and Artificial Intelligence in Industry; Tuovinen, T., Periaux, J., Neittaanmäki, P., Eds.; Springer: Cham, Switzerland, 2022; pp. 109–122. [Google Scholar]
  8. Karmitsa, N.; Taheri, S.; Bagirov, A.M.; Mäkinen, P. Missing value imputation via clusterwise linear regression. IEEE Trans. Knowl. Data Eng. 2022, 34, 1889–1901. [Google Scholar] [CrossRef]
  9. Airola, A.; Pahikkala, T. Fast Kronecker product kernel methods via generalized vec trick. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 3374–3387. [Google Scholar]
  10. Bian, W.; Chen, X. Neural network for nonsmooth, nonconvex constrained minimization via smooth approximation. IEEE Trans. Neural Netw. Learn. Syst. 2014, 25, 545–556. [Google Scholar] [CrossRef] [PubMed]
  11. JunRu, L.; Hong, Q.; Bo, Z. Learning with smooth Hinge losses. Neurocomputing 2021, 463, 379–387. [Google Scholar]
  12. Griewank, A.; Rojas, A. Treating Artificial Neural Net Training as a Nonsmooth Global Optimization Problem. In Machine Learning, Optimization, and Data Science. LOD 2019; Nicosia, G., Pardalos, P., Umeton, R., Giuffrida, G., Sciacca, V., Eds.; Springer: Cham, Switzerland, 2019; Volume 11943. [Google Scholar]
  13. Griewank, A.; Rojas, A. Generalized Abs-Linear Learning by Mixed Binary Quadratic Optimization. In Proceedings of African Conference on Research in Computer Science CARI 2020, Thes, Senegal, 14–17 October 2020; Available online: https://hal.science/hal-02945038 (accessed on 10 September 2023).
  14. Yang, T.; Mahdavi, M.; Jin, R.; Zhu, S. An efficient primal dual prox method for non-smooth optimization. Mach. Learn. 2015, 98, 369–406. [Google Scholar] [CrossRef]
  15. Astorino, A.; Gaudioso, M. Ellipsoidal separation for classification problems. Optim. Methods Softw. 2005, 20, 267–276. [Google Scholar] [CrossRef]
  16. Bagirov, A.M.; Taheri, S.; Karmitsa, N.; Sultanova, N.; Asadi, S. Robust piecewise linear L1-regression via nonsmooth DC optimization. Optim. Methods Softw. 2022, 37, 1289–1309. [Google Scholar] [CrossRef]
  17. Gaudioso, M.; Giallombardo, G.; Miglionico, G.; Vocaturo, E. Classification in the multiple instance learning framework via spherical separation. Soft Comput. 2020, 24, 5071–5077. [Google Scholar] [CrossRef]
  18. Astorino, A.; Fuduli, A. Support vector machine polyhedral separability in semisupervised learning. J. Optim. Theory Appl. 2015, 164, 1039–1050. [Google Scholar] [CrossRef]
  19. Astorino, A.; Fuduli, A. The proximal trajectory algorithm in SVM cross validation. IEEE Trans. Neural Netw. Learn. Syst. 2015, 27, 966–977. [Google Scholar] [CrossRef] [PubMed]
  20. Joki, K.; Bagirov, A.M.; Karmitsa, N.; Mäkelä, M.M.; Taheri, S. Clusterwise support vector linear regression. Eur. J. Oper. Res. 2020, 287, 19–35. [Google Scholar] [CrossRef]
  21. Selmic, R.R.; Lewis, F.L. Neural-network approximation of piecewise continuous functions: Application to friction compensation. IEEE Trans. Neural Netw. 2002, 13, 745–751. [Google Scholar] [CrossRef] [PubMed]
  22. Imaizumi, M.; Fukumizu, K. Deep Neural Networks Learn Non-Smooth Functions Effectively. In Proceedings of the Machine Learning Research, Naha, Okinawa, Japan, 16–18 April 2019; Volume 89, pp. 869–878. [Google Scholar]
  23. Davies, D.; Drusvyatskiy, D.; Kakade, S.; Lee, J. Stochastic subgradient method converges on tame functions. Found. Comput. Math. 2020, 20, 119–154. [Google Scholar] [CrossRef]
  24. Aggarwal, C. Neural Networks and Deep Learning; Springer: Berlin, Germany, 2018. [Google Scholar]
  25. Rumelhart, D.; Hinton, G.; Williams, R. Learning representations by back-propagating errors. Nature 1988, 323, 533–536. [Google Scholar] [CrossRef]
  26. Huang, G.B. Learning capability and storage capacity of two-hidden-layer feedforward networks. IEEE Trans. Neural Netw. 2003, 14, 274–281. [Google Scholar] [CrossRef]
  27. Reed, R.; Marks, R.J. Neural Smithing: Supervised Learning in Feedforward Artificial Neural Networks; The MIT Press: Cambridge, MA, USA, 1998. [Google Scholar]
  28. Vicoveanu, P.; Vasilache, I.; Scripcariu, I.; Nemescu, D.; Carauleanu, A.; Vicoveanu, D.; Covali, A.; Filip, C.; Socolov, D. Use of a feed-forward back propagation network for the prediction of small for gestational age newborns in a cohort of pregnant patients with thrombophilia. Diagnostics 2022, 12, 1009. [Google Scholar] [CrossRef]
  29. Broomhead, D.; Lowe, D. Radial Basis Functions, Multi-Variable Functional Interpolation and Adaptive Networks; Royals Signals and Radar Establishment: Great Malvern, UK, 1988. [Google Scholar]
  30. Olusola, A.O.; Ashiribo, S.W.; Mazzara, M. A machine learning prediction of academic performance of secondary school students using radial basis function neural network. Trends Neurosci. Educ. 2022, 22, 100190. [Google Scholar]
  31. Zhang, D.; Zhang, N.; Ye, N.; Fang, J.; Han, X. Hybrid learning algorithm of radial basis function networks for reliability analysis. IEEE Trans. Reliab. 2021, 70, 887–900. [Google Scholar] [CrossRef]
  32. Haykin, S. Neural Networks: A Comprehensive Foundation; Prentice Hall: Upper Saddle River, NJ, USA, 2007. [Google Scholar]
  33. Faris, H.; Mirjalili, S.; Aljarah, I. Automatic selection of hidden neurons and weights in neural networks using grey wolf optimizer based on a hybrid encoding scheme. Int. J. Mach. Learn. Cybern. 2019, 10, 2901–2920. [Google Scholar] [CrossRef]
  34. Huang, D.S.; Du, J.X. A constructive hybrid structure optimization methodology for radial basis probabilistic neural networks. IEEE Trans. Neural Netw. 2008, 19, 2099–2115. [Google Scholar] [CrossRef] [PubMed]
  35. Odikwa, H.; Ifeanyi-Reuben, N.; Thom-Manuel, O.M. An improved approach for hidden nodes selection in artificial neural network. Int. J. Appl. Inf. Syst. 2020, 12, 7–14. [Google Scholar]
  36. Leung, F.F.; Lam, H.K.; Ling, S.H.; Tam, P.S. Tuning of the structure and parameters of a neural network using an improved genetic algorithm. IEEE Trans. Neural Netw. 2003, 11, 79–88. [Google Scholar] [CrossRef] [PubMed]
  37. Stathakis, D. How many hidden layers and nodes? Int. J. Remote Sens. 2009, 30, 2133–2147. [Google Scholar] [CrossRef]
  38. Tsai, J.T.; Chou, J.H.; Liu, T.K. Tuning the structure and parameters of a neural network by using hybrid Taguchi-genetic algorithm. IEEE Trans. Neural Netw. 2006, 17, 69–80. [Google Scholar] [CrossRef]
  39. Bagirov, A.M.; Karmitsa, N.; Mäkelä, M.M. Introduction to Nonsmooth Optimization: Theory, Practice and Software; Springer: Cham, Switzerland, 2014. [Google Scholar]
  40. Clarke, F.H. Optimization and Nonsmooth Analysis; Wiley-Interscience: New York, NY, USA, 1983. [Google Scholar]
  41. Wilamowski, B.M. Neural Network Architectures. In The Industrial Electronics Handbook; CRC Press: Boca Raton, FL, USA, 2011. [Google Scholar]
  42. Kärkkäinen, T.; Heikkola, E. Robust formulations for training multilayer perceptrons. Neural Comput. 2004, 16, 837–862. [Google Scholar] [CrossRef]
  43. Karmitsa, N.; Taheri, S.; Joki, K.; Mäkinen, P.; Bagirov, A.; Mäkelä, M.M. Hyperparameter-Free NN Algorithm for Large-Scale Regression Problems; TUCS Technical Report, No. 1213; Turku Centre for Computer Science: Turku, Finland, 2020; Available online: https://napsu.karmitsa.fi/publications/lmbnnr_tucs.pdf (accessed on 10 September 2023).
  44. Zhang, H.; Hager, W. A nonmonotone line search technique and its application to unconstrained optimization. SIAM J. Optim. 2004, 14, 1043–1056. [Google Scholar] [CrossRef]
  45. Byrd, R.H.; Nocedal, J.; Schnabel, R.B. Representations of quasi-Newton matrices and their use in limited memory methods. Math. Program. 1994, 63, 129–156. [Google Scholar] [CrossRef]
  46. Kiwiel, K.C. Methods of Descent for Nondifferentiable Optimization; Lecture Notes in Mathematics 1133; Springer: Berlin, Germany, 1985. [Google Scholar]
  47. Bihain, A. Optimization of upper semidifferentiable functions. J. Optim. Theory Appl. 1984, 4, 545–568. [Google Scholar] [CrossRef]
  48. Huang, G.B.; Zhou, H.; Ding, X.; Zhang, R. Extreme learning machine for regression and multiclass classification. IEEE Trans. Syst. Man Cybern. 2011, 42, 513–529. [Google Scholar] [CrossRef] [PubMed]
  49. Lang, B. Monotonic Multi-Layer Perceptron Networks as Universal Approximators. In Artificial Neural Networks: Formal Models and Their Applications—ICANN 2005; Duch, W., Kacprzyk, J., Oja, E., Zadroźny, S., Eds.; Springer: Berlin/Heidelberg, Germany, 2005; Volume 3697. [Google Scholar]
  50. Tüfekci, P. Prediction of full load electrical power output of a base load operated combined cycle power plant using machine learning methods. Int. J. Electr. Power Energy Syst. 2014, 60, 126–140. [Google Scholar] [CrossRef]
  51. Kaya, H.; Tüfekci, P.; Gürgen, S.F. Local and Global Learning Methods for Predicting Power of a Combined Gas & Steam Turbine. In Proceedings of the International Conference on Emerging Trends in Computer and Electronics Engineering ICETCEE 2012, Dubai, United Arab Emirates, 24–25 March 2012; pp. 13–18. [Google Scholar]
  52. Dua, D.; Karra Taniskidou, E. UCI Machine Learning Repository. 2017. Available online: http://archive.ics.uci.edu/ml (accessed on 25 November 2020).
  53. Yeh, I. Modeling of strength of high performance concrete using artificial neural networks. Cem. Concr. Res. 1998, 28, 1797–1808. [Google Scholar] [CrossRef]
  54. Harrison, D.; Rubinfeld, D. Hedonic prices and the demand for clean air. J. Environ. Econ. Manag. 1978, 5, 81–102. [Google Scholar] [CrossRef]
  55. Paredes, E.; Ballester-Ripoll, R. SGEMM GPU kernel performance (2018). In UCI Machine Learning Repository. Available online: https://doi.org/10.24432/C5MK70 (accessed on 10 September 2023).
  56. Nugteren, C.; Codreanu, V. CLTune: A Generic Auto-Tuner for OpenCL Kernels. In Proceedings of the MCSoC: 9th International Symposium on Embedded Multicore/Many-core Systems-on-Chip, Turin, Italy, 23–25 September 2015. [Google Scholar]
  57. Fernandes, K.; Vinagre, P.; Cortez, P. A Proactive Intelligent Decision Support System for Predicting the Popularity of Online News. In Proceedings of the 17th EPIA 2015—Portuguese Conference on Artificial Intelligence, Coimbra, Portugal, 8–11 September 2015. [Google Scholar]
  58. Rafiei, M.; Adeli, H. A novel machine learning model for estimation of sale prices of real estate units. ASCE J. Constr. Eng. Manag. 2015, 142, 04015066. [Google Scholar] [CrossRef]
  59. Buza, K. Feedback Prediction for Blogs. In Data Analysis, Machine Learning and Knowledge Discovery; Springer International Publishing: Cham, Switzerland, 2014; pp. 145–152. [Google Scholar]
  60. Krizhevsky, A. Learning Multiple Layers of Features from Tiny Images. 2009. Available online: https://www.cs.toronto.edu/~kriz/cifar.html (accessed on 14 November 2021).
  61. Lucas, D.; Yver Kwok, C.; Cameron-Smith, P.; Graven, H.; Bergmann, D.; Guilderson, T.; Weiss, R.; Keeling, R. Designing optimal greenhouse gas observing networks that consider performance and cost. Geosci. Instrum. Methods Data Syst. 2015, 4, 121–137. [Google Scholar] [CrossRef]
  62. Diaz, M.; Grimmer, B. Optimal convergence rates for the proximal bundle method. SIAM J. Optim. 2023, 33, 424–454. [Google Scholar] [CrossRef]
Figure 1. A simple NN model with one output, one hidden layer and H hidden nodes.
Figure 1. A simple NN model with one output, one hidden layer and H hidden nodes.
Algorithms 16 00444 g001
Figure 2. Construction of model by LMBNNR.
Figure 2. Construction of model by LMBNNR.
Algorithms 16 00444 g002
Table 1. Brief description of data sets.
Table 1. Brief description of data sets.
Data SetNo. of SamplesNo. of FeaturesReference
Combined cycle power plant95685[50,51]
Airfoil self-noise15036[52]
Concrete compressive strength10309[53]
Physicochemical properties
      of protein tertiary structure45,73010[52]
Boston housing data50614[54]
SGEMM GPU kernel performance 1241,60015[55,56]
Online news popularity39,64459[57]
Residential building data set 1372108[58]
BlogFeedback 252,397281[59]
ISOLET7797618[52]
CIFAR-1060,0003073[60]
Greenhouse gas observing network29215232[61]
1 Used with the first output feature. 2 Only the training data set.
Table 2. Results with relatively small numbers of features.
Table 2. Results with relatively small numbers of features.
LMBNNRMBGDBGDSGDELMMONMLP
H RMSECPURMSECPURMSECPURMSECPURMSECPURMSECPU
Combined cycle power plant
24.4330.148.4721.4519.0610.514.5821036.4218.9870.006.81117.10
54.2670.616.0910.9720.5470.584.415790.275.1920.006.69240.72
104.2162.025.7840.7817.8090.474.278650.955.1920.026.66797.91
504.14012.724.7121.0418.7900.664.215679.025.1920.179.4241347.99
1004.13917.704.6120.6517.2130.384.167639.655.1920.4917.0071154.44
2004.13931.924.5660.6617.6490.694.190691.765.1922.34t-lim
500 4.5341.0016.5140.634.168761.365.1928.03t-lim
5000 4.5331.8116.7101.534.205897.525.1921347.99t-lim
Best:4.1388.14 H = 29
ASP:4.13913.73 H = 61
Airfoil self-noise
24.9070.026.7560.3610.1290.294.387102.6587.6220.005.7094.98
54.4070.106.5800.359.4650.303.688102.1330.2740.005.6349.05
104.4100.305.8880.358.9500.332.804101.0130.2740.005.64819.72
504.3447.055.5950.347.1140.302.386100.8730.2740.006.308231.75
1001.89719.805.5160.406.8400.332.018102.9130.2740.056.727224.88
2001.89729.045.3460.386.9670.322.044108.7330.2740.176.723890.39
500 5.3190.356.7430.302.194116.6030.2741.27t-lim
5000 5.2990.426.6890.372.035133.0030.27446.66t-lim
Best:1.89717.29 H = 83
ASP:1.89720.15 H = 105
Concrete compressive strength
211.6570.0217.5540.3730.3050.337.65970.0016.5110.006.3395.43
56.8080.1015.7360.3723.2200.326.37274.1914.1730.005.02615.21
106.1040.3415.1560.3823.9490.365.57771.789.8610.004.33833.68
505.2356.9413.7290.3917.8240.354.91874.429.8610.0215.042139.85
1005.22415.6013.2770.3918.0690.315.01972.879.8610.0515.899472.58
2005.22422.4212.4720.3617.3650.314.80776.919.8610.2515.8951573.22
500 12.3790.4016.5150.424.77179.929.8610.57t-lim
5000 12.3620.4516.4960.414.46894.339.8619.89t-lim
Best:5.22415.06 H = 96
ASP:5.22513.22 H = 74
Physicochemical properties of protein
25.3041.065.2232.297.8380.715.127 ( 1 ) 3368.546.7030.004.962111.60
55.0954.895.1222.148.3660.72NaN6.4570.034.798296.25
105.06217.765.0822.098.1680.734.856 ( 2 ) 3356.055.2150.084.947771.12
504.794109.344.9752.056.7840.73NaN5.2150.72t-lim
1004.793151.024.9521.866.4660.69NaN5.2152.56t-lim
2004.793259.124.9412.316.1550.77NaN5.21510.80t-lim
500 4.9723.066.1301.07NaN5.21526.48t-lim
5000 4.9425.796.0434.68NaN5.2152761.84t-lim
Best:4.793114.31 H = 56
ASP:4.793131.93 H = 79
( 1 ) 2/10 runs led to NaN loss function value.
( 2 ) 8/10 runs led to NaN loss function value.
Boston housing
25.5390.0110.9830.4713.3350.334.67540.029.8830.012.7734.72
55.5340.049.3450.4912.2030.334.01933.658.9220.002.4327.90
104.5510.228.7530.5112.6630.323.93634.166.4330.002.59117.06
504.2284.038.0610.4910.8310.333.78435.724.7860.003.96780.14
1004.2287.757.7020.5410.6360.293.79437.224.7860.055.842275.42
2004.22812.377.4980.3310.5160.553.60938.104.7860.197.588967.62
500 7.6990.4710.4100.323.65638.444.7860.22t-lim
5000 7.6250.509.9560.373.66947.454.7862.31t-lim
Best:4.2281.87 H = 30
ASP:4.2282.97 H = 39
SGEMM GPU kernel performance
2223.2581.37168.9718.11515.6472.15148.5311536.53327.4040.0591.07970.11
5118.6975.87128.1867.98549.2012.18105.8511812.15310.1720.1568.932551.58
1094.82619.60116.1958.02502.7932.1478.3091839.19289.0190.3681.376779.13
5093.36275.16108.8678.09429.3322.3243.9591669.78285.9534.56t-lim
10093.362234.57106.4408.09402.9352.3141.3231809.20285.95315.92t-lim
20093.362887.41104.3078.44385.3602.5237.3111835.75285.95360.72t-lim
500 102.03310.12383.1673.2534.2831679.94285.953147.90t-lim
5000 101.88623.86367.07423.8328.5242243.39t-limt-lim
Best:93.36226.94 H = 19
ASP:94.27025.78 H = 18
Table 3. Results with large numbers of features.
Table 3. Results with large numbers of features.
LMBNNRMBGDBGDSGDELMMONMLP
H RMSECPURMSECPURMSECPURMSECPURMSECPURMSECPU
Online news popularity
212,387.460.7512,301.201.6815,701.630.6312,475.79 ( 1 ) 2664.1111,228.420.0310,626.67692.14
512,377.173.7812,304.051.6518,562.270.65NaN11,189.210.059500.962310.14
1012,365.3813.2712,351.971.6218,095.870.69NaN11,098.650.099548.602554.58
5012,365.1952.3812,670.861.7716,461.910.67NaN11,046.390.94t-lim
10012,365.19202.5012,714.961.7814,731.390.68NaN11,043.693.07t-lim
20012,365.19782.5912,937.681.8013,977.350.74NaN11,043.6910.28t-lim
500 12,519.332.1013,089.170.91NaN11,043.6923.44t-lim
5000 12,322.536.2712,364.063.52NaNt-limt-lim
Best:12,365.1916.49 H = 19
ASP:12,365.1916.49 H = 19
( 1 ) 7/10 runs led to NaN loss function value.
Residential building
2239.140.091163.960.541635.840.32236.87 ( 1 ) 25.921838.860.0060.9322.39
5240.630.371048.950.351350.820.33192.68 ( 2 ) 25.931832.780.00399.0725.40
10183.461.401046.230.361524.370.31155.14 ( 3 ) 26.91976.840.001200.1349.06
50183.118.341014.720.351217.390.32148.19 ( 3 ) 27.61170.170.001200.32966.99
100183.1112.93998.600.361119.600.33NaN128.270.03t-lim
200183.1123.77956.510.351018.660.31NaN128.270.06t-lim
500 972.950.35998.180.33NaN128.270.14t-lim
5000 775.800.45916.630.40NaN128.271.38t-lim
Best:182.941.69 H = 11
ASP:183.126.18 H = 27
( 1 ) 2/10 runs led to NaN loss function value.
( 2 ) 1/10 runs led to NaN loss function value.
( 3 ) 4/10 runs led to NaN loss function value.
BlogFeedback
229.143.83NaNNaNNaN32.510.0728.615877.94
529.1417.59NaNNaNNaN32.030.13t-lim
1027.0268.04NaNNaNNaN28.260.17t-lim
5027.02405.20NaNNaNNaN27.531.08t-lim
10027.021449.46NaNNaNNaN27.503.45t-lim
20027.025598.74NaNNaNNaN27.3718.83t-lim
500 NaNNaNNaN27.3796.81t-lim
5000 NaNNaNNaN27.374178.65t-lim
Best:27.0268.04 H = 10
ASP:27.0275.07 H = 12
ISOLET
24.821.715.320.719.250.42NaN13.010.062.833190.22
54.208.534.910.6811.010.43NaN10.530.09t-lim
103.8831.194.760.689.900.42NaN9.080.07t-lim
503.89128.684.690.749.750.45NaN6.550.54t-lim
1003.89409.634.660.789.650.46NaN6.001.33t-lim
2003.891953.214.660.879.040.50NaN4.953.54t-lim
500 4.831.209.230.66NaN4.346.97t-lim
5000 NaN8.222.90NaN4.09467.00t-lim
Best:3.8831.19 H = 10
ASP:3.8935.48 H = 14
CIFAR-10
22.8374.22NaNNaNNaN5.054.89t-lim
52.80345.60NaNNaNNaN3.185.10t-lim
102.78683.83NaNNaNNaN3.187.06t-lim
502.774845.04NaNNaNNaN2.9825.42t-lim
100t-limNaNNaNNaN2.9645.22t-lim
200t-limNaNNaNNaN2.9388.11t-lim
500 NaNNaNNaN2.92250.08t-lim
5000 NaNNaNNaNt-limt-lim
Best:2.77755.45 H = 12
ASP:2.77968.08 H = 17
Greenhouse gas observing network
225.075.0967.65 ( 1 ) 0.70362.120.58NaN75.040.18t-lim
523.5125.60NaN315.620.54NaN52.220.34t-lim
1023.1894.54NaN506.360.59NaN47.970.50t-lim
5023.18404.90NaN488.830.84NaN27.911.73t-lim
10023.181338.38NaN455.620.68NaN23.883.55t-lim
20023.186084.58NaN493.250.80NaN20.656.75t-lim
500 NaN576.691.55NaN17.397.30t-lim
5000 NaN1189.7110.83NaNfailt-lim
Best:23.1894.54 H = 10
ASP:23.18100.38 H = 12
( 1 ) 4/10 runs led to NaN loss function value.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Karmitsa, N.; Taheri, S.; Joki, K.; Paasivirta, P.; Bagirov, A.M.; Mäkelä, M.M. Nonsmooth Optimization-Based Hyperparameter-Free Neural Networks for Large-Scale Regression. Algorithms 2023, 16, 444. https://doi.org/10.3390/a16090444

AMA Style

Karmitsa N, Taheri S, Joki K, Paasivirta P, Bagirov AM, Mäkelä MM. Nonsmooth Optimization-Based Hyperparameter-Free Neural Networks for Large-Scale Regression. Algorithms. 2023; 16(9):444. https://doi.org/10.3390/a16090444

Chicago/Turabian Style

Karmitsa, Napsu, Sona Taheri, Kaisa Joki, Pauliina Paasivirta, Adil M. Bagirov, and Marko M. Mäkelä. 2023. "Nonsmooth Optimization-Based Hyperparameter-Free Neural Networks for Large-Scale Regression" Algorithms 16, no. 9: 444. https://doi.org/10.3390/a16090444

APA Style

Karmitsa, N., Taheri, S., Joki, K., Paasivirta, P., Bagirov, A. M., & Mäkelä, M. M. (2023). Nonsmooth Optimization-Based Hyperparameter-Free Neural Networks for Large-Scale Regression. Algorithms, 16(9), 444. https://doi.org/10.3390/a16090444

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop