Algorithms doi: 10.3390/a11020023

Authors: Hou-Ping Dai Dong-Dong Chen Zhou-Shun Zheng

Particle swarm optimization (PSO) algorithm is generally improved by adaptively adjusting the inertia weight or combining with other evolution algorithms. However, in most modified PSO algorithms, the random values are always generated by uniform distribution in the range of [0, 1]. In this study, the random values, which are generated by uniform distribution in the ranges of [0, 1] and [−1, 1], and Gauss distribution with mean 0 and variance 1 (U[0,1], U[−1,1] ,and G(0, 1)), are respectively used in the standard PSO and linear decreasing inertia weight (LDIW) PSO algorithms. For comparison, the deterministic PSO algorithm, in which the random values are set as 0.5, is also investigated in this study. Some benchmark functions and the pressure vessel design problem are selected to test these algorithms with different types of random values in three space dimensions (10, 30, and 100). The experimental results show that the standard PSO and LDIW-PSO algorithms with random values generated by U[−1,1] or G(0, 1) are more likely to avoid falling into local optima and quickly obtain the global optima. This is because the large-scale random values can expand the range of particle velocity to make the particle more likely to escape from local optima and obtain the global optima. Although the random values generated by U[−1,1] or are beneficial to improve the global searching ability, the local searching ability for a low dimensional practical optimization problem may be decreased due to the finite particles.

]]>Algorithms doi: 10.3390/a11020022

Authors: Chen Zhou Xinhui Liu Feixiang Xu

Two cylinders arranged symmetrically on a frame have become a major form of steering mechanism for articulated off-road vehicles (AORVs). However, the differences of stroke and arm lead to pressure fluctuation, vibration noise, and a waste of torque. In this paper, the differences of stroke and arm are reduced based on a genetic algorithm (GA). First, the mathematical model of the steering mechanism is put forward. Then, the difference of stroke and arm are optimized using a GA. Finally, a FW50GLwheel loader is used as an example to demonstrate the proposed GA-based optimization method, and its effectiveness is verified by means of automatic dynamic analysis of mechanical systems (ADAMS). The stroke difference of the steering hydraulic cylinders was reduced by 92% and the arm difference reached a decrease of 78% through GA optimization, in comparison with unoptimized structures. The simulation result shows that the steering mechanism optimized by GA behaved better than by previous methods.

]]>Algorithms doi: 10.3390/a11020021

Authors: Lijun Zhang Junyu Tao

Aiming at the pitting fault of deep groove ball bearing during service, this paper uses the vibration signal of five different states of deep groove ball bearing and extracts the relevant features, then uses a neural network to model the degradation for identifying and classifying the fault type. By comparing the effects of training samples with different capacities through performance indexes such as the accuracy and convergence speed, it is proven that an increase in the sample size can improve the performance of the model. Based on the polynomial fitting principle and Pearson correlation coefficient, fusion features based on the skewness index are proposed, and the performance improvement of the model after incorporating the fusion features is also validated. A comparison of the performance of the support vector machine (SVM) model and the neural network model on this dataset is given. The research shows that neural networks have more potential for complex and high-volume datasets.

]]>Algorithms doi: 10.3390/a11020019

Authors: Oliver Lemke Bettina Keller

Cluster analyses are often conducted with the goal to characterize an underlying probability density, for which the data-point density serves as an estimate for this probability density. We here test and benchmark the common nearest neighbor (CNN) cluster algorithm. This algorithm assigns a spherical neighborhood R to each data point and estimates the data-point density between two data points as the number of data points N in the overlapping region of their neighborhoods (step 1). The main principle in the CNN cluster algorithm is cluster growing. This grows the clusters by sequentially adding data points and thereby effectively positions the border of the clusters along an iso-surface of the underlying probability density. This yields a strict partitioning with outliers, for which the cluster represents peaks in the underlying probability density—termed core sets (step 2). The removal of the outliers on the basis of a threshold criterion is optional (step 3). The benchmark datasets address a series of typical challenges, including datasets with a very high dimensional state space and datasets in which the cluster centroids are aligned along an underlying structure (Birch sets). The performance of the CNN algorithm is evaluated with respect to these challenges. The results indicate that the CNN cluster algorithm can be useful in a wide range of settings. Cluster algorithms are particularly important for the analysis of molecular dynamics (MD) simulations. We demonstrate how the CNN cluster results can be used as a discretization of the molecular state space for the construction of a core-set model of the MD improving the accuracy compared to conventional full-partitioning models. The software for the CNN clustering is available on GitHub.

]]>Algorithms doi: 10.3390/a11020020

Authors: Amer Mouawad Naomi Nishimura Venkatesh Raman Sebastian Siebertz

In the Vertex Cover Reconfiguration (VCR) problem, given a graph G, positive integers k and ℓ and two vertex covers S and T of G of size at most k, we determine whether S can be transformed into T by a sequence of at most ℓ vertex additions or removals such that every operation results in a vertex cover of size at most k. Motivated by results establishing the W [ 1 ] -hardness of VCR when parameterized by ℓ, we delineate the complexity of the problem restricted to various graph classes. In particular, we show that VCR remains W [ 1 ] -hard on bipartite graphs, is NP -hard, but fixed-parameter tractable on (regular) graphs of bounded degree and more generally on nowhere dense graphs and is solvable in polynomial time on trees and (with some additional restrictions) on cactus graphs.

]]>Algorithms doi: 10.3390/a11020018

Authors: Hongliang Zhang Youcai Fang Ruilin Pan Chuanming Ge

To improve energy efficiency and maintain the stability of the power grid, time-of-use (TOU) electricity tariffs have been widely used around the world, which bring both opportunities and challenges to the energy-efficient scheduling problems. Single machine scheduling problems under TOU electricity tariffs are of great significance both in theory and practice. Although methods based on discrete-time or continuous-time models have been put forward for addressing this problem, they are deficient in solution quality or time complexity, especially when dealing with large-size instances. To address large-scale problems more efficiently, a new greedy insertion heuristic algorithm with a multi-stage filtering mechanism including coarse granularity and fine granularity filtering is developed in this paper. Based on the concentration and diffusion strategy, the algorithm can quickly filter out many impossible positions in the coarse granularity filtering stage, and then, each job can find its optimal position in a relatively large space in the fine granularity filtering stage. To show the effectiveness and computational process of the proposed algorithm, a real case study is provided. Furthermore, two sets of contrast experiments are conducted, aiming to demonstrate the good application of the algorithm. The experiments indicate that the small-size instances can be solved within 0.02 s using our algorithm, and the accuracy is further improved. For the large-size instances, the computation speed of our algorithm is improved greatly compared with the classic greedy insertion heuristic algorithm.

]]>Algorithms doi: 10.3390/a11020017

Authors: Xinen Lv Huiling Chen Qian Zhang Xujie Li Hui Huang Gang Wang

It is of great clinical significance to establish an accurate intelligent model to diagnose the somatization disorder of community correctional personnel. In this study, a novel machine learning framework is proposed to predict the severity of somatization disorder in community correction personnel. The core of this framework is to adopt the improved bacterial foraging optimization (IBFO) to optimize two key parameters (penalty coefficient and the kernel width) of a kernel extreme learning machine (KELM) and build an IBFO-based KELM (IBFO-KELM) for the diagnosis of somatization disorder patients. The main innovation point of the IBFO-KELM model is the introduction of opposition-based learning strategies in traditional bacteria foraging optimization, which increases the diversity of bacterial species, keeps a uniform distribution of individuals of initial population, and improves the convergence rate of the BFO optimization process as well as the probability of escaping from the local optimal solution. In order to verify the effectiveness of the method proposed in this study, a 10-fold cross-validation method based on data from a symptom self-assessment scale (SCL-90) is used to make comparison among IBFO-KELM, BFO-KELM (model based on the original bacterial foraging optimization model), GA-KELM (model based on genetic algorithm), PSO-KELM (model based on particle swarm optimization algorithm) and Grid-KELM (model based on grid search method). The experimental results show that the proposed IBFO-KELM prediction model has better performance than other methods in terms of classification accuracy, Matthews correlation coefficient (MCC), sensitivity and specificity. It can distinguish very well between severe somatization disorder and mild somatization and assist the psychological doctor with clinical diagnosis.

]]>Algorithms doi: 10.3390/a11020016

Authors: Liping Liu Ning Wang Zhigang Chen Lin Guo

Cognitive radio is a promising technology for improving spectrum utilization, which allows cognitive users access to the licensed spectrum while primary users are absent. In this paper, we design a resource allocation framework based on graph theory for spectrum assignment in cognitive radio networks. The framework takes into account the constraints that interference for primary users and possible collision among cognitive users. Based on the proposed model, we formulate a system utility function to maximize the system benefit. Based on the proposed model and objective problem, we design an improved ant colony optimization algorithm (IACO) from two aspects: first, we introduce differential evolution (DE) process to accelerate convergence speed by monitoring mechanism; then we design a variable neighborhood search (VNS) process to avoid the algorithm falling into the local optimal. Simulation results demonstrate that the improved algorithm achieves better performance.

]]>Algorithms doi: 10.3390/a11020015

Authors: Wei Nai Lu Liu Shaoyin Wang Decun Dong

Credit card holders from different age groups have different usage behaviors, so deeply investigating the credit card usage condition and properly modeling the usage trend of all customers in different age groups from time series data is meaningful for financial institutions as well as banks. Until now, related research in trend analysis of credit card usage has mostly been focused on specific group of people, such as the behavioral tendencies of the elderly or college students, or certain behaviors, such as the increasing number of cards owned and the rise in personal card debt or bankruptcy, in which the only analysis methods employed are simply enumerating or classifying raw data; thus, there is a lack of support in specific mathematical models based on usage behavioral time series data. Considering that few systematic modeling methods have been introduced, in this paper, a novel usage trend analysis method for credit card holders in different age groups based on singular spectrum analysis (SSA) has been proposed, using the time series data from the Survey of Consumer Payment Choice (SCPC). The decomposition and reconstruction process in the method is proposed. The results show that the credit card usage frequency falls down from the age of 26 to the lowest point at around the age of 58 and then begins to increase again. At last, future work is discussed.

]]>Algorithms doi: 10.3390/a11020014

Authors: Guangyuan Wu Zhigang Chen Lin Guo Jia Wu

In Body Area Networks (BANs), how to achieve energy management to extend the lifetime of the body area networks system is one of the most critical problems. In this paper, we design a body area network system powered by renewable energy, in which the sensors carried by patient with energy harvesting module can transmit data to a personal device. We do not require any a priori knowledge of the stochastic nature of energy harvesting and energy consumption. We formulate a user utility optimization problem. We use Lyapunov Optimization techniques to decompose the problem into three sub-problems, i.e., battery management, collecting rate control and transmission power allocation. We propose an online resource allocation algorithm to achieve two major goals: (1) balancing sensors’ energy harvesting and energy consumption while stabilizing the BANs system; and (2) maximizing the user utility. Performance analysis addresses required battery capacity, bounded data queue length and optimality of the proposed algorithm. Simulation results verify the optimization of algorithm.

]]>Algorithms doi: 10.3390/a11020013

Authors: Stefano Boldrini Luca De Nardis Giuseppe Caso Mai Le Jocelyn Fiorina Maria-Gabriella Di Benedetto

Multi-armed bandit (MAB) models are a viable approach to describe the problem of best wireless network selection by a multi-Radio Access Technology (multi-RAT) device, with the goal of maximizing the quality perceived by the final user. The classical MAB model does not allow, however, to properly describe the problem of wireless network selection by a multi-RAT device, in which a device typically performs a set of measurements in order to collect information on available networks, before a selection takes place. The MAB model foresees in fact only one possible action for the player, which is the selection of one among different arms at each time step; existing arm selection algorithms thus mainly differ in the rule according to which a specific arm is selected. This work proposes a new MAB model, named measure-use-MAB (muMAB), aiming at providing a higher flexibility, and thus a better accuracy in describing the network selection problem. The muMAB model extends the classical MAB model in a twofold manner; first, it foresees two different actions: to measure and to use; second, it allows actions to span over multiple time steps. Two new algorithms designed to take advantage of the higher flexibility provided by the muMAB model are also introduced. The first one, referred to as measure-use-UCB1 (muUCB1) is derived from the well known UCB1 algorithm, while the second one, referred to as Measure with Logarithmic Interval (MLI), is appositely designed for the new model so to take advantage of the new measure action, while aggressively using the best arm. The new algorithms are compared against existing ones from the literature in the context of the muMAB model, by means of computer simulations using both synthetic and captured data. Results show that the performance of the algorithms heavily depends on the Probability Density Function (PDF) of the reward received on each arm, with different algorithms leading to the best performance depending on the PDF. Results highlight, however, that as the ratio between the time required for using an arm and the time required to measure increases, the proposed algorithms guarantee the best performance, with muUCB1 emerging as the best candidate when the arms are characterized by similar mean rewards, and MLI prevailing when an arm is significantly more rewarding than others. This calls thus for the introduction of an adaptive approach capable of adjusting the behavior of the algorithm or of switching algorithm altogether, depending on the acquired knowledge on the PDF of the reward on each arm.

]]>Algorithms doi: 10.3390/a11020012

Authors: Farong Kou Jiafeng Du Zhe Wang Dong Li Jianan Xu

In order to coordinate the damping performance and energy regenerative performance of energy regenerative suspension, this paper proposes a structure of a vehicle semi-active energy regenerative suspension with an electro-hydraulic actuator (EHA). In light of the proposed concept, a specific energy regenerative scheme is designed and a mechanical properties test is carried out. Based on the test results, the parameter identification for the system model is conducted using a recursive least squares algorithm. On the basis of the system principle, the nonlinear model of the semi-active energy regenerative suspension with an EHA is built. Meanwhile, linear-quadratic-Gaussian control strategy of the system is designed. Then, the influence of the main parameters of the EHA on the damping performance and energy regenerative performance of the suspension is analyzed. Finally, the main parameters of the EHA are optimized via the genetic algorithm. The test results show that when a sinusoidal is input at the frequency of 2 Hz and the amplitude of 30 mm, the spring mass acceleration root meam square value of the optimized EHA semi-active energy regenerative suspension is reduced by 22.23% and the energy regenerative power RMS value is increased by 40.51%, which means that while meeting the requirements of vehicle ride comfort and driving safety, the energy regenerative performance is improved significantly.

]]>Algorithms doi: 10.3390/a11010010

Authors: Pasin Manurangsi

The Small Set Expansion Hypothesis is a conjecture which roughly states that it is NP-hard to distinguish between a graph with a small subset of vertices whose (edge) expansion is almost zero and one in which all small subsets of vertices have expansion almost one. In this work, we prove conditional inapproximability results with essentially optimal ratios for the following graph problems based on this hypothesis: Maximum Edge Biclique, Maximum Balanced Biclique, Minimum k-Cut and Densest At-Least-k-Subgraph. Our hardness results for the two biclique problems are proved by combining a technique developed by Raghavendra, Steurer and Tulsiani to avoid locality of gadget reductions with a generalization of Bansal and Khot’s long code test whereas our results for Minimum k-Cut and Densest At-Least-k-Subgraph are shown via elementary reductions.

]]>Algorithms doi: 10.3390/a11010011

Authors: Algorithms Editorial Office

Peer review is an essential part in the publication process, ensuring that Algorithms maintains high quality standards for its published papers.[...]

]]>Algorithms doi: 10.3390/a11010009

Authors: Wei Cui Qi Zhou Zhendong Zheng

Variation in the format and classification requirements for remote sensing data makes establishing a standard remote sensing sample dataset difficult. As a result, few remote sensing deep neural network models have been widely accepted. We propose a hybrid deep neural network model based on a convolutional auto-encoder and a complementary convolutional neural network to solve this problem. The convolutional auto-encoder supports feature extraction and data dimension reduction of remote sensing data. The extracted features are input into the convolutional neural network and subsequently classified. Experimental results show that in the proposed model, the classification accuracy increases from 0.916 to 0.944, compared to a traditional convolutional neural network model; furthermore, the number of training runs is reduced from 40,000 to 22,000, and the number of labelled samples can be reduced by more than half, all while ensuring a classification accuracy of no less than 0.9, which suggests the effectiveness and feasibility of the proposed model.

]]>Algorithms doi: 10.3390/a11010008

Authors: Yossi Peretz

In this article we suggest a randomized algorithm for the LQR (Linear Quadratic Regulator) optimal-control problem via static-output-feedback. The suggested algorithm is based on the recently introduced randomized optimization method called the Ray-Shooting Method that efficiently solves the global minimization problem of continuous functions over compact non-convex unconnected regions. The algorithm presented here is a randomized algorithm with a proof of convergence in probability. Its practical implementation has good performance in terms of the quality of controllers obtained and the percentage of success.

]]>Algorithms doi: 10.3390/a11010007

Authors: Lianchao Sheng Wei Li

In order to improve the control precision and robustness of the existing proportion integration differentiation (PID) controller of a 3-Revolute–Revolute–Revolute (3-RRR) parallel robot, a variable PID parameter controller optimized by a genetic algorithm controller is proposed in this paper. Firstly, the inverse kinematics model of the 3-RRR parallel robot was established according to the vector method, and the motor conversion matrix was deduced. Then, the error square integral was chosen as the fitness function, and the genetic algorithm controller was designed. Finally, the control precision of the new controller was verified through the simulation model of the 3-RRR planar parallel robot—built in SimMechanics—and the robustness of the new controller was verified by adding interference. The results show that compared with the traditional PID controller, the new controller designed in this paper has better control precision and robustness, which provides the basis for practical application.

]]>Algorithms doi: 10.3390/a11010006

Authors: Kaimeng Ding Shiping Chen Fan Meng

The perceptual hash algorithm is a technique to authenticate the integrity of images. While a few scholars have worked on mono-spectral image perceptual hashing, there is limited research on multispectral image perceptual hashing. In this paper, we propose a perceptual hash algorithm for the content authentication of a multispectral remote sensing image based on the synthetic characteristics of each band: firstly, the multispectral remote sensing image is preprocessed with band clustering and grid partition; secondly, the edge feature of the band subsets is extracted by band fusion-based edge feature extraction; thirdly, the perceptual feature of the same region of the band subsets is compressed and normalized to generate the perceptual hash value. The authentication procedure is achieved via the normalized Hamming distance between the perceptual hash value of the recomputed perceptual hash value and the original hash value. The experiments indicated that our proposed algorithm is robust compared to content-preserved operations and it efficiently authenticates the integrity of multispectral remote sensing images.

]]>Algorithms doi: 10.3390/a11010005

Authors: Xiyue Tang Yuhan Huang Guiwu Wei

In this paper, we investigate multiple-attribute decision-making (MADM) with Pythagorean 2-tuple linguistic numbers (P2TLNs). Then, we combine the weighted Bonferroni mean (WBM) operator and weighted geometric Bonferroni mean (WGBM) operator with P2TLNs to propose the Pythagorean 2-tuple linguistic WBM (P2TLWBM) operator and Pythagorean 2-tuple linguistic WGBM (P2TLWGBM) operator; MADM methods are then developed based on these two operators. Finally, a practical example for green supplier selection is given to verify the developed approach and to demonstrate its practicality and effectiveness.

]]>Algorithms doi: 10.3390/a11010004

Authors: Daoyu Lin Yang Wang Guangluan Xu Jun Li Kun Fu

Recently, inspired by the power of deep learning, convolution neural networks can produce fantastic images at the pixel level. However, a significant limiting factor for previous approaches is that they focus on some simple datasets such as faces and bedrooms. In this paper, we propose a multiscale deep neural network to transform sketches into Chinese paintings. To synthesize more realistic imagery, we train the generative network by using both L1 loss and adversarial loss. Additionally, users can control the process of the synthesis since the generative network is feed-forward. This network can also be treated as neural style transfer by adding an edge detector. Furthermore, additional experiments on image colorization and image super-resolution demonstrate the universality of our proposed approach.

]]>Algorithms doi: 10.3390/a11010003

Authors: Guillaume Filion

Seeding heuristics are the most widely used strategies to speed up sequence alignment in bioinformatics. Such strategies are most successful if they are calibrated, so that the speed-versus-accuracy trade-off can be properly tuned. In the widely used case of read mapping, it has been so far impossible to predict the success rate of competing seeding strategies for lack of a theoretical framework. Here, we present an approach to estimate such quantities based on the theory of analytic combinatorics. The strategy is to specify a combinatorial construction of reads where the seeding heuristic fails, translate this specification into a generating function using formal rules, and finally extract the probabilities of interest from the singularities of the generating function. The generating function can also be used to set up a simple recurrence to compute the probabilities with greater precision. We use this approach to construct simple estimators of the success rate of the seeding heuristic under different types of sequencing errors, and we show that the estimates are accurate in practical situations. More generally, this work shows novel strategies based on analytic combinatorics to compute probabilities of interest in bioinformatics.

]]>Algorithms doi: 10.3390/a11010002

Authors: Jie Wang Xiyue Tang Guiwu Wei

In this article, we expand the dual generalized weighted BM (DGWBM) and dual generalized weighted geometric Bonferroni mean (DGWGBM) operator with single valued neutrosophic numbers (SVNNs) to propose the dual generalized single-valued neutrosophic number WBM (DGSVNNWBM) operator and dual generalized single-valued neutrosophic numbers WGBM (DGSVNNWGBM) operator. Then, the multiple attribute decision making (MADM) methods are proposed with these operators. In the end, we utilize an applicable example for strategic suppliers selection to prove the proposed methods.

]]>Algorithms doi: 10.3390/a11010001

Authors: Anton Kasprzhitskii Georgy Lazorenko Victor Yavna

Investigation of the interaction of electromagnetic radiation with molecular systems provides most of the information on their structure and properties. Interpretation of experimental data is directly determined by the knowledge of the structure of energy levels and its change in the transition of these systems to an excited state. A key task of the methods for calculating the molecular orbitals of excited states is to accurately describe the emerging vacancies of the molecular core, leading to radial relaxation of the electron density. We propose an iterative scheme for solving a system of coupled integro-differential equations for obtaining molecular orbitals of electron configurations with excited/ionized deep and subvalent shells in a single-center representation. The numerical procedure of the iterative scheme is reduced to solving a boundary value problem based on a combination of the three-point difference scheme of Numerov and Thomas algorithm. To increase the rate of convergence of the computational procedure, an accurate account is taken of the behavior of the electron density near the nuclei of the molecular system. The realization of the algorithm of the computational scheme is considered on the example of a diatomic hydrogen fluoride molecule. The energy characteristics of the ground and ionized states of the molecule are estimated, and also the spatial distribution of the electron density is presented for the example of the σ-symmetry shell.

]]>Algorithms doi: 10.3390/a10040140

Authors: Ekaterina Auer Luise Senkel Stefan Kiel Andreas Rauh

In this paper, we take a look at the analysis and parameter identification for control-oriented, dynamic models for the thermal subsystem of solid oxide fuel cells (SOFC) from the systematized point of view of verification and validation (V&amp;V). First, we give a possible classification of models according to their verification degree which depends, for example, on the kind of arithmetic used for both formulation and simulation. Typical SOFC models, consisting of several coupled differential equations for gas preheaters and the temperature distribution in the stack module, do not have analytical solutions because of spatial nonlinearity. Therefore, in the next part of the paper, we describe in detail two possible ways to simplify such models so that the underlying differential equations can be solved analytically while still being sufficiently accurate to serve as the basis for control synthesis. The simplifying assumption is to approximate the heat capacities of the gases by zero-order polynomials (or first-oder polynomials, respectively) in the temperature. In the last, application-oriented part of the paper, we identify the parameters of these models as well as compare their performance and their ability to reflect the reality with the corresponding characteristics of models in which the heat capacities are represented by quadratic polynomials (the usual case). For this purpose, the framework UniVerMeC (Unified Framework for Verified GeoMetric Computations) is used, which allows us to employ different kinds of arithmetics including the interval one. This latter possibility ensures a high level of reliability of simulations and of the subsequent validation. Besides, it helps to take into account bounded uncertainty in measurements.

]]>Algorithms doi: 10.3390/a10040139

Authors: Wei Nai Lu Liu Shaoyin Wang Decun Dong

The ever-increasing air traffic demand in China has brought huge pressure on the planning and management of, and investment in, air terminals as well as airline companies. In this context, accurate and adequate short-term air traffic forecasting is essential for the operations of those entities. In consideration of such a problem, a hybrid air traffic forecasting model based on empirical mode decomposition (EMD) and seasonal auto regressive integrated moving average (SARIMA) has been proposed in this paper. The model proposed decomposes the original time series into components at first, and models each component with the SARIMA forecasting model, then integrates all the models together to form the final combined forecast result. By using the monthly air cargo and passenger flow data from the years 2006 to 2014 available at the official website of the Civil Aviation Administration of China (CAAC), the effectiveness in forecasting of the model proposed has been demonstrated, and by a horizontal performance comparison between several other widely used forecasting models, the advantage of the proposed model has also been proved.

]]>Algorithms doi: 10.3390/a10040138

Authors: Shou Feng Ping Fu Wenbin Zheng

Gene function prediction is a complicated and challenging hierarchical multi-label classification (HMC) task, in which genes may have many functions at the same time and these functions are organized in a hierarchy. This paper proposed a novel HMC algorithm for solving this problem based on the Gene Ontology (GO), the hierarchy of which is a directed acyclic graph (DAG) and is more difficult to tackle. In the proposed algorithm, the HMC task is firstly changed into a set of binary classification tasks. Then, two measures are implemented in the algorithm to enhance the HMC performance by considering the hierarchy structure during the learning procedures. Firstly, negative instances selecting policy associated with the SMOTE approach are proposed to alleviate the imbalanced data set problem. Secondly, a nodes interaction method is introduced to combine the results of binary classifiers. It can guarantee that the predictions are consistent with the hierarchy constraint. The experiments on eight benchmark yeast data sets annotated by the Gene Ontology show the promising performance of the proposed algorithm compared with other state-of-the-art algorithms.

]]>Algorithms doi: 10.3390/a10040137

Authors: Jorge Hernández-Gómez Carlos Couder-Castañeda Israel Herrera-Díaz Norberto Flores-Guzmán Enrique Gómez-Cruz

Numerical estimation of Lyapunov exponents in non-linear dynamical systems results in a very high computational cost. This is due to the large-scale computational cost of several Runge–Kutta problems that need to be calculated. In this work we introduce a parallel implementation based on MPI (Message Passing Interface) for the calculation of the Lyapunov exponents for a multidimensional dynamical system, considering a weakly coupled algorithm. Since we work on an academic high-latency cluster interconnected with a gigabit switch, the design has to be oriented to reduce the number of messages required. With the design introduced in this work, the computing time is drastically reduced, and the obtained performance leads to close to optimal speed-up ratios. The implemented parallelisation allows us to carry out many experiments for the calculation of several Lyapunov exponents with a low-cost cluster. The numerical experiments showed a high scalability, which we showed with up to 68 cores.

]]>Algorithms doi: 10.3390/a10040135

Authors: Sarah Berkemer Christian Höner zu Siederdissen Peter Stadler

Where string grammars describe how to generate and parse strings, tree grammars describe how to generate and parse trees. We show how to extend generalized algebraic dynamic programming to tree grammars. The resulting dynamic programming algorithms are efficient and provide the complete feature set available to string grammars, including automatic generation of outside parsers and algebra products for efficient backtracking. The complete parsing infrastructure is available as an embedded domain-specific language in Haskell. In addition to the formal framework, we provide implementations for both tree alignment and tree editing. Both algorithms are in active use in, among others, the area of bioinformatics, where optimization problems on trees are of considerable practical importance. This framework and the accompanying algorithms provide a beneficial starting point for developing complex grammars with tree- and forest-based inputs.

]]>Algorithms doi: 10.3390/a10040136

Authors: Yi Yang Chu Pan

The detection of composite miRNA functional module (CMFM) is of tremendous significance and helps in understanding the organization, regulation and execution of cell processes in cancer, but how to identify functional CMFMs is still a computational challenge. In this paper we propose a novel module detection method called MBCFM (detecting Composite Function Modules based on Maximal Biclique enumeration), specifically designed to bicluster miRNAs and target messenger RNAs (mRNAs) on the basis of multiple biological interaction information and topical network features. In this method, we employ algorithm MICA to enumerate all maximal bicliques and further extract R-pairs from the miRNA-mRNA regulatory network. Compared with two existing methods, Mirsynergy and SNMNMF on ovarian cancer dataset, the proposed method of MBCFM is not only able to extract cohesiveness-preserved CMFMs but also has high efficiency in running time. More importantly, MBCFM can be applied to detect other cancer-associated miRNA functional modules.

]]>Algorithms doi: 10.3390/a10040134

Authors: Shuai Zhang Xiao Qi

A novel stability analysis for the interval time-delay systems is proposed by employing a new series of integral inequalities for single and double integrals. Different from the recently introduced Wirtinger-based inequalities, refined Jensen inequalities and auxiliary function-based inequalities, the proposed ones can provide more accurate bounds for the cross terms in derivatives of the Lyapunov–Krasovskii functional (LKF) without involving additional slack variables. Based on the augmented LKF with triple-integral terms, their applications to stability analysis for interval time-delay systems are provided. By virtue of the newly derived inequalities, the resulting criteria are less conservative than some existing literature. Finally, numerical examples are provided to verify the effectiveness and improvement of the proposed approaches.

]]>Algorithms doi: 10.3390/a10040133

Authors: Jun Ye

A neutrosophic number (NN) presented by Smarandache can express determinate and/or indeterminate information in real life. NN (z = a + uI) consists of the determinate part a and the indeterminate part uI for a, u ∈ R (R is all real numbers) and indeterminacy I, and is very suitable for representing and handling problems with both determinate and indeterminate information. Based on the concept of NNs, this paper presents for first time the concepts of neutrosophic linear equations and the neutrosophic matrix, and introduces the neutrosophic matrix operations. Then, we propose some solving methods, including the substitution method, the addition method, and the inverse matrix method, for the system of neutrosophic linear equations or the neutrosophic matrix equation. Finally, an applied example about a traffic flow problem is provided to illustrate the application and effectiveness of handling the indeterminate traffic flow problem by using the system of neutrosophic linear equations.

]]>Algorithms doi: 10.3390/a10040132

Authors: Raffaele Pizzolante Bruno Carpentieri

Airborne and spaceborne hyperspectral sensors collect information which is derived from the electromagnetic spectrum of an observed area. Hyperspectral data are used in several studies and they are an important aid in different real-life applications (e.g., mining and geology applications, ecology, surveillance, etc.). A hyperspectral image has a three-dimensional structure (a sort of datacube): it can be considered as a sequence of narrow and contiguous spectral channels (bands). The objective of this paper is to present a framework permits the efficient storage/transmission of an input hyperspectral image, and its protection. The proposed framework relies on a reversible invisible watermarking scheme and an efficient lossless compression algorithm. The reversible watermarking scheme is used in conjunction with digital signature techniques in order to permit the verification of the integrity of a hyperspectral image by the receiver.

]]>Algorithms doi: 10.3390/a10040131

Authors: Su-Ting Chen Chuang Zhang Peng Li Yan-Yan Zhang Liang-Bao Jiao

The Approximate Point-In-Triangulation (APIT) localization algorithm is a widely used indoor positioning technology due to its simplicity and low power consumption. However, in practice, In-to-Out misjudgments exist regularly in APIT, and a considerable amount of nodes cannot be positioned due to the low node density. To tackle this issue, a Collaborative Coefficient-triangle APIT Localization (CCAL) algorithm is proposed. Firstly, an effective triangle criterion is put forward to reduce the probability of In-to-Out misjudgment and reduce the computational complexity. Then, a further Received Signal Strength Indicator (RSSI) location and weighted triangle coordinate calculation method is adopted to reduce the positioning error. Meanwhile, the idea of iterative collaborative positioning of the positioned unknown nodes is introduced to remarkably expand the localization coverage rate. Simulation results show that the proposed algorithm outperforms APIT, RSSI, and other improved algorithms in terms of both node location error and localization coverage rate.

]]>Algorithms doi: 10.3390/a10040130

Authors: Seyedeh Eftekharian Mohammad Shojafar Shahaboddin Shamshirband

Portfolio optimization is a serious challenge for financial engineering and has pulled down special attention among investors. It has two objectives: to maximize the reward that is calculated by expected return and to minimize the risk. Variance has been considered as a risk measure. There are many constraints in the world that ultimately lead to a non–convex search space such as cardinality constraint. In conclusion, parametric quadratic programming could not be applied and it seems essential to apply multi-objective evolutionary algorithm (MOEA). In this paper, a new efficient multi-objective portfolio optimization algorithm called 2-phase NSGA II algorithm is developed and the results of this algorithm are compared with the NSGA II algorithm. It was found that 2-phase NSGA II significantly outperformed NSGA II algorithm.

]]>Algorithms doi: 10.3390/a10040129

Authors: Albert Podusenko Vsevolod Nikulin Ivan Tanev Katsunori Shimohara

We investigate the feasibility of classifying (inferring) the emergency braking situations in road vehicles from the motion pattern of the accelerator pedal. We trained and compared several classifiers and employed genetic algorithms to tune their associated hyperparameters. Using offline time series data of the dynamics of the accelerator pedal as the test set, the experimental results suggest that the evolved classifiers detect the emergency braking situation with at least 93% accuracy. The best performing classifier could be integrated into the agent that perceives the dynamics of the accelerator pedal in real time and—if emergency braking is detected—acts by applying full brakes well before the driver would have been able to apply them.

]]>Algorithms doi: 10.3390/a10040128

Authors: Feng Du Qiao-Yue Dong Hong-Shuang Li

This paper presents a global optimization method for structural design optimization, which integrates subset simulation optimization (SSO) and the dynamic augmented Lagrangian multiplier method (DALMM). The proposed method formulates the structural design optimization as a series of unconstrained optimization sub-problems using DALMM and makes use of SSO to find the global optimum. The combined strategy guarantees that the proposed method can automatically detect active constraints and provide global optimal solutions with finite penalty parameters. The accuracy and robustness of the proposed method are demonstrated by four classical truss sizing problems. The results are compared with those reported in the literature, and show a remarkable statistical performance based on 30 independent runs.

]]>Algorithms doi: 10.3390/a10040126

Authors: Amrapali Zaveri Gökhan Ertaylan

Massive amounts of data are currently available and being produced at an unprecedented rate in all domains of life sciences worldwide. However, this data is disparately stored and is in different and unstructured formats making it very hard to integrate. In this review, we examine the state of the art and propose the use of the Linked Data (LD) paradigm, which is a set of best practices for publishing and connecting structured data on the Web in a semantically meaningful format. We argue that utilizing LD in the life sciences will make data sets better Findable, Accessible, Interoperable, and Reusable. We identify three tiers of the research cycle in life sciences, namely (i) systematic review of the existing body of knowledge, (ii) meta-analysis of data, and (iii) knowledge discovery of novel links across different evidence streams to primarily utilize the proposed LD paradigm. Finally, we demonstrate the use of LD in three use case scenarios along the same research question and discuss the future of data/knowledge integration in life sciences and the challenges ahead.

]]>Algorithms doi: 10.3390/a10040127

Authors: Jianming Zhang Manting Huang Xiaokang Jin Xudong Li

Traffic sign detection is an important task in traffic sign recognition systems. Chinese traffic signs have their unique features compared with traffic signs of other countries. Convolutional neural networks (CNNs) have achieved a breakthrough in computer vision tasks and made great success in traffic sign classification. In this paper, we present a Chinese traffic sign detection algorithm based on a deep convolutional network. To achieve real-time Chinese traffic sign detection, we propose an end-to-end convolutional network inspired by YOLOv2. In view of the characteristics of traffic signs, we take the multiple 1 × 1 convolutional layers in intermediate layers of the network and decrease the convolutional layers in top layers to reduce the computational complexity. For effectively detecting small traffic signs, we divide the input images into dense grids to obtain finer feature maps. Moreover, we expand the Chinese traffic sign dataset (CTSD) and improve the marker information, which is available online. All experimental results evaluated according to our expanded CTSD and German Traffic Sign Detection Benchmark (GTSDB) indicate that the proposed method is the faster and more robust. The fastest detection speed achieved was 0.017 s per image.

]]>Algorithms doi: 10.3390/a10040125

Authors: Jian Tang Kejun Zhu Haixiang Guo Can Liao Shuwen Zhang

In this paper, we optimize the search and rescue (SAR) in disaster relief through agent-based simulation. We simulate rescue teams’ search behaviors with the improved Truncated Lévy walks. Then we propose a cooperative rescue plan based on a distributed auction mechanism, and illustrate it with the case of landslide disaster relief. The simulation is conducted in three scenarios, including “fatal”, “serious” and “normal”. Compared with the non-cooperative rescue plan, the proposed rescue plan in this paper would increase victims’ relative survival probability by 7–15%, increase the ratio of survivors getting rescued by 5.3–12.9%, and decrease the average elapsed time for one site getting rescued by 16.6–21.6%. The robustness analysis shows that search radius can affect the rescue efficiency significantly, while the scope of cooperation cannot. The sensitivity analysis shows that the two parameters, the time limit for completing rescue operations in one buried site and the maximum turning angle for next step, both have a great influence on rescue efficiency, and there exists optimal value for both of them in view of rescue efficiency.

]]>Algorithms doi: 10.3390/a10040124

Authors: Yingying Wang Yibin Li Yong Song Xuewen Rong Shuaishuai Zhang

The decision tree algorithm is a core technology in data classification mining, and ID3 (Iterative Dichotomiser 3) algorithm is a famous one, which has achieved good results in the field of classification mining. Nevertheless, there exist some disadvantages of ID3 such as attributes biasing multi-values, high complexity, large scales, etc. In this paper, an improved ID3 algorithm is proposed that combines the simplified information entropy based on different weights with coordination degree in rough set theory. The traditional ID3 algorithm and the proposed one are fairly compared by using three common data samples as well as the decision tree classifiers. It is shown that the proposed algorithm has a better performance in the running time and tree structure, but not in accuracy than the ID3 algorithm, for the first two sample sets, which are small. For the third sample set that is large, the proposed algorithm improves the ID3 algorithm for all of the running time, tree structure and accuracy. The experimental results show that the proposed algorithm is effective and viable.

]]>Algorithms doi: 10.3390/a10040123

Authors: Adam Chehouri Rafic Younes Jihan Khoder Jean Perron Adrian Ilinca

This article presents a newly proposed selection process for genetic algorithms on a class of unconstrained optimization problems. The k-means genetic algorithm selection process (KGA) is composed of four essential stages: clustering, membership phase, fitness scaling and selection. Inspired from the hypothesis that clustering the population helps to preserve a selection pressure throughout the evolution of the population, a membership probability index is assigned to each individual following the clustering phase. Fitness scaling converts the membership scores in a range suitable for the selection function which selects the parents of the next generation. Two versions of the KGA process are presented: using a fixed number of clusters K (KGAf) and via an optimal partitioning Kopt (KGAo) determined by two different internal validity indices. The performance of each method is tested on seven benchmark problems.

]]>Algorithms doi: 10.3390/a10040121

Authors: Laxmi Bewoor V. Chandra Prakash Sagar Sapkal

The no-wait flow shop is a flowshop in which the scheduling of jobs is continuous and simultaneous through all machines without waiting for any consecutive machines. The scheduling of a no-wait flow shop requires finding an appropriate sequence of jobs for scheduling, which in turn reduces total processing time. The classical brute force method for finding the probabilities of scheduling for improving the utilization of resources may become trapped in local optima, and this problem can hence be observed as a typical NP-hard combinatorial optimization problem that requires finding a near optimal solution with heuristic and metaheuristic techniques. This paper proposes an effective hybrid Particle Swarm Optimization (PSO) metaheuristic algorithm for solving no-wait flow shop scheduling problems with the objective of minimizing the total flow time of jobs. This Proposed Hybrid Particle Swarm Optimization (PHPSO) algorithm presents a solution by the random key representation rule for converting the continuous position information values of particles to a discrete job permutation. The proposed algorithm initializes population efficiently with the Nawaz-Enscore-Ham (NEH) heuristic technique and uses an evolutionary search guided by the mechanism of PSO, as well as simulated annealing based on a local neighborhood search to avoid getting stuck in local optima and to provide the appropriate balance of global exploration and local exploitation. Extensive computational experiments are carried out based on Taillard’s benchmark suite. Computational results and comparisons with existing metaheuristics show that the PHPSO algorithm outperforms the existing methods in terms of quality search and robustness for the problem considered. The improvement in solution quality is confirmed by statistical tests of significance.

]]>Algorithms doi: 10.3390/a10040122

Authors: Sean Yaw Brendan Mumey

This paper examines an important problem in smart grid energy scheduling; peaks in power demand are proportionally more expensive to generate and provision for. The issue is exacerbated in local microgrids that do not benefit from the aggregate smoothing experienced by large grids. Demand-side scheduling can reduce these peaks by taking advantage of the fact that there is often flexibility in job start times. We focus attention on the case where the jobs are non-preemptible, meaning once started, they run to completion. The associated optimization problem is called the peak demand minimization problem, and has been previously shown to be NP-hard. Our results include an optimal fixed-parameter tractable algorithm, a polynomial-time approximation algorithm, as well as an effective heuristic that can also be used in an online setting of the problem. Simulation results show that these methods can reduce peak demand by up to 50% versus on-demand scheduling for household power jobs.

]]>Algorithms doi: 10.3390/a10040119

Authors: Zijian Cao Lei Wang

In this paper, we propose a novel nature-inspired meta-heuristic algorithm for continuous global optimization, named the phase transition-based optimization algorithm (PTBO). It mimics three completely different kinds of motion characteristics of elements in three different phases, which are the unstable phase, the meta-stable phase, and the stable phase. Three corresponding operators, which are the stochastic operator of the unstable phase, the shrinkage operator in the meta-stable phase, and the vibration operator of the stable phase, are designed in the proposed algorithm. In PTBO, the three different phases of elements dynamically execute different search tasks according to their phase in each generation. It makes it such that PTBO not only has a wide range of exploration capabilities, but also has the ability to quickly exploit them. Numerical experiments are carried out on twenty-eight functions of the CEC 2013 benchmark suite. The simulation results demonstrate its better performance compared with that of other state-of-the-art optimization algorithms.

]]>Algorithms doi: 10.3390/a10040120

Authors: Abdulbaset Saad Zuomin Dong Meysam Karimi

Advanced global optimization algorithms have been continuously introduced and improved to solve various complex design optimization problems for which the objective and constraint functions can only be evaluated through computation intensive numerical analyses or simulations with a large number of design variables. The often implicit, multimodal, and ill-shaped objective and constraint functions in high-dimensional and “black-box” forms demand the search to be carried out using low number of function evaluations with high search efficiency and good robustness. This work investigates the performance of six recently introduced, nature-inspired global optimization methods: Artificial Bee Colony (ABC), Firefly Algorithm (FFA), Cuckoo Search (CS), Bat Algorithm (BA), Flower Pollination Algorithm (FPA) and Grey Wolf Optimizer (GWO). These approaches are compared in terms of search efficiency and robustness in solving a set of representative benchmark problems in smooth-unimodal, non-smooth unimodal, smooth multimodal, and non-smooth multimodal function forms. In addition, four classic engineering optimization examples and a real-life complex mechanical system design optimization problem, floating offshore wind turbines design optimization, are used as additional test cases representing computationally-expensive black-box global optimization problems. Results from this comparative study show that the ability of these global optimization methods to obtain a good solution diminishes as the dimension of the problem, or number of design variables increases. Although none of these methods is universally capable, the study finds that GWO and ABC are more efficient on average than the other four in obtaining high quality solutions efficiently and consistently, solving 86% and 80% of the tested benchmark problems, respectively. The research contributes to future improvements of global optimization methods.

]]>Algorithms doi: 10.3390/a10040118

Authors: Siyu Liu Ling Xu Feng Ding

This paper focuses on the iterative parameter estimation algorithms for dual-frequency signal models that are disturbed by stochastic noise. The key of the work is to overcome the difficulty that the signal model is a highly nonlinear function with respect to frequencies. A gradient-based iterative (GI) algorithm is presented based on the gradient search. In order to improve the estimation accuracy of the GI algorithm, a Newton iterative algorithm and a moving data window gradient-based iterative algorithm are proposed based on the moving data window technique. Comparative simulation results are provided to illustrate the effectiveness of the proposed approaches for estimating the parameters of signal models.

]]>Algorithms doi: 10.3390/a10040117

Authors: Babar Khan Zhijie Wang Fang Han Ather Iqbal Rana Masood

Usually, a fabric weave pattern is recognized using methods which identify the warp floats and weft floats. Although these methods perform well for uniform or repetitive weave patterns, in the case of complex weave patterns, these methods become computationally complex and the classification error rates are comparatively higher. Furthermore, the fault-tolerance (invariance) and stability (selectivity) of the existing methods are still to be enhanced. We present a novel biologically-inspired method to invariantly recognize the fabric weave pattern (fabric texture) and yarn color from the color image input. We proposed a model in which the fabric weave pattern descriptor is based on the HMAX model for computer vision inspired by the hierarchy in the visual cortex, the color descriptor is based on the opponent color channel inspired by the classical opponent color theory of human vision, and the classification stage is composed of a multi-layer (deep) extreme learning machine. Since the weave pattern descriptor, yarn color descriptor, and the classification stage are all biologically inspired, we propose a method which is completely biologically plausible. The classification performance of the proposed algorithm indicates that the biologically-inspired computer-aided-vision models might provide accurate, fast, reliable and cost-effective solution to industrial automation.

]]>Algorithms doi: 10.3390/a10040116

Authors: Liangliang Li Yujuan Si Zhenhong Jia

In this paper, a novel remote sensing image enhancement technique based on a non-local means filter in a nonsubsampled contourlet transform (NSCT) domain is proposed. The overall flow of the approach can be divided into the following steps: Firstly, the image is decomposed into one low-frequency sub-band and several high-frequency sub-bands with NSCT. Secondly, contrast stretching is adopted to deal with the low-frequency sub-band coefficients, and the non-local means filter is applied to suppress the noise contained in the first high-frequency sub-band coefficients. Thirdly, the processed coefficients are reconstructed with the inverse NSCT transform. Finally, the unsharp filter is used to enhance the details of the image. The simulation results show that the proposed algorithm has better performance in remote sensing image enhancement than the existing approaches.

]]>Algorithms doi: 10.3390/a10040115

Authors: Huanqing Cui Jian Niu Chuanai Zhou Minglei Shu

A graph is a very important structure to describe many applications in the real world. In many applications, such as dependency graphs and debt graphs, it is an important problem to find and remove cycles to make these graphs be cycle-free. The common algorithm often leads to an out-of-memory exception in commodity personal computer, and it cannot leverage the advantage of multicore computers. This paper introduces a new problem, cycle detection and removal with vertex priority. It proposes a multithreading iterative algorithm to solve this problem for large-scale graphs on personal computers. The algorithm includes three main steps: simplification to decrease the scale of graph, calculation of strongly connected components, and cycle detection and removal according to a pre-defined priority in parallel. This algorithm avoids the out-of-memory exception by simplification and iteration, and it leverages the advantage of multicore computers by multithreading parallelism. Five different versions of the proposed algorithm are compared by experiments, and the results show that the parallel iterative algorithm outperforms the others, and simplification can effectively improve the algorithm's performance.

]]>Algorithms doi: 10.3390/a10040114

Authors: Hristos Tyralis Georgia Papacharalampous

Time series forecasting using machine learning algorithms has gained popularity recently. Random forest is a machine learning algorithm implemented in time series forecasting; however, most of its forecasting properties have remained unexplored. Here we focus on assessing the performance of random forests in one-step forecasting using two large datasets of short time series with the aim to suggest an optimal set of predictor variables. Furthermore, we compare its performance to benchmarking methods. The first dataset is composed by 16,000 simulated time series from a variety of Autoregressive Fractionally Integrated Moving Average (ARFIMA) models. The second dataset consists of 135 mean annual temperature time series. The highest predictive performance of RF is observed when using a low number of recent lagged predictor variables. This outcome could be useful in relevant future applications, with the prospect to achieve higher predictive accuracy.

]]>Algorithms doi: 10.3390/a10040113

Authors: Shahram Shahinpour Shirin Shirvani Zeynep Ertem Sergiy Butenko

Given a simple, undirected graph G, a biclique is a subset of vertices inducing a complete bipartite subgraph in G. In this paper, we consider two associated optimization problems, the maximum biclique problem, which asks for a biclique of the maximum cardinality in the graph, and the maximum edge biclique problem, aiming to find a biclique with the maximum number of edges in the graph. These NP-hard problems find applications in biclustering-type tasks arising in complex network analysis. Real-life instances of these problems often involve massive, but sparse networks. We develop exact approaches for detecting optimal bicliques in large-scale graphs that combine effective scale reduction techniques with integer programming methodology. Results of computational experiments with numerous real-life network instances demonstrate the performance of the proposed approach.

]]>Algorithms doi: 10.3390/a10040111

Authors: Guanci Yang

Many approaches that model specific intelligent behaviors perform excellently in solving complex optimization problems. Game theory is widely recognized as an important tool in many fields. This paper introduces a game theory-inspired evolutionary algorithm for global optimization (GameEA). A formulation to estimate payoff expectations is provided, which is a mechanism to make a player become a rational decision-maker. GameEA has one population (i.e., set of players) and generates new offspring only through an imitation operator and a belief-learning operator. An imitation operator adopts learning strategies and actions from other players to improve its competitiveness and applies these strategies to future games where one player updates its chromosome by strategically copying segments of gene sequences from a competitor. Belief learning refers to models in which a player adjusts his/her strategies, behavior or chromosomes by analyzing the current history information to improve solution quality. Experimental results on various classes of problems show that GameEA outperforms the other four algorithms on stability, robustness, and accuracy.

]]>Algorithms doi: 10.3390/a10040112

Authors: Daniel Edler Ludvig Bohlin and Rosvall

Comprehending complex systems by simplifying and highlighting important dynamical patterns requires modeling and mapping higher-order network flows. However, complex systems come in many forms and demand a range of representations, including memory and multilayer networks, which in turn call for versatile community-detection algorithms to reveal important modular regularities in the flows. Here we show that various forms of higher-order network flows can be represented in a unified way with networks that distinguish physical nodes for representing a complex system’s objects from state nodes for describing flows between the objects. Moreover, these so-called sparse memory networks allow the information-theoretic community detection method known as the map equation to identify overlapping and nested flow modules in data from a range of different higher-order interactions such as multistep, multi-source, and temporal data. We derive the map equation applied to sparse memory networks and describe its search algorithm Infomap, which can exploit the flexibility of sparse memory networks. Together they provide a general solution to reveal overlapping modular patterns in higher-order flows through complex systems.

]]>Algorithms doi: 10.3390/a10040109

Authors: Fatemeh Salehi Rizi Michael Granitzer

Embedding social network data into a low-dimensional vector space has shown promising performance for many real-world applications, such as node classification, node clustering, link prediction and network visualization. However, the information contained in these vector embeddings remains abstract and hard to interpret. Methods for inspecting embeddings usually rely on visualization methods, which do not work on a larger scale and do not give concrete interpretations of vector embeddings in terms of preserved network properties (e.g., centrality or betweenness measures). In this paper, we study and investigate network properties preserved by recent random walk-based embedding procedures like node2vec, DeepWalk or LINE. We propose a method that applies learning to rank in order to relate embeddings to network centralities. We evaluate our approach with extensive experiments on real-world and artificial social networks. Experiments show that each embedding method learns different network properties. In addition, we show that our graph embeddings in combination with neural networks provide a computationally efficient way to approximate the Closeness Centrality measure in social networks.

]]>Algorithms doi: 10.3390/a10040110

Authors: Rana Masood Dao Wang Zain Ali Babar Khan

In this article, a novel hybrid control scheme is proposed for controlling the position of a three-phase brushless direct current (BLDC) motor. The hybrid controller consists of discrete time sliding mode control (SMC) with model free adaptive control (MFAC) to make a new data-driven control (DDC) strategy that is able to reduce the simulation time and complexity of a nonlinear system. The proposed hybrid algorithm is also suitable for controlling the speed variations of a BLDC motor, and is also applicable for the real time simulation of platforms such as a gimbal platform. The DDC method does not require any system model because it depends on data collected by the system about its Inputs/Outputs (IOS). However, the model-based control (MBC) method is difficult to apply from a practical point of view and is time-consuming because we need to linearize the system model. The above proposed method is verified by multiple simulations using MATLAB Simulink. It shows that the proposed controller has better performance, more precise tracking, and greater robustness compared with the classical proportional integral derivative (PID) controller, MFAC, and model free learning adaptive control (MFLAC).

]]>Algorithms doi: 10.3390/a10030106

Authors: Krassimir Atanassov

A comparison between type-1 fuzzy sets (T1FSs) and intuitionistic fuzzy sets (IFSs) is made. The operators defined over IFSs that do not have analogues in T1FSs are shown, and such analogues are introduced whenever possible.

]]>Algorithms doi: 10.3390/a10030107

Authors: Shifeng Chen Rong Chen Jian Gao

The dynamic vehicle routing problem (DVRP) is a variant of the Vehicle Routing Problem (VRP) in which customers appear dynamically. The objective is to determine a set of routes that minimizes the total travel distance. In this paper, we propose a monarch butterfly optimization (MBO) algorithm to solve DVRPs, utilizing a greedy strategy. Both migration operation and the butterfly adjusting operator only accept the offspring of butterfly individuals that have better fitness than their parents. To improve performance, a later perturbation procedure is implemented, to maintain a balance between global diversification and local intensification. The computational results indicate that the proposed technique outperforms the existing approaches in the literature for average performance by at least 9.38%. In addition, 12 new best solutions were found. This shows that this proposed technique consistently produces high-quality solutions and outperforms other published heuristics for the DVRP.

]]>Algorithms doi: 10.3390/a10030108

Authors: Deyun Wang Chenqiang Yue Shuai Wei Jun Lv

Agricultural commodity futures prices play a significant role in the change tendency of these spot prices and the supply–demand relationship of global agricultural product markets. Due to the nonlinear and nonstationary nature of this kind of time series data, it is inevitable for price forecasting research to take this nature into consideration. Therefore, we aim to enrich the existing research literature and offer a new way of thinking about forecasting agricultural commodity futures prices, so that four hybrid models are proposed based on the back propagation neural network (BPNN) optimized by the particle swarm optimization (PSO) algorithm and four decomposition methods: empirical mode decomposition (EMD), wavelet packet transform (WPT), intrinsic time-scale decomposition (ITD) and variational mode decomposition (VMD). In order to verify the applicability and validity of these hybrid models, we select three futures prices of wheat, corn and soybean to conduct the experiment. The experimental results show that (1) all the hybrid models combined with decomposition technique have a better performance than the single PSO–BPNN model; (2) VMD contributes the most in improving the forecasting ability of the PSO–BPNN model, while WPT ranks second; (3) ITD performs better than EMD in both cases of corn and soybean; and (4) the proposed models perform well in the forecasting of agricultural commodity futures prices.

]]>Algorithms doi: 10.3390/a10030105

Authors: Joonas Hämäläinen Susanne Jauhiainen Tommi Kärkkäinen

Clustering is an unsupervised machine learning and pattern recognition method. In general, in addition to revealing hidden groups of similar observations and clusters, their number needs to be determined. Internal clustering validation indices estimate this number without any external information. The purpose of this article is to evaluate, empirically, characteristics of a representative set of internal clustering validation indices with many datasets. The prototype-based clustering framework includes multiple, classical and robust, statistical estimates of cluster location so that the overall setting of the paper is novel. General observations on the quality of validation indices and on the behavior of different variants of clustering algorithms will be given.

]]>Algorithms doi: 10.3390/a10030104

Authors: Nan Zhao Menglin Fan Chao Tian Pengfei Fan

Mobile crowdsourcing networks (MCNs) are a promising method of data collecting and processing by leveraging the mobile devices’ sensing and computing capabilities. However, because of the selfish characteristics of the service provider (SP) and mobile users (MUs), crowdsourcing participants only aim to maximize their own benefits. This paper investigates the incentive mechanism between the above two parties to create mutual benefits. By modeling MCNs as a labor market, a contract-based crowdsourcing model with moral hazard is proposed under the asymmetric information scenario. In order to incentivize the potential MUs to participate in crowdsourcing tasks, the optimization problem is formulated to maximize the SP’s utility by jointly examining the crowdsourcing participants’ risk preferences. The impact of crowdsourcing participants’ attitudes of risks on the incentive mechanism has been studied analytically and experimentally. Numerical simulation results demonstrate the effectiveness of the proposed contract design scheme for the crowdsourcing incentive.

]]>Algorithms doi: 10.3390/a10030103

Authors: Jingbo Zhang Henan Yu Shufang Zhang

The recent development of maritime transport has resulted in the demand for a wider communication bandwidth being more intense. Cognitive radios can dynamically manage resources in a spectrum. Thus, building a new type of maritime cognitive radio communication system (MCRCS) is an effective solution. In this paper, the enhanced dynamic spectrum allocation algorithm (EDSAA) is proposed, which is based on the Cournot game model. In EDSAA, the decision-making center (DC) sets the weights according to the detection capability of the secondary user (SU), before adding these weighting coefficients in the price function. Furthermore, the willingness of the SU will reduce after meeting their basic communication needs when it continues to increase the leasable spectrum by adding the elastic model in the SU’s revenue function. On this basis, the profit function is established. The simulation results show that the EDSAA has Nash equilibrium and conforms to the actual situation. It shows that the results of spectrum allocation are fair, efficient and reasonable.

]]>Algorithms doi: 10.3390/a10030102

Authors: Eisha Nathan Anita Zakrzewska Jason Riedy David Bader

Analyzing massive graphs poses challenges due to the vast amount of data available. Extracting smaller relevant subgraphs allows for further visualization and analysis that would otherwise be too computationally intensive. Furthermore, many real data sets are constantly changing, and require algorithms to update as the graph evolves. This work addresses the topic of local community detection, or seed set expansion, using personalized centrality measures, specifically PageRank and Katz centrality. We present a method to efficiently update local communities in dynamic graphs. By updating the personalized ranking vectors, we can incrementally update the corresponding local community. Applying our methods to real-world graphs, we are able to obtain speedups of up to 60× compared to static recomputation while maintaining an average recall of 0.94 of the highly ranked vertices returned. Next, we investigate how approximations of a centrality vector affect the resulting local community. Specifically, our method guarantees that the vertices returned in the community are the highly ranked vertices from a personalized centrality metric.

]]>Algorithms doi: 10.3390/a10030101

Authors: Frumen Olivas Leticia Amador-Angulo Jonathan Perez Camilo Caraveo Fevrier Valdez Oscar Castillo

In this paper, a comparison among Particle swarm optimization (PSO), Bee Colony Optimization (BCO) and the Bat Algorithm (BA) is presented. In addition, a modification to the main parameters of each algorithm through an interval type-2 fuzzy logic system is presented. The main aim of using interval type-2 fuzzy systems is providing dynamic parameter adaptation to the algorithms. These algorithms (original and modified versions) are compared with the design of fuzzy systems used for controlling the trajectory of an autonomous mobile robot. Simulation results reveal that PSO algorithm outperforms the results of the BCO and BA algorithms.

]]>Algorithms doi: 10.3390/a10030099

Authors: Mauricio Sanchez Juan Castro Violeta Ocegueda-Miramontes Leticia Cervantes

This work is focused on creating fuzzy granular classification models based on general type-2 fuzzy logic systems when consequents are represented by interval type-2 TSK linear functions. Due to the complexity of general type-2 TSK fuzzy logic systems, a hybrid learning approach is proposed, where the principle of justifiable granularity is heuristically used to define an amount of uncertainty in the system, which in turn is used to define the parameters in the interval type-2 TSK linear functions via a dual LSE algorithm. Multiple classification benchmark datasets were tested in order to assess the quality of the formed granular models; its performance is also compared against other common classification algorithms. Shown results conclude that classification performance in general is better than results obtained by other techniques, and in general, all achieved results, when averaged, have a better performance rate than compared techniques, demonstrating the stability of the proposed hybrid learning technique.

]]>Algorithms doi: 10.3390/a10030100

Authors: Tao Ye Ziqiang Yang Siling Feng

The portfolio optimization problem is the central problem of modern economics and decision theory; there is the Mean-Variance Model and Stochastic Dominance Model for solving this problem. In this paper, based on the second order stochastic dominance constraints, we propose the improved biogeography-based optimization algorithm to optimize the portfolio, which we called ε BBO. In order to test the computing power of ε BBO, we carry out two numerical experiments in several kinds of constraints. In experiment 1, comparing the Stochastic Approximation (SA) method with the Level Function (LF) algorithm and Genetic Algorithm (GA), we get a similar optimal solution by ε BBO in [ 0 , 0 . 6 ] and [ 0 , 1 ] constraints with the return of 1.174% and 1.178%. In [ - 1 , 2 ] constraint, we get the optimal return of 1.3043% by ε BBO, while the return of SA and LF is 1.23% and 1.26%. In experiment 2, we get the optimal return of 0.1325% and 0.3197% by ε BBO in [ 0 , 0 . 1 ] and [ - 0 . 05 , 0 . 15 ] constraints. As a comparison, the return of FTSE100 Index portfolio is 0.0937%. The results prove that ε BBO algorithm has great potential in the field of financial decision-making, it also shows that ε BBO algorithm has a better performance in optimization problem.

]]>Algorithms doi: 10.3390/a10030098

Authors: Fei Chen Xiaohong Bi Ruimin Lyu Zhongwei Hua Yuan Liu Xiaoting Zhang

Over the past decades, vehicular ad hoc networks (VANETs) have been a core networking technology to provide drivers and passengers with safety and convenience. As a new emerging technology, the vehicular cloud computing (VCC) can provide cloud services for various data-intensive applications in VANETs, such as multimedia streaming. However, the vehicle mobility and intermittent connectivity present challenges to the large-scale data dissemination with underlying computing and networking architecture. In this paper, we will explore the service scheduling of virtual RSUs for diverse request demands in the dynamic traffic flow in vehicular cloud environment. Specifically, we formulate the RSU allocation problem as maximum service capacity with multiple-source and multiple-destination, and propose a bidirectional RSU allocation strategy. In addition, we formulate the content replication in distributed RSUs as the minimum replication set coverage problem in a two-layer mapping model, and analyze the solutions in different scenarios. Numerical results further prove the superiority of our proposed solution, as well as the scalability to various traffic condition variations.

]]>Algorithms doi: 10.3390/a10030097

Authors: Seifemichael B. Amsalu Abdollah Homaifar Albert Esterline

In this paper, a new algorithm for sensitivity analysis of discrete hidden Markov models (HMMs) is proposed. Sensitivity analysis is a general technique for investigating the robustness of the output of a system model. Sensitivity analysis of probabilistic networks has recently been studied extensively. This has resulted in the development of mathematical relations between a parameter and an output probability of interest and also methods for establishing the effects of parameter variations on decisions. Sensitivity analysis in HMMs has usually been performed by taking small perturbations in parameter values and re-computing the output probability of interest. As recent studies show, the sensitivity analysis of an HMM can be performed using a functional relationship that describes how an output probability varies as the network’s parameters of interest change. To derive this sensitivity function, existing Bayesian network algorithms have been employed for HMMs. These algorithms are computationally inefficient as the length of the observation sequence and the number of parameters increases. In this study, a simplified efficient matrix-based algorithm for computing the coefficients of the sensitivity function for all hidden states and all time steps is proposed and an example is presented.

]]>Algorithms doi: 10.3390/a10030095

Authors: Manyu Xiao Quanyi Lv Zhuo Xing Yingchun Zhang

In this paper we propose a parallel two-stage iteration algorithm for solving large-scale continuous Sylvester equations. By splitting the coefficient matrices, the original linear system is transformed into a symmetric linear system which is then solved by using the SYMMLQ algorithm. In order to improve the relative parallel efficiency, an adjusting strategy is explored during the iteration calculation of the SYMMLQ algorithm to decrease the degree of the reduce-operator from two to one communications at each step. Moreover, the convergence of the iteration scheme is discussed, and finally numerical results are reported showing that the proposed method is an efficient and robust algorithm for this class of continuous Sylvester equations on a parallel machine.

]]>Algorithms doi: 10.3390/a10030096

Authors: Xueying Diao Qianhui Dong Zijian Yang Yibing Li

Spectrum sensing is of great importance in the cognitive radio (CR) networks. Compared with individual spectrum sensing, cooperative spectrum sensing (CSS) has been shown to greatly improve the accuracy of the detection. However, the existing CSS algorithms are sensitive to noise uncertainty and are inaccurate in low signal-to-noise ratio (SNR) detection. To address this, we propose a double-threshold CSS algorithm based on Sevcik fractal dimension (SFD) in this paper. The main idea of the presented scheme is to sense the presence of primary users in the local spectrum sensing by analyzing different characteristics of the SFD between signals and noise. Considering the stochastic fluctuation characteristic of the noise SFD in a certain range, we adopt the double-threshold method in the multi-cognitive user CSS so as to improve the detection accuracy, where thresholds are set according to the maximum and minimum values of the noise SFD. After obtaining the detection results, the cognitive user sends local detection results to the fusion center for reliability fusion. Simulation results demonstrate that the proposed method is insensitive to noise uncertainty. Simulations also show that the algorithm presented in this paper can achieve high detection performance at the low SNR region.

]]>Algorithms doi: 10.3390/a10030094

Authors: Peng Sun Zhiming Yang Yang Yu Junbao Li Xiyuan Peng

As technology scales, negative bias temperature instability (NBTI) becomes one of the primary failure mechanisms for Very Large Scale Integration (VLSI) circuits. Meanwhile, the leakage power increases dramatically as the supply/threshold voltage continues to scale down. These two issues pose severe reliability problems for complementary metal oxide semiconductor (CMOS) devices. Because both the NBTI and leakage are dependent on the input vector of the circuit, we present an input vector control (IVC) method based on a linear programming algorithm, which can co-optimize circuit aging and power dissipation simultaneously. In addition, our proposed IVC method is combined with the supply voltage assignment technique to further reduce delay degradation and leakage power. Experimental results on various circuits show the effectiveness of the proposed combination method.

]]>Algorithms doi: 10.3390/a10030093

Authors: William Weir Scott Emmons Ryan Gibson Dane Taylor Peter Mucha

We introduce the Convex Hull of Admissible Modularity Partitions (CHAMP) algorithm to prune and prioritize different network community structures identified across multiple runs of possibly various computational heuristics. Given a set of partitions, CHAMP identifies the domain of modularity optimization for each partition—i.e., the parameter-space domain where it has the largest modularity relative to the input set—discarding partitions with empty domains to obtain the subset of partitions that are “admissible” candidate community structures that remain potentially optimal over indicated parameter domains. Importantly, CHAMP can be used for multi-dimensional parameter spaces, such as those for multilayer networks where one includes a resolution parameter and interlayer coupling. Using the results from CHAMP, a user can more appropriately select robust community structures by observing the sizes of domains of optimization and the pairwise comparisons between partitions in the admissible subset. We demonstrate the utility of CHAMP with several example networks. In these examples, CHAMP focuses attention onto pruned subsets of admissible partitions that are 20-to-1785 times smaller than the sets of unique partitions obtained by community detection heuristics that were input into CHAMP.

]]>Algorithms doi: 10.3390/a10030092

Authors: Lijin Xie Qun Wan

Higher-order cyclic cumulants (CCs) have been widely adopted for automatic modulation recognition (AMR) in cognitive radio. However, the CC-based AMR suffers greatly from the requirement of high-rate sampling. To overcome this limit, we resort to the theory of compressive sensing (CS). By exploiting the sparsity of CCs, recognition features can be extracted from a small amount of compressive measurements via a rough CS reconstruction algorithm. Accordingly, a CS-based AMR scheme is formulated. Simulation results demonstrate the availability and robustness of the proposed approach.

]]>Algorithms doi: 10.3390/a10030091

Authors: Chengyuan Chen Qiang Shen

In support of reasoning with sparse rule bases, fuzzy rule interpolation (FRI) offers a helpful inference mechanism for deriving an approximate conclusion when a given observation has no overlap with any rule in the existing rule base. One of the recent and popular FRI approaches is the scale and move transformation-based rule interpolation, known as T-FRI in the literature. It supports both interpolation and extrapolation with multiple multi-antecedent rules. However, the difficult problem of defining the precise-valued membership functions required in the representation of fuzzy rules, or of the observations, restricts its applications. Fortunately, this problem can be alleviated through the use of type-2 fuzzy sets, owing to the fact that the membership functions of such fuzzy sets are themselves fuzzy, providing a more flexible means of modelling. This paper therefore, extends the existing T-FRI approach using interval type-2 fuzzy sets, which covers the original T-FRI as its specific instance. The effectiveness of this extension is demonstrated by experimental investigations and, also, by a practical application in comparison to the state-of-the-art alternative approach developed using rough-fuzzy sets.

]]>Algorithms doi: 10.3390/a10030090

Authors: Michael Hamann Eike Röhrs Dorothea Wagner

Community detection aims to find dense subgraphs in a network. We consider the problem of finding a community locally around a seed node both in unweighted and weighted networks. This is a faster alternative to algorithms that detect communities that cover the whole network when actually only a single community is required. Further, many overlapping community detection algorithms use local community detection algorithms as basic building block. We provide a broad comparison of different existing strategies of expanding a seed node greedily into a community. For this, we conduct an extensive experimental evaluation both on synthetic benchmark graphs as well as real world networks. We show that results both on synthetic as well as real-world networks can be significantly improved by starting from the largest clique in the neighborhood of the seed node. Further, our experiments indicate that algorithms using scores based on triangles outperform other algorithms in most cases. We provide theoretical descriptions as well as open source implementations of all algorithms used.

]]>Algorithms doi: 10.3390/a10030088

Authors: Francesco Mezzadri Emanuele Galligani

In this paper, we extend the analysis of the Lagged Diffusivity Method for nonlinear, non-steady reaction-convection-diffusion equations. In particular, we describe how the method can be used to solve the systems arising from different discretization schemes, recalling some results on the convergence of the method itself. Moreover, we also analyze the behavior of the method in case of problems presenting boundary layers or blow-up solutions.

]]>Algorithms doi: 10.3390/a10030089

Authors: José Ezquerro Miguel Hernández-Verón

The well-known Kantorovich technique based on majorizing sequences is used to analyse the convergence of Newton’s method when it is used to solve nonlinear Fredholm integral equations. In addition, we obtain information about the domains of existence and uniqueness of a solution for these equations. Finally, we illustrate the above with two particular Fredholm integral equations.

]]>Algorithms doi: 10.3390/a10030084

Authors: Li Xie Huizhong Yang

Due to the lack of powerful model description methods, the identification of Hammerstein systems based on the non-uniform input-output dataset remains a challenging problem. This paper introduces a time-varying backward shift operator to describe periodically non-uniformly sampled-data Hammerstein systems, which can simplify the structure of the lifted models using the traditional lifting technique. Furthermore, an auxiliary model-based multi-innovation stochastic gradient algorithm is presented to estimate the parameters involved in the linear and nonlinear blocks. The simulation results confirm that the proposed algorithm is effective and can achieve a high estimation performance.

]]>Algorithms doi: 10.3390/a10030087

Authors: Mozhdeh Shahbazi Gunho Sohn Jérôme Théau

In this paper, a robust technique based on a genetic algorithm is proposed for estimating two-view epipolar-geometry of uncalibrated perspective stereo images from putative correspondences containing a high percentage of outliers. The advantages of this technique are three-fold: (i) replacing random search with evolutionary search applying new strategies of encoding and guided sampling; (ii) robust and fast estimation of the epipolar geometry via detecting a more-than-enough set of inliers without making any assumptions about the probability distribution of the residuals; (iii) determining the inlier-outlier threshold based on the uncertainty of the estimated model. The proposed method was evaluated both on synthetic data and real images. The results were compared with the most popular techniques from the state-of-the-art, including RANSAC (random sample consensus), MSAC, MLESAC, Cov-RANSAC, LO-RANSAC, StaRSAC, Multi-GS RANSAC and least median of squares (LMedS). Experimental results showed that the proposed approach performed better than other methods regarding the accuracy of inlier detection and epipolar-geometry estimation, as well as the computational efficiency for datasets majorly contaminated by outliers and noise.

]]>Algorithms doi: 10.3390/a10030085

Authors: Camilo Caraveo Fevrier Valdez Oscar Castillo

Fuzzy logic is a soft computing technique that has been very successful in recent years when it is used as a complement to improve meta-heuristic optimization. In this paper, we present a new variant of the bio-inspired optimization algorithm based on the self-defense mechanisms of plants in the nature. The optimization algorithm proposed in this work is based on the predator-prey model originally presented by Lotka and Volterra, where two populations interact with each other and the objective is to maintain a balance. The system of predator-prey equations use four variables (α, β, λ, δ) and the values of these variables are very important since they are in charge of maintaining a balance between the pair of equations. In this work, we propose the use of Type-2 fuzzy logic for the dynamic adaptation of the variables of the system. This time a fuzzy controller is in charge of finding the optimal values for the model variables, the use of this technique will allow the algorithm to have a higher performance and accuracy in the exploration of the values.

]]>Algorithms doi: 10.3390/a10030086

Authors: Wei Zheng Yanyan Tan Xiaonan Fang Shengtao Li

MOEA/D is a promising multi-objective evolutionary algorithm based on decomposition, and it has been used to solve many multi-objective optimization problems very well. However, there is a class of multi-objective problems, called many-objective optimization problems, but the original MOEA/D cannot solve them well. In this paper, an improved MOEA/D with optimal differential evolution (oDE) schemes is proposed, called MOEA/D-oDE, aiming to solve many-objective optimization problems. Compared with MOEA/D, MOEA/D-oDE has two distinguishing points. On the one hand, MOEA/D-oDE adopts a newly-introduced decomposition approach to decompose the many-objective optimization problems, which combines the advantages of the weighted sum approach and the Tchebycheff approach. On the other hand, a kind of combination mechanism for DE operators is designed for finding the best child solution so as to do the a posteriori computing. In our experimental study, six continuous test instances with 4–6 objectives comparing NSGA-II (nondominated sorting genetic algorithm II) and MOEA/D as accompanying experiments are applied. Additionally, the final results indicate that MOEA/D-oDE outperforms NSGA-II and MOEA/D in almost all cases, particularly in those problems that have complicated Pareto shapes and higher dimensional objectives, where its advantages are more obvious.

]]>Algorithms doi: 10.3390/a10030083

Authors: Juan Barraza Patricia Melin Fevrier Valdez Claudia Gonzalez

The main goal of this paper is to improve the performance of the Fireworks Algorithm (FWA). To improve the performance of the FWA we propose three modifications: the first modification is to change the stopping criteria, this is to say, previously, the number of function evaluations was utilized as a stopping criteria, and we decided to change this to specify a particular number of iterations; the second and third modifications consist on introducing a dispersion metric (dispersion percent), and both modifications were made with the goal of achieving dynamic adaptation of the two parameters in the algorithm. The parameters that were controlled are the explosion amplitude and the number of sparks, and it is worth mentioning that the control of these parameters is based on a fuzzy logic approach. To measure the impact of these modifications, we perform experiments with 14 benchmark functions and a comparative study shows the advantage of the proposed approach. We decided to call the proposed algorithms Iterative Fireworks Algorithm (IFWA) and two variants of the Dispersion Percent Iterative Fuzzy Fireworks Algorithm (DPIFWA-I and DPIFWA-II, respectively).

]]>Algorithms doi: 10.3390/a10030082

Authors: Cinthia Peraza Fevrier Valdez Patricia Melin

This article focuses on the dynamic parameter adaptation in the harmony search algorithm using Type-1 and interval Type-2 fuzzy logic. In particular, this work focuses on the adaptation of the parameters of the original harmony search algorithm. At present there are several types of algorithms that can solve complex real-world problems with uncertainty management. In this case the proposed method is in charge of optimizing the membership functions of three benchmark control problems (water tank, shower, and mobile robot). The main goal is to find the best parameters for the membership functions in the controller to follow a desired trajectory. Noise experiments are performed to test the efficacy of the method.

]]>Algorithms doi: 10.3390/a10030081

Authors: Xiaocong Wei Hongfei Lin Yuhai Yu Liang Yang

The literature [-5]contains several reports evaluating the abilities of deep neural networks in text transfer learning. To our knowledge, however, there have been few efforts to fully realize the potential of deep neural networks in cross-domain product review sentiment classification. In this paper, we propose a two-layer convolutional neural network (CNN) for cross-domain product review sentiment classification (LM-CNN-LB). Transfer learning research into product review sentiment classification based on deep neural networks has been limited by the lack of a large-scale corpus; we sought to remedy this problem using a large-scale auxiliary cross-domain dataset collected from Amazon product reviews. Our proposed framework exhibits the dramatic transferability of deep neural networks for cross-domain product review sentiment classification and achieves state-of-the-art performance. The framework also outperforms complex engineered features used with a non-deep neural network method. The experiments demonstrate that introducing large-scale data from similar domains is an effective way to resolve the lack of training data. The LM-CNN-LB trained on the multi-source related domain dataset outperformed the one trained on a single similar domain.

]]>Algorithms doi: 10.3390/a10030080

Authors: Yasser El Khamlichi Abderrahim Tahiri Anouar Abtoy Inmaculada Medina-Bulo Francisco Palomo-Lozano

Wireless sensor network (WSN) applications are rapidly growing and are widely used in various disciplines. Deployment is one of the key issues to be solved in WSNs, since the sensor nodes’ positioning affects highly the system performance. An optimal WSN deployment should maximize the collection of the desired interest phenomena, guarantee the required coverage and connectivity, extend the network lifetime, and minimize the network cost in terms of energy consumption. Most of the research effort in this area aims to solve the deployment issue, without minimizing the network cost by reducing unnecessary working nodes in the network. In this paper, we propose a deployment approach based on the gradient method and the Simulated Annealing algorithm to solve the sensor deployment problem with the minimum number of sensor nodes. The proposed algorithm is able to heuristically optimize the number of sensors and their positions in order to achieve the desired application requirements.

]]>Algorithms doi: 10.3390/a10030079

Authors: Juan Carlos Guzman Patricia Melin German Prado-Arechiga

A neuro fuzzy hybrid model (NFHM) is proposed as a new artificial intelligence method to classify blood pressure (BP). The NFHM uses techniques such as neural networks, fuzzy logic and evolutionary computation, and in the last case genetic algorithms (GAs) are used. The main goal is to model the behavior of blood pressure based on monitoring data of 24 h per patient and based on this to obtain the trend, which is classified using a fuzzy system based on rules provided by an expert, and these rules are optimized by a genetic algorithm to obtain the best possible number of rules for the classifier with the lowest classification error. Simulation results are presented to show the advantage of the proposed model.

]]>Algorithms doi: 10.3390/a10030078

Authors: Yunqiu Shen Tjalling Ypma

The nonlinear least squares problem m i n y , z ∥ A ( y ) z + b ( y ) ∥ , where A ( y ) is a full-rank ( N + ℓ ) × N matrix, y ∈ R n , z ∈ R N and b ( y ) ∈ R N + ℓ with ℓ ≥ n , can be solved by first solving a reduced problem m i n y ∥ f ( y ) ∥ to find the optimal value y * of y, and then solving the resulting linear least squares problem m i n z ∥ A ( y * ) z + b ( y * ) ∥ to find the optimal value z * of z. We have previously justified the use of the reduced function f ( y ) = C T ( y ) b ( y ) , where C ( y ) is a matrix whose columns form an orthonormal basis for the nullspace of A T ( y ) , and presented a quadratically convergent Gauss–Newton type method for solving m i n y ∥ C T ( y ) b ( y ) ∥ based on the use of QR factorization. In this note, we show how LU factorization can replace the QR factorization in those computations, halving the associated computational cost while also providing opportunities to exploit sparsity and thus further enhance computational efficiency.

]]>Algorithms doi: 10.3390/a10030077

Authors: Emanuel Ontiveros-Robles Patricia Melin Oscar Castillo

Interval Type-2 fuzzy systems allow the possibility of considering uncertainty in models based on fuzzy systems, and enable an increase of robustness in solutions to applications, but also increase the complexity of the fuzzy system design. Several attempts have been previously proposed to reduce the computational cost of the type-reduction stage, as this process requires a lot of computing time because it is basically a numerical approximation based on sampling, and the computational cost is proportional to the number of samples, but also the error is inversely proportional to the number of samples. Several works have focused on reducing the computational cost of type-reduction by developing strategies to reduce the number of operations. The first type-reduction method was proposed by Karnik and Mendel (KM), and then was followed by its enhanced version called EKM. Then continuous versions were called CKM and CEKM, and there were variations of this and also other types of variations that eliminate the type-reduction process reducing the computational cost to a Type-1 defuzzification, such as the Nie-Tan versions and similar enhancements. In this work we analyzed and proposed a variant of CEKM by viewing this process as solving a root-finding problem, in this way taking advantage of existing numerical methods to solve the type-reduction problem, the main objective being eliminating the type-reduction process and also providing a continuous solution of the defuzzification.

]]>Algorithms doi: 10.3390/a10030076

Authors: Wenke Zang Weining Zhang Wenqian Zhang Xiyu Liu

As one of the evolutionary heuristics methods, genetic algorithms (GAs) have shown a promising ability to solve complex optimization problems. However, existing GAs still have difficulties in finding the global optimum and avoiding premature convergence. To further improve the search efficiency and convergence rate of evolution algorithms, inspired by the mechanism of biological DNA genetic information and evolution, we present a new genetic algorithm, called GA-TNE+DRO, which uses a novel triplet nucleotide coding scheme to encode potential solutions and a set of new genetic operators to search for globally optimal solutions. The coding scheme represents potential solutions as a sequence of triplet nucleotides and the DNA reproduction operations mimic the DNA reproduction process more vividly than existing DNA-GAs. We compared our algorithm with several existing GA and DNA-based GA algorithms using a benchmark of eight unconstrained optimization functions. Our experimental results show that the proposed algorithm can converge to solutions much closer to the global optimal solutions in a much lower number of iterations than the existing algorithms. A complexity analysis also shows that our algorithm is computationally more efficient than the existing algorithms.

]]>Algorithms doi: 10.3390/a10030075

Authors: Stefan Maruster

We investigate the efficiency of multi-step Newton method (the classical Newton method in which the first derivative is re-evaluated periodically after m steps) for solving nonlinear equations, F ( x ) = 0 , F : D ⊆ R n → R n . We highlight the following property of multi-step Newton method with respect to some other Newton-type method: for a given n, there exist thresholds of m, that is an interval ( m i , m s ) , such that for m inside of this interval, the efficiency index of multi-step Newton method is better than that of other Newton-type method. We also search for optimal values of m.

]]>Algorithms doi: 10.3390/a10030074

Authors: Annabell Berger Berit Schreck

The discrepancy BR for an m × n 0, 1-matrix from Brualdi and Sanderson in 1998 is defined as the minimum number of 1 s that need to be shifted in each row to the left to achieve its Ferrers matrix, i.e., each row consists of consecutive 1 s followed by consecutive 0 s. For ecological bipartite networks, BR describes a nested set of relationships. Since two different labelled networks can be isomorphic, but possess different discrepancies due to different adjacency matrices, we define a metric determining the minimum discrepancy in an isomorphic class. We give a reduction to k ≤ n minimum weighted perfect matching problems. We show on 289 ecological matrices (given as a benchmark by Atmar and Patterson in 1995) that classical discrepancy can underestimate the nestedness by up to 30%.

]]>Algorithms doi: 10.3390/a10030073

Authors: Huanyu Chen Tong Chen Zhihao Zhang Guangyuan Liu

Variable selection is a key step for eliminating redundant information in spectroscopy. Among various variable selection methods, the physarum network (PN) is a newly-introduced and efficient one. However, the whole spectrum has to be equally divided into sub-spectral bands in PN. These division criteria limit the selecting ability and prediction performance. In this paper, we transform the spectrum division problem into a clustering problem and solve the problem by using an affinity propagation (AP) algorithm, an adaptive clustering method, to find the optimized number of sub-spectral bands and the number of wavelengths in each sub-spectral band. Experimental results show that combining AP and PN together can achieve similar prediction accuracy with much less wavelength than what PN alone can achieve.

]]>Algorithms doi: 10.3390/a10030072

Authors: Jie Yang Jian Xiong Guan Gui Rongfang Song Wang Luo Xianzhong Long

Video quality assessment (VQA) plays an important role in video applications for quality evaluation and resource allocation. It aims to evaluate video quality in a way that is consistent with human perception. In this letter, a hierarchical gradient similarity based VQA metric is proposed inspired by the structure of the primate visual cortex, in which visual information is processed through sequential visual areas. These areas are modeled with the corresponding measures to evaluate the overall perceptual quality. Experimental results on the LIVE database show that the proposed VQA metric significantly outperforms most of the state-of-the-art VQA metrics.

]]>Algorithms doi: 10.3390/a10020071

Authors: Qixuan Bi Wenhao Gui

In this paper, we consider the problem of estimating stress-strength reliability for inverse Weibull lifetime models having the same shape parameters but different scale parameters. We obtain the maximum likelihood estimator and its asymptotic distribution. Since the classical estimator doesn’t hold explicit forms, we propose an approximate maximum likelihood estimator. The asymptotic confidence interval and two bootstrap intervals are obtained. Using the Gibbs sampling technique, Bayesian estimator and the corresponding credible interval are obtained. The Metropolis-Hastings algorithm is used to generate random variates. Monte Carlo simulations are conducted to compare the proposed methods. Analysis of a real dataset is performed.

]]>Algorithms doi: 10.3390/a10020070

Authors: Ying Mei Guanzheng Tan Zhentao Liu

Classification is an important task of machine intelligence in the field of information. The artificial neural network (ANN) is widely used for classification. However, the traditional ANN shows slow training speed, and it is hard to meet the real-time requirement for large-scale applications. In this paper, an improved brain-inspired emotional learning (BEL) algorithm is proposed for fast classification. The BEL algorithm was put forward to mimic the high speed of the emotional learning mechanism in mammalian brain, which has the superior features of fast learning and low computational complexity. To improve the accuracy of BEL in classification, the genetic algorithm (GA) is adopted for optimally tuning the weights and biases of amygdala and orbitofrontal cortex in the BEL neural network. The combinational algorithm named as GA-BEL has been tested on eight University of California at Irvine (UCI) datasets and two well-known databases (Japanese Female Facial Expression, Cohn–Kanade). The comparisons of experiments indicate that the proposed GA-BEL is more accurate than the original BEL algorithm, and it is much faster than the traditional algorithm.

]]>Algorithms doi: 10.3390/a10020069

Authors: Anak Agung Putri Ratna Prima Dewi Purnamasari Boma Anantasatya Adhi F. Astha Ekadiyanto Muhammad Salman Mardiyah Mardiyah Darien Jonathan Winata

Computerized cross-language plagiarism detection has recently become essential. With the scarcity of scientific publications in Bahasa Indonesia, many Indonesian authors frequently consult publications in English in order to boost the quantity of scientific publications in Bahasa Indonesia (which is currently rising). Due to the syntax disparity between Bahasa Indonesia and English, most of the existing methods for automated cross-language plagiarism detection do not provide satisfactory results. This paper analyses the probability of developing Latent Semantic Analysis (LSA) for a computerized cross-language plagiarism detector for two languages with different syntax. To improve performance, various alterations in LSA are suggested. By using a linear vector quantization (LVQ) classifier in the LSA and taking into account the Frobenius norm, output has reached up to 65.98% in accuracy. The results of the experiments showed that the best accuracy achieved is 87% with a document size of 6 words, and the document definition size must be kept below 10 words in order to maintain high accuracy. Additionally, based on experimental results, this paper suggests utilizing the frequency occurrence method as opposed to the binary method for the term–document matrix construction.

]]>Algorithms doi: 10.3390/a10020068

Authors: Radu-Emil Precup Radu-Codrut David Alexandra-Iulia Szedlak-Stinean Emil M. Petriu Florin Dragan

This paper proposes an easily understandable Grey Wolf Optimizer (GWO) applied to the optimal tuning of the parameters of Takagi-Sugeno proportional-integral fuzzy controllers (T-S PI-FCs). GWO is employed for solving optimization problems focused on the minimization of discrete-time objective functions defined as the weighted sum of the absolute value of the control error and of the squared output sensitivity function, and the vector variable consists of the tuning parameters of the T-S PI-FCs. Since the sensitivity functions are introduced with respect to the parametric variations of the process, solving these optimization problems is important as it leads to fuzzy control systems with a reduced process parametric sensitivity obtained by a GWO-based fuzzy controller tuning approach. GWO algorithms applied with this regard are formulated in easily understandable terms for both vector and scalar operations, and discussions on stability, convergence, and parameter settings are offered. The controlled processes referred to in the course of this paper belong to a family of nonlinear servo systems, which are modeled by second order dynamics plus a saturation and dead zone static nonlinearity. Experimental results concerning the angular position control of a laboratory servo system are included for validating the proposed method.

]]>Algorithms doi: 10.3390/a10020067

Authors: Yancai Xiao Yujia Wang Huan Mu Na Kang

The misalignment of the drive system of the DFIG (Doubly Fed Induction Generator) wind turbine is one of the important factors that cause damage to the gears, bearings of the high-speed gearbox and the generator bearings. How to use the limited information to accurately determine the type of failure has become a difficult study for the scholars. In this paper, the time-domain indexes and frequency-domain indexes are extracted by using the vibration signals of various misaligned simulation conditions of the wind turbine drive system, and the time-frequency domain features—energy entropy are also extracted by the IEMD (Improved Empirical Mode Decomposition). A mixed-domain feature set is constructed by them. Then, SVM (Support Vector Machine) is used as the classifier, the mixed-domain features are used as the inputs of SVM, and PSO (Particle Swarm Optimization) is used to optimize the parameters of SVM. The fault types of misalignment are classified successfully. Compared with other methods, the accuracy of the given fault isolation model is improved.

]]>Algorithms doi: 10.3390/a10020066

Authors: Hamid Hassannejad Guido Matrella Paolo Ciampolini Ilaria Munari Monica Mordonini Stefano Cagnoni

A balanced diet is the key to a healthy lifestyle and is crucial for preventing or dealing with many chronic diseases such as diabetes and obesity. Therefore, monitoring diet can be an effective way of improving people’s health. However, manual reporting of food intake has been shown to be inaccurate and often impractical. This paper presents a new approach to food intake quantity estimation using image-based modeling. The modeling method consists of three steps: firstly, a short video of the food is taken by the user’s smartphone. From such a video, six frames are selected based on the pictures’ viewpoints as determined by the smartphone’s orientation sensors. Secondly, the user marks one of the frames to seed an interactive segmentation algorithm. Segmentation is based on a Gaussian Mixture Model alongside the graph-cut algorithm. Finally, a customized image-based modeling algorithm generates a point-cloud to model the food. At the same time, a stochastic object-detection method locates a checkerboard used as size/ground reference. The modeling algorithm is optimized such that the use of six input images still results in an acceptable computation cost. In our evaluation procedure, we achieved an average accuracy of 92 % on a test set that includes images of different kinds of pasta and bread, with an average processing time of about 23 s.

]]>Algorithms doi: 10.3390/a10020065

Authors: Xin Tian Song Li

We introduce a seismic signal compression method based on nonparametric Bayesian dictionary learning method via clustering. The seismic data is compressed patch by patch, and the dictionary is learned online. Clustering is introduced for dictionary learning. A set of dictionaries could be generated, and each dictionary is used for one cluster’s sparse coding. In this way, the signals in one cluster could be well represented by their corresponding dictionaries. A nonparametric Bayesian dictionary learning method is used to learn the dictionaries, which naturally infers an appropriate dictionary size for each cluster. A uniform quantizer and an adaptive arithmetic coding algorithm are adopted to code the sparse coefficients. With comparisons to other state-of-the art approaches, the effectiveness of the proposed method could be validated in the experiments.

]]>Algorithms doi: 10.3390/a10020064

Authors: Sergio Amat Ioannis Argyros Miguel Hernández-Verón Natalia Romero

This paper is devoted to the semilocal convergence of a Househölder-like method for nonlinear equations. The method includes many of the studied third order iterative methods. In the present study, we use our new idea of restricted convergence domains leading to smaller γ -parameters, which in turn lead to the following advantages over earlier works (and under the same computational cost): larger convergence domain; tighter error bounds on the distances involved, and at least as precise information on the location of the solution.

]]>