Algorithms doi: 10.3390/a11080128

Authors: Shuhei Denzumi Jun Kawahara Koji Tsuda Hiroki Arimura Shin-ichi Minato Kunihiko Sadakane

In this article, we propose a succinct data structure of zero-suppressed binary decision diagrams (ZDDs). A ZDD represents sets of combinations efficiently and we can perform various set operations on the ZDD without explicitly extracting combinations. Thanks to these features, ZDDs have been applied to web information retrieval, information integration, and data mining. However, to support rich manipulation of sets of combinations and update ZDDs in the future, ZDDs need too much space, which means that there is still room to be compressed. The paper introduces a new succinct data structure, called DenseZDD, for further compressing a ZDD when we do not need to conduct set operations on the ZDD but want to examine whether a given set is included in the family represented by the ZDD, and count the number of elements in the family. We also propose a hybrid method, which combines DenseZDDs with ordinary ZDDs. By numerical experiments, we show that the sizes of our data structures are three times smaller than those of ordinary ZDDs, and membership operations and random sampling on DenseZDDs are about ten times and three times faster than those on ordinary ZDDs for some datasets, respectively.

]]>Algorithms doi: 10.3390/a11080127

Authors: Mingbin Zeng Xu Yang Mengxing Wang Bangjiang Xu

In recent years, Intelligent Transportation Systems (ITS) have developed a lot. More and more sensors and communication technologies (e.g., cloud computing) are being integrated into cars, which opens up a new design space for vehicular-based applications. In this paper, we present the Spatial Optimized Dynamic Path Planning algorithm. Our contributions are, firstly, to enhance the effective of loading mechanism for road maps by dividing the connected sub-net, and building a spatial index; and secondly, to enhance the effect of the dynamic path planning by optimizing the search direction. We use the real road network and real-time traffic flow data of Karamay city to simulate the effect of our algorithm. Experiments show that our Spatial Optimized Dynamic Path Planning algorithm can significantly reduce the time complexity, and is better suited for use as a real-time navigation system. The algorithm can achieve superior real-time performance and obtain the optimal solution in dynamic path planning.

]]>Algorithms doi: 10.3390/a11080126

Authors: Zhiguo Song Jifeng Sun Jialin Yu Shengqing Liu

Appearance models play an important role in visual tracking. Effective modeling of the appearance of tracked objects is still a challenging problem because of object appearance changes caused by factors, such as partial occlusion, illumination variation and deformation, etc. In this paper, we propose a tracking method based on the patch descriptor and the structural local sparse representation. In our method, the object is firstly divided into multiple non-overlapped patches, and the patch sparse coefficients are obtained by structural local sparse representation. Secondly, each patch is further decomposed into several sub-patches. The patch descriptors are defined as the proportion of sub-patches, of which the reconstruction error is less than the given threshold. Finally, the appearance of an object is modeled by the patch descriptors and the patch sparse coefficients. Furthermore, in order to adapt to appearance changes of an object and alleviate the model drift, an outlier-aware template update scheme is introduced. Experimental results on a large benchmark dataset demonstrate the effectiveness of the proposed method.

]]>Algorithms doi: 10.3390/a11080125

Authors: Yeqing Yan Zhigang Chen Jia Wu Leilei Wang

With the popularization of mobile communication equipment, human activities have an increasing impact on the structure of networks, and so the social characteristics of opportunistic networks become increasingly obvious. Opportunistic networks are increasingly used in social situations. However, existing routing algorithms are not suitable for opportunistic social networks, because traditional opportunistic network routing does not consider participation in human activities, which usually causes a high ratio of transmission delay and routing overhead. Therefore, this research proposes an effective data transmission algorithm based on social relationships (ESR), which considers the community characteristics of opportunistic mobile social networks. This work uses the idea of the faction to divide the nodes in the network into communities, reduces the number of inefficient nodes in the community, and performs another contraction of the structure. Simulation results show that the ESR algorithm, through community transmission, is not only faster and safer, but also has lower transmission delay and routing overhead compared with the spray and wait algorithm, SCR algorithm and the EMIST algorithm.

]]>Algorithms doi: 10.3390/a11080124

Authors: Yihong Li Fangzheng Liu Zhenyu Du Dubing Zhang

In the malware detection process, obfuscated malicious codes cannot be efficiently and accurately detected solely in the dynamic or static feature space. Aiming at this problem, an integrative feature extraction algorithm based on simhash was proposed, which combines the static information e.g., API (Application Programming Interface) calls and dynamic information (such as file, registry and network behaviors) of malicious samples to form integrative features. The experiment extracts the integrative features of some static information and dynamic information, and then compares the classification, time and obfuscated-detection performance of the static, dynamic and integrated features, respectively, by using several common machine learning algorithms. The results show that the integrative features have better time performance than the static features, and better classification performance than the dynamic features, and almost the same obfuscated-detection performance as the dynamic features. This algorithm can provide some support for feature extraction of malware detection.

]]>Algorithms doi: 10.3390/a11080123

Authors: Yamur K. Al-Douri Hussan Hamodi Jan Lundberg

The aim of this study has been to develop a novel two-level multi-objective genetic algorithm (GA) to optimize time series forecasting data for fans used in road tunnels by the Swedish Transport Administration (Trafikverket). Level 1 is for the process of forecasting time series cost data, while level 2 evaluates the forecasting. Level 1 implements either a multi-objective GA based on the ARIMA model or a multi-objective GA based on the dynamic regression model. Level 2 utilises a multi-objective GA based on different forecasting error rates to identify a proper forecasting. Our method is compared with using the ARIMA model only. The results show the drawbacks of time series forecasting using only the ARIMA model. In addition, the results of the two-level model show the drawbacks of forecasting using a multi-objective GA based on the dynamic regression model. A multi-objective GA based on the ARIMA model produces better forecasting results. In level 2, five forecasting accuracy functions help in selecting the best forecasting. Selecting a proper methodology for forecasting is based on the averages of the forecasted data, the historical data, the actual data and the polynomial trends. The forecasted data can be used for life cycle cost (LCC) analysis.

]]>Algorithms doi: 10.3390/a11080122

Authors: Chi-Yi Tsai Kuang-Jui Hsu Humaira Nisar

Three-Dimensional (3D) object pose estimation plays a crucial role in computer vision because it is an essential function in many practical applications. In this paper, we propose a real-time model-based object pose estimation algorithm, which integrates template matching and Perspective-n-Point (PnP) pose estimation methods to deal with this issue efficiently. The proposed method firstly extracts and matches keypoints of the scene image and the object reference image. Based on the matched keypoints, a two-dimensional (2D) planar transformation between the reference image and the detected object can be formulated by a homography matrix, which can initialize a template tracking algorithm efficiently. Based on the template tracking result, the correspondence between image features and control points of the Computer-Aided Design (CAD) model of the object can be determined efficiently, thus leading to a fast 3D pose tracking result. Finally, the 3D pose of the object with respect to the camera is estimated by a PnP solver based on the tracked 2D-3D correspondences, which improves the accuracy of the pose estimation. Experimental results show that the proposed method not only achieves real-time performance in tracking multiple objects, but also provides accurate pose estimation results. These advantages make the proposed method suitable for many practical applications, such as augmented reality.

]]>Algorithms doi: 10.3390/a11080121

Authors: Feilai Pan Jun Li Bendong Tan Ciling Zeng Xinfan Jiang Li Liu Jun Yang

With the interconnection between large power grids, the issue of security and stability has become increasingly prominent. At present, data-driven power system adaptive transient stability assessment methods have achieved excellent performances by balancing speed and accuracy, but the complicated construction and parameters are difficult to obtain. This paper proposes a stacked-GRU (Gated Recurrent Unit)-based transient stability intelligent assessment method, which builds a stacked-GRU model based on time-dependent parameter sharing and spatial stacking. By using the time series data after power system failure, the offline training is performed to obtain the optimal parameters of stacked-GRU. When the application is online, it is assessed by framework of confidence. Basing on New England power system, the performance of proposed adaptive transient stability assessment method is investigated. Simulation results show that the proposed model realizes reliable and accurate assessment of transient stability and it has the advantages of short assessment time with less complex model structure to leave time for emergency control.

]]>Algorithms doi: 10.3390/a11080120

Authors: Wenying Wu Ying Li Zhiwei Ni Feifei Jin Xuhui Zhu

Based on the probabilistic interval-valued hesitant fuzzy information aggregation operators, this paper investigates a novel multi-attribute group decision making (MAGDM) model to address the serious loss of information in a hesitant fuzzy information environment. Firstly, the definition of probabilistic interval-valued hesitant fuzzy set will be introduced, and then, using Archimedean norm, some new probabilistic interval-valued hesitant fuzzy operations are defined. Secondly, based on these operations, the generalized probabilistic interval-valued hesitant fuzzy ordered weighted averaging (GPIVHFOWA) operator, and the generalized probabilistic interval-valued hesitant fuzzy ordered weighted geometric (GPIVHFOWG) operator are proposed, and their desirable properties are discussed. We further study their common forms and analyze the relationship among these proposed operators. Finally, a new probabilistic interval-valued hesitant fuzzy MAGDM model is constructed, and the feasibility and effectiveness of the proposed model are verified by using an example of supplier selection.

]]>Algorithms doi: 10.3390/a11080119

Authors: Yucheng Lin Zhigang Chen Jia Wu Leilei Wang

The mobility of nodes leads to dynamic changes in topology structure, which makes the traditional routing algorithms of a wireless network difficult to apply to the opportunistic network. In view of the problems existing in the process of information forwarding, this paper proposed a routing algorithm based on the cosine similarity of data packets between nodes (cosSim). The cosine distance, an algorithm for calculating the similarity between text data, is used to calculate the cosine similarity of data packets between nodes. The data packet set of nodes are expressed in the form of vectors, thereby facilitating the calculation of the similarity between the nodes. Through the definition of the upper and lower thresholds, the similarity between the nodes is filtered according to certain rules, and finally obtains a plurality of relatively reliable transmission paths. Simulation experiments show that compared with the traditional opportunistic network routing algorithm, such as the Spray and Wait (S&amp;W) algorithm and Epidemic algorithm, the cosSim algorithm has a better transmission effect, which can not only improve the delivery ratio, but also reduce the network transmission delay and decline the routing overhead.

]]>Algorithms doi: 10.3390/a11080118

Authors: Andrej Brodnik Matevž Jekovec

We consider a sliding window W over a stream of characters from some alphabet of constant size. We want to look up a pattern in the current sliding window content and obtain all positions of the matches. We present an indexed version of the sliding window, based on a suffix tree. The data structure of size Θ(|W|) has optimal time queries Θ(m+occ) and amortized constant time updates, where m is the length of the query string and occ is its number of occurrences.

]]>Algorithms doi: 10.3390/a11080117

Authors: Yanzhu Hu Song Wang Xinbo Ai

This paper aims to improve the source tracking efficiency of distributed vibration signals generated by phase-sensitive optical time-domain reflectometry (&Phi;-OTDR). Considering the two dimensions (time and length) of &Phi;-OTDR signals, the authors saved and processed these signals as images after particle filtering. The filtering method could save 0.1% of hard drive space without sacrificing the original features of the signals. Then, an integrated feature extraction method was proposed to further process the generated image. The method combines three individual extraction methods, namely, texture feature extraction, shape feature extraction and intrinsic feature extraction. Subsequently, the signal of each frame image was recognized to track the vibration source. To verify the effect of the proposed method, several experiments were carried out to compare it with popular and traditional approaches. The results show that: Hard drive space is greatly conserved by saving the distributed vibration signals as images; the proposed particle filter is a desirable way to screen the vibration signals for monitoring; the integrated feature extraction outperforms the individual extraction methods for texture features, shape features and intrinsic features; the proposed method has a better effect than other popular integrated feature extraction methods; and, the signal source tracking method has little impact on the positioning accuracy of the vibration source. The research findings provide important insights into the source tracking of &Phi;-OTDR signals.

]]>Algorithms doi: 10.3390/a11080116

Authors: Huamei Qi Fengqi Liu Tailong Xiao Jiang Su

In an Ad hoc sensor network, nodes have characteristics of limited battery energy, self-organization and low mobility. Due to the mobility and heterogeneity of the energy consumption in the hierarchical network, the cluster head and topology are changed dynamically. Therefore, topology control and energy consumption are growing to be critical in enhancing the stability and prolonging the lifetime of the network. In order to improve the survivability of Ad hoc network effectively, this paper proposes a new algorithm named the robust, energy-efficient weighted clustering algorithm (RE2WCA). For the homogeneous of the energy consumption; the proposed clustering algorithm takes the residual energy and group mobility into consideration by restricting minimum iteration times. In addition, a distributed fault detection algorithm and cluster head backup mechanism are presented to achieve the periodic and real-time topology maintenance to enhance the robustness of the network. The network is analyzed and the simulations are performed to compare the performance of this new clustering algorithm with the similar algorithms in terms of cluster characteristics, lifetime, throughput and energy consumption of the network. The result shows that the proposed algorithm provides better performance than others.

]]>Algorithms doi: 10.3390/a11080115

Authors: Jing Wang Lidong Wang Xiaodong Liu Yan Ren Ye Yuan

The goal of object retrieval is to rank a set of images by their similarity compared with a query image. Nowadays, content-based image retrieval is a hot research topic, and color features play an important role in this procedure. However, it is important to establish a measure of image similarity in advance. The innovation point of this paper lies in the following. Firstly, the idea of the proximity space theory is utilized to retrieve the relevant images between the query image and images of database, and we use the color histogram of an image to obtain the Top-ranked colors, which can be regard as the object set. Secondly, the similarity is calculated based on an improved dominance granule structure similarity method. Thus, we propose a color-based image retrieval method by using proximity space theory. To detect the feasibility of this method, we conducted an experiment on COIL-20 image database and Corel-1000 database. Experimental results demonstrate the effectiveness of the proposed framework and its applications.

]]>Algorithms doi: 10.3390/a11080114

Authors: Mihaly Mezei

The steady growth of the Protein Data Bank (PDB) suggests the periodic repetition of searches for sequences that form different secondary structures in different protein structures; these are called chameleon sequences. This paper presents a fast (nlog(n)) algorithm for such searches and presents the results on all protein structures in the PDB. The longest such sequence found consists of 20 residues.

]]>Algorithms doi: 10.3390/a11080113

Authors: Xiangfeng Su Huaiqing Zhang Lin Chen Ling Qin Lili Yu

Envelope current signals are increasingly emerging in power systems, and their parameter identification is particularly necessary for accurate measurement of electrical energy. In order to analyze the envelope current signal, the harmonic parameters, as well as the envelope parameters, need to be calculated. The interpolation fast Fourier transform (FFT) is a widely used approach which can estimate the signal frequency with high precision, but it cannot calculate the envelope parameters of the signal. Therefore, this paper proposes an improved method based on windowed interpolation FFT (WIFFT) and differential evolution (DE). The amplitude and phase parameters obtained through WIFFT and the envelope parameters estimated by the envelope analysis are optimized using the DE algorithm, which makes full use of the performance advantage of DE. The simulation results show that the proposed method can improve the accuracy of the harmonic parameters and the envelope parameter significantly. In addition, it has good anti-noise ability and high precision.

]]>Algorithms doi: 10.3390/a11080112

Authors: Ruhua Wang Ling Li Jun Li

In this paper, damage detection/identification for a seven-storey steel structure is investigated via using the vibration signals and deep learning techniques. Vibration characteristics, such as natural frequencies and mode shapes are captured and utilized as input for a deep learning network while the output vector represents the structural damage associated with locations. The deep auto-encoder with sparsity constraint is used for effective feature extraction for different types of signals and another deep auto-encoder is used to learn the relationship of different signals for final regression. The existing SAF model in a recent research study for the same problem processed all signals in one serial auto-encoder model. That kind of models have the following difficulties: (1) the natural frequencies and mode shapes are in different magnitude scales and it is not logical to normalize them in the same scale in building the models with training samples; (2) some frequencies and mode shapes may not be related to each other and it is not fair to use them for dimension reduction together. To tackle the above-mentioned problems for the multi-scale dataset in SHM, a novel parallel auto-encoder framework (Para-AF) is proposed in this paper. It processes the frequency signals and mode shapes separately for feature selection via dimension reduction and then combine these features together in relationship learning for regression. Furthermore, we introduce sparsity constraint in model reduction stage for performance improvement. Two experiments are conducted on performance evaluation and our results show the significant advantages of the proposed model in comparison with the existing approaches.

]]>Algorithms doi: 10.3390/a11080111

Authors: David Völgyes Anne Catrine Trægde Martinsen Arne Stray-Pedersen Dag Waaler Marius Pedersen

Computed Tomography (CT) images have a high dynamic range, which makes visualization challenging. Histogram equalization methods either use spatially invariant weights or limited kernel size due to the complexity of pairwise contribution calculation. We present a weighted histogram equalization-based tone mapping algorithm which utilizes Fast Fourier Transform for distance-dependent contribution calculation and distance-based weights. The weights follow power-law without distance-based cut-off. The resulting images have good local contrast without noticeable artefacts. The results are compared to eight popular tone mapping operators.

]]>Algorithms doi: 10.3390/a11080109

Authors: Liu Liu Kaile Liu Zhenghai Cong Jiali Zhao Yefei Ji Jun He

The exponential increase in online reviews and recommendations makes document classification and sentiment analysis a hot topic in academic and industrial research. Traditional deep learning based document classification methods require the use of full textual information to extract features. In this paper, in order to tackle long document, we proposed three methods that use local convolutional feature aggregation to implement document classification. The first proposed method randomly draws blocks of continuous words in the full document. Each block is then fed into the convolution neural network to extract features and then are concatenated together to output the classification probability through a classifier. The second model improves the first by capturing the contextual order information of the sampled blocks with a recurrent neural network. The third model is inspired by the recurrent attention model (RAM), in which a reinforcement learning module is introduced to act as a controller for selecting the next block position based on the recurrent state. Experiments on our collected four-class arXiv paper dataset show that the three proposed models all perform well, and the RAM model achieves the best test accuracy with the least information.

]]>Algorithms doi: 10.3390/a11080110

Authors: David Völgyes Anne Catrine Trægde Martinsen Arne Stray-Pedersen Dag Waaler Marius Pedersen

Discretized image signals might have a lower dynamic range than the display. Because of this, false contours might appear when the image has the same pixel value for a larger region and the distance between pixel levels reaches the noticeable difference threshold. There have been several methods aimed at approximating the high bit depth of the original signal. Our method models a region with a bended plate model, which leads to the biharmonic equation. This method addresses several new aspects: the reconstruction of non-continuous regions when foreground objects split the area into separate regions; the incorporation of confidence about pixel levels, making the model tunable; and the method gives a physics-inspired way to handle local maximal/minimal regions. The solution of the biharmonic equation yields a smooth high-order signal approximation and handles the local maxima/minima problems.

]]>Algorithms doi: 10.3390/a11070108

Authors: Natalia Alekseeva Ivan Tanev Katsunori Shimohara

The most important characteristics of autonomous vehicles are their safety and their ability to adapt to various traffic situations and road conditions. In our research, we focused on the development of controllers for automated steering of a realistically simulated car in slippery road conditions. We comparatively investigated three implementations of such controllers: a proportional-derivative (PD) controller built in accordance with the canonical servo-control model of steering, a PID controller as an extension of the servo-control, and a controller designed heuristically via the most versatile evolutionary computing paradigm: genetic programming (GP). The experimental results suggest that the controller evolved via GP offers the best quality of control of the car in all of the tested slippery (rainy, snowy, and icy) road conditions.

]]>Algorithms doi: 10.3390/a11070107

Authors: Rui Yang Shuliang Xu Lin Feng

Extreme learning machine (ELM) is a single hidden layer feedforward neural network (SLFN). Because ELM has a fast speed for classification, it is widely applied in data stream classification tasks. In this paper, a new ensemble extreme learning machine is presented. Different from traditional ELM methods, a concept drift detection method is embedded; it uses online sequence learning strategy to handle gradual concept drift and uses updating classifier to deal with abrupt concept drift, so both gradual concept drift and abrupt concept drift can be detected in this paper. The experimental results showed the new ELM algorithm not only can improve the accuracy of classification result, but also can adapt to new concept in a short time.

]]>Algorithms doi: 10.3390/a11070106

Authors: Gerardo Navarro-Guerrero Yu Tang

The design of a fractional-order closed-loop model reference adaptive control (FOCMRAC) for anesthesia based on a fractional-order model (FOM) is proposed in the paper. This proposed model gets around many difficulties, namely, unknown parameters, lack of state measurement, inter and intra-patient variability, and variable time-delay, encountered in controller designs based on the PK/PD model commonly used for control of anesthesia, and allows to design a simple adaptive controller based on the Lyapunov analysis. Simulations illustrate the effectiveness and robustness of the proposed control.

]]>Algorithms doi: 10.3390/a11070105

Authors: Guillaume Damiand Aldo Gonzalez-Lorenzo Florence Zara Florent Dupont

We propose a new strategy for the parallelization of mesh processing algorithms. Our main contribution is the definition of distributed combinatorial maps (called n-dmaps), which allow us to represent the topology of big meshes by splitting them into independent parts. Our mathematical definition ensures the global consistency of the meshes at their interfaces. Thus, an n-dmap can be used to represent a mesh, to traverse it, or to modify it by using different mesh processing algorithms. Moreover, an nD mesh with a huge number of elements can be considered, which is not possible with a sequential approach and a regular data structure. We illustrate the interest of our solution by presenting a parallel adaptive subdivision method of a 3D hexahedral mesh, implemented in a distributed version. We report space and time performance results that show the interest of our approach for parallel processing of huge meshes.

]]>Algorithms doi: 10.3390/a11070104

Authors: Igor Gribanov Rocky Taylor Robert Sarracino

Computation of the distance between point and triangle in 3D is a common task in numerical analysis. The input values of the algorithm are coordinates of three points of the triangle and one point from which the distance is determined. An existing algorithm is extended to compute the gradient and the Hessian of that distance with respect to coordinates of involved points. Derivation of exact expressions for gradient and Hessian is presented, and numerical accuracy is evaluated for various cases. The algorithm has O(1) time and space complexity. The included open-source code may be used in applications where derivatives of point-triangle distance are required.

]]>Algorithms doi: 10.3390/a11070103

Authors: Jocelyn Sabatier

This paper analyses algorithms currently found in the literature for the approximation of fractional order models and based on recursive pole and zero distributions. The analysis focuses on the sub-optimality of the approximations obtained and stability issues that may appear after approximation depending on the pole location of the initial fractional order model. Solutions are proposed to reduce this sub-optimality and to avoid stability issues.

]]>Algorithms doi: 10.3390/a11070102

Authors: Tin-Chih Toly Chen Cheng-Li Liu Hong-Dar Lin

Artificial neural networks (ANNs) have been extensively applied to a wide range of disciplines, such as system identification and control, decision making, pattern recognition, medical diagnosis, finance, data mining, visualization, and others. With advances in computing and networking technologies, more complicated forms of ANNs are expected to emerge, requiring the design of advanced learning algorithms. This Special Issue is intended to provide technical details of the construction and training of advanced ANNs.

]]>Algorithms doi: 10.3390/a11070101

Authors: Bachir Bourouba Samir Ladaci

In this paper, a new adaptive fuzzy sliding mode control (AFSMC) design strategy is proposed for the control of a special class of three-dimensional fractional order chaotic systems with uncertainties and external disturbance. The design methodology is developed in two stages: first, an adaptive sliding mode control law is proposed for the class of fractional order chaotic systems without uncertainties, and then a fuzzy logic system is used to estimate the control compensation effort to be added in the case of uncertainties on the system&rsquo;s model. Based on the Lyapunov theory, the stability analysis of both control laws is provided with elimination of the chattering action in the control signal. The developed control scheme is simple to implement and the overall control scheme guarantees the global asymptotic stability in the Lyapunov sense if all the involved signals are uniformly bounded. In the present work, simulation studies on fractional-order Chen chaotic systems are carried out to show the efficiency of the proposed fractional adaptive controllers.

]]>Algorithms doi: 10.3390/a11070100

Authors: Maxim A. Dulebenets Masoud Kavoosi Olumide Abioye Junayed Pasha

Since ancient times, maritime transportation has played a very important role for the global trade and economy of many countries. The volumes of all major types of cargo, which are transported by vessels, has substantially increased in recent years. Considering a rapid growth of waterborne trade, marine container terminal operators should focus on upgrading the existing terminal infrastructure and improving operations planning. This study aims to assist marine container terminal operators with improving the seaside operations and primarily focuses on the berth scheduling problem. The problem is formulated as a mixed integer linear programming model, minimizing the total weighted vessel turnaround time and the total weighted vessel late departures. A self-adaptive Evolutionary Algorithm is proposed to solve the problem, where the crossover and mutation probabilities are encoded in the chromosomes. Numerical experiments are conducted to evaluate performance of the developed solution algorithm against the alternative Evolutionary Algorithms, which rely on the deterministic parameter control, adaptive parameter control, and parameter tuning strategies, respectively. Results indicate that all the considered solution algorithms demonstrate a relatively low variability in terms of the objective function values at termination from one replication to another and can maintain the adequate population diversity. However, application of the self-adaptive parameter control strategy substantially improves the objective function values at termination without a significant impact on the computational time.

]]>Algorithms doi: 10.3390/a11070099

Authors: Jiao Feng Xiaofei Zhang Peng Li Dongshun Hu

For Massive multiple-input multiple output (MIMO) systems, many algorithms have been proposed for detecting spatially multiplexed signals, such as reactive tabu search (RTS), minimum mean square error (MMSE), etc. As a heuristic neighborhood search algorithm, RTS is particularly suitable for signal detection in systems with large number of antennas. In this paper, we propose a strategy to reduce the neighborhood searching space of the traditional RTS algorithms. For this, we introduce a constellation constraints (CC) structure to determine whether including a candidate vector into the RTS searching neighborhood. By setting a pre-defined threshold on the symbol constellation, the Euclidean distance between the estimated signal and its nearest constellation points are calculated, and the threshold and distance are compared to separate the reliable estimated signal from unreliable ones. With this structure, the proposed CC-RTS algorithm may ignore a significant number of unnecessary candidates in the RTS neighborhood searching space and greatly reduce the computational complexity of the traditional RTS algorithm. Simulation results show that the BER performance of the proposed CC-RTS algorithm is very close to that of the traditional RTS algorithm, and with about 50% complexity reduction with the same signal-to-noise (SNR) ratio.

]]>Algorithms doi: 10.3390/a11070098

Authors: Li-Hsuan Chen Felix Reidl Peter Rossmanith Fernando Sánchez Villaamil

Treedepth is a well-established width measure which has recently seen a resurgence of interest. Since graphs of bounded treedepth are more restricted than graphs of bounded tree- or pathwidth, we are interested in the algorithmic utility of this additional structure. On the negative side, we show with a novel approach that the space consumption of any (single-pass) dynamic programming algorithm on treedepth decompositions of depth d cannot be bounded by (2&minus;ϵ)d&middot;logO(1)n for Vertex Cover, (3&minus;ϵ)d&middot;logO(1)n for 3-Coloring and (3&minus;ϵ)d&middot;logO(1)n for Dominating Set for any ϵ&gt;0. This formalizes the common intuition that dynamic programming algorithms on graph decompositions necessarily consume a lot of space and complements known results of the time-complexity of problems restricted to low-treewidth classes. We then show that treedepth lends itself to the design of branching algorithms. Specifically, we design two novel algorithms for Dominating Set on graphs of treedepth d: A pure branching algorithm that runs in time dO(d2)&middot;n and uses space O(d3logd+dlogn) and a hybrid of branching and dynamic programming that achieves a running time of O(3dlogd&middot;n) while using O(2ddlogd+dlogn) space.

]]>Algorithms doi: 10.3390/a11070097

Authors: Lin Tang Lin Liu Jianhou Gan

The events location and real-time computational performance of crowd scenes continuously challenge the field of video mining. In this paper, we address these two problems based on a regional topic model. In the process of video topic modeling, region topic model can simultaneously cluster motion words of video into motion topics, and the locations of motion into motion regions, where each motion topic associates with its region. Meanwhile, a hybrid stochastic variational Gibbs sampling algorithm is developed for inference of our region topic model, which has the ability of inferring in real time with massive video stream dataset. We evaluate our method on simulate and real datasets. The comparison with the Gibbs sampling algorithm shows the superiorities of proposed model and its online inference algorithm in terms of anomaly detection.

]]>Algorithms doi: 10.3390/a11070096

Authors: Asma Al-Saleh Mohamed El Bachir Menai

With advances in information technology, people face the problem of dealing with tremendous amounts of information and need ways to save time and effort by summarizing the most important and relevant information. Thus, automatic text summarization has become necessary to reduce the information overload. This article proposes a novel extractive graph-based approach to solve the multi-document summarization (MDS) problem. To optimize the coverage of information in the output summary, the problem is formulated as an orienteering problem and heuristically solved by an ant colony system algorithm. The performance of the implemented system (MDS-OP) was evaluated on DUC 2004 (Task 2) and MultiLing 2015 (MMS task) benchmark corpora using several ROUGE metrics, as well as other methods. Its comparison with the performances of 26 systems shows that MDS-OP achieved the best F-measure scores on both tasks in terms of ROUGE-1 and ROUGE-L (DUC 2004), ROUGE-SU4, and three other evaluation methods (MultiLing 2015). Overall, MDS-OP ranked among the best 3 systems.

]]>Algorithms doi: 10.3390/a11070095

Authors: Cristina I. Muresan Cosmin Copot Isabela Birs Robin De Keyser Steve Vanlanduit Clara M. Ionescu

Classical fractional order controller tuning techniques usually consider the frequency domain specifications (phase margin, gain crossover frequency, iso-damping) and are based on knowledge of a process model, as well as solving a system of nonlinear equations to determine the controller parameters. In this paper, a novel auto-tuning method is used to tune a fractional order PI controller. The advantages of the proposed auto-tuning method are two-fold: There is no need for a process model, neither to solve the system of nonlinear equations. The tuning is based on defining a forbidden region in the Nyquist plane using the phase margin requirement and determining the parameters of the fractional order controller such that the loop frequency response remains out of the forbidden region. Additionally, the final controller parameters are those that minimize the difference between the slope of the loop frequency response and the slope of the forbidden region border, to ensure the iso-damping property. To validate the proposed method, a case study has been used consisting of a pick and place movement of an UR10 robot. The experimental results, considering two different robot configurations, demonstrate that the designed fractional order PI controller is indeed robust.

]]>Algorithms doi: 10.3390/a11070094

Authors: Dongxu Wei Andong Wang Xiaoqin Feng Boyu Wang Bo Wang

Many tasks in computer vision suffer from missing values in tensor data, i.e., multi-way data array. The recently proposed tensor tubal nuclear norm (TNN) has shown superiority in imputing missing values in 3D visual data, like color images and videos. However, by interpreting in a circulant way, TNN only exploits tube (often carrying temporal/channel information) redundancy in a circulant way while preserving the row and column (often carrying spatial information) relationship. In this paper, a new tensor norm named the triple tubal nuclear norm (TriTNN) is proposed to simultaneously exploit tube, row and column redundancy in a circulant way by using a weighted sum of three TNNs. Thus, more spatial-temporal information can be mined. Further, a TriTNN-based tensor completion model with an ADMM solver is developed. Experiments on color images, videos and LiDAR datasets show the superiority of the proposed TriTNN against state-of-the-art nuclear norm-based tensor norms.

]]>Algorithms doi: 10.3390/a11070093

Authors: Bhadrachalam Chitturi Srijith Balachander Sandeep Satheesh Krithic Puthiyoppil

The computation of distances between strings has applications in molecular biology, music theory and pattern recognition. One such measure, called short reversal distance, has applications in evolutionary distance computation. It has been shown that this problem can be reduced to the computation of a maximum independent set on the corresponding graph that is constructed from the given input strings. The constructed graphs primarily fall into a class that we call layered graphs. In a layered graph, each layer refers to a subgraph containing, at most, some k vertices. The inter-layer edges are restricted to the vertices in adjacent layers. We study the MIS, MVC, MDS, MCV and MCD problems on layered graphs where MIS computes the maximum independent set; MVC computes the minimum vertex cover; MDS computes the minimum dominating set; MCV computes the minimum connected vertex cover; and MCD computes the minimum connected dominating set. MIS, MVC and MDS run in polynomial time if k=Θ(log|V|). MCV and MCD run in polynomial time ifk=O((log|V|)α), where α&lt;1. If k=Θ((log|V|)1+ϵ), for ϵ&gt;0, then MIS, MVC and MDS run in quasi-polynomial time. If k=Θ(log|V|), then MCV and MCD run in quasi-polynomial time.

]]>Algorithms doi: 10.3390/a11070092

Authors: Guoliang Yang Haitao Yi Chunhua Chai Bingxu Huang Yuna Zhang Zhe Chen

A topology structure based on boost three-level converters (BTL converters) and T-type three-level inverters for a direct-drive wind turbine in a wind power generation system is proposed. In this structure, the generator-side control can be realized by the boost-TL converter. Compared with the conventional boost converter, the boost-TL converter has a low inductor current ripple, which reduces the torque ripple of the generator, increases the converter&rsquo;s capacity, and minimizes switching losses. The boost-TL converter can boost the DC output from the rectifier at low speeds. The principles of the boost-TL converter and the T-type three-level inverter are separately introduced. Based on the cascaded structure of the proposed BTL converter and three-level inverter, a model predictive current control (MPCC) method is adopted, and the optimization of the MPCC is presented. The prediction model is derived, and the simulation and experimental research are carried out. The results show that the algorithm based on the proposed cascaded structure is feasible and superior.

]]>Algorithms doi: 10.3390/a11070091

Authors: Christian Himpe

System Gramian matrices are a well-known encoding for properties of input-output systems such as controllability, observability or minimality. These so-called system Gramians were developed in linear system theory for applications such as model order reduction of control systems. Empirical Gramians are an extension to the system Gramians for parametric and nonlinear systems as well as a data-driven method of computation. The empirical Gramian framework - emgr - implements the empirical Gramians in a uniform and configurable manner, with applications such as Gramian-based (nonlinear) model reduction, decentralized control, sensitivity analysis, parameter identification and combined state and parameter reduction.

]]>Algorithms doi: 10.3390/a11070090

Authors: Kai Liu YangQuan Chen Paweł D. Domański Xi Zhang

The significant task for control performance assessment (CPA) is to review and evaluate the performance of the control system. The control system in the semiconductor industry exhibits a complex dynamic behavior, which is hard to analyze. This paper investigates the interesting crossover properties of Hurst exponent estimations and proposes a novel method for feature extraction of the nonlinear multi-input multi-output (MIMO) systems. At first, coupled data from real industry are analyzed by multifractal detrended fluctuation analysis (MFDFA) and the resultant multifractal spectrum is obtained. Secondly, the crossover points with spline fit in the scale-law curve are located and then employed to segment the entire scale-law curve into several different scaling regions, in which a single Hurst exponent can be estimated. Thirdly, to further ascertain the origin of the multifractality of control signals, the generalized Hurst exponents of the original series are compared with shuffled data. At last, non-Gaussian statistical properties, multifractal properties and Hurst exponents of the process control variables are derived and compared with different sets of tuning parameters. The results have shown that CPA of the MIMO system can be better employed with the help of fractional order signal processing (FOSP).

]]>Algorithms doi: 10.3390/a11070089

Authors: Qing Li Steven Y. Liang

Timely maintenance and accurate fault prediction of rotating machinery are essential for ensuring system availability, minimizing downtime, and contributing to sustainable production. This paper proposes a novel approach based on long-range dependence (LRD) and particle filter (PF) for degradation trend prediction of rotating machinery, taking the rolling bearing as an example. In this work, the degradation prediction is evaluated based on two health indicators time series; i.e., equivalent vibration severity (EVI) time series and kurtosis time series. Specifically, the degradation trend prediction issues here addressed have the following two distinctive features: (i) EVI time series with weak LRD property and (ii) kurtosis time series with sharp transition points (STPs) in the forecasted region. The core idea is that the parameters distribution of the LRD model can be updated recursively by the particle filter algorithm; i.e., the parameters degradation of the LRD model are restrained, and thus the prognostic results could be generated real-time, wherein the initial LRD model is designed randomly. The prediction results demonstrate that the significant improvements in prediction accuracy are obtained with the proposed method compared to some state-of-the-art approaches such as the autoregressive&ndash;moving-average (ARMA) model and the fractional order characteristic (FOC) model, etc.

]]>Algorithms doi: 10.3390/a11060088

Authors: Qingge Ji Wenjie He Jie Huang Yankui Sun

We present an automatic method based on transfer learning for the identification of dry age-related macular degeneration (AMD) and diabetic macular edema (DME) from retinal optical coherence tomography (OCT) images. The algorithm aims to improve the classification performance of retinal OCT images and shorten the training time. Firstly, we remove the last several layers from the pre-trained Inception V3 model and regard the remaining part as a fixed feature extractor. Then, the features are used as input of a convolutional neural network (CNN) designed to learn the feature space shifts. The experimental results on two different retinal OCT images datasets demonstrate the effectiveness of the proposed method.

]]>Algorithms doi: 10.3390/a11060087

Authors: Frank Werner Larysa Burtseva Yuri N. Sotskov

This special issue of Algorithms is devoted to the development of scheduling algorithms based on innovative approaches for solving hard scheduling problems either exactly or approximately. Submissions were welcome both for traditional scheduling problems as well as for new practical applications. The main topics include sequencing and scheduling with additional constraints (setup times or costs, precedence constraints, resource constraints, and batch production environment) and production planning and scheduling problems arising in real-world applications.

]]>Algorithms doi: 10.3390/a11060086

Authors: Christer Dalen David Di Ruscio

A method for tuning PI controller parameters, a prescribed maximum time delay error or a relative time delay error is presented. The method is based on integrator plus time delay models. The integral time constant is linear in the relative time delay error, and the proportional constant is seen inversely proportional to the relative time delay error. The keystone in the method is the method product parameter, i.e., the product of the PI controller proportional constant, the integral time constant, and the integrator plus time delay model, velocity gain. The method product parameter is found to be constant for various PI controller tuning methods. Optimal suggestions are given for choosing the method product parameter, i.e., optimal such that the integrated absolute error or, more interestingly, the Pareto performance objective (i.e., integrated absolute error for combined step changes in output and input disturbances) is minimised. Variants of the presented tuning method are demonstrated for tuning PI controllers for motivated (possible) higher order process model examples, i.e., the presented method is combined with the model reduction step (process–reaction curve) in Ziegler–Nichols.

]]>Algorithms doi: 10.3390/a11060085

Authors: Yong-Hong Lan Zhe-Min Cui

This paper presents a second order P-type iterative learning control (ILC) scheme with initial state learning for a class of fractional order linear distributed parameter systems. First, by analyzing the control and learning processes, a discrete system for P-type ILC is established, and the ILC design problem is then converted to a stability problem for such a discrete system. Next, a sufficient condition for the convergence of the control input and the tracking errors is obtained by introducing a new norm and using the generalized Gronwall inequality, which is less conservative than the existing one. Finally, the validity of the proposed method is verified by a numerical example.

]]>Algorithms doi: 10.3390/a11060084

Authors: Alessandro Epasto Eli Upfal

Covering the edges of a bipartite graph by a minimum set of bipartite complete graphs (bicliques) is a basic graph theoretic problem, with numerous applications. In particular, it is used to characterize parsimonious models of a set of observations (each biclique corresponds to a factor or feature that relates the observations in the two sets of nodes connected by the biclique). The decision version of the minimum biclique cover problem is NP-Complete, and unless P=NP, the cover size cannot be approximated in general within less than a sub-linear factor of the number of nodes (or edges) in the graph. In this work, we consider two natural restrictions to the problem, motivated by practical applications. In the first case, we restrict the number of bicliques a node can belong to. We show that when this number is at least 5, the problem is still NP-hard. In contrast, we show that when nodes belong to no more than two bicliques, the problem has efficient approximations. The second model we consider corresponds to observing a set of independent samples from an unknown model, governed by a possibly large number of factors. The model is defined by a bipartite graph G=(L,R,E), where each node in L is assigned to an arbitrary subset of up to a constant f factors, while the nodes in R (the independent observations) are assigned to random subsets of the set of k factors where k can grow with size of the graph. We show that this practical version of the biclique cover problem is amenable to efficient approximations.

]]>Algorithms doi: 10.3390/a11060083

Authors: David Pang Tomohiko Igasaki

Long-term heart rate variability (HRV) analysis is useful as a noninvasive technique for autonomic nervous system activity assessment. It provides a method for assessing many physiological and pathological factors that modulate the normal heartbeat. The performance of HRV analysis systems heavily depends on a reliable and accurate detection of the R peak of the QRS complex. Ectopic beats caused by misdetection or arrhythmic events can introduce bias into HRV results, resulting in significant problems in their interpretation. This study presents a novel method for long-term detection of normal R peaks (which represent the normal heartbeat in electrocardiographic signals), intended specifically for HRV analysis. The very low computational complexity of the proposed method, which combines and exploits the advantages of syntactical and statistical approaches, enables real-time applications. The approach was validated using the Massachusetts Institute of Technology&ndash;Beth Israel Hospital Normal Sinus Rhythm and the Fantasia database, and has a sensitivity, positive predictivity, detection error rate, and accuracy of 99.998, 99.999, 0.003, and 99.996%, respectively.

]]>Algorithms doi: 10.3390/a11060082

Authors: Yi Wei Yaokun Yue

Since the traditional fault diagnosis method of the marine fuel system has a low accuracy of identification, the algorithm solution can easily fall into local optimum, and they are not fit for the research on the fault diagnosis of a marine fuel system. Hence, a fault diagnosis method for a marine fuel system based on the SaDE-ELM algorithm is proposed. First, the parameters of initializing extreme learning machine are adopted by a differential evolution algorithm. Second, the fault diagnosis of the marine fuel system is realized by the fault diagnosis model corresponding to the state training of marine fuel system. Based on the obtained fault data of a marine fuel system, the proposed method is verified. The experimental results show that this method produces higher recognition accuracy and faster recognition speed that are superior to the traditional BP neural network, SVM support vector machine diagnosis algorithm, and the un-optimized extreme learning machine algorithm. The results have important significance relevant to fault diagnosis for a marine fuel system. The algorithm based on SaDE-ELM is an effective and practical method of fault diagnosis for a marine fuel system.

]]>Algorithms doi: 10.3390/a11060081

Authors: Yossi Peretz

A randomized algorithm is suggested for the syntheses of optimal PID controllers for MIMO coupled systems, where the optimality is with respect to the H &infin; -norm, the H 2 -norm and the LQR functional, with possible system-performance specifications defined by regional pole-placement. Other notions of optimality (e.g., mixed H 2 / H &infin; design, controller norm or controller sparsity) can be handled similarly with the suggested algorithm. The suggested method is direct and thus can be applied to continuous-time systems as well as to discrete-time systems with the obvious minor changes. The presented algorithm is a randomized algorithm, which has a proof of convergence (in probability) to a global optimum.

]]>Algorithms doi: 10.3390/a11060080

Authors: Nodari Vakhania

We study a scheduling problem in which jobs with release times and due dates are to be processed on a single machine. With the primary objective to minimize the maximum job lateness, the problem is strongly NP-hard. We describe a general algorithmic scheme to minimize the maximum job lateness, with the secondary objective to minimize the maximum job completion time. The problem of finding the Pareto-optimal set of feasible solutions with these two objective criteria is strongly NP-hard. We give the dominance properties and conditions when the Pareto-optimal set can be formed in polynomial time. These properties, together with our general framework, provide the theoretical background, so that the basic framework can be expanded to (exponential-time) implicit enumeration algorithms and polynomial-time approximation algorithms (generating the Pareto sub-optimal frontier with a fair balance between the two objectives). Some available in the literature experimental results confirm the practical efficiency of the proposed framework.

]]>Algorithms doi: 10.3390/a11060079

Authors: Yongming Bian Meng Yang Xuying Fan Yuchao Liu

Automatic fire detection, which can detect and raise the alarm for fire early, is expected to help reduce the loss of life and property as much as possible. Due to its advantages over traditional methods, image processing technology has been applied gradually in fire detection. In this paper, a novel algorithm is proposed to achieve fire image detection, combined with Tchebichef (sometimes referred to as Chebyshev) moment invariants (TMIs) and particle swarm optimization-support vector machine (PSO-SVM). According to the correlation between geometric moments and Tchebichef moments, the translation, rotation, and scaling (TRS) invariants of Tchebichef moments are obtained first. Then, the TMIs of candidate images are calculated to construct feature vectors. To gain the best detection performance, a PSO-SVM model is proposed, where the kernel parameter and penalty factor of support vector machine (SVM) are optimized by particle swarm optimization (PSO). Then, the PSO-SVM model is utilized to identify the fire images. Compared with algorithms based on Hu moment invariants (HMIs) and Zernike moment invariants (ZMIs), the experimental results show that the proposed algorithm can improve the detection accuracy, achieving the highest detection rate of 98.18%. Moreover, it still exhibits the best performance even if the size of the training sample set is small and the images are transformed by TRS.

]]>Algorithms doi: 10.3390/a11060078

Authors: Bao Pang Yong Song Chengjin Zhang Hongling Wang Runtao Yang

Artificial bee colony (ABC) algorithm, a novel category of bionic intelligent optimization algorithm, was achieved for solving complex nonlinear optimization problems. Previous studies have shown that ABC algorithm is competitive to other biological-inspired optimization algorithms, but there still exist several insufficiencies due to the inefficient solution search equation (SSE), which does well in exploration but poorly in exploitation. To improve accuracy of the solutions, this paper proposes a modified ABC algorithm based on the self-learning mechanism (SLABC) with five SSEs as the candidate operator pool; among them, one is good at exploration and two of them are good at exploitation; another SSE intends to balance exploration and exploitation; moreover, the last SSE with L&eacute;vy flight step-size which can generate smaller step-size with high frequency and bigger step-size occasionally not only can balance exploration and exploitation but also possesses the ability to escape from the local optimum. This paper proposes a simple self-learning mechanism, wherein the SSE is selected according to the previous success ratio in generating promising solutions at each iteration. Experiments on a set of 9 benchmark functions are carried out with the purpose of evaluating the performance of the proposed method. The experimental results illustrated that the SLABC algorithm achieves significant improvement compared with other competitive algorithms.

]]>Algorithms doi: 10.3390/a11050077

Authors: Wei Xu Yi Li Jinghong Miao Jiaxiang Zhao Xin Gao

Cosine-modulated filter banks play a major role in digital signal processing. Sparse FIR filter banks have lower implementation complexity than full filter banks, while keeping a good performance level. This paper presents a fast design paradigm for sparse nearly perfect-reconstruction (NPR) cosine-modulated filter banks. First, an approximation function is introduced to reduce the non-convex quadratically constrained optimization problem to a linearly constrained optimization problem. Then, the desired sparse linear phase FIR prototype filter is derived through the orthogonal matching pursuit (OMP) performed under the weighted l 2 norm. The simulation results demonstrate that the proposed scheme is an effective paradigm to design sparse NPR cosine-modulated filter banks.

]]>Algorithms doi: 10.3390/a11050076

Authors: Alexander Yu. Drozdov Andrei Tchernykh Sergey V. Novikov Victor E. Vladislavlev Raul Rivera-Rodriguez

We address image processing workflow scheduling problems on a multicore digital signal processor cluster. We present an experimental study of scheduling strategies that include task labeling, prioritization, resource selection, and digital signal processor scheduling. We apply these strategies in the context of executing the Ligo and Montage applications. To provide effective guidance in choosing a good strategy, we present a joint analysis of three conflicting goals based on performance degradation. A case study is given, and experimental results demonstrate that a pessimistic scheduling approach provides the best optimization criteria trade-offs. The Pessimistic Heterogeneous Earliest Finish Time scheduling algorithm performs well in different scenarios with a variety of workloads and cluster configurations.

]]>Algorithms doi: 10.3390/a11050075

Authors: Guohui Wang Yuanbo Chu

3D shape reconstruction from images has been an important topic in the field of robot vision. Shape-From-Shading (SFS) is a classical method for determining the shape of a 3D surface from a one intensity image. The Lambertian reflectance is a fundamental assumption in conventional SFS approaches. Unfortunately, when applied to characterize the reflection attribute of the diffuse reflection, the Lambertian model is tested to be inexact. In this paper, we present a new SFS approach for 3D reconstruction of diffuse surfaces whose reflection attribute is approximated by the Oren&ndash;Nayar reflection model. The partial differential Image Irradiance Equation (IIR) is set up with this model under a single distant point light source and an orthographic camera projection whose direction coincides with the light source. Then, the IIR is converted into an eikonal equation by solving a quadratic equation that includes the 3D surface shape. The viscosity solution of the resulting eikonal equation is approximated by using the high-order Godunov-based scheme that is accelerated by means of an alternating sweeping strategy. We conduct the experiments on synthetic and real-world images, and the experimental results illustrate the effectiveness of the presented approach.

]]>Algorithms doi: 10.3390/a11050074

Authors: Petr Stodola

This article deals with the modified Multi-Depot Vehicle Routing Problem (MDVRP). The modification consists of altering the optimization criterion. The optimization criterion of the standard MDVRP is to minimize the total sum of routes of all vehicles, whereas the criterion of modified MDVRP (M-MDVRP) is to minimize the longest route of all vehicles, i.e., the time to conduct the routing operation is as short as possible. For this problem, a metaheuristic algorithm&mdash;based on the Ant Colony Optimization (ACO) theory and developed by the author for solving the classic MDVRP instances&mdash;has been modified and adapted for M-MDVRP. In this article, an additional deterministic optimization process which further enhances the original ACO algorithm has been proposed. For evaluation of results, Cordeau&rsquo;s benchmark instances are used.

]]>Algorithms doi: 10.3390/a11050073

Authors: Hendrik Santosa Xuetong Zhai Frank Fishburn Theodore Huppert

Functional near-infrared spectroscopy (fNIRS) is a noninvasive neuroimaging technique that uses low-levels of light (650&ndash;900 nm) to measure changes in cerebral blood volume and oxygenation. Over the last several decades, this technique has been utilized in a growing number of functional and resting-state brain studies. The lower operation cost, portability, and versatility of this method make it an alternative to methods such as functional magnetic resonance imaging for studies in pediatric and special populations and for studies without the confining limitations of a supine and motionless acquisition setup. However, the analysis of fNIRS data poses several challenges stemming from the unique physics of the technique, the unique statistical properties of data, and the growing diversity of non-traditional experimental designs being utilized in studies due to the flexibility of this technology. For these reasons, specific analysis methods for this technology must be developed. In this paper, we introduce the NIRS Brain AnalyzIR toolbox as an open-source Matlab-based analysis package for fNIRS data management, pre-processing, and first- and second-level (i.e., single subject and group-level) statistical analysis. Here, we describe the basic architectural format of this toolbox, which is based on the object-oriented programming paradigm. We also detail the algorithms for several of the major components of the toolbox including statistical analysis, probe registration, image reconstruction, and region-of-interest based statistics.

]]>Algorithms doi: 10.3390/a11050072

Authors: Yixuan Ren Tao Ye Mengxing Huang Siling Feng

In the field of investment, how to construct a suitable portfolio based on historical data is still an important issue. The second-order stochastic dominant constraint is a branch of the stochastic dominant constraint theory. However, only considering the second-order stochastic dominant constraints does not conform to the investment environment under realistic conditions. Therefore, we added a series of constraints into basic portfolio optimization model, which reflect the realistic investment environment, such as skewness and kurtosis. In addition, we consider two kinds of risk measures: conditional value at risk and value at risk. Most important of all, in this paper, we introduce Gray Wolf Optimization (GWO) algorithm into portfolio optimization model, which simulates the gray wolf’s social hierarchy and predatory behavior. In the numerical experiments, we compare the GWO algorithm with Particle Swarm Optimization (PSO) algorithm and Genetic Algorithm (GA). The experimental results show that GWO algorithm not only shows better optimization ability and optimization efficiency, but also the portfolio optimized by GWO algorithm has a better performance than FTSE100 index, which prove that GWO algorithm has a great potential in portfolio optimization.

]]>Algorithms doi: 10.3390/a11050071

Authors: Hui Hu Zhaoquan Cai Song Hu Yingxue Cai Jia Chen Sibo Huang

Inspired by the migration behavior of monarch butterflies in nature, Wang et al. proposed a novel, promising, intelligent swarm-based algorithm, monarch butterfly optimization (MBO), for tackling global optimization problems. In the basic MBO algorithm, the butterflies in land 1 (subpopulation 1) and land 2 (subpopulation 2) are calculated according to the parameter p, which is unchanged during the entire optimization process. In our present work, a self-adaptive strategy is introduced to dynamically adjust the butterflies in land 1 and 2. Accordingly, the population size in subpopulation 1 and 2 are dynamically changed as the algorithm evolves in a linear way. After introducing the concept of a self-adaptive strategy, an improved MBO algorithm, called monarch butterfly optimization with self-adaptive population (SPMBO), is put forward. In SPMBO, only generated individuals who are better than before can be accepted as new individuals for the next generations in the migration operation. Finally, the proposed SPMBO algorithm is benchmarked by thirteen standard test functions with dimensions of 30 and 60. The experimental results indicate that the search ability of the proposed SPMBO approach significantly outperforms the basic MBO algorithm on most test functions. This also implies the self-adaptive strategy is an effective way to improve the performance of the basic MBO algorithm.

]]>Algorithms doi: 10.3390/a11050070

Authors: Samuel Montero-Hernandez Felipe Orihuela-Espina Luis Sucar Paola Pinti Antonia Hamilton Paul Burgess Ilias Tachtsidis

Functional Near InfraRed Spectroscopy (fNIRS) connectivity analysis is often performed using the measured oxy-haemoglobin (HbO2) signal, while the deoxy-haemoglobin (HHb) is largely ignored. The in-common information of the connectivity networks of both HbO2 and HHb is not regularly reported, or worse, assumed to be similar. Here we describe a methodology that allows the estimation of the symmetry between the functional connectivity (FC) networks of HbO2 and HHb and propose a differential symmetry index (DSI) indicative of the in-common physiological information. Our hypothesis is that the symmetry between FC networks associated with HbO2 and HHb is above what should be expected from random networks. FC analysis was done in fNIRS data collected from six freely-moving healthy volunteers over 16 locations on the prefrontal cortex during a real-world task in an out-of-the-lab environment. In addition, systemic data including breathing rate (BR) and heart rate (HR) were also synchronously collected and used within the FC analysis. FC networks for HbO2 and HHb were established independently using a Bayesian networks analysis. The DSI between both haemoglobin (Hb) networks with and without systemic influence was calculated. The relationship between the symmetry of HbO2 and HHb networks, including the segregational and integrational characteristics of the networks (modularity and global efficiency respectively) were further described. Consideration of systemic information increases the path lengths of the connectivity networks by 3%. Sparse networks exhibited higher asymmetry than dense networks. Importantly, our experimental connectivity networks symmetry between HbO2 and HHb departs from random (t-test: t(509) = 26.39, p &lt; 0.0001). The DSI distribution suggests a threshold of 0.2 to decide whether both HbO2 and HHb FC networks ought to be studied. For sparse FC networks, analysis of both haemoglobin species is strongly recommended. Our DSI can provide a quantifiable guideline for deciding whether to proceed with single or both Hb networks in FC analysis.

]]>Algorithms doi: 10.3390/a11050069

Authors: Lucia Cassettari Melissa Demartini Roberto Mosca Roberto Revetria Flavio Tonelli

The Vehicle Routing Problem (VRP) is one of the most optimized tasks studied and it is implemented in a huge variety of industrial applications. The objective is to design a set of minimum cost paths for each vehicle in order to serve a given set of customers. Our attention is focused on a variant of VRP, the capacitated vehicle routing problem when applied to natural gas distribution networks. Managing natural gas distribution networks includes facing a variety of decisions ranging from human resources and material resources to facilities, infrastructures, and carriers. Despite the numerous papers available on vehicle routing problem, there are only a few that study and analyze the problems occurring in capillary distribution operations such as those found in a metropolitan area. Therefore, this work introduces a new algorithm based on the Saving Algorithm heuristic approach which aims to solve a Capacitated Vehicle Routing Problem with time and distance constraints. This joint algorithm minimizes the transportation costs and maximizes the workload according to customer demand within the constraints of a time window. Results from a real case study in a natural gas distribution network demonstrates the effectiveness of the approach.

]]>Algorithms doi: 10.3390/a11050068

Authors: Victor Yaurima-Basaldua Andrei Tchernykh Francisco Villalobos-Rodríguez Ricardo Salomon-Torres

We address a scheduling problem in an actual environment of the tortilla industry. Since the problem is NP hard, we focus on suboptimal scheduling solutions. We concentrate on a complex multistage, multiproduct, multimachine, and batch production environment considering completion time and energy consumption optimization criteria. The production of wheat-based and corn-based tortillas of different styles is considered. The proposed bi-objective algorithm is based on the known Nondominated Sorting Genetic Algorithm II (NSGA-II). To tune it up, we apply statistical analysis of multifactorial variance. A branch and bound algorithm is used to assert obtained performance. We show that the proposed algorithms can be efficiently used in a real production environment. The mono-objective and bi-objective analyses provide a good compromise between saving energy and efficiency. To demonstrate the practical relevance of the results, we examine our solution on real data. We find that it can save 48% of production time and 47% of electricity consumption over the actual production.

]]>Algorithms doi: 10.3390/a11050067

Authors: Lia M. Hocke Ibukunoluwa K. Oni Chris C. Duszynski Alex V. Corrigan Blaise deB. Frederick Jeff F. Dunn

With the rapid increase in new fNIRS users employing commercial software, there is a concern that many studies are biased by suboptimal processing methods. The purpose of this study is to provide a visual reference showing the effects of different processing methods, to help inform researchers in setting up and evaluating a processing pipeline. We show the significant impact of pre- and post-processing choices and stress again how important it is to combine data from both hemoglobin species in order to make accurate inferences about the activation site.

]]>Algorithms doi: 10.3390/a11050066

Authors: Yuri N. Sotskov Natalja G. Egorova

We consider a single machine scheduling problem with uncertain durations of the given jobs. The objective function is minimizing the sum of the job completion times. We apply the stability approach to the considered uncertain scheduling problem using a relative perimeter of the optimality box as a stability measure of the optimal job permutation. We investigated properties of the optimality box and developed algorithms for constructing job permutations that have the largest relative perimeters of the optimality box. Computational results for constructing such permutations showed that they provided the average error less than 0 . 74 % for the solved uncertain problems.

]]>Algorithms doi: 10.3390/a11050065

Authors: Pengzhan Chen Zhiqiang He Chuanxi Chen Jiahong Xu

We developed a novel control strategy of speed servo systems based on deep reinforcement learning. The control parameters of speed servo systems are difficult to regulate for practical applications, and problems of moment disturbance and inertia mutation occur during the operation process. A class of reinforcement learning agents for speed servo systems is designed based on the deep deterministic policy gradient algorithm. The agents are trained by a significant number of system data. After learning completion, they can automatically adjust the control parameters of servo systems and compensate for current online. Consequently, a servo system can always maintain good control performance. Numerous experiments are conducted to verify the proposed control strategy. Results show that the proposed method can achieve proportional&ndash;integral&ndash;derivative automatic tuning and effectively overcome the effects of inertia mutation and torque disturbance.

]]>Algorithms doi: 10.3390/a11050064

Authors: Ming Lan Fu Hao Wang Bao Fu Fang

This paper focuses on the rational distribution of task utilities in coalition skill games, which is a restricted form of coalition game, where each service agent has a set of skills and each task agent needs a set of skills in order to be completed. These two types of agents are assumed to be self-interested. Given the task selection strategy of service agents, the utility distribution strategies of task agents play an important role in improving their individual revenues and system total revenue. The problem that needs to be resolved is how to design the task selection strategies of the service agents and the utility distribution strategies of the task agents to make the self-interested decisions improve the system whole performance. However, to the best of our knowledge, this problem has been the topic of very few studies and has not been properly addressed. To address this problem, a task allocation algorithm for self-interested agents in a coalition skill game is proposed, it distributes the utilities of tasks to the needed skills according to the powers of the service agents that possess the corresponding skills. The final simulation results verify the effectiveness of the algorithm.

]]>Algorithms doi: 10.3390/a11050063

Authors: Minghui Shao Yan Song Biao Wu Yanjie Chang

Supplier selection is an important decision-making link in bidding activity. When overall scores of several suppliers are similar, it is hard to obtain an accurate ranking of these suppliers. Applying the Diversity Factors Method (Diversity Factors Method, DFM) may lead to over correction of weights, which would degrade the capability of indexes to reflect the importance. A Limited Diversity Factors Method (Limited Diversity Factors Method, LDFM) based on entropy is presented in this paper in order to adjust the weights, in order to relieve the over correction in DFM and to improve the capability of identification of indexes in supplier selection. An example of salvage ship bidding demonstrates the advantages of the LDFM, in which the raking of overall scores of suppliers is more accurate.

]]>Algorithms doi: 10.3390/a11050062

Authors: Minghua Xie Decheng Wang Lili Xie

Support Vector Regression (SVR), which converts the original low-dimensional problem to a high-dimensional kernel space linear problem by introducing kernel functions, has been successfully applied in system modeling. Regarding the classical SVR algorithm, the value of the features has been taken into account, while its contribution to the model output is omitted. Therefore, the construction of the kernel space may not be reasonable. In the paper, a Feature-Weighted SVR (FW-SVR) is presented. The range of the feature is matched with its contribution by properly assigning the weight of the input features in data pre-processing. FW-SVR optimizes the distribution of the sample points in the kernel space to make the minimizing of the structural risk more reasonable. Four synthetic datasets and seven real datasets are applied. A superior generalization ability is obtained by the proposed method.

]]>Algorithms doi: 10.3390/a11050061

Authors: Edward Talmage Jennifer L. Welch

In the quest for higher-performance shared data structures, weakening consistency conditions and relaxing the sequential specifications of data types are two of the primary tools available in the literature today. In this paper, we show that these two approaches are in many cases different ways to specify the same sets of allowed concurrent behaviors of a given shared data object. This equivalence allows us to use whichever description is clearer, simpler, or easier to achieve equivalent guarantees. Specifically, for three common data type relaxations, we define consistency conditions such that the combination of the new consistency condition and an unrelaxed type allows the same behaviors as Linearizability and the relaxed version of the data type. Conversely, for the consistency condition k-Atomicity, we define a new data type relaxation such that the behaviors allowed by the relaxed version of a data type, combined with Linearizability, are the same as those allowed by k-Atomicity and the original type. As an example of the possibilities opened by our new equivalence, we use standard techniques from the literature on consistency conditions to prove that the three data type relaxations we consider are not comparable to one another or to several similar known conditions. Finally, we show a particular class of data types where one of our newly-defined consistency conditions is comparable to, and stronger than, one of the known consistency conditions we consider.

]]>Algorithms doi: 10.3390/a11050060

Authors: Molin Sun Zhongyi Zheng

Uncertainty analysis is considered to be a necessary step in the process of vessel traffic risk assessment. The purpose of this study is to propose the uncertainty analysis algorithm which can be used to investigate the reliability of the risk assessment result. Probability and possibility distributions are used to quantify the two types of uncertainty identified in the risk assessment process. In addition, the algorithm for appropriate time window selection is chosen by considering the uncertainty of vessel traffic accident occurrence and the variation trend of the vessel traffic risk caused by maritime rules becoming operative. Vessel traffic accident data from the United Kingdom&rsquo;s marine accident investigation branch are used for the case study. Based on a comparison with the common method of estimating the vessel traffic risk and the algorithm for uncertainty quantification without considering the time window selection, the availability of the proposed algorithms is verified, which can provide guidance for vessel traffic risk management.

]]>Algorithms doi: 10.3390/a11050059

Authors: Muhammad Akram Nabeela Ishfaq Sidra Sayed Florentin Smarandache

Rough set theory and neutrosophic set theory are mathematical models to deal with incomplete and vague information. These two theories can be combined into a framework for modeling and processing incomplete information in information systems. Thus, the neutrosophic rough set hybrid model gives more precision, flexibility and compatibility to the system as compared to the classic and fuzzy models. In this research study, we develop neutrosophic rough digraphs based on the neutrosophic rough hybrid model. Moreover, we discuss regular neutrosophic rough digraphs, and we solve decision-making problems by using our proposed hybrid model. Finally, we give a comparison analysis of two hybrid models, namely, neutrosophic rough digraphs and rough neutrosophic digraphs.

]]>Algorithms doi: 10.3390/a11050058

Authors: Volker Turau

The analysis of self-stabilizing algorithms is often limited to the worst case stabilization time starting from an arbitrary state, i.e., a state resulting from a sequence of faults. Considering the fact that these algorithms are intended to provide fault tolerance in the long run, this is not the most relevant metric. A common situation is that a running system is an a legitimate state when hit by a single fault. This event has a much higher probability than multiple concurrent faults. Therefore, the worst case time to recover from a single fault is more relevant than the recovery time from a large number of faults. This paper presents techniques to derive upper bounds for the mean time to recover from a single fault for self-stabilizing algorithms based on Markov chains in combination with lumping. To illustrate the applicability of the techniques they are applied to a new self-stabilizing coloring algorithm.

]]>Algorithms doi: 10.3390/a11050057

Authors: Boris Sokolov Alexandre Dolgui Dmitry Ivanov

Current literature presents optimal control computational algorithms with regard to state, control, and conjunctive variable spaces. This paper first analyses the advantages and limitations of different optimal control computational methods and algorithms which can be used for short-term scheduling. Second, it develops an optimal control computational algorithm that allows for the solution of short-term scheduling in an optimal manner. Moreover, qualitative and quantitative analysis of the manufacturing system scheduling problem is presented. Results highlight computer experiments with a scheduling software prototype as well as potential future research avenues.

]]>Algorithms doi: 10.3390/a11050056

Authors: J. Senthilnath Sumanth Simha C Nagaraj G Meenakumari Thapa Indiramma M

This paper proposes the Bayesian Extreme Learning Machine Kohonen Network (BELMKN) framework to solve the clustering problem. The BELMKN framework uses three levels in processing nonlinearly separable datasets to obtain efficient clustering in terms of accuracy. In the first level, the Extreme Learning Machine (ELM)-based feature learning approach captures the nonlinearity in the data distribution by mapping it onto a d-dimensional space. In the second level, ELM-based feature extracted data is used as an input for Bayesian Information Criterion (BIC) to predict the number of clusters termed as a cluster prediction. In the final level, feature-extracted data along with the cluster prediction is passed to the Kohonen Network to obtain improved clustering accuracy. The main advantage of the proposed method is to overcome the problem of having a priori identifiers or class labels for the data; it is difficult to obtain labels in most of the cases for the real world datasets. The BELMKN framework is applied to 3 synthetic datasets and 10 benchmark datasets from the UCI machine learning repository and compared with the state-of-the-art clustering methods. The experimental results show that the proposed BELMKN-based clustering outperforms other clustering algorithms for the majority of the datasets. Hence, the BELMKN framework can be used to improve the clustering accuracy of the nonlinearly separable datasets.

]]>Algorithms doi: 10.3390/a11040055

Authors: Omid Gholami Johanna Törnquist Krasemann

Effectiveness in managing disturbances and disruptions in railway traffic networks, when they inevitably do occur, is a significant challenge, both from a practical and theoretical perspective. In this paper, we propose a heuristic approach for solving the real-time train traffic re-scheduling problem. This problem is here interpreted as a blocking job-shop scheduling problem, and a hybrid of the mixed graph and alternative graph is used for modelling the infrastructure and traffic dynamics on a mesoscopic level. A heuristic algorithm is developed and applied to resolve the conflicts by re-timing, re-ordering, and locally re-routing the trains. A part of the Southern Swedish railway network from Karlskrona centre to Malm&ouml; city is considered for an experimental performance assessment of the approach. The network consists of 290 block sections, and for a one-hour time horizon with around 80 active trains, the algorithm generates a solution in less than ten seconds. A benchmark with the corresponding mixed-integer program formulation, solved by commercial state-of-the-art solver Gurobi, is also conducted to assess the optimality of the generated solutions.

]]>Algorithms doi: 10.3390/a11040054

Authors: D. G. Mogale Geet Lahoti Shashi Bhushan Jha Manish Shukla Narasimha Kamath Manoj Kumar Tiwari

A number of markets, geographically separated, with different demand characteristics for different products that share a common component, are analyzed. This common component can either be manufactured locally in each of the markets or transported between the markets to fulfill the demand. However, final assemblies are localized to the respective markets. The decision making challenge is whether to manufacture the common component centrally or locally. To formulate the underlying setting, a newsvendor modeling based approach is considered. The developed model is solved using Frank-Wolfe linearization technique along with Benders&rsquo; decomposition method. Further, the propensity of decision makers in each market to make suboptimal decisions leading to bounded rationality is considered. The results obtained for both the cases are compared.

]]>Algorithms doi: 10.3390/a11040053

Authors: Luís M. S. Russo Andreia Sofia Teixeira Alexandre P. Francisco

We consider the problem of uniformly generating a spanning tree for an undirected connected graph. This process is useful for computing statistics, namely for phylogenetic trees. We describe a Markov chain for producing these trees. For cycle graphs, we prove that this approach significantly outperforms existing algorithms. For general graphs, experimental results show that the chain converges quickly. This yields an efficient algorithm due to the use of proper fast data structures. To obtain the mixing time of the chain we describe a coupling, which we analyze for cycle graphs and simulate for other graphs.

]]>Algorithms doi: 10.3390/a11040052

Authors: Naomi Nishimura

Reconfiguration is concerned with relationships among solutions to a problem instance, where the reconfiguration of one solution to another is a sequence of steps such that each step produces an intermediate feasible solution. The solution space can be represented as a reconfiguration graph, where two vertices representing solutions are adjacent if one can be formed from the other in a single step. Work in the area encompasses both structural questions (Is the reconfiguration graph connected?) and algorithmic ones (How can one find the shortest sequence of steps between two solutions?) This survey discusses techniques, results, and future directions in the area.

]]>Algorithms doi: 10.3390/a11040051

Authors: Xinxin Lv Qi Zhu

Spectrum sensing is the prerequisite of the realization of cognitive radio. So it is a significant part of cognitive radio. In order to stimulate the SUs to sense the spectrum, we combine the incentive mechanism of crowd-sensing with cooperative spectrum sensing effectively, and put forward a crowd cooperative spectrum sensing algorithm with optimal utility of secondary users (SUs) under non-ideal channel which we define SUs&rsquo; utility expectation functions related to rewards, sensing time and transmission power. Then, we construct the optimization problem of maximizing the utilities of SUs by optimizing the sensing time and the transmission power, and prove that this problem is a convex optimization problem. The optimal sensing time and transmission power are obtained by using the Karush-Kuhn-Tucker (KKT) conditions. The numerical simulation results show that the spectrum detection performance of algorithm, which we put forward, is improved.

]]>Algorithms doi: 10.3390/a11040050

Authors: Alexander A. Lazarev Ivan Nekrasov Nikolay Pravdivets

We consider one approach to formalize the Resource-Constrained Project Scheduling Problem (RCPSP) in terms of combinatorial optimization theory. The transformation of the original problem into combinatorial setting is based on interpreting each operation as an atomic entity that has a defined duration and has to be resided on the continuous time axis meeting additional restrictions. The simplest case of continuous-time scheduling assumes one-to-one correspondence of resources and operations and corresponds to the linear programming problem setting. However, real scheduling problems include many-to-one relations which leads to the additional combinatorial component in the formulation due to operations competition. We research how to apply several typical algorithms to solve the resulted combinatorial optimization problem: enumeration including branch-and-bound method, gradient algorithm, random search technique.

]]>Algorithms doi: 10.3390/a11040049

Authors: Qing Tian Weihang Zhao Yun Wei Liping Pang

With the improvement of China&rsquo;s metro carrying capacity, people in big cities are inclined to travel by metro. The carrying load of these metros is huge during the morning and evening rush hours. Coupled with the increase in numbers of summer tourists, the thermal environmental quality in early metro stations will decline badly. Therefore, it is necessary to analyze the factors that affect the thermal environment in metro stations and establish a thermal environment change model. This will help to support the prediction and analysis of the thermal environment in such limited underground spaces. In order to achieve relatively accurate and rapid on-line modeling, this paper proposes a thermal environment modeling method based on a Random Vector Functional Link Neural Network (RVFLNN). This modeling method has the advantages of fast modeling speed and relatively accurate prediction results. Once the preprocessed data is input into this RVFLNN for training, the metro station thermal environment model will be quickly established. The study results show that the thermal model based on the RVFLNN method can effectively predict the temperature inside the metro station.

]]>Algorithms doi: 10.3390/a11040048

Authors: Xinxin Wang Xiaoqiang Yan Donghai Li Li Sun

In this paper, a new tuning method is proposed, based on the desired dynamics equation (DDE) and the generalized frequency method (GFM), for a two-degree-of-freedom proportional-integral-derivative (PID) controller. The DDE method builds a quantitative relationship between the performance and the two-degree-of-freedom PID controller parameters and guarantees the desired dynamic, but it cannot guarantee the stability margin. So, we have developed the proposed tuning method, which guarantees not only the desired dynamic but also the stability margin. Based on the DDE and the GFM, several simple formulas are deduced to calculate directly the controller parameters. In addition, it performs almost no overshooting setpoint response. Compared with Panagopoulos’ method, the proposed methodology is proven to be effective.

]]>Algorithms doi: 10.3390/a11040047

Authors: Yanzhen Xing Donghui Wang Leiou Wang

To enhance the convergence speed and calculation precision of the grey wolf optimization algorithm (GWO), this paper proposes a dynamic generalized opposition-based grey wolf optimization algorithm (DOGWO). A dynamic generalized opposition-based learning strategy enhances the diversity of search populations and increases the potential of finding better solutions which can accelerate the convergence speed, improve the calculation precision, and avoid local optima to some extent. Furthermore, 23 benchmark functions were employed to evaluate the DOGWO algorithm. Experimental results show that the proposed DOGWO algorithm could provide very competitive results compared with other analyzed algorithms, with a faster convergence speed, higher calculation precision, and stronger stability.

]]>Algorithms doi: 10.3390/a11040046

Authors: Mirella Rodriguez Daniel Jeske

Continuous sampling plans are used to ensure a high level of quality for items produced in long-run contexts. The basic idea of these plans is to alternate between 100% inspection and a reduced rate of inspection frequency. Any inspected item that is found to be defective is replaced with a non-defective item. Because not all items are inspected, some defective items will escape to the customer. Analytical formulas have been developed that measure both the customer perceived quality and also the level of inspection effort. The analysis of continuous sampling plans does not apply to short-run contexts, where only a finite-size batch of items is to be produced. In this paper, a simulation algorithm is designed and implemented to analyze the customer perceived quality and the level of inspection effort for short-run contexts. A parameter representing the effectiveness of the test used during inspection is introduced to the analysis, and an analytical approximation is discussed. An application of the simulation algorithm that helped answer questions for the U.S. Navy is discussed.

]]>Algorithms doi: 10.3390/a11040045

Authors: Rolf Klein Christos Levcopoulos Andrzej Lingas

Let R denote a connected region inside a simple polygon, P. By building barriers (typically straight-line segments) in P \ R , we want to separate from R part(s) of P of maximum area. All edges of the boundary of P are assumed to be already constructed or natural barriers. In this paper we introduce two versions of this problem. In the budget fence version the region R is static, and there is an upper bound on the total length of barriers we may build. In the basic geometric firefighter version we assume that R represents a fire that is spreading over P at constant speed (varying speed can also be handled). Building a barrier takes time proportional to its length, and each barrier must be completed before the fire arrives. In this paper we are assuming that barriers are chosen from a given set B that satisfies certain conditions. Even for simple cases (e.g., P is a convex polygon and B the set of all diagonals), both problems are shown to be NP-hard. Our main result is an efficient ≈11.65 approximation algorithm for the firefighter problem, where the set B of allowed barriers is any set of straight-line segments with all endpoints on the boundary of P and pairwise disjoint interiors. Since this algorithm solves a much more general problem—a hybrid of scheduling and maximum coverage—it may find wider applications. We also provide a polynomial-time approximation scheme for the budget fence problem, for the case where barriers chosen from a set of straight-line cuts of the polygon must not cross.

]]>Algorithms doi: 10.3390/a11040044

Authors: Hong-Mei Zhang Ming-Long Li Le Yang

The A* algorithm has been widely investigated and applied in path planning problems, but it does not fully consider the safety and smoothness of the path. Therefore, an improved A* algorithm is presented in this paper. Firstly, a new environment modeling method is proposed in which the evaluation function of A* algorithm is improved by taking the safety cost into account. This results in a safer path which can stay farther away from obstacles. Then a new path smoothing method is proposed, which introduces a path evaluation mechanism into the smoothing process. This method is then applied to smoothing the path without safety reduction. Secondly, with respect to path planning problems in complex terrains, a complex terrain environment model is established in which the distance and safety cost of the evaluation function of the A* algorithm are converted into time cost. This results in a unification of units as well as a clarity in their physical meanings. The simulation results show that the improved A* algorithm can greatly improve the safety and smoothness of the planned path and the movement time of the robot in complex terrain is greatly reduced.

]]>Algorithms doi: 10.3390/a11040043

Authors: Helio Fuchigami Ruhul Sarker Socorro Rangel

The number of just-in-time jobs maximization in a permutation flow shop scheduling problem is considered. A mixed integer linear programming model to represent the problem as well as solution approaches based on enumeration and constructive heuristics were proposed and computationally implemented. Instances with up to 10 jobs and five machines are solved by the mathematical model in an acceptable running time (3.3 min on average) while the enumeration method consumes, on average, 1.5 s. The 10 constructive heuristics proposed show they are practical especially for large-scale instances (up to 100 jobs and 20 machines), with very good-quality results and efficient running times. The best two heuristics obtain near-optimal solutions, with only 0.6% and 0.8% average relative deviations. They prove to be better than adaptations of the NEH heuristic (well-known for providing very good solutions for makespan minimization in flow shop) for the considered problem.

]]>Algorithms doi: 10.3390/a11040042

Authors: Muneki Yasuda

The machine learning techniques for Markov random fields are fundamental in various fields involving pattern recognition, image processing, sparse modeling, and earth science, and a Boltzmann machine is one of the most important models in Markov random fields. However, the inference and learning problems in the Boltzmann machine are NP-hard. The investigation of an effective learning algorithm for the Boltzmann machine is one of the most important challenges in the field of statistical machine learning. In this paper, we study Boltzmann machine learning based on the (first-order) spatial Monte Carlo integration method, referred to as the 1-SMCI learning method, which was proposed in the author&rsquo;s previous paper. In the first part of this paper, we compare the method with the maximum pseudo-likelihood estimation (MPLE) method using a theoretical and a numerical approaches, and show the 1-SMCI learning method is more effective than the MPLE. In the latter part, we compare the 1-SMCI learning method with other effective methods, ratio matching and minimum probability flow, using a numerical experiment, and show the 1-SMCI learning method outperforms them.

]]>Algorithms doi: 10.3390/a11040041

Authors: Zhenwen He Xiaogang Ma

Timelines have been used for centuries and have become more and more widely used with the development of social media in recent years. Every day, various smart phones and other instruments on the internet of things generate massive data related to time. Most of these data can be managed in the way of timelines. However, it is still a challenge to effectively and efficiently store, query, and process big timeline data, especially the instant recommendation based on timeline similarities. Most existing studies have focused on indexing spatial and interval datasets rather than the timeline dataset. In addition, many of them are designed for a centralized system. A timeline index structure adapting to parallel and distributed computation framework is in urgent need. In this research, we have defined the timeline similarity query and developed a novel timeline index in the distributed system, called the Distributed Triangle Increment Tree (DTI-Tree), to support the similarity query. The DTI-Tree consists of one T-Tree and one or more TI-Trees based on a triangle increment partition strategy with the Apache Spark. Furthermore, we have provided an open source timeline benchmark data generator, named TimelineGenerator, to generate various timeline test datasets for different conditions. The experiments for DTI-Tree’s construction, insertion, deletion, and similarity queries have been executed on a cluster with two benchmark datasets that are generated by TimelineGenerator. The experimental results show that the DTI-tree provides an effective and efficient distributed index solution to big timeline data.

]]>Algorithms doi: 10.3390/a11040040

Authors: Ruth Haas Gary MacGillivray

A k-colouring of a graph G with colours 1 , 2 , &hellip; , k is canonical with respect to an ordering &pi; = v 1 , v 2 , &hellip; , v n of the vertices of G if adjacent vertices are assigned different colours and, for 1 &le; c &le; k , whenever colour c is assigned to a vertex v i , each colour less than c has been assigned to a vertex that precedes v i in &pi; . The canonical k-colouring graph of G with respect to &pi; is the graph Can k &pi; ( G ) with vertex set equal to the set of canonical k-colourings of G with respect to &pi; , with two of these being adjacent if and only if they differ in the colour assigned to exactly one vertex. Connectivity and Hamiltonicity of canonical colouring graphs of bipartite and complete multipartite graphs is studied. It is shown that for complete multipartite graphs, and bipartite graphs there exists a vertex ordering &pi; such that Can k &pi; ( G ) is connected for large enough values of k. It is proved that a canonical colouring graph of a complete multipartite graph usually does not have a Hamilton cycle, and that there exists a vertex ordering &pi; such that Can k &pi; ( K m , n ) has a Hamilton path for all k &ge; 3 . The paper concludes with a detailed consideration of Can k &pi; ( K 2 , 2 , &hellip; , 2 ) . For each k &ge; &chi; and all vertex orderings &pi; , it is proved that Can k &pi; ( K 2 , 2 , &hellip; , 2 ) is either disconnected or isomorphic to a particular tree.

]]>Algorithms doi: 10.3390/a11040039

Authors: Maimaitiyiming Hasimu Wushour Silamu

Text on the Internet is written in different languages and scripts that can be divided into different language groups. Most of the errors in language identification occur with similar languages. To improve the performance of short-text language identification, we propose four different levels of hierarchical language identification methods and conducted comparative tests in this paper. The efficiency of the algorithms was evaluated on sentences from 97 languages, and its macro-averaged F1-score reached in four-stage language identification was 0.9799. The experimental results verified that, after script identification, language group identification and similar language group identification, the performance of the language identification algorithm improved with each stage. Notably, the language identification accuracy between similar languages improved substantially. We also investigated how foreign content in a language affects language identification.

]]>Algorithms doi: 10.3390/a11040038

Authors: Christos Papalitsas Panayiotis Karakostas Theodore Andronikos Spyros Sioutas Konstantinos Giannakis

General variable neighborhood search (GVNS) is a well known and widely used metaheuristic for efficiently solving many NP-hard combinatorial optimization problems. We propose a novel extension of the conventional GVNS. Our approach incorporates ideas and techniques from the field of quantum computation during the shaking phase. The travelling salesman problem (TSP) is a well known NP-hard problem which has broadly been used for modelling many real life routing cases. As a consequence, TSP can be used as a basis for modelling and finding routes via the Global Positioning System (GPS). In this paper, we examine the potential use of this method for the GPS system of garbage trucks. Specifically, we provide a thorough presentation of our method accompanied with extensive computational results. The experimental data accumulated on a plethora of TSP instances, which are shown in a series of figures and tables, allow us to conclude that the novel GVNS algorithm can provide an efficient solution for this type of geographical problem.

]]>Algorithms doi: 10.3390/a11040037

Authors: Waldemar Kaiser Johannes Popp Michael Rinderle Tim Albes Alessio Gagliardi

In this paper, we present our generalized kinetic Monte Carlo (kMC) framework for the simulation of organic semiconductors and electronic devices such as solar cells (OSCs) and light-emitting diodes (OLEDs). Our model generalizes the geometrical representation of the multifaceted properties of the organic material by the use of a non-cubic, generalized Voronoi tessellation and a model that connects sites to polymer chains. Herewith, we obtain a realistic model for both amorphous and crystalline domains of small molecules and polymers. Furthermore, we generalize the excitonic processes and include triplet exciton dynamics, which allows an enhanced investigation of OSCs and OLEDs. We outline the developed methods of our generalized kMC framework and give two exemplary studies of electrical and optical properties inside an organic semiconductor.

]]>Algorithms doi: 10.3390/a11040036

Authors: Yu Feng Jianzhong Zhou Li Mo Chao Wang Zhe Yuan Jiang Wu

In this paper, a gradient-based cuckoo search algorithm (GCS) is proposed to solve a reservoir-scheduling problem. The classical cuckoo search (CS) is first improved by a self-adaptive solution-generation technique, together with a differential strategy for Lévy flight. This improved CS is then employed to solve the reservoir-scheduling problem, and a two-way solution-correction strategy is introduced to handle variants’ constraints. Moreover, a gradient-based search strategy is developed to improve the search speed and accuracy. Finally, the proposed GCS is used to obtain optimal schemes for cascade reservoirs in the Jinsha River, China. Results show that the mean and standard deviation of power generation obtained by GCS are much better than other methods. The converging speed of GCS is also faster. In the optimal results, the fluctuation of the water level obtained by GCS is small, indicating the proposed GCS’s effectiveness in dealing with reservoir-scheduling problems.

]]>Algorithms doi: 10.3390/a11040035

Authors: Boris Kriheli Eugene Levner

This paper considers a graph model of hierarchical supply chains. The goal is to measure the complexity of links between different components of the chain, for instance, between the principal equipment manufacturer (a root node) and its suppliers (preceding supply nodes). The information entropy is used to serve as a measure of knowledge about the complexity of shortages and pitfalls in relationship between the supply chain components under uncertainty. The concept of conditional (relative) entropy is introduced which is a generalization of the conventional (non-relative) entropy. An entropy-based algorithm providing efficient assessment of the supply chain complexity as a function of the SC size is developed.

]]>Algorithms doi: 10.3390/a11040034

Authors: Jianghong Zhu Rui Wang Yanlai Li

Failure mode and effects analysis is an effective and powerful risk evaluation technique in the field of risk management, and it has been extensively used in various industries for identifying and decreasing known and potential failure modes in systems, processes, products, and services. Traditionally, a risk priority number is applied to capture the ranking order of failure modes in failure mode and effects analysis. However, this method has several drawbacks and deficiencies, which need to be improved for enhancing its application capability. For instance, this method ignores the consensus-reaching process and the correlations among the experts’ preferences. Therefore, the aim of this study was to present a new risk priority method to determine the risk priority of failure modes under an interval-valued Pythagorean fuzzy environment, which combines the extended Geometric Bonferroni mean operator, a consensus-reaching process, and an improved Multi-Attributive Border Approximation area Comparison approach. Finally, a case study concerning product development is described to demonstrate the feasibility and effectiveness of the proposed method. The results show that the risk priority of failure modes obtained by the proposed method is more reasonable in practical application compared with other failure mode and effects analysis methods.

]]>Algorithms doi: 10.3390/a11030033

Authors: Feiyan Qin Weimin Li Yue Hu Guoqing Xu

Hybrid electric vehicles are a compromise between traditional vehicles and pure electric vehicles and can be part of the solution to the energy shortage problem. Energy management strategies (EMSs) are highly related to energy utilization in HEVs’ fuel economy. In this research, we have employed a neuro-dynamic programming (NDP) method to simultaneously optimize fuel economy and battery state of charge (SOC). In this NDP method, the critic network is a multi-resolution wavelet neural network based on the Meyer wavelet function, and the action network is a conventional wavelet neural network based on the Morlet function. The weights and parameters of both networks are obtained by an algorithm of backpropagation type. The NDP-based EMS has been applied to a parallel HEV and compared with a previously reported NDP EMS and a stochastic dynamic programing-based method. Simulation results under ADVISOR2002 have shown that the proposed NDP approach achieves better performance than both the methods. These indicate that the proposed NDP EMS, and the CWNN and MRWNN, are effective in approximating a nonlinear system.

]]>Algorithms doi: 10.3390/a11030032

Authors: Temitope Jaiyeola Florentin Smarandache

This paper is the first study of the neutrosophic triplet loop (NTL) which was originally introduced by Floretin Smarandache. NTL originated from the neutrosophic triplet set X: a collection of triplets ( x , n e u t ( x ) , a n t i ( x ) ) for an x ∈ X which obeys some axioms (existence of neutral(s) and opposite(s)). NTL can be informally said to be a neutrosophic triplet group that is not associative. That is, a neutrosophic triplet group is an NTL that is associative. In this study, NTL with inverse properties such as: right inverse property (RIP), left inverse property (LIP), right cross inverse property (RCIP), left cross inverse property (LCIP), right weak inverse property (RWIP), left weak inverse property (LWIP), automorphic inverse property (AIP), and anti-automorphic inverse property are introduced and studied. The research was carried out with the following assumptions: the inverse property (IP) is the RIP and LIP, cross inverse property (CIP) is the RCIP and LCIP, weak inverse property (WIP) is the RWIP and LWIP. The algebraic properties of neutrality and opposite in the aforementioned inverse property NTLs were investigated, and they were found to share some properties with the neutrosophic triplet group. The following were established: (1) In a CIPNTL (IPNTL), RIP (RCIP) and LIP (LCIP) were equivalent; (2) In an RIPNTL (LIPNTL), the CIP was equivalent to commutativity; (3) In a commutative NTL, the RIP, LIP, RCIP, and LCIP were found to be equivalent; (4) In an NTL, IP implied anti-automorphic inverse property and WIP, RCIP implied AIP and RWIP, while LCIP implied AIP and LWIP; (5) An NTL has the IP (CIP) if and only if it has the WIP and anti-automorphic inverse property (AIP); (6) A CIPNTL or an IPNTL was a quasigroup; (7) An LWIPNTL (RWIPNTL) was a left (right) quasigroup. The algebraic behaviours of an element, its neutral and opposite in the associator and commutator of a CIPNTL or an IPNTL were investigated. It was shown that ( Z p , ∗ ) where x ∗ y = ( p − 1 ) ( x + y ) , for any prime p, is a non-associative commutative CIPNTL and IPNTL. The application of some of these varieties of inverse property NTLs to cryptography is discussed.

]]>Algorithms doi: 10.3390/a11030031

Authors: A. Ahmed Ji Sun

The classical capacitated vehicle routing problem (CVRP) is a very popular combinatorial optimization problem in the field of logistics and supply chain management. Although CVRP has drawn interests of many researchers, no standard way has been established yet to obtain best known solutions for all the different problem sets. We propose an efficient algorithm Bilayer Local Search-based Particle Swarm Optimization (BLS-PSO) along with a novel decoding method to solve CVRP. Decoding method is important to relate the encoded particle position to a feasible CVRP solution. In bilayer local search, one layer of local search is for the whole population in any iteration whereas another one is applied only on the pool of the best particles generated in different generations. Such searching strategies help the BLS-PSO to perform better than the existing proposals by obtaining best known solutions for most of the existing benchmark problems within very reasonable computational time. Computational results also show that the performance achieved by the proposed algorithm outperforms other PSO-based approaches.

]]>Algorithms doi: 10.3390/a11030030

Authors: Liping Liu Xiaobo Liu Ning Wang Peijun Zou

Cuckoo Search (CS) is a Meta-heuristic method, which exhibits several advantages such as easier to application and fewer tuning parameters. However, it has proven to very easily fall into local optimal solutions and has a slow rate of convergence. Therefore, we propose Modified cuckoo search algorithm with variational parameter and logistic map (VLCS) to ameliorate these defects. To balance the exploitation and exploration of the VLCS algorithm, we not only use the coefficient function to change step size α and probability of detection p a at next generation, but also use logistic map of each dimension to initialize host nest location and update the location of host nest beyond the boundary. With fifteen benchmark functions, the simulations demonstrate that the VLCS algorithm can over come the disadvantages of the CS algorithm.In addition, the VLCS algorithm is good at dealing with high dimension problems and low dimension problems.

]]>Algorithms doi: 10.3390/a11030029

Authors: Lilian Shi Jun Ye

The neutrosophic cubic set can describe complex decision-making problems with its single-valued neutrosophic numbers and interval neutrosophic numbers simultaneously. The Dombi operations have the advantage of good flexibility with the operational parameter. In order to solve decision-making problems with flexible operational parameter under neutrosophic cubic environments, the paper extends the Dombi operations to neutrosophic cubic sets and proposes a neutrosophic cubic Dombi weighted arithmetic average (NCDWAA) operator and a neutrosophic cubic Dombi weighted geometric average (NCDWGA) operator. Then, we propose a multiple attribute decision-making (MADM) method based on the NCDWAA and NCDWGA operators. Finally, we provide two illustrative examples of MADM to demonstrate the application and effectiveness of the established method.

]]>