Previous Issue
Volume 12, May-1
 
 

Mathematics, Volume 12, Issue 10 (May-2 2024) – 175 articles

  • Issues are regarded as officially published after their release is announced to the table of contents alert mailing list.
  • You may sign up for e-mail alerts to receive table of contents of newly released issues.
  • PDF is the official format for papers published in both, html and pdf forms. To view the papers in pdf format, click on the "PDF Full-text" link, and use the free Adobe Reader to open them.
Order results
Result details
Section
Select all
Export citation of selected articles as:
31 pages, 2841 KiB  
Article
Performance Evaluation of Railway Infrastructure Managers: A Novel Hybrid Fuzzy MCDM Model
by Aida Kalem, Snežana Tadić, Mladen Krstić, Nermin Čabrić and Nedžad Branković
Mathematics 2024, 12(10), 1590; https://doi.org/10.3390/math12101590 (registering DOI) - 19 May 2024
Viewed by 146
Abstract
Modern challenges such as the liberalization of the railway sector and growing demands for sustainability, high-quality services, and user satisfaction set new standards in railway operations. In this context, railway infrastructure managers (RIMs) play a crucial role in ensuring innovative approaches that will [...] Read more.
Modern challenges such as the liberalization of the railway sector and growing demands for sustainability, high-quality services, and user satisfaction set new standards in railway operations. In this context, railway infrastructure managers (RIMs) play a crucial role in ensuring innovative approaches that will strengthen the position of railways in the market by enhancing efficiency and competitiveness. Evaluating their performance is essential for assessing the achieved objectives, and it is conducted through a wide range of key performance indicators (KPIs), which encompass various dimensions of operations. Monitoring and analyzing KPIs are crucial for improving service quality, achieving sustainability, and establishing a foundation for research and development of new strategies in the railway sector. This paper provides a detailed overview and evaluation of KPIs for RIMs. This paper creates a framework for RIM evaluation using various scientific methods, from identifying KPIs to applying complex analysis methods. A novel hybrid model, which integrates the fuzzy Delphi method for aggregating expert opinions on the KPIs’ importance, the extended fuzzy analytic hierarchy process (AHP) method for determining the relative weights of these KPIs, and the ADAM method for ranking RIMs, has been developed in this paper. This approach enables a detailed analysis and comparison of RIMs and their performances, providing the basis for informed decision-making and the development of new strategies within the railway sector. The analysis results provide insight into the current state of railway infrastructure and encourage further efforts to improve the railway sector by identifying key areas for enhancement. The main contributions of the research include a detailed overview of KPIs for RIMs and the development of a hybrid multi-criteria decision making (MCDM) model. The hybrid model represents a significant step in RIM performance analysis, providing a basis for future research in this area. The model is universal and, as such, represents a valuable contribution to MCDM theory. Full article
(This article belongs to the Special Issue Multi-criteria Optimization Models and Methods for Smart Cities)
Show Figures

Figure 1

22 pages, 8339 KiB  
Article
Micro-Grinding Parameter Control of Hard and Brittle Materials Based on Kinematic Analysis of Material Removal
by Hisham Manea, Hong Lu, Qi Liu, Junbiao Xiao and Kefan Yang
Mathematics 2024, 12(10), 1589; https://doi.org/10.3390/math12101589 (registering DOI) - 19 May 2024
Viewed by 110
Abstract
This article explores the intricacies of micro-grinding parameter control for hard and brittle materials, with a specific focus on Zirconia ceramics (ZrO2) and Optical Glass (BK7). Given the increasing demand and application of these materials in various high-precision industries, this study [...] Read more.
This article explores the intricacies of micro-grinding parameter control for hard and brittle materials, with a specific focus on Zirconia ceramics (ZrO2) and Optical Glass (BK7). Given the increasing demand and application of these materials in various high-precision industries, this study aims to provide a comprehensive kinematic analysis of material removal during the micro-grinding process. According to the grinding parameters selected to be analyzed in this study, the ac-max values are between (9.55 nm ~ 67.58 nm). Theoretical modeling of the grinding force considering the brittle and ductile removal phase, frictional effects, the possibility of grit to cut materials, and grinding conditions is very important in order to control and optimize the surface grinding process. This research introduces novel models for predicting and optimizing micro-grinding forces effectively. The primary objective is to establish a micro-grinding force model that facilitates the easy manipulation of micro-grinding parameters, thereby optimizing the machining process for these challenging materials. Through experimental investigations conducted on Zirconia ceramics, the paper evaluates a mathematical model of the grinding force, highlighting its significance in predicting and controlling the forces involved in micro-grinding. The suggested model underwent thorough testing to assess its validity, revealing an accuracy with average variances of 6.616% for the normal force and 5.752% for the tangential force. Additionally, the study delves into the coefficient of friction within the grinding process, suggesting a novel frictional force model. This model is assessed through a series of experiments on Optical Glass BK7, aiming to accurately characterize the frictional forces at play during grinding. The empirical results obtained from both sets of experiments—on Zirconia ceramics and Optical Glass BK7—substantiate the efficacy of the proposed models. These findings confirm the models’ capability to accurately describe the force dynamics in the micro-grinding of hard and brittle materials. The research not only contributes to the theoretical understanding of micro-grinding processes but also offers practical insights for enhancing the efficiency and effectiveness of machining operations involving hard and brittle materials. Full article
Show Figures

Figure 1

8 pages, 259 KiB  
Article
Persistence and Stochastic Extinction in a Lotka–Volterra Predator–Prey Stochastically Perturbed Model
by Leonid Shaikhet and Andrei Korobeinikov
Mathematics 2024, 12(10), 1588; https://doi.org/10.3390/math12101588 (registering DOI) - 19 May 2024
Viewed by 124
Abstract
The classical Lotka–Volterra predator–prey model is globally stable and uniformly persistent. However, in real-life biosystems, the extinction of species due to stochastic effects is possible and may occur if the magnitudes of the stochastic effects are large enough. In this paper, we consider [...] Read more.
The classical Lotka–Volterra predator–prey model is globally stable and uniformly persistent. However, in real-life biosystems, the extinction of species due to stochastic effects is possible and may occur if the magnitudes of the stochastic effects are large enough. In this paper, we consider the classical Lotka–Volterra predator–prey model under stochastic perturbations. For this model, using an analytical technique based on the direct Lyapunov method and a development of the ideas of R.Z. Khasminskii, we find the precise sufficient conditions for the stochastic extinction of one and both species and, thus, the precise necessary conditions for the stochastic system’s persistence. The stochastic extinction occurs via a process known as the stabilization by noise of the Khasminskii type. Therefore, in order to establish the sufficient conditions for extinction, we found the conditions for this stabilization. The analytical results are illustrated by numerical simulations. Full article
(This article belongs to the Special Issue Stochastic Models in Mathematical Biology, 2nd Edition)
18 pages, 607 KiB  
Article
Reliability Assessment of Bridge Structure Using Bilal Distribution
by Ahmed T. Ramadan, Osama Abdulaziz Alamri and Ahlam H. Tolba
Mathematics 2024, 12(10), 1587; https://doi.org/10.3390/math12101587 (registering DOI) - 19 May 2024
Viewed by 98
Abstract
Reliability assessments are pivotal in evaluating system quality and have found extensive application in manufacturing. This research delves into a system comprising five components, one of which is a bridge network. The components are presumed to follow a Bilal lifetime distribution with a [...] Read more.
Reliability assessments are pivotal in evaluating system quality and have found extensive application in manufacturing. This research delves into a system comprising five components, one of which is a bridge network. The components are presumed to follow a Bilal lifetime distribution with a failure rate that changes over time. Four distinct methods are employed to enhance the components within the system. This study involves the computation of δ-fractiles and reliability equivalence factors (REFs). Additionally, a numerical case study is provided to elucidate the theoretical findings. Full article
(This article belongs to the Section Engineering Mathematics)
13 pages, 6907 KiB  
Article
Inverse Scheme to Locally Determine Nonlinear Magnetic Material Properties: Numerical Case Study
by Manfred Kaltenbacher, Andreas Gschwentner, Barbara Kaltenbacher, Stefan Ulbrich and Alice Reinbacher-Köstinger
Mathematics 2024, 12(10), 1586; https://doi.org/10.3390/math12101586 (registering DOI) - 19 May 2024
Viewed by 96
Abstract
We are interested in the determination of the local nonlinear magnetic material behaviour in electrical steel sheets due to cutting and punching effects. For this purpose, the inverse problem has to be solved, where the objective function, which penalises the difference between the [...] Read more.
We are interested in the determination of the local nonlinear magnetic material behaviour in electrical steel sheets due to cutting and punching effects. For this purpose, the inverse problem has to be solved, where the objective function, which penalises the difference between the measured and the simulated magnetic flux density, has to be minimised under a constraint defined according to the corresponding partial differential equation model. We use the adjoint method to efficiently obtain the gradients of the objective function with respect to the material parameters. The optimisation algorithm is low-memory Broyden–Fletcher–Goldfarb–Shanno (BFGS), the forward and adjoint formulations are solved using the finite element (FE) method and the ill-posedness is handled via Tikhonov regularisation, in combination with the discrepancy principle. Realistic numerical case studies show promising results. Full article
(This article belongs to the Special Issue Numerical Optimization for Electromagnetic Problems)
19 pages, 7906 KiB  
Article
Abundant New Optical Soliton Solutions to the Biswas–Milovic Equation with Sensitivity Analysis for Optimization
by Md Nur Hossain, Faisal Alsharif, M. Mamun Miah and Mohammad Kanan
Mathematics 2024, 12(10), 1585; https://doi.org/10.3390/math12101585 (registering DOI) - 19 May 2024
Viewed by 133
Abstract
This study extensively explores the Biswas–Milovic equation (BME) with Kerr and power law nonlinearity to extract the unique characteristics of optical soliton solutions. These optical soliton solutions have different applications in the field of precision in optical switching, applications in waveguide design, exploration [...] Read more.
This study extensively explores the Biswas–Milovic equation (BME) with Kerr and power law nonlinearity to extract the unique characteristics of optical soliton solutions. These optical soliton solutions have different applications in the field of precision in optical switching, applications in waveguide design, exploration of nonlinear optical effects, imaging precision, reduced intensity fluctuations, suitability for optical signal processing in optical physics, etc. Through the powerful (G/G, 1/G)-expansion analytical method, a variety of soliton solutions are expressed in three distinct forms: trigonometric, hyperbolic, and rational expressions. Rigorous validation using Mathematica software ensures precision, while dynamic visual representations vividly portray various soliton patterns such as kink, anti-kink, singular soliton, hyperbolic, dark soliton, and periodic bright soliton solutions. Indeed, a sensitivity analysis was conducted to assess how changes in parameters affect the exact solutions, aiding in the understanding of system behavior and informing decision-making, especially in accurately designing or analyzing real-world optical phenomena. This investigation reveals the significant influence of parameters λ, τ, c, B, and Κ on the precise solutions in Kerr and power law nonlinearities within the BME. Notably, parameter λ exhibits consistently high sensitivity across all scenarios, while parameters τ and c demonstrate pronounced sensitivity in scenario III. The outcomes derived from this method are distinctive and carry significant implications for the dynamics of optical fibers and wave phenomena across various optical systems. Full article
(This article belongs to the Special Issue Exact Solutions and Numerical Solutions of Differential Equations)
Show Figures

Figure 1

28 pages, 13072 KiB  
Article
An Inverse Problem for Estimating Spatially and Temporally Dependent Surface Heat Flux with Thermography Techniques
by Cheng-Hung Huang and Kuan-Chieh Fang
Mathematics 2024, 12(10), 1584; https://doi.org/10.3390/math12101584 (registering DOI) - 19 May 2024
Viewed by 95
Abstract
In this study, an inverse conjugate heat transfer problem is examined to estimate temporally and spatially the dependent unknown surface heat flux using thermography techniques with a thermal camera in a three-dimensional domain. Thermography techniques encompass a broad set of methods and procedures [...] Read more.
In this study, an inverse conjugate heat transfer problem is examined to estimate temporally and spatially the dependent unknown surface heat flux using thermography techniques with a thermal camera in a three-dimensional domain. Thermography techniques encompass a broad set of methods and procedures used for capturing and analyzing thermal data, while thermal cameras are specific tools used within those techniques to capture thermal images. In the present study, the interface conditions of the plate and air domains are obtained using perfect thermal contact conditions, and therefore we define the problem studied as an inverse conjugate heat transfer problem. Achieving the simultaneous solution of the continuity, Navier–Stokes, and energy equations within the air domain, alongside the heat conduction equation in the plate domain, presents a more intricate challenge compared to conventional inverse heat conduction problems. The validity of our inverse solutions was verified through numerical simulations, considering various inlet air velocities and plate thicknesses. Notably, it was found that due to the singularity of the gradient of the cost function at the final time point, the estimated results near the final time must be discarded, and exact measurements consistently produce accurate boundary heat fluxes under thin-plate conditions, with air velocity exhibiting no significant impact on the estimates. Additionally, an analysis of measurement errors and their influence on the inverse solutions was conducted. The numerical results conclusively demonstrated that the maximum error when estimating heat flux consistently remained below 3% and higher measurement noise resulted in the accuracy of the heat flux estimation decreasing. This underscores the inherent challenges associated with inverse problems and highlights the importance of obtaining accurate measurement data in the problem domain. Full article
(This article belongs to the Special Issue Computational and Analytical Methods for Inverse Problems)
18 pages, 288 KiB  
Article
A New Approach of Complex Fuzzy Ideals in BCK/BCI-Algebras
by Manivannan Balamurugan, Thukkaraman Ramesh, Anas Al-Masarwah and Kholood Alsager
Mathematics 2024, 12(10), 1583; https://doi.org/10.3390/math12101583 (registering DOI) - 18 May 2024
Viewed by 210
Abstract
The concept of complex fuzzy sets, where the unit disk of the complex plane acts as the codomain of the membership function, as an extension of fuzzy sets. The objective of this article is to use complex fuzzy sets in BCK/BCI-algebras. We present [...] Read more.
The concept of complex fuzzy sets, where the unit disk of the complex plane acts as the codomain of the membership function, as an extension of fuzzy sets. The objective of this article is to use complex fuzzy sets in BCK/BCI-algebras. We present the concept of a complex fuzzy subalgebra in a BCK/BCI-algebra and explore their properties. Furthermore, we discuss the modal and level operators of these complex fuzzy subalgebras, highlighting their importance in BCK/BCI-algebras. We study various operations, and the laws of a complex fuzzy system, including union, intersection, complement, simple differences, and bounded differences of complex fuzzy ideals within BCK/BCI-algebras. Finally, we generate a computer programming algorithm that implements our complex fuzzy subalgebras/ideals in BCK/BCI-algebras procedure for ease of lengthy calculations. Full article
(This article belongs to the Special Issue Advanced Methods in Fuzzy Control and Their Applications)
19 pages, 2235 KiB  
Article
Consumer Default Risk Portrait: An Intelligent Management Framework of Online Consumer Credit Default Risk
by Miao Zhu, Ben-Chang Shia, Meng Su and Jialin Liu
Mathematics 2024, 12(10), 1582; https://doi.org/10.3390/math12101582 (registering DOI) - 18 May 2024
Viewed by 255
Abstract
Online consumer credit services play a vital role in the contemporary consumer market. To foster their sustainable development, it is essential to establish and strengthen the relevant risk management mechanism. This study proposes an intelligent management framework called the consumer default risk portrait [...] Read more.
Online consumer credit services play a vital role in the contemporary consumer market. To foster their sustainable development, it is essential to establish and strengthen the relevant risk management mechanism. This study proposes an intelligent management framework called the consumer default risk portrait (CDRP) to mitigate the default risks associated with online consumer loans. The CDRP framework combines traditional credit information and Internet platform data to depict the portrait of consumer default risks. It consists of four modules: addressing data imbalances, establishing relationships between user characteristics and the default risk, analyzing the influence of different variables on default, and ultimately presenting personalized consumer profiles. Empirical findings reveal that “Repayment Periods”, “Loan Amount”, and “Debt to Income Type” emerge as the three variables with the most significant impact on default. “Re-payment Periods” and “Debt to Income Type” demonstrate a positive correlation with default probability, while a lower “Loan Amount” corresponds to a higher likelihood of default. Additionally, our verification highlights that the significance of variables varies across different samples, thereby presenting a personalized portrait from a single sample. In conclusion, the proposed framework provides valuable suggestions and insights for financial institutions and Internet platform managers to improve the market environment of online consumer credit services. Full article
Show Figures

Figure 1

16 pages, 516 KiB  
Article
A Negative Sample-Free Graph Contrastive Learning Algorithm
by Dongming Chen, Mingshuo Nie, Zhen Wang, Huilin Chen and Dongqi Wang
Mathematics 2024, 12(10), 1581; https://doi.org/10.3390/math12101581 (registering DOI) - 18 May 2024
Viewed by 147
Abstract
Self-supervised learning is a new machine learning method that does not rely on manually labeled data, and learns from rich unlabeled data itself by designing agent tasks using the input data as supervision to obtain a more generalized representation for application in downstream [...] Read more.
Self-supervised learning is a new machine learning method that does not rely on manually labeled data, and learns from rich unlabeled data itself by designing agent tasks using the input data as supervision to obtain a more generalized representation for application in downstream tasks. However, the current self-supervised learning suffers from the problem of relying on the selection and number of negative samples and the problem of sample bias phenomenon after graph data augmentation. In this paper, we investigate the above problems and propose a corresponding solution, proposing a graph contrastive learning algorithm without negative samples. The model uses matrix sketching in the implicit space for feature augmentation to reduce sample bias and iteratively trains the mutual correlation matrix of two viewpoints by drawing closer to the distance of the constant matrix as the objective function. This method does not require techniques such as negative samples, gradient stopping, and momentum updating to prevent self-supervised model collapse. This method is compared with 10 graph representation learning algorithms on four datasets for node classification tasks, and the experimental results show that the algorithm proposed in this paper achieves good results. Full article
(This article belongs to the Special Issue Complex Network Modeling in Artificial Intelligence Applications)
Show Figures

Figure 1

14 pages, 10874 KiB  
Article
A Stochastic Semi-Parametric SEIR Model with Infectivity in an Incubation Period
by Jing Zhang and Tong Jin
Mathematics 2024, 12(10), 1580; https://doi.org/10.3390/math12101580 (registering DOI) - 18 May 2024
Viewed by 136
Abstract
This paper introduces stochastic disturbances into a semi-parametric SEIR model with infectivity in an incubation period. The model combines the randomness of disease transmission and the nonlinearity of transmission rate, providing a flexible framework for more accurate description of the process of infectious [...] Read more.
This paper introduces stochastic disturbances into a semi-parametric SEIR model with infectivity in an incubation period. The model combines the randomness of disease transmission and the nonlinearity of transmission rate, providing a flexible framework for more accurate description of the process of infectious disease transmission. On the basis of the discussion of the deterministic model, the stochastic semi-parametric SEIR model is studied. Firstly, we use Lyapunov analysis to prove the existence and uniqueness of global positive solutions for the model. Secondly, the conditions for disease extinction are established, and appropriate stochastic Lyapunov functions are constructed to discuss the asymptotic behavior of the model’s solution at the disease-free equilibrium point of the deterministic model. Finally, the specific transmission functions are enumerated, and the accuracy of the results are demonstrated through numerical simulations. Full article
27 pages, 1488 KiB  
Article
Strategic Queueing Behavior of Two Groups of Patients in a Healthcare System
by Youxin Liu, Liwei Liu, Tao Jiang and Xudong Chai
Mathematics 2024, 12(10), 1579; https://doi.org/10.3390/math12101579 (registering DOI) - 18 May 2024
Viewed by 136
Abstract
Long waiting times and crowded services are the current medical situation in China. Especially in hierarchic healthcare systems, as high-quality medical resources are mainly concentrated in comprehensive hospitals, patients are too concentrated in these hospitals, which leads to overcrowding. This paper constructs a [...] Read more.
Long waiting times and crowded services are the current medical situation in China. Especially in hierarchic healthcare systems, as high-quality medical resources are mainly concentrated in comprehensive hospitals, patients are too concentrated in these hospitals, which leads to overcrowding. This paper constructs a game-theoretical queueing model to analyze the strategic queueing behavior of patients. In such hospitals, patients are divided into first-visit and referred patients, and the hospitals provide patients with two service phases of “diagnosis” and “treatment”. We first obtain the expected sojourn time. By defining the patience level of patients, the queueing behavior of patients in equilibrium is studied. The results suggest that as long as the patients with low patience levels join the queue, the patients with high patience levels also join the queue. As more patients arrive at the hospitals, the queueing behavior of patients with high patience levels may have a negative effect on that of patients with low patience levels. The numerical results also show that the equilibrium behavior deviates from a socially optimal solution; therefore, to reach maximal social welfare, the social planner should adopt some regulatory policies to control the arrival rates of patients. Full article
(This article belongs to the Special Issue Queueing Systems Models and Their Applications)
21 pages, 2245 KiB  
Article
The Lomax-Exponentiated Odds Ratio–G Distribution and Its Applications
by Sudakshina Singha Roy, Hannah Knehr, Declan McGurk, Xinyu Chen, Achraf Cohen and Shusen Pu
Mathematics 2024, 12(10), 1578; https://doi.org/10.3390/math12101578 (registering DOI) - 18 May 2024
Viewed by 128
Abstract
This paper introduces the Lomax-exponentiated odds ratio–G (L-EOR–G) distribution, a novel framework designed to adeptly navigate the complexities of modern datasets. It blends theoretical rigor with practical application to surpass the limitations of traditional models in capturing complex data attributes such as heavy [...] Read more.
This paper introduces the Lomax-exponentiated odds ratio–G (L-EOR–G) distribution, a novel framework designed to adeptly navigate the complexities of modern datasets. It blends theoretical rigor with practical application to surpass the limitations of traditional models in capturing complex data attributes such as heavy tails, shaped curves, and multimodality. Through a comprehensive examination of its theoretical foundations and empirical data analysis, this study lays down a systematic theoretical framework by detailing its statistical properties and validates the distribution’s efficacy and robustness in parameter estimation via Monte Carlo simulations. Empirical evidence from real-world datasets further demonstrates the distribution’s superior modeling capabilities, supported by compelling various goodness-of-fit tests. The convergence of theoretical precision and practical utility heralds the L-EOR–G distribution as a groundbreaking advancement in statistical modeling, significantly enhancing precision and adaptability. The new model not only addresses a critical need within statistical modeling but also opens avenues for future research, including the development of more sophisticated estimation methods and the adaptation of the model for various data types, thereby promising to refine statistical analysis and interpretation across a wide array of disciplines. Full article
(This article belongs to the Special Issue New Advances in Applied Probability and Stochastic Processes)
13 pages, 2219 KiB  
Article
Utilizing Artificial Neural Networks for Geometric Bone Model Reconstruction in Mandibular Prognathism Patients
by Jelena Mitić, Nikola Vitković, Miroslav Trajanović, Filip Górski, Ancuţa Păcurar, Cristina Borzan, Emilia Sabău and Răzvan Păcurar
Mathematics 2024, 12(10), 1577; https://doi.org/10.3390/math12101577 (registering DOI) - 18 May 2024
Viewed by 147
Abstract
Patient-specific 3D models of the human mandible are finding increasing utility in medical fields such as oral and maxillofacial surgery, orthodontics, dentistry, and forensic sciences. The efficient creation of personalized 3D bone models poses a key challenge in these applications. Existing solutions often [...] Read more.
Patient-specific 3D models of the human mandible are finding increasing utility in medical fields such as oral and maxillofacial surgery, orthodontics, dentistry, and forensic sciences. The efficient creation of personalized 3D bone models poses a key challenge in these applications. Existing solutions often rely on 3D statistical models of human bone, offering advantages in rapid bone geometry adaptation and flexibility by capturing a range of anatomical variations, but also a disadvantage in terms of reduced precision in representing specific shapes. Considering this, the proposed parametric model allows for precise manipulation using morphometric parameters acquired from medical images. This paper highlights the significance of employing the parametric model in the creation of a personalized bone model, exemplified through a case study targeting mandibular prognathism reconstruction. A personalized model is described as 3D point cloud determined through the utilization of series of parametric functions, determined by the application of geometrical morphometrics, morphology properties, and artificial neural networks in the input dataset of human mandible samples. With 95.05% of the personalized model’s surface area displaying deviations within −1.00–1.00 mm relative to the input polygonal model, and a maximum deviation of 2.52 mm, this research accentuates the benefits of the parametric approach, particularly in the preoperative planning of mandibular deformity surgeries. Full article
Show Figures

Figure 1

21 pages, 2864 KiB  
Article
The Rise of the Superstars: Uncovering the Composition Effect of International Trade that Cements the Dominant Position of Big Businesses
by Chara Vavoura
Mathematics 2024, 12(10), 1576; https://doi.org/10.3390/math12101576 (registering DOI) - 18 May 2024
Viewed by 137
Abstract
International markets are extremely polarised, with a few big superstar businesses operating alongside numerous small competitors, and globalisation has been highlighted as a powerful force behind the superstars’ increasingly dominant presence. The empirical literature has established that superstars are more efficient compared to [...] Read more.
International markets are extremely polarised, with a few big superstar businesses operating alongside numerous small competitors, and globalisation has been highlighted as a powerful force behind the superstars’ increasingly dominant presence. The empirical literature has established that superstars are more efficient compared to their smaller counterparts, and, unlike them, they exhibit strategic behaviour. Building on this evidence, we develop a model to examine how an initial productivity advantage allows a select few firms to expand, via innovation, to the extent that it becomes optimal to adopt strategic behaviour, and show how polarised markets emerge endogenously as the unique subgame perfect equilibrium in pure strategies. We then introduce international trade and show that, in polarised markets, trade liberalisation puts into motion a novel composition effect, reallocating market share from smaller to larger rivals and raising large firms’ profits. This effect suppresses the pro-competitive welfare gains from trade and cements the dominant position of big businesses, who come out as the big winners of globalisation. We find that, although trade increases welfare, by reducing average markup and markup heterogeneity, in the presence of a handful of large powerful firms, welfare gains are severely diminished, and subsidising smaller enterprises may turn out to be welfare-enhancing. Full article
(This article belongs to the Special Issue Mathematical Economics and Its Applications)
18 pages, 2561 KiB  
Article
Ensemble Machine Learning Approach for Parkinson’s Disease Detection Using Speech Signals
by Syed Nisar Hussain Bukhari and Kingsley A. Ogudo
Mathematics 2024, 12(10), 1575; https://doi.org/10.3390/math12101575 (registering DOI) - 18 May 2024
Viewed by 189
Abstract
The detection of Parkinson’s disease (PD) is vital as it affects the population worldwide and decreases the quality of life. The disability and death rate due to PD is increasing at an unprecedented rate, more than any other neurological disorder. To this date, [...] Read more.
The detection of Parkinson’s disease (PD) is vital as it affects the population worldwide and decreases the quality of life. The disability and death rate due to PD is increasing at an unprecedented rate, more than any other neurological disorder. To this date, no diagnostic procedures exist for this disease. However, several computational approaches have proven successful in detecting PD at early stages, overcoming the disadvantages of traditional methods of diagnosis. In this study, a machine learning (ML) detection system based on the voice signals of PD patients is proposed. The AdaBoost classifier has been utilized to construct the model and trained on a dataset obtained from the machine learning repository of the University of California, Irvine (UCI). This dataset includes voice attributes such as time-frequency features, Mel frequency cepstral coefficients, wavelet transform features, vocal fold features, and tremor waveform quality time. The model demonstrated promising performance, achieving high accuracy, precision, recall, F1 score, and AUC score of 0.96, 0.98, 0.93, 0.95, and 0.99, respectively. Furthermore, the robustness of the proposed model is rigorously assessed through cross-validation, revealing consistent performance across all iterations. The overarching objective of this study is to contribute to the scientific community by furnishing a robust system for the detection of PD. Full article
(This article belongs to the Special Issue Artificial Intelligence Solutions in Healthcare)
Show Figures

Figure 1

15 pages, 653 KiB  
Article
Preserving Global Information for Graph Clustering with Masked Autoencoders
by Rui Chen
Mathematics 2024, 12(10), 1574; https://doi.org/10.3390/math12101574 - 17 May 2024
Viewed by 259
Abstract
Graph clustering aims to divide nodes into different clusters without labels and has attracted great attention due to the success of graph neural networks (GNNs). Traditional GNN-based clustering methods are based on the homophilic assumption, i.e., connected nodes belong to the same clusters. [...] Read more.
Graph clustering aims to divide nodes into different clusters without labels and has attracted great attention due to the success of graph neural networks (GNNs). Traditional GNN-based clustering methods are based on the homophilic assumption, i.e., connected nodes belong to the same clusters. However, this assumption is not always true, as heterophilic graphs are also ubiquitous in the real world, which limits the application of GNNs. Furthermore, these methods overlook global positions, which can result in erroneous clustering. To solve the aforementioned problems, we propose a novel model called Preserving Global Information for Graph Clustering with Masked Autoencoders (GCMA). We first propose a low–high-pass filter to capture meaningful low- and high-frequency information. Then, we propose a graph diffusion method to obtain the global position. Specifically, a parameterized Laplacian matrix is proposed to better control the global direction. To further enhance the learning ability of the autoencoders, we design a model with a masking strategy that enhances the learning ability. Extensive experiments on both homophilic and heterophilic graphs demonstrate GCMA’s advantages over state-of-the-art baselines. Full article
(This article belongs to the Special Issue Advances in Data Mining, Neural Networks and Deep Graph Learning)
Show Figures

Figure 1

16 pages, 1674 KiB  
Article
Novel Automatic Classification of Human Adult Lung Alveolar Type II Cells Infected with SARS-CoV-2 through the Deep Transfer Learning Approach
by Turki Turki, Sarah Al Habib and Y-h. Taguchi
Mathematics 2024, 12(10), 1573; https://doi.org/10.3390/math12101573 - 17 May 2024
Viewed by 243
Abstract
Transmission electron microscopy imaging provides a unique opportunity to inspect the detailed structure of infected lung cells with SARS-CoV-2. Unlike previous studies, this novel study aims to investigate COVID-19 classification at the lung cellular level in response to SARS-CoV-2. Particularly, differentiating between healthy [...] Read more.
Transmission electron microscopy imaging provides a unique opportunity to inspect the detailed structure of infected lung cells with SARS-CoV-2. Unlike previous studies, this novel study aims to investigate COVID-19 classification at the lung cellular level in response to SARS-CoV-2. Particularly, differentiating between healthy and infected human alveolar type II (hAT2) cells with SARS-CoV-2. Hence, we explore the feasibility of deep transfer learning (DTL) and introduce a highly accurate approach that works as follows: First, we downloaded and processed 286 images pertaining to healthy and infected hAT2 cells obtained from the electron microscopy public image archive. Second, we provided processed images to two DTL computations to induce ten DTL models. The first DTL computation employs five pre-trained models (including DenseNet201 and ResNet152V2) trained on more than one million images from the ImageNet database to extract features from hAT2 images. Then, it flattens and provides the output feature vectors to a trained, densely connected classifier with the Adam optimizer. The second DTL computation works in a similar manner, with a minor difference in that we freeze the first layers for feature extraction in pre-trained models while unfreezing and jointly training the next layers. The results using five-fold cross-validation demonstrated that TFeDenseNet201 is 12.37× faster and superior, yielding the highest average ACC of 0.993 (F1 of 0.992 and MCC of 0.986) with statistical significance ( from a t-test) compared to an average ACC of 0.937 (F1 of 0.938 and MCC of 0.877) for the counterpart (TFtDenseNet201), showing no significance results ( from a t-test). Full article
(This article belongs to the Special Issue Advanced Applications of Deep Learning Methods in Medical Diagnosis)
26 pages, 1307 KiB  
Article
Incorporating Multi-Source Market Sentiment and Price Data for Stock Price Prediction
by Kui Fu and Yanbin Zhang
Mathematics 2024, 12(10), 1572; https://doi.org/10.3390/math12101572 - 17 May 2024
Viewed by 237
Abstract
The problem of stock price prediction has been a hot research issue. Stock price is influenced by various factors at the same time, and market sentiment is one of the most critical factors. Financial texts such as news and investor comments reflect investor [...] Read more.
The problem of stock price prediction has been a hot research issue. Stock price is influenced by various factors at the same time, and market sentiment is one of the most critical factors. Financial texts such as news and investor comments reflect investor sentiment in the stock market and influence market movements. Previous research models have struggled to accurately mine multiple sources of market sentiment information originating from the Internet and traditional sentiment analysis models are challenging to quantify and combine indicator data from market data and multi-source sentiment data. Therefore, we propose a BERT-LLA stock price prediction model incorporating multi-source market sentiment and technical analysis. In the sentiment analysis module, we propose a semantic similarity and sector heat-based model to screen for related sectors and use fine-tuned BERT models to calculate the text sentiment index, transforming the text data into sentiment index time series data. In the technical indicator calculation module, technical indicator time series are calculated using market data. Finally, in the prediction module, we combine the sentiment index time series and technical indicator time series and employ a two-layer LSTM network prediction model with an integrated attention mechanism to predict stock close price. Our experiment results show that the BERT-LLA model can accurately capture market sentiment and has a strong practicality and forecasting ability in analyzing market sentiment and stock price prediction. Full article
14 pages, 297 KiB  
Article
Markov Chains and Kinetic Theory: A Possible Application to Socio-Economic Problems
by Bruno Carbonaro and Marco Menale
Mathematics 2024, 12(10), 1571; https://doi.org/10.3390/math12101571 - 17 May 2024
Viewed by 281
Abstract
A very important class of models widely used nowadays to describe and predict, at least in stochastic terms, the behavior of many-particle systems (where the word “particle” is not meant in the purely mechanical sense: particles can be cells of a living tissue, [...] Read more.
A very important class of models widely used nowadays to describe and predict, at least in stochastic terms, the behavior of many-particle systems (where the word “particle” is not meant in the purely mechanical sense: particles can be cells of a living tissue, or cars in a traffic flow, or even members of an animal or human population) is the Kinetic Theory for Active Particles, i.e., a scheme of possible generalizations and re-interpretations of the Boltzmann equation. Now, though in the literature on the subject this point is systematically disregarded, this scheme is based on Markov Chains, which are special stochastic processes with important properties they share with many natural processes. This circumstance is here carefully discussed not only to suggest the different ways in which Markov Chains can intervene in equations describing the stochastic behavior of any many-particle system, but also, as a preliminary methodological step, to point out the way in which the notion of a Markov Chain can be suitably generalized to this aim. As a final result of the discussion, we find how to develop new very plausible and likely ways to take into account possible effects of the external world on a non-isolated many-particle system, with particular attention paid to socio-economic problems. Full article
(This article belongs to the Special Issue Kinetic Models of Collective Phenomena and Data Science)
17 pages, 7500 KiB  
Article
Parameterizations of Delaunay Surfaces from Scratch
by Clementina D. Mladenova and Ivaïlo M. Mladenov
Mathematics 2024, 12(10), 1570; https://doi.org/10.3390/math12101570 - 17 May 2024
Viewed by 192
Abstract
Starting with the most fundamental differential-geometric principles we derive here new explicit parameterizations of the Delaunay surfaces of revolution which depend on two real parameters with fixed ranges. Besides, we have proved that these parameters have very clear geometrical meanings. The first one [...] Read more.
Starting with the most fundamental differential-geometric principles we derive here new explicit parameterizations of the Delaunay surfaces of revolution which depend on two real parameters with fixed ranges. Besides, we have proved that these parameters have very clear geometrical meanings. The first one is responsible for the size of the surface under consideration and the second one specifies its shape. Depending on the concrete ranges of these parameters any of the Delaunay surfaces which is neither a cylinder nor the plane is classified unambiguously either as a first or a second kind Delaunay surface. According to this classification spheres are Delaunay surfaces of first kind while the catenoids are Delaunay surfaces of second kind. Geometry of both classes is established meaning that the coefficients of their fundamental forms are found in explicit form. Full article
(This article belongs to the Special Issue Differentiable Manifolds and Geometric Structures)
25 pages, 1736 KiB  
Article
Robust-MBDL: A Robust Multi-Branch Deep-Learning-Based Model for Remaining Useful Life Prediction of Rotating Machines
by Khoa Tran, Hai-Canh Vu, Lam Pham, Nassim Boudaoud and Ho-Si-Hung Nguyen
Mathematics 2024, 12(10), 1569; https://doi.org/10.3390/math12101569 - 17 May 2024
Viewed by 321
Abstract
Predictive maintenance (PdM) is one of the most powerful maintenance techniques based on the estimation of the remaining useful life (RUL) of machines. Accurately estimating the RUL is crucial to ensure the effectiveness of PdM. However, current methods have limitations in fully exploring [...] Read more.
Predictive maintenance (PdM) is one of the most powerful maintenance techniques based on the estimation of the remaining useful life (RUL) of machines. Accurately estimating the RUL is crucial to ensure the effectiveness of PdM. However, current methods have limitations in fully exploring condition monitoring data, particularly vibration signals, for RUL estimation. To address these challenges, this research presents a novel Robust Multi-Branch Deep Learning (Robust-MBDL) model. Robust-MBDL stands out by leveraging diverse data sources, including raw vibration signals, time–frequency representations, and multiple feature domains. To achieve this, it adopts a specialized three-branch architecture inspired by efficient network designs. The model seamlessly integrates information from these branches using an advanced attention-based Bi-LSTM network. Furthermore, recognizing the importance of data quality, Robust-MBDL incorporates an unsupervised LSTM-Autoencoder for noise reduction in raw vibration data. This comprehensive approach not only overcomes the limitations of existing methods but also leads to superior performance. Experimental evaluations on benchmark datasets such as XJTU-SY and PRONOSTIA showcase Robust-MBDL’s efficacy, particularly in rotating machine health prognostics. These results underscore its potential for real-world applications, heralding a new era in predictive maintenance practices. Full article
Show Figures

Figure 1

17 pages, 1784 KiB  
Article
Robustness of Real-World Networks after Weight Thresholding with Strong Link Removal
by Jisha Mariyam John, Michele Bellingeri, Divya Sindhu Lekha, Davide Cassi and Roberto Alfieri
Mathematics 2024, 12(10), 1568; https://doi.org/10.3390/math12101568 - 17 May 2024
Viewed by 200
Abstract
Weight thresholding (WT) is a method intended to decrease the number of links within weighted networks that may otherwise be excessively dense for network science applications. WT aims to remove links to simplify the network by holding most of the features [...] Read more.
Weight thresholding (WT) is a method intended to decrease the number of links within weighted networks that may otherwise be excessively dense for network science applications. WT aims to remove links to simplify the network by holding most of the features of the original network. Here, we test the robustness and the efficacy of the node attack strategies on real-world networks subjected to WT that remove links of higher weight (strong links). We measure the network robustness along node removal with the largest connected component (LCC). We find that the real-world networks under study are generally robust when subjected to WT. Nonetheless, WT with strong link removal changes the efficacy of the attack strategies and the rank of node centralities. Also, WT with strong link removal may trigger a more significant change in the node centrality rank than WT by removing weak links. Network science research with the aim to find important/influential nodes in the network has to consider that simplifying the network with WT methodologies may change the node centrality. Full article
(This article belongs to the Special Issue Big Data and Complex Networks)
23 pages, 5800 KiB  
Article
The Impact of Missing Continuous Blood Glucose Samples on Machine Learning Models for Predicting Postprandial Hypoglycemia: An Experimental Analysis
by Najib Ur Rehman, Ivan Contreras, Aleix Beneyto and Josep Vehi
Mathematics 2024, 12(10), 1567; https://doi.org/10.3390/math12101567 - 17 May 2024
Viewed by 240
Abstract
This study investigates how missing data samples in continuous blood glucose data affect the prediction of postprandial hypoglycemia, which is crucial for diabetes management. We analyzed the impact of missing samples at different times before meals using two datasets: virtual patient data and [...] Read more.
This study investigates how missing data samples in continuous blood glucose data affect the prediction of postprandial hypoglycemia, which is crucial for diabetes management. We analyzed the impact of missing samples at different times before meals using two datasets: virtual patient data and real patient data. The study uses six commonly used machine learning models under varying conditions of missing samples, including custom and random patterns reflective of device failures and arbitrary data loss, with different levels of data removal before mealtimes. Additionally, the study explored different interpolation techniques to counter the effects of missing data samples. The research shows that missing samples generally reduce the model performance, but random forest is more robust to missing samples. The study concludes that the adverse effects of missing samples can be mitigated by leveraging complementary and informative non-point features. Consequently, our research highlights the importance of strategically handling missing data, selecting appropriate machine learning models, and considering feature types to enhance the performance of postprandial hypoglycemia predictions, thereby improving diabetes management. Full article
(This article belongs to the Special Issue Artificial Intelligence Solutions in Healthcare)
23 pages, 16746 KiB  
Article
Analysis of Soil Slope Stability under Underground Coal Seam Mining Using Improved Radial Movement Optimization with Lévy Flight
by Haotian Li, Liangxing Jin and Pingting Liu
Mathematics 2024, 12(10), 1566; https://doi.org/10.3390/math12101566 - 17 May 2024
Viewed by 233
Abstract
Underground coal seam mining significantly reduces the stability of slopes, especially soil slopes, and an accurate evaluation of the stability of soil slopes under underground mining conditions is crucial for mining safety. In this study, the impact of coal seam mining is considered [...] Read more.
Underground coal seam mining significantly reduces the stability of slopes, especially soil slopes, and an accurate evaluation of the stability of soil slopes under underground mining conditions is crucial for mining safety. In this study, the impact of coal seam mining is considered as the additional horizontal and vertical stresses acting on the slope, and an equation for calculating the safety factor of soil slopes under underground mining conditions is derived based on the rigorous Janbu method. Then, the Improved Radial Movement Optimization (IRMO) algorithm is introduced and combined with Lévy flight optimization to conduct global optimization searches, obtaining the critical sliding surface and corresponding safety factor of the soil slope under underground coal seam mining. Through comparisons with the numerical simulation results in three different case studies, the feasibility of applying the IRMO algorithm with Lévy flight to analyze the stability of soil slopes under underground mining is demonstrated. This ensures the accuracy and stability of the calculation results while maintaining a high convergence efficiency. Furthermore, the effects of the mining thickness and mining direction on slope stability are analyzed, and the results indicate that a smaller mining thickness and mining along the slope are advantageous for slope stability. The method proposed in this study provides valuable insights for preventing the slope instability hazards caused by underground coal seam mining. Full article
Show Figures

Figure 1

10 pages, 2733 KiB  
Article
Use of the Adaptive Cross Approximation for the Efficient Computation of the Reduced Matrix with the Characteristic Basis Function Method
by Eliseo García, Carlos Delgado and Felipe Cátedra
Mathematics 2024, 12(10), 1565; https://doi.org/10.3390/math12101565 - 17 May 2024
Viewed by 250
Abstract
A technique for the reduction in the CPU-time in the analysis of electromagnetic problems using the Characteristic Basis Function Method (CBFM) is presented here, allowing for analysis of electrically large cases where an iterative solution process cannot be avoided. This technique is based [...] Read more.
A technique for the reduction in the CPU-time in the analysis of electromagnetic problems using the Characteristic Basis Function Method (CBFM) is presented here, allowing for analysis of electrically large cases where an iterative solution process cannot be avoided. This technique is based on the use of the Adaptive Cross Approximation (ACA) for the fast computation of the coupling matrix between CBFs belonging to adjacent blocks, as well as the Multilevel Fast Multipole Method (MLFMM) for the computation of matrix−vector products in the solution of the full system. This combination allows for a noticeable reduction in the computational resources during the analysis of electrically large and complex scenarios while maintaining a very good degree of accuracy. A number of test cases serve to validate the presented approach in terms of accuracy, memory and CPU-time compared with conventional techniques. Full article
(This article belongs to the Special Issue Mathematical Applications in Electrical Engineering)
Show Figures

Figure 1

18 pages, 305 KiB  
Article
Using Game Theory to Explore the Multinational Supply Chain Production Inventory Models of Various Carbon Emission Policy Combinations
by Jialiang Pan, Kun-Shan Wu, Chih-Te Yang, Chi-Jie Lu and Shin Lu
Mathematics 2024, 12(10), 1564; https://doi.org/10.3390/math12101564 - 17 May 2024
Viewed by 250
Abstract
This study uses Stackelberg game theory, considering different combinations of carbon emission reduction policies and that high-carbon-emission enterprises may face various carbon emission reduction regulations, to explore the production inventory problems in a multinational supply chain system. The purpose is to determine the [...] Read more.
This study uses Stackelberg game theory, considering different combinations of carbon emission reduction policies and that high-carbon-emission enterprises may face various carbon emission reduction regulations, to explore the production inventory problems in a multinational supply chain system. The purpose is to determine the manufacturer’s optimal production, shipping, carbon reduction investment, and the retailer’s replenishment under the equilibrium for different carbon emission policy combinations. To develop the production inventory models, this study first develops the total profit and carbon emission functions of the supply chain members, respectively, and then obtains the optimal solutions and total profits of the manufacturer and the retailer under different carbon emission policy combinations through the mathematical analysis method. Further, this study used several numerical examples to solve and compare the proposed models. The results of numerical analysis show that regardless of the increase in carbon price or carbon tax, the manufacturer and retailer will adjust their decisions to reduce carbon emissions. Specifically, an increase in the carbon price contributes to an increase in the total profit of manufacturers, while an increase in the carbon tax reduces the total profit of manufacturers. This study also explores a sensitivity analysis on the main parameters and has yielded meaningful management insights. For instance, in cases where low-carbonization strategies are required, the manufacturer or retailer can effectively reduce the carbon emissions resulting from production or purchasing activities, thereby significantly reducing overall carbon emissions. It is believed that the results of this study can provide enterprises/supply chains with reference to their respective production, transportation, carbon reduction investment, and inventory decisions under carbon emission policies, as well as information on partner selection and how to adjust decisions under environmental changes. Full article
(This article belongs to the Section Engineering Mathematics)
14 pages, 1402 KiB  
Article
Remote Sensing Image Classification Based on Neural Networks Designed Using an Efficient Neural Architecture Search Methodology
by Lan Song, Lixin Ding, Mengjia Yin, Wei Ding, Zhigao Zeng and Chunxia Xiao
Mathematics 2024, 12(10), 1563; https://doi.org/10.3390/math12101563 - 17 May 2024
Viewed by 192
Abstract
Successful applications of machine learning for the analysis of remote sensing images remain limited by the difficulty of designing neural networks manually. However, while the development of neural architecture search offers the unique potential for discovering new and more effective network architectures, existing [...] Read more.
Successful applications of machine learning for the analysis of remote sensing images remain limited by the difficulty of designing neural networks manually. However, while the development of neural architecture search offers the unique potential for discovering new and more effective network architectures, existing neural architecture search algorithms are computationally intensive methods requiring a large amount of data and computational resources and are therefore challenging to apply for developing optimal neural network architectures for remote sensing image classification. Our proposed method uses a differentiable neural architecture search approach for remote sensing image classification. We utilize a binary gate strategy for partial channel connections to reduce the sizes of the network parameters, creating a sparse connection pattern that lowers memory consumption and NAS computational costs. Experimental results indicate that our method achieves a 15.1% increase in validation accuracy during the search phase compared to DDSAS, although slightly lower (by 4.5%) than DARTS. However, we reduced the search time by 88% and network parameter size by 84% compared to DARTS. In the architecture evaluation phase, our method demonstrates a 2.79% improvement in validation accuracy over a manually configured CNN network. Full article
(This article belongs to the Special Issue Deep Learning and Adaptive Control, 2nd Edition)
Show Figures

Figure 1

14 pages, 892 KiB  
Article
CC-DETR: DETR with Hybrid Context and Multi-Scale Coordinate Convolution for Crowd Counting
by Yanhong Gu, Tao Zhang, Yuxia Hu and Fudong Nian
Mathematics 2024, 12(10), 1562; https://doi.org/10.3390/math12101562 - 17 May 2024
Viewed by 225
Abstract
Prevailing crowd counting approaches primarily rely on density map regression methods. Despite wonderful progress, significant scale variations and complex background interference within the same image remain challenges. To address these issues, in this paper we propose a novel DETR-based crowd counting framework called [...] Read more.
Prevailing crowd counting approaches primarily rely on density map regression methods. Despite wonderful progress, significant scale variations and complex background interference within the same image remain challenges. To address these issues, in this paper we propose a novel DETR-based crowd counting framework called Crowd Counting DETR (CC-DETR), which aims to extend the state-of-the-art DETR object detection framework to the crowd counting task. In CC-DETR, a DETR-like encoder–decoder structure (Hybrid Context DETR, i.e., HCDETR) is proposed to tackle complex visual information by fusing features from hybrid semantic levels through a transformer. In addition, we design a Coordinate Dilated Convolution Module (CDCM) to effectively employ position-sensitive context information in different scales. Extensive experiments on three challenging crowd counting datasets (ShanghaiTech, UCF-QNRF, and NWPU) demonstrate that our model is effective and competitive when compared against SOTA crowd counting models. Full article
Show Figures

Figure 1

11 pages, 255 KiB  
Article
The Adjoint of α-Times-Integrated C-Regularized Semigroups
by Mostafa Bachar
Mathematics 2024, 12(10), 1561; https://doi.org/10.3390/math12101561 - 16 May 2024
Viewed by 231
Abstract
We consider an operator {S(t)}t0 on a Banach space X with generator A, characterized by being an α-times-integrated C-regularized semigroup. The adjoint family [...] Read more.
We consider an operator {S(t)}t0 on a Banach space X with generator A, characterized by being an α-times-integrated C-regularized semigroup. The adjoint family S*(t):X*X* is introduced for analysis. {S*(t)}t0 maintains the characteristics of an α-times-integrated C-regularized semigroup, though with strong continuity and Bochner integrals being substituted by weak* continuity and weak* integrals, respectively. Our investigation focuses on the closed subspace X, where {S*(t)}t0 exhibits strong continuity. Additionally, a comparison between the adjoint A* of A and the generator of the adjoint family is conducted. Full article
Previous Issue
Back to TopTop