Numerical examples illustrate the efficiency of the considered algorithms in
Section 4. The tested systems, which are represented by
, are generated by synthetic datasets of random numbers and real-world counterparts of large-scale measurements. As used in
Section 2, the systems of interest are in discrete-time LTI forms. The entire computation was conducted on MATLAB R2022a, and CVX 2.2 [
85,
86]; the MOSEK solver was used for the SDP-based selection, as noted in
Appendix A. The
dlyap function was used for the solutions of the discrete-time Lyapunov equations, such as Equations (9) and (18), adopting the subroutine libraries from the Subroutine Library in Control Theory (SLICOT) [
87,
88,
89,
90]. It should also be noted that there is another approach to solving the equation, such as [
91]. The MATLAB programs are available through the GitHub repository of the present authors [
92].
4.1. Results of the Randomly Generated System
The characteristics of the sensor selection methods are investigated for different sets of system dimension parameters
by varying one parameter while fixing the others. The abscissa of
Figure 3a,b is the rank
, which is the dimension of the reduced state variable vector and, thus, of the Gramian. In
Figure 3c,d,
denote the sizes of node members comprising the original measurements, and accordingly the number of sensor candidates. In
Figure 3e,f,
represents the varying number of sensors selected.
The problem setting for the system of random numbers is found in [
39]. First, the conjugate complex numbers, of which the real parts are negative, are assigned to the eigenvalues of a damping continuous-time system matrix
. A discrete-time system matrix can be obtained by
, which is stable and full-rank, whereas
is the time step of the discrete system. The observation matrix
is a column-orthogonal matrix generated by the singular value decomposition for a matrix of Gaussian random numbers of appropriate dimensions. Sensors up to
p are selected using the algorithms presented in the previous section, and the objective function,
, is calculated for each selected subset.
Figure 3 shows the performance of the selection methods applied to the system of random numbers.
Figure 3a,c,e illustrate the total computation time of each algorithm in the tests, where the gradient greedy method is the least time-consuming, followed by the pure greedy and the SDP-based methods, which produce similar results. Unfortunately, the proposed approximate convex relaxation method took the longest time to solve the optimization in almost all of the conditions tested despite its acceleration owing to the customized randomization. However, it is also expected that the computation time of the SDP-based method and the pure greedy selection will exceed the others for even larger
r and
p, respectively.
The empirical orders of the computation time with regard to each parameter are analyzed herein. The growth rates of the computation time of each algorithm are evaluated by solely changing the system dimension parameters
r,
n, and
p, as summarized in
Table 2, based on the results shown in
Figure 3a,c,e. First, the gradient greedy method ran in time proportional to
n, but it is not clear regarding
r. Since the number of sensor candidates,
n, is much larger than
r in the first experiment, the term with
is not significant for the gradient greedy method. The increase against
p is on the order of unity, which is clearly reasonable. Second, the empirical orders of the pure greedy method when solely changing
r,
n, and
p are
,
n, and
p, respectively, which agrees with the expected leading order. One can see a term with
becomes dominant as
r grows in
Figure 3a. Meanwhile, the estimated orders of the SDP-based method when solely changing
r and
n are
and
n, while those of the approximate convex relaxation method are
and
, respectively. Here, let the notation
stand for (despite an obvious abuse) a real number
x bounded by
and
j for an arbitrary positive integer
j. These noninteger orders may be due to the optimized arithmetic employed in the software, such as the Strassen-like algorithm [
93]. The dependency regarding
p was not clear in the SDP-based method, while a slight increase was observed for the approximate convex relaxation method.
The interesting aspect is that the dominating order of the SDP-based method with respect to
n was not
, which was initially expected but is approximately proportional to
n. This is perhaps because the constraint posed as
in
Appendix A was simplified in the tested implementation as a mere diagonal block of the semidefinite linear matrix inequality (LMI), and the MOSEK solver should have taken advantage of its structure during the Newton step calculation, despite the large
n assumed. It should also be noted that this efficacy was not observed for other solvers, such as the SDPT3, although the CVX parser does not seem to change its output. Nonetheless, the complexity of solving the SDP would be enormous, as expected, if the LMI included such a semidefinite relaxation [
61,
94] of the selection variable vector
, such as
being a semidefinite matrix. This agrees well with the observations from the experiment for the SDP-based method regarding
r, where the LMI of the dimension of
is included.
In addition to the computation time per step, the iteration numbers before convergence, as shown in
Table 3, also illustrate the computationally friendly features for the high-dimensional problems. The iteration numbers of the SDP-based and the approximate convex relaxation methods do not increase significantly as
r,
n, and
p individually increase. Interestingly, the iteration numbers of these convex relaxation methods change in a different fashion. That of the SDP-based method slightly increases with increases in
n and
p, and slightly decreases with an increase in
r, while that of the approximate convex relaxation shows opposite results. This leads to similar growth rates against
r between these two methods, as shown in
Figure 3a, where the empirical computational complexity is less than
for the SDP-based method and more than
for the approximate convex relaxation method. Moreover,
Figure 3c shows that this difference leads to growth rates that are slightly larger than
for the SDP-based method and smaller than
for the approximate convex relaxation method, against
n. This unexpected efficiency of the SDP-based method leads to better scalability with respect to
n over the approximate convex relaxation, which is the other convex relaxation method. With regard to the increase in
p, the iteration numbers of the SDP-based method increased only slightly, while those of the approximate convex relaxation method exhibited an increase around
, corresponding to the increase in the total run time in
Figure 3e.
As shown in
Figure 3b,d,f, almost the same objective values were obtained by the SDP-based and the approximate convex relaxation methods, which implies a good agreement between the solutions of these relaxation methods. They yielded better or comparable objective function values as compared to the greedy methods, which were up to twice as high, except for
cases. Those obtained by the gradient greedy method, on the other hand, give an inferior impression as a selection method as compared with the other methods for
cases. This is possibly due to the difficulty in ensuring the observability with a limited number of sensors, especially
, which should lead to an unstable calculation of the gradient Equation (20). The gradient greedy algorithm is concluded to have poor performance in achieving observability, especially when the convex objective function is hardly approximated by its linear tangent due to large
r.
The discussion above illustrates that the use of the SDP-based method seems to be the most favorable among the compared methods in this experiment. The use of the pure greedy method is also a reasonable choice due to its shorter computation time, straightforward implementation, and reasonable performance, which returns half the objective function value of the convex relaxation methods in the experiments.
4.2. Results for Data-Driven System Derived from Real-World Experiment
An example of a practical application was conducted using the experimental dataset of flow velocity distribution around an airfoil. The data used herein are found in [
95,
96], which were previously acquired by the particle image velocimetry (PIV) in a wind tunnel. A brief description of the experiment is shown in
Table 4; refer to the original article by the authors for more specific information. The snapshots taken in the experiment quantify velocity vectors that span the visualized plane, i.e., two components on a two-dimensional grid of
points, as depicted in
Figure 4.
Only the streamwise components (the direction shown by the arrow in
Figure 4) are used, and the ensemble averages over
m snapshots are subtracted at each measurement location; that is, averages for each pixel of the calculated velocity image.
As attempted in
Section 4.1, a linear representation
is derived first. The data-driven system identification procedure is based on the modeling method of [
30,
78]. Here, the snapshots of the velocity field are reshaped to
-dimensional
m vectors,
, then the data matrix is defined by
. The proper orthogonal decomposition (POD) is then adopted and the data are projected onto the subspace of the leading
r POD modes [
37], resulting in
(
, respectively). The measurement matrix
is defined by
, which consists of
r left singular vectors related to the largest
r singular values. These left singular vectors illustrate the dominant spatial coherent structures, while the right vectors represent temporal coherence, and the singular values are the amplitudes of these modes. The use of POD in our study is intended for a more fundamental discussion based on a linear system representation where arbitrary-order low-rank systems are derived from high-dimensional measurement data.
The
r-dimensional state variable vector
is given by
After introducing
and
, the system matrix is then expressed using linear least squares via the pseudo-inverse operation
. The system matrix
is manipulated to ensure that its eigenvalues are less than unity and, therefore, the considered systems are stable.
As discussed in
Section 4.1, the comparisons are illustrated using the objective function values obtained. The sensor candidate set corresponds to each pixel obtained by the experimental visualization and, thus, its size is fixed to
. The state variable vector is set to have
components as a result of the order reduction at different truncation thresholds, while
p is fixed to 20. Moreover, the results with respect to various
p values are also provided, where
with
being fixed. The original dataset consisting of the
snapshots is divided into
consecutive snapshots and, therefore, the following results are the five-fold average.
The values of the objective function show similar trends to
Figure 3b,f for the gradient greedy algorithm, while a large degradation is observed in the convex relaxation methods for
. Surprisingly, those of the convex relaxation methods are 2–100 times lower than that of the pure greedy method. This trend is also partially observed for the
case in
Figure 5b, where the pure greedy selection corresponds to the true optimum for this configuration by definition. This clear degradation in the convex relaxations has not yet been explained in this study and may be due to the ill condition of the low-order approximation or the numerical error. A similar discussion was conducted for the previous study [
66]. The cause should be investigated in detail for a practical application with the real-world data in a future study.
The results above, together with those in the previous subsection, illustrate that the pure greedy method shows a more stable performance, while the convex relaxation methods show wide variations in their performance. Therefore, the pure greedy method can be considered the most appropriate method, taking into account the computation time. It should be noted that the present study also shows the prominent performance of the SDP-based method in terms of the objective function and the computation time in the randomly generated, general problem settings. As long as the additional computation time is acceptable, the present authors recommend trying both the pure greedy and the SDP-based methods and to use a better sensor set when selecting the sensors based on the observability Gramian.