Next Article in Journal
Formation of Photo-Responsive Liquid Crystalline Emulsion by Using Microfluidics Device
Next Article in Special Issue
Exact Renormalization Groups As a Form of Entropic Dynamics
Previous Article in Journal
Assessment of Component Selection Strategies in Hyperspectral Imagery
Previous Article in Special Issue
EXONEST: The Bayesian Exoplanetary Explorer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

L1-Minimization Algorithm for Bayesian Online Compressed Sensing

1
Latam Experian DataLab , São Paulo-SP 04547-130, Brazil
2
Department of General Physics, Institute of Physics, University of São Paulo, São Paulo-SP 05508-090, Brazil
3
Department of Applied Mathematics, Institute of Mathematics and Statistics, University of São Paulo, São Paulo-SP 05508-090, Brazil
*
Author to whom correspondence should be addressed.
Entropy 2017, 19(12), 667; https://doi.org/10.3390/e19120667
Submission received: 3 November 2017 / Revised: 25 November 2017 / Accepted: 1 December 2017 / Published: 5 December 2017

Abstract

:
In this work, we propose a Bayesian online reconstruction algorithm for sparse signals based on Compressed Sensing and inspired by L1-regularization schemes. A previous work has introduced a mean-field approximation for the Bayesian online algorithm and has shown that it is possible to saturate the offline performance in the presence of Gaussian measurement noise when the signal generating distribution is known. Here, we build on these results and show that reconstruction is possible even if prior knowledge about the generation of the signal is limited, by introduction of a Laplace prior and of an extra Kullback–Leibler divergence minimization step for hyper-parameter learning.

1. Introduction

It has become commonplace to talk about the recent “information explosion” or “data deluge”. These expressions refer to a much faster growth in data production compared to all available data storage and to the even more evident divergence between the volume of data produced and the general data processing capacity. This state of things begs for more efficient ways of storing and analyzing data. Natural and man-made signals tend to be compressible, which means their innate redundancies allow them to be expressed as a small number of combinations of its components—they are sparse in some basis. In the last few decades a lot of effort has been directed to the development of compression techniques that allow the sampled data to be rewritten to a reduced number of bits. However, these techniques mean superfluous costs on the gathering and processing of the data. When talking about medical imaging that includes radiation, unnecessary data gathering almost certainly signifies collateral and unwanted health effects in the long run for the patients.
Compressed sensing (CS) is a framework for signal processing that presents itself as a successful resource that has produced many techniques for efficient information extraction [1]. CS utilizes the fact that real-world signals can typically be represented by a small combination of elemental components [2,3] to eliminate the data redundancy directly in its acquisition (in contrast with what occurs in a digital camera, for example, that captures millions of point measurements each time a picture is taken only to discard a large portion of this data immediately after its acquisition through a image compression algorithm). In a standard scenario, CS utilizes the sparsity property to enable the recovery of various signals from much fewer samples of linear measurements than required by the Nyquist–Shannon theorem [4,5]. The Nyquist rate [6,7] is a concept present in virtually all signal acquisition protocols used in consumer electronics and medical imaging devices, among others [2]. Unfortunately, it implies the necessity of a high sampling frequency, which means an especially high demand if we consider the now pervasive high definition registers. Between 2004 and 2006, Candès, Tao and Donoho [4,8,9] showed that sparse signals could be exactly reconstructed with below-the-Nyquist-rate sampling. Moreover (and perhaps what made it so intriguing), the reconstruction could be done in non-adaptive fashion through L1-minimization. These seminal works gave rise to CS [5]. It should be noted that “sampling” here has a twist, though, in CS, a sample of the signal is obtained by calculating its inner product against different test functions. All in all, the popularity of CS in the last few years made it pervasive through various disciplines, from Computer Science to Statistical Physics.
In this work, we consider the scenario of a signal generated by an unknown distribution and its recovery by means of a Bayesian online CS scheme. There is a vast literature on Bayesian schemes for CS (e.g., [10,11,12]) and their reconstruction have been thoroughly examined, not least by the Statistical Physics community [13,14,15]. Online reconstruction schemes can be invaluable in situations where real-time inference is needed (e.g., magnetic resonance imaging), when the data set is too large to be entirely stored on a hard disk (e.g., modern simulation science) or in the presence of time-variant signals (e.g., data streaming). Algorithms for Bayesian online learning are well-known in the neural networks community [16,17,18] and, in many fields of engineering such as optimal control theory, where they are known as Kalman filters [19,20,21]. In addition, an interesting compromise between online and offline learning is the so-called mini-batch learning [22], which has recently also found its way to the CS scenario [23]. In direct connection with this letter, previous work [24] has already established the possibility of online CS reconstruction, but it suffered from an important limitation, namely the necessity of an exact knowledge about the generating distribution of the signal.
This paper is organized as follows: in Section 2, the basic CS problem is defined, together with two reconstruction methods—the L1-minimization problem called LASSO (Least Absolute Shrinkage and Selection Operator) and the Bayesian scheme. Section 3 introduces the Bayesian online CS framework as presented in [24]. In Section 4, the main contribution of this work, an extension for the online framework in which imperfect knowledge about the signal generating distribution is allowed, is described. Section 5 presents the main results and Section 6 summarizes and discusses the main findings of this work.

2. Problem Setup

Let x 0 = ( x i 0 ) R N be a real N-dimensional signal where each component is generated by a sparsity-inducing distribution ϕ ( x ) = ( 1 ρ ) δ ( x ) + ρ g ( x ) . It is assumed that 0 < ρ < 1 and that g ( x ) does not have a finite mass at the origin. Sequential linear projections of the signal with known independent random measurement vectors A μ = ( A i μ ) R N , with A i μ N ( 0 , N 1 ) are obtained at instants μ = 1 , 2 , . Let these random projections u μ A μ · x 0 pass through a (possibly noisy) channel before their result becomes available—this way, the value of each measurement, y μ , will be distributed according to P ( y μ | u μ ) (Figure 1).
A common goal in Compressed Sensing is to accurately recover the signal x 0 based on the knowledge of D t = { ( A 1 , y 1 ) , ( A 2 , y 2 ) , , ( A t , y t ) } and of the functional form of the channel P ( y | A · x ) . Since the beginnings of CS, the staple technique in the field [5] has been the LASSO, an optimization problem with an L1-regularization term [25],
x ^ t = arg min x y A x 2 2 + λ x 1 ,
where y = ( y μ ) μ = 1 t , A is a matrix with all vectors A μ as rows and the Lp-norms are defined as z p = ( i | z i | p ) 1 / p . The seminal works [4,8,9] were pivotal in proving that perfect signal reconstruction in absence of noise is possible by means of (1) even in subsampling scenarios (i.e., for ρ < α < 1 , with α t / N ) . The Statistical Physics community is no stranger for these results, with typical behaviour analysis for large systems, where N [26,27] being accepted as guarantees for the validity of the method.
More recently, Bayesian techniques for signal reconstruction have also been proposed (e.g., [10,28]). They consist in estimating x 0 from the posterior distribution
P ( x | D t ) = P ( D t | x ) i = 1 N ϕ ( x i ) d x P ( D t | x ) i = 1 N ϕ ( x i ) .
It has been shown [11] that the Bayesian reconstruction typically performs better than the L1-minimization scheme for fewer measurements. In fact, the minimum mean squared error (mmse) estimator x ^ ( D t ) = d x x P ( x | D t ) is guaranteed [14] to minimize the mean square error mse defined for any possible estimate x as:
mse N 1 x x 0 2 D t , x 0 .

3. Bayesian Online Compressed Sensing

In this work, reconstruction of the signal is done in an online manner. That is, measurements are used one at a time and discarded thereafter (in opposition to the offline/batch scenarios presented above where the whole dataset D t is available during the entire reconstruction). In addition, we would like to achieve this objective with limited knowledge about the generating distribution ϕ ( x ) .
Bayes’ Theorem (2) can be easily transformed into an online update recursion [16]. Since all measurements are independent by design, the likelihood can be factored as
P ( D t | x ) = μ = 1 t P ( y μ | A μ · x ) = P ( y t | A t · x ) μ = 1 t 1 P ( y μ | A μ · x ) .
This way, the posterior distribution P ( x | D t ) for the signal x after t measurements can be written
P ( x | D t ) P ( y t | A t · x ) P ( x | D t 1 ) .
In general, distribution P ( x | D t ) can be complicated, so the calculation of x ^ ( D t ) is potentially difficult. Consider for now the Compressed Sensing scenario where the prior distribution is exactly the same as the known signal generating distribution ϕ ( x ) . The δ ( x ) factor here adds a singularity to the prior ϕ ( x ) , so that the posterior should not approximate a Gaussian distribution even for t , which was a crucial hypothesis in previous works [16] and would greatly simplify asymptotic calculations. In absence of such a simplification, a previous work [24] introduced the following mean-field approximation:
P ˜ ( x | D t ) i = 1 N e a i t x i 2 / 2 + h i t x i ϕ ( x i ) Z ( a i t , h i t )
for the posterior distribution P ( x | D t ) in Label (5) as a device to facilitate marginal computations and (as a consequence) their expected values. In this expression, { ( a i t , h i t ) } is a set of natural parameters and
Z ( a i t , h i t ) = d x i e a i t x i 2 / 2 + h i t x i ϕ ( x i )
is the normalization of the marginal distribution P ˜ i ( x i | D t ) = d x \ i P ˜ ( x | D t ) , where d x \ i denotes integration over all { x j } j i . So as not to introduce any biases, we define h i 0 = a i 0 = 0 , i , which results in P ˜ i ( x i | D 0 ) = i = 1 N ϕ ( x i ) . The mmse estimate of the signal and its variance can then be readily calculated from the approximate posterior as
m i t = ( / h i t ) ln Z ( a i t , h i t ) ,
v i t = ( 2 / ( h i t ) 2 ) ln Z ( a i t , h i t ) ,
respectively. Equations (5) and (6) correspond to a sequence of update and projection steps—the update step adds new information obtained with the most recent measurement t, but transforms the posterior into a possibly intractable distribution; the projection step then returns this distribution to an easier exponential representation with independent components (Figure 2). The full scheme can be summarized as 2 N update rules for the parameters { ( a i t , h i t ) } [24]. Defining Δ t i = 1 N A i t m i t 1 and χ t i = 1 N ( A i t ) 2 v i t 1 , these rules are
a i t + 1 = a i t ( A i t + 1 ) 2 2 ( Δ t + 1 ) 2 ln Ω t ( Δ t + 1 ) ,
h i t + 1 = h i t + A i t + 1 Δ t + 1 ln Ω t ( Δ t + 1 ) m i t ( A i t + 1 ) 2 2 ( Δ t + 1 ) 2 ln Ω t ( Δ t + 1 ) ,
where Ω t ( Δ t + 1 ) = D z P ( y t + 1 | Δ t + 1 + χ t + 1 z ) and D z = ( d z / 2 π ) exp ( z 2 / 2 ) is the standard Gaussian measure. It has been shown [24] that this online algorithm has an asymptotic error decay comparable to the offline reconstruction scheme when additive measurement noise is present. In the noiseless scenario, the offline Bayesian algorithm achieves zero error for finite t, while the online version decays exponentially.

4. Mismatched Priors and L1-Minimization Based Reconstruction

A limitation of the framework introduced above is the simple fact that, in a real-world scenario, one does not know the exact parameters of the generating distribution, and possibly not even some of its major characteristics except for the fact that the generated signals are somewhat sparse. Using this algorithm as a canvas, we propose a strategy based on a Bayesian formulation of the L1-regularized problem, relying on the use of the sparsity-inducing Laplace prior
ϕ λ ( x ) = ( 2 / λ ) exp ( λ | x | )
and a learning scheme for λ concurrent with the measurements. The use of Laplace-like priors is not new in the literature [29,30,31], but, to the best of our knowledge, this has not been introduced as part of an online algorithm yet.
We observe that the L1-estimator (1) can also be written in a Bayesian manner as x ^ t = d x x lim β P L 1 ( x | D t ; β , λ ) , where
P L 1 ( x | D t ; β , λ ) = Z β , λ 1 exp β 1 2 σ n 2 μ = 1 t ( y μ A μ · x ) 2 + λ x 1 ,
with Z β , λ a normalization factor dependent on β and λ . This formulation suggests the replacement of the generating distribution ϕ ( x ) in Label (6) for ϕ λ ( x ) , effectively altering the projection step of the Bayesian online algorithm to the similar expression
P ˜ λ ( x | D t ) i = 1 N e a i t x i 2 / 2 + h i t x i λ | x i | Z ( a i t , h i t ; λ ) ,
where Z ( a , h ; λ ) : = d x exp a x 2 / 2 + h x λ | x | . The first and second moments of this alternate approximate distribution are, respectively,
m i t = ( / h i t ) ln Z ( a i t , h i t ; λ ) ,
v i t = ( 2 / ( h i t ) 2 ) ln Z ( a i t , h i t ; λ ) ,
paralleling Equations (8) and (9). Note that the natural parameters update Equations (10) and (11) remain unchanged, since they are independent of our choice of prior, be it ϕ ( x ) , ϕ λ ( x ) , or any other.
A common practical issue when making use of the L1-regularization scheme (1) is that the best value of the shrinkage parameter λ is data-dependent and not always clear. Nevertheless, its value is highly influential on the reconstructed model—there is a direct relation between λ and the number of non-zero coefficients in the estimate, which is reflected on the prediction accuracy of the model [32]. Therefore, since it is crucial to estimate accurate values of the shrinkage parameter, usual strategies in offline scenarios include calculation of its best value through cross-validation procedures [25] or straight calculation of an ideal number of non-zero covariates by adding one at a time in a recursive manner [32]. Unfortunately, these strategies are of little use in our present setting, where each measurement has to be used only once in an online manner.
Given any two distributions q ( x ) and r ( x ) , the Kullback–Leibler divergence [33], D K L ( q | | r ) = d x q ( x ) ln [ q ( x ) / r ( x ) ] is a common measure for their dissimilarity. Also called relative entropy, it has the interesting property that D K L ( q | | r ) 0 always and it is zero if and only if q = r . Now, whenever distribution r ( x ) is of the exponential family, i.e., f ( x ) k ξ k s k ( x ) , minimization of D K L ( q | | r ) with respect to ξ k is equivalent to matching the moments d x q ( x ) s k ( x ) and d x r ( x ) s k ( x ) . In particular, for the exact posterior P as defined in the left-hand side of Label (2), minimization of D K L ( P | | P ˜ λ ) with respect to λ after measurement t, i.e., finding λ t such that
λ t = arg min λ D K L ( P | | P ˜ λ )
is the same as solving the implicit equation
| x | P = | x | P ˜ λ
for λ . Here, | x | P is the best possible estimate for the true sparsity of the original signal, Q 0 d x | x | ϕ ( x ) . Our proposal is that λ t is, at any instant t, an optimal estimate for the shrinkage parameter. Moreover, it is not static. With the natural parameters { ( a i t , h i t ) } evolving, λ t must change as well to maintain | x | P ˜ λ Q 0 .
Alas, the left-hand side of expression (18) is unknown. To overcome this difficulty, we make the approximation | x | P | m t | . An argument can be made in its defense: as more and more measurements are acquired and the approximate posterior distribution (14) concentrates around the true values, two limits are obtained: (a) | x | P ˜ λ t | x | P and (b) | x | P ˜ λ t | x P ˜ λ | . The first limit has been proved in the large signal size limit N in [24] for an exact prior; here, we assume its validity also in the present scenario. As for the second limit, Minkowski inequality guarantees that | x | P ˜ λ | m t | ; with the collapse of the distribution P ˜ i ( x i ) around the true value x i 0 at t , no probability mass crosses the x i = 0 axis, so that the equality limit is met. Therefore, these results mean a good estimate for the solution of (18) is given by λ ^ t that solves
| x | P ˜ λ | m t |
for λ , where the right-hand side can be directly calculated from P ˜ λ t 1 ( x | D t ) . This implicit expression corresponds to an extra step in the online algorithm for the recovery of the signal, which is shown in full below (Algorithm 1).
Algorithm 1 Online L 1 -based signal recovery for CS.
1: Initialize P λ ( x ) a i 0 , h i 0 = 0 ( i ) ; λ 0 = λ
2: while t < t max do
3:   Obtain new measurement y t P ( y t | A t · x 0 )
4:   Update { ( a i t , h i t ) } ▹ Equations (10) and (11)
5:   Find λ ^ t ▹ Equation (19)
6:   Update λ t = ( 1 γ ) λ t 1 + γ λ ^ t
7:   Estimate signal means m i t and variances v i t ▹ Equations (15) and (16)
8: end while
9: return m t

5. Results and Discussion

We tested the algorithm on sparse signals generated by two different distributions of the form ϕ ( x ) = i = 1 N ϕ ( x i ) , with ϕ ( x i ) = ( 1 ρ ) δ ( x i ) + ρ g ( x i ) . The δ ( x ) factor induces sparsity by making every component equal to zero with probability 1 ρ . The non-sparse components could either come from a Gaussian distribution g ( x ) = ( 1 / 2 π ) exp ( x 2 / 2 ) or from a binary distribution g ( x ) = ( δ ( x + 1 ) + δ ( x 1 ) ) / 2 . To better evaluate the accuracy losses incurred due to approximation (19), we also tested the algorithm for the ideal situation where Q 0 is known. To guarantee practical usefulness, all results correspond to the standard CS scenario y μ = u μ + ξ μ with Gaussian noise ξ μ N ( 0 , σ n 2 ) , i.e., P ( y | u ) = ( 1 / 2 π σ n 2 ) exp [ ( y u ) 2 / 2 σ n 2 ] . Its noiseless counterpart can be obtained by taking the limit σ n 2 0 , which leads to P ( y | u ) = δ ( y u ) . Figure 3 shows the average mean squared error normalized by x 0 2 2 resulting of simulations of the L1-based online CS algorithm with λ learning. Simulations for finite signal length size N showed decreasing error for increasing N—in order to correct for finite size effects, the curves in Figure 3 were obtained by extrapolating the results for finite N = 200 , 500 , 1000 and 2000 to N by means of a quadratic fit. As expected, approximation (19) induces a significant accuracy loss, but, for α 1 , the reconstruction error decays approximately exponentially, proving the usefulness of the reconstruction scheme.
In all the examples shown, the solution to Equation (19) was obtained by numerically searching the value λ ^ t which best approximates both sides, i.e., by finding λ ^ t = arg min λ [ | x | P ˜ λ | m t | ] 2 . The empirical results prove what could have been a reasonable assumption: Figure 4 shows that in the early stages when only a few measurements have been obtained, even for an exact prior, | m t | is not an excellent approximation of | x | P . It also shows that the approximation gets progressively better with more measurements. For this reason, we found that the use of a small damping factor γ in the update of λ provides faster and more accurate learning.
Finding an optimal value for the sparsity-inducing parameter λ is not trivial. Ideally, one desires to find λ such that the P ˜ λ ( x ) would represent the actual sparsity of the original signal | x i 0 | . The limit λ is undesirable, since the soft constraints in P ˜ λ ( x ) due to the finite value of the natural parameters { ( a i t , h i t ) } would not be enough to prevent the prior distribution to completely dominate the entire approximate posterior, leading all signal estimates to zero. At the same time, consider as a representative example the update Equations (10) and (11) for the noisy standard CS scenario, which can explicitly written as a i t + 1 = a i t + ( A i t + 1 ) 2 / ( σ n 2 + χ t + 1 ) and h i t + 1 = h i t + A i t + 1 ( y t + 1 Δ t + 1 ) / ( σ n 2 + χ t + 1 ) + m i ( A i t + 1 ) 2 / ( σ n 2 + χ t + 1 ) . Taking the average of d a i μ = a i μ a i μ 1 and d h i μ = h i μ h i μ 1 over the measurement vectors A μ results in
d a i μ A μ N 1 / ( σ n 2 + χ μ ) , d h i μ A μ N 1 m i / ( σ n 2 + χ μ ) ,
which means that all the natural parameters a i t = μ = 1 t d a i μ typically grow with more and more measurements. Similarly, whenever x i 0 is different from zero, any parameters h i t = μ = 1 t d h i μ exhibit a growth in absolute value lead by the m i factor. For any fixed λ , these considerations mean that the regularizing effect of the prior ϕ λ ( x ) in Label (14) would vanish asymptotically in what would be a typical case of a prior being dominated by the likelihood. It is noteworthy that the introduction of a step for the minimization of the KL-divergence in the Bayesian online algorithm results in estimates λ ^ t that typically increase with α as well, in a scale similar to the other natural parameters (Figure 5).

6. Conclusions

In this paper, we proposed an online algorithm for Compressed Sensing. Based on an earlier version that expected exact knowledge of the signal generating distribution [24], the present adaptation made use of a double-exponential prior (12) to get rid of this limitation. In order to keep the signal sparsity always in focus during the signal reconstruction, we introduced in this work an extra step where an optimal value for the sparsity-inducing parameter λ can be estimated as a function of { ( a i t , h i t ) } (i.e., a function of the data D t ). This was achieved through minimization of the KL-divergence between P ˜ λ ( x ) and the (unavailable) full posterior distribution (2), which is approximated by the signal estimate at the previous step.
It is clear that the method is strongly dependent on the validity of the mean-field approximation (6) for the posterior to obtain good reconstruction accuracy. This approximation was introduced in order to simplify the calculations of the moments and to reduce the number of parameters, which should be updated at every step (and, as a consequence, to reduce computational effort and memory requirements). Nevertheless, it is not essential and can be easily replaced by a full Gaussian distribution if needed. In all examples shown here, which made use of signals with fully independent components, the algorithm was shown to give accurate predictions for different signal generating distributions, with and without the presence of additive measurement noise.

Acknowledgments

Paulo V. Rossi would like to thank Yoshiyuki Kabashima for all the discussions that originated this work and Yoshiyuki Kabashima, Andre Manoel and Rafael Calsaverini for all the talks during its production. Renato Vicente would like to thank Julio Stern for many insightful interactions and Paulo V. Rossi and Renato Vicente would like to thank the reviewers for their constructive feedback. This work was partially funded by FAPESP 2014/00792-7 (Paulo V. Rossi). Renato Vicente and Paulo V. Rossi acknowledge the Experian Latam DataLab for supporting this work.

Author Contributions

Paulo V. Rossi and Renato Vicente conceived the algorithm; Paulo V. Rossi performed the simulations; Paulo V. Rossi and Renato Vicente wrote the paper. Both authors have read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Strohmer, T. Measure What Should be Measured: Progress and Challenges in Compressive Sensing. IEEE Signal Process. Lett. 2012, 19, 887–893. [Google Scholar] [CrossRef]
  2. Candès, E.J.; Wakin, M.B. An Introduction to Compressive Sampling. IEEE Signal Process. Mag. 2008, 25, 21–30. [Google Scholar] [CrossRef]
  3. Holtz, O. Compressive sensing: A paradigm shift in signal processing. arXiv, 2008; arXiv:0812.3137. [Google Scholar]
  4. Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  5. Eldar, Y.C.; Kutyniok, G. Compressed Sensing: Theory and Applications; Cambridge University Press: Cambridge, UK, 2012. [Google Scholar]
  6. Nyquist, H. Certain topics in telegraph transmission theory. Trans. AIEE 1928, 47, 617–644. [Google Scholar] [CrossRef]
  7. Shannon, C. Communication in the presence of noise. Proc. Inst. Radio Eng. 1949, 37, 10–21. [Google Scholar] [CrossRef]
  8. Candès, E.; Tao, T. Near Optimal Signal Recovery from Random Projections: Universal Encoding Strategies? IEEE Trans. Inf. Theory 2006, 52, 5406–5425. [Google Scholar] [CrossRef]
  9. Candès, E.J.; Romberg, J.; Tao, T. Robust Uncertainty Principles: Exact Signal Reconstruction from Highly Incomplete Frequency Information. IEEE Trans. Inf. Theory 2006, 52, 489–509. [Google Scholar] [CrossRef]
  10. Rangan, S. Generalized approximate message passing for estimation with random linear mixing. In Proceedings of the 2011 IEEE International Symposium on Information Theory Proceedings (ISIT), St. Petersburg, Russia, 31 July–5 August 2011. [Google Scholar]
  11. Krzakala, F.; Mezard, M.; Sausset, F.; Zdeborova, L. Statistical-Physics-Based Reconstruction in Compressed Sensing. Phys. Rev. X 2012, 2, 021005. [Google Scholar] [CrossRef]
  12. Tramel, E.W.; Manoel, A.; Caltagirone, F.; Gabrié, M.; Krzakala, F. Inferring sparsity: Compressed sensing using generalized restricted Boltzmann machines. In Proceedings of the 2016 IEEE Information Theory Workshop (ITW), Cambridge, UK, 11–14 September 2016; pp. 265–269. [Google Scholar]
  13. Krzakala, F.; Mezard, M.; Sausset, F.; Sun, Y.; Zdeborova, L. Probabilistic reconstruction in compressed sensing: Algorithms, phase diagrams, and threshold achieving matrices. J. Stat. Mech. Theory Exp. 2012, P08009. [Google Scholar] [CrossRef]
  14. Xu, Y.; Kabashima, Y.; Zdeborova, L. Bayesian signal reconstruction for 1-bit Compressed Sensing. J. Stat. Mech. Theory Exp. 2014. [Google Scholar] [CrossRef]
  15. Rangan, S.; Fletcher, A.K.; Goyal, V.K. Asymptotic Analysis of MAP Estimation via the Replica Method and Applications to Compressed Sensing. IEEE Trans. Inf. Theory 2012, 58, 1902–1923. [Google Scholar] [CrossRef] [Green Version]
  16. Opper, M.; Winther, O. Chapter A Bayesian approach to on-line learning. In On-Line Learning in Neural Networks; Cambridge University Press: Cambridge, UK, 1998; pp. 363–378. [Google Scholar]
  17. De Oliveira, E.A.; Caticha, N. Inference from aging information. IEEE Trans. Neural Netw. 2010, 21, 1015–1020. [Google Scholar] [CrossRef] [PubMed]
  18. Vicente, R.; Kinouchi, O.; Caticha, N. Statistical mechanics of online learning of drifting concepts: A variational approach. Mach. Learn. 1998, 32, 179–201. [Google Scholar] [CrossRef]
  19. Kalman, R.E. A new approach to linear filtering and prediction problems. J. Basic Eng. 1960, 82, 35–45. [Google Scholar] [CrossRef]
  20. Stengel, R.F. Optimal Control and Estimation; Courier Corporation: North Chelmsford, MA, USA, 2012. [Google Scholar]
  21. Särkkä, S. Bayesian Filtering and Smoothing; Cambridge University Press: Cambridge, UK, 2013; Volume 3. [Google Scholar]
  22. Broderick, T.; Boyd, N.; Wibisono, A.; Wilson, A.C.; Jordan, M.I. Streaming variational bayes. In Advances in Neural Information Processing Systems; Curran: Red Hook, NY, USA, 2013; pp. 1727–1735. [Google Scholar]
  23. Manoel, A.; Krzakala, F.; Tramel, E.W.; Zdeborová, L. Streaming Bayesian inference: Theoretical limits and mini-batch approximate message-passing. arXiv, 2017; arXiv:1706.00705. [Google Scholar]
  24. Rossi, P.V.; Kabashima, Y.; Inoue, J. Bayesian online compressed sensing. Phys. Rev. E 2016, 94, 022137. [Google Scholar] [CrossRef] [PubMed]
  25. Tibshirani, R. Regression Shrinkage and Selection via the Lasso. J. R. Stat. Soc. Ser. B (Methodological) 1996, 58, 267–288. [Google Scholar]
  26. Ganguli, S.; Sompolinsky, H. Statistical Mechanics of Compressed Sensing. Phys. Rev. Lett. 2010, 104, 188701. [Google Scholar] [CrossRef] [PubMed]
  27. Kabashima, Y.; Wadayama, T.; Tanaka, T. A typical reconstruction limit for compressed sensing based on Lp-norm minimization. J. Stat. Mech. Theory Exp. 2009, 9, L09003. [Google Scholar]
  28. Baron, D.; Sarvotham, S.; Baraniuk, R.G. Bayesian Compressive Sensing Via Belief Propagation. IEEE Trans. Signal Process. 2010, 58, 269–280. [Google Scholar] [CrossRef]
  29. Hans, C. Bayesian lasso regression. Biometrika 2009, 96, 835–845. [Google Scholar] [CrossRef]
  30. Park, T.; Casella, G. The bayesian lasso. J. Am. Stat. Assoc. 2008, 103, 681–686. [Google Scholar] [CrossRef]
  31. Figueiredo, M.A. Adaptive sparseness for supervised learning. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 1150–1159. [Google Scholar] [CrossRef]
  32. Efron, B.; Hastie, T.; Johnstone, I.; Tibshirani, R. Least angle regression. Ann. Stat. 2004, 32, 407–499. [Google Scholar]
  33. Barber, D. Bayesian Reasoning and Machine Learning; Cambridge University Press: Cambridge, UK, 2012. [Google Scholar]
Figure 1. Schematic representation of the measurement process in Compressed Sensing. The sparse signal is projected onto a (known) random measurement vector A . The result u = A · x goes through a channel giving rise to the value y P ( y | u ) .
Figure 1. Schematic representation of the measurement process in Compressed Sensing. The sparse signal is projected onto a (known) random measurement vector A . The result u = A · x goes through a channel giving rise to the value y P ( y | u ) .
Entropy 19 00667 g001
Figure 2. Update and projection steps.
Figure 2. Update and projection steps.
Entropy 19 00667 g002
Figure 3. L1-based reconstruction for CS with ρ = 0.1 for signals generated by ϕ ( x ) = i = 1 N [ ( 1 ρ ) δ ( x i ) + ρ g ( x i ) ] . (a) and (b): g ( x ) = ( 1 / 2 π ) exp ( x 2 / 2 ) ; (c) and (d): g ( x ) = ( δ ( x + 1 ) + δ ( x 1 ) ) / 2 . The top row consists of the noiseless standard CS scenario; the bottom row corresponds to the noisy scenario with σ n 2 = 10 4 . In all figures, dashed lines were obtained with known Q 0 and full lines with Q 0 estimated through (19). In (a), the average error of 100 simulations each of N = 250 , 500 , 1000 and 2000 for chosen values of α are shown as crosses (unknown Q 0 ) and triangles (known Q 0 ). The error decreases with increasing N. All lines in all pictures then correspond to the extrapolation of these finite N values to N by means of a quadratic fit. All pictures: γ = 10 3 .
Figure 3. L1-based reconstruction for CS with ρ = 0.1 for signals generated by ϕ ( x ) = i = 1 N [ ( 1 ρ ) δ ( x i ) + ρ g ( x i ) ] . (a) and (b): g ( x ) = ( 1 / 2 π ) exp ( x 2 / 2 ) ; (c) and (d): g ( x ) = ( δ ( x + 1 ) + δ ( x 1 ) ) / 2 . The top row consists of the noiseless standard CS scenario; the bottom row corresponds to the noisy scenario with σ n 2 = 10 4 . In all figures, dashed lines were obtained with known Q 0 and full lines with Q 0 estimated through (19). In (a), the average error of 100 simulations each of N = 250 , 500 , 1000 and 2000 for chosen values of α are shown as crosses (unknown Q 0 ) and triangles (known Q 0 ). The error decreases with increasing N. All lines in all pictures then correspond to the extrapolation of these finite N values to N by means of a quadratic fit. All pictures: γ = 10 3 .
Entropy 19 00667 g003
Figure 4. Difference between | x | P ˜ (from Label (6), where the exact prior is considered) and the true signal sparsity | x i 0 | | . The full line corresponds to the average of 100 simulations of the noisy standard CS scenario with N = 1000 , ρ = 0.2 , σ n 2 = 10 4 and g ( x ) = ( 1 / 2 π ) exp ( x 2 / 2 ) . The dashed line is the true sparsity of the original signal Q 0 = d x 0 | x 0 | ϕ ( x 0 ) . Inset: Mean variance. Notice that the absolute value approximation gets progressively better with growing α . Indeed, its accuracy matches the diminishing of the posterior distribution’s variance.
Figure 4. Difference between | x | P ˜ (from Label (6), where the exact prior is considered) and the true signal sparsity | x i 0 | | . The full line corresponds to the average of 100 simulations of the noisy standard CS scenario with N = 1000 , ρ = 0.2 , σ n 2 = 10 4 and g ( x ) = ( 1 / 2 π ) exp ( x 2 / 2 ) . The dashed line is the true sparsity of the original signal Q 0 = d x 0 | x 0 | ϕ ( x 0 ) . Inset: Mean variance. Notice that the absolute value approximation gets progressively better with growing α . Indeed, its accuracy matches the diminishing of the posterior distribution’s variance.
Entropy 19 00667 g004
Figure 5. An example of the online CS algorithm where λ has been adjusted after all measurements so that the inferred signal sparsity was equal to the known value Q 0 . Note that λ t , just like the natural parameters { ( a i t , h i t ) } , typically grows exponentially. Here, N = 1000 , ρ = 0.1 , σ n 2 = 0 and g ( x ) = ( 1 / 2 π ) exp ( x 2 / 2 ) .
Figure 5. An example of the online CS algorithm where λ has been adjusted after all measurements so that the inferred signal sparsity was equal to the known value Q 0 . Note that λ t , just like the natural parameters { ( a i t , h i t ) } , typically grows exponentially. Here, N = 1000 , ρ = 0.1 , σ n 2 = 0 and g ( x ) = ( 1 / 2 π ) exp ( x 2 / 2 ) .
Entropy 19 00667 g005

Share and Cite

MDPI and ACS Style

Rossi, P.V.; Vicente, R. L1-Minimization Algorithm for Bayesian Online Compressed Sensing. Entropy 2017, 19, 667. https://doi.org/10.3390/e19120667

AMA Style

Rossi PV, Vicente R. L1-Minimization Algorithm for Bayesian Online Compressed Sensing. Entropy. 2017; 19(12):667. https://doi.org/10.3390/e19120667

Chicago/Turabian Style

Rossi, Paulo V., and Renato Vicente. 2017. "L1-Minimization Algorithm for Bayesian Online Compressed Sensing" Entropy 19, no. 12: 667. https://doi.org/10.3390/e19120667

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop