Abstract
Compressed sensing is a new sampling technique which can exactly reconstruct sparse signal from a few measurements. In this article, we consider the block-sparse compressed sensing with special structure assumption about the signal. A novel non-convex model is proposed to reconstruct the block-sparse signals. In addition, the conditions of the proposed model for recovering the block-sparse noise or noise-free signals are presented. The experimental results demonstrate that the proposed non-convex method surpasses the convex method (the mixed -norm optimization) and some algorithms without considering the block-sparse structure (the
- and
-norm optimization).
1. Introduction
The traditional signal acquisition must follow the Nyquist–Shannon sampling theorem. Based on this theorem, the sample rate is at least twice of the signal bandwidth for exact reconstruction. In many recent applications, the Nyquist rate may result in high storage and transmission pressure. In addition, the traditional signal acquisition model does not consider the signal structure known as the prior information. A breakthrough in data acquisition, namely compressed sensing, was brought by Donoho Citation1 and Candès et al. Citation2,Citation3, which has been widely applied in magnetic resonance imaging Citation4 and radar imaging Citation5. Compressed sensing assumes that the original signal is sparse or approximately sparse under some bases. Without loss of generality, we assume that the original signal is sparse under the canonical basis. Then, the -sparse signal means that there are no more than
nonzero entries in the signal. Let
be the
-sparse signal. The task of compressed sensing is to recover
from the measurements
calculated as
(1)
Here, is the measurement matrix with more columns than rows. Obviously, Equation (1) is underdetermined. However, the theory of compressed sensing indicates that the sparse or compressible signal can be exactly reconstructed from a few measurements
through the following optimization problem:
(2)
where
denotes the
-norm counting the number of nonzero entries. However, the problem (2) is a combinatorial optimization problem. Solving it directly is general NP-hard. Therefore, some sub-optimization strategies are proposed to approximate the exact solution.
Popular sub-optimization strategies include greedy algorithms (such as the orthogonal matching pursuit Citation6), and relaxation methods (such as convex and non-convex relaxation methods Citation7–9). The representative convex method is the -optimization problem which is called basis pursuit (BP) Citation7
(3)
where
is the
-norm of the vector
. In the noise case, the problem (3) can be modified as the basis pursuit denoising (BPDN) Citation7
(4)
where
is a tradeoff parameter. The non-convex relaxation method has been studied in Citation8,Citation9, i.e. the following non-convex optimization problem:
(5)
where
. In Citation8, the experimental results show that the non-convex problem (5) can reconstruct the sparse signal from fewer measurements than the convex method acquiring. The restricted isometry property (RIP) Citation3 plays a crucial role in the theory of compressed sensing. The previous results show that if the measurement matrix has the RIP with a certain parameters, the convex and non-convex optimization methods can recover the sparse signal exactly.
The existed analysis of compressed sensing does not consider the special inherent structures of signals. This article considers a variation of the compressed sensing problem named the block-sparse compressed sensing. In block-sparse compressed sensing, the block structure of sparse signal is investigated, i.e. the nonzero entries of the signal occur in clusters. Various practical signals exhibit the block structure, such as DNA microarrays Citation10, multiband signal Citation11 and magnetoencephalography (MEG) Citation12. The framework of block-sparse compressed sensing was introduced and studied first in refs Citation13–15. The mixed -norm optimization is developed to recover the block-sparse signal in Citation13, which is a convex method. In this article, we propose a novel non-convex model termed as the mixed
-norm optimization for block-sparse signal recovery, which can greatly improve the performance of the mixed
-norm optimization.
The remainder of this article is organized as follows. In Section 2, we briefly review the block-sparse compressed sensing. In Section 3, the mixed -norm optimization problem is proposed, and the stability and uniqueness of the proposed model is investigated. In Section 4, the re-weighted
-algorithm is used to solve mixed
-norm optimization problem, and some numerical experiments are reported. Finally, Section 5 concludes this article.
2. Block-sparse compressed sensing and RIP
In this section, we briefly review block sparsity and the block-sparse compressed sensing.
In the block-sparse compressed sensing, the signal is the concatenation of some blocks. Without loss of generality, we assume that there are n blocks with the block size
in x. Then, the signal x can be rewritten as
(6)
where
,
. Then, the signal x is called K-block-sparse if no more than K blocks over
are nonzero. In Citation13, a convex optimization problem termed as the mixed
-norm optimization is proposed to reconstruct the block-sparse signal, i.e. the following model:
(7)
where
. In noisy measurements case, the measurements can be calculated as
(8)
where
is unknown noise. In this article, we assume
, and the standard deviation
denotes the noise level. In this noise case, the block-sparse signal can be reconstructed via the following optimization problem:
(9)
To investigate the theory of compressed sensing, Candès and Tao Citation3 introduce the RIP. The matrix is said to have the RIP with parameter
, if the inequality
(10)
holds for every K-sparse signal x. Many random matrices have been proven to have the RIP with overwhelming probability, such as Gaussian, Bernoulli and partial Fourier matrices Citation16. To study the mixed
-norm optimization problems (7) and (9), Eldar and Mishali Citation13 introduce the block RIP (B-RIP) which is the block-sparse version of RIP. The matrix
has the B-RIP with parameter
, if the inequality
(11)
holds for every K-block-sparse signal x, where
is the B-RIP constant over
. In Citation13 Eldar and Mishali show that the Gaussian random matrix has the B-RIP with high probability.
3. Non-convex recovery model
In this section, the motivation of proposed non-convex method is stated first. Then, the theoretical results about the uniqueness and robust of the proposed non-convex model to recover the block-sparse noised and noise-free signals are studied. At last, an iterative re-weighted algorithm is presented to solve the non-convex method.
3.1. Motivation
In Citation13, Eldar and Mishali claim that the block-sparse signal x can be exactly reconstructed from the measurements by the following optimization problem:
(12)
where
By introducing auxiliary variables
and
, the objective function of problem (12) can be equivalently transformed as
. It is most accurate to measure the sparsity by the
-norm. Unfortunately, the
-norm optimization problem is combinatorial optimization, which cannot be solved directly. An effective relaxation strategy of the
-norm to measure the sparsity is
-norm, i.e.
. illustrates the one-dimensional trend of
for varying p. Some previous results have demonstrated that the
-norm can approximate the
-norm effectively Citation3,Citation13. However, with
, the
-norm tends to the
-norm. Obviously, the non-convex
-norm can give more powerful sparsity measure than the
-norm. Therefore, we focus on non-convex method in our work. Corresponding to the block-sparse signal x, we study the following non-convex method termed as the mixed
-norm optimization:
(13)
where
3.2. Exact recovery
In this section, we show that the proposed non-convex model (13) can recover the block-sparse signal x exactly in the noise-free case with a certain condition of the measurement matrix .
Theorem 1
Let be the measurements of the K-block-sparse signal x. If the measurement matrix
has the B-RIP with the condition
(14)
where
, then x is the unique solution of the problem (13).
3.3. Robust recovery
This section considers the noise case. Assume x be an arbitrary signal, and the measurements are computed with noise, i.e.
(15)
where
is the Gaussian noise. In this case, the optimization problem (13) must be modified as
(16)
The following Theorems 2 and 3 state the robustness of the proposed non-convex model. In addition, we prove that the reconstruction error of problem (16) is the linear function of the noise energy .
First, the case of the block-sparse signal with noise is studied.
Theorem 2
Let be the noisy measurements of the K-block-sparse signal x. If the measurement matrix
has the B-RIP and the below inequality
(17)
holds, where
. Assume
be the solution of the problem (16). Then,
where
Afterwards, let x and be an arbitrary signal and the K-block-sparse approximation of x, respectively. Assume T0 be the index set corresponding to the first K largest blocks in terms of
-norm. Then, we have the following theory for general case.
Theorem 3
Let be the noisy measurements of any signal x. If the condition (17) holds, then the solution
of problem (16) obeys
where
Proof
Following the method of proof stated in Citation9, let be the difference between the true signal and the reconstructed signal. The set
can be divided as
, where
is the indices set which is make up of
blocks, chosen such that the
-norm of
over
is largest, and the
-norm of
over
is the second largest, and so on. So, we can decompose h as
.
Based on the B-RIP and the triangle inequality of , we obtain the lower bound of
as follows:
(18)
Next, we will give the bounds of and
.
For all ,
holds. Following this inequality, we have
Thus, we obtain
Due to ,
Therefore,
Applying the Holder inequality, the is bounded as the
-norm, i.e.
Thus, we have
(19)
Combining (18) and (19), the following inequality holds
where
Furthermore, has the upper bound as
Then, we have . That is, the upper bound of
is
(20)
By the condition of (17), we know that . Hence, the inequality (20) holds.
Let be the
th largest block of
in
-norm. The below inequality
can be obtained easily. For
, the
is bounded as
Finally, we have
▪
Remark 1
Theorems 1 and 2 are the special cases of Theorem 3. When the original signal x is a K-block-sparse signal, then yields Theorem 2. Furthermore, Theorem 1 is the K-block-sparse signal and noise-free case of Theorem 3.
3.4. The iterative re-weighted
-algorithm
In this section, we propose the re-weighted -algorithm for solving the
-norm optimization deriving from the re-weighted
-algorithm Citation17. Due to
the main steps of the iterative re-weighted
-algorithm are summarized in Algorithm 1.
Algorithm 1
The iterative re-weighted -algorithm
Input: | |||||
For i=1: Num | |||||
While | |||||
(1) Solve the weighted |
(2) Update the weights |
End while |
End for | |||||
Output: The block-sparse signal x. |
The weighted -norm optimization
can be equivalently transformed to a second-order cone programming (SOCP):
Furthermore, the ‘cvx’ software packages (the code is available at http://cvxr.com/cvx/) can be used to solve the above SOCP.
4. Experiments
In this section, some numerical experiments are presented to evaluate the performance of the mixed -norm optimization. First, the experimental settings are presented. Then, the performance of the mixed
-norm optimization is compared with three popular methods, including the mixed
-norm optimization,
-norm optimization and the
-norm optimization.
4.1. Experimental settings
Three numerical experiments are performed to assess the capacity of the mixed -norm optimization: (a) the noise-free case; (b) the noise case; (c) the large scale problem. Therefore, in this section, we describe the experiments setting, including the generation of the experiment data and the parameters setting.
We assume that the synthetic signal has n blocks. Then, the synthetic block-sparse signals are generated as follows: (1) the K blocks are selected from these n blocks randomly; (2) the entries restricted to these K blocks are generated from a standard Gaussian distribution and the entries out of the selected K blocks are set to zero. In our experiments, the Gaussian random matrix is used as the measurement matrix. For different experiment cases, we set the parameters of the re-weighted -algorithm listed in . In addition, the parameters for other tested algorithms are set according to refs Citation7,Citation8,Citation13. The relative error and probability of exact reconstruction are used as the evaluation indices. The relative error between the reconstructed signal
and the original signal x is calculated as
Table 1. The parameters of the re-weighted
-algorithm for three different experiments.
The signal is said to be exactly reconstructed, if the relative error between the reconstructed signal and the original signal x obeys
.
4.2. Experimental results
Firstly, we consider the noise-free case. The dimension of signal is , and the block size is
. The comparison experiments about recovering the block-sparse signal from the
measurements are performed. The active blocks vary from 1 to 18. For each block-sparsity level, the tested algorithms are all repeated 500 times. The probability of exact reconstruction is used to valuate the performance of the algorithms. The proposed method is compared with the mixed
-norm optimization, BP and the
-method. The exact reconstruction performances are presented in . The horizontal axis of denotes the block-sparsity level K, and the vertical axis is the probability of exact reconstruction. When the block-sparse level is greater than 6, the BP,
and mixed
-norm optimization methods cannot recover the signal exactly, but the proposed method can recover the signal exactly up to 9-block-sparse level. In addition, when the block-sparse level is greater than 11, the BP,
and mixed
-norm optimization methods cannot recover the signal, but the proposed non-convex method can also recover the signal exactly with high probability. Therefore, from , it can be seen that the proposed non-convex method is more accurate and effective than other three methods. Especially, the capacity of the mixed
-norm optimization is improved with the decreasing p.
Secondly, the case of noisy measurements is studied. The dimension of the simulation signal is , and the block number is
. The relative error is used to valuate the performance of the test algorithms from
measurements. The noise is assumed as i.i.d. Gaussian with zero mean and standard deviation
. The experimental results about three block-sparsity levels and four noise levels are listed in . The best results are labeled in bold. The experimental results indicate that the mixed
-norm optimization provides lower relative error than the
, BPDN and the mixed
-norm methods in the presence of noise. Especially, for the less block sparse
and
, when the noise level is small, the performance of mixed
-norm optimization is more efficient with a little smaller p. However, if the noise level is large, the reconstruction capacity decreases with small p. For block-sparsity level
, when p is around 0.5, the mixed
-norm optimization have better reconstruction performance. Especially, comparing with the convex method (the mixed
-norm optimization), the proposed method generates smaller reconstruction error. These experimental results demonstrate that the proposed method is more robust than the mixed
-norm optimization in the case of noisy measurements.
Table 2. The comparison results (
,
; average of 20 runs).
Lastly, the large-scale problem is studied. The length of the block-sparse signal is set to , the number of blocks is 64, and the block-sparsity level is fixed as
. illustrates the original true signal and the reconstructed signal of different algorithms from
measurements with the noise level
. The black diamond is the original true signal and the blue cross is the reconstructed signal. shows the results of
, BPDN,
and
, respectively. The relative errors are listed simultaneously. Obviously, the mixed
-norm optimization generates the lowest relative error, which shows the effectiveness of proposed non-convex method for large scale block-sparse signal recovery.
5. Conclusion
In this article, we propose a non-convex model for block-sparse signals recovery termed as the mixed -norm optimization. Based on the B-RIP, we prove that this non-convex method can recover the block-sparse signals effectively in noise-free and noise cases. The re-weighted
-algorithm is developed to solve the proposed non-convex optimization problem. The experimental results show that the performance of the mixed
-norm optimization surpasses the mixed
-norm optimization and some reconstruction algorithms without considering the block structure. Our future works will focus on the applications of block-sparse compressed sensing including the DNA microarrays, multiband signals and MEG.
Acknowledgements
This article is supported by the National Natural Science Foundation of China (no. 61172161), the Fundamental Research Funds for the Central Universities, Hunan University, the Scholarship Award for Excellent Doctoral Student granted by Chinese Ministry of Education, the Independent Research Funds of the State Key Laboratory of Advanced Design and Manufacturing for Vehicle Body (Hunan University, no. 71165002) and the Hunan Provincial Innovation Foundation For Postgraduate (no. CX2011B131).
References
- Donoho, D, 2006. Compressed sensing, IEEE Trans. Inform. Theory 52 (2006), pp. 1289–1306.
- Candès, E, Romberg, J, and Tao, T, 2006. Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information, IEEE Trans. Inform. Theory 52 (2006), pp. 489–509.
- Candès, E, and Tao, T, 2005. Decoding by linear programming, IEEE Trans. Inform. Theory 51 (2005), pp. 4203–4215.
- Qu, X, Zhang, W, Guo, D, Cai, C, Cai, S, and Chen, Z, 2010. Iterative thresholding compressed sensing MRI based on contourlet transform, Inverse Probl. Sci. Eng. 18 (2010), pp. 737–758.
- Potter, LC, Ertin, E, Parker, JT, and Cetin, M, 2010. Sparsity and compressed sensing in radar imaging, Proc. IEEE 98 (2010), pp. 1006–1020.
- Mallat, SG, and Zhang, Z, 1993. Matching pursuits with time-frequency dictionaries, IEEE Trans. Signal Process. 41 (1993), pp. 3397–3415.
- Chen, S, Donoho, D, and Saunders, M, 2001. Atomic decomposition by basis pursuit, SIAM Rev. Soc. Ind. Appl. Math. 43 (2001), pp. 129–159.
- Chartrand, R, 2007. Exact reconstruction of sparse signals via nonconvex minimization, IEEE Signal Process. Lett. 14 (2007), pp. 707–710.
- Chartrand, R, and Staneva, V, 2008. Restricted isometry properties and nonconvex compressive sensing, Inverse Probl. 24 (2008), pp. 1–14.
- Parvaresh, F, Vikalo, H, Misra, S, and Hassibi, B, 2008. Recovering sparse signals using sparse measurement matrices in compressed DNA microarrays, IEEE J. Sel. Top. Signal Process. 2 (2008), pp. 275–285.
- Mishali, M, and Eldar, YC, 2009. Blind multi-band signal reconstruction: Compressed sensing for analog signals, IEEE Trans. Signal Process. 57 (2009), pp. 993–1009.
- Phillips, JW, Leahy, RM, and Mosher, JC, 1997. Meg-based imaging of focal neuronal current sources, IEEE Trans. Med. Imaging 16 (1997), pp. 338–348.
- Eldar, YC, and Mishali, M, 2009. Robust recovery of signals from a structured union of subspaces, IEEE Trans. Inform. Theory 55 (2009), pp. 5302–5316.
- Baraniuk, RG, Cevher, V, Duarte, MF, and Hegde, C, 2010. Model-based compressive sensing, IEEE Trans. Inform. Theory 56 (2010), pp. 1982–2001.
- Eldar, YC, Kuppinger, P, and Bolcskei, H, 2010. Block-sparse signals: Uncertainty relations and efficient recovery, IEEE Trans. Signal Process. 58 (2010), pp. 3042–3054.
- Candès, E, Romberg, J, and Tao, T, 2006. Stable signal recovery from incomplete and inaccurate measurements, Commun. Pure Appl. Math. 59 (2006), pp. 1207–1223.
- Candès, E, Wakin, M, and Boyd, S, 2008. Enhancing sparsity by reweighted l1 minimization, J. Fourier Anal. Appl. 14 (2008), pp. 877–905.